mirror of
https://github.com/octoleo/restic.git
synced 2024-11-29 16:23:59 +00:00
Merge pull request #178 from restic/refactor-server
Rename Server -> Repository
This commit is contained in:
commit
b2dcdf00e3
36
archiver.go
36
archiver.go
@ -15,7 +15,7 @@ import (
|
|||||||
"github.com/restic/restic/debug"
|
"github.com/restic/restic/debug"
|
||||||
"github.com/restic/restic/pack"
|
"github.com/restic/restic/pack"
|
||||||
"github.com/restic/restic/pipe"
|
"github.com/restic/restic/pipe"
|
||||||
"github.com/restic/restic/server"
|
"github.com/restic/restic/repo"
|
||||||
|
|
||||||
"github.com/juju/errors"
|
"github.com/juju/errors"
|
||||||
)
|
)
|
||||||
@ -30,7 +30,7 @@ var archiverAllowAllFiles = func(string, os.FileInfo) bool { return true }
|
|||||||
|
|
||||||
// Archiver is used to backup a set of directories.
|
// Archiver is used to backup a set of directories.
|
||||||
type Archiver struct {
|
type Archiver struct {
|
||||||
s *server.Server
|
repo *repo.Repo
|
||||||
|
|
||||||
blobToken chan struct{}
|
blobToken chan struct{}
|
||||||
|
|
||||||
@ -39,9 +39,9 @@ type Archiver struct {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// NewArchiver returns a new archiver.
|
// NewArchiver returns a new archiver.
|
||||||
func NewArchiver(s *server.Server) *Archiver {
|
func NewArchiver(repo *repo.Repo) *Archiver {
|
||||||
arch := &Archiver{
|
arch := &Archiver{
|
||||||
s: s,
|
repo: repo,
|
||||||
blobToken: make(chan struct{}, maxConcurrentBlobs),
|
blobToken: make(chan struct{}, maxConcurrentBlobs),
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -55,18 +55,18 @@ func NewArchiver(s *server.Server) *Archiver {
|
|||||||
return arch
|
return arch
|
||||||
}
|
}
|
||||||
|
|
||||||
// Save stores a blob read from rd in the server.
|
// Save stores a blob read from rd in the repository.
|
||||||
func (arch *Archiver) Save(t pack.BlobType, id backend.ID, length uint, rd io.Reader) error {
|
func (arch *Archiver) Save(t pack.BlobType, id backend.ID, length uint, rd io.Reader) error {
|
||||||
debug.Log("Archiver.Save", "Save(%v, %v)\n", t, id.Str())
|
debug.Log("Archiver.Save", "Save(%v, %v)\n", t, id.Str())
|
||||||
|
|
||||||
// test if this blob is already known
|
// test if this blob is already known
|
||||||
if arch.s.Index().Has(id) {
|
if arch.repo.Index().Has(id) {
|
||||||
debug.Log("Archiver.Save", "(%v, %v) already saved\n", t, id.Str())
|
debug.Log("Archiver.Save", "(%v, %v) already saved\n", t, id.Str())
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// otherwise save blob
|
// otherwise save blob
|
||||||
err := arch.s.SaveFrom(t, id, length, rd)
|
err := arch.repo.SaveFrom(t, id, length, rd)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
debug.Log("Archiver.Save", "Save(%v, %v): error %v\n", t, id.Str(), err)
|
debug.Log("Archiver.Save", "Save(%v, %v): error %v\n", t, id.Str(), err)
|
||||||
return err
|
return err
|
||||||
@ -76,7 +76,7 @@ func (arch *Archiver) Save(t pack.BlobType, id backend.ID, length uint, rd io.Re
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// SaveTreeJSON stores a tree in the server.
|
// SaveTreeJSON stores a tree in the repository.
|
||||||
func (arch *Archiver) SaveTreeJSON(item interface{}) (backend.ID, error) {
|
func (arch *Archiver) SaveTreeJSON(item interface{}) (backend.ID, error) {
|
||||||
data, err := json.Marshal(item)
|
data, err := json.Marshal(item)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@ -86,11 +86,11 @@ func (arch *Archiver) SaveTreeJSON(item interface{}) (backend.ID, error) {
|
|||||||
|
|
||||||
// check if tree has been saved before
|
// check if tree has been saved before
|
||||||
id := backend.Hash(data)
|
id := backend.Hash(data)
|
||||||
if arch.s.Index().Has(id) {
|
if arch.repo.Index().Has(id) {
|
||||||
return id, nil
|
return id, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
return arch.s.SaveJSON(pack.Tree, item)
|
return arch.repo.SaveJSON(pack.Tree, item)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (arch *Archiver) reloadFileIfChanged(node *Node, file *os.File) (*Node, error) {
|
func (arch *Archiver) reloadFileIfChanged(node *Node, file *os.File) (*Node, error) {
|
||||||
@ -184,7 +184,7 @@ func (arch *Archiver) SaveFile(p *Progress, node *Node) error {
|
|||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
chnker := chunker.New(file, arch.s.Config.ChunkerPolynomial, sha256.New())
|
chnker := chunker.New(file, arch.repo.Config.ChunkerPolynomial, sha256.New())
|
||||||
resultChannels := [](<-chan saveResult){}
|
resultChannels := [](<-chan saveResult){}
|
||||||
|
|
||||||
for {
|
for {
|
||||||
@ -254,7 +254,7 @@ func (arch *Archiver) fileWorker(wg *sync.WaitGroup, p *Progress, done <-chan st
|
|||||||
// check if all content is still available in the repository
|
// check if all content is still available in the repository
|
||||||
contentMissing := false
|
contentMissing := false
|
||||||
for _, blob := range oldNode.blobs {
|
for _, blob := range oldNode.blobs {
|
||||||
if ok, err := arch.s.Test(backend.Data, blob.Storage.String()); !ok || err != nil {
|
if ok, err := arch.repo.Test(backend.Data, blob.Storage.String()); !ok || err != nil {
|
||||||
debug.Log("Archiver.fileWorker", " %v not using old data, %v (%v) is missing", e.Path(), blob.ID.Str(), blob.Storage.Str())
|
debug.Log("Archiver.fileWorker", " %v not using old data, %v (%v) is missing", e.Path(), blob.ID.Str(), blob.Storage.Str())
|
||||||
contentMissing = true
|
contentMissing = true
|
||||||
break
|
break
|
||||||
@ -557,14 +557,14 @@ func (arch *Archiver) Snapshot(p *Progress, paths []string, parentID backend.ID)
|
|||||||
sn.Parent = parentID
|
sn.Parent = parentID
|
||||||
|
|
||||||
// load parent snapshot
|
// load parent snapshot
|
||||||
parent, err := LoadSnapshot(arch.s, parentID)
|
parent, err := LoadSnapshot(arch.repo, parentID)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, nil, err
|
return nil, nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
// start walker on old tree
|
// start walker on old tree
|
||||||
ch := make(chan WalkTreeJob)
|
ch := make(chan WalkTreeJob)
|
||||||
go WalkTree(arch.s, parent.Tree, done, ch)
|
go WalkTree(arch.repo, parent.Tree, done, ch)
|
||||||
jobs.Old = ch
|
jobs.Old = ch
|
||||||
} else {
|
} else {
|
||||||
// use closed channel
|
// use closed channel
|
||||||
@ -622,7 +622,7 @@ func (arch *Archiver) Snapshot(p *Progress, paths []string, parentID backend.ID)
|
|||||||
sn.Tree = root.Subtree
|
sn.Tree = root.Subtree
|
||||||
|
|
||||||
// save snapshot
|
// save snapshot
|
||||||
id, err := arch.s.SaveJSONUnpacked(backend.Snapshot, sn)
|
id, err := arch.repo.SaveJSONUnpacked(backend.Snapshot, sn)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, nil, err
|
return nil, nil, err
|
||||||
}
|
}
|
||||||
@ -631,14 +631,14 @@ func (arch *Archiver) Snapshot(p *Progress, paths []string, parentID backend.ID)
|
|||||||
sn.id = id
|
sn.id = id
|
||||||
debug.Log("Archiver.Snapshot", "saved snapshot %v", id.Str())
|
debug.Log("Archiver.Snapshot", "saved snapshot %v", id.Str())
|
||||||
|
|
||||||
// flush server
|
// flush repository
|
||||||
err = arch.s.Flush()
|
err = arch.repo.Flush()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, nil, err
|
return nil, nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
// save index
|
// save index
|
||||||
indexID, err := arch.s.SaveIndex()
|
indexID, err := arch.repo.SaveIndex()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
debug.Log("Archiver.Snapshot", "error saving index: %v", err)
|
debug.Log("Archiver.Snapshot", "error saving index: %v", err)
|
||||||
return nil, nil, err
|
return nil, nil, err
|
||||||
|
@ -51,8 +51,8 @@ func BenchmarkChunkEncrypt(b *testing.B) {
|
|||||||
data := Random(23, 10<<20) // 10MiB
|
data := Random(23, 10<<20) // 10MiB
|
||||||
rd := bytes.NewReader(data)
|
rd := bytes.NewReader(data)
|
||||||
|
|
||||||
s := SetupBackend(b)
|
s := SetupRepo(b)
|
||||||
defer TeardownBackend(b, s)
|
defer TeardownRepo(b, s)
|
||||||
|
|
||||||
buf := make([]byte, chunker.MaxSize)
|
buf := make([]byte, chunker.MaxSize)
|
||||||
buf2 := make([]byte, chunker.MaxSize)
|
buf2 := make([]byte, chunker.MaxSize)
|
||||||
@ -82,8 +82,8 @@ func benchmarkChunkEncryptP(b *testing.PB, buf []byte, rd Rdr, key *crypto.Key)
|
|||||||
}
|
}
|
||||||
|
|
||||||
func BenchmarkChunkEncryptParallel(b *testing.B) {
|
func BenchmarkChunkEncryptParallel(b *testing.B) {
|
||||||
s := SetupBackend(b)
|
s := SetupRepo(b)
|
||||||
defer TeardownBackend(b, s)
|
defer TeardownRepo(b, s)
|
||||||
|
|
||||||
data := Random(23, 10<<20) // 10MiB
|
data := Random(23, 10<<20) // 10MiB
|
||||||
|
|
||||||
@ -101,10 +101,10 @@ func BenchmarkChunkEncryptParallel(b *testing.B) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func archiveDirectory(b testing.TB) {
|
func archiveDirectory(b testing.TB) {
|
||||||
server := SetupBackend(b)
|
repo := SetupRepo(b)
|
||||||
defer TeardownBackend(b, server)
|
defer TeardownRepo(b, repo)
|
||||||
|
|
||||||
arch := restic.NewArchiver(server)
|
arch := restic.NewArchiver(repo)
|
||||||
|
|
||||||
_, id, err := arch.Snapshot(nil, []string{*benchArchiveDirectory}, nil)
|
_, id, err := arch.Snapshot(nil, []string{*benchArchiveDirectory}, nil)
|
||||||
OK(b, err)
|
OK(b, err)
|
||||||
@ -135,8 +135,8 @@ func archiveWithDedup(t testing.TB) {
|
|||||||
t.Skip("benchdir not set, skipping TestArchiverDedup")
|
t.Skip("benchdir not set, skipping TestArchiverDedup")
|
||||||
}
|
}
|
||||||
|
|
||||||
server := SetupBackend(t)
|
repo := SetupRepo(t)
|
||||||
defer TeardownBackend(t, server)
|
defer TeardownRepo(t, repo)
|
||||||
|
|
||||||
var cnt struct {
|
var cnt struct {
|
||||||
before, after, after2 struct {
|
before, after, after2 struct {
|
||||||
@ -145,24 +145,24 @@ func archiveWithDedup(t testing.TB) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// archive a few files
|
// archive a few files
|
||||||
sn := SnapshotDir(t, server, *benchArchiveDirectory, nil)
|
sn := SnapshotDir(t, repo, *benchArchiveDirectory, nil)
|
||||||
t.Logf("archived snapshot %v", sn.ID().Str())
|
t.Logf("archived snapshot %v", sn.ID().Str())
|
||||||
|
|
||||||
// get archive stats
|
// get archive stats
|
||||||
cnt.before.packs = server.Count(backend.Data)
|
cnt.before.packs = repo.Count(backend.Data)
|
||||||
cnt.before.dataBlobs = server.Index().Count(pack.Data)
|
cnt.before.dataBlobs = repo.Index().Count(pack.Data)
|
||||||
cnt.before.treeBlobs = server.Index().Count(pack.Tree)
|
cnt.before.treeBlobs = repo.Index().Count(pack.Tree)
|
||||||
t.Logf("packs %v, data blobs %v, tree blobs %v",
|
t.Logf("packs %v, data blobs %v, tree blobs %v",
|
||||||
cnt.before.packs, cnt.before.dataBlobs, cnt.before.treeBlobs)
|
cnt.before.packs, cnt.before.dataBlobs, cnt.before.treeBlobs)
|
||||||
|
|
||||||
// archive the same files again, without parent snapshot
|
// archive the same files again, without parent snapshot
|
||||||
sn2 := SnapshotDir(t, server, *benchArchiveDirectory, nil)
|
sn2 := SnapshotDir(t, repo, *benchArchiveDirectory, nil)
|
||||||
t.Logf("archived snapshot %v", sn2.ID().Str())
|
t.Logf("archived snapshot %v", sn2.ID().Str())
|
||||||
|
|
||||||
// get archive stats again
|
// get archive stats again
|
||||||
cnt.after.packs = server.Count(backend.Data)
|
cnt.after.packs = repo.Count(backend.Data)
|
||||||
cnt.after.dataBlobs = server.Index().Count(pack.Data)
|
cnt.after.dataBlobs = repo.Index().Count(pack.Data)
|
||||||
cnt.after.treeBlobs = server.Index().Count(pack.Tree)
|
cnt.after.treeBlobs = repo.Index().Count(pack.Tree)
|
||||||
t.Logf("packs %v, data blobs %v, tree blobs %v",
|
t.Logf("packs %v, data blobs %v, tree blobs %v",
|
||||||
cnt.after.packs, cnt.after.dataBlobs, cnt.after.treeBlobs)
|
cnt.after.packs, cnt.after.dataBlobs, cnt.after.treeBlobs)
|
||||||
|
|
||||||
@ -173,13 +173,13 @@ func archiveWithDedup(t testing.TB) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// archive the same files again, with a parent snapshot
|
// archive the same files again, with a parent snapshot
|
||||||
sn3 := SnapshotDir(t, server, *benchArchiveDirectory, sn2.ID())
|
sn3 := SnapshotDir(t, repo, *benchArchiveDirectory, sn2.ID())
|
||||||
t.Logf("archived snapshot %v, parent %v", sn3.ID().Str(), sn2.ID().Str())
|
t.Logf("archived snapshot %v, parent %v", sn3.ID().Str(), sn2.ID().Str())
|
||||||
|
|
||||||
// get archive stats again
|
// get archive stats again
|
||||||
cnt.after2.packs = server.Count(backend.Data)
|
cnt.after2.packs = repo.Count(backend.Data)
|
||||||
cnt.after2.dataBlobs = server.Index().Count(pack.Data)
|
cnt.after2.dataBlobs = repo.Index().Count(pack.Data)
|
||||||
cnt.after2.treeBlobs = server.Index().Count(pack.Tree)
|
cnt.after2.treeBlobs = repo.Index().Count(pack.Tree)
|
||||||
t.Logf("packs %v, data blobs %v, tree blobs %v",
|
t.Logf("packs %v, data blobs %v, tree blobs %v",
|
||||||
cnt.after2.packs, cnt.after2.dataBlobs, cnt.after2.treeBlobs)
|
cnt.after2.packs, cnt.after2.dataBlobs, cnt.after2.treeBlobs)
|
||||||
|
|
||||||
@ -199,8 +199,8 @@ func BenchmarkLoadTree(t *testing.B) {
|
|||||||
t.Skip("benchdir not set, skipping TestArchiverDedup")
|
t.Skip("benchdir not set, skipping TestArchiverDedup")
|
||||||
}
|
}
|
||||||
|
|
||||||
s := SetupBackend(t)
|
s := SetupRepo(t)
|
||||||
defer TeardownBackend(t, s)
|
defer TeardownRepo(t, s)
|
||||||
|
|
||||||
// archive a few files
|
// archive a few files
|
||||||
arch := restic.NewArchiver(s)
|
arch := restic.NewArchiver(s)
|
||||||
|
14
cache.go
14
cache.go
@ -10,21 +10,21 @@ import (
|
|||||||
|
|
||||||
"github.com/restic/restic/backend"
|
"github.com/restic/restic/backend"
|
||||||
"github.com/restic/restic/debug"
|
"github.com/restic/restic/debug"
|
||||||
"github.com/restic/restic/server"
|
"github.com/restic/restic/repo"
|
||||||
)
|
)
|
||||||
|
|
||||||
// Cache is used to locally cache items from a server.
|
// Cache is used to locally cache items from a repository.
|
||||||
type Cache struct {
|
type Cache struct {
|
||||||
base string
|
base string
|
||||||
}
|
}
|
||||||
|
|
||||||
func NewCache(s *server.Server) (*Cache, error) {
|
func NewCache(repo *repo.Repo) (*Cache, error) {
|
||||||
cacheDir, err := getCacheDir()
|
cacheDir, err := getCacheDir()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
basedir := filepath.Join(cacheDir, s.Config.ID)
|
basedir := filepath.Join(cacheDir, repo.Config.ID)
|
||||||
debug.Log("Cache.New", "opened cache at %v", basedir)
|
debug.Log("Cache.New", "opened cache at %v", basedir)
|
||||||
|
|
||||||
return &Cache{base: basedir}, nil
|
return &Cache{base: basedir}, nil
|
||||||
@ -105,8 +105,8 @@ func (c *Cache) purge(t backend.Type, subtype string, id backend.ID) error {
|
|||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
// Clear removes information from the cache that isn't present in the server any more.
|
// Clear removes information from the cache that isn't present in the repository any more.
|
||||||
func (c *Cache) Clear(s *server.Server) error {
|
func (c *Cache) Clear(repo *repo.Repo) error {
|
||||||
list, err := c.list(backend.Snapshot)
|
list, err := c.list(backend.Snapshot)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
@ -115,7 +115,7 @@ func (c *Cache) Clear(s *server.Server) error {
|
|||||||
for _, entry := range list {
|
for _, entry := range list {
|
||||||
debug.Log("Cache.Clear", "found entry %v", entry)
|
debug.Log("Cache.Clear", "found entry %v", entry)
|
||||||
|
|
||||||
if ok, err := s.Test(backend.Snapshot, entry.ID.String()); !ok || err != nil {
|
if ok, err := repo.Test(backend.Snapshot, entry.ID.String()); !ok || err != nil {
|
||||||
debug.Log("Cache.Clear", "snapshot %v doesn't exist any more, removing %v", entry.ID, entry)
|
debug.Log("Cache.Clear", "snapshot %v doesn't exist any more, removing %v", entry.ID, entry)
|
||||||
|
|
||||||
err = c.purge(backend.Snapshot, entry.Subtype, entry.ID)
|
err = c.purge(backend.Snapshot, entry.Subtype, entry.ID)
|
||||||
|
@ -8,13 +8,13 @@ import (
|
|||||||
)
|
)
|
||||||
|
|
||||||
func TestCache(t *testing.T) {
|
func TestCache(t *testing.T) {
|
||||||
server := SetupBackend(t)
|
repo := SetupRepo(t)
|
||||||
defer TeardownBackend(t, server)
|
defer TeardownRepo(t, repo)
|
||||||
|
|
||||||
_, err := restic.NewCache(server)
|
_, err := restic.NewCache(repo)
|
||||||
OK(t, err)
|
OK(t, err)
|
||||||
|
|
||||||
arch := restic.NewArchiver(server)
|
arch := restic.NewArchiver(repo)
|
||||||
|
|
||||||
// archive some files, this should automatically cache all blobs from the snapshot
|
// archive some files, this should automatically cache all blobs from the snapshot
|
||||||
_, _, err = arch.Snapshot(nil, []string{*benchArchiveDirectory}, nil)
|
_, _, err = arch.Snapshot(nil, []string{*benchArchiveDirectory}, nil)
|
||||||
|
@ -11,7 +11,7 @@ import (
|
|||||||
"github.com/restic/restic/backend"
|
"github.com/restic/restic/backend"
|
||||||
"github.com/restic/restic/debug"
|
"github.com/restic/restic/debug"
|
||||||
"github.com/restic/restic/pack"
|
"github.com/restic/restic/pack"
|
||||||
"github.com/restic/restic/server"
|
"github.com/restic/restic/repo"
|
||||||
)
|
)
|
||||||
|
|
||||||
type CmdCat struct{}
|
type CmdCat struct{}
|
||||||
@ -107,7 +107,7 @@ func (cmd CmdCat) Execute(args []string) error {
|
|||||||
|
|
||||||
dec := json.NewDecoder(rd)
|
dec := json.NewDecoder(rd)
|
||||||
|
|
||||||
var key server.Key
|
var key repo.Key
|
||||||
err = dec.Decode(&key)
|
err = dec.Decode(&key)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
|
@ -8,7 +8,7 @@ import (
|
|||||||
"github.com/restic/restic"
|
"github.com/restic/restic"
|
||||||
"github.com/restic/restic/backend"
|
"github.com/restic/restic/backend"
|
||||||
"github.com/restic/restic/debug"
|
"github.com/restic/restic/debug"
|
||||||
"github.com/restic/restic/server"
|
"github.com/restic/restic/repo"
|
||||||
)
|
)
|
||||||
|
|
||||||
type findResult struct {
|
type findResult struct {
|
||||||
@ -59,9 +59,9 @@ func parseTime(str string) (time.Time, error) {
|
|||||||
return time.Time{}, fmt.Errorf("unable to parse time: %q", str)
|
return time.Time{}, fmt.Errorf("unable to parse time: %q", str)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (c CmdFind) findInTree(s *server.Server, id backend.ID, path string) ([]findResult, error) {
|
func (c CmdFind) findInTree(repo *repo.Repo, id backend.ID, path string) ([]findResult, error) {
|
||||||
debug.Log("restic.find", "checking tree %v\n", id)
|
debug.Log("restic.find", "checking tree %v\n", id)
|
||||||
tree, err := restic.LoadTree(s, id)
|
tree, err := restic.LoadTree(repo, id)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
@ -93,7 +93,7 @@ func (c CmdFind) findInTree(s *server.Server, id backend.ID, path string) ([]fin
|
|||||||
}
|
}
|
||||||
|
|
||||||
if node.Type == "dir" {
|
if node.Type == "dir" {
|
||||||
subdirResults, err := c.findInTree(s, id, filepath.Join(path, node.Name))
|
subdirResults, err := c.findInTree(repo, id, filepath.Join(path, node.Name))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
@ -105,7 +105,7 @@ func (c CmdFind) findInTree(s *server.Server, id backend.ID, path string) ([]fin
|
|||||||
return results, nil
|
return results, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (c CmdFind) findInSnapshot(s *server.Server, name string) error {
|
func (c CmdFind) findInSnapshot(repo *repo.Repo, name string) error {
|
||||||
debug.Log("restic.find", "searching in snapshot %s\n for entries within [%s %s]", name, c.oldest, c.newest)
|
debug.Log("restic.find", "searching in snapshot %s\n for entries within [%s %s]", name, c.oldest, c.newest)
|
||||||
|
|
||||||
id, err := backend.ParseID(name)
|
id, err := backend.ParseID(name)
|
||||||
@ -113,12 +113,12 @@ func (c CmdFind) findInSnapshot(s *server.Server, name string) error {
|
|||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
sn, err := restic.LoadSnapshot(s, id)
|
sn, err := restic.LoadSnapshot(repo, id)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
results, err := c.findInTree(s, sn.Tree, "")
|
results, err := c.findInTree(repo, sn.Tree, "")
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
@ -10,7 +10,7 @@ import (
|
|||||||
"github.com/restic/restic/crypto"
|
"github.com/restic/restic/crypto"
|
||||||
"github.com/restic/restic/debug"
|
"github.com/restic/restic/debug"
|
||||||
"github.com/restic/restic/pack"
|
"github.com/restic/restic/pack"
|
||||||
"github.com/restic/restic/server"
|
"github.com/restic/restic/repo"
|
||||||
)
|
)
|
||||||
|
|
||||||
type CmdFsck struct {
|
type CmdFsck struct {
|
||||||
@ -34,7 +34,7 @@ func init() {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func fsckFile(opts CmdFsck, s *server.Server, IDs []backend.ID) (uint64, error) {
|
func fsckFile(opts CmdFsck, repo *repo.Repo, IDs []backend.ID) (uint64, error) {
|
||||||
debug.Log("restic.fsckFile", "checking file %v", IDs)
|
debug.Log("restic.fsckFile", "checking file %v", IDs)
|
||||||
var bytes uint64
|
var bytes uint64
|
||||||
|
|
||||||
@ -42,7 +42,7 @@ func fsckFile(opts CmdFsck, s *server.Server, IDs []backend.ID) (uint64, error)
|
|||||||
debug.Log("restic.fsck", " checking data blob %v\n", id)
|
debug.Log("restic.fsck", " checking data blob %v\n", id)
|
||||||
|
|
||||||
// test if blob is in the index
|
// test if blob is in the index
|
||||||
packID, tpe, _, length, err := s.Index().Lookup(id)
|
packID, tpe, _, length, err := repo.Index().Lookup(id)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return 0, fmt.Errorf("storage for blob %v (%v) not found", id, tpe)
|
return 0, fmt.Errorf("storage for blob %v (%v) not found", id, tpe)
|
||||||
}
|
}
|
||||||
@ -52,13 +52,13 @@ func fsckFile(opts CmdFsck, s *server.Server, IDs []backend.ID) (uint64, error)
|
|||||||
|
|
||||||
if opts.CheckData {
|
if opts.CheckData {
|
||||||
// load content
|
// load content
|
||||||
_, err := s.LoadBlob(pack.Data, id)
|
_, err := repo.LoadBlob(pack.Data, id)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return 0, err
|
return 0, err
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
// test if data blob is there
|
// test if data blob is there
|
||||||
ok, err := s.Test(backend.Data, packID.String())
|
ok, err := repo.Test(backend.Data, packID.String())
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return 0, err
|
return 0, err
|
||||||
}
|
}
|
||||||
@ -77,10 +77,10 @@ func fsckFile(opts CmdFsck, s *server.Server, IDs []backend.ID) (uint64, error)
|
|||||||
return bytes, nil
|
return bytes, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func fsckTree(opts CmdFsck, s *server.Server, id backend.ID) error {
|
func fsckTree(opts CmdFsck, repo *repo.Repo, id backend.ID) error {
|
||||||
debug.Log("restic.fsckTree", "checking tree %v", id.Str())
|
debug.Log("restic.fsckTree", "checking tree %v", id.Str())
|
||||||
|
|
||||||
tree, err := restic.LoadTree(s, id)
|
tree, err := restic.LoadTree(repo, id)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
@ -122,7 +122,7 @@ func fsckTree(opts CmdFsck, s *server.Server, id backend.ID) error {
|
|||||||
}
|
}
|
||||||
|
|
||||||
debug.Log("restic.fsckTree", "check file %v (%v)", node.Name, id.Str())
|
debug.Log("restic.fsckTree", "check file %v (%v)", node.Name, id.Str())
|
||||||
bytes, err := fsckFile(opts, s, node.Content)
|
bytes, err := fsckFile(opts, repo, node.Content)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
@ -139,7 +139,7 @@ func fsckTree(opts CmdFsck, s *server.Server, id backend.ID) error {
|
|||||||
// record id
|
// record id
|
||||||
seenIDs.Insert(node.Subtree)
|
seenIDs.Insert(node.Subtree)
|
||||||
|
|
||||||
err = fsckTree(opts, s, node.Subtree)
|
err = fsckTree(opts, repo, node.Subtree)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
firstErr = err
|
firstErr = err
|
||||||
fmt.Fprintf(os.Stderr, "%v\n", err)
|
fmt.Fprintf(os.Stderr, "%v\n", err)
|
||||||
@ -157,15 +157,15 @@ func fsckTree(opts CmdFsck, s *server.Server, id backend.ID) error {
|
|||||||
return firstErr
|
return firstErr
|
||||||
}
|
}
|
||||||
|
|
||||||
func fsckSnapshot(opts CmdFsck, s *server.Server, id backend.ID) error {
|
func fsckSnapshot(opts CmdFsck, repo *repo.Repo, id backend.ID) error {
|
||||||
debug.Log("restic.fsck", "checking snapshot %v\n", id)
|
debug.Log("restic.fsck", "checking snapshot %v\n", id)
|
||||||
|
|
||||||
sn, err := restic.LoadSnapshot(s, id)
|
sn, err := restic.LoadSnapshot(repo, id)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("loading snapshot %v failed: %v", id, err)
|
return fmt.Errorf("loading snapshot %v failed: %v", id, err)
|
||||||
}
|
}
|
||||||
|
|
||||||
err = fsckTree(opts, s, sn.Tree)
|
err = fsckTree(opts, repo, sn.Tree)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
debug.Log("restic.fsck", " checking tree %v for snapshot %v\n", sn.Tree, id)
|
debug.Log("restic.fsck", " checking tree %v for snapshot %v\n", sn.Tree, id)
|
||||||
fmt.Fprintf(os.Stderr, "snapshot %v:\n error for tree %v:\n %v\n", id, sn.Tree, err)
|
fmt.Fprintf(os.Stderr, "snapshot %v:\n error for tree %v:\n %v\n", id, sn.Tree, err)
|
||||||
|
@ -6,7 +6,7 @@ import (
|
|||||||
"os"
|
"os"
|
||||||
|
|
||||||
"github.com/restic/restic/backend"
|
"github.com/restic/restic/backend"
|
||||||
"github.com/restic/restic/server"
|
"github.com/restic/restic/repo"
|
||||||
)
|
)
|
||||||
|
|
||||||
type CmdKey struct{}
|
type CmdKey struct{}
|
||||||
@ -21,7 +21,7 @@ func init() {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func listKeys(s *server.Server) error {
|
func listKeys(s *repo.Repo) error {
|
||||||
tab := NewTable()
|
tab := NewTable()
|
||||||
tab.Header = fmt.Sprintf(" %-10s %-10s %-10s %s", "ID", "User", "Host", "Created")
|
tab.Header = fmt.Sprintf(" %-10s %-10s %-10s %s", "ID", "User", "Host", "Created")
|
||||||
tab.RowFormat = "%s%-10s %-10s %-10s %s"
|
tab.RowFormat = "%s%-10s %-10s %-10s %s"
|
||||||
@ -35,7 +35,7 @@ func listKeys(s *server.Server) error {
|
|||||||
defer close(done)
|
defer close(done)
|
||||||
|
|
||||||
for name := range s.List(backend.Key, done) {
|
for name := range s.List(backend.Key, done) {
|
||||||
k, err := server.LoadKey(s, name)
|
k, err := repo.LoadKey(s, name)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
fmt.Fprintf(os.Stderr, "LoadKey() failed: %v\n", err)
|
fmt.Fprintf(os.Stderr, "LoadKey() failed: %v\n", err)
|
||||||
continue
|
continue
|
||||||
@ -56,7 +56,7 @@ func listKeys(s *server.Server) error {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func addKey(s *server.Server) error {
|
func addKey(s *repo.Repo) error {
|
||||||
pw := readPassword("RESTIC_NEWPASSWORD", "enter password for new key: ")
|
pw := readPassword("RESTIC_NEWPASSWORD", "enter password for new key: ")
|
||||||
pw2 := readPassword("RESTIC_NEWPASSWORD", "enter password again: ")
|
pw2 := readPassword("RESTIC_NEWPASSWORD", "enter password again: ")
|
||||||
|
|
||||||
@ -64,7 +64,7 @@ func addKey(s *server.Server) error {
|
|||||||
return errors.New("passwords do not match")
|
return errors.New("passwords do not match")
|
||||||
}
|
}
|
||||||
|
|
||||||
id, err := server.AddKey(s, pw, s.Key())
|
id, err := repo.AddKey(s, pw, s.Key())
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("creating new key failed: %v\n", err)
|
return fmt.Errorf("creating new key failed: %v\n", err)
|
||||||
}
|
}
|
||||||
@ -74,12 +74,12 @@ func addKey(s *server.Server) error {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func deleteKey(s *server.Server, name string) error {
|
func deleteKey(repo *repo.Repo, name string) error {
|
||||||
if name == s.KeyName() {
|
if name == repo.KeyName() {
|
||||||
return errors.New("refusing to remove key currently used to access repository")
|
return errors.New("refusing to remove key currently used to access repository")
|
||||||
}
|
}
|
||||||
|
|
||||||
err := s.Remove(backend.Key, name)
|
err := repo.Remove(backend.Key, name)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
@ -88,7 +88,7 @@ func deleteKey(s *server.Server, name string) error {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func changePassword(s *server.Server) error {
|
func changePassword(s *repo.Repo) error {
|
||||||
pw := readPassword("RESTIC_NEWPASSWORD", "enter password for new key: ")
|
pw := readPassword("RESTIC_NEWPASSWORD", "enter password for new key: ")
|
||||||
pw2 := readPassword("RESTIC_NEWPASSWORD", "enter password again: ")
|
pw2 := readPassword("RESTIC_NEWPASSWORD", "enter password again: ")
|
||||||
|
|
||||||
@ -97,7 +97,7 @@ func changePassword(s *server.Server) error {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// add new key
|
// add new key
|
||||||
id, err := server.AddKey(s, pw, s.Key())
|
id, err := repo.AddKey(s, pw, s.Key())
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("creating new key failed: %v\n", err)
|
return fmt.Errorf("creating new key failed: %v\n", err)
|
||||||
}
|
}
|
||||||
|
@ -7,7 +7,7 @@ import (
|
|||||||
|
|
||||||
"github.com/restic/restic"
|
"github.com/restic/restic"
|
||||||
"github.com/restic/restic/backend"
|
"github.com/restic/restic/backend"
|
||||||
"github.com/restic/restic/server"
|
"github.com/restic/restic/repo"
|
||||||
)
|
)
|
||||||
|
|
||||||
type CmdLs struct{}
|
type CmdLs struct{}
|
||||||
@ -38,8 +38,8 @@ func printNode(prefix string, n *restic.Node) string {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func printTree(prefix string, s *server.Server, id backend.ID) error {
|
func printTree(prefix string, repo *repo.Repo, id backend.ID) error {
|
||||||
tree, err := restic.LoadTree(s, id)
|
tree, err := restic.LoadTree(repo, id)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
@ -48,7 +48,7 @@ func printTree(prefix string, s *server.Server, id backend.ID) error {
|
|||||||
fmt.Println(printNode(prefix, entry))
|
fmt.Println(printNode(prefix, entry))
|
||||||
|
|
||||||
if entry.Type == "dir" && entry.Subtree != nil {
|
if entry.Type == "dir" && entry.Subtree != nil {
|
||||||
err = printTree(filepath.Join(prefix, entry.Name), s, entry.Subtree)
|
err = printTree(filepath.Join(prefix, entry.Name), repo, entry.Subtree)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
@ -14,7 +14,7 @@ import (
|
|||||||
"github.com/restic/restic/backend/local"
|
"github.com/restic/restic/backend/local"
|
||||||
"github.com/restic/restic/backend/sftp"
|
"github.com/restic/restic/backend/sftp"
|
||||||
"github.com/restic/restic/debug"
|
"github.com/restic/restic/debug"
|
||||||
"github.com/restic/restic/server"
|
"github.com/restic/restic/repo"
|
||||||
)
|
)
|
||||||
|
|
||||||
var version = "compiled manually"
|
var version = "compiled manually"
|
||||||
@ -72,7 +72,7 @@ func (cmd CmdInit) Execute(args []string) error {
|
|||||||
os.Exit(1)
|
os.Exit(1)
|
||||||
}
|
}
|
||||||
|
|
||||||
s := server.NewServer(be)
|
s := repo.New(be)
|
||||||
err = s.Init(pw)
|
err = s.Init(pw)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
fmt.Fprintf(os.Stderr, "creating key in backend at %s failed: %v\n", opts.Repo, err)
|
fmt.Fprintf(os.Stderr, "creating key in backend at %s failed: %v\n", opts.Repo, err)
|
||||||
@ -133,7 +133,7 @@ func create(u string) (backend.Backend, error) {
|
|||||||
return sftp.Create(url.Path[1:], "ssh", args...)
|
return sftp.Create(url.Path[1:], "ssh", args...)
|
||||||
}
|
}
|
||||||
|
|
||||||
func OpenRepo() (*server.Server, error) {
|
func OpenRepo() (*repo.Repo, error) {
|
||||||
if opts.Repo == "" {
|
if opts.Repo == "" {
|
||||||
return nil, errors.New("Please specify repository location (-r)")
|
return nil, errors.New("Please specify repository location (-r)")
|
||||||
}
|
}
|
||||||
@ -143,7 +143,7 @@ func OpenRepo() (*server.Server, error) {
|
|||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
s := server.NewServer(be)
|
s := repo.New(be)
|
||||||
|
|
||||||
err = s.SearchKey(readPassword("RESTIC_PASSWORD", "enter password for repository: "))
|
err = s.SearchKey(readPassword("RESTIC_PASSWORD", "enter password for repository: "))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
12
node.go
12
node.go
@ -14,7 +14,7 @@ import (
|
|||||||
"github.com/restic/restic/backend"
|
"github.com/restic/restic/backend"
|
||||||
"github.com/restic/restic/debug"
|
"github.com/restic/restic/debug"
|
||||||
"github.com/restic/restic/pack"
|
"github.com/restic/restic/pack"
|
||||||
"github.com/restic/restic/server"
|
"github.com/restic/restic/repo"
|
||||||
)
|
)
|
||||||
|
|
||||||
// Node is a file, directory or other item in a backup.
|
// Node is a file, directory or other item in a backup.
|
||||||
@ -43,7 +43,7 @@ type Node struct {
|
|||||||
|
|
||||||
path string
|
path string
|
||||||
err error
|
err error
|
||||||
blobs server.Blobs
|
blobs repo.Blobs
|
||||||
}
|
}
|
||||||
|
|
||||||
func (node Node) String() string {
|
func (node Node) String() string {
|
||||||
@ -103,14 +103,14 @@ func nodeTypeFromFileInfo(fi os.FileInfo) string {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// CreateAt creates the node at the given path and restores all the meta data.
|
// CreateAt creates the node at the given path and restores all the meta data.
|
||||||
func (node *Node) CreateAt(path string, s *server.Server) error {
|
func (node *Node) CreateAt(path string, repo *repo.Repo) error {
|
||||||
switch node.Type {
|
switch node.Type {
|
||||||
case "dir":
|
case "dir":
|
||||||
if err := node.createDirAt(path); err != nil {
|
if err := node.createDirAt(path); err != nil {
|
||||||
return errors.Annotate(err, "createDirAt")
|
return errors.Annotate(err, "createDirAt")
|
||||||
}
|
}
|
||||||
case "file":
|
case "file":
|
||||||
if err := node.createFileAt(path, s); err != nil {
|
if err := node.createFileAt(path, repo); err != nil {
|
||||||
return errors.Annotate(err, "createFileAt")
|
return errors.Annotate(err, "createFileAt")
|
||||||
}
|
}
|
||||||
case "symlink":
|
case "symlink":
|
||||||
@ -176,7 +176,7 @@ func (node Node) createDirAt(path string) error {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (node Node) createFileAt(path string, s *server.Server) error {
|
func (node Node) createFileAt(path string, repo *repo.Repo) error {
|
||||||
f, err := os.OpenFile(path, os.O_CREATE|os.O_WRONLY, 0600)
|
f, err := os.OpenFile(path, os.O_CREATE|os.O_WRONLY, 0600)
|
||||||
defer f.Close()
|
defer f.Close()
|
||||||
|
|
||||||
@ -185,7 +185,7 @@ func (node Node) createFileAt(path string, s *server.Server) error {
|
|||||||
}
|
}
|
||||||
|
|
||||||
for _, id := range node.Content {
|
for _, id := range node.Content {
|
||||||
buf, err := s.LoadBlob(pack.Data, id)
|
buf, err := repo.LoadBlob(pack.Data, id)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return errors.Annotate(err, "Load")
|
return errors.Annotate(err, "Load")
|
||||||
}
|
}
|
||||||
|
@ -1,4 +1,4 @@
|
|||||||
package server
|
package repo
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"bytes"
|
"bytes"
|
2
repo/doc.go
Normal file
2
repo/doc.go
Normal file
@ -0,0 +1,2 @@
|
|||||||
|
// Package repo implements a restic repository on top of a backend.
|
||||||
|
package repo
|
@ -1,4 +1,4 @@
|
|||||||
package server
|
package repo
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"encoding/json"
|
"encoding/json"
|
@ -1,4 +1,4 @@
|
|||||||
package server_test
|
package repo_test
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"bytes"
|
"bytes"
|
||||||
@ -8,7 +8,7 @@ import (
|
|||||||
|
|
||||||
"github.com/restic/restic/backend"
|
"github.com/restic/restic/backend"
|
||||||
"github.com/restic/restic/pack"
|
"github.com/restic/restic/pack"
|
||||||
"github.com/restic/restic/server"
|
"github.com/restic/restic/repo"
|
||||||
. "github.com/restic/restic/test"
|
. "github.com/restic/restic/test"
|
||||||
)
|
)
|
||||||
|
|
||||||
@ -30,7 +30,7 @@ func TestIndexSerialize(t *testing.T) {
|
|||||||
}
|
}
|
||||||
tests := []testEntry{}
|
tests := []testEntry{}
|
||||||
|
|
||||||
idx := server.NewIndex()
|
idx := repo.NewIndex()
|
||||||
|
|
||||||
// create 50 packs with 20 blobs each
|
// create 50 packs with 20 blobs each
|
||||||
for i := 0; i < 50; i++ {
|
for i := 0; i < 50; i++ {
|
||||||
@ -58,7 +58,7 @@ func TestIndexSerialize(t *testing.T) {
|
|||||||
err := idx.Encode(wr)
|
err := idx.Encode(wr)
|
||||||
OK(t, err)
|
OK(t, err)
|
||||||
|
|
||||||
idx2, err := server.DecodeIndex(wr)
|
idx2, err := repo.DecodeIndex(wr)
|
||||||
OK(t, err)
|
OK(t, err)
|
||||||
Assert(t, idx2 != nil,
|
Assert(t, idx2 != nil,
|
||||||
"nil returned for decoded index")
|
"nil returned for decoded index")
|
||||||
@ -113,7 +113,7 @@ func TestIndexSerialize(t *testing.T) {
|
|||||||
err = idx2.Encode(wr3)
|
err = idx2.Encode(wr3)
|
||||||
OK(t, err)
|
OK(t, err)
|
||||||
|
|
||||||
idx3, err := server.DecodeIndex(wr3)
|
idx3, err := repo.DecodeIndex(wr3)
|
||||||
OK(t, err)
|
OK(t, err)
|
||||||
Assert(t, idx3 != nil,
|
Assert(t, idx3 != nil,
|
||||||
"nil returned for decoded index")
|
"nil returned for decoded index")
|
||||||
@ -138,7 +138,7 @@ func TestIndexSerialize(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func TestIndexSize(t *testing.T) {
|
func TestIndexSize(t *testing.T) {
|
||||||
idx := server.NewIndex()
|
idx := repo.NewIndex()
|
||||||
|
|
||||||
packs := 200
|
packs := 200
|
||||||
blobs := 100
|
blobs := 100
|
||||||
@ -210,7 +210,7 @@ var exampleTests = []struct {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func TestIndexUnserialize(t *testing.T) {
|
func TestIndexUnserialize(t *testing.T) {
|
||||||
idx, err := server.DecodeIndex(bytes.NewReader(docExample))
|
idx, err := repo.DecodeIndex(bytes.NewReader(docExample))
|
||||||
OK(t, err)
|
OK(t, err)
|
||||||
|
|
||||||
for _, test := range exampleTests {
|
for _, test := range exampleTests {
|
@ -1,4 +1,4 @@
|
|||||||
package server
|
package repo
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"crypto/rand"
|
"crypto/rand"
|
||||||
@ -49,12 +49,12 @@ type Key struct {
|
|||||||
|
|
||||||
// createMasterKey creates a new master key in the given backend and encrypts
|
// createMasterKey creates a new master key in the given backend and encrypts
|
||||||
// it with the password.
|
// it with the password.
|
||||||
func createMasterKey(s *Server, password string) (*Key, error) {
|
func createMasterKey(s *Repo, password string) (*Key, error) {
|
||||||
return AddKey(s, password, nil)
|
return AddKey(s, password, nil)
|
||||||
}
|
}
|
||||||
|
|
||||||
// OpenKey tries do decrypt the key specified by name with the given password.
|
// OpenKey tries do decrypt the key specified by name with the given password.
|
||||||
func OpenKey(s *Server, name string, password string) (*Key, error) {
|
func OpenKey(s *Repo, name string, password string) (*Key, error) {
|
||||||
k, err := LoadKey(s, name)
|
k, err := LoadKey(s, name)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
@ -94,7 +94,7 @@ func OpenKey(s *Server, name string, password string) (*Key, error) {
|
|||||||
|
|
||||||
// SearchKey tries to decrypt all keys in the backend with the given password.
|
// SearchKey tries to decrypt all keys in the backend with the given password.
|
||||||
// If none could be found, ErrNoKeyFound is returned.
|
// If none could be found, ErrNoKeyFound is returned.
|
||||||
func SearchKey(s *Server, password string) (*Key, error) {
|
func SearchKey(s *Repo, password string) (*Key, error) {
|
||||||
// try all keys in repo
|
// try all keys in repo
|
||||||
done := make(chan struct{})
|
done := make(chan struct{})
|
||||||
defer close(done)
|
defer close(done)
|
||||||
@ -111,7 +111,7 @@ func SearchKey(s *Server, password string) (*Key, error) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// LoadKey loads a key from the backend.
|
// LoadKey loads a key from the backend.
|
||||||
func LoadKey(s *Server, name string) (*Key, error) {
|
func LoadKey(s *Repo, name string) (*Key, error) {
|
||||||
// extract data from repo
|
// extract data from repo
|
||||||
rd, err := s.be.Get(backend.Key, name)
|
rd, err := s.be.Get(backend.Key, name)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@ -131,7 +131,7 @@ func LoadKey(s *Server, name string) (*Key, error) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// AddKey adds a new key to an already existing repository.
|
// AddKey adds a new key to an already existing repository.
|
||||||
func AddKey(s *Server, password string, template *crypto.Key) (*Key, error) {
|
func AddKey(s *Repo, password string, template *crypto.Key) (*Key, error) {
|
||||||
// fill meta data about key
|
// fill meta data about key
|
||||||
newkey := &Key{
|
newkey := &Key{
|
||||||
Created: time.Now(),
|
Created: time.Now(),
|
@ -1,4 +1,4 @@
|
|||||||
package server
|
package repo
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"sync"
|
"sync"
|
@ -1,4 +1,4 @@
|
|||||||
package server
|
package repo
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"bytes"
|
"bytes"
|
||||||
@ -26,8 +26,8 @@ type Config struct {
|
|||||||
ChunkerPolynomial chunker.Pol `json:"chunker_polynomial"`
|
ChunkerPolynomial chunker.Pol `json:"chunker_polynomial"`
|
||||||
}
|
}
|
||||||
|
|
||||||
// Server is used to access a repository in a backend.
|
// Repo is used to access a repository in a backend.
|
||||||
type Server struct {
|
type Repo struct {
|
||||||
be backend.Backend
|
be backend.Backend
|
||||||
Config Config
|
Config Config
|
||||||
key *crypto.Key
|
key *crypto.Key
|
||||||
@ -38,8 +38,8 @@ type Server struct {
|
|||||||
packs []*pack.Packer
|
packs []*pack.Packer
|
||||||
}
|
}
|
||||||
|
|
||||||
func NewServer(be backend.Backend) *Server {
|
func New(be backend.Backend) *Repo {
|
||||||
return &Server{
|
return &Repo{
|
||||||
be: be,
|
be: be,
|
||||||
idx: NewIndex(),
|
idx: NewIndex(),
|
||||||
}
|
}
|
||||||
@ -48,31 +48,31 @@ func NewServer(be backend.Backend) *Server {
|
|||||||
// Find loads the list of all blobs of type t and searches for names which start
|
// Find loads the list of all blobs of type t and searches for names which start
|
||||||
// with prefix. If none is found, nil and ErrNoIDPrefixFound is returned. If
|
// with prefix. If none is found, nil and ErrNoIDPrefixFound is returned. If
|
||||||
// more than one is found, nil and ErrMultipleIDMatches is returned.
|
// more than one is found, nil and ErrMultipleIDMatches is returned.
|
||||||
func (s *Server) Find(t backend.Type, prefix string) (string, error) {
|
func (s *Repo) Find(t backend.Type, prefix string) (string, error) {
|
||||||
return backend.Find(s.be, t, prefix)
|
return backend.Find(s.be, t, prefix)
|
||||||
}
|
}
|
||||||
|
|
||||||
// FindSnapshot takes a string and tries to find a snapshot whose ID matches
|
// FindSnapshot takes a string and tries to find a snapshot whose ID matches
|
||||||
// the string as closely as possible.
|
// the string as closely as possible.
|
||||||
func (s *Server) FindSnapshot(name string) (string, error) {
|
func (s *Repo) FindSnapshot(name string) (string, error) {
|
||||||
return backend.FindSnapshot(s.be, name)
|
return backend.FindSnapshot(s.be, name)
|
||||||
}
|
}
|
||||||
|
|
||||||
// PrefixLength returns the number of bytes required so that all prefixes of
|
// PrefixLength returns the number of bytes required so that all prefixes of
|
||||||
// all IDs of type t are unique.
|
// all IDs of type t are unique.
|
||||||
func (s *Server) PrefixLength(t backend.Type) (int, error) {
|
func (s *Repo) PrefixLength(t backend.Type) (int, error) {
|
||||||
return backend.PrefixLength(s.be, t)
|
return backend.PrefixLength(s.be, t)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Load tries to load and decrypt content identified by t and id from the
|
// Load tries to load and decrypt content identified by t and id from the
|
||||||
// backend.
|
// backend.
|
||||||
func (s *Server) Load(t backend.Type, id backend.ID) ([]byte, error) {
|
func (s *Repo) Load(t backend.Type, id backend.ID) ([]byte, error) {
|
||||||
debug.Log("Server.Load", "load %v with id %v", t, id.Str())
|
debug.Log("Repo.Load", "load %v with id %v", t, id.Str())
|
||||||
|
|
||||||
// load blob from pack
|
// load blob from pack
|
||||||
rd, err := s.be.Get(t, id.String())
|
rd, err := s.be.Get(t, id.String())
|
||||||
if err != nil {
|
if err != nil {
|
||||||
debug.Log("Server.Load", "error loading %v: %v", id.Str(), err)
|
debug.Log("Repo.Load", "error loading %v: %v", id.Str(), err)
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -102,26 +102,26 @@ func (s *Server) Load(t backend.Type, id backend.ID) ([]byte, error) {
|
|||||||
|
|
||||||
// LoadBlob tries to load and decrypt content identified by t and id from a
|
// LoadBlob tries to load and decrypt content identified by t and id from a
|
||||||
// pack from the backend.
|
// pack from the backend.
|
||||||
func (s *Server) LoadBlob(t pack.BlobType, id backend.ID) ([]byte, error) {
|
func (s *Repo) LoadBlob(t pack.BlobType, id backend.ID) ([]byte, error) {
|
||||||
debug.Log("Server.LoadBlob", "load %v with id %v", t, id.Str())
|
debug.Log("Repo.LoadBlob", "load %v with id %v", t, id.Str())
|
||||||
// lookup pack
|
// lookup pack
|
||||||
packID, tpe, offset, length, err := s.idx.Lookup(id)
|
packID, tpe, offset, length, err := s.idx.Lookup(id)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
debug.Log("Server.LoadBlob", "id %v not found in index: %v", id.Str(), err)
|
debug.Log("Repo.LoadBlob", "id %v not found in index: %v", id.Str(), err)
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
if tpe != t {
|
if tpe != t {
|
||||||
debug.Log("Server.LoadBlob", "wrong type returned for %v: wanted %v, got %v", id.Str(), t, tpe)
|
debug.Log("Repo.LoadBlob", "wrong type returned for %v: wanted %v, got %v", id.Str(), t, tpe)
|
||||||
return nil, fmt.Errorf("blob has wrong type %v (wanted: %v)", tpe, t)
|
return nil, fmt.Errorf("blob has wrong type %v (wanted: %v)", tpe, t)
|
||||||
}
|
}
|
||||||
|
|
||||||
debug.Log("Server.LoadBlob", "id %v found in pack %v at offset %v (length %d)", id.Str(), packID.Str(), offset, length)
|
debug.Log("Repo.LoadBlob", "id %v found in pack %v at offset %v (length %d)", id.Str(), packID.Str(), offset, length)
|
||||||
|
|
||||||
// load blob from pack
|
// load blob from pack
|
||||||
rd, err := s.be.GetReader(backend.Data, packID.String(), offset, length)
|
rd, err := s.be.GetReader(backend.Data, packID.String(), offset, length)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
debug.Log("Server.LoadBlob", "error loading pack %v for %v: %v", packID.Str(), id.Str(), err)
|
debug.Log("Repo.LoadBlob", "error loading pack %v for %v: %v", packID.Str(), id.Str(), err)
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -151,7 +151,7 @@ func (s *Server) LoadBlob(t pack.BlobType, id backend.ID) ([]byte, error) {
|
|||||||
|
|
||||||
// LoadJSONUnpacked decrypts the data and afterwards calls json.Unmarshal on
|
// LoadJSONUnpacked decrypts the data and afterwards calls json.Unmarshal on
|
||||||
// the item.
|
// the item.
|
||||||
func (s *Server) LoadJSONUnpacked(t backend.Type, id backend.ID, item interface{}) error {
|
func (s *Repo) LoadJSONUnpacked(t backend.Type, id backend.ID, item interface{}) error {
|
||||||
// load blob from backend
|
// load blob from backend
|
||||||
rd, err := s.be.Get(t, id.String())
|
rd, err := s.be.Get(t, id.String())
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@ -178,7 +178,7 @@ func (s *Server) LoadJSONUnpacked(t backend.Type, id backend.ID, item interface{
|
|||||||
|
|
||||||
// LoadJSONPack calls LoadBlob() to load a blob from the backend, decrypt the
|
// LoadJSONPack calls LoadBlob() to load a blob from the backend, decrypt the
|
||||||
// data and afterwards call json.Unmarshal on the item.
|
// data and afterwards call json.Unmarshal on the item.
|
||||||
func (s *Server) LoadJSONPack(t pack.BlobType, id backend.ID, item interface{}) error {
|
func (s *Repo) LoadJSONPack(t pack.BlobType, id backend.ID, item interface{}) error {
|
||||||
// lookup pack
|
// lookup pack
|
||||||
packID, _, offset, length, err := s.idx.Lookup(id)
|
packID, _, offset, length, err := s.idx.Lookup(id)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@ -215,16 +215,16 @@ const maxPackers = 200
|
|||||||
|
|
||||||
// findPacker returns a packer for a new blob of size bytes. Either a new one is
|
// findPacker returns a packer for a new blob of size bytes. Either a new one is
|
||||||
// created or one is returned that already has some blobs.
|
// created or one is returned that already has some blobs.
|
||||||
func (s *Server) findPacker(size uint) (*pack.Packer, error) {
|
func (s *Repo) findPacker(size uint) (*pack.Packer, error) {
|
||||||
s.pm.Lock()
|
s.pm.Lock()
|
||||||
defer s.pm.Unlock()
|
defer s.pm.Unlock()
|
||||||
|
|
||||||
// search for a suitable packer
|
// search for a suitable packer
|
||||||
if len(s.packs) > 0 {
|
if len(s.packs) > 0 {
|
||||||
debug.Log("Server.findPacker", "searching packer for %d bytes\n", size)
|
debug.Log("Repo.findPacker", "searching packer for %d bytes\n", size)
|
||||||
for i, p := range s.packs {
|
for i, p := range s.packs {
|
||||||
if p.Size()+size < maxPackSize {
|
if p.Size()+size < maxPackSize {
|
||||||
debug.Log("Server.findPacker", "found packer %v", p)
|
debug.Log("Repo.findPacker", "found packer %v", p)
|
||||||
// remove from list
|
// remove from list
|
||||||
s.packs = append(s.packs[:i], s.packs[i+1:]...)
|
s.packs = append(s.packs[:i], s.packs[i+1:]...)
|
||||||
return p, nil
|
return p, nil
|
||||||
@ -237,22 +237,22 @@ func (s *Server) findPacker(size uint) (*pack.Packer, error) {
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
debug.Log("Server.findPacker", "create new pack %p", blob)
|
debug.Log("Repo.findPacker", "create new pack %p", blob)
|
||||||
return pack.NewPacker(s.key, blob), nil
|
return pack.NewPacker(s.key, blob), nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// insertPacker appends p to s.packs.
|
// insertPacker appends p to s.packs.
|
||||||
func (s *Server) insertPacker(p *pack.Packer) {
|
func (s *Repo) insertPacker(p *pack.Packer) {
|
||||||
s.pm.Lock()
|
s.pm.Lock()
|
||||||
defer s.pm.Unlock()
|
defer s.pm.Unlock()
|
||||||
|
|
||||||
s.packs = append(s.packs, p)
|
s.packs = append(s.packs, p)
|
||||||
debug.Log("Server.insertPacker", "%d packers\n", len(s.packs))
|
debug.Log("Repo.insertPacker", "%d packers\n", len(s.packs))
|
||||||
}
|
}
|
||||||
|
|
||||||
// savePacker stores p in the backend.
|
// savePacker stores p in the backend.
|
||||||
func (s *Server) savePacker(p *pack.Packer) error {
|
func (s *Repo) savePacker(p *pack.Packer) error {
|
||||||
debug.Log("Server.savePacker", "save packer with %d blobs\n", p.Count())
|
debug.Log("Repo.savePacker", "save packer with %d blobs\n", p.Count())
|
||||||
_, err := p.Finalize()
|
_, err := p.Finalize()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
@ -262,15 +262,15 @@ func (s *Server) savePacker(p *pack.Packer) error {
|
|||||||
sid := p.ID()
|
sid := p.ID()
|
||||||
err = p.Writer().(backend.Blob).Finalize(backend.Data, sid.String())
|
err = p.Writer().(backend.Blob).Finalize(backend.Data, sid.String())
|
||||||
if err != nil {
|
if err != nil {
|
||||||
debug.Log("Server.savePacker", "blob Finalize() error: %v", err)
|
debug.Log("Repo.savePacker", "blob Finalize() error: %v", err)
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
debug.Log("Server.savePacker", "saved as %v", sid.Str())
|
debug.Log("Repo.savePacker", "saved as %v", sid.Str())
|
||||||
|
|
||||||
// update blobs in the index
|
// update blobs in the index
|
||||||
for _, b := range p.Blobs() {
|
for _, b := range p.Blobs() {
|
||||||
debug.Log("Server.savePacker", " updating blob %v to pack %v", b.ID.Str(), sid.Str())
|
debug.Log("Repo.savePacker", " updating blob %v to pack %v", b.ID.Str(), sid.Str())
|
||||||
s.idx.Store(b.Type, b.ID, sid, b.Offset, uint(b.Length))
|
s.idx.Store(b.Type, b.ID, sid, b.Offset, uint(b.Length))
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -278,7 +278,7 @@ func (s *Server) savePacker(p *pack.Packer) error {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// countPacker returns the number of open (unfinished) packers.
|
// countPacker returns the number of open (unfinished) packers.
|
||||||
func (s *Server) countPacker() int {
|
func (s *Repo) countPacker() int {
|
||||||
s.pm.Lock()
|
s.pm.Lock()
|
||||||
defer s.pm.Unlock()
|
defer s.pm.Unlock()
|
||||||
|
|
||||||
@ -287,13 +287,13 @@ func (s *Server) countPacker() int {
|
|||||||
|
|
||||||
// Save encrypts data and stores it to the backend as type t. If data is small
|
// Save encrypts data and stores it to the backend as type t. If data is small
|
||||||
// enough, it will be packed together with other small blobs.
|
// enough, it will be packed together with other small blobs.
|
||||||
func (s *Server) Save(t pack.BlobType, data []byte, id backend.ID) (backend.ID, error) {
|
func (s *Repo) Save(t pack.BlobType, data []byte, id backend.ID) (backend.ID, error) {
|
||||||
if id == nil {
|
if id == nil {
|
||||||
// compute plaintext hash
|
// compute plaintext hash
|
||||||
id = backend.Hash(data)
|
id = backend.Hash(data)
|
||||||
}
|
}
|
||||||
|
|
||||||
debug.Log("Server.Save", "save id %v (%v, %d bytes)", id.Str(), t, len(data))
|
debug.Log("Repo.Save", "save id %v (%v, %d bytes)", id.Str(), t, len(data))
|
||||||
|
|
||||||
// get buf from the pool
|
// get buf from the pool
|
||||||
ciphertext := getBuf()
|
ciphertext := getBuf()
|
||||||
@ -317,12 +317,12 @@ func (s *Server) Save(t pack.BlobType, data []byte, id backend.ID) (backend.ID,
|
|||||||
// add this id to the index, although we don't know yet in which pack it
|
// add this id to the index, although we don't know yet in which pack it
|
||||||
// will be saved, the entry will be updated when the pack is written.
|
// will be saved, the entry will be updated when the pack is written.
|
||||||
s.idx.Store(t, id, nil, 0, 0)
|
s.idx.Store(t, id, nil, 0, 0)
|
||||||
debug.Log("Server.Save", "saving stub for %v (%v) in index", id.Str, t)
|
debug.Log("Repo.Save", "saving stub for %v (%v) in index", id.Str, t)
|
||||||
|
|
||||||
// if the pack is not full enough and there are less than maxPackers
|
// if the pack is not full enough and there are less than maxPackers
|
||||||
// packers, put back to the list
|
// packers, put back to the list
|
||||||
if packer.Size() < minPackSize && s.countPacker() < maxPackers {
|
if packer.Size() < minPackSize && s.countPacker() < maxPackers {
|
||||||
debug.Log("Server.Save", "pack is not full enough (%d bytes)", packer.Size())
|
debug.Log("Repo.Save", "pack is not full enough (%d bytes)", packer.Size())
|
||||||
s.insertPacker(packer)
|
s.insertPacker(packer)
|
||||||
return id, nil
|
return id, nil
|
||||||
}
|
}
|
||||||
@ -332,8 +332,8 @@ func (s *Server) Save(t pack.BlobType, data []byte, id backend.ID) (backend.ID,
|
|||||||
}
|
}
|
||||||
|
|
||||||
// SaveFrom encrypts data read from rd and stores it in a pack in the backend as type t.
|
// SaveFrom encrypts data read from rd and stores it in a pack in the backend as type t.
|
||||||
func (s *Server) SaveFrom(t pack.BlobType, id backend.ID, length uint, rd io.Reader) error {
|
func (s *Repo) SaveFrom(t pack.BlobType, id backend.ID, length uint, rd io.Reader) error {
|
||||||
debug.Log("Server.SaveFrom", "save id %v (%v, %d bytes)", id.Str(), t, length)
|
debug.Log("Repo.SaveFrom", "save id %v (%v, %d bytes)", id.Str(), t, length)
|
||||||
if id == nil {
|
if id == nil {
|
||||||
return errors.New("id is nil")
|
return errors.New("id is nil")
|
||||||
}
|
}
|
||||||
@ -353,8 +353,8 @@ func (s *Server) SaveFrom(t pack.BlobType, id backend.ID, length uint, rd io.Rea
|
|||||||
|
|
||||||
// SaveJSON serialises item as JSON and encrypts and saves it in a pack in the
|
// SaveJSON serialises item as JSON and encrypts and saves it in a pack in the
|
||||||
// backend as type t.
|
// backend as type t.
|
||||||
func (s *Server) SaveJSON(t pack.BlobType, item interface{}) (backend.ID, error) {
|
func (s *Repo) SaveJSON(t pack.BlobType, item interface{}) (backend.ID, error) {
|
||||||
debug.Log("Server.SaveJSON", "save %v blob", t)
|
debug.Log("Repo.SaveJSON", "save %v blob", t)
|
||||||
buf := getBuf()[:0]
|
buf := getBuf()[:0]
|
||||||
defer freeBuf(buf)
|
defer freeBuf(buf)
|
||||||
|
|
||||||
@ -372,13 +372,13 @@ func (s *Server) SaveJSON(t pack.BlobType, item interface{}) (backend.ID, error)
|
|||||||
|
|
||||||
// SaveJSONUnpacked serialises item as JSON and encrypts and saves it in the
|
// SaveJSONUnpacked serialises item as JSON and encrypts and saves it in the
|
||||||
// backend as type t, without a pack. It returns the storage hash.
|
// backend as type t, without a pack. It returns the storage hash.
|
||||||
func (s *Server) SaveJSONUnpacked(t backend.Type, item interface{}) (backend.ID, error) {
|
func (s *Repo) SaveJSONUnpacked(t backend.Type, item interface{}) (backend.ID, error) {
|
||||||
// create file
|
// create file
|
||||||
blob, err := s.be.Create()
|
blob, err := s.be.Create()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
debug.Log("Server.SaveJSONUnpacked", "create new file %p", blob)
|
debug.Log("Repo.SaveJSONUnpacked", "create new file %p", blob)
|
||||||
|
|
||||||
// hash
|
// hash
|
||||||
hw := backend.NewHashingWriter(blob, sha256.New())
|
hw := backend.NewHashingWriter(blob, sha256.New())
|
||||||
@ -409,11 +409,11 @@ func (s *Server) SaveJSONUnpacked(t backend.Type, item interface{}) (backend.ID,
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Flush saves all remaining packs.
|
// Flush saves all remaining packs.
|
||||||
func (s *Server) Flush() error {
|
func (s *Repo) Flush() error {
|
||||||
s.pm.Lock()
|
s.pm.Lock()
|
||||||
defer s.pm.Unlock()
|
defer s.pm.Unlock()
|
||||||
|
|
||||||
debug.Log("Server.Flush", "manually flushing %d packs", len(s.packs))
|
debug.Log("Repo.Flush", "manually flushing %d packs", len(s.packs))
|
||||||
|
|
||||||
for _, p := range s.packs {
|
for _, p := range s.packs {
|
||||||
err := s.savePacker(p)
|
err := s.savePacker(p)
|
||||||
@ -426,23 +426,23 @@ func (s *Server) Flush() error {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *Server) Backend() backend.Backend {
|
func (s *Repo) Backend() backend.Backend {
|
||||||
return s.be
|
return s.be
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *Server) Index() *Index {
|
func (s *Repo) Index() *Index {
|
||||||
return s.idx
|
return s.idx
|
||||||
}
|
}
|
||||||
|
|
||||||
// SetIndex instructs the server to use the given index.
|
// SetIndex instructs the repository to use the given index.
|
||||||
func (s *Server) SetIndex(i *Index) {
|
func (s *Repo) SetIndex(i *Index) {
|
||||||
s.idx = i
|
s.idx = i
|
||||||
}
|
}
|
||||||
|
|
||||||
// SaveIndex saves all new packs in the index in the backend, returned is the
|
// SaveIndex saves all new packs in the index in the backend, returned is the
|
||||||
// storage ID.
|
// storage ID.
|
||||||
func (s *Server) SaveIndex() (backend.ID, error) {
|
func (s *Repo) SaveIndex() (backend.ID, error) {
|
||||||
debug.Log("Server.SaveIndex", "Saving index")
|
debug.Log("Repo.SaveIndex", "Saving index")
|
||||||
|
|
||||||
// create blob
|
// create blob
|
||||||
blob, err := s.be.Create()
|
blob, err := s.be.Create()
|
||||||
@ -450,7 +450,7 @@ func (s *Server) SaveIndex() (backend.ID, error) {
|
|||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
debug.Log("Server.SaveIndex", "create new pack %p", blob)
|
debug.Log("Repo.SaveIndex", "create new pack %p", blob)
|
||||||
|
|
||||||
// hash
|
// hash
|
||||||
hw := backend.NewHashingWriter(blob, sha256.New())
|
hw := backend.NewHashingWriter(blob, sha256.New())
|
||||||
@ -476,15 +476,15 @@ func (s *Server) SaveIndex() (backend.ID, error) {
|
|||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
debug.Log("Server.SaveIndex", "Saved index as %v", sid.Str())
|
debug.Log("Repo.SaveIndex", "Saved index as %v", sid.Str())
|
||||||
|
|
||||||
return sid, nil
|
return sid, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// LoadIndex loads all index files from the backend and merges them with the
|
// LoadIndex loads all index files from the backend and merges them with the
|
||||||
// current index.
|
// current index.
|
||||||
func (s *Server) LoadIndex() error {
|
func (s *Repo) LoadIndex() error {
|
||||||
debug.Log("Server.LoadIndex", "Loading index")
|
debug.Log("Repo.LoadIndex", "Loading index")
|
||||||
done := make(chan struct{})
|
done := make(chan struct{})
|
||||||
defer close(done)
|
defer close(done)
|
||||||
|
|
||||||
@ -498,8 +498,8 @@ func (s *Server) LoadIndex() error {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// loadIndex loads the index id and merges it with the currently used index.
|
// loadIndex loads the index id and merges it with the currently used index.
|
||||||
func (s *Server) loadIndex(id string) error {
|
func (s *Repo) loadIndex(id string) error {
|
||||||
debug.Log("Server.loadIndex", "Loading index %v", id[:8])
|
debug.Log("Repo.loadIndex", "Loading index %v", id[:8])
|
||||||
before := len(s.idx.pack)
|
before := len(s.idx.pack)
|
||||||
|
|
||||||
rd, err := s.be.Get(backend.Index, id)
|
rd, err := s.be.Get(backend.Index, id)
|
||||||
@ -517,22 +517,22 @@ func (s *Server) loadIndex(id string) error {
|
|||||||
|
|
||||||
idx, err := DecodeIndex(decryptRd)
|
idx, err := DecodeIndex(decryptRd)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
debug.Log("Server.loadIndex", "error while decoding index %v: %v", id, err)
|
debug.Log("Repo.loadIndex", "error while decoding index %v: %v", id, err)
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
s.idx.Merge(idx)
|
s.idx.Merge(idx)
|
||||||
|
|
||||||
after := len(s.idx.pack)
|
after := len(s.idx.pack)
|
||||||
debug.Log("Server.loadIndex", "Loaded index %v, added %v blobs", id[:8], after-before)
|
debug.Log("Repo.loadIndex", "Loaded index %v, added %v blobs", id[:8], after-before)
|
||||||
|
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
const repositoryIDSize = sha256.Size
|
const repositoryIDSize = sha256.Size
|
||||||
const RepositoryVersion = 1
|
const RepoVersion = 1
|
||||||
|
|
||||||
func createConfig(s *Server) (err error) {
|
func createConfig(s *Repo) (err error) {
|
||||||
s.Config.ChunkerPolynomial, err = chunker.RandomPolynomial()
|
s.Config.ChunkerPolynomial, err = chunker.RandomPolynomial()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
@ -545,21 +545,21 @@ func createConfig(s *Server) (err error) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
s.Config.ID = hex.EncodeToString(newID)
|
s.Config.ID = hex.EncodeToString(newID)
|
||||||
s.Config.Version = RepositoryVersion
|
s.Config.Version = RepoVersion
|
||||||
|
|
||||||
debug.Log("Server.createConfig", "New config: %#v", s.Config)
|
debug.Log("Repo.createConfig", "New config: %#v", s.Config)
|
||||||
|
|
||||||
_, err = s.SaveJSONUnpacked(backend.Config, s.Config)
|
_, err = s.SaveJSONUnpacked(backend.Config, s.Config)
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *Server) loadConfig(cfg *Config) error {
|
func (s *Repo) loadConfig(cfg *Config) error {
|
||||||
err := s.LoadJSONUnpacked(backend.Config, nil, cfg)
|
err := s.LoadJSONUnpacked(backend.Config, nil, cfg)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
if cfg.Version != RepositoryVersion {
|
if cfg.Version != RepoVersion {
|
||||||
return errors.New("unsupported repository version")
|
return errors.New("unsupported repository version")
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -572,7 +572,7 @@ func (s *Server) loadConfig(cfg *Config) error {
|
|||||||
|
|
||||||
// SearchKey finds a key with the supplied password, afterwards the config is
|
// SearchKey finds a key with the supplied password, afterwards the config is
|
||||||
// read and parsed.
|
// read and parsed.
|
||||||
func (s *Server) SearchKey(password string) error {
|
func (s *Repo) SearchKey(password string) error {
|
||||||
key, err := SearchKey(s, password)
|
key, err := SearchKey(s, password)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
@ -585,7 +585,7 @@ func (s *Server) SearchKey(password string) error {
|
|||||||
|
|
||||||
// Init creates a new master key with the supplied password and initializes the
|
// Init creates a new master key with the supplied password and initializes the
|
||||||
// repository config.
|
// repository config.
|
||||||
func (s *Server) Init(password string) error {
|
func (s *Repo) Init(password string) error {
|
||||||
has, err := s.Test(backend.Config, "")
|
has, err := s.Test(backend.Config, "")
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
@ -604,32 +604,32 @@ func (s *Server) Init(password string) error {
|
|||||||
return createConfig(s)
|
return createConfig(s)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *Server) Decrypt(ciphertext []byte) ([]byte, error) {
|
func (s *Repo) Decrypt(ciphertext []byte) ([]byte, error) {
|
||||||
if s.key == nil {
|
if s.key == nil {
|
||||||
return nil, errors.New("key for server not set")
|
return nil, errors.New("key for repository not set")
|
||||||
}
|
}
|
||||||
|
|
||||||
return crypto.Decrypt(s.key, nil, ciphertext)
|
return crypto.Decrypt(s.key, nil, ciphertext)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *Server) Encrypt(ciphertext, plaintext []byte) ([]byte, error) {
|
func (s *Repo) Encrypt(ciphertext, plaintext []byte) ([]byte, error) {
|
||||||
if s.key == nil {
|
if s.key == nil {
|
||||||
return nil, errors.New("key for server not set")
|
return nil, errors.New("key for repository not set")
|
||||||
}
|
}
|
||||||
|
|
||||||
return crypto.Encrypt(s.key, ciphertext, plaintext)
|
return crypto.Encrypt(s.key, ciphertext, plaintext)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *Server) Key() *crypto.Key {
|
func (s *Repo) Key() *crypto.Key {
|
||||||
return s.key
|
return s.key
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *Server) KeyName() string {
|
func (s *Repo) KeyName() string {
|
||||||
return s.keyName
|
return s.keyName
|
||||||
}
|
}
|
||||||
|
|
||||||
// Count returns the number of blobs of a given type in the backend.
|
// Count returns the number of blobs of a given type in the backend.
|
||||||
func (s *Server) Count(t backend.Type) (n uint) {
|
func (s *Repo) Count(t backend.Type) (n uint) {
|
||||||
for _ = range s.be.List(t, nil) {
|
for _ = range s.be.List(t, nil) {
|
||||||
n++
|
n++
|
||||||
}
|
}
|
||||||
@ -639,27 +639,27 @@ func (s *Server) Count(t backend.Type) (n uint) {
|
|||||||
|
|
||||||
// Proxy methods to backend
|
// Proxy methods to backend
|
||||||
|
|
||||||
func (s *Server) Get(t backend.Type, name string) (io.ReadCloser, error) {
|
func (s *Repo) Get(t backend.Type, name string) (io.ReadCloser, error) {
|
||||||
return s.be.Get(t, name)
|
return s.be.Get(t, name)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *Server) List(t backend.Type, done <-chan struct{}) <-chan string {
|
func (s *Repo) List(t backend.Type, done <-chan struct{}) <-chan string {
|
||||||
return s.be.List(t, done)
|
return s.be.List(t, done)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *Server) Test(t backend.Type, name string) (bool, error) {
|
func (s *Repo) Test(t backend.Type, name string) (bool, error) {
|
||||||
return s.be.Test(t, name)
|
return s.be.Test(t, name)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *Server) Remove(t backend.Type, name string) error {
|
func (s *Repo) Remove(t backend.Type, name string) error {
|
||||||
return s.be.Remove(t, name)
|
return s.be.Remove(t, name)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *Server) Close() error {
|
func (s *Repo) Close() error {
|
||||||
return s.be.Close()
|
return s.be.Close()
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *Server) Delete() error {
|
func (s *Repo) Delete() error {
|
||||||
if b, ok := s.be.(backend.Deleter); ok {
|
if b, ok := s.be.(backend.Deleter); ok {
|
||||||
return b.Delete()
|
return b.Delete()
|
||||||
}
|
}
|
||||||
@ -667,6 +667,6 @@ func (s *Server) Delete() error {
|
|||||||
return errors.New("Delete() called for backend that does not implement this method")
|
return errors.New("Delete() called for backend that does not implement this method")
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *Server) Location() string {
|
func (s *Repo) Location() string {
|
||||||
return s.be.Location()
|
return s.be.Location()
|
||||||
}
|
}
|
@ -1,4 +1,4 @@
|
|||||||
package server_test
|
package repo_test
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"bytes"
|
"bytes"
|
||||||
@ -23,21 +23,21 @@ type testJSONStruct struct {
|
|||||||
Baz []byte
|
Baz []byte
|
||||||
}
|
}
|
||||||
|
|
||||||
var serverTests = []testJSONStruct{
|
var repoTests = []testJSONStruct{
|
||||||
testJSONStruct{Foo: 23, Bar: "Teststring", Baz: []byte("xx")},
|
testJSONStruct{Foo: 23, Bar: "Teststring", Baz: []byte("xx")},
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestSaveJSON(t *testing.T) {
|
func TestSaveJSON(t *testing.T) {
|
||||||
server := SetupBackend(t)
|
repo := SetupRepo(t)
|
||||||
defer TeardownBackend(t, server)
|
defer TeardownRepo(t, repo)
|
||||||
|
|
||||||
for _, obj := range serverTests {
|
for _, obj := range repoTests {
|
||||||
data, err := json.Marshal(obj)
|
data, err := json.Marshal(obj)
|
||||||
OK(t, err)
|
OK(t, err)
|
||||||
data = append(data, '\n')
|
data = append(data, '\n')
|
||||||
h := sha256.Sum256(data)
|
h := sha256.Sum256(data)
|
||||||
|
|
||||||
id, err := server.SaveJSON(pack.Tree, obj)
|
id, err := repo.SaveJSON(pack.Tree, obj)
|
||||||
OK(t, err)
|
OK(t, err)
|
||||||
|
|
||||||
Assert(t, bytes.Equal(h[:], id),
|
Assert(t, bytes.Equal(h[:], id),
|
||||||
@ -47,10 +47,10 @@ func TestSaveJSON(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func BenchmarkSaveJSON(t *testing.B) {
|
func BenchmarkSaveJSON(t *testing.B) {
|
||||||
server := SetupBackend(t)
|
repo := SetupRepo(t)
|
||||||
defer TeardownBackend(t, server)
|
defer TeardownRepo(t, repo)
|
||||||
|
|
||||||
obj := serverTests[0]
|
obj := repoTests[0]
|
||||||
|
|
||||||
data, err := json.Marshal(obj)
|
data, err := json.Marshal(obj)
|
||||||
OK(t, err)
|
OK(t, err)
|
||||||
@ -60,7 +60,7 @@ func BenchmarkSaveJSON(t *testing.B) {
|
|||||||
t.ResetTimer()
|
t.ResetTimer()
|
||||||
|
|
||||||
for i := 0; i < t.N; i++ {
|
for i := 0; i < t.N; i++ {
|
||||||
id, err := server.SaveJSON(pack.Tree, obj)
|
id, err := repo.SaveJSON(pack.Tree, obj)
|
||||||
OK(t, err)
|
OK(t, err)
|
||||||
|
|
||||||
Assert(t, bytes.Equal(h[:], id),
|
Assert(t, bytes.Equal(h[:], id),
|
||||||
@ -72,8 +72,8 @@ func BenchmarkSaveJSON(t *testing.B) {
|
|||||||
var testSizes = []int{5, 23, 2<<18 + 23, 1 << 20}
|
var testSizes = []int{5, 23, 2<<18 + 23, 1 << 20}
|
||||||
|
|
||||||
func TestSave(t *testing.T) {
|
func TestSave(t *testing.T) {
|
||||||
server := SetupBackend(t)
|
repo := SetupRepo(t)
|
||||||
defer TeardownBackend(t, server)
|
defer TeardownRepo(t, repo)
|
||||||
|
|
||||||
for _, size := range testSizes {
|
for _, size := range testSizes {
|
||||||
data := make([]byte, size)
|
data := make([]byte, size)
|
||||||
@ -83,15 +83,15 @@ func TestSave(t *testing.T) {
|
|||||||
id := backend.Hash(data)
|
id := backend.Hash(data)
|
||||||
|
|
||||||
// save
|
// save
|
||||||
sid, err := server.Save(pack.Data, data, nil)
|
sid, err := repo.Save(pack.Data, data, nil)
|
||||||
OK(t, err)
|
OK(t, err)
|
||||||
|
|
||||||
Equals(t, id, sid)
|
Equals(t, id, sid)
|
||||||
|
|
||||||
OK(t, server.Flush())
|
OK(t, repo.Flush())
|
||||||
|
|
||||||
// read back
|
// read back
|
||||||
buf, err := server.LoadBlob(pack.Data, id)
|
buf, err := repo.LoadBlob(pack.Data, id)
|
||||||
|
|
||||||
Assert(t, len(buf) == len(data),
|
Assert(t, len(buf) == len(data),
|
||||||
"number of bytes read back does not match: expected %d, got %d",
|
"number of bytes read back does not match: expected %d, got %d",
|
||||||
@ -104,8 +104,8 @@ func TestSave(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func TestSaveFrom(t *testing.T) {
|
func TestSaveFrom(t *testing.T) {
|
||||||
server := SetupBackend(t)
|
repo := SetupRepo(t)
|
||||||
defer TeardownBackend(t, server)
|
defer TeardownRepo(t, repo)
|
||||||
|
|
||||||
for _, size := range testSizes {
|
for _, size := range testSizes {
|
||||||
data := make([]byte, size)
|
data := make([]byte, size)
|
||||||
@ -115,13 +115,13 @@ func TestSaveFrom(t *testing.T) {
|
|||||||
id := backend.Hash(data)
|
id := backend.Hash(data)
|
||||||
|
|
||||||
// save
|
// save
|
||||||
err = server.SaveFrom(pack.Data, id[:], uint(size), bytes.NewReader(data))
|
err = repo.SaveFrom(pack.Data, id[:], uint(size), bytes.NewReader(data))
|
||||||
OK(t, err)
|
OK(t, err)
|
||||||
|
|
||||||
OK(t, server.Flush())
|
OK(t, repo.Flush())
|
||||||
|
|
||||||
// read back
|
// read back
|
||||||
buf, err := server.LoadBlob(pack.Data, id[:])
|
buf, err := repo.LoadBlob(pack.Data, id[:])
|
||||||
|
|
||||||
Assert(t, len(buf) == len(data),
|
Assert(t, len(buf) == len(data),
|
||||||
"number of bytes read back does not match: expected %d, got %d",
|
"number of bytes read back does not match: expected %d, got %d",
|
||||||
@ -134,8 +134,8 @@ func TestSaveFrom(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func BenchmarkSaveFrom(t *testing.B) {
|
func BenchmarkSaveFrom(t *testing.B) {
|
||||||
server := SetupBackend(t)
|
repo := SetupRepo(t)
|
||||||
defer TeardownBackend(t, server)
|
defer TeardownRepo(t, repo)
|
||||||
|
|
||||||
size := 4 << 20 // 4MiB
|
size := 4 << 20 // 4MiB
|
||||||
|
|
||||||
@ -150,48 +150,48 @@ func BenchmarkSaveFrom(t *testing.B) {
|
|||||||
|
|
||||||
for i := 0; i < t.N; i++ {
|
for i := 0; i < t.N; i++ {
|
||||||
// save
|
// save
|
||||||
err = server.SaveFrom(pack.Data, id[:], uint(size), bytes.NewReader(data))
|
err = repo.SaveFrom(pack.Data, id[:], uint(size), bytes.NewReader(data))
|
||||||
OK(t, err)
|
OK(t, err)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestLoadJSONPack(t *testing.T) {
|
func TestLoadJSONPack(t *testing.T) {
|
||||||
if *benchTestDir == "" {
|
if *benchTestDir == "" {
|
||||||
t.Skip("benchdir not set, skipping TestServerStats")
|
t.Skip("benchdir not set, skipping")
|
||||||
}
|
}
|
||||||
|
|
||||||
server := SetupBackend(t)
|
repo := SetupRepo(t)
|
||||||
defer TeardownBackend(t, server)
|
defer TeardownRepo(t, repo)
|
||||||
|
|
||||||
// archive a few files
|
// archive a few files
|
||||||
sn := SnapshotDir(t, server, *benchTestDir, nil)
|
sn := SnapshotDir(t, repo, *benchTestDir, nil)
|
||||||
OK(t, server.Flush())
|
OK(t, repo.Flush())
|
||||||
|
|
||||||
tree := restic.NewTree()
|
tree := restic.NewTree()
|
||||||
err := server.LoadJSONPack(pack.Tree, sn.Tree, &tree)
|
err := repo.LoadJSONPack(pack.Tree, sn.Tree, &tree)
|
||||||
OK(t, err)
|
OK(t, err)
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestLoadJSONUnpacked(t *testing.T) {
|
func TestLoadJSONUnpacked(t *testing.T) {
|
||||||
if *benchTestDir == "" {
|
if *benchTestDir == "" {
|
||||||
t.Skip("benchdir not set, skipping TestServerStats")
|
t.Skip("benchdir not set, skipping")
|
||||||
}
|
}
|
||||||
|
|
||||||
server := SetupBackend(t)
|
repo := SetupRepo(t)
|
||||||
defer TeardownBackend(t, server)
|
defer TeardownRepo(t, repo)
|
||||||
|
|
||||||
// archive a snapshot
|
// archive a snapshot
|
||||||
sn := restic.Snapshot{}
|
sn := restic.Snapshot{}
|
||||||
sn.Hostname = "foobar"
|
sn.Hostname = "foobar"
|
||||||
sn.Username = "test!"
|
sn.Username = "test!"
|
||||||
|
|
||||||
id, err := server.SaveJSONUnpacked(backend.Snapshot, &sn)
|
id, err := repo.SaveJSONUnpacked(backend.Snapshot, &sn)
|
||||||
OK(t, err)
|
OK(t, err)
|
||||||
|
|
||||||
var sn2 restic.Snapshot
|
var sn2 restic.Snapshot
|
||||||
|
|
||||||
// restore
|
// restore
|
||||||
err = server.LoadJSONUnpacked(backend.Snapshot, id, &sn2)
|
err = repo.LoadJSONUnpacked(backend.Snapshot, id, &sn2)
|
||||||
OK(t, err)
|
OK(t, err)
|
||||||
|
|
||||||
Equals(t, sn.Hostname, sn2.Hostname)
|
Equals(t, sn.Hostname, sn2.Hostname)
|
18
restorer.go
18
restorer.go
@ -7,15 +7,15 @@ import (
|
|||||||
"syscall"
|
"syscall"
|
||||||
|
|
||||||
"github.com/restic/restic/backend"
|
"github.com/restic/restic/backend"
|
||||||
"github.com/restic/restic/server"
|
"github.com/restic/restic/repo"
|
||||||
|
|
||||||
"github.com/juju/errors"
|
"github.com/juju/errors"
|
||||||
)
|
)
|
||||||
|
|
||||||
// Restorer is used to restore a snapshot to a directory.
|
// Restorer is used to restore a snapshot to a directory.
|
||||||
type Restorer struct {
|
type Restorer struct {
|
||||||
s *server.Server
|
repo *repo.Repo
|
||||||
sn *Snapshot
|
sn *Snapshot
|
||||||
|
|
||||||
Error func(dir string, node *Node, err error) error
|
Error func(dir string, node *Node, err error) error
|
||||||
Filter func(item string, dstpath string, node *Node) bool
|
Filter func(item string, dstpath string, node *Node) bool
|
||||||
@ -24,12 +24,12 @@ type Restorer struct {
|
|||||||
var restorerAbortOnAllErrors = func(str string, node *Node, err error) error { return err }
|
var restorerAbortOnAllErrors = func(str string, node *Node, err error) error { return err }
|
||||||
|
|
||||||
// NewRestorer creates a restorer preloaded with the content from the snapshot id.
|
// NewRestorer creates a restorer preloaded with the content from the snapshot id.
|
||||||
func NewRestorer(s *server.Server, id backend.ID) (*Restorer, error) {
|
func NewRestorer(repo *repo.Repo, id backend.ID) (*Restorer, error) {
|
||||||
r := &Restorer{s: s, Error: restorerAbortOnAllErrors}
|
r := &Restorer{repo: repo, Error: restorerAbortOnAllErrors}
|
||||||
|
|
||||||
var err error
|
var err error
|
||||||
|
|
||||||
r.sn, err = LoadSnapshot(s, id)
|
r.sn, err = LoadSnapshot(repo, id)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, errors.Annotate(err, "load snapshot for restorer")
|
return nil, errors.Annotate(err, "load snapshot for restorer")
|
||||||
}
|
}
|
||||||
@ -38,7 +38,7 @@ func NewRestorer(s *server.Server, id backend.ID) (*Restorer, error) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func (res *Restorer) restoreTo(dst string, dir string, treeID backend.ID) error {
|
func (res *Restorer) restoreTo(dst string, dir string, treeID backend.ID) error {
|
||||||
tree, err := LoadTree(res.s, treeID)
|
tree, err := LoadTree(res.repo, treeID)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return res.Error(dir, nil, errors.Annotate(err, "LoadTree"))
|
return res.Error(dir, nil, errors.Annotate(err, "LoadTree"))
|
||||||
}
|
}
|
||||||
@ -74,7 +74,7 @@ func (res *Restorer) restoreNodeTo(node *Node, dir string, dst string) error {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
err := node.CreateAt(dstPath, res.s)
|
err := node.CreateAt(dstPath, res.repo)
|
||||||
|
|
||||||
// Did it fail because of ENOENT?
|
// Did it fail because of ENOENT?
|
||||||
if pe, ok := errors.Cause(err).(*os.PathError); ok {
|
if pe, ok := errors.Cause(err).(*os.PathError); ok {
|
||||||
@ -83,7 +83,7 @@ func (res *Restorer) restoreNodeTo(node *Node, dir string, dst string) error {
|
|||||||
// Create parent directories and retry
|
// Create parent directories and retry
|
||||||
err = os.MkdirAll(filepath.Dir(dstPath), 0700)
|
err = os.MkdirAll(filepath.Dir(dstPath), 0700)
|
||||||
if err == nil || err == os.ErrExist {
|
if err == nil || err == os.ErrExist {
|
||||||
err = node.CreateAt(dstPath, res.s)
|
err = node.CreateAt(dstPath, res.repo)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -1,2 +0,0 @@
|
|||||||
// Package server implements a restic repository on top of a backend.
|
|
||||||
package server
|
|
@ -9,7 +9,7 @@ import (
|
|||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/restic/restic/backend"
|
"github.com/restic/restic/backend"
|
||||||
"github.com/restic/restic/server"
|
"github.com/restic/restic/repo"
|
||||||
)
|
)
|
||||||
|
|
||||||
type Snapshot struct {
|
type Snapshot struct {
|
||||||
@ -50,9 +50,9 @@ func NewSnapshot(paths []string) (*Snapshot, error) {
|
|||||||
return sn, nil
|
return sn, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func LoadSnapshot(s *server.Server, id backend.ID) (*Snapshot, error) {
|
func LoadSnapshot(repo *repo.Repo, id backend.ID) (*Snapshot, error) {
|
||||||
sn := &Snapshot{id: id}
|
sn := &Snapshot{id: id}
|
||||||
err := s.LoadJSONUnpacked(backend.Snapshot, id, sn)
|
err := repo.LoadJSONUnpacked(backend.Snapshot, id, sn)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
@ -8,8 +8,8 @@ import (
|
|||||||
)
|
)
|
||||||
|
|
||||||
func TestNewSnapshot(t *testing.T) {
|
func TestNewSnapshot(t *testing.T) {
|
||||||
s := SetupBackend(t)
|
s := SetupRepo(t)
|
||||||
defer TeardownBackend(t, s)
|
defer TeardownRepo(t, s)
|
||||||
|
|
||||||
paths := []string{"/home/foobar"}
|
paths := []string{"/home/foobar"}
|
||||||
|
|
||||||
|
@ -10,14 +10,14 @@ import (
|
|||||||
"github.com/restic/restic"
|
"github.com/restic/restic"
|
||||||
"github.com/restic/restic/backend"
|
"github.com/restic/restic/backend"
|
||||||
"github.com/restic/restic/backend/local"
|
"github.com/restic/restic/backend/local"
|
||||||
"github.com/restic/restic/server"
|
"github.com/restic/restic/repo"
|
||||||
)
|
)
|
||||||
|
|
||||||
var TestPassword = flag.String("test.password", "geheim", `use this password for repositories created during tests (default: "geheim")`)
|
var TestPassword = flag.String("test.password", "geheim", `use this password for repositories created during tests (default: "geheim")`)
|
||||||
var TestCleanup = flag.Bool("test.cleanup", true, "clean up after running tests (remove local backend directory with all content)")
|
var TestCleanup = flag.Bool("test.cleanup", true, "clean up after running tests (remove local backend directory with all content)")
|
||||||
var TestTempDir = flag.String("test.tempdir", "", "use this directory for temporary storage (default: system temp dir)")
|
var TestTempDir = flag.String("test.tempdir", "", "use this directory for temporary storage (default: system temp dir)")
|
||||||
|
|
||||||
func SetupBackend(t testing.TB) *server.Server {
|
func SetupRepo(t testing.TB) *repo.Repo {
|
||||||
tempdir, err := ioutil.TempDir(*TestTempDir, "restic-test-")
|
tempdir, err := ioutil.TempDir(*TestTempDir, "restic-test-")
|
||||||
OK(t, err)
|
OK(t, err)
|
||||||
|
|
||||||
@ -29,23 +29,23 @@ func SetupBackend(t testing.TB) *server.Server {
|
|||||||
err = os.Setenv("RESTIC_CACHE", filepath.Join(tempdir, "cache"))
|
err = os.Setenv("RESTIC_CACHE", filepath.Join(tempdir, "cache"))
|
||||||
OK(t, err)
|
OK(t, err)
|
||||||
|
|
||||||
s := server.NewServer(b)
|
repo := repo.New(b)
|
||||||
OK(t, s.Init(*TestPassword))
|
OK(t, repo.Init(*TestPassword))
|
||||||
return s
|
return repo
|
||||||
}
|
}
|
||||||
|
|
||||||
func TeardownBackend(t testing.TB, s *server.Server) {
|
func TeardownRepo(t testing.TB, repo *repo.Repo) {
|
||||||
if !*TestCleanup {
|
if !*TestCleanup {
|
||||||
l := s.Backend().(*local.Local)
|
l := repo.Backend().(*local.Local)
|
||||||
t.Logf("leaving local backend at %s\n", l.Location())
|
t.Logf("leaving local backend at %s\n", l.Location())
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
OK(t, s.Delete())
|
OK(t, repo.Delete())
|
||||||
}
|
}
|
||||||
|
|
||||||
func SnapshotDir(t testing.TB, server *server.Server, path string, parent backend.ID) *restic.Snapshot {
|
func SnapshotDir(t testing.TB, repo *repo.Repo, path string, parent backend.ID) *restic.Snapshot {
|
||||||
arch := restic.NewArchiver(server)
|
arch := restic.NewArchiver(repo)
|
||||||
sn, _, err := arch.Snapshot(nil, []string{path}, parent)
|
sn, _, err := arch.Snapshot(nil, []string{path}, parent)
|
||||||
OK(t, err)
|
OK(t, err)
|
||||||
return sn
|
return sn
|
||||||
|
6
tree.go
6
tree.go
@ -8,7 +8,7 @@ import (
|
|||||||
"github.com/restic/restic/backend"
|
"github.com/restic/restic/backend"
|
||||||
"github.com/restic/restic/debug"
|
"github.com/restic/restic/debug"
|
||||||
"github.com/restic/restic/pack"
|
"github.com/restic/restic/pack"
|
||||||
"github.com/restic/restic/server"
|
"github.com/restic/restic/repo"
|
||||||
)
|
)
|
||||||
|
|
||||||
type Tree struct {
|
type Tree struct {
|
||||||
@ -30,9 +30,9 @@ func (t Tree) String() string {
|
|||||||
return fmt.Sprintf("Tree<%d nodes>", len(t.Nodes))
|
return fmt.Sprintf("Tree<%d nodes>", len(t.Nodes))
|
||||||
}
|
}
|
||||||
|
|
||||||
func LoadTree(s *server.Server, id backend.ID) (*Tree, error) {
|
func LoadTree(repo *repo.Repo, id backend.ID) (*Tree, error) {
|
||||||
tree := &Tree{}
|
tree := &Tree{}
|
||||||
err := s.LoadJSONPack(pack.Tree, id, tree)
|
err := repo.LoadJSONPack(pack.Tree, id, tree)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
10
tree_test.go
10
tree_test.go
@ -93,19 +93,19 @@ func TestNodeComparison(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func TestLoadTree(t *testing.T) {
|
func TestLoadTree(t *testing.T) {
|
||||||
server := SetupBackend(t)
|
repo := SetupRepo(t)
|
||||||
defer TeardownBackend(t, server)
|
defer TeardownRepo(t, repo)
|
||||||
|
|
||||||
// save tree
|
// save tree
|
||||||
tree := restic.NewTree()
|
tree := restic.NewTree()
|
||||||
id, err := server.SaveJSON(pack.Tree, tree)
|
id, err := repo.SaveJSON(pack.Tree, tree)
|
||||||
OK(t, err)
|
OK(t, err)
|
||||||
|
|
||||||
// save packs
|
// save packs
|
||||||
OK(t, server.Flush())
|
OK(t, repo.Flush())
|
||||||
|
|
||||||
// load tree again
|
// load tree again
|
||||||
tree2, err := restic.LoadTree(server, id)
|
tree2, err := restic.LoadTree(repo, id)
|
||||||
OK(t, err)
|
OK(t, err)
|
||||||
|
|
||||||
Assert(t, tree.Equals(tree2),
|
Assert(t, tree.Equals(tree2),
|
||||||
|
12
walk.go
12
walk.go
@ -5,7 +5,7 @@ import (
|
|||||||
|
|
||||||
"github.com/restic/restic/backend"
|
"github.com/restic/restic/backend"
|
||||||
"github.com/restic/restic/debug"
|
"github.com/restic/restic/debug"
|
||||||
"github.com/restic/restic/server"
|
"github.com/restic/restic/repo"
|
||||||
)
|
)
|
||||||
|
|
||||||
type WalkTreeJob struct {
|
type WalkTreeJob struct {
|
||||||
@ -16,10 +16,10 @@ type WalkTreeJob struct {
|
|||||||
Tree *Tree
|
Tree *Tree
|
||||||
}
|
}
|
||||||
|
|
||||||
func walkTree(s *server.Server, path string, treeID backend.ID, done chan struct{}, jobCh chan<- WalkTreeJob) {
|
func walkTree(repo *repo.Repo, path string, treeID backend.ID, done chan struct{}, jobCh chan<- WalkTreeJob) {
|
||||||
debug.Log("walkTree", "start on %q (%v)", path, treeID.Str())
|
debug.Log("walkTree", "start on %q (%v)", path, treeID.Str())
|
||||||
|
|
||||||
t, err := LoadTree(s, treeID)
|
t, err := LoadTree(repo, treeID)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
jobCh <- WalkTreeJob{Path: path, Error: err}
|
jobCh <- WalkTreeJob{Path: path, Error: err}
|
||||||
return
|
return
|
||||||
@ -28,7 +28,7 @@ func walkTree(s *server.Server, path string, treeID backend.ID, done chan struct
|
|||||||
for _, node := range t.Nodes {
|
for _, node := range t.Nodes {
|
||||||
p := filepath.Join(path, node.Name)
|
p := filepath.Join(path, node.Name)
|
||||||
if node.Type == "dir" {
|
if node.Type == "dir" {
|
||||||
walkTree(s, p, node.Subtree, done, jobCh)
|
walkTree(repo, p, node.Subtree, done, jobCh)
|
||||||
} else {
|
} else {
|
||||||
jobCh <- WalkTreeJob{Path: p, Node: node}
|
jobCh <- WalkTreeJob{Path: p, Node: node}
|
||||||
}
|
}
|
||||||
@ -41,9 +41,9 @@ func walkTree(s *server.Server, path string, treeID backend.ID, done chan struct
|
|||||||
// WalkTree walks the tree specified by id recursively and sends a job for each
|
// WalkTree walks the tree specified by id recursively and sends a job for each
|
||||||
// file and directory it finds. When the channel done is closed, processing
|
// file and directory it finds. When the channel done is closed, processing
|
||||||
// stops.
|
// stops.
|
||||||
func WalkTree(server *server.Server, id backend.ID, done chan struct{}, jobCh chan<- WalkTreeJob) {
|
func WalkTree(repo *repo.Repo, id backend.ID, done chan struct{}, jobCh chan<- WalkTreeJob) {
|
||||||
debug.Log("WalkTree", "start on %v", id.Str())
|
debug.Log("WalkTree", "start on %v", id.Str())
|
||||||
walkTree(server, "", id, done, jobCh)
|
walkTree(repo, "", id, done, jobCh)
|
||||||
close(jobCh)
|
close(jobCh)
|
||||||
debug.Log("WalkTree", "done")
|
debug.Log("WalkTree", "done")
|
||||||
}
|
}
|
||||||
|
12
walk_test.go
12
walk_test.go
@ -16,22 +16,22 @@ func TestWalkTree(t *testing.T) {
|
|||||||
dirs, err := filepath.Glob(*testWalkDirectory)
|
dirs, err := filepath.Glob(*testWalkDirectory)
|
||||||
OK(t, err)
|
OK(t, err)
|
||||||
|
|
||||||
server := SetupBackend(t)
|
repo := SetupRepo(t)
|
||||||
defer TeardownBackend(t, server)
|
defer TeardownRepo(t, repo)
|
||||||
|
|
||||||
// archive a few files
|
// archive a few files
|
||||||
arch := restic.NewArchiver(server)
|
arch := restic.NewArchiver(repo)
|
||||||
sn, _, err := arch.Snapshot(nil, dirs, nil)
|
sn, _, err := arch.Snapshot(nil, dirs, nil)
|
||||||
OK(t, err)
|
OK(t, err)
|
||||||
|
|
||||||
// flush server, write all packs
|
// flush repo, write all packs
|
||||||
OK(t, server.Flush())
|
OK(t, repo.Flush())
|
||||||
|
|
||||||
done := make(chan struct{})
|
done := make(chan struct{})
|
||||||
|
|
||||||
// start tree walker
|
// start tree walker
|
||||||
treeJobs := make(chan restic.WalkTreeJob)
|
treeJobs := make(chan restic.WalkTreeJob)
|
||||||
go restic.WalkTree(server, sn.Tree, done, treeJobs)
|
go restic.WalkTree(repo, sn.Tree, done, treeJobs)
|
||||||
|
|
||||||
// start filesystem walker
|
// start filesystem walker
|
||||||
fsJobs := make(chan pipe.Job)
|
fsJobs := make(chan pipe.Job)
|
||||||
|
Loading…
Reference in New Issue
Block a user