diff --git a/archiver.go b/archiver.go index f4f474563..286cd0d8f 100644 --- a/archiver.go +++ b/archiver.go @@ -15,7 +15,7 @@ import ( "github.com/restic/restic/debug" "github.com/restic/restic/pack" "github.com/restic/restic/pipe" - "github.com/restic/restic/repo" + "github.com/restic/restic/repository" "github.com/juju/errors" ) @@ -30,7 +30,7 @@ var archiverAllowAllFiles = func(string, os.FileInfo) bool { return true } // Archiver is used to backup a set of directories. type Archiver struct { - repo *repo.Repo + repo *repository.Repository blobToken chan struct{} @@ -39,7 +39,7 @@ type Archiver struct { } // NewArchiver returns a new archiver. -func NewArchiver(repo *repo.Repo) *Archiver { +func NewArchiver(repo *repository.Repository) *Archiver { arch := &Archiver{ repo: repo, blobToken: make(chan struct{}, maxConcurrentBlobs), diff --git a/cache.go b/cache.go index 47bfbac76..5e419bf2f 100644 --- a/cache.go +++ b/cache.go @@ -10,7 +10,7 @@ import ( "github.com/restic/restic/backend" "github.com/restic/restic/debug" - "github.com/restic/restic/repo" + "github.com/restic/restic/repository" ) // Cache is used to locally cache items from a repository. @@ -18,7 +18,7 @@ type Cache struct { base string } -func NewCache(repo *repo.Repo) (*Cache, error) { +func NewCache(repo *repository.Repository) (*Cache, error) { cacheDir, err := getCacheDir() if err != nil { return nil, err @@ -106,7 +106,7 @@ func (c *Cache) purge(t backend.Type, subtype string, id backend.ID) error { } // Clear removes information from the cache that isn't present in the repository any more. -func (c *Cache) Clear(repo *repo.Repo) error { +func (c *Cache) Clear(repo *repository.Repository) error { list, err := c.list(backend.Snapshot) if err != nil { return err diff --git a/cmd/restic/cmd_cat.go b/cmd/restic/cmd_cat.go index b90d85d45..e10a9071b 100644 --- a/cmd/restic/cmd_cat.go +++ b/cmd/restic/cmd_cat.go @@ -11,7 +11,7 @@ import ( "github.com/restic/restic/backend" "github.com/restic/restic/debug" "github.com/restic/restic/pack" - "github.com/restic/restic/repo" + "github.com/restic/restic/repository" ) type CmdCat struct{} @@ -107,7 +107,7 @@ func (cmd CmdCat) Execute(args []string) error { dec := json.NewDecoder(rd) - var key repo.Key + var key repository.Key err = dec.Decode(&key) if err != nil { return err diff --git a/cmd/restic/cmd_find.go b/cmd/restic/cmd_find.go index cdb2b4c18..10c1de4bd 100644 --- a/cmd/restic/cmd_find.go +++ b/cmd/restic/cmd_find.go @@ -8,7 +8,7 @@ import ( "github.com/restic/restic" "github.com/restic/restic/backend" "github.com/restic/restic/debug" - "github.com/restic/restic/repo" + "github.com/restic/restic/repository" ) type findResult struct { @@ -59,7 +59,7 @@ func parseTime(str string) (time.Time, error) { return time.Time{}, fmt.Errorf("unable to parse time: %q", str) } -func (c CmdFind) findInTree(repo *repo.Repo, id backend.ID, path string) ([]findResult, error) { +func (c CmdFind) findInTree(repo *repository.Repository, id backend.ID, path string) ([]findResult, error) { debug.Log("restic.find", "checking tree %v\n", id) tree, err := restic.LoadTree(repo, id) if err != nil { @@ -105,7 +105,7 @@ func (c CmdFind) findInTree(repo *repo.Repo, id backend.ID, path string) ([]find return results, nil } -func (c CmdFind) findInSnapshot(repo *repo.Repo, name string) error { +func (c CmdFind) findInSnapshot(repo *repository.Repository, name string) error { debug.Log("restic.find", "searching in snapshot %s\n for entries within [%s %s]", name, c.oldest, c.newest) id, err := backend.ParseID(name) diff --git a/cmd/restic/cmd_fsck.go b/cmd/restic/cmd_fsck.go index 6fbb4ddca..efd28956a 100644 --- a/cmd/restic/cmd_fsck.go +++ b/cmd/restic/cmd_fsck.go @@ -10,7 +10,7 @@ import ( "github.com/restic/restic/crypto" "github.com/restic/restic/debug" "github.com/restic/restic/pack" - "github.com/restic/restic/repo" + "github.com/restic/restic/repository" ) type CmdFsck struct { @@ -34,7 +34,7 @@ func init() { } } -func fsckFile(opts CmdFsck, repo *repo.Repo, IDs []backend.ID) (uint64, error) { +func fsckFile(opts CmdFsck, repo *repository.Repository, IDs []backend.ID) (uint64, error) { debug.Log("restic.fsckFile", "checking file %v", IDs) var bytes uint64 @@ -77,7 +77,7 @@ func fsckFile(opts CmdFsck, repo *repo.Repo, IDs []backend.ID) (uint64, error) { return bytes, nil } -func fsckTree(opts CmdFsck, repo *repo.Repo, id backend.ID) error { +func fsckTree(opts CmdFsck, repo *repository.Repository, id backend.ID) error { debug.Log("restic.fsckTree", "checking tree %v", id.Str()) tree, err := restic.LoadTree(repo, id) @@ -157,7 +157,7 @@ func fsckTree(opts CmdFsck, repo *repo.Repo, id backend.ID) error { return firstErr } -func fsckSnapshot(opts CmdFsck, repo *repo.Repo, id backend.ID) error { +func fsckSnapshot(opts CmdFsck, repo *repository.Repository, id backend.ID) error { debug.Log("restic.fsck", "checking snapshot %v\n", id) sn, err := restic.LoadSnapshot(repo, id) diff --git a/cmd/restic/cmd_key.go b/cmd/restic/cmd_key.go index b1abdad69..d962fdffc 100644 --- a/cmd/restic/cmd_key.go +++ b/cmd/restic/cmd_key.go @@ -6,7 +6,7 @@ import ( "os" "github.com/restic/restic/backend" - "github.com/restic/restic/repo" + "github.com/restic/restic/repository" ) type CmdKey struct{} @@ -21,7 +21,7 @@ func init() { } } -func listKeys(s *repo.Repo) error { +func listKeys(s *repository.Repository) error { tab := NewTable() tab.Header = fmt.Sprintf(" %-10s %-10s %-10s %s", "ID", "User", "Host", "Created") tab.RowFormat = "%s%-10s %-10s %-10s %s" @@ -35,7 +35,7 @@ func listKeys(s *repo.Repo) error { defer close(done) for name := range s.List(backend.Key, done) { - k, err := repo.LoadKey(s, name) + k, err := repository.LoadKey(s, name) if err != nil { fmt.Fprintf(os.Stderr, "LoadKey() failed: %v\n", err) continue @@ -56,7 +56,7 @@ func listKeys(s *repo.Repo) error { return nil } -func addKey(s *repo.Repo) error { +func addKey(s *repository.Repository) error { pw := readPassword("RESTIC_NEWPASSWORD", "enter password for new key: ") pw2 := readPassword("RESTIC_NEWPASSWORD", "enter password again: ") @@ -64,7 +64,7 @@ func addKey(s *repo.Repo) error { return errors.New("passwords do not match") } - id, err := repo.AddKey(s, pw, s.Key()) + id, err := repository.AddKey(s, pw, s.Key()) if err != nil { return fmt.Errorf("creating new key failed: %v\n", err) } @@ -74,7 +74,7 @@ func addKey(s *repo.Repo) error { return nil } -func deleteKey(repo *repo.Repo, name string) error { +func deleteKey(repo *repository.Repository, name string) error { if name == repo.KeyName() { return errors.New("refusing to remove key currently used to access repository") } @@ -88,7 +88,7 @@ func deleteKey(repo *repo.Repo, name string) error { return nil } -func changePassword(s *repo.Repo) error { +func changePassword(s *repository.Repository) error { pw := readPassword("RESTIC_NEWPASSWORD", "enter password for new key: ") pw2 := readPassword("RESTIC_NEWPASSWORD", "enter password again: ") @@ -97,7 +97,7 @@ func changePassword(s *repo.Repo) error { } // add new key - id, err := repo.AddKey(s, pw, s.Key()) + id, err := repository.AddKey(s, pw, s.Key()) if err != nil { return fmt.Errorf("creating new key failed: %v\n", err) } diff --git a/cmd/restic/cmd_ls.go b/cmd/restic/cmd_ls.go index c9d67ed43..ee07bcf0d 100644 --- a/cmd/restic/cmd_ls.go +++ b/cmd/restic/cmd_ls.go @@ -7,7 +7,7 @@ import ( "github.com/restic/restic" "github.com/restic/restic/backend" - "github.com/restic/restic/repo" + "github.com/restic/restic/repository" ) type CmdLs struct{} @@ -38,7 +38,7 @@ func printNode(prefix string, n *restic.Node) string { } } -func printTree(prefix string, repo *repo.Repo, id backend.ID) error { +func printTree(prefix string, repo *repository.Repository, id backend.ID) error { tree, err := restic.LoadTree(repo, id) if err != nil { return err diff --git a/cmd/restic/main.go b/cmd/restic/main.go index f34ecd4ac..805caf236 100644 --- a/cmd/restic/main.go +++ b/cmd/restic/main.go @@ -14,7 +14,7 @@ import ( "github.com/restic/restic/backend/local" "github.com/restic/restic/backend/sftp" "github.com/restic/restic/debug" - "github.com/restic/restic/repo" + "github.com/restic/restic/repository" ) var version = "compiled manually" @@ -72,7 +72,7 @@ func (cmd CmdInit) Execute(args []string) error { os.Exit(1) } - s := repo.New(be) + s := repository.New(be) err = s.Init(pw) if err != nil { fmt.Fprintf(os.Stderr, "creating key in backend at %s failed: %v\n", opts.Repo, err) @@ -133,7 +133,7 @@ func create(u string) (backend.Backend, error) { return sftp.Create(url.Path[1:], "ssh", args...) } -func OpenRepo() (*repo.Repo, error) { +func OpenRepo() (*repository.Repository, error) { if opts.Repo == "" { return nil, errors.New("Please specify repository location (-r)") } @@ -143,7 +143,7 @@ func OpenRepo() (*repo.Repo, error) { return nil, err } - s := repo.New(be) + s := repository.New(be) err = s.SearchKey(readPassword("RESTIC_PASSWORD", "enter password for repository: ")) if err != nil { diff --git a/node.go b/node.go index 517cbb31b..7785e7099 100644 --- a/node.go +++ b/node.go @@ -14,7 +14,7 @@ import ( "github.com/restic/restic/backend" "github.com/restic/restic/debug" "github.com/restic/restic/pack" - "github.com/restic/restic/repo" + "github.com/restic/restic/repository" ) // Node is a file, directory or other item in a backup. @@ -43,7 +43,7 @@ type Node struct { path string err error - blobs repo.Blobs + blobs repository.Blobs } func (node Node) String() string { @@ -103,7 +103,7 @@ func nodeTypeFromFileInfo(fi os.FileInfo) string { } // CreateAt creates the node at the given path and restores all the meta data. -func (node *Node) CreateAt(path string, repo *repo.Repo) error { +func (node *Node) CreateAt(path string, repo *repository.Repository) error { switch node.Type { case "dir": if err := node.createDirAt(path); err != nil { @@ -176,7 +176,7 @@ func (node Node) createDirAt(path string) error { return nil } -func (node Node) createFileAt(path string, repo *repo.Repo) error { +func (node Node) createFileAt(path string, repo *repository.Repository) error { f, err := os.OpenFile(path, os.O_CREATE|os.O_WRONLY, 0600) defer f.Close() diff --git a/repo/blob.go b/repository/blob.go similarity index 97% rename from repo/blob.go rename to repository/blob.go index 81c8a629f..12e1ed64b 100644 --- a/repo/blob.go +++ b/repository/blob.go @@ -1,4 +1,4 @@ -package repo +package repository import ( "bytes" diff --git a/repo/doc.go b/repository/doc.go similarity index 78% rename from repo/doc.go rename to repository/doc.go index bbc8a6f8e..b69b865e7 100644 --- a/repo/doc.go +++ b/repository/doc.go @@ -1,2 +1,2 @@ // Package repo implements a restic repository on top of a backend. -package repo +package repository diff --git a/repo/index.go b/repository/index.go similarity index 99% rename from repo/index.go rename to repository/index.go index 2854859c1..be7904c66 100644 --- a/repo/index.go +++ b/repository/index.go @@ -1,4 +1,4 @@ -package repo +package repository import ( "encoding/json" diff --git a/repo/index_test.go b/repository/index_test.go similarity index 94% rename from repo/index_test.go rename to repository/index_test.go index 2cc4a971c..9e3749ed3 100644 --- a/repo/index_test.go +++ b/repository/index_test.go @@ -1,4 +1,4 @@ -package repo_test +package repository_test import ( "bytes" @@ -8,7 +8,7 @@ import ( "github.com/restic/restic/backend" "github.com/restic/restic/pack" - "github.com/restic/restic/repo" + "github.com/restic/restic/repository" . "github.com/restic/restic/test" ) @@ -30,7 +30,7 @@ func TestIndexSerialize(t *testing.T) { } tests := []testEntry{} - idx := repo.NewIndex() + idx := repository.NewIndex() // create 50 packs with 20 blobs each for i := 0; i < 50; i++ { @@ -58,7 +58,7 @@ func TestIndexSerialize(t *testing.T) { err := idx.Encode(wr) OK(t, err) - idx2, err := repo.DecodeIndex(wr) + idx2, err := repository.DecodeIndex(wr) OK(t, err) Assert(t, idx2 != nil, "nil returned for decoded index") @@ -113,7 +113,7 @@ func TestIndexSerialize(t *testing.T) { err = idx2.Encode(wr3) OK(t, err) - idx3, err := repo.DecodeIndex(wr3) + idx3, err := repository.DecodeIndex(wr3) OK(t, err) Assert(t, idx3 != nil, "nil returned for decoded index") @@ -138,7 +138,7 @@ func TestIndexSerialize(t *testing.T) { } func TestIndexSize(t *testing.T) { - idx := repo.NewIndex() + idx := repository.NewIndex() packs := 200 blobs := 100 @@ -210,7 +210,7 @@ var exampleTests = []struct { } func TestIndexUnserialize(t *testing.T) { - idx, err := repo.DecodeIndex(bytes.NewReader(docExample)) + idx, err := repository.DecodeIndex(bytes.NewReader(docExample)) OK(t, err) for _, test := range exampleTests { diff --git a/repo/key.go b/repository/key.go similarity index 92% rename from repo/key.go rename to repository/key.go index d99a1408e..8a2ec4530 100644 --- a/repo/key.go +++ b/repository/key.go @@ -1,4 +1,4 @@ -package repo +package repository import ( "crypto/rand" @@ -49,12 +49,12 @@ type Key struct { // createMasterKey creates a new master key in the given backend and encrypts // it with the password. -func createMasterKey(s *Repo, password string) (*Key, error) { +func createMasterKey(s *Repository, password string) (*Key, error) { return AddKey(s, password, nil) } // OpenKey tries do decrypt the key specified by name with the given password. -func OpenKey(s *Repo, name string, password string) (*Key, error) { +func OpenKey(s *Repository, name string, password string) (*Key, error) { k, err := LoadKey(s, name) if err != nil { return nil, err @@ -94,7 +94,7 @@ func OpenKey(s *Repo, name string, password string) (*Key, error) { // SearchKey tries to decrypt all keys in the backend with the given password. // If none could be found, ErrNoKeyFound is returned. -func SearchKey(s *Repo, password string) (*Key, error) { +func SearchKey(s *Repository, password string) (*Key, error) { // try all keys in repo done := make(chan struct{}) defer close(done) @@ -111,7 +111,7 @@ func SearchKey(s *Repo, password string) (*Key, error) { } // LoadKey loads a key from the backend. -func LoadKey(s *Repo, name string) (*Key, error) { +func LoadKey(s *Repository, name string) (*Key, error) { // extract data from repo rd, err := s.be.Get(backend.Key, name) if err != nil { @@ -131,7 +131,7 @@ func LoadKey(s *Repo, name string) (*Key, error) { } // AddKey adds a new key to an already existing repository. -func AddKey(s *Repo, password string, template *crypto.Key) (*Key, error) { +func AddKey(s *Repository, password string, template *crypto.Key) (*Key, error) { // fill meta data about key newkey := &Key{ Created: time.Now(), diff --git a/repo/pool.go b/repository/pool.go similarity index 93% rename from repo/pool.go rename to repository/pool.go index 32100a3bd..258d08d92 100644 --- a/repo/pool.go +++ b/repository/pool.go @@ -1,4 +1,4 @@ -package repo +package repository import ( "sync" diff --git a/repo/repository.go b/repository/repository.go similarity index 84% rename from repo/repository.go rename to repository/repository.go index 80d392f89..16c3d8a14 100644 --- a/repo/repository.go +++ b/repository/repository.go @@ -1,4 +1,4 @@ -package repo +package repository import ( "bytes" @@ -26,8 +26,8 @@ type Config struct { ChunkerPolynomial chunker.Pol `json:"chunker_polynomial"` } -// Repo is used to access a repository in a backend. -type Repo struct { +// Repository is used to access a repository in a backend. +type Repository struct { be backend.Backend Config Config key *crypto.Key @@ -38,8 +38,8 @@ type Repo struct { packs []*pack.Packer } -func New(be backend.Backend) *Repo { - return &Repo{ +func New(be backend.Backend) *Repository { + return &Repository{ be: be, idx: NewIndex(), } @@ -48,25 +48,25 @@ func New(be backend.Backend) *Repo { // Find loads the list of all blobs of type t and searches for names which start // with prefix. If none is found, nil and ErrNoIDPrefixFound is returned. If // more than one is found, nil and ErrMultipleIDMatches is returned. -func (s *Repo) Find(t backend.Type, prefix string) (string, error) { +func (s *Repository) Find(t backend.Type, prefix string) (string, error) { return backend.Find(s.be, t, prefix) } // FindSnapshot takes a string and tries to find a snapshot whose ID matches // the string as closely as possible. -func (s *Repo) FindSnapshot(name string) (string, error) { +func (s *Repository) FindSnapshot(name string) (string, error) { return backend.FindSnapshot(s.be, name) } // PrefixLength returns the number of bytes required so that all prefixes of // all IDs of type t are unique. -func (s *Repo) PrefixLength(t backend.Type) (int, error) { +func (s *Repository) PrefixLength(t backend.Type) (int, error) { return backend.PrefixLength(s.be, t) } // Load tries to load and decrypt content identified by t and id from the // backend. -func (s *Repo) Load(t backend.Type, id backend.ID) ([]byte, error) { +func (s *Repository) Load(t backend.Type, id backend.ID) ([]byte, error) { debug.Log("Repo.Load", "load %v with id %v", t, id.Str()) // load blob from pack @@ -102,7 +102,7 @@ func (s *Repo) Load(t backend.Type, id backend.ID) ([]byte, error) { // LoadBlob tries to load and decrypt content identified by t and id from a // pack from the backend. -func (s *Repo) LoadBlob(t pack.BlobType, id backend.ID) ([]byte, error) { +func (s *Repository) LoadBlob(t pack.BlobType, id backend.ID) ([]byte, error) { debug.Log("Repo.LoadBlob", "load %v with id %v", t, id.Str()) // lookup pack packID, tpe, offset, length, err := s.idx.Lookup(id) @@ -151,7 +151,7 @@ func (s *Repo) LoadBlob(t pack.BlobType, id backend.ID) ([]byte, error) { // LoadJSONUnpacked decrypts the data and afterwards calls json.Unmarshal on // the item. -func (s *Repo) LoadJSONUnpacked(t backend.Type, id backend.ID, item interface{}) error { +func (s *Repository) LoadJSONUnpacked(t backend.Type, id backend.ID, item interface{}) error { // load blob from backend rd, err := s.be.Get(t, id.String()) if err != nil { @@ -178,7 +178,7 @@ func (s *Repo) LoadJSONUnpacked(t backend.Type, id backend.ID, item interface{}) // LoadJSONPack calls LoadBlob() to load a blob from the backend, decrypt the // data and afterwards call json.Unmarshal on the item. -func (s *Repo) LoadJSONPack(t pack.BlobType, id backend.ID, item interface{}) error { +func (s *Repository) LoadJSONPack(t pack.BlobType, id backend.ID, item interface{}) error { // lookup pack packID, _, offset, length, err := s.idx.Lookup(id) if err != nil { @@ -215,7 +215,7 @@ const maxPackers = 200 // findPacker returns a packer for a new blob of size bytes. Either a new one is // created or one is returned that already has some blobs. -func (s *Repo) findPacker(size uint) (*pack.Packer, error) { +func (s *Repository) findPacker(size uint) (*pack.Packer, error) { s.pm.Lock() defer s.pm.Unlock() @@ -242,7 +242,7 @@ func (s *Repo) findPacker(size uint) (*pack.Packer, error) { } // insertPacker appends p to s.packs. -func (s *Repo) insertPacker(p *pack.Packer) { +func (s *Repository) insertPacker(p *pack.Packer) { s.pm.Lock() defer s.pm.Unlock() @@ -251,7 +251,7 @@ func (s *Repo) insertPacker(p *pack.Packer) { } // savePacker stores p in the backend. -func (s *Repo) savePacker(p *pack.Packer) error { +func (s *Repository) savePacker(p *pack.Packer) error { debug.Log("Repo.savePacker", "save packer with %d blobs\n", p.Count()) _, err := p.Finalize() if err != nil { @@ -278,7 +278,7 @@ func (s *Repo) savePacker(p *pack.Packer) error { } // countPacker returns the number of open (unfinished) packers. -func (s *Repo) countPacker() int { +func (s *Repository) countPacker() int { s.pm.Lock() defer s.pm.Unlock() @@ -287,7 +287,7 @@ func (s *Repo) countPacker() int { // Save encrypts data and stores it to the backend as type t. If data is small // enough, it will be packed together with other small blobs. -func (s *Repo) Save(t pack.BlobType, data []byte, id backend.ID) (backend.ID, error) { +func (s *Repository) Save(t pack.BlobType, data []byte, id backend.ID) (backend.ID, error) { if id == nil { // compute plaintext hash id = backend.Hash(data) @@ -332,7 +332,7 @@ func (s *Repo) Save(t pack.BlobType, data []byte, id backend.ID) (backend.ID, er } // SaveFrom encrypts data read from rd and stores it in a pack in the backend as type t. -func (s *Repo) SaveFrom(t pack.BlobType, id backend.ID, length uint, rd io.Reader) error { +func (s *Repository) SaveFrom(t pack.BlobType, id backend.ID, length uint, rd io.Reader) error { debug.Log("Repo.SaveFrom", "save id %v (%v, %d bytes)", id.Str(), t, length) if id == nil { return errors.New("id is nil") @@ -353,7 +353,7 @@ func (s *Repo) SaveFrom(t pack.BlobType, id backend.ID, length uint, rd io.Reade // SaveJSON serialises item as JSON and encrypts and saves it in a pack in the // backend as type t. -func (s *Repo) SaveJSON(t pack.BlobType, item interface{}) (backend.ID, error) { +func (s *Repository) SaveJSON(t pack.BlobType, item interface{}) (backend.ID, error) { debug.Log("Repo.SaveJSON", "save %v blob", t) buf := getBuf()[:0] defer freeBuf(buf) @@ -372,7 +372,7 @@ func (s *Repo) SaveJSON(t pack.BlobType, item interface{}) (backend.ID, error) { // SaveJSONUnpacked serialises item as JSON and encrypts and saves it in the // backend as type t, without a pack. It returns the storage hash. -func (s *Repo) SaveJSONUnpacked(t backend.Type, item interface{}) (backend.ID, error) { +func (s *Repository) SaveJSONUnpacked(t backend.Type, item interface{}) (backend.ID, error) { // create file blob, err := s.be.Create() if err != nil { @@ -409,7 +409,7 @@ func (s *Repo) SaveJSONUnpacked(t backend.Type, item interface{}) (backend.ID, e } // Flush saves all remaining packs. -func (s *Repo) Flush() error { +func (s *Repository) Flush() error { s.pm.Lock() defer s.pm.Unlock() @@ -426,22 +426,22 @@ func (s *Repo) Flush() error { return nil } -func (s *Repo) Backend() backend.Backend { +func (s *Repository) Backend() backend.Backend { return s.be } -func (s *Repo) Index() *Index { +func (s *Repository) Index() *Index { return s.idx } // SetIndex instructs the repository to use the given index. -func (s *Repo) SetIndex(i *Index) { +func (s *Repository) SetIndex(i *Index) { s.idx = i } // SaveIndex saves all new packs in the index in the backend, returned is the // storage ID. -func (s *Repo) SaveIndex() (backend.ID, error) { +func (s *Repository) SaveIndex() (backend.ID, error) { debug.Log("Repo.SaveIndex", "Saving index") // create blob @@ -483,7 +483,7 @@ func (s *Repo) SaveIndex() (backend.ID, error) { // LoadIndex loads all index files from the backend and merges them with the // current index. -func (s *Repo) LoadIndex() error { +func (s *Repository) LoadIndex() error { debug.Log("Repo.LoadIndex", "Loading index") done := make(chan struct{}) defer close(done) @@ -498,7 +498,7 @@ func (s *Repo) LoadIndex() error { } // loadIndex loads the index id and merges it with the currently used index. -func (s *Repo) loadIndex(id string) error { +func (s *Repository) loadIndex(id string) error { debug.Log("Repo.loadIndex", "Loading index %v", id[:8]) before := len(s.idx.pack) @@ -532,7 +532,7 @@ func (s *Repo) loadIndex(id string) error { const repositoryIDSize = sha256.Size const RepoVersion = 1 -func createConfig(s *Repo) (err error) { +func createConfig(s *Repository) (err error) { s.Config.ChunkerPolynomial, err = chunker.RandomPolynomial() if err != nil { return err @@ -553,7 +553,7 @@ func createConfig(s *Repo) (err error) { return err } -func (s *Repo) loadConfig(cfg *Config) error { +func (s *Repository) loadConfig(cfg *Config) error { err := s.LoadJSONUnpacked(backend.Config, nil, cfg) if err != nil { return err @@ -572,7 +572,7 @@ func (s *Repo) loadConfig(cfg *Config) error { // SearchKey finds a key with the supplied password, afterwards the config is // read and parsed. -func (s *Repo) SearchKey(password string) error { +func (s *Repository) SearchKey(password string) error { key, err := SearchKey(s, password) if err != nil { return err @@ -585,7 +585,7 @@ func (s *Repo) SearchKey(password string) error { // Init creates a new master key with the supplied password and initializes the // repository config. -func (s *Repo) Init(password string) error { +func (s *Repository) Init(password string) error { has, err := s.Test(backend.Config, "") if err != nil { return err @@ -604,7 +604,7 @@ func (s *Repo) Init(password string) error { return createConfig(s) } -func (s *Repo) Decrypt(ciphertext []byte) ([]byte, error) { +func (s *Repository) Decrypt(ciphertext []byte) ([]byte, error) { if s.key == nil { return nil, errors.New("key for repository not set") } @@ -612,7 +612,7 @@ func (s *Repo) Decrypt(ciphertext []byte) ([]byte, error) { return crypto.Decrypt(s.key, nil, ciphertext) } -func (s *Repo) Encrypt(ciphertext, plaintext []byte) ([]byte, error) { +func (s *Repository) Encrypt(ciphertext, plaintext []byte) ([]byte, error) { if s.key == nil { return nil, errors.New("key for repository not set") } @@ -620,16 +620,16 @@ func (s *Repo) Encrypt(ciphertext, plaintext []byte) ([]byte, error) { return crypto.Encrypt(s.key, ciphertext, plaintext) } -func (s *Repo) Key() *crypto.Key { +func (s *Repository) Key() *crypto.Key { return s.key } -func (s *Repo) KeyName() string { +func (s *Repository) KeyName() string { return s.keyName } // Count returns the number of blobs of a given type in the backend. -func (s *Repo) Count(t backend.Type) (n uint) { +func (s *Repository) Count(t backend.Type) (n uint) { for _ = range s.be.List(t, nil) { n++ } @@ -639,27 +639,27 @@ func (s *Repo) Count(t backend.Type) (n uint) { // Proxy methods to backend -func (s *Repo) Get(t backend.Type, name string) (io.ReadCloser, error) { +func (s *Repository) Get(t backend.Type, name string) (io.ReadCloser, error) { return s.be.Get(t, name) } -func (s *Repo) List(t backend.Type, done <-chan struct{}) <-chan string { +func (s *Repository) List(t backend.Type, done <-chan struct{}) <-chan string { return s.be.List(t, done) } -func (s *Repo) Test(t backend.Type, name string) (bool, error) { +func (s *Repository) Test(t backend.Type, name string) (bool, error) { return s.be.Test(t, name) } -func (s *Repo) Remove(t backend.Type, name string) error { +func (s *Repository) Remove(t backend.Type, name string) error { return s.be.Remove(t, name) } -func (s *Repo) Close() error { +func (s *Repository) Close() error { return s.be.Close() } -func (s *Repo) Delete() error { +func (s *Repository) Delete() error { if b, ok := s.be.(backend.Deleter); ok { return b.Delete() } @@ -667,6 +667,6 @@ func (s *Repo) Delete() error { return errors.New("Delete() called for backend that does not implement this method") } -func (s *Repo) Location() string { +func (s *Repository) Location() string { return s.be.Location() } diff --git a/repo/repository_test.go b/repository/repository_test.go similarity index 99% rename from repo/repository_test.go rename to repository/repository_test.go index 5620b923f..fe9f5c9d4 100644 --- a/repo/repository_test.go +++ b/repository/repository_test.go @@ -1,4 +1,4 @@ -package repo_test +package repository_test import ( "bytes" diff --git a/restorer.go b/restorer.go index be0e6e347..9511d3bc9 100644 --- a/restorer.go +++ b/restorer.go @@ -7,14 +7,14 @@ import ( "syscall" "github.com/restic/restic/backend" - "github.com/restic/restic/repo" + "github.com/restic/restic/repository" "github.com/juju/errors" ) // Restorer is used to restore a snapshot to a directory. type Restorer struct { - repo *repo.Repo + repo *repository.Repository sn *Snapshot Error func(dir string, node *Node, err error) error @@ -24,7 +24,7 @@ type Restorer struct { var restorerAbortOnAllErrors = func(str string, node *Node, err error) error { return err } // NewRestorer creates a restorer preloaded with the content from the snapshot id. -func NewRestorer(repo *repo.Repo, id backend.ID) (*Restorer, error) { +func NewRestorer(repo *repository.Repository, id backend.ID) (*Restorer, error) { r := &Restorer{repo: repo, Error: restorerAbortOnAllErrors} var err error diff --git a/snapshot.go b/snapshot.go index 776e35790..7558f15c3 100644 --- a/snapshot.go +++ b/snapshot.go @@ -9,7 +9,7 @@ import ( "time" "github.com/restic/restic/backend" - "github.com/restic/restic/repo" + "github.com/restic/restic/repository" ) type Snapshot struct { @@ -50,7 +50,7 @@ func NewSnapshot(paths []string) (*Snapshot, error) { return sn, nil } -func LoadSnapshot(repo *repo.Repo, id backend.ID) (*Snapshot, error) { +func LoadSnapshot(repo *repository.Repository, id backend.ID) (*Snapshot, error) { sn := &Snapshot{id: id} err := repo.LoadJSONUnpacked(backend.Snapshot, id, sn) if err != nil { diff --git a/test/backend.go b/test/backend.go index 6886fe6fd..173ecee9d 100644 --- a/test/backend.go +++ b/test/backend.go @@ -10,14 +10,14 @@ import ( "github.com/restic/restic" "github.com/restic/restic/backend" "github.com/restic/restic/backend/local" - "github.com/restic/restic/repo" + "github.com/restic/restic/repository" ) var TestPassword = flag.String("test.password", "geheim", `use this password for repositories created during tests (default: "geheim")`) var TestCleanup = flag.Bool("test.cleanup", true, "clean up after running tests (remove local backend directory with all content)") var TestTempDir = flag.String("test.tempdir", "", "use this directory for temporary storage (default: system temp dir)") -func SetupRepo(t testing.TB) *repo.Repo { +func SetupRepo(t testing.TB) *repository.Repository { tempdir, err := ioutil.TempDir(*TestTempDir, "restic-test-") OK(t, err) @@ -29,12 +29,12 @@ func SetupRepo(t testing.TB) *repo.Repo { err = os.Setenv("RESTIC_CACHE", filepath.Join(tempdir, "cache")) OK(t, err) - repo := repo.New(b) + repo := repository.New(b) OK(t, repo.Init(*TestPassword)) return repo } -func TeardownRepo(t testing.TB, repo *repo.Repo) { +func TeardownRepo(t testing.TB, repo *repository.Repository) { if !*TestCleanup { l := repo.Backend().(*local.Local) t.Logf("leaving local backend at %s\n", l.Location()) @@ -44,7 +44,7 @@ func TeardownRepo(t testing.TB, repo *repo.Repo) { OK(t, repo.Delete()) } -func SnapshotDir(t testing.TB, repo *repo.Repo, path string, parent backend.ID) *restic.Snapshot { +func SnapshotDir(t testing.TB, repo *repository.Repository, path string, parent backend.ID) *restic.Snapshot { arch := restic.NewArchiver(repo) sn, _, err := arch.Snapshot(nil, []string{path}, parent) OK(t, err) diff --git a/tree.go b/tree.go index 0fccf5082..8afa6cc64 100644 --- a/tree.go +++ b/tree.go @@ -8,7 +8,7 @@ import ( "github.com/restic/restic/backend" "github.com/restic/restic/debug" "github.com/restic/restic/pack" - "github.com/restic/restic/repo" + "github.com/restic/restic/repository" ) type Tree struct { @@ -30,7 +30,7 @@ func (t Tree) String() string { return fmt.Sprintf("Tree<%d nodes>", len(t.Nodes)) } -func LoadTree(repo *repo.Repo, id backend.ID) (*Tree, error) { +func LoadTree(repo *repository.Repository, id backend.ID) (*Tree, error) { tree := &Tree{} err := repo.LoadJSONPack(pack.Tree, id, tree) if err != nil { diff --git a/walk.go b/walk.go index f666d5d38..24b4a5294 100644 --- a/walk.go +++ b/walk.go @@ -5,7 +5,7 @@ import ( "github.com/restic/restic/backend" "github.com/restic/restic/debug" - "github.com/restic/restic/repo" + "github.com/restic/restic/repository" ) type WalkTreeJob struct { @@ -16,7 +16,7 @@ type WalkTreeJob struct { Tree *Tree } -func walkTree(repo *repo.Repo, path string, treeID backend.ID, done chan struct{}, jobCh chan<- WalkTreeJob) { +func walkTree(repo *repository.Repository, path string, treeID backend.ID, done chan struct{}, jobCh chan<- WalkTreeJob) { debug.Log("walkTree", "start on %q (%v)", path, treeID.Str()) t, err := LoadTree(repo, treeID) @@ -41,7 +41,7 @@ func walkTree(repo *repo.Repo, path string, treeID backend.ID, done chan struct{ // WalkTree walks the tree specified by id recursively and sends a job for each // file and directory it finds. When the channel done is closed, processing // stops. -func WalkTree(repo *repo.Repo, id backend.ID, done chan struct{}, jobCh chan<- WalkTreeJob) { +func WalkTree(repo *repository.Repository, id backend.ID, done chan struct{}, jobCh chan<- WalkTreeJob) { debug.Log("WalkTree", "start on %v", id.Str()) walkTree(repo, "", id, done, jobCh) close(jobCh)