mirror of
https://github.com/octoleo/restic.git
synced 2024-12-22 02:48:55 +00:00
Rename 'Repo' -> 'Repository'
This commit is contained in:
parent
232c472836
commit
95536e8a21
@ -30,7 +30,7 @@ var archiverAllowAllFiles = func(string, os.FileInfo) bool { return true }
|
||||
|
||||
// Archiver is used to backup a set of directories.
|
||||
type Archiver struct {
|
||||
repo *repository.Repo
|
||||
repo *repository.Repository
|
||||
|
||||
blobToken chan struct{}
|
||||
|
||||
@ -39,7 +39,7 @@ type Archiver struct {
|
||||
}
|
||||
|
||||
// NewArchiver returns a new archiver.
|
||||
func NewArchiver(repo *repository.Repo) *Archiver {
|
||||
func NewArchiver(repo *repository.Repository) *Archiver {
|
||||
arch := &Archiver{
|
||||
repo: repo,
|
||||
blobToken: make(chan struct{}, maxConcurrentBlobs),
|
||||
|
4
cache.go
4
cache.go
@ -18,7 +18,7 @@ type Cache struct {
|
||||
base string
|
||||
}
|
||||
|
||||
func NewCache(repo *repository.Repo) (*Cache, error) {
|
||||
func NewCache(repo *repository.Repository) (*Cache, error) {
|
||||
cacheDir, err := getCacheDir()
|
||||
if err != nil {
|
||||
return nil, err
|
||||
@ -106,7 +106,7 @@ func (c *Cache) purge(t backend.Type, subtype string, id backend.ID) error {
|
||||
}
|
||||
|
||||
// Clear removes information from the cache that isn't present in the repository any more.
|
||||
func (c *Cache) Clear(repo *repository.Repo) error {
|
||||
func (c *Cache) Clear(repo *repository.Repository) error {
|
||||
list, err := c.list(backend.Snapshot)
|
||||
if err != nil {
|
||||
return err
|
||||
|
@ -59,7 +59,7 @@ func parseTime(str string) (time.Time, error) {
|
||||
return time.Time{}, fmt.Errorf("unable to parse time: %q", str)
|
||||
}
|
||||
|
||||
func (c CmdFind) findInTree(repo *repository.Repo, id backend.ID, path string) ([]findResult, error) {
|
||||
func (c CmdFind) findInTree(repo *repository.Repository, id backend.ID, path string) ([]findResult, error) {
|
||||
debug.Log("restic.find", "checking tree %v\n", id)
|
||||
tree, err := restic.LoadTree(repo, id)
|
||||
if err != nil {
|
||||
@ -105,7 +105,7 @@ func (c CmdFind) findInTree(repo *repository.Repo, id backend.ID, path string) (
|
||||
return results, nil
|
||||
}
|
||||
|
||||
func (c CmdFind) findInSnapshot(repo *repository.Repo, name string) error {
|
||||
func (c CmdFind) findInSnapshot(repo *repository.Repository, name string) error {
|
||||
debug.Log("restic.find", "searching in snapshot %s\n for entries within [%s %s]", name, c.oldest, c.newest)
|
||||
|
||||
id, err := backend.ParseID(name)
|
||||
|
@ -34,7 +34,7 @@ func init() {
|
||||
}
|
||||
}
|
||||
|
||||
func fsckFile(opts CmdFsck, repo *repository.Repo, IDs []backend.ID) (uint64, error) {
|
||||
func fsckFile(opts CmdFsck, repo *repository.Repository, IDs []backend.ID) (uint64, error) {
|
||||
debug.Log("restic.fsckFile", "checking file %v", IDs)
|
||||
var bytes uint64
|
||||
|
||||
@ -77,7 +77,7 @@ func fsckFile(opts CmdFsck, repo *repository.Repo, IDs []backend.ID) (uint64, er
|
||||
return bytes, nil
|
||||
}
|
||||
|
||||
func fsckTree(opts CmdFsck, repo *repository.Repo, id backend.ID) error {
|
||||
func fsckTree(opts CmdFsck, repo *repository.Repository, id backend.ID) error {
|
||||
debug.Log("restic.fsckTree", "checking tree %v", id.Str())
|
||||
|
||||
tree, err := restic.LoadTree(repo, id)
|
||||
@ -157,7 +157,7 @@ func fsckTree(opts CmdFsck, repo *repository.Repo, id backend.ID) error {
|
||||
return firstErr
|
||||
}
|
||||
|
||||
func fsckSnapshot(opts CmdFsck, repo *repository.Repo, id backend.ID) error {
|
||||
func fsckSnapshot(opts CmdFsck, repo *repository.Repository, id backend.ID) error {
|
||||
debug.Log("restic.fsck", "checking snapshot %v\n", id)
|
||||
|
||||
sn, err := restic.LoadSnapshot(repo, id)
|
||||
|
@ -21,7 +21,7 @@ func init() {
|
||||
}
|
||||
}
|
||||
|
||||
func listKeys(s *repository.Repo) error {
|
||||
func listKeys(s *repository.Repository) error {
|
||||
tab := NewTable()
|
||||
tab.Header = fmt.Sprintf(" %-10s %-10s %-10s %s", "ID", "User", "Host", "Created")
|
||||
tab.RowFormat = "%s%-10s %-10s %-10s %s"
|
||||
@ -56,7 +56,7 @@ func listKeys(s *repository.Repo) error {
|
||||
return nil
|
||||
}
|
||||
|
||||
func addKey(s *repository.Repo) error {
|
||||
func addKey(s *repository.Repository) error {
|
||||
pw := readPassword("RESTIC_NEWPASSWORD", "enter password for new key: ")
|
||||
pw2 := readPassword("RESTIC_NEWPASSWORD", "enter password again: ")
|
||||
|
||||
@ -74,7 +74,7 @@ func addKey(s *repository.Repo) error {
|
||||
return nil
|
||||
}
|
||||
|
||||
func deleteKey(repo *repository.Repo, name string) error {
|
||||
func deleteKey(repo *repository.Repository, name string) error {
|
||||
if name == repo.KeyName() {
|
||||
return errors.New("refusing to remove key currently used to access repository")
|
||||
}
|
||||
@ -88,7 +88,7 @@ func deleteKey(repo *repository.Repo, name string) error {
|
||||
return nil
|
||||
}
|
||||
|
||||
func changePassword(s *repository.Repo) error {
|
||||
func changePassword(s *repository.Repository) error {
|
||||
pw := readPassword("RESTIC_NEWPASSWORD", "enter password for new key: ")
|
||||
pw2 := readPassword("RESTIC_NEWPASSWORD", "enter password again: ")
|
||||
|
||||
|
@ -38,7 +38,7 @@ func printNode(prefix string, n *restic.Node) string {
|
||||
}
|
||||
}
|
||||
|
||||
func printTree(prefix string, repo *repository.Repo, id backend.ID) error {
|
||||
func printTree(prefix string, repo *repository.Repository, id backend.ID) error {
|
||||
tree, err := restic.LoadTree(repo, id)
|
||||
if err != nil {
|
||||
return err
|
||||
|
@ -133,7 +133,7 @@ func create(u string) (backend.Backend, error) {
|
||||
return sftp.Create(url.Path[1:], "ssh", args...)
|
||||
}
|
||||
|
||||
func OpenRepo() (*repository.Repo, error) {
|
||||
func OpenRepo() (*repository.Repository, error) {
|
||||
if opts.Repo == "" {
|
||||
return nil, errors.New("Please specify repository location (-r)")
|
||||
}
|
||||
|
4
node.go
4
node.go
@ -103,7 +103,7 @@ func nodeTypeFromFileInfo(fi os.FileInfo) string {
|
||||
}
|
||||
|
||||
// CreateAt creates the node at the given path and restores all the meta data.
|
||||
func (node *Node) CreateAt(path string, repo *repository.Repo) error {
|
||||
func (node *Node) CreateAt(path string, repo *repository.Repository) error {
|
||||
switch node.Type {
|
||||
case "dir":
|
||||
if err := node.createDirAt(path); err != nil {
|
||||
@ -176,7 +176,7 @@ func (node Node) createDirAt(path string) error {
|
||||
return nil
|
||||
}
|
||||
|
||||
func (node Node) createFileAt(path string, repo *repository.Repo) error {
|
||||
func (node Node) createFileAt(path string, repo *repository.Repository) error {
|
||||
f, err := os.OpenFile(path, os.O_CREATE|os.O_WRONLY, 0600)
|
||||
defer f.Close()
|
||||
|
||||
|
@ -49,12 +49,12 @@ type Key struct {
|
||||
|
||||
// createMasterKey creates a new master key in the given backend and encrypts
|
||||
// it with the password.
|
||||
func createMasterKey(s *Repo, password string) (*Key, error) {
|
||||
func createMasterKey(s *Repository, password string) (*Key, error) {
|
||||
return AddKey(s, password, nil)
|
||||
}
|
||||
|
||||
// OpenKey tries do decrypt the key specified by name with the given password.
|
||||
func OpenKey(s *Repo, name string, password string) (*Key, error) {
|
||||
func OpenKey(s *Repository, name string, password string) (*Key, error) {
|
||||
k, err := LoadKey(s, name)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
@ -94,7 +94,7 @@ func OpenKey(s *Repo, name string, password string) (*Key, error) {
|
||||
|
||||
// SearchKey tries to decrypt all keys in the backend with the given password.
|
||||
// If none could be found, ErrNoKeyFound is returned.
|
||||
func SearchKey(s *Repo, password string) (*Key, error) {
|
||||
func SearchKey(s *Repository, password string) (*Key, error) {
|
||||
// try all keys in repo
|
||||
done := make(chan struct{})
|
||||
defer close(done)
|
||||
@ -111,7 +111,7 @@ func SearchKey(s *Repo, password string) (*Key, error) {
|
||||
}
|
||||
|
||||
// LoadKey loads a key from the backend.
|
||||
func LoadKey(s *Repo, name string) (*Key, error) {
|
||||
func LoadKey(s *Repository, name string) (*Key, error) {
|
||||
// extract data from repo
|
||||
rd, err := s.be.Get(backend.Key, name)
|
||||
if err != nil {
|
||||
@ -131,7 +131,7 @@ func LoadKey(s *Repo, name string) (*Key, error) {
|
||||
}
|
||||
|
||||
// AddKey adds a new key to an already existing repository.
|
||||
func AddKey(s *Repo, password string, template *crypto.Key) (*Key, error) {
|
||||
func AddKey(s *Repository, password string, template *crypto.Key) (*Key, error) {
|
||||
// fill meta data about key
|
||||
newkey := &Key{
|
||||
Created: time.Now(),
|
||||
|
@ -26,8 +26,8 @@ type Config struct {
|
||||
ChunkerPolynomial chunker.Pol `json:"chunker_polynomial"`
|
||||
}
|
||||
|
||||
// Repo is used to access a repository in a backend.
|
||||
type Repo struct {
|
||||
// Repository is used to access a repository in a backend.
|
||||
type Repository struct {
|
||||
be backend.Backend
|
||||
Config Config
|
||||
key *crypto.Key
|
||||
@ -38,8 +38,8 @@ type Repo struct {
|
||||
packs []*pack.Packer
|
||||
}
|
||||
|
||||
func New(be backend.Backend) *Repo {
|
||||
return &Repo{
|
||||
func New(be backend.Backend) *Repository {
|
||||
return &Repository{
|
||||
be: be,
|
||||
idx: NewIndex(),
|
||||
}
|
||||
@ -48,25 +48,25 @@ func New(be backend.Backend) *Repo {
|
||||
// Find loads the list of all blobs of type t and searches for names which start
|
||||
// with prefix. If none is found, nil and ErrNoIDPrefixFound is returned. If
|
||||
// more than one is found, nil and ErrMultipleIDMatches is returned.
|
||||
func (s *Repo) Find(t backend.Type, prefix string) (string, error) {
|
||||
func (s *Repository) Find(t backend.Type, prefix string) (string, error) {
|
||||
return backend.Find(s.be, t, prefix)
|
||||
}
|
||||
|
||||
// FindSnapshot takes a string and tries to find a snapshot whose ID matches
|
||||
// the string as closely as possible.
|
||||
func (s *Repo) FindSnapshot(name string) (string, error) {
|
||||
func (s *Repository) FindSnapshot(name string) (string, error) {
|
||||
return backend.FindSnapshot(s.be, name)
|
||||
}
|
||||
|
||||
// PrefixLength returns the number of bytes required so that all prefixes of
|
||||
// all IDs of type t are unique.
|
||||
func (s *Repo) PrefixLength(t backend.Type) (int, error) {
|
||||
func (s *Repository) PrefixLength(t backend.Type) (int, error) {
|
||||
return backend.PrefixLength(s.be, t)
|
||||
}
|
||||
|
||||
// Load tries to load and decrypt content identified by t and id from the
|
||||
// backend.
|
||||
func (s *Repo) Load(t backend.Type, id backend.ID) ([]byte, error) {
|
||||
func (s *Repository) Load(t backend.Type, id backend.ID) ([]byte, error) {
|
||||
debug.Log("Repo.Load", "load %v with id %v", t, id.Str())
|
||||
|
||||
// load blob from pack
|
||||
@ -102,7 +102,7 @@ func (s *Repo) Load(t backend.Type, id backend.ID) ([]byte, error) {
|
||||
|
||||
// LoadBlob tries to load and decrypt content identified by t and id from a
|
||||
// pack from the backend.
|
||||
func (s *Repo) LoadBlob(t pack.BlobType, id backend.ID) ([]byte, error) {
|
||||
func (s *Repository) LoadBlob(t pack.BlobType, id backend.ID) ([]byte, error) {
|
||||
debug.Log("Repo.LoadBlob", "load %v with id %v", t, id.Str())
|
||||
// lookup pack
|
||||
packID, tpe, offset, length, err := s.idx.Lookup(id)
|
||||
@ -151,7 +151,7 @@ func (s *Repo) LoadBlob(t pack.BlobType, id backend.ID) ([]byte, error) {
|
||||
|
||||
// LoadJSONUnpacked decrypts the data and afterwards calls json.Unmarshal on
|
||||
// the item.
|
||||
func (s *Repo) LoadJSONUnpacked(t backend.Type, id backend.ID, item interface{}) error {
|
||||
func (s *Repository) LoadJSONUnpacked(t backend.Type, id backend.ID, item interface{}) error {
|
||||
// load blob from backend
|
||||
rd, err := s.be.Get(t, id.String())
|
||||
if err != nil {
|
||||
@ -178,7 +178,7 @@ func (s *Repo) LoadJSONUnpacked(t backend.Type, id backend.ID, item interface{})
|
||||
|
||||
// LoadJSONPack calls LoadBlob() to load a blob from the backend, decrypt the
|
||||
// data and afterwards call json.Unmarshal on the item.
|
||||
func (s *Repo) LoadJSONPack(t pack.BlobType, id backend.ID, item interface{}) error {
|
||||
func (s *Repository) LoadJSONPack(t pack.BlobType, id backend.ID, item interface{}) error {
|
||||
// lookup pack
|
||||
packID, _, offset, length, err := s.idx.Lookup(id)
|
||||
if err != nil {
|
||||
@ -215,7 +215,7 @@ const maxPackers = 200
|
||||
|
||||
// findPacker returns a packer for a new blob of size bytes. Either a new one is
|
||||
// created or one is returned that already has some blobs.
|
||||
func (s *Repo) findPacker(size uint) (*pack.Packer, error) {
|
||||
func (s *Repository) findPacker(size uint) (*pack.Packer, error) {
|
||||
s.pm.Lock()
|
||||
defer s.pm.Unlock()
|
||||
|
||||
@ -242,7 +242,7 @@ func (s *Repo) findPacker(size uint) (*pack.Packer, error) {
|
||||
}
|
||||
|
||||
// insertPacker appends p to s.packs.
|
||||
func (s *Repo) insertPacker(p *pack.Packer) {
|
||||
func (s *Repository) insertPacker(p *pack.Packer) {
|
||||
s.pm.Lock()
|
||||
defer s.pm.Unlock()
|
||||
|
||||
@ -251,7 +251,7 @@ func (s *Repo) insertPacker(p *pack.Packer) {
|
||||
}
|
||||
|
||||
// savePacker stores p in the backend.
|
||||
func (s *Repo) savePacker(p *pack.Packer) error {
|
||||
func (s *Repository) savePacker(p *pack.Packer) error {
|
||||
debug.Log("Repo.savePacker", "save packer with %d blobs\n", p.Count())
|
||||
_, err := p.Finalize()
|
||||
if err != nil {
|
||||
@ -278,7 +278,7 @@ func (s *Repo) savePacker(p *pack.Packer) error {
|
||||
}
|
||||
|
||||
// countPacker returns the number of open (unfinished) packers.
|
||||
func (s *Repo) countPacker() int {
|
||||
func (s *Repository) countPacker() int {
|
||||
s.pm.Lock()
|
||||
defer s.pm.Unlock()
|
||||
|
||||
@ -287,7 +287,7 @@ func (s *Repo) countPacker() int {
|
||||
|
||||
// Save encrypts data and stores it to the backend as type t. If data is small
|
||||
// enough, it will be packed together with other small blobs.
|
||||
func (s *Repo) Save(t pack.BlobType, data []byte, id backend.ID) (backend.ID, error) {
|
||||
func (s *Repository) Save(t pack.BlobType, data []byte, id backend.ID) (backend.ID, error) {
|
||||
if id == nil {
|
||||
// compute plaintext hash
|
||||
id = backend.Hash(data)
|
||||
@ -332,7 +332,7 @@ func (s *Repo) Save(t pack.BlobType, data []byte, id backend.ID) (backend.ID, er
|
||||
}
|
||||
|
||||
// SaveFrom encrypts data read from rd and stores it in a pack in the backend as type t.
|
||||
func (s *Repo) SaveFrom(t pack.BlobType, id backend.ID, length uint, rd io.Reader) error {
|
||||
func (s *Repository) SaveFrom(t pack.BlobType, id backend.ID, length uint, rd io.Reader) error {
|
||||
debug.Log("Repo.SaveFrom", "save id %v (%v, %d bytes)", id.Str(), t, length)
|
||||
if id == nil {
|
||||
return errors.New("id is nil")
|
||||
@ -353,7 +353,7 @@ func (s *Repo) SaveFrom(t pack.BlobType, id backend.ID, length uint, rd io.Reade
|
||||
|
||||
// SaveJSON serialises item as JSON and encrypts and saves it in a pack in the
|
||||
// backend as type t.
|
||||
func (s *Repo) SaveJSON(t pack.BlobType, item interface{}) (backend.ID, error) {
|
||||
func (s *Repository) SaveJSON(t pack.BlobType, item interface{}) (backend.ID, error) {
|
||||
debug.Log("Repo.SaveJSON", "save %v blob", t)
|
||||
buf := getBuf()[:0]
|
||||
defer freeBuf(buf)
|
||||
@ -372,7 +372,7 @@ func (s *Repo) SaveJSON(t pack.BlobType, item interface{}) (backend.ID, error) {
|
||||
|
||||
// SaveJSONUnpacked serialises item as JSON and encrypts and saves it in the
|
||||
// backend as type t, without a pack. It returns the storage hash.
|
||||
func (s *Repo) SaveJSONUnpacked(t backend.Type, item interface{}) (backend.ID, error) {
|
||||
func (s *Repository) SaveJSONUnpacked(t backend.Type, item interface{}) (backend.ID, error) {
|
||||
// create file
|
||||
blob, err := s.be.Create()
|
||||
if err != nil {
|
||||
@ -409,7 +409,7 @@ func (s *Repo) SaveJSONUnpacked(t backend.Type, item interface{}) (backend.ID, e
|
||||
}
|
||||
|
||||
// Flush saves all remaining packs.
|
||||
func (s *Repo) Flush() error {
|
||||
func (s *Repository) Flush() error {
|
||||
s.pm.Lock()
|
||||
defer s.pm.Unlock()
|
||||
|
||||
@ -426,22 +426,22 @@ func (s *Repo) Flush() error {
|
||||
return nil
|
||||
}
|
||||
|
||||
func (s *Repo) Backend() backend.Backend {
|
||||
func (s *Repository) Backend() backend.Backend {
|
||||
return s.be
|
||||
}
|
||||
|
||||
func (s *Repo) Index() *Index {
|
||||
func (s *Repository) Index() *Index {
|
||||
return s.idx
|
||||
}
|
||||
|
||||
// SetIndex instructs the repository to use the given index.
|
||||
func (s *Repo) SetIndex(i *Index) {
|
||||
func (s *Repository) SetIndex(i *Index) {
|
||||
s.idx = i
|
||||
}
|
||||
|
||||
// SaveIndex saves all new packs in the index in the backend, returned is the
|
||||
// storage ID.
|
||||
func (s *Repo) SaveIndex() (backend.ID, error) {
|
||||
func (s *Repository) SaveIndex() (backend.ID, error) {
|
||||
debug.Log("Repo.SaveIndex", "Saving index")
|
||||
|
||||
// create blob
|
||||
@ -483,7 +483,7 @@ func (s *Repo) SaveIndex() (backend.ID, error) {
|
||||
|
||||
// LoadIndex loads all index files from the backend and merges them with the
|
||||
// current index.
|
||||
func (s *Repo) LoadIndex() error {
|
||||
func (s *Repository) LoadIndex() error {
|
||||
debug.Log("Repo.LoadIndex", "Loading index")
|
||||
done := make(chan struct{})
|
||||
defer close(done)
|
||||
@ -498,7 +498,7 @@ func (s *Repo) LoadIndex() error {
|
||||
}
|
||||
|
||||
// loadIndex loads the index id and merges it with the currently used index.
|
||||
func (s *Repo) loadIndex(id string) error {
|
||||
func (s *Repository) loadIndex(id string) error {
|
||||
debug.Log("Repo.loadIndex", "Loading index %v", id[:8])
|
||||
before := len(s.idx.pack)
|
||||
|
||||
@ -532,7 +532,7 @@ func (s *Repo) loadIndex(id string) error {
|
||||
const repositoryIDSize = sha256.Size
|
||||
const RepoVersion = 1
|
||||
|
||||
func createConfig(s *Repo) (err error) {
|
||||
func createConfig(s *Repository) (err error) {
|
||||
s.Config.ChunkerPolynomial, err = chunker.RandomPolynomial()
|
||||
if err != nil {
|
||||
return err
|
||||
@ -553,7 +553,7 @@ func createConfig(s *Repo) (err error) {
|
||||
return err
|
||||
}
|
||||
|
||||
func (s *Repo) loadConfig(cfg *Config) error {
|
||||
func (s *Repository) loadConfig(cfg *Config) error {
|
||||
err := s.LoadJSONUnpacked(backend.Config, nil, cfg)
|
||||
if err != nil {
|
||||
return err
|
||||
@ -572,7 +572,7 @@ func (s *Repo) loadConfig(cfg *Config) error {
|
||||
|
||||
// SearchKey finds a key with the supplied password, afterwards the config is
|
||||
// read and parsed.
|
||||
func (s *Repo) SearchKey(password string) error {
|
||||
func (s *Repository) SearchKey(password string) error {
|
||||
key, err := SearchKey(s, password)
|
||||
if err != nil {
|
||||
return err
|
||||
@ -585,7 +585,7 @@ func (s *Repo) SearchKey(password string) error {
|
||||
|
||||
// Init creates a new master key with the supplied password and initializes the
|
||||
// repository config.
|
||||
func (s *Repo) Init(password string) error {
|
||||
func (s *Repository) Init(password string) error {
|
||||
has, err := s.Test(backend.Config, "")
|
||||
if err != nil {
|
||||
return err
|
||||
@ -604,7 +604,7 @@ func (s *Repo) Init(password string) error {
|
||||
return createConfig(s)
|
||||
}
|
||||
|
||||
func (s *Repo) Decrypt(ciphertext []byte) ([]byte, error) {
|
||||
func (s *Repository) Decrypt(ciphertext []byte) ([]byte, error) {
|
||||
if s.key == nil {
|
||||
return nil, errors.New("key for repository not set")
|
||||
}
|
||||
@ -612,7 +612,7 @@ func (s *Repo) Decrypt(ciphertext []byte) ([]byte, error) {
|
||||
return crypto.Decrypt(s.key, nil, ciphertext)
|
||||
}
|
||||
|
||||
func (s *Repo) Encrypt(ciphertext, plaintext []byte) ([]byte, error) {
|
||||
func (s *Repository) Encrypt(ciphertext, plaintext []byte) ([]byte, error) {
|
||||
if s.key == nil {
|
||||
return nil, errors.New("key for repository not set")
|
||||
}
|
||||
@ -620,16 +620,16 @@ func (s *Repo) Encrypt(ciphertext, plaintext []byte) ([]byte, error) {
|
||||
return crypto.Encrypt(s.key, ciphertext, plaintext)
|
||||
}
|
||||
|
||||
func (s *Repo) Key() *crypto.Key {
|
||||
func (s *Repository) Key() *crypto.Key {
|
||||
return s.key
|
||||
}
|
||||
|
||||
func (s *Repo) KeyName() string {
|
||||
func (s *Repository) KeyName() string {
|
||||
return s.keyName
|
||||
}
|
||||
|
||||
// Count returns the number of blobs of a given type in the backend.
|
||||
func (s *Repo) Count(t backend.Type) (n uint) {
|
||||
func (s *Repository) Count(t backend.Type) (n uint) {
|
||||
for _ = range s.be.List(t, nil) {
|
||||
n++
|
||||
}
|
||||
@ -639,27 +639,27 @@ func (s *Repo) Count(t backend.Type) (n uint) {
|
||||
|
||||
// Proxy methods to backend
|
||||
|
||||
func (s *Repo) Get(t backend.Type, name string) (io.ReadCloser, error) {
|
||||
func (s *Repository) Get(t backend.Type, name string) (io.ReadCloser, error) {
|
||||
return s.be.Get(t, name)
|
||||
}
|
||||
|
||||
func (s *Repo) List(t backend.Type, done <-chan struct{}) <-chan string {
|
||||
func (s *Repository) List(t backend.Type, done <-chan struct{}) <-chan string {
|
||||
return s.be.List(t, done)
|
||||
}
|
||||
|
||||
func (s *Repo) Test(t backend.Type, name string) (bool, error) {
|
||||
func (s *Repository) Test(t backend.Type, name string) (bool, error) {
|
||||
return s.be.Test(t, name)
|
||||
}
|
||||
|
||||
func (s *Repo) Remove(t backend.Type, name string) error {
|
||||
func (s *Repository) Remove(t backend.Type, name string) error {
|
||||
return s.be.Remove(t, name)
|
||||
}
|
||||
|
||||
func (s *Repo) Close() error {
|
||||
func (s *Repository) Close() error {
|
||||
return s.be.Close()
|
||||
}
|
||||
|
||||
func (s *Repo) Delete() error {
|
||||
func (s *Repository) Delete() error {
|
||||
if b, ok := s.be.(backend.Deleter); ok {
|
||||
return b.Delete()
|
||||
}
|
||||
@ -667,6 +667,6 @@ func (s *Repo) Delete() error {
|
||||
return errors.New("Delete() called for backend that does not implement this method")
|
||||
}
|
||||
|
||||
func (s *Repo) Location() string {
|
||||
func (s *Repository) Location() string {
|
||||
return s.be.Location()
|
||||
}
|
||||
|
@ -14,7 +14,7 @@ import (
|
||||
|
||||
// Restorer is used to restore a snapshot to a directory.
|
||||
type Restorer struct {
|
||||
repo *repository.Repo
|
||||
repo *repository.Repository
|
||||
sn *Snapshot
|
||||
|
||||
Error func(dir string, node *Node, err error) error
|
||||
@ -24,7 +24,7 @@ type Restorer struct {
|
||||
var restorerAbortOnAllErrors = func(str string, node *Node, err error) error { return err }
|
||||
|
||||
// NewRestorer creates a restorer preloaded with the content from the snapshot id.
|
||||
func NewRestorer(repo *repository.Repo, id backend.ID) (*Restorer, error) {
|
||||
func NewRestorer(repo *repository.Repository, id backend.ID) (*Restorer, error) {
|
||||
r := &Restorer{repo: repo, Error: restorerAbortOnAllErrors}
|
||||
|
||||
var err error
|
||||
|
@ -50,7 +50,7 @@ func NewSnapshot(paths []string) (*Snapshot, error) {
|
||||
return sn, nil
|
||||
}
|
||||
|
||||
func LoadSnapshot(repo *repository.Repo, id backend.ID) (*Snapshot, error) {
|
||||
func LoadSnapshot(repo *repository.Repository, id backend.ID) (*Snapshot, error) {
|
||||
sn := &Snapshot{id: id}
|
||||
err := repo.LoadJSONUnpacked(backend.Snapshot, id, sn)
|
||||
if err != nil {
|
||||
|
@ -17,7 +17,7 @@ var TestPassword = flag.String("test.password", "geheim", `use this password for
|
||||
var TestCleanup = flag.Bool("test.cleanup", true, "clean up after running tests (remove local backend directory with all content)")
|
||||
var TestTempDir = flag.String("test.tempdir", "", "use this directory for temporary storage (default: system temp dir)")
|
||||
|
||||
func SetupRepo(t testing.TB) *repository.Repo {
|
||||
func SetupRepo(t testing.TB) *repository.Repository {
|
||||
tempdir, err := ioutil.TempDir(*TestTempDir, "restic-test-")
|
||||
OK(t, err)
|
||||
|
||||
@ -34,7 +34,7 @@ func SetupRepo(t testing.TB) *repository.Repo {
|
||||
return repo
|
||||
}
|
||||
|
||||
func TeardownRepo(t testing.TB, repo *repository.Repo) {
|
||||
func TeardownRepo(t testing.TB, repo *repository.Repository) {
|
||||
if !*TestCleanup {
|
||||
l := repo.Backend().(*local.Local)
|
||||
t.Logf("leaving local backend at %s\n", l.Location())
|
||||
@ -44,7 +44,7 @@ func TeardownRepo(t testing.TB, repo *repository.Repo) {
|
||||
OK(t, repo.Delete())
|
||||
}
|
||||
|
||||
func SnapshotDir(t testing.TB, repo *repository.Repo, path string, parent backend.ID) *restic.Snapshot {
|
||||
func SnapshotDir(t testing.TB, repo *repository.Repository, path string, parent backend.ID) *restic.Snapshot {
|
||||
arch := restic.NewArchiver(repo)
|
||||
sn, _, err := arch.Snapshot(nil, []string{path}, parent)
|
||||
OK(t, err)
|
||||
|
2
tree.go
2
tree.go
@ -30,7 +30,7 @@ func (t Tree) String() string {
|
||||
return fmt.Sprintf("Tree<%d nodes>", len(t.Nodes))
|
||||
}
|
||||
|
||||
func LoadTree(repo *repository.Repo, id backend.ID) (*Tree, error) {
|
||||
func LoadTree(repo *repository.Repository, id backend.ID) (*Tree, error) {
|
||||
tree := &Tree{}
|
||||
err := repo.LoadJSONPack(pack.Tree, id, tree)
|
||||
if err != nil {
|
||||
|
4
walk.go
4
walk.go
@ -16,7 +16,7 @@ type WalkTreeJob struct {
|
||||
Tree *Tree
|
||||
}
|
||||
|
||||
func walkTree(repo *repository.Repo, path string, treeID backend.ID, done chan struct{}, jobCh chan<- WalkTreeJob) {
|
||||
func walkTree(repo *repository.Repository, path string, treeID backend.ID, done chan struct{}, jobCh chan<- WalkTreeJob) {
|
||||
debug.Log("walkTree", "start on %q (%v)", path, treeID.Str())
|
||||
|
||||
t, err := LoadTree(repo, treeID)
|
||||
@ -41,7 +41,7 @@ func walkTree(repo *repository.Repo, path string, treeID backend.ID, done chan s
|
||||
// WalkTree walks the tree specified by id recursively and sends a job for each
|
||||
// file and directory it finds. When the channel done is closed, processing
|
||||
// stops.
|
||||
func WalkTree(repo *repository.Repo, id backend.ID, done chan struct{}, jobCh chan<- WalkTreeJob) {
|
||||
func WalkTree(repo *repository.Repository, id backend.ID, done chan struct{}, jobCh chan<- WalkTreeJob) {
|
||||
debug.Log("WalkTree", "start on %v", id.Str())
|
||||
walkTree(repo, "", id, done, jobCh)
|
||||
close(jobCh)
|
||||
|
Loading…
Reference in New Issue
Block a user