mirror of
https://github.com/octoleo/restic.git
synced 2024-11-17 02:25:12 +00:00
Merge pull request #511 from restic/cleanups
Cleanups and test functions
This commit is contained in:
commit
deae1e7e29
@ -45,8 +45,7 @@ type CIEnvironment interface {
|
|||||||
|
|
||||||
// TravisEnvironment is the environment in which Travis tests run.
|
// TravisEnvironment is the environment in which Travis tests run.
|
||||||
type TravisEnvironment struct {
|
type TravisEnvironment struct {
|
||||||
goxArch []string
|
goxOSArch []string
|
||||||
goxOS []string
|
|
||||||
minio string
|
minio string
|
||||||
|
|
||||||
minioSrv *Background
|
minioSrv *Background
|
||||||
@ -175,24 +174,27 @@ func (env *TravisEnvironment) Prepare() error {
|
|||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
if runtime.GOOS == "linux" {
|
if runtime.GOOS == "linux" {
|
||||||
env.goxArch = []string{"386", "amd64"}
|
env.goxOSArch = []string{
|
||||||
|
"linux/386", "linux/amd64",
|
||||||
|
"windows/386", "windows/amd64",
|
||||||
|
"darwin/386", "darwin/amd64",
|
||||||
|
"freebsd/386", "freebsd/amd64",
|
||||||
|
"opendbsd/386", "opendbsd/amd64",
|
||||||
|
}
|
||||||
if !strings.HasPrefix(runtime.Version(), "go1.3") {
|
if !strings.HasPrefix(runtime.Version(), "go1.3") {
|
||||||
env.goxArch = append(env.goxArch, "arm")
|
env.goxOSArch = append(env.goxOSArch,
|
||||||
|
"linux/arm", "freebsd/arm")
|
||||||
}
|
}
|
||||||
|
|
||||||
env.goxOS = []string{"linux", "darwin", "freebsd", "openbsd", "windows"}
|
|
||||||
} else {
|
} else {
|
||||||
env.goxArch = []string{runtime.GOARCH}
|
env.goxOSArch = []string{runtime.GOOS + "/" + runtime.GOARCH}
|
||||||
env.goxOS = []string{runtime.GOOS}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
msg("gox: OS %v, ARCH %v\n", env.goxOS, env.goxArch)
|
msg("gox: OS/ARCH %v\n", env.goxOSArch)
|
||||||
|
|
||||||
v := runtime.Version()
|
v := runtime.Version()
|
||||||
if !strings.HasPrefix(v, "go1.5") && !strings.HasPrefix(v, "go1.6") {
|
if !strings.HasPrefix(v, "go1.5") && !strings.HasPrefix(v, "go1.6") {
|
||||||
err := run("gox", "-build-toolchain",
|
err := run("gox", "-build-toolchain",
|
||||||
"-os", strings.Join(env.goxOS, " "),
|
"-osarch", strings.Join(env.goxOSArch, " "))
|
||||||
"-arch", strings.Join(env.goxArch, " "))
|
|
||||||
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
@ -320,8 +322,7 @@ func (env *TravisEnvironment) RunTests() error {
|
|||||||
// compile for all target architectures with tags
|
// compile for all target architectures with tags
|
||||||
for _, tags := range []string{"release", "debug"} {
|
for _, tags := range []string{"release", "debug"} {
|
||||||
runWithEnv(env.env, "gox", "-verbose",
|
runWithEnv(env.env, "gox", "-verbose",
|
||||||
"-os", strings.Join(env.goxOS, " "),
|
"-osarch", strings.Join(env.goxOSArch, " "),
|
||||||
"-arch", strings.Join(env.goxArch, " "),
|
|
||||||
"-tags", tags,
|
"-tags", tags,
|
||||||
"-output", "/tmp/{{.Dir}}_{{.OS}}_{{.Arch}}",
|
"-output", "/tmp/{{.Dir}}_{{.OS}}_{{.Arch}}",
|
||||||
"cmds/restic")
|
"cmds/restic")
|
||||||
|
@ -33,17 +33,7 @@ func loadBlobsFromPacks(repo *repository.Repository) (packs map[backend.ID][]pac
|
|||||||
defer close(done)
|
defer close(done)
|
||||||
|
|
||||||
f := func(job worker.Job, done <-chan struct{}) (interface{}, error) {
|
f := func(job worker.Job, done <-chan struct{}) (interface{}, error) {
|
||||||
id := job.Data.(backend.ID)
|
return repo.ListPack(job.Data.(backend.ID))
|
||||||
|
|
||||||
h := backend.Handle{Type: backend.Data, Name: id.String()}
|
|
||||||
rd := backend.NewReadSeeker(repo.Backend(), h)
|
|
||||||
|
|
||||||
unpacker, err := pack.NewUnpacker(repo.Key(), rd)
|
|
||||||
if err != nil {
|
|
||||||
return nil, err
|
|
||||||
}
|
|
||||||
|
|
||||||
return unpacker.Entries, nil
|
|
||||||
}
|
}
|
||||||
|
|
||||||
jobCh := make(chan worker.Job)
|
jobCh := make(chan worker.Job)
|
||||||
|
@ -1,7 +1,6 @@
|
|||||||
package restic
|
package restic
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"bytes"
|
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
"fmt"
|
"fmt"
|
||||||
"io"
|
"io"
|
||||||
@ -92,7 +91,7 @@ func (arch *Archiver) isKnownBlob(id backend.ID) bool {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Save stores a blob read from rd in the repository.
|
// Save stores a blob read from rd in the repository.
|
||||||
func (arch *Archiver) Save(t pack.BlobType, id backend.ID, length uint, rd io.Reader) error {
|
func (arch *Archiver) Save(t pack.BlobType, data []byte, id backend.ID) error {
|
||||||
debug.Log("Archiver.Save", "Save(%v, %v)\n", t, id.Str())
|
debug.Log("Archiver.Save", "Save(%v, %v)\n", t, id.Str())
|
||||||
|
|
||||||
if arch.isKnownBlob(id) {
|
if arch.isKnownBlob(id) {
|
||||||
@ -100,7 +99,7 @@ func (arch *Archiver) Save(t pack.BlobType, id backend.ID, length uint, rd io.Re
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
err := arch.repo.SaveFrom(t, &id, length, rd)
|
_, err := arch.repo.SaveAndEncrypt(t, data, &id)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
debug.Log("Archiver.Save", "Save(%v, %v): error %v\n", t, id.Str(), err)
|
debug.Log("Archiver.Save", "Save(%v, %v): error %v\n", t, id.Str(), err)
|
||||||
return err
|
return err
|
||||||
@ -160,7 +159,7 @@ func (arch *Archiver) saveChunk(chunk chunker.Chunk, p *Progress, token struct{}
|
|||||||
defer freeBuf(chunk.Data)
|
defer freeBuf(chunk.Data)
|
||||||
|
|
||||||
id := backend.Hash(chunk.Data)
|
id := backend.Hash(chunk.Data)
|
||||||
err := arch.Save(pack.Data, id, chunk.Length, bytes.NewReader(chunk.Data))
|
err := arch.Save(pack.Data, chunk.Data, id)
|
||||||
// TODO handle error
|
// TODO handle error
|
||||||
if err != nil {
|
if err != nil {
|
||||||
panic(err)
|
panic(err)
|
||||||
|
@ -1,7 +1,6 @@
|
|||||||
package restic_test
|
package restic_test
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"bytes"
|
|
||||||
"crypto/rand"
|
"crypto/rand"
|
||||||
"errors"
|
"errors"
|
||||||
"io"
|
"io"
|
||||||
@ -108,7 +107,7 @@ func testArchiverDuplication(t *testing.T) {
|
|||||||
|
|
||||||
buf := make([]byte, 50)
|
buf := make([]byte, 50)
|
||||||
|
|
||||||
err := arch.Save(pack.Data, id, uint(len(buf)), bytes.NewReader(buf))
|
err := arch.Save(pack.Data, buf, id)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
t.Fatal(err)
|
t.Fatal(err)
|
||||||
}
|
}
|
||||||
|
@ -267,7 +267,7 @@ func testParallelSaveWithDuplication(t *testing.T, seed int) {
|
|||||||
|
|
||||||
id := backend.Hash(c.Data)
|
id := backend.Hash(c.Data)
|
||||||
time.Sleep(time.Duration(id[0]))
|
time.Sleep(time.Duration(id[0]))
|
||||||
err := arch.Save(pack.Data, id, c.Length, bytes.NewReader(c.Data))
|
err := arch.Save(pack.Data, c.Data, id)
|
||||||
<-barrier
|
<-barrier
|
||||||
errChan <- err
|
errChan <- err
|
||||||
}(c, errChan)
|
}(c, errChan)
|
||||||
|
@ -534,6 +534,7 @@ func filterTrees(backlog backend.IDs, loaderChan chan<- backend.ID, in <-chan tr
|
|||||||
inCh = nil
|
inCh = nil
|
||||||
|
|
||||||
case outCh <- job:
|
case outCh <- job:
|
||||||
|
debug.Log("checker.FilterTrees", "tree sent to check: %v", job.ID.Str())
|
||||||
outCh = nil
|
outCh = nil
|
||||||
inCh = in
|
inCh = in
|
||||||
}
|
}
|
||||||
@ -581,6 +582,14 @@ func (c *Checker) checkTree(id backend.ID, tree *restic.Tree) (errs []error) {
|
|||||||
for _, node := range tree.Nodes {
|
for _, node := range tree.Nodes {
|
||||||
switch node.Type {
|
switch node.Type {
|
||||||
case "file":
|
case "file":
|
||||||
|
if node.Content == nil {
|
||||||
|
errs = append(errs, Error{TreeID: id, Err: fmt.Errorf("file %q has nil blob list", node.Name)})
|
||||||
|
}
|
||||||
|
|
||||||
|
if node.Mode == 0 {
|
||||||
|
errs = append(errs, Error{TreeID: id, Err: fmt.Errorf("file %q has invalid mode: %v", node.Name, node.Mode)})
|
||||||
|
}
|
||||||
|
|
||||||
for b, blobID := range node.Content {
|
for b, blobID := range node.Content {
|
||||||
if blobID.IsNull() {
|
if blobID.IsNull() {
|
||||||
errs = append(errs, Error{TreeID: id, Err: fmt.Errorf("file %q blob %d has null ID", node.Name, b)})
|
errs = append(errs, Error{TreeID: id, Err: fmt.Errorf("file %q blob %d has null ID", node.Name, b)})
|
||||||
@ -598,6 +607,16 @@ func (c *Checker) checkTree(id backend.ID, tree *restic.Tree) (errs []error) {
|
|||||||
errs = append(errs, Error{TreeID: id, Err: fmt.Errorf("dir node %q subtree id is null", node.Name)})
|
errs = append(errs, Error{TreeID: id, Err: fmt.Errorf("dir node %q subtree id is null", node.Name)})
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
|
|
||||||
|
case "symlink":
|
||||||
|
// nothing to check
|
||||||
|
|
||||||
|
default:
|
||||||
|
errs = append(errs, Error{TreeID: id, Err: fmt.Errorf("node %q with invalid type %q", node.Name, node.Type)})
|
||||||
|
}
|
||||||
|
|
||||||
|
if node.Name == "" {
|
||||||
|
errs = append(errs, Error{TreeID: id, Err: errors.New("node with empty name")})
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -19,7 +19,7 @@ type Progress struct {
|
|||||||
start time.Time
|
start time.Time
|
||||||
c *time.Ticker
|
c *time.Ticker
|
||||||
cancel chan struct{}
|
cancel chan struct{}
|
||||||
o sync.Once
|
o *sync.Once
|
||||||
d time.Duration
|
d time.Duration
|
||||||
lastUpdate time.Time
|
lastUpdate time.Time
|
||||||
|
|
||||||
@ -52,7 +52,7 @@ func (p *Progress) Start() {
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
p.o = sync.Once{}
|
p.o = &sync.Once{}
|
||||||
p.cancel = make(chan struct{})
|
p.cancel = make(chan struct{})
|
||||||
p.running = true
|
p.running = true
|
||||||
p.Reset()
|
p.Reset()
|
||||||
|
@ -6,7 +6,6 @@ import (
|
|||||||
"errors"
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
"io"
|
"io"
|
||||||
"io/ioutil"
|
|
||||||
"os"
|
"os"
|
||||||
|
|
||||||
"restic/backend"
|
"restic/backend"
|
||||||
@ -212,26 +211,6 @@ func (r *Repository) SaveAndEncrypt(t pack.BlobType, data []byte, id *backend.ID
|
|||||||
return *id, r.savePacker(packer)
|
return *id, r.savePacker(packer)
|
||||||
}
|
}
|
||||||
|
|
||||||
// SaveFrom encrypts data read from rd and stores it in a pack in the backend as type t.
|
|
||||||
func (r *Repository) SaveFrom(t pack.BlobType, id *backend.ID, length uint, rd io.Reader) error {
|
|
||||||
debug.Log("Repo.SaveFrom", "save id %v (%v, %d bytes)", id.Str(), t, length)
|
|
||||||
if id == nil {
|
|
||||||
return errors.New("id is nil")
|
|
||||||
}
|
|
||||||
|
|
||||||
buf, err := ioutil.ReadAll(rd)
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
|
|
||||||
_, err = r.SaveAndEncrypt(t, buf, id)
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// SaveJSON serialises item as JSON and encrypts and saves it in a pack in the
|
// SaveJSON serialises item as JSON and encrypts and saves it in a pack in the
|
||||||
// backend as type t.
|
// backend as type t.
|
||||||
func (r *Repository) SaveJSON(t pack.BlobType, item interface{}) (backend.ID, error) {
|
func (r *Repository) SaveJSON(t pack.BlobType, item interface{}) (backend.ID, error) {
|
||||||
@ -539,6 +518,19 @@ func (r *Repository) List(t backend.Type, done <-chan struct{}) <-chan backend.I
|
|||||||
return outCh
|
return outCh
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// ListPack returns the list of blobs saved in the pack id.
|
||||||
|
func (r *Repository) ListPack(id backend.ID) ([]pack.Blob, error) {
|
||||||
|
h := backend.Handle{Type: backend.Data, Name: id.String()}
|
||||||
|
rd := backend.NewReadSeeker(r.Backend(), h)
|
||||||
|
|
||||||
|
unpacker, err := pack.NewUnpacker(r.Key(), rd)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
return unpacker.Entries, nil
|
||||||
|
}
|
||||||
|
|
||||||
// Delete calls backend.Delete() if implemented, and returns an error
|
// Delete calls backend.Delete() if implemented, and returns an error
|
||||||
// otherwise.
|
// otherwise.
|
||||||
func (r *Repository) Delete() error {
|
func (r *Repository) Delete() error {
|
||||||
|
@ -117,8 +117,9 @@ func TestSaveFrom(t *testing.T) {
|
|||||||
id := backend.Hash(data)
|
id := backend.Hash(data)
|
||||||
|
|
||||||
// save
|
// save
|
||||||
err = repo.SaveFrom(pack.Data, &id, uint(size), bytes.NewReader(data))
|
id2, err := repo.SaveAndEncrypt(pack.Data, data, &id)
|
||||||
OK(t, err)
|
OK(t, err)
|
||||||
|
Equals(t, id, id2)
|
||||||
|
|
||||||
OK(t, repo.Flush())
|
OK(t, repo.Flush())
|
||||||
|
|
||||||
@ -136,7 +137,7 @@ func TestSaveFrom(t *testing.T) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func BenchmarkSaveFrom(t *testing.B) {
|
func BenchmarkSaveAndEncrypt(t *testing.B) {
|
||||||
repo := SetupRepo()
|
repo := SetupRepo()
|
||||||
defer TeardownRepo(repo)
|
defer TeardownRepo(repo)
|
||||||
|
|
||||||
@ -153,7 +154,7 @@ func BenchmarkSaveFrom(t *testing.B) {
|
|||||||
|
|
||||||
for i := 0; i < t.N; i++ {
|
for i := 0; i < t.N; i++ {
|
||||||
// save
|
// save
|
||||||
err = repo.SaveFrom(pack.Data, &id, uint(size), bytes.NewReader(data))
|
_, err = repo.SaveAndEncrypt(pack.Data, data, &id)
|
||||||
OK(t, err)
|
OK(t, err)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
63
src/restic/repository/testing.go
Normal file
63
src/restic/repository/testing.go
Normal file
@ -0,0 +1,63 @@
|
|||||||
|
package repository
|
||||||
|
|
||||||
|
import (
|
||||||
|
"os"
|
||||||
|
"restic/backend"
|
||||||
|
"restic/backend/local"
|
||||||
|
"restic/backend/mem"
|
||||||
|
"testing"
|
||||||
|
)
|
||||||
|
|
||||||
|
// TestBackend returns a fully configured in-memory backend.
|
||||||
|
func TestBackend(t testing.TB) (be backend.Backend, cleanup func()) {
|
||||||
|
return mem.New(), func() {}
|
||||||
|
}
|
||||||
|
|
||||||
|
// TestPassword is used for all repositories created by the Test* functions.
|
||||||
|
const TestPassword = "geheim"
|
||||||
|
|
||||||
|
// TestRepositoryWithBackend returns a repository initialized with a test
|
||||||
|
// password. If be is nil, an in-memory backend is used.
|
||||||
|
func TestRepositoryWithBackend(t testing.TB, be backend.Backend) (r *Repository, cleanup func()) {
|
||||||
|
var beCleanup func()
|
||||||
|
if be == nil {
|
||||||
|
be, beCleanup = TestBackend(t)
|
||||||
|
}
|
||||||
|
|
||||||
|
r = New(be)
|
||||||
|
|
||||||
|
err := r.Init(TestPassword)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatalf("TestRepopository(): initialize repo failed: %v", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
return r, func() {
|
||||||
|
if beCleanup != nil {
|
||||||
|
beCleanup()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// TestRepository returns a repository initialized with a test password on an
|
||||||
|
// in-memory backend. When the environment variable RESTIC_TEST_REPO is set to
|
||||||
|
// a non-existing directory, a local backend is created there and this is used
|
||||||
|
// instead. The directory is not removed.
|
||||||
|
func TestRepository(t testing.TB) (r *Repository, cleanup func()) {
|
||||||
|
dir := os.Getenv("RESTIC_TEST_REPO")
|
||||||
|
if dir != "" {
|
||||||
|
_, err := os.Stat(dir)
|
||||||
|
if err != nil {
|
||||||
|
be, err := local.Create(dir)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatalf("error creating local backend at %v: %v", dir, err)
|
||||||
|
}
|
||||||
|
return TestRepositoryWithBackend(t, be)
|
||||||
|
}
|
||||||
|
|
||||||
|
if err == nil {
|
||||||
|
t.Logf("directory at %v already exists, using mem backend", dir)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return TestRepositoryWithBackend(t, nil)
|
||||||
|
}
|
@ -11,6 +11,7 @@ import (
|
|||||||
"restic/repository"
|
"restic/repository"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
// Snapshot is the state of a resource at one point in time.
|
||||||
type Snapshot struct {
|
type Snapshot struct {
|
||||||
Time time.Time `json:"time"`
|
Time time.Time `json:"time"`
|
||||||
Parent *backend.ID `json:"parent,omitempty"`
|
Parent *backend.ID `json:"parent,omitempty"`
|
||||||
@ -25,6 +26,8 @@ type Snapshot struct {
|
|||||||
id *backend.ID // plaintext ID, used during restore
|
id *backend.ID // plaintext ID, used during restore
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// NewSnapshot returns an initialized snapshot struct for the current user and
|
||||||
|
// time.
|
||||||
func NewSnapshot(paths []string) (*Snapshot, error) {
|
func NewSnapshot(paths []string) (*Snapshot, error) {
|
||||||
for i, path := range paths {
|
for i, path := range paths {
|
||||||
if p, err := filepath.Abs(path); err != nil {
|
if p, err := filepath.Abs(path); err != nil {
|
||||||
@ -50,6 +53,7 @@ func NewSnapshot(paths []string) (*Snapshot, error) {
|
|||||||
return sn, nil
|
return sn, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// LoadSnapshot loads the snapshot with the id and returns it.
|
||||||
func LoadSnapshot(repo *repository.Repository, id backend.ID) (*Snapshot, error) {
|
func LoadSnapshot(repo *repository.Repository, id backend.ID) (*Snapshot, error) {
|
||||||
sn := &Snapshot{id: &id}
|
sn := &Snapshot{id: &id}
|
||||||
err := repo.LoadJSONUnpacked(backend.Snapshot, id, sn)
|
err := repo.LoadJSONUnpacked(backend.Snapshot, id, sn)
|
||||||
@ -60,10 +64,28 @@ func LoadSnapshot(repo *repository.Repository, id backend.ID) (*Snapshot, error)
|
|||||||
return sn, nil
|
return sn, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// LoadAllSnapshots returns a list of all snapshots in the repo.
|
||||||
|
func LoadAllSnapshots(repo *repository.Repository) (snapshots []*Snapshot, err error) {
|
||||||
|
done := make(chan struct{})
|
||||||
|
defer close(done)
|
||||||
|
|
||||||
|
for id := range repo.List(backend.Snapshot, done) {
|
||||||
|
sn, err := LoadSnapshot(repo, id)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
snapshots = append(snapshots, sn)
|
||||||
|
}
|
||||||
|
|
||||||
|
return snapshots, nil
|
||||||
|
}
|
||||||
|
|
||||||
func (sn Snapshot) String() string {
|
func (sn Snapshot) String() string {
|
||||||
return fmt.Sprintf("<Snapshot of %v at %s>", sn.Paths, sn.Time)
|
return fmt.Sprintf("<Snapshot of %v at %s>", sn.Paths, sn.Time)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// ID retuns the snapshot's ID.
|
||||||
func (sn Snapshot) ID() *backend.ID {
|
func (sn Snapshot) ID() *backend.ID {
|
||||||
return sn.id
|
return sn.id
|
||||||
}
|
}
|
||||||
|
182
src/restic/testing.go
Normal file
182
src/restic/testing.go
Normal file
@ -0,0 +1,182 @@
|
|||||||
|
package restic
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
"io"
|
||||||
|
"math/rand"
|
||||||
|
"restic/backend"
|
||||||
|
"restic/pack"
|
||||||
|
"restic/repository"
|
||||||
|
"testing"
|
||||||
|
"time"
|
||||||
|
|
||||||
|
"github.com/restic/chunker"
|
||||||
|
)
|
||||||
|
|
||||||
|
type randReader struct {
|
||||||
|
rnd *rand.Rand
|
||||||
|
buf []byte
|
||||||
|
}
|
||||||
|
|
||||||
|
func newRandReader(rnd *rand.Rand) io.Reader {
|
||||||
|
return &randReader{rnd: rnd, buf: make([]byte, 0, 7)}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (rd *randReader) read(p []byte) (n int, err error) {
|
||||||
|
if len(p)%7 != 0 {
|
||||||
|
panic("invalid buffer length, not multiple of 7")
|
||||||
|
}
|
||||||
|
|
||||||
|
rnd := rd.rnd
|
||||||
|
for i := 0; i < len(p); i += 7 {
|
||||||
|
val := rnd.Int63()
|
||||||
|
|
||||||
|
p[i+0] = byte(val >> 0)
|
||||||
|
p[i+1] = byte(val >> 8)
|
||||||
|
p[i+2] = byte(val >> 16)
|
||||||
|
p[i+3] = byte(val >> 24)
|
||||||
|
p[i+4] = byte(val >> 32)
|
||||||
|
p[i+5] = byte(val >> 40)
|
||||||
|
p[i+6] = byte(val >> 48)
|
||||||
|
}
|
||||||
|
|
||||||
|
return len(p), nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (rd *randReader) Read(p []byte) (int, error) {
|
||||||
|
// first, copy buffer to p
|
||||||
|
pos := copy(p, rd.buf)
|
||||||
|
copy(rd.buf, rd.buf[pos:])
|
||||||
|
|
||||||
|
// shorten buf and p accordingly
|
||||||
|
rd.buf = rd.buf[:len(rd.buf)-pos]
|
||||||
|
p = p[pos:]
|
||||||
|
|
||||||
|
// if this is enough to fill p, return
|
||||||
|
if len(p) == 0 {
|
||||||
|
return pos, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// load multiple of 7 byte
|
||||||
|
l := (len(p) / 7) * 7
|
||||||
|
n, err := rd.read(p[:l])
|
||||||
|
pos += n
|
||||||
|
if err != nil {
|
||||||
|
return pos, err
|
||||||
|
}
|
||||||
|
p = p[n:]
|
||||||
|
|
||||||
|
// load 7 byte to temp buffer
|
||||||
|
rd.buf = rd.buf[:7]
|
||||||
|
n, err = rd.read(rd.buf)
|
||||||
|
if err != nil {
|
||||||
|
return pos, err
|
||||||
|
}
|
||||||
|
|
||||||
|
// copy the remaining bytes from the buffer to p
|
||||||
|
n = copy(p, rd.buf)
|
||||||
|
pos += n
|
||||||
|
|
||||||
|
// save the remaining bytes in rd.buf
|
||||||
|
n = copy(rd.buf, rd.buf[n:])
|
||||||
|
rd.buf = rd.buf[:n]
|
||||||
|
|
||||||
|
return pos, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// fakeFile returns a reader which yields deterministic pseudo-random data.
|
||||||
|
func fakeFile(t testing.TB, seed, size int64) io.Reader {
|
||||||
|
return io.LimitReader(newRandReader(rand.New(rand.NewSource(seed))), size)
|
||||||
|
}
|
||||||
|
|
||||||
|
// saveFile reads from rd and saves the blobs in the repository. The list of
|
||||||
|
// IDs is returned.
|
||||||
|
func saveFile(t testing.TB, repo *repository.Repository, rd io.Reader) (blobs backend.IDs) {
|
||||||
|
ch := chunker.New(rd, repo.Config.ChunkerPolynomial)
|
||||||
|
|
||||||
|
for {
|
||||||
|
chunk, err := ch.Next(getBuf())
|
||||||
|
if err == io.EOF {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
|
||||||
|
if err != nil {
|
||||||
|
t.Fatalf("unabel to save chunk in repo: %v", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
id, err := repo.SaveAndEncrypt(pack.Data, chunk.Data, nil)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatalf("error saving chunk: %v", err)
|
||||||
|
}
|
||||||
|
blobs = append(blobs, id)
|
||||||
|
}
|
||||||
|
|
||||||
|
return blobs
|
||||||
|
}
|
||||||
|
|
||||||
|
const maxFileSize = 1500000
|
||||||
|
const maxSeed = 100
|
||||||
|
|
||||||
|
// saveTree saves a tree of fake files in the repo and returns the ID.
|
||||||
|
func saveTree(t testing.TB, repo *repository.Repository, seed int64) backend.ID {
|
||||||
|
rnd := rand.NewSource(seed)
|
||||||
|
numNodes := int(rnd.Int63() % 64)
|
||||||
|
t.Logf("create %v nodes", numNodes)
|
||||||
|
|
||||||
|
var tree Tree
|
||||||
|
for i := 0; i < numNodes; i++ {
|
||||||
|
seed := rnd.Int63() % maxSeed
|
||||||
|
size := rnd.Int63() % maxFileSize
|
||||||
|
|
||||||
|
node := &Node{
|
||||||
|
Name: fmt.Sprintf("file-%v", seed),
|
||||||
|
Type: "file",
|
||||||
|
Mode: 0644,
|
||||||
|
Size: uint64(size),
|
||||||
|
}
|
||||||
|
|
||||||
|
node.Content = saveFile(t, repo, fakeFile(t, seed, size))
|
||||||
|
tree.Nodes = append(tree.Nodes, node)
|
||||||
|
}
|
||||||
|
|
||||||
|
id, err := repo.SaveJSON(pack.Tree, tree)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
return id
|
||||||
|
}
|
||||||
|
|
||||||
|
// TestCreateSnapshot creates a snapshot filled with fake data. The
|
||||||
|
// fake data is generated deterministically from the timestamp `at`, which is
|
||||||
|
// also used as the snapshot's timestamp.
|
||||||
|
func TestCreateSnapshot(t testing.TB, repo *repository.Repository, at time.Time) backend.ID {
|
||||||
|
fakedir := fmt.Sprintf("fakedir-at-%v", at.Format("2006-01-02 15:04:05"))
|
||||||
|
snapshot, err := NewSnapshot([]string{fakedir})
|
||||||
|
if err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
snapshot.Time = at
|
||||||
|
|
||||||
|
treeID := saveTree(t, repo, at.UnixNano())
|
||||||
|
snapshot.Tree = &treeID
|
||||||
|
|
||||||
|
id, err := repo.SaveJSONUnpacked(backend.Snapshot, snapshot)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
t.Logf("saved snapshot %v", id.Str())
|
||||||
|
|
||||||
|
err = repo.Flush()
|
||||||
|
if err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
err = repo.SaveIndex()
|
||||||
|
if err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
return id
|
||||||
|
}
|
71
src/restic/testing_test.go
Normal file
71
src/restic/testing_test.go
Normal file
@ -0,0 +1,71 @@
|
|||||||
|
package restic_test
|
||||||
|
|
||||||
|
import (
|
||||||
|
"restic"
|
||||||
|
"restic/checker"
|
||||||
|
"restic/repository"
|
||||||
|
"testing"
|
||||||
|
"time"
|
||||||
|
)
|
||||||
|
|
||||||
|
var testSnapshotTime = time.Unix(1460289341, 207401672)
|
||||||
|
|
||||||
|
const testCreateSnapshots = 3
|
||||||
|
|
||||||
|
func TestCreateSnapshot(t *testing.T) {
|
||||||
|
repo, cleanup := repository.TestRepository(t)
|
||||||
|
defer cleanup()
|
||||||
|
|
||||||
|
for i := 0; i < testCreateSnapshots; i++ {
|
||||||
|
restic.TestCreateSnapshot(t, repo, testSnapshotTime.Add(time.Duration(i)*time.Second))
|
||||||
|
}
|
||||||
|
|
||||||
|
snapshots, err := restic.LoadAllSnapshots(repo)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(snapshots) != testCreateSnapshots {
|
||||||
|
t.Fatalf("got %d snapshots, expected %d", len(snapshots), 1)
|
||||||
|
}
|
||||||
|
|
||||||
|
sn := snapshots[0]
|
||||||
|
if sn.Time.Before(testSnapshotTime) || sn.Time.After(testSnapshotTime.Add(testCreateSnapshots*time.Second)) {
|
||||||
|
t.Fatalf("timestamp %v is outside of the allowed time range", sn.Time)
|
||||||
|
}
|
||||||
|
|
||||||
|
if sn.Tree == nil {
|
||||||
|
t.Fatalf("tree id is nil")
|
||||||
|
}
|
||||||
|
|
||||||
|
if sn.Tree.IsNull() {
|
||||||
|
t.Fatalf("snapshot has zero tree ID")
|
||||||
|
}
|
||||||
|
|
||||||
|
chkr := checker.New(repo)
|
||||||
|
|
||||||
|
hints, errs := chkr.LoadIndex()
|
||||||
|
if len(errs) != 0 {
|
||||||
|
t.Fatalf("errors loading index: %v", errs)
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(hints) != 0 {
|
||||||
|
t.Fatalf("errors loading index: %v", hints)
|
||||||
|
}
|
||||||
|
|
||||||
|
done := make(chan struct{})
|
||||||
|
defer close(done)
|
||||||
|
errChan := make(chan error)
|
||||||
|
go chkr.Structure(errChan, done)
|
||||||
|
|
||||||
|
for err := range errChan {
|
||||||
|
t.Error(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
errChan = make(chan error)
|
||||||
|
go chkr.ReadData(nil, errChan, done)
|
||||||
|
|
||||||
|
for err := range errChan {
|
||||||
|
t.Error(err)
|
||||||
|
}
|
||||||
|
}
|
Loading…
Reference in New Issue
Block a user