restic/internal/restic/testing.go

210 lines
4.7 KiB
Go
Raw Normal View History

2016-04-10 14:52:15 +00:00
package restic
import (
2017-06-04 09:16:55 +00:00
"context"
"encoding/json"
2016-04-10 14:52:15 +00:00
"fmt"
"io"
"math/rand"
"testing"
"time"
"github.com/restic/chunker"
"golang.org/x/sync/errgroup"
2016-04-10 14:52:15 +00:00
)
// fakeFile returns a reader which yields deterministic pseudo-random data.
func fakeFile(seed, size int64) io.Reader {
return io.LimitReader(rand.New(rand.NewSource(seed)), size)
2016-04-10 14:52:15 +00:00
}
type fakeFileSystem struct {
t testing.TB
2016-08-31 18:29:54 +00:00
repo Repository
duplication float32
buf []byte
chunker *chunker.Chunker
rand *rand.Rand
}
2016-04-10 14:52:15 +00:00
// saveFile reads from rd and saves the blobs in the repository. The list of
// IDs is returned.
2017-06-04 09:16:55 +00:00
func (fs *fakeFileSystem) saveFile(ctx context.Context, rd io.Reader) (blobs IDs) {
if fs.buf == nil {
fs.buf = make([]byte, chunker.MaxSize)
}
2016-04-10 14:52:15 +00:00
if fs.chunker == nil {
fs.chunker = chunker.New(rd, fs.repo.Config().ChunkerPolynomial)
} else {
fs.chunker.Reset(rd, fs.repo.Config().ChunkerPolynomial)
}
blobs = IDs{}
2016-04-10 14:52:15 +00:00
for {
chunk, err := fs.chunker.Next(fs.buf)
if err == io.EOF {
2016-04-10 14:52:15 +00:00
break
}
if err != nil {
fs.t.Fatalf("unable to save chunk in repo: %v", err)
2016-04-10 14:52:15 +00:00
}
2016-08-31 18:29:54 +00:00
id := Hash(chunk.Data)
2020-11-05 21:18:00 +00:00
if !fs.blobIsKnown(BlobHandle{ID: id, Type: DataBlob}) {
2022-05-01 12:26:57 +00:00
_, _, _, err := fs.repo.SaveBlob(ctx, DataBlob, chunk.Data, id, true)
if err != nil {
fs.t.Fatalf("error saving chunk: %v", err)
}
2016-04-10 14:52:15 +00:00
}
2016-04-10 14:52:15 +00:00
blobs = append(blobs, id)
}
return blobs
}
2016-07-31 08:58:09 +00:00
const (
2018-03-11 20:42:39 +00:00
maxFileSize = 20000
maxSeed = 32
2018-03-11 20:42:39 +00:00
maxNodes = 15
2016-07-31 08:58:09 +00:00
)
2016-04-10 15:25:32 +00:00
func (fs *fakeFileSystem) treeIsKnown(tree *Tree) (bool, []byte, ID) {
data, err := json.Marshal(tree)
if err != nil {
fs.t.Fatalf("json.Marshal(tree) returned error: %v", err)
2016-09-03 18:55:22 +00:00
return false, nil, ID{}
}
data = append(data, '\n')
2016-08-31 18:29:54 +00:00
id := Hash(data)
2020-11-05 21:18:00 +00:00
return fs.blobIsKnown(BlobHandle{ID: id, Type: TreeBlob}), data, id
}
2020-11-05 21:18:00 +00:00
func (fs *fakeFileSystem) blobIsKnown(bh BlobHandle) bool {
if fs.rand.Float32() < fs.duplication {
return false
}
2020-11-05 21:18:00 +00:00
if fs.repo.Index().Has(bh) {
return true
}
return false
}
// saveTree saves a tree of fake files in the repo and returns the ID.
2017-06-04 09:16:55 +00:00
func (fs *fakeFileSystem) saveTree(ctx context.Context, seed int64, depth int) ID {
2016-04-10 14:52:15 +00:00
rnd := rand.NewSource(seed)
2016-07-31 08:58:09 +00:00
numNodes := int(rnd.Int63() % maxNodes)
2016-04-10 14:52:15 +00:00
var tree Tree
for i := 0; i < numNodes; i++ {
2016-07-31 08:58:09 +00:00
// randomly select the type of the node, either tree (p = 1/4) or file (p = 3/4).
if depth > 1 && rnd.Int63()%4 == 0 {
treeSeed := rnd.Int63() % maxSeed
2017-06-04 09:16:55 +00:00
id := fs.saveTree(ctx, treeSeed, depth-1)
2016-07-31 08:58:09 +00:00
node := &Node{
2016-09-01 19:20:03 +00:00
Name: fmt.Sprintf("dir-%v", treeSeed),
Type: "dir",
Mode: 0755,
Subtree: &id,
2016-07-31 08:58:09 +00:00
}
tree.Nodes = append(tree.Nodes, node)
continue
}
fileSeed := rnd.Int63() % maxSeed
fileSize := (maxFileSize / maxSeed) * fileSeed
2016-04-10 15:25:32 +00:00
node := &Node{
2016-09-01 19:20:03 +00:00
Name: fmt.Sprintf("file-%v", fileSeed),
Type: "file",
Mode: 0644,
Size: uint64(fileSize),
2016-04-10 15:25:32 +00:00
}
2016-04-10 14:52:15 +00:00
node.Content = fs.saveFile(ctx, fakeFile(fileSeed, fileSize))
2016-04-10 14:52:15 +00:00
tree.Nodes = append(tree.Nodes, node)
}
2016-09-03 18:55:22 +00:00
known, buf, id := fs.treeIsKnown(&tree)
if known {
return id
}
2022-05-01 12:26:57 +00:00
_, _, _, err := fs.repo.SaveBlob(ctx, TreeBlob, buf, id, false)
2016-04-10 14:52:15 +00:00
if err != nil {
fs.t.Fatal(err)
2016-04-10 14:52:15 +00:00
}
return id
}
// TestCreateSnapshot creates a snapshot filled with fake data. The
2016-04-10 14:52:15 +00:00
// fake data is generated deterministically from the timestamp `at`, which is
2016-07-31 08:58:09 +00:00
// also used as the snapshot's timestamp. The tree's depth can be specified
// with the parameter depth. The parameter duplication is a probability that
// the same blob will saved again.
2016-08-31 18:29:54 +00:00
func TestCreateSnapshot(t testing.TB, repo Repository, at time.Time, depth int, duplication float32) *Snapshot {
2016-07-31 08:58:09 +00:00
seed := at.Unix()
t.Logf("create fake snapshot at %s with seed %d", at, seed)
2016-04-10 14:52:15 +00:00
fakedir := fmt.Sprintf("fakedir-at-%v", at.Format("2006-01-02 15:04:05"))
snapshot, err := NewSnapshot([]string{fakedir}, []string{"test"}, "foo", time.Now())
2016-04-10 14:52:15 +00:00
if err != nil {
t.Fatal(err)
}
snapshot.Time = at
fs := fakeFileSystem{
t: t,
repo: repo,
duplication: duplication,
rand: rand.New(rand.NewSource(seed)),
}
var wg errgroup.Group
repo.StartPackUploader(context.TODO(), &wg)
2017-06-04 09:16:55 +00:00
treeID := fs.saveTree(context.TODO(), seed, depth)
2016-04-10 14:52:15 +00:00
snapshot.Tree = &treeID
err = repo.Flush(context.Background())
if err != nil {
t.Fatal(err)
}
id, err := SaveSnapshot(context.TODO(), repo, snapshot)
2016-04-10 14:52:15 +00:00
if err != nil {
t.Fatal(err)
}
snapshot.id = &id
2016-04-10 14:52:15 +00:00
t.Logf("saved snapshot %v", id.Str())
return snapshot
2016-04-10 14:52:15 +00:00
}
2016-08-31 18:29:54 +00:00
// TestParseID parses s as a ID and panics if that fails.
func TestParseID(s string) ID {
id, err := ParseID(s)
if err != nil {
panic(fmt.Sprintf("unable to parse string %q as ID: %v", s, err))
}
2016-08-31 18:29:54 +00:00
return id
}
// TestParseHandle parses s as a ID, panics if that fails and creates a BlobHandle with t.
func TestParseHandle(s string, t BlobType) BlobHandle {
return BlobHandle{ID: TestParseID(s), Type: t}
}