2
2
mirror of https://github.com/octoleo/restic.git synced 2024-11-17 02:25:12 +00:00
restic/archiver_test.go

277 lines
6.3 KiB
Go
Raw Normal View History

2014-12-05 20:45:49 +00:00
package restic_test
2014-11-16 20:41:05 +00:00
import (
"bytes"
"flag"
2014-11-16 20:41:05 +00:00
"io"
"math/rand"
"testing"
2014-12-05 20:45:49 +00:00
"github.com/restic/restic"
2015-02-17 21:39:44 +00:00
"github.com/restic/restic/backend"
2014-12-05 20:45:49 +00:00
"github.com/restic/restic/chunker"
2014-11-16 20:41:05 +00:00
)
2015-03-10 14:58:23 +00:00
var benchArchiveDirectory = flag.String("test.benchdir", ".", "benchmark archiving a real directory (default: .)")
2014-11-16 20:41:05 +00:00
func get_random(seed, count int) []byte {
buf := make([]byte, count)
rnd := rand.New(rand.NewSource(int64(seed)))
for i := 0; i < count; i++ {
buf[i] = byte(rnd.Uint32())
2014-11-16 20:41:05 +00:00
}
return buf
}
func randomReader(seed, size int) *bytes.Reader {
return bytes.NewReader(get_random(seed, size))
}
const bufSize = chunker.MiB
2015-02-17 19:02:43 +00:00
type Rdr interface {
io.ReadSeeker
io.ReaderAt
2015-02-17 19:02:43 +00:00
}
2015-02-17 19:02:43 +00:00
func benchmarkChunkEncrypt(b testing.TB, buf []byte, rd Rdr, key *restic.Key) {
ch := restic.GetChunker("BenchmarkChunkEncrypt")
rd.Seek(0, 0)
ch.Reset(rd)
for {
chunk, err := ch.Next()
if err == io.EOF {
break
}
ok(b, err)
2015-02-17 19:02:43 +00:00
// reduce length of buf
buf = buf[:chunk.Length]
n, err := io.ReadFull(chunk.Reader(rd), buf)
ok(b, err)
assert(b, uint(n) == chunk.Length, "invalid length: got %d, expected %d", n, chunk.Length)
2015-02-17 19:02:43 +00:00
_, err = key.Encrypt(buf, buf)
ok(b, err)
}
restic.FreeChunker("BenchmarkChunkEncrypt", ch)
}
2014-11-16 20:41:05 +00:00
func BenchmarkChunkEncrypt(b *testing.B) {
data := get_random(23, 10<<20) // 10MiB
rd := bytes.NewReader(data)
2014-11-16 20:41:05 +00:00
be := setupBackend(b)
defer teardownBackend(b, be)
key := setupKey(b, be, "geheim")
2015-02-17 19:02:43 +00:00
buf := restic.GetChunkBuf("BenchmarkChunkEncrypt")
2014-11-16 20:41:05 +00:00
b.ResetTimer()
b.SetBytes(int64(len(data)))
for i := 0; i < b.N; i++ {
2015-02-17 19:02:43 +00:00
benchmarkChunkEncrypt(b, buf, rd, key)
}
2014-11-16 20:41:05 +00:00
2015-02-17 19:02:43 +00:00
restic.FreeChunkBuf("BenchmarkChunkEncrypt", buf)
}
2015-02-17 19:02:43 +00:00
func benchmarkChunkEncryptP(b *testing.PB, buf []byte, rd Rdr, key *restic.Key) {
ch := restic.GetChunker("BenchmarkChunkEncryptP")
rd.Seek(0, 0)
ch.Reset(rd)
2014-11-16 20:41:05 +00:00
for {
chunk, err := ch.Next()
if err == io.EOF {
break
}
2014-11-16 20:41:05 +00:00
// reduce length of chunkBuf
2015-02-17 19:02:43 +00:00
buf = buf[:chunk.Length]
io.ReadFull(chunk.Reader(rd), buf)
key.Encrypt(buf, buf)
}
2014-11-16 20:41:05 +00:00
restic.FreeChunker("BenchmarkChunkEncryptP", ch)
}
func BenchmarkChunkEncryptParallel(b *testing.B) {
be := setupBackend(b)
defer teardownBackend(b, be)
key := setupKey(b, be, "geheim")
data := get_random(23, 10<<20) // 10MiB
2015-02-17 19:02:43 +00:00
buf := restic.GetChunkBuf("BenchmarkChunkEncryptParallel")
b.ResetTimer()
b.SetBytes(int64(len(data)))
b.RunParallel(func(pb *testing.PB) {
for pb.Next() {
rd := bytes.NewReader(data)
2015-02-17 19:02:43 +00:00
benchmarkChunkEncryptP(pb, buf, rd, key)
2014-11-16 20:41:05 +00:00
}
})
2015-02-17 19:02:43 +00:00
restic.FreeChunkBuf("BenchmarkChunkEncryptParallel", buf)
2014-11-16 20:41:05 +00:00
}
func BenchmarkArchiveDirectory(b *testing.B) {
if *benchArchiveDirectory == "" {
b.Skip("benchdir not set, skipping BenchmarkArchiveDirectory")
}
be := setupBackend(b)
defer teardownBackend(b, be)
key := setupKey(b, be, "geheim")
server := restic.NewServerWithKey(be, key)
2015-02-21 13:23:49 +00:00
arch, err := restic.NewArchiver(server)
ok(b, err)
2015-03-02 13:48:47 +00:00
_, id, err := arch.Snapshot(nil, []string{*benchArchiveDirectory}, nil)
b.Logf("snapshot archived as %v", id)
}
2015-02-17 21:39:44 +00:00
2015-03-09 21:58:17 +00:00
func snapshot(t testing.TB, server restic.Server, path string, parent backend.ID) *restic.Snapshot {
2015-02-21 13:23:49 +00:00
arch, err := restic.NewArchiver(server)
2015-02-17 21:39:44 +00:00
ok(t, err)
2015-03-09 21:26:39 +00:00
ok(t, arch.Preload())
2015-03-09 21:58:17 +00:00
sn, _, err := arch.Snapshot(nil, []string{path}, parent)
2015-02-17 21:39:44 +00:00
ok(t, err)
return sn
}
2015-03-10 14:58:23 +00:00
func countBlobs(t testing.TB, server restic.Server) (trees int, data int) {
list, err := server.List(backend.Tree)
ok(t, err)
2015-02-17 21:39:44 +00:00
2015-03-10 14:58:23 +00:00
trees = len(list)
list, err = server.List(backend.Data)
2015-02-17 21:39:44 +00:00
ok(t, err)
2015-03-10 14:58:23 +00:00
data = len(list)
return
2015-02-17 21:39:44 +00:00
}
func archiveWithPreload(t testing.TB) {
2015-02-17 22:40:37 +00:00
if *benchArchiveDirectory == "" {
t.Skip("benchdir not set, skipping TestArchiverPreload")
}
2015-02-17 21:39:44 +00:00
be := setupBackend(t)
defer teardownBackend(t, be)
key := setupKey(t, be, "geheim")
server := restic.NewServerWithKey(be, key)
// archive a few files
2015-03-09 21:58:17 +00:00
sn := snapshot(t, server, *benchArchiveDirectory, nil)
t.Logf("archived snapshot %v", sn.ID().Str())
2015-02-17 21:39:44 +00:00
// get archive stats
2015-03-10 14:58:23 +00:00
beforeTrees, beforeData := countBlobs(t, server)
t.Logf("found %v trees, %v data blobs", beforeTrees, beforeData)
2015-02-17 21:39:44 +00:00
2015-03-09 21:58:17 +00:00
// archive the same files again, without parent snapshot
sn2 := snapshot(t, server, *benchArchiveDirectory, nil)
t.Logf("archived snapshot %v", sn2.ID().Str())
2015-02-17 21:39:44 +00:00
// get archive stats
2015-03-10 14:58:23 +00:00
afterTrees2, afterData2 := countBlobs(t, server)
t.Logf("found %v trees, %v data blobs", afterTrees2, afterData2)
2015-02-17 21:39:44 +00:00
2015-03-10 14:58:23 +00:00
// if there are more blobs, something is wrong
if afterData2 > beforeData {
t.Fatalf("TestArchiverPreload: too many data blobs in repository: before %d, after %d",
beforeData, afterData2)
2015-03-09 21:58:17 +00:00
}
// archive the same files again, with a parent snapshot
sn3 := snapshot(t, server, *benchArchiveDirectory, sn2.ID())
t.Logf("archived snapshot %v, parent %v", sn3.ID().Str(), sn2.ID().Str())
// get archive stats
2015-03-10 14:58:23 +00:00
afterTrees3, afterData3 := countBlobs(t, server)
t.Logf("found %v trees, %v data blobs", afterTrees3, afterData3)
2015-03-09 21:58:17 +00:00
2015-03-10 14:58:23 +00:00
// if there are more blobs, something is wrong
if afterData3 > beforeData {
t.Fatalf("TestArchiverPreload: too many data blobs in repository: before %d, after %d",
beforeData, afterData3)
}
}
func TestArchivePreload(t *testing.T) {
archiveWithPreload(t)
}
func BenchmarkPreload(t *testing.B) {
if *benchArchiveDirectory == "" {
t.Skip("benchdir not set, skipping TestArchiverPreload")
}
be := setupBackend(t)
defer teardownBackend(t, be)
key := setupKey(t, be, "geheim")
server := restic.NewServerWithKey(be, key)
// archive a few files
2015-02-21 13:23:49 +00:00
arch, err := restic.NewArchiver(server)
ok(t, err)
2015-03-02 13:48:47 +00:00
sn, _, err := arch.Snapshot(nil, []string{*benchArchiveDirectory}, nil)
ok(t, err)
t.Logf("archived snapshot %v", sn.ID())
// start benchmark
t.ResetTimer()
for i := 0; i < t.N; i++ {
// create new archiver and preload
2015-02-21 13:23:49 +00:00
arch2, err := restic.NewArchiver(server)
ok(t, err)
2015-03-09 21:26:39 +00:00
ok(t, arch2.Preload())
2015-02-17 21:39:44 +00:00
}
}
2015-02-21 23:09:57 +00:00
func BenchmarkLoadTree(t *testing.B) {
if *benchArchiveDirectory == "" {
t.Skip("benchdir not set, skipping TestArchiverPreload")
}
be := setupBackend(t)
defer teardownBackend(t, be)
key := setupKey(t, be, "geheim")
server := restic.NewServerWithKey(be, key)
// archive a few files
arch, err := restic.NewArchiver(server)
ok(t, err)
2015-03-02 13:48:47 +00:00
sn, _, err := arch.Snapshot(nil, []string{*benchArchiveDirectory}, nil)
2015-02-21 23:09:57 +00:00
ok(t, err)
t.Logf("archived snapshot %v", sn.ID())
// start benchmark
t.ResetTimer()
list, err := server.List(backend.Tree)
ok(t, err)
list = list[:10]
for i := 0; i < t.N; i++ {
_, err := restic.LoadTree(server, list[0])
ok(t, err)
}
}