2
2
mirror of https://github.com/octoleo/restic.git synced 2024-11-26 23:06:32 +00:00

Merge pull request #745 from restic/fix-742

Fix restore/fuse with larger files
This commit is contained in:
Alexander Neumann 2017-01-24 12:12:35 +01:00
commit 17d7af6ccc
7 changed files with 170 additions and 136 deletions

View File

@ -7,6 +7,7 @@ import (
"fmt" "fmt"
"io" "io"
"io/ioutil" "io/ioutil"
mrand "math/rand"
"os" "os"
"path/filepath" "path/filepath"
"regexp" "regexp"
@ -739,6 +740,30 @@ func TestRestoreFilter(t *testing.T) {
}) })
} }
func TestRestore(t *testing.T) {
withTestEnvironment(t, func(env *testEnvironment, gopts GlobalOptions) {
testRunInit(t, gopts)
for i := 0; i < 10; i++ {
p := filepath.Join(env.testdata, fmt.Sprintf("foo/bar/testfile%v", i))
OK(t, os.MkdirAll(filepath.Dir(p), 0755))
OK(t, appendRandomData(p, uint(mrand.Intn(5<<21))))
}
opts := BackupOptions{}
testRunBackup(t, []string{env.testdata}, opts, gopts)
testRunCheck(t, gopts)
// Restore latest without any filters
restoredir := filepath.Join(env.base, "restore")
testRunRestoreLatest(t, gopts, restoredir, nil, "")
Assert(t, directoriesEqualContents(env.testdata, filepath.Join(restoredir, filepath.Base(env.testdata))),
"directories are not equal")
})
}
func TestRestoreLatest(t *testing.T) { func TestRestoreLatest(t *testing.T) {
withTestEnvironment(t, func(env *testEnvironment, gopts GlobalOptions) { withTestEnvironment(t, func(env *testEnvironment, gopts GlobalOptions) {

View File

@ -9,7 +9,6 @@ import (
"restic" "restic"
"restic/archiver" "restic/archiver"
"restic/backend/mem"
"restic/checker" "restic/checker"
"restic/repository" "restic/repository"
"restic/test" "restic/test"
@ -249,19 +248,15 @@ func induceError(data []byte) {
} }
func TestCheckerModifiedData(t *testing.T) { func TestCheckerModifiedData(t *testing.T) {
be := mem.New() repo, cleanup := repository.TestRepository(t)
defer cleanup()
repository.TestUseLowSecurityKDFParameters(t)
repo := repository.New(be)
test.OK(t, repo.Init(test.TestPassword))
arch := archiver.New(repo) arch := archiver.New(repo)
_, id, err := arch.Snapshot(nil, []string{"."}, nil, nil) _, id, err := arch.Snapshot(nil, []string{"."}, nil, nil)
test.OK(t, err) test.OK(t, err)
t.Logf("archived as %v", id.Str()) t.Logf("archived as %v", id.Str())
beError := &errorBackend{Backend: be} beError := &errorBackend{Backend: repo.Backend()}
checkRepo := repository.New(beError) checkRepo := repository.New(beError)
test.OK(t, checkRepo.SearchKey(test.TestPassword, 5)) test.OK(t, checkRepo.SearchKey(test.TestPassword, 5))

View File

@ -4,8 +4,6 @@
package fuse package fuse
import ( import (
"sync"
"restic/errors" "restic/errors"
"restic" "restic"
@ -35,29 +33,23 @@ type file struct {
node *restic.Node node *restic.Node
ownerIsRoot bool ownerIsRoot bool
sizes []uint sizes []int
blobs [][]byte blobs [][]byte
} }
const defaultBlobSize = 128 * 1024 const defaultBlobSize = 128 * 1024
var blobPool = sync.Pool{
New: func() interface{} {
return make([]byte, defaultBlobSize)
},
}
func newFile(repo BlobLoader, node *restic.Node, ownerIsRoot bool) (*file, error) { func newFile(repo BlobLoader, node *restic.Node, ownerIsRoot bool) (*file, error) {
debug.Log("create new file for %v with %d blobs", node.Name, len(node.Content)) debug.Log("create new file for %v with %d blobs", node.Name, len(node.Content))
var bytes uint64 var bytes uint64
sizes := make([]uint, len(node.Content)) sizes := make([]int, len(node.Content))
for i, id := range node.Content { for i, id := range node.Content {
size, err := repo.LookupBlobSize(id, restic.DataBlob) size, err := repo.LookupBlobSize(id, restic.DataBlob)
if err != nil { if err != nil {
return nil, err return nil, err
} }
sizes[i] = size sizes[i] = int(size)
bytes += uint64(size) bytes += uint64(size)
} }
@ -99,16 +91,7 @@ func (f *file) getBlobAt(i int) (blob []byte, err error) {
return f.blobs[i], nil return f.blobs[i], nil
} }
buf := blobPool.Get().([]byte) buf := restic.NewBlobBuffer(f.sizes[i])
buf = buf[:cap(buf)]
if uint(len(buf)) < f.sizes[i] {
if len(buf) > defaultBlobSize {
blobPool.Put(buf)
}
buf = make([]byte, f.sizes[i])
}
n, err := f.repo.LoadBlob(restic.DataBlob, f.node.Content[i], buf) n, err := f.repo.LoadBlob(restic.DataBlob, f.node.Content[i], buf)
if err != nil { if err != nil {
debug.Log("LoadBlob(%v, %v) failed: %v", f.node.Name, f.node.Content[i], err) debug.Log("LoadBlob(%v, %v) failed: %v", f.node.Name, f.node.Content[i], err)
@ -169,10 +152,7 @@ func (f *file) Read(ctx context.Context, req *fuse.ReadRequest, resp *fuse.ReadR
func (f *file) Release(ctx context.Context, req *fuse.ReleaseRequest) error { func (f *file) Release(ctx context.Context, req *fuse.ReleaseRequest) error {
for i := range f.blobs { for i := range f.blobs {
if f.blobs[i] != nil {
blobPool.Put(f.blobs[i])
f.blobs[i] = nil f.blobs[i] = nil
} }
}
return nil return nil
} }

View File

@ -9,7 +9,9 @@ import (
"testing" "testing"
"time" "time"
"restic/errors" "golang.org/x/net/context"
"restic/repository"
"bazil.org/fuse" "bazil.org/fuse"
@ -17,108 +19,96 @@ import (
. "restic/test" . "restic/test"
) )
type MockRepo struct { func testRead(t testing.TB, f *file, offset, length int, data []byte) {
blobs map[restic.ID][]byte ctx, cancel := context.WithCancel(context.Background())
} defer cancel()
func NewMockRepo(content map[restic.ID][]byte) *MockRepo {
return &MockRepo{blobs: content}
}
func (m *MockRepo) LookupBlobSize(id restic.ID, t restic.BlobType) (uint, error) {
buf, ok := m.blobs[id]
if !ok {
return 0, errors.New("blob not found")
}
return uint(len(buf)), nil
}
func (m *MockRepo) LoadBlob(t restic.BlobType, id restic.ID, buf []byte) (int, error) {
size, err := m.LookupBlobSize(id, t)
if err != nil {
return 0, err
}
if uint(len(buf)) < size {
return 0, errors.New("buffer too small")
}
buf = buf[:size]
copy(buf, m.blobs[id])
return int(size), nil
}
type MockContext struct{}
func (m MockContext) Deadline() (time.Time, bool) { return time.Now(), false }
func (m MockContext) Done() <-chan struct{} { return nil }
func (m MockContext) Err() error { return nil }
func (m MockContext) Value(key interface{}) interface{} { return nil }
var testContent = genTestContent()
var testContentLengths = []uint{
4646 * 1024,
655 * 1024,
378 * 1024,
8108 * 1024,
558 * 1024,
}
var testMaxFileSize uint
func genTestContent() map[restic.ID][]byte {
m := make(map[restic.ID][]byte)
for _, length := range testContentLengths {
buf := Random(int(length), int(length))
id := restic.Hash(buf)
m[id] = buf
testMaxFileSize += length
}
return m
}
const maxBufSize = 20 * 1024 * 1024
func testRead(t *testing.T, f *file, offset, length int, data []byte) {
ctx := MockContext{}
req := &fuse.ReadRequest{ req := &fuse.ReadRequest{
Offset: int64(offset), Offset: int64(offset),
Size: length, Size: length,
} }
resp := &fuse.ReadResponse{ resp := &fuse.ReadResponse{
Data: make([]byte, length), Data: data,
} }
OK(t, f.Read(ctx, req, resp)) OK(t, f.Read(ctx, req, resp))
} }
var offsetReadsTests = []struct { func firstSnapshotID(t testing.TB, repo restic.Repository) (first restic.ID) {
offset, length int done := make(chan struct{})
}{ defer close(done)
{0, 5 * 1024 * 1024}, for id := range repo.List(restic.SnapshotFile, done) {
{4000 * 1024, 1000 * 1024}, if first.IsNull() {
first = id
}
}
return first
}
func loadFirstSnapshot(t testing.TB, repo restic.Repository) *restic.Snapshot {
id := firstSnapshotID(t, repo)
sn, err := restic.LoadSnapshot(repo, id)
OK(t, err)
return sn
}
func loadTree(t testing.TB, repo restic.Repository, id restic.ID) *restic.Tree {
tree, err := repo.LoadTree(id)
OK(t, err)
return tree
} }
func TestFuseFile(t *testing.T) { func TestFuseFile(t *testing.T) {
repo := NewMockRepo(testContent) repo, cleanup := repository.TestRepository(t)
ctx := MockContext{} defer cleanup()
memfile := make([]byte, 0, maxBufSize) ctx, cancel := context.WithCancel(context.Background())
defer cancel()
timestamp, err := time.Parse(time.RFC3339, "2017-01-24T10:42:56+01:00")
OK(t, err)
restic.TestCreateSnapshot(t, repo, timestamp, 2, 0.1)
sn := loadFirstSnapshot(t, repo)
tree := loadTree(t, repo, *sn.Tree)
var content restic.IDs
for _, node := range tree.Nodes {
content = append(content, node.Content...)
}
t.Logf("tree loaded, content: %v", content)
var (
filesize uint64
memfile []byte
)
for _, id := range content {
size, err := repo.LookupBlobSize(id, restic.DataBlob)
OK(t, err)
filesize += uint64(size)
buf := restic.NewBlobBuffer(int(size))
n, err := repo.LoadBlob(restic.DataBlob, id, buf)
OK(t, err)
if uint(n) != size {
t.Fatalf("not enough bytes read for id %v: want %v, got %v", id.Str(), size, n)
}
if uint(len(buf)) != size {
t.Fatalf("buffer has wrong length for id %v: want %v, got %v", id.Str(), size, len(buf))
}
var ids restic.IDs
for id, buf := range repo.blobs {
ids = append(ids, id)
memfile = append(memfile, buf...) memfile = append(memfile, buf...)
} }
t.Logf("filesize is %v, memfile has size %v", filesize, len(memfile))
node := &restic.Node{ node := &restic.Node{
Name: "foo", Name: "foo",
Inode: 23, Inode: 23,
Mode: 0742, Mode: 0742,
Size: 42, Size: filesize,
Content: ids, Content: content,
} }
f, err := newFile(repo, node, false) f, err := newFile(repo, node, false)
OK(t, err) OK(t, err)
@ -131,28 +121,19 @@ func TestFuseFile(t *testing.T) {
Equals(t, node.Size, attr.Size) Equals(t, node.Size, attr.Size)
Equals(t, (node.Size/uint64(attr.BlockSize))+1, attr.Blocks) Equals(t, (node.Size/uint64(attr.BlockSize))+1, attr.Blocks)
for i, test := range offsetReadsTests {
b := memfile[test.offset : test.offset+test.length]
buf := make([]byte, test.length)
testRead(t, f, test.offset, test.length, buf)
if !bytes.Equal(b, buf) {
t.Errorf("test %d failed, wrong data returned", i)
}
}
for i := 0; i < 200; i++ { for i := 0; i < 200; i++ {
length := rand.Intn(int(testMaxFileSize) / 2) offset := rand.Intn(int(filesize))
offset := rand.Intn(int(testMaxFileSize)) length := rand.Intn(int(filesize)-offset) + 100
if length+offset > int(testMaxFileSize) {
diff := length + offset - int(testMaxFileSize)
length -= diff
}
b := memfile[offset : offset+length] b := memfile[offset : offset+length]
buf := make([]byte, length) buf := make([]byte, length)
testRead(t, f, offset, length, buf) testRead(t, f, offset, length, buf)
if !bytes.Equal(b, buf) { if !bytes.Equal(b, buf) {
t.Errorf("test %d failed (offset %d, length %d), wrong data returned", i, offset, length) t.Errorf("test %d failed, wrong data returned (offset %v, length %v)", i, offset, length)
} }
} }
OK(t, f.Release(ctx, nil))
} }

View File

@ -2,6 +2,7 @@ package restic
import ( import (
"io" "io"
"restic/debug"
) )
type backendReaderAt struct { type backendReaderAt struct {
@ -20,6 +21,7 @@ func ReaderAt(be Backend, h Handle) io.ReaderAt {
// ReadAt reads from the backend handle h at the given position. // ReadAt reads from the backend handle h at the given position.
func ReadAt(be Backend, h Handle, offset int64, p []byte) (n int, err error) { func ReadAt(be Backend, h Handle, offset int64, p []byte) (n int, err error) {
debug.Log("ReadAt(%v) at %v, len %v", h, offset, len(p))
rd, err := be.Load(h, len(p), offset) rd, err := be.Load(h, len(p), offset)
if err != nil { if err != nil {
return 0, err return 0, err
@ -31,5 +33,7 @@ func ReadAt(be Backend, h Handle, offset int64, p []byte) (n int, err error) {
err = e err = e
} }
debug.Log("ReadAt(%v) ReadFull returned %v bytes", h, n)
return n, err return n, err
} }

View File

@ -77,7 +77,7 @@ func (r *Repository) LoadAndDecrypt(t restic.FileType, id restic.ID) ([]byte, er
// pack from the backend, the result is stored in plaintextBuf, which must be // pack from the backend, the result is stored in plaintextBuf, which must be
// large enough to hold the complete blob. // large enough to hold the complete blob.
func (r *Repository) loadBlob(id restic.ID, t restic.BlobType, plaintextBuf []byte) (int, error) { func (r *Repository) loadBlob(id restic.ID, t restic.BlobType, plaintextBuf []byte) (int, error) {
debug.Log("load %v with id %v (buf %p, len %d)", t, id.Str(), plaintextBuf, len(plaintextBuf)) debug.Log("load %v with id %v (buf len %v, cap %d)", t, id.Str(), len(plaintextBuf), cap(plaintextBuf))
// lookup packs // lookup packs
blobs, err := r.idx.Lookup(id, t) blobs, err := r.idx.Lookup(id, t)
@ -96,7 +96,12 @@ func (r *Repository) loadBlob(id restic.ID, t restic.BlobType, plaintextBuf []by
// load blob from pack // load blob from pack
h := restic.Handle{Type: restic.DataFile, Name: blob.PackID.String()} h := restic.Handle{Type: restic.DataFile, Name: blob.PackID.String()}
plaintextBuf = plaintextBuf[:cap(plaintextBuf)]
if uint(cap(plaintextBuf)) < blob.Length {
return 0, errors.Errorf("buffer is too small: %v < %v", cap(plaintextBuf), blob.Length)
}
plaintextBuf = plaintextBuf[:blob.Length]
n, err := restic.ReadAt(r.be, h, int64(blob.Offset), plaintextBuf) n, err := restic.ReadAt(r.be, h, int64(blob.Offset), plaintextBuf)
if err != nil { if err != nil {
@ -520,15 +525,14 @@ func (r *Repository) Close() error {
// be large enough to hold the encrypted blob, since it is used as scratch // be large enough to hold the encrypted blob, since it is used as scratch
// space. // space.
func (r *Repository) LoadBlob(t restic.BlobType, id restic.ID, buf []byte) (int, error) { func (r *Repository) LoadBlob(t restic.BlobType, id restic.ID, buf []byte) (int, error) {
debug.Log("load blob %v into buf %p", id.Str(), buf) debug.Log("load blob %v into buf (len %v, cap %v)", id.Str(), len(buf), cap(buf))
size, err := r.idx.LookupSize(id, t) size, err := r.idx.LookupSize(id, t)
if err != nil { if err != nil {
return 0, err return 0, err
} }
buf = buf[:cap(buf)] if cap(buf) < restic.CiphertextLength(int(size)) {
if len(buf) < restic.CiphertextLength(int(size)) { return 0, errors.Errorf("buffer is too small for data blob (%d < %d)", cap(buf), restic.CiphertextLength(int(size)))
return 0, errors.Errorf("buffer is too small for data blob (%d < %d)", len(buf), restic.CiphertextLength(int(size)))
} }
n, err := r.loadBlob(id, t, buf) n, err := r.loadBlob(id, t, buf)

View File

@ -147,6 +147,51 @@ func BenchmarkLoadTree(t *testing.B) {
} }
} }
func TestLoadBlob(t *testing.T) {
repo, cleanup := repository.TestRepository(t)
defer cleanup()
length := 1000000
buf := restic.NewBlobBuffer(length)
_, err := io.ReadFull(rnd, buf)
OK(t, err)
id, err := repo.SaveBlob(restic.DataBlob, buf, restic.ID{})
OK(t, err)
OK(t, repo.Flush())
// first, test with buffers that are too small
for _, testlength := range []int{length - 20, length, restic.CiphertextLength(length) - 1} {
buf = make([]byte, 0, testlength)
n, err := repo.LoadBlob(restic.DataBlob, id, buf)
if err == nil {
t.Errorf("LoadBlob() did not return an error for a buffer that is too small to hold the blob")
continue
}
if n != 0 {
t.Errorf("LoadBlob() returned an error and n > 0")
continue
}
}
// then use buffers that are large enough
base := restic.CiphertextLength(length)
for _, testlength := range []int{base, base + 7, base + 15, base + 1000} {
buf = make([]byte, 0, testlength)
n, err := repo.LoadBlob(restic.DataBlob, id, buf)
if err != nil {
t.Errorf("LoadBlob() returned an error for buffer size %v: %v", testlength, err)
continue
}
if n != length {
t.Errorf("LoadBlob() returned the wrong number of bytes: want %v, got %v", length, n)
continue
}
}
}
func BenchmarkLoadBlob(b *testing.B) { func BenchmarkLoadBlob(b *testing.B) {
repo, cleanup := repository.TestRepository(b) repo, cleanup := repository.TestRepository(b)
defer cleanup() defer cleanup()