2014-12-05 20:45:49 +00:00
|
|
|
package restic
|
2014-09-23 20:39:12 +00:00
|
|
|
|
|
|
|
import (
|
2015-01-10 22:40:10 +00:00
|
|
|
"encoding/json"
|
2014-11-30 21:49:14 +00:00
|
|
|
"fmt"
|
2014-11-17 22:28:51 +00:00
|
|
|
"io"
|
2014-09-23 20:39:12 +00:00
|
|
|
"os"
|
2015-02-21 13:23:49 +00:00
|
|
|
"path/filepath"
|
2015-03-02 13:48:47 +00:00
|
|
|
"sort"
|
2014-11-16 21:50:20 +00:00
|
|
|
"sync"
|
2015-10-12 21:59:17 +00:00
|
|
|
"time"
|
2014-09-23 20:39:12 +00:00
|
|
|
|
2016-08-21 15:46:23 +00:00
|
|
|
"github.com/pkg/errors"
|
|
|
|
|
2016-02-14 14:29:28 +00:00
|
|
|
"restic/debug"
|
2016-03-28 13:26:04 +00:00
|
|
|
"restic/fs"
|
2016-02-14 14:29:28 +00:00
|
|
|
"restic/pack"
|
|
|
|
"restic/pipe"
|
2015-05-01 23:29:54 +00:00
|
|
|
|
2016-02-22 20:09:21 +00:00
|
|
|
"github.com/restic/chunker"
|
2014-09-23 20:39:12 +00:00
|
|
|
)
|
|
|
|
|
2014-11-16 21:50:20 +00:00
|
|
|
const (
|
2015-05-02 14:39:24 +00:00
|
|
|
maxConcurrentBlobs = 32
|
|
|
|
maxConcurrency = 10
|
2014-11-16 21:50:20 +00:00
|
|
|
)
|
|
|
|
|
2015-04-30 01:41:51 +00:00
|
|
|
var archiverAbortOnAllErrors = func(str string, fi os.FileInfo, err error) error { return err }
|
|
|
|
var archiverAllowAllFiles = func(string, os.FileInfo) bool { return true }
|
|
|
|
|
2015-05-02 13:51:40 +00:00
|
|
|
// Archiver is used to backup a set of directories.
|
2014-09-23 20:39:12 +00:00
|
|
|
type Archiver struct {
|
2016-08-31 18:29:54 +00:00
|
|
|
repo Repository
|
2016-02-01 22:50:56 +00:00
|
|
|
knownBlobs struct {
|
2016-08-31 18:29:54 +00:00
|
|
|
IDSet
|
2016-02-01 22:50:56 +00:00
|
|
|
sync.Mutex
|
|
|
|
}
|
2014-09-23 20:39:12 +00:00
|
|
|
|
2014-11-22 21:05:39 +00:00
|
|
|
blobToken chan struct{}
|
2014-11-16 21:50:20 +00:00
|
|
|
|
2015-07-19 22:13:39 +00:00
|
|
|
Error func(dir string, fi os.FileInfo, err error) error
|
|
|
|
SelectFilter pipe.SelectFunc
|
2015-07-22 20:33:23 +00:00
|
|
|
Excludes []string
|
2014-09-23 20:39:12 +00:00
|
|
|
}
|
|
|
|
|
2015-05-02 13:51:40 +00:00
|
|
|
// NewArchiver returns a new archiver.
|
2016-08-31 18:29:54 +00:00
|
|
|
func NewArchiver(repo Repository) *Archiver {
|
2014-11-16 21:50:20 +00:00
|
|
|
arch := &Archiver{
|
2015-05-09 11:32:52 +00:00
|
|
|
repo: repo,
|
2014-11-22 21:05:39 +00:00
|
|
|
blobToken: make(chan struct{}, maxConcurrentBlobs),
|
2016-02-01 22:50:56 +00:00
|
|
|
knownBlobs: struct {
|
2016-08-31 18:29:54 +00:00
|
|
|
IDSet
|
2016-02-01 22:50:56 +00:00
|
|
|
sync.Mutex
|
|
|
|
}{
|
2016-08-31 18:29:54 +00:00
|
|
|
IDSet: NewIDSet(),
|
2016-02-01 22:50:56 +00:00
|
|
|
},
|
2014-11-16 21:50:20 +00:00
|
|
|
}
|
|
|
|
|
2014-11-22 21:05:39 +00:00
|
|
|
for i := 0; i < maxConcurrentBlobs; i++ {
|
|
|
|
arch.blobToken <- struct{}{}
|
|
|
|
}
|
|
|
|
|
2015-04-30 01:41:51 +00:00
|
|
|
arch.Error = archiverAbortOnAllErrors
|
2015-07-19 22:13:39 +00:00
|
|
|
arch.SelectFilter = archiverAllowAllFiles
|
2014-09-23 20:39:12 +00:00
|
|
|
|
2015-04-30 01:41:51 +00:00
|
|
|
return arch
|
2015-01-10 22:40:10 +00:00
|
|
|
}
|
|
|
|
|
2016-02-01 22:50:56 +00:00
|
|
|
// isKnownBlob returns true iff the blob is not yet in the list of known blobs.
|
|
|
|
// When the blob is not known, false is returned and the blob is added to the
|
|
|
|
// list. This means that the caller false is returned to is responsible to save
|
|
|
|
// the blob to the backend.
|
2016-08-31 18:29:54 +00:00
|
|
|
func (arch *Archiver) isKnownBlob(id ID, t pack.BlobType) bool {
|
2016-02-01 22:50:56 +00:00
|
|
|
arch.knownBlobs.Lock()
|
|
|
|
defer arch.knownBlobs.Unlock()
|
|
|
|
|
|
|
|
if arch.knownBlobs.Has(id) {
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
|
|
|
|
arch.knownBlobs.Insert(id)
|
|
|
|
|
2016-08-03 20:38:05 +00:00
|
|
|
_, err := arch.repo.Index().Lookup(id, t)
|
2016-02-01 22:50:56 +00:00
|
|
|
if err == nil {
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
2015-05-09 11:35:55 +00:00
|
|
|
// Save stores a blob read from rd in the repository.
|
2016-08-31 18:29:54 +00:00
|
|
|
func (arch *Archiver) Save(t pack.BlobType, data []byte, id ID) error {
|
2015-01-14 21:08:48 +00:00
|
|
|
debug.Log("Archiver.Save", "Save(%v, %v)\n", t, id.Str())
|
2015-01-10 22:40:10 +00:00
|
|
|
|
2016-08-03 20:38:05 +00:00
|
|
|
if arch.isKnownBlob(id, pack.Data) {
|
2016-02-01 22:50:56 +00:00
|
|
|
debug.Log("Archiver.Save", "blob %v is known\n", id.Str())
|
2015-04-26 15:44:38 +00:00
|
|
|
return nil
|
2015-01-05 20:40:43 +00:00
|
|
|
}
|
2014-09-23 20:39:12 +00:00
|
|
|
|
2016-05-08 11:13:29 +00:00
|
|
|
_, err := arch.repo.SaveAndEncrypt(t, data, &id)
|
2015-04-26 15:44:38 +00:00
|
|
|
if err != nil {
|
|
|
|
debug.Log("Archiver.Save", "Save(%v, %v): error %v\n", t, id.Str(), err)
|
|
|
|
return err
|
2014-09-23 20:39:12 +00:00
|
|
|
}
|
|
|
|
|
2015-04-26 15:44:38 +00:00
|
|
|
debug.Log("Archiver.Save", "Save(%v, %v): new blob\n", t, id.Str())
|
|
|
|
return nil
|
2014-09-23 20:39:12 +00:00
|
|
|
}
|
|
|
|
|
2015-05-09 11:35:55 +00:00
|
|
|
// SaveTreeJSON stores a tree in the repository.
|
2016-08-31 18:29:54 +00:00
|
|
|
func (arch *Archiver) SaveTreeJSON(item interface{}) (ID, error) {
|
2015-01-10 22:40:10 +00:00
|
|
|
data, err := json.Marshal(item)
|
2014-09-23 20:39:12 +00:00
|
|
|
if err != nil {
|
2016-08-31 18:29:54 +00:00
|
|
|
return ID{}, errors.Wrap(err, "Marshal")
|
2014-09-23 20:39:12 +00:00
|
|
|
}
|
2015-04-30 01:41:51 +00:00
|
|
|
data = append(data, '\n')
|
2014-09-23 20:39:12 +00:00
|
|
|
|
2015-01-10 22:40:10 +00:00
|
|
|
// check if tree has been saved before
|
2016-08-31 18:29:54 +00:00
|
|
|
id := Hash(data)
|
2016-08-03 20:38:05 +00:00
|
|
|
if arch.isKnownBlob(id, pack.Tree) {
|
2015-04-26 15:44:38 +00:00
|
|
|
return id, nil
|
2015-01-10 22:40:10 +00:00
|
|
|
}
|
2014-09-23 20:39:12 +00:00
|
|
|
|
2015-05-09 11:32:52 +00:00
|
|
|
return arch.repo.SaveJSON(pack.Tree, item)
|
2014-09-23 20:39:12 +00:00
|
|
|
}
|
|
|
|
|
2016-03-28 13:26:04 +00:00
|
|
|
func (arch *Archiver) reloadFileIfChanged(node *Node, file fs.File) (*Node, error) {
|
2015-01-04 21:39:30 +00:00
|
|
|
fi, err := file.Stat()
|
|
|
|
if err != nil {
|
2016-08-29 19:38:34 +00:00
|
|
|
return nil, errors.Wrap(err, "Stat")
|
2015-01-04 21:39:30 +00:00
|
|
|
}
|
|
|
|
|
2015-05-02 05:05:49 +00:00
|
|
|
if fi.ModTime() == node.ModTime {
|
|
|
|
return node, nil
|
2014-11-17 22:28:51 +00:00
|
|
|
}
|
|
|
|
|
2015-05-02 05:05:49 +00:00
|
|
|
err = arch.Error(node.path, fi, errors.New("file has changed"))
|
|
|
|
if err != nil {
|
|
|
|
return nil, err
|
2015-04-26 15:44:38 +00:00
|
|
|
}
|
2014-11-17 22:28:51 +00:00
|
|
|
|
2015-05-02 05:05:49 +00:00
|
|
|
node, err = NodeFromFileInfo(node.path, fi)
|
|
|
|
if err != nil {
|
|
|
|
debug.Log("Archiver.SaveFile", "NodeFromFileInfo returned error for %v: %v", node.path, err)
|
|
|
|
return nil, err
|
|
|
|
}
|
2014-11-17 22:28:51 +00:00
|
|
|
|
2015-05-02 05:05:49 +00:00
|
|
|
return node, nil
|
|
|
|
}
|
2014-11-30 21:49:14 +00:00
|
|
|
|
2015-05-02 05:05:49 +00:00
|
|
|
type saveResult struct {
|
2016-08-31 18:29:54 +00:00
|
|
|
id ID
|
2015-05-02 05:05:49 +00:00
|
|
|
bytes uint64
|
|
|
|
}
|
2014-11-22 21:05:39 +00:00
|
|
|
|
2016-03-28 13:26:04 +00:00
|
|
|
func (arch *Archiver) saveChunk(chunk chunker.Chunk, p *Progress, token struct{}, file fs.File, resultChannel chan<- saveResult) {
|
2016-02-22 20:09:21 +00:00
|
|
|
defer freeBuf(chunk.Data)
|
2015-07-25 15:05:45 +00:00
|
|
|
|
2016-08-31 18:29:54 +00:00
|
|
|
id := Hash(chunk.Data)
|
2016-05-08 11:13:29 +00:00
|
|
|
err := arch.Save(pack.Data, chunk.Data, id)
|
2015-05-02 05:05:49 +00:00
|
|
|
// TODO handle error
|
|
|
|
if err != nil {
|
|
|
|
panic(err)
|
|
|
|
}
|
2014-11-17 22:28:51 +00:00
|
|
|
|
2015-05-02 05:05:49 +00:00
|
|
|
p.Report(Stat{Bytes: uint64(chunk.Length)})
|
|
|
|
arch.blobToken <- token
|
2015-07-25 15:05:45 +00:00
|
|
|
resultChannel <- saveResult{id: id, bytes: uint64(chunk.Length)}
|
2015-05-02 05:05:49 +00:00
|
|
|
}
|
2014-11-22 21:05:39 +00:00
|
|
|
|
2015-05-02 05:05:49 +00:00
|
|
|
func waitForResults(resultChannels [](<-chan saveResult)) ([]saveResult, error) {
|
|
|
|
results := []saveResult{}
|
2014-11-17 22:28:51 +00:00
|
|
|
|
2015-05-02 05:05:49 +00:00
|
|
|
for _, ch := range resultChannels {
|
2015-04-26 15:44:38 +00:00
|
|
|
results = append(results, <-ch)
|
2015-02-08 21:54:45 +00:00
|
|
|
}
|
2014-11-30 21:49:14 +00:00
|
|
|
|
2015-05-02 05:05:49 +00:00
|
|
|
if len(results) != len(resultChannels) {
|
2016-08-21 15:48:36 +00:00
|
|
|
return nil, errors.Errorf("chunker returned %v chunks, but only %v blobs saved", len(resultChannels), len(results))
|
2014-09-23 20:39:12 +00:00
|
|
|
}
|
|
|
|
|
2015-05-02 05:05:49 +00:00
|
|
|
return results, nil
|
|
|
|
}
|
2014-12-07 12:30:16 +00:00
|
|
|
|
2015-05-02 05:05:49 +00:00
|
|
|
func updateNodeContent(node *Node, results []saveResult) error {
|
2015-02-08 21:54:45 +00:00
|
|
|
debug.Log("Archiver.Save", "checking size for file %s", node.path)
|
2015-05-02 05:05:49 +00:00
|
|
|
|
|
|
|
var bytes uint64
|
2016-08-31 18:29:54 +00:00
|
|
|
node.Content = make([]ID, len(results))
|
2015-05-02 05:05:49 +00:00
|
|
|
|
2015-04-26 15:44:38 +00:00
|
|
|
for i, b := range results {
|
|
|
|
node.Content[i] = b.id
|
|
|
|
bytes += b.bytes
|
2015-02-08 21:54:45 +00:00
|
|
|
|
2015-04-26 15:44:38 +00:00
|
|
|
debug.Log("Archiver.Save", " adding blob %s, %d bytes", b.id.Str(), b.bytes)
|
2014-12-07 12:30:16 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
if bytes != node.Size {
|
2016-08-21 15:48:36 +00:00
|
|
|
return errors.Errorf("errors saving node %q: saved %d bytes, wanted %d bytes", node.path, bytes, node.Size)
|
2014-09-23 20:39:12 +00:00
|
|
|
}
|
|
|
|
|
2015-04-26 15:44:38 +00:00
|
|
|
debug.Log("Archiver.SaveFile", "SaveFile(%q): %v blobs\n", node.path, len(results))
|
2015-01-10 22:40:10 +00:00
|
|
|
|
2015-04-26 15:44:38 +00:00
|
|
|
return nil
|
2014-09-23 20:39:12 +00:00
|
|
|
}
|
|
|
|
|
2015-05-02 05:05:49 +00:00
|
|
|
// SaveFile stores the content of the file on the backend as a Blob by calling
|
|
|
|
// Save for each chunk.
|
|
|
|
func (arch *Archiver) SaveFile(p *Progress, node *Node) error {
|
2016-03-28 13:26:04 +00:00
|
|
|
file, err := fs.Open(node.path)
|
2015-05-02 05:05:49 +00:00
|
|
|
defer file.Close()
|
|
|
|
if err != nil {
|
2016-08-29 19:38:34 +00:00
|
|
|
return errors.Wrap(err, "Open")
|
2015-05-02 05:05:49 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
node, err = arch.reloadFileIfChanged(node, file)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
2016-08-31 18:29:54 +00:00
|
|
|
chnker := chunker.New(file, arch.repo.Config().ChunkerPolynomial())
|
2015-05-02 05:05:49 +00:00
|
|
|
resultChannels := [](<-chan saveResult){}
|
|
|
|
|
|
|
|
for {
|
2016-02-22 20:09:21 +00:00
|
|
|
chunk, err := chnker.Next(getBuf())
|
2016-08-29 17:18:57 +00:00
|
|
|
if errors.Cause(err) == io.EOF {
|
2015-05-02 05:05:49 +00:00
|
|
|
break
|
|
|
|
}
|
|
|
|
|
|
|
|
if err != nil {
|
2016-08-29 19:38:34 +00:00
|
|
|
return errors.Wrap(err, "chunker.Next")
|
2015-05-02 05:05:49 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
resCh := make(chan saveResult, 1)
|
|
|
|
go arch.saveChunk(chunk, p, <-arch.blobToken, file, resCh)
|
|
|
|
resultChannels = append(resultChannels, resCh)
|
|
|
|
}
|
|
|
|
|
|
|
|
results, err := waitForResults(resultChannels)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
err = updateNodeContent(node, results)
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
2015-03-02 13:48:47 +00:00
|
|
|
func (arch *Archiver) fileWorker(wg *sync.WaitGroup, p *Progress, done <-chan struct{}, entCh <-chan pipe.Entry) {
|
2015-03-07 10:53:32 +00:00
|
|
|
defer func() {
|
|
|
|
debug.Log("Archiver.fileWorker", "done")
|
|
|
|
wg.Done()
|
|
|
|
}()
|
2015-03-02 13:48:47 +00:00
|
|
|
for {
|
|
|
|
select {
|
|
|
|
case e, ok := <-entCh:
|
|
|
|
if !ok {
|
|
|
|
// channel is closed
|
|
|
|
return
|
|
|
|
}
|
2015-01-11 13:09:44 +00:00
|
|
|
|
2015-03-07 10:53:32 +00:00
|
|
|
debug.Log("Archiver.fileWorker", "got job %v", e)
|
|
|
|
|
2015-03-15 11:20:30 +00:00
|
|
|
// check for errors
|
|
|
|
if e.Error() != nil {
|
|
|
|
debug.Log("Archiver.fileWorker", "job %v has errors: %v", e.Path(), e.Error())
|
2015-03-15 14:48:05 +00:00
|
|
|
// TODO: integrate error reporting
|
|
|
|
fmt.Fprintf(os.Stderr, "error for %v: %v\n", e.Path(), e.Error())
|
|
|
|
// ignore this file
|
|
|
|
e.Result() <- nil
|
2015-04-26 01:54:35 +00:00
|
|
|
p.Report(Stat{Errors: 1})
|
2015-03-15 14:48:05 +00:00
|
|
|
continue
|
2015-03-15 11:20:30 +00:00
|
|
|
}
|
|
|
|
|
2015-03-07 10:53:32 +00:00
|
|
|
node, err := NodeFromFileInfo(e.Fullpath(), e.Info())
|
2015-03-02 13:48:47 +00:00
|
|
|
if err != nil {
|
2015-03-21 13:43:33 +00:00
|
|
|
// TODO: integrate error reporting
|
|
|
|
debug.Log("Archiver.fileWorker", "NodeFromFileInfo returned error for %v: %v", node.path, err)
|
|
|
|
e.Result() <- nil
|
2015-04-26 01:54:35 +00:00
|
|
|
p.Report(Stat{Errors: 1})
|
2015-03-21 13:43:33 +00:00
|
|
|
continue
|
2015-03-02 13:48:47 +00:00
|
|
|
}
|
2014-11-23 08:22:18 +00:00
|
|
|
|
2015-03-07 10:53:32 +00:00
|
|
|
// try to use old node, if present
|
|
|
|
if e.Node != nil {
|
|
|
|
debug.Log("Archiver.fileWorker", " %v use old data", e.Path())
|
|
|
|
|
|
|
|
oldNode := e.Node.(*Node)
|
|
|
|
// check if all content is still available in the repository
|
|
|
|
contentMissing := false
|
|
|
|
for _, blob := range oldNode.blobs {
|
2016-08-31 18:29:54 +00:00
|
|
|
if ok, err := arch.repo.Backend().Test(DataFile, blob.Storage.String()); !ok || err != nil {
|
2015-03-07 10:53:32 +00:00
|
|
|
debug.Log("Archiver.fileWorker", " %v not using old data, %v (%v) is missing", e.Path(), blob.ID.Str(), blob.Storage.Str())
|
|
|
|
contentMissing = true
|
|
|
|
break
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if !contentMissing {
|
|
|
|
node.Content = oldNode.Content
|
|
|
|
node.blobs = oldNode.blobs
|
|
|
|
debug.Log("Archiver.fileWorker", " %v content is complete", e.Path())
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
debug.Log("Archiver.fileWorker", " %v no old data", e.Path())
|
|
|
|
}
|
|
|
|
|
|
|
|
// otherwise read file normally
|
2016-08-31 17:10:10 +00:00
|
|
|
if node.FileType == "file" && len(node.Content) == 0 {
|
2015-03-07 10:53:32 +00:00
|
|
|
debug.Log("Archiver.fileWorker", " read and save %v, content: %v", e.Path(), node.Content)
|
2015-04-26 15:44:38 +00:00
|
|
|
err = arch.SaveFile(p, node)
|
2015-03-02 13:48:47 +00:00
|
|
|
if err != nil {
|
2015-03-15 14:48:05 +00:00
|
|
|
// TODO: integrate error reporting
|
|
|
|
fmt.Fprintf(os.Stderr, "error for %v: %v\n", node.path, err)
|
|
|
|
// ignore this file
|
|
|
|
e.Result() <- nil
|
2015-04-26 01:54:35 +00:00
|
|
|
p.Report(Stat{Errors: 1})
|
2015-03-15 14:48:05 +00:00
|
|
|
continue
|
2015-03-02 13:48:47 +00:00
|
|
|
}
|
2015-03-07 10:53:32 +00:00
|
|
|
} else {
|
|
|
|
// report old data size
|
|
|
|
p.Report(Stat{Bytes: node.Size})
|
2015-03-02 13:48:47 +00:00
|
|
|
}
|
|
|
|
|
2015-03-07 10:53:32 +00:00
|
|
|
debug.Log("Archiver.fileWorker", " processed %v, %d/%d blobs", e.Path(), len(node.Content), len(node.blobs))
|
|
|
|
e.Result() <- node
|
2015-03-02 13:48:47 +00:00
|
|
|
p.Report(Stat{Files: 1})
|
|
|
|
case <-done:
|
|
|
|
// pipeline was cancelled
|
|
|
|
return
|
|
|
|
}
|
2014-12-21 16:20:49 +00:00
|
|
|
}
|
2015-03-02 13:48:47 +00:00
|
|
|
}
|
2014-12-21 16:20:49 +00:00
|
|
|
|
2015-03-02 13:48:47 +00:00
|
|
|
func (arch *Archiver) dirWorker(wg *sync.WaitGroup, p *Progress, done <-chan struct{}, dirCh <-chan pipe.Dir) {
|
2015-11-06 18:41:57 +00:00
|
|
|
debug.Log("Archiver.dirWorker", "start")
|
2015-03-07 10:53:32 +00:00
|
|
|
defer func() {
|
|
|
|
debug.Log("Archiver.dirWorker", "done")
|
|
|
|
wg.Done()
|
|
|
|
}()
|
2015-03-02 13:48:47 +00:00
|
|
|
for {
|
|
|
|
select {
|
|
|
|
case dir, ok := <-dirCh:
|
|
|
|
if !ok {
|
|
|
|
// channel is closed
|
|
|
|
return
|
|
|
|
}
|
2015-11-06 18:41:57 +00:00
|
|
|
debug.Log("Archiver.dirWorker", "save dir %v (%d entries), error %v\n", dir.Path(), len(dir.Entries), dir.Error())
|
2014-11-16 20:29:11 +00:00
|
|
|
|
2015-11-07 10:42:28 +00:00
|
|
|
// ignore dir nodes with errors
|
|
|
|
if dir.Error() != nil {
|
2016-02-07 21:18:00 +00:00
|
|
|
fmt.Fprintf(os.Stderr, "error walking dir %v: %v\n", dir.Path(), dir.Error())
|
2015-11-07 10:42:28 +00:00
|
|
|
dir.Result() <- nil
|
|
|
|
p.Report(Stat{Errors: 1})
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
2015-03-02 13:48:47 +00:00
|
|
|
tree := NewTree()
|
2015-02-15 13:44:54 +00:00
|
|
|
|
2015-03-02 13:48:47 +00:00
|
|
|
// wait for all content
|
|
|
|
for _, ch := range dir.Entries {
|
2015-11-06 18:41:57 +00:00
|
|
|
debug.Log("Archiver.dirWorker", "receiving result from %v", ch)
|
2015-03-15 14:48:05 +00:00
|
|
|
res := <-ch
|
|
|
|
|
|
|
|
// if we get a nil pointer here, an error has happened while
|
|
|
|
// processing this entry. Ignore it for now.
|
|
|
|
if res == nil {
|
2016-02-07 22:22:06 +00:00
|
|
|
debug.Log("Archiver.dirWorker", "got nil result?")
|
2015-03-15 14:48:05 +00:00
|
|
|
continue
|
|
|
|
}
|
|
|
|
|
|
|
|
// else insert node
|
|
|
|
node := res.(*Node)
|
2015-03-02 13:48:47 +00:00
|
|
|
tree.Insert(node)
|
2015-02-15 13:44:54 +00:00
|
|
|
|
2016-08-31 17:10:10 +00:00
|
|
|
if node.FileType == "dir" {
|
2016-02-07 22:22:06 +00:00
|
|
|
debug.Log("Archiver.dirWorker", "got tree node for %s: %v", node.path, node.Subtree)
|
2015-10-11 18:45:42 +00:00
|
|
|
|
|
|
|
if node.Subtree.IsNull() {
|
|
|
|
panic("invalid null subtree ID")
|
|
|
|
}
|
2015-02-15 13:44:54 +00:00
|
|
|
}
|
2015-03-02 13:48:47 +00:00
|
|
|
}
|
2015-02-15 13:44:54 +00:00
|
|
|
|
2015-11-06 18:41:57 +00:00
|
|
|
node := &Node{}
|
|
|
|
|
|
|
|
if dir.Path() != "" && dir.Info() != nil {
|
|
|
|
n, err := NodeFromFileInfo(dir.Path(), dir.Info())
|
2015-03-08 19:57:21 +00:00
|
|
|
if err != nil {
|
2015-11-06 18:41:57 +00:00
|
|
|
n.Error = err.Error()
|
|
|
|
dir.Result() <- n
|
2015-03-08 19:57:21 +00:00
|
|
|
continue
|
|
|
|
}
|
2015-11-06 18:41:57 +00:00
|
|
|
node = n
|
|
|
|
}
|
|
|
|
|
|
|
|
if err := dir.Error(); err != nil {
|
|
|
|
node.Error = err.Error()
|
2015-03-02 13:48:47 +00:00
|
|
|
}
|
|
|
|
|
2015-04-26 15:44:38 +00:00
|
|
|
id, err := arch.SaveTreeJSON(tree)
|
2015-03-02 13:48:47 +00:00
|
|
|
if err != nil {
|
|
|
|
panic(err)
|
2015-02-15 13:44:54 +00:00
|
|
|
}
|
2015-04-26 15:44:38 +00:00
|
|
|
debug.Log("Archiver.dirWorker", "save tree for %s: %v", dir.Path(), id.Str())
|
2015-10-11 18:45:42 +00:00
|
|
|
if id.IsNull() {
|
|
|
|
panic("invalid null subtree ID return from SaveTreeJSON()")
|
|
|
|
}
|
2015-03-02 13:48:47 +00:00
|
|
|
|
2015-07-25 15:05:45 +00:00
|
|
|
node.Subtree = &id
|
2015-03-02 13:48:47 +00:00
|
|
|
|
2015-11-06 18:41:57 +00:00
|
|
|
debug.Log("Archiver.dirWorker", "sending result to %v", dir.Result())
|
|
|
|
|
2015-03-07 10:53:32 +00:00
|
|
|
dir.Result() <- node
|
2015-04-25 19:40:42 +00:00
|
|
|
if dir.Path() != "" {
|
|
|
|
p.Report(Stat{Dirs: 1})
|
|
|
|
}
|
2015-03-02 13:48:47 +00:00
|
|
|
case <-done:
|
|
|
|
// pipeline was cancelled
|
|
|
|
return
|
2015-02-15 13:44:54 +00:00
|
|
|
}
|
|
|
|
}
|
2015-03-02 13:48:47 +00:00
|
|
|
}
|
2015-02-15 13:44:54 +00:00
|
|
|
|
2015-05-02 13:31:31 +00:00
|
|
|
type archivePipe struct {
|
2015-03-07 10:53:32 +00:00
|
|
|
Old <-chan WalkTreeJob
|
|
|
|
New <-chan pipe.Job
|
|
|
|
}
|
|
|
|
|
|
|
|
func copyJobs(done <-chan struct{}, in <-chan pipe.Job, out chan<- pipe.Job) {
|
|
|
|
var (
|
2015-05-02 14:36:48 +00:00
|
|
|
// disable sending on the outCh until we received a job
|
|
|
|
outCh chan<- pipe.Job
|
|
|
|
// enable receiving from in
|
|
|
|
inCh = in
|
|
|
|
job pipe.Job
|
|
|
|
ok bool
|
2015-03-07 10:53:32 +00:00
|
|
|
)
|
2015-05-02 14:36:48 +00:00
|
|
|
|
2015-03-07 10:53:32 +00:00
|
|
|
for {
|
|
|
|
select {
|
|
|
|
case <-done:
|
|
|
|
return
|
2015-05-02 14:36:48 +00:00
|
|
|
case job, ok = <-inCh:
|
2015-03-07 10:53:32 +00:00
|
|
|
if !ok {
|
2015-05-02 14:36:48 +00:00
|
|
|
// input channel closed, we're done
|
|
|
|
debug.Log("copyJobs", "input channel closed, we're done")
|
2015-03-07 10:53:32 +00:00
|
|
|
return
|
|
|
|
}
|
2015-05-02 14:36:48 +00:00
|
|
|
inCh = nil
|
|
|
|
outCh = out
|
|
|
|
case outCh <- job:
|
|
|
|
outCh = nil
|
|
|
|
inCh = in
|
2015-03-07 10:53:32 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
type archiveJob struct {
|
|
|
|
hasOld bool
|
|
|
|
old WalkTreeJob
|
|
|
|
new pipe.Job
|
|
|
|
}
|
|
|
|
|
2015-05-02 13:31:31 +00:00
|
|
|
func (a *archivePipe) compare(done <-chan struct{}, out chan<- pipe.Job) {
|
2015-03-07 10:53:32 +00:00
|
|
|
defer func() {
|
|
|
|
close(out)
|
|
|
|
debug.Log("ArchivePipe.compare", "done")
|
|
|
|
}()
|
|
|
|
|
|
|
|
debug.Log("ArchivePipe.compare", "start")
|
|
|
|
var (
|
|
|
|
loadOld, loadNew bool = true, true
|
|
|
|
ok bool
|
|
|
|
oldJob WalkTreeJob
|
|
|
|
newJob pipe.Job
|
|
|
|
)
|
|
|
|
|
|
|
|
for {
|
|
|
|
if loadOld {
|
|
|
|
oldJob, ok = <-a.Old
|
|
|
|
// if the old channel is closed, just pass through the new jobs
|
|
|
|
if !ok {
|
|
|
|
debug.Log("ArchivePipe.compare", "old channel is closed, copy from new channel")
|
|
|
|
|
|
|
|
// handle remaining newJob
|
|
|
|
if !loadNew {
|
|
|
|
out <- archiveJob{new: newJob}.Copy()
|
|
|
|
}
|
|
|
|
|
|
|
|
copyJobs(done, a.New, out)
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
loadOld = false
|
|
|
|
}
|
|
|
|
|
|
|
|
if loadNew {
|
|
|
|
newJob, ok = <-a.New
|
|
|
|
// if the new channel is closed, there are no more files in the current snapshot, return
|
|
|
|
if !ok {
|
|
|
|
debug.Log("ArchivePipe.compare", "new channel is closed, we're done")
|
|
|
|
return
|
|
|
|
}
|
|
|
|
|
|
|
|
loadNew = false
|
|
|
|
}
|
|
|
|
|
|
|
|
debug.Log("ArchivePipe.compare", "old job: %v", oldJob.Path)
|
|
|
|
debug.Log("ArchivePipe.compare", "new job: %v", newJob.Path())
|
|
|
|
|
|
|
|
// at this point we have received an old job as well as a new job, compare paths
|
|
|
|
file1 := oldJob.Path
|
|
|
|
file2 := newJob.Path()
|
|
|
|
|
|
|
|
dir1 := filepath.Dir(file1)
|
|
|
|
dir2 := filepath.Dir(file2)
|
|
|
|
|
|
|
|
if file1 == file2 {
|
|
|
|
debug.Log("ArchivePipe.compare", " same filename %q", file1)
|
|
|
|
|
|
|
|
// send job
|
|
|
|
out <- archiveJob{hasOld: true, old: oldJob, new: newJob}.Copy()
|
|
|
|
loadOld = true
|
|
|
|
loadNew = true
|
|
|
|
continue
|
|
|
|
} else if dir1 < dir2 {
|
|
|
|
debug.Log("ArchivePipe.compare", " %q < %q, file %q added", dir1, dir2, file2)
|
|
|
|
// file is new, send new job and load new
|
|
|
|
loadNew = true
|
|
|
|
out <- archiveJob{new: newJob}.Copy()
|
|
|
|
continue
|
2015-03-09 21:56:23 +00:00
|
|
|
} else if dir1 == dir2 {
|
|
|
|
if file1 < file2 {
|
|
|
|
debug.Log("ArchivePipe.compare", " %q < %q, file %q removed", file1, file2, file1)
|
|
|
|
// file has been removed, load new old
|
|
|
|
loadOld = true
|
|
|
|
continue
|
|
|
|
} else {
|
|
|
|
debug.Log("ArchivePipe.compare", " %q > %q, file %q added", file1, file2, file2)
|
|
|
|
// file is new, send new job and load new
|
|
|
|
loadNew = true
|
|
|
|
out <- archiveJob{new: newJob}.Copy()
|
|
|
|
continue
|
|
|
|
}
|
2015-03-07 10:53:32 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
debug.Log("ArchivePipe.compare", " %q > %q, file %q removed", file1, file2, file1)
|
|
|
|
// file has been removed, throw away old job and load new
|
|
|
|
loadOld = true
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func (j archiveJob) Copy() pipe.Job {
|
|
|
|
if !j.hasOld {
|
|
|
|
return j.new
|
|
|
|
}
|
|
|
|
|
|
|
|
// handle files
|
2015-07-08 01:42:13 +00:00
|
|
|
if isRegularFile(j.new.Info()) {
|
2015-03-07 10:53:32 +00:00
|
|
|
debug.Log("archiveJob.Copy", " job %v is file", j.new.Path())
|
|
|
|
|
|
|
|
// if type has changed, return new job directly
|
|
|
|
if j.old.Node == nil {
|
|
|
|
return j.new
|
|
|
|
}
|
|
|
|
|
|
|
|
// if file is newer, return the new job
|
|
|
|
if j.old.Node.isNewer(j.new.Fullpath(), j.new.Info()) {
|
|
|
|
debug.Log("archiveJob.Copy", " job %v is newer", j.new.Path())
|
|
|
|
return j.new
|
|
|
|
}
|
|
|
|
|
|
|
|
debug.Log("archiveJob.Copy", " job %v add old data", j.new.Path())
|
|
|
|
// otherwise annotate job with old data
|
|
|
|
e := j.new.(pipe.Entry)
|
|
|
|
e.Node = j.old.Node
|
|
|
|
return e
|
|
|
|
}
|
|
|
|
|
|
|
|
// dirs and other types are just returned
|
|
|
|
return j.new
|
|
|
|
}
|
|
|
|
|
2015-10-12 21:59:17 +00:00
|
|
|
const saveIndexTime = 30 * time.Second
|
|
|
|
|
|
|
|
// saveIndexes regularly queries the master index for full indexes and saves them.
|
|
|
|
func (arch *Archiver) saveIndexes(wg *sync.WaitGroup, done <-chan struct{}) {
|
|
|
|
defer wg.Done()
|
|
|
|
|
|
|
|
ticker := time.NewTicker(saveIndexTime)
|
|
|
|
defer ticker.Stop()
|
|
|
|
|
|
|
|
for {
|
|
|
|
select {
|
|
|
|
case <-done:
|
|
|
|
return
|
|
|
|
case <-ticker.C:
|
|
|
|
debug.Log("Archiver.saveIndexes", "saving full indexes")
|
|
|
|
err := arch.repo.SaveFullIndex()
|
|
|
|
if err != nil {
|
|
|
|
debug.Log("Archiver.saveIndexes", "save indexes returned an error: %v", err)
|
|
|
|
fmt.Fprintf(os.Stderr, "error saving preliminary index: %v\n", err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2016-02-07 18:35:35 +00:00
|
|
|
// unique returns a slice that only contains unique strings.
|
|
|
|
func unique(items []string) []string {
|
|
|
|
seen := make(map[string]struct{})
|
|
|
|
for _, item := range items {
|
|
|
|
seen[item] = struct{}{}
|
|
|
|
}
|
|
|
|
|
|
|
|
items = items[:0]
|
|
|
|
for item := range seen {
|
|
|
|
items = append(items, item)
|
|
|
|
}
|
|
|
|
return items
|
|
|
|
}
|
|
|
|
|
2016-05-09 12:46:14 +00:00
|
|
|
// baseNameSlice allows sorting paths by basename.
|
2016-05-09 12:44:03 +00:00
|
|
|
//
|
2016-05-09 12:32:17 +00:00
|
|
|
// Snapshots have contents sorted by basename, but we receive full paths.
|
|
|
|
// For the archivePipe to advance them in pairs, we traverse the given
|
|
|
|
// paths in the same order as the snapshot.
|
2016-05-09 12:46:14 +00:00
|
|
|
type baseNameSlice []string
|
2016-05-09 12:32:17 +00:00
|
|
|
|
2016-05-09 12:46:14 +00:00
|
|
|
func (p baseNameSlice) Len() int { return len(p) }
|
|
|
|
func (p baseNameSlice) Less(i, j int) bool { return filepath.Base(p[i]) < filepath.Base(p[j]) }
|
|
|
|
func (p baseNameSlice) Swap(i, j int) { p[i], p[j] = p[j], p[i] }
|
2016-05-09 12:32:17 +00:00
|
|
|
|
2015-05-02 13:51:40 +00:00
|
|
|
// Snapshot creates a snapshot of the given paths. If parentID is set, this is
|
2015-05-02 15:01:31 +00:00
|
|
|
// used to compare the files to the ones archived at the time this snapshot was
|
|
|
|
// taken.
|
2016-08-31 18:29:54 +00:00
|
|
|
func (arch *Archiver) Snapshot(p *Progress, paths []string, parentID *ID) (*Snapshot, ID, error) {
|
2016-02-07 18:35:35 +00:00
|
|
|
paths = unique(paths)
|
2016-05-09 12:46:14 +00:00
|
|
|
sort.Sort(baseNameSlice(paths))
|
2016-02-07 18:35:35 +00:00
|
|
|
|
2015-03-02 13:48:47 +00:00
|
|
|
debug.Log("Archiver.Snapshot", "start for %v", paths)
|
2015-02-15 13:44:54 +00:00
|
|
|
|
2015-06-21 15:12:38 +00:00
|
|
|
debug.RunHook("Archiver.Snapshot", nil)
|
2015-02-15 13:44:54 +00:00
|
|
|
|
2015-03-07 10:53:32 +00:00
|
|
|
// signal the whole pipeline to stop
|
|
|
|
done := make(chan struct{})
|
|
|
|
var err error
|
|
|
|
|
2015-03-02 13:48:47 +00:00
|
|
|
p.Start()
|
|
|
|
defer p.Done()
|
2015-02-15 13:44:54 +00:00
|
|
|
|
2015-03-07 10:53:32 +00:00
|
|
|
// create new snapshot
|
2015-03-02 13:48:47 +00:00
|
|
|
sn, err := NewSnapshot(paths)
|
|
|
|
if err != nil {
|
2016-08-31 18:29:54 +00:00
|
|
|
return nil, ID{}, err
|
2015-02-15 13:44:54 +00:00
|
|
|
}
|
2015-07-22 20:33:23 +00:00
|
|
|
sn.Excludes = arch.Excludes
|
2015-02-15 13:44:54 +00:00
|
|
|
|
2015-05-02 13:31:31 +00:00
|
|
|
jobs := archivePipe{}
|
2015-03-02 13:48:47 +00:00
|
|
|
|
2015-03-07 10:53:32 +00:00
|
|
|
// use parent snapshot (if some was given)
|
2015-05-02 13:51:40 +00:00
|
|
|
if parentID != nil {
|
|
|
|
sn.Parent = parentID
|
2015-03-07 10:53:32 +00:00
|
|
|
|
|
|
|
// load parent snapshot
|
2015-07-25 15:05:45 +00:00
|
|
|
parent, err := LoadSnapshot(arch.repo, *parentID)
|
2015-03-07 10:53:32 +00:00
|
|
|
if err != nil {
|
2016-08-31 18:29:54 +00:00
|
|
|
return nil, ID{}, err
|
2015-03-07 10:53:32 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// start walker on old tree
|
|
|
|
ch := make(chan WalkTreeJob)
|
2015-07-25 15:05:45 +00:00
|
|
|
go WalkTree(arch.repo, *parent.Tree, done, ch)
|
2015-03-07 10:53:32 +00:00
|
|
|
jobs.Old = ch
|
|
|
|
} else {
|
|
|
|
// use closed channel
|
|
|
|
ch := make(chan WalkTreeJob)
|
|
|
|
close(ch)
|
|
|
|
jobs.Old = ch
|
|
|
|
}
|
|
|
|
|
|
|
|
// start walker
|
|
|
|
pipeCh := make(chan pipe.Job)
|
|
|
|
resCh := make(chan pipe.Result, 1)
|
|
|
|
go func() {
|
2015-11-06 18:41:57 +00:00
|
|
|
pipe.Walk(paths, arch.SelectFilter, done, pipeCh, resCh)
|
2015-03-07 10:53:32 +00:00
|
|
|
debug.Log("Archiver.Snapshot", "pipe.Walk done")
|
|
|
|
}()
|
|
|
|
jobs.New = pipeCh
|
2015-03-02 13:48:47 +00:00
|
|
|
|
2015-03-07 10:53:32 +00:00
|
|
|
ch := make(chan pipe.Job)
|
|
|
|
go jobs.compare(done, ch)
|
2015-03-02 13:48:47 +00:00
|
|
|
|
2015-02-15 13:44:54 +00:00
|
|
|
var wg sync.WaitGroup
|
2015-03-02 13:48:47 +00:00
|
|
|
entCh := make(chan pipe.Entry)
|
|
|
|
dirCh := make(chan pipe.Dir)
|
|
|
|
|
|
|
|
// split
|
|
|
|
wg.Add(1)
|
|
|
|
go func() {
|
2015-03-07 10:53:32 +00:00
|
|
|
pipe.Split(ch, dirCh, entCh)
|
|
|
|
debug.Log("Archiver.Snapshot", "split done")
|
2015-03-02 13:48:47 +00:00
|
|
|
close(dirCh)
|
|
|
|
close(entCh)
|
|
|
|
wg.Done()
|
|
|
|
}()
|
|
|
|
|
|
|
|
// run workers
|
2015-02-15 13:44:54 +00:00
|
|
|
for i := 0; i < maxConcurrency; i++ {
|
|
|
|
wg.Add(2)
|
2015-03-02 13:48:47 +00:00
|
|
|
go arch.fileWorker(&wg, p, done, entCh)
|
|
|
|
go arch.dirWorker(&wg, p, done, dirCh)
|
2015-02-15 13:44:54 +00:00
|
|
|
}
|
|
|
|
|
2015-10-12 21:59:17 +00:00
|
|
|
// run index saver
|
|
|
|
var wgIndexSaver sync.WaitGroup
|
|
|
|
stopIndexSaver := make(chan struct{})
|
|
|
|
wgIndexSaver.Add(1)
|
|
|
|
go arch.saveIndexes(&wgIndexSaver, stopIndexSaver)
|
|
|
|
|
2015-02-15 13:44:54 +00:00
|
|
|
// wait for all workers to terminate
|
2015-03-02 13:48:47 +00:00
|
|
|
debug.Log("Archiver.Snapshot", "wait for workers")
|
2015-02-15 13:44:54 +00:00
|
|
|
wg.Wait()
|
|
|
|
|
2015-10-12 21:59:17 +00:00
|
|
|
// stop index saver
|
|
|
|
close(stopIndexSaver)
|
|
|
|
wgIndexSaver.Wait()
|
|
|
|
|
2015-03-02 13:48:47 +00:00
|
|
|
debug.Log("Archiver.Snapshot", "workers terminated")
|
2015-02-15 13:44:54 +00:00
|
|
|
|
2015-03-08 19:57:21 +00:00
|
|
|
// receive the top-level tree
|
|
|
|
root := (<-resCh).(*Node)
|
2015-04-26 15:44:38 +00:00
|
|
|
debug.Log("Archiver.Snapshot", "root node received: %v", root.Subtree.Str())
|
|
|
|
sn.Tree = root.Subtree
|
2014-11-23 21:26:01 +00:00
|
|
|
|
2014-09-23 20:39:12 +00:00
|
|
|
// save snapshot
|
2016-08-31 18:29:54 +00:00
|
|
|
id, err := arch.repo.SaveJSONUnpacked(SnapshotFile, sn)
|
2014-09-23 20:39:12 +00:00
|
|
|
if err != nil {
|
2016-08-31 18:29:54 +00:00
|
|
|
return nil, ID{}, err
|
2014-09-23 20:39:12 +00:00
|
|
|
}
|
|
|
|
|
2015-03-09 21:56:44 +00:00
|
|
|
// store ID in snapshot struct
|
2015-07-25 15:05:45 +00:00
|
|
|
sn.id = &id
|
2015-04-26 15:44:38 +00:00
|
|
|
debug.Log("Archiver.Snapshot", "saved snapshot %v", id.Str())
|
2015-03-09 21:56:44 +00:00
|
|
|
|
2015-05-09 11:35:55 +00:00
|
|
|
// flush repository
|
2015-05-09 11:32:52 +00:00
|
|
|
err = arch.repo.Flush()
|
2015-04-26 15:44:38 +00:00
|
|
|
if err != nil {
|
2016-08-31 18:29:54 +00:00
|
|
|
return nil, ID{}, err
|
2015-04-26 15:44:38 +00:00
|
|
|
}
|
2015-03-09 21:26:39 +00:00
|
|
|
|
2015-04-26 15:44:38 +00:00
|
|
|
// save index
|
2015-10-12 20:34:12 +00:00
|
|
|
err = arch.repo.SaveIndex()
|
2015-03-09 21:26:39 +00:00
|
|
|
if err != nil {
|
2015-04-26 15:44:38 +00:00
|
|
|
debug.Log("Archiver.Snapshot", "error saving index: %v", err)
|
2016-08-31 18:29:54 +00:00
|
|
|
return nil, ID{}, err
|
2015-03-09 21:26:39 +00:00
|
|
|
}
|
|
|
|
|
2015-10-12 20:34:12 +00:00
|
|
|
debug.Log("Archiver.Snapshot", "saved indexes")
|
2015-04-26 15:44:38 +00:00
|
|
|
|
|
|
|
return sn, id, nil
|
2014-09-23 20:39:12 +00:00
|
|
|
}
|
2015-02-21 13:23:49 +00:00
|
|
|
|
2015-07-08 01:42:13 +00:00
|
|
|
func isRegularFile(fi os.FileInfo) bool {
|
2015-03-07 10:53:32 +00:00
|
|
|
if fi == nil {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
2015-02-21 13:23:49 +00:00
|
|
|
return fi.Mode()&(os.ModeType|os.ModeCharDevice) == 0
|
|
|
|
}
|
|
|
|
|
2015-05-02 13:51:40 +00:00
|
|
|
// Scan traverses the dirs to collect Stat information while emitting progress
|
|
|
|
// information with p.
|
2015-07-19 22:13:39 +00:00
|
|
|
func Scan(dirs []string, filter pipe.SelectFunc, p *Progress) (Stat, error) {
|
2015-02-21 13:23:49 +00:00
|
|
|
p.Start()
|
|
|
|
defer p.Done()
|
|
|
|
|
|
|
|
var stat Stat
|
|
|
|
|
2015-03-02 13:48:47 +00:00
|
|
|
for _, dir := range dirs {
|
2015-03-15 11:20:30 +00:00
|
|
|
debug.Log("Scan", "Start for %v", dir)
|
Fix 567 (#570)
* Patch for https://github.com/restic/restic/issues/567
Backup also files on windows with longer pathnames than 255 chars (e.g. from node).
as fd0 says "So, as far as I can see, we need to have custom methods for all functions that accept a path, so that on Windows we can substitute the normal (possibly relative) path used within restic by an (absolute) UNC path, and only then call the underlying functions like os.Stat(), os.Lstat(), os.Open() and so on.
I've already thought about adding a generic abstraction for the file system (so we can mock this easier in tests), and this looks like a good opportunity to build it."
* fixed building tests
* Restructured patches
Add Wrapper for filepath.Walk
* using \\?\ requires absolute pathes to be used.
Now all tests run
* used gofmt on the code
* Restructured Code. No patches dir, integrate the file functions into restic/fs/
There is still an issue, because restic.fs.Open has a different api the os.Open, which returns the result of OpenFile, but takes only a string
* Changed the last os.Open() calls to fs.Open() after extending the File interface
* fixed name-clash of restic.fs and fuse.fs detected by travis
* fixed fmt with gofmt
* c&p failure: removed fixpath() call.
* missing include
* fixed includes in linux variant
* Fix for Linux. Fd() is required on File interface
* done gofmt
2016-08-15 19:59:13 +00:00
|
|
|
err := fs.Walk(dir, func(str string, fi os.FileInfo, err error) error {
|
2015-03-21 13:43:33 +00:00
|
|
|
// TODO: integrate error reporting
|
|
|
|
if err != nil {
|
|
|
|
fmt.Fprintf(os.Stderr, "error for %v: %v\n", str, err)
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
if fi == nil {
|
|
|
|
fmt.Fprintf(os.Stderr, "error for %v: FileInfo is nil\n", str)
|
|
|
|
return nil
|
|
|
|
}
|
2015-07-19 22:13:39 +00:00
|
|
|
|
|
|
|
if !filter(str, fi) {
|
|
|
|
debug.Log("Scan.Walk", "path %v excluded", str)
|
|
|
|
if fi.IsDir() {
|
|
|
|
return filepath.SkipDir
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2015-03-02 13:48:47 +00:00
|
|
|
s := Stat{}
|
2015-07-08 01:42:13 +00:00
|
|
|
if fi.IsDir() {
|
2015-03-02 13:48:47 +00:00
|
|
|
s.Dirs++
|
2015-07-08 01:42:13 +00:00
|
|
|
} else {
|
|
|
|
s.Files++
|
|
|
|
|
|
|
|
if isRegularFile(fi) {
|
|
|
|
s.Bytes += uint64(fi.Size())
|
|
|
|
}
|
2015-03-02 13:48:47 +00:00
|
|
|
}
|
2015-02-21 13:23:49 +00:00
|
|
|
|
2015-03-02 13:48:47 +00:00
|
|
|
p.Report(s)
|
|
|
|
stat.Add(s)
|
2015-02-21 13:23:49 +00:00
|
|
|
|
2015-03-02 13:48:47 +00:00
|
|
|
// TODO: handle error?
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
|
2015-03-15 11:20:30 +00:00
|
|
|
debug.Log("Scan", "Done for %v, err: %v", dir, err)
|
2015-03-02 13:48:47 +00:00
|
|
|
if err != nil {
|
2016-08-29 19:38:34 +00:00
|
|
|
return Stat{}, errors.Wrap(err, "fs.Walk")
|
2015-03-02 13:48:47 +00:00
|
|
|
}
|
|
|
|
}
|
2015-02-21 13:23:49 +00:00
|
|
|
|
2015-03-02 13:48:47 +00:00
|
|
|
return stat, nil
|
2015-02-21 13:23:49 +00:00
|
|
|
}
|