mirror of
https://github.com/octoleo/restic.git
synced 2024-11-26 23:06:32 +00:00
434c79354e
Count errors separately and not as files that are "done"
181 lines
3.2 KiB
Go
181 lines
3.2 KiB
Go
package restic
|
|
|
|
import (
|
|
"fmt"
|
|
"sync"
|
|
"time"
|
|
)
|
|
|
|
type Progress struct {
|
|
OnStart func()
|
|
OnUpdate ProgressFunc
|
|
OnDone ProgressFunc
|
|
fnM sync.Mutex
|
|
|
|
cur Stat
|
|
curM sync.Mutex
|
|
start time.Time
|
|
c *time.Ticker
|
|
cancel chan struct{}
|
|
o sync.Once
|
|
d time.Duration
|
|
|
|
running bool
|
|
}
|
|
|
|
type Stat struct {
|
|
Files uint64
|
|
Dirs uint64
|
|
Bytes uint64
|
|
Trees uint64
|
|
Blobs uint64
|
|
Errors uint64
|
|
}
|
|
|
|
type ProgressFunc func(s Stat, runtime time.Duration, ticker bool)
|
|
|
|
// NewProgress returns a new progress reporter. When Start() called, the
|
|
// function OnStart is executed once. Afterwards the function OnUpdate is
|
|
// called when new data arrives or at least every d interval. The function
|
|
// OnDone is called when Done() is called. Both functions are called
|
|
// synchronously and can use shared state.
|
|
func NewProgress(d time.Duration) *Progress {
|
|
return &Progress{d: d}
|
|
}
|
|
|
|
// Start runs resets and runs the progress reporter.
|
|
func (p *Progress) Start() {
|
|
if p == nil || p.running {
|
|
return
|
|
}
|
|
|
|
p.o = sync.Once{}
|
|
p.cancel = make(chan struct{})
|
|
p.running = true
|
|
p.Reset()
|
|
p.start = time.Now()
|
|
p.c = time.NewTicker(p.d)
|
|
|
|
if p.OnStart != nil {
|
|
p.OnStart()
|
|
}
|
|
|
|
go p.reporter()
|
|
}
|
|
|
|
// Report adds the statistics from s to the current state and tries to report
|
|
// the accumulated statistics via the feedback channel.
|
|
func (p *Progress) Report(s Stat) {
|
|
if p == nil {
|
|
return
|
|
}
|
|
|
|
if !p.running {
|
|
panic("reporting in a non-running Progress")
|
|
}
|
|
|
|
p.curM.Lock()
|
|
p.cur.Add(s)
|
|
cur := p.cur
|
|
p.curM.Unlock()
|
|
|
|
// update progress
|
|
if p.OnUpdate != nil {
|
|
p.fnM.Lock()
|
|
p.OnUpdate(cur, time.Since(p.start), false)
|
|
p.fnM.Unlock()
|
|
}
|
|
}
|
|
|
|
func (p *Progress) reporter() {
|
|
if p == nil {
|
|
return
|
|
}
|
|
|
|
for {
|
|
select {
|
|
case <-p.c.C:
|
|
p.curM.Lock()
|
|
cur := p.cur
|
|
p.curM.Unlock()
|
|
|
|
if p.OnUpdate != nil {
|
|
p.fnM.Lock()
|
|
p.OnUpdate(cur, time.Since(p.start), true)
|
|
p.fnM.Unlock()
|
|
}
|
|
case <-p.cancel:
|
|
p.c.Stop()
|
|
return
|
|
}
|
|
}
|
|
}
|
|
|
|
// Reset resets all statistic counters to zero.
|
|
func (p *Progress) Reset() {
|
|
if p == nil {
|
|
return
|
|
}
|
|
|
|
if !p.running {
|
|
panic("resetting a non-running Progress")
|
|
}
|
|
|
|
p.curM.Lock()
|
|
p.cur = Stat{}
|
|
p.curM.Unlock()
|
|
}
|
|
|
|
// Done closes the progress report.
|
|
func (p *Progress) Done() {
|
|
if p == nil || !p.running {
|
|
return
|
|
}
|
|
|
|
if p.running {
|
|
p.running = false
|
|
p.o.Do(func() {
|
|
close(p.cancel)
|
|
})
|
|
|
|
cur := p.cur
|
|
|
|
if p.OnDone != nil {
|
|
p.fnM.Lock()
|
|
p.OnDone(cur, time.Since(p.start), false)
|
|
p.fnM.Unlock()
|
|
}
|
|
}
|
|
}
|
|
|
|
// Add accumulates other into s.
|
|
func (s *Stat) Add(other Stat) {
|
|
s.Bytes += other.Bytes
|
|
s.Dirs += other.Dirs
|
|
s.Files += other.Files
|
|
s.Trees += other.Trees
|
|
s.Blobs += other.Blobs
|
|
s.Errors += other.Errors
|
|
}
|
|
|
|
func (s Stat) String() string {
|
|
b := float64(s.Bytes)
|
|
var str string
|
|
|
|
switch {
|
|
case s.Bytes > 1<<40:
|
|
str = fmt.Sprintf("%.3f TiB", b/(1<<40))
|
|
case s.Bytes > 1<<30:
|
|
str = fmt.Sprintf("%.3f GiB", b/(1<<30))
|
|
case s.Bytes > 1<<20:
|
|
str = fmt.Sprintf("%.3f MiB", b/(1<<20))
|
|
case s.Bytes > 1<<10:
|
|
str = fmt.Sprintf("%.3f KiB", b/(1<<10))
|
|
default:
|
|
str = fmt.Sprintf("%dB", s.Bytes)
|
|
}
|
|
|
|
return fmt.Sprintf("Stat(%d files, %d dirs, %v trees, %v blobs, %d errors, %v)",
|
|
s.Files, s.Dirs, s.Trees, s.Blobs, s.Errors, str)
|
|
}
|