mirror of
https://github.com/octoleo/syncthing.git
synced 2024-11-19 11:35:10 +00:00
442 lines
14 KiB
Go
442 lines
14 KiB
Go
// Copyright (C) 2015 The Syncthing Authors.
|
|
//
|
|
// This Source Code Form is subject to the terms of the Mozilla Public
|
|
// License, v. 2.0. If a copy of the MPL was not distributed with this file,
|
|
// You can obtain one at https://mozilla.org/MPL/2.0/.
|
|
|
|
//go:generate -command counterfeiter go run github.com/maxbrunsfeld/counterfeiter/v6
|
|
//go:generate counterfeiter -o mocks/folderSummaryService.go --fake-name FolderSummaryService . FolderSummaryService
|
|
|
|
package model
|
|
|
|
import (
|
|
"context"
|
|
"fmt"
|
|
"strings"
|
|
"time"
|
|
|
|
"github.com/thejerf/suture/v4"
|
|
|
|
"github.com/syncthing/syncthing/lib/config"
|
|
"github.com/syncthing/syncthing/lib/db"
|
|
"github.com/syncthing/syncthing/lib/events"
|
|
"github.com/syncthing/syncthing/lib/protocol"
|
|
"github.com/syncthing/syncthing/lib/svcutil"
|
|
"github.com/syncthing/syncthing/lib/sync"
|
|
)
|
|
|
|
const maxDurationSinceLastEventReq = time.Minute
|
|
|
|
type FolderSummaryService interface {
|
|
suture.Service
|
|
Summary(folder string) (*FolderSummary, error)
|
|
OnEventRequest()
|
|
}
|
|
|
|
// The folderSummaryService adds summary information events (FolderSummary and
|
|
// FolderCompletion) into the event stream at certain intervals.
|
|
type folderSummaryService struct {
|
|
*suture.Supervisor
|
|
|
|
cfg config.Wrapper
|
|
model Model
|
|
id protocol.DeviceID
|
|
evLogger events.Logger
|
|
immediate chan string
|
|
|
|
// For keeping track of folders to recalculate for
|
|
foldersMut sync.Mutex
|
|
folders map[string]struct{}
|
|
|
|
// For keeping track of when the last event request on the API was
|
|
lastEventReq time.Time
|
|
lastEventReqMut sync.Mutex
|
|
}
|
|
|
|
func NewFolderSummaryService(cfg config.Wrapper, m Model, id protocol.DeviceID, evLogger events.Logger) FolderSummaryService {
|
|
service := &folderSummaryService{
|
|
Supervisor: suture.New("folderSummaryService", svcutil.SpecWithDebugLogger(l)),
|
|
cfg: cfg,
|
|
model: m,
|
|
id: id,
|
|
evLogger: evLogger,
|
|
immediate: make(chan string),
|
|
folders: make(map[string]struct{}),
|
|
foldersMut: sync.NewMutex(),
|
|
lastEventReqMut: sync.NewMutex(),
|
|
}
|
|
|
|
service.Add(svcutil.AsService(service.listenForUpdates, fmt.Sprintf("%s/listenForUpdates", service)))
|
|
service.Add(svcutil.AsService(service.calculateSummaries, fmt.Sprintf("%s/calculateSummaries", service)))
|
|
|
|
return service
|
|
}
|
|
|
|
func (c *folderSummaryService) String() string {
|
|
return fmt.Sprintf("FolderSummaryService@%p", c)
|
|
}
|
|
|
|
// FolderSummary replaces the previously used map[string]interface{}, and needs
|
|
// to keep the structure/naming for api backwards compatibility
|
|
type FolderSummary struct {
|
|
Errors int `json:"errors"`
|
|
PullErrors int `json:"pullErrors"` // deprecated
|
|
|
|
Invalid string `json:"invalid"` // deprecated
|
|
|
|
GlobalFiles int `json:"globalFiles"`
|
|
GlobalDirectories int `json:"globalDirectories"`
|
|
GlobalSymlinks int `json:"globalSymlinks"`
|
|
GlobalDeleted int `json:"globalDeleted"`
|
|
GlobalBytes int64 `json:"globalBytes"`
|
|
GlobalTotalItems int `json:"globalTotalItems"`
|
|
|
|
LocalFiles int `json:"localFiles"`
|
|
LocalDirectories int `json:"localDirectories"`
|
|
LocalSymlinks int `json:"localSymlinks"`
|
|
LocalDeleted int `json:"localDeleted"`
|
|
LocalBytes int64 `json:"localBytes"`
|
|
LocalTotalItems int `json:"localTotalItems"`
|
|
|
|
NeedFiles int `json:"needFiles"`
|
|
NeedDirectories int `json:"needDirectories"`
|
|
NeedSymlinks int `json:"needSymlinks"`
|
|
NeedDeletes int `json:"needDeletes"`
|
|
NeedBytes int64 `json:"needBytes"`
|
|
NeedTotalItems int `json:"needTotalItems"`
|
|
|
|
ReceiveOnlyChangedFiles int `json:"receiveOnlyChangedFiles"`
|
|
ReceiveOnlyChangedDirectories int `json:"receiveOnlyChangedDirectories"`
|
|
ReceiveOnlyChangedSymlinks int `json:"receiveOnlyChangedSymlinks"`
|
|
ReceiveOnlyChangedDeletes int `json:"receiveOnlyChangedDeletes"`
|
|
ReceiveOnlyChangedBytes int64 `json:"receiveOnlyChangedBytes"`
|
|
ReceiveOnlyTotalItems int `json:"receiveOnlyTotalItems"`
|
|
|
|
InSyncFiles int `json:"inSyncFiles"`
|
|
InSyncBytes int64 `json:"inSyncBytes"`
|
|
|
|
State string `json:"state"`
|
|
StateChanged time.Time `json:"stateChanged"`
|
|
Error string `json:"error"`
|
|
|
|
Version int64 `json:"version"` // deprecated
|
|
Sequence int64 `json:"sequence"`
|
|
|
|
IgnorePatterns bool `json:"ignorePatterns"`
|
|
WatchError string `json:"watchError"`
|
|
}
|
|
|
|
func (c *folderSummaryService) Summary(folder string) (*FolderSummary, error) {
|
|
res := new(FolderSummary)
|
|
|
|
var local, global, need, ro db.Counts
|
|
var ourSeq, remoteSeq int64
|
|
errors, err := c.model.FolderErrors(folder)
|
|
if err == nil {
|
|
var snap *db.Snapshot
|
|
if snap, err = c.model.DBSnapshot(folder); err == nil {
|
|
global = snap.GlobalSize()
|
|
local = snap.LocalSize()
|
|
need = snap.NeedSize(protocol.LocalDeviceID)
|
|
ro = snap.ReceiveOnlyChangedSize()
|
|
ourSeq = snap.Sequence(protocol.LocalDeviceID)
|
|
remoteSeq = snap.Sequence(protocol.GlobalDeviceID)
|
|
snap.Release()
|
|
}
|
|
}
|
|
// For API backwards compatibility (SyncTrayzor needs it) an empty folder
|
|
// summary is returned for not running folders, an error might actually be
|
|
// more appropriate
|
|
if err != nil && err != ErrFolderPaused && err != ErrFolderNotRunning {
|
|
return nil, err
|
|
}
|
|
|
|
res.Errors = len(errors)
|
|
res.PullErrors = len(errors) // deprecated
|
|
|
|
res.Invalid = "" // Deprecated, retains external API for now
|
|
|
|
res.GlobalFiles, res.GlobalDirectories, res.GlobalSymlinks, res.GlobalDeleted, res.GlobalBytes, res.GlobalTotalItems = global.Files, global.Directories, global.Symlinks, global.Deleted, global.Bytes, global.TotalItems()
|
|
|
|
res.LocalFiles, res.LocalDirectories, res.LocalSymlinks, res.LocalDeleted, res.LocalBytes, res.LocalTotalItems = local.Files, local.Directories, local.Symlinks, local.Deleted, local.Bytes, local.TotalItems()
|
|
|
|
fcfg, haveFcfg := c.cfg.Folder(folder)
|
|
|
|
if haveFcfg && fcfg.IgnoreDelete {
|
|
need.Deleted = 0
|
|
}
|
|
|
|
need.Bytes -= c.model.FolderProgressBytesCompleted(folder)
|
|
// This may happen if we are in progress of pulling files that were
|
|
// deleted globally after the pull started.
|
|
if need.Bytes < 0 {
|
|
need.Bytes = 0
|
|
}
|
|
res.NeedFiles, res.NeedDirectories, res.NeedSymlinks, res.NeedDeletes, res.NeedBytes, res.NeedTotalItems = need.Files, need.Directories, need.Symlinks, need.Deleted, need.Bytes, need.TotalItems()
|
|
|
|
if haveFcfg && (fcfg.Type == config.FolderTypeReceiveOnly || fcfg.Type == config.FolderTypeReceiveEncrypted) {
|
|
// Add statistics for things that have changed locally in a receive
|
|
// only or receive encrypted folder.
|
|
res.ReceiveOnlyChangedFiles = ro.Files
|
|
res.ReceiveOnlyChangedDirectories = ro.Directories
|
|
res.ReceiveOnlyChangedSymlinks = ro.Symlinks
|
|
res.ReceiveOnlyChangedDeletes = ro.Deleted
|
|
res.ReceiveOnlyChangedBytes = ro.Bytes
|
|
res.ReceiveOnlyTotalItems = ro.TotalItems()
|
|
}
|
|
|
|
res.InSyncFiles, res.InSyncBytes = global.Files-need.Files, global.Bytes-need.Bytes
|
|
|
|
res.State, res.StateChanged, err = c.model.State(folder)
|
|
if err != nil {
|
|
res.Error = err.Error()
|
|
}
|
|
|
|
res.Version = ourSeq + remoteSeq // legacy
|
|
res.Sequence = ourSeq + remoteSeq // new name
|
|
|
|
ignorePatterns, _, _ := c.model.CurrentIgnores(folder)
|
|
res.IgnorePatterns = false
|
|
for _, line := range ignorePatterns {
|
|
if len(line) > 0 && !strings.HasPrefix(line, "//") {
|
|
res.IgnorePatterns = true
|
|
break
|
|
}
|
|
}
|
|
|
|
err = c.model.WatchError(folder)
|
|
if err != nil {
|
|
res.WatchError = err.Error()
|
|
}
|
|
|
|
return res, nil
|
|
}
|
|
|
|
func (c *folderSummaryService) OnEventRequest() {
|
|
c.lastEventReqMut.Lock()
|
|
c.lastEventReq = time.Now()
|
|
c.lastEventReqMut.Unlock()
|
|
}
|
|
|
|
// listenForUpdates subscribes to the event bus and makes note of folders that
|
|
// need their data recalculated.
|
|
func (c *folderSummaryService) listenForUpdates(ctx context.Context) error {
|
|
sub := c.evLogger.Subscribe(events.LocalIndexUpdated | events.RemoteIndexUpdated | events.StateChanged | events.RemoteDownloadProgress | events.DeviceConnected | events.ClusterConfigReceived | events.FolderWatchStateChanged | events.DownloadProgress)
|
|
defer sub.Unsubscribe()
|
|
|
|
for {
|
|
// This loop needs to be fast so we don't miss too many events.
|
|
|
|
select {
|
|
case ev, ok := <-sub.C():
|
|
if !ok {
|
|
<-ctx.Done()
|
|
return ctx.Err()
|
|
}
|
|
c.processUpdate(ev)
|
|
case <-ctx.Done():
|
|
return ctx.Err()
|
|
}
|
|
}
|
|
}
|
|
|
|
func (c *folderSummaryService) processUpdate(ev events.Event) {
|
|
var folder string
|
|
|
|
switch ev.Type {
|
|
case events.DeviceConnected, events.ClusterConfigReceived:
|
|
// When a device connects we schedule a refresh of all
|
|
// folders shared with that device.
|
|
|
|
var deviceID protocol.DeviceID
|
|
if ev.Type == events.DeviceConnected {
|
|
data := ev.Data.(map[string]string)
|
|
deviceID, _ = protocol.DeviceIDFromString(data["id"])
|
|
} else {
|
|
data := ev.Data.(ClusterConfigReceivedEventData)
|
|
deviceID = data.Device
|
|
}
|
|
|
|
c.foldersMut.Lock()
|
|
nextFolder:
|
|
for _, folder := range c.cfg.Folders() {
|
|
for _, dev := range folder.Devices {
|
|
if dev.DeviceID == deviceID {
|
|
c.folders[folder.ID] = struct{}{}
|
|
continue nextFolder
|
|
}
|
|
}
|
|
}
|
|
c.foldersMut.Unlock()
|
|
|
|
return
|
|
|
|
case events.DownloadProgress:
|
|
data := ev.Data.(map[string]map[string]*pullerProgress)
|
|
c.foldersMut.Lock()
|
|
for folder := range data {
|
|
c.folders[folder] = struct{}{}
|
|
}
|
|
c.foldersMut.Unlock()
|
|
return
|
|
|
|
case events.StateChanged:
|
|
data := ev.Data.(map[string]interface{})
|
|
if data["to"].(string) != "idle" {
|
|
return
|
|
}
|
|
if from := data["from"].(string); from != "syncing" && from != "sync-preparing" {
|
|
return
|
|
}
|
|
|
|
// The folder changed to idle from syncing. We should do an
|
|
// immediate refresh to update the GUI. The send to
|
|
// c.immediate must be nonblocking so that we can continue
|
|
// handling events.
|
|
|
|
folder = data["folder"].(string)
|
|
select {
|
|
case c.immediate <- folder:
|
|
c.foldersMut.Lock()
|
|
delete(c.folders, folder)
|
|
c.foldersMut.Unlock()
|
|
return
|
|
default:
|
|
// Refresh whenever we do the next summary.
|
|
}
|
|
|
|
default:
|
|
// The other events all have a "folder" attribute that they
|
|
// affect. Whenever the local or remote index is updated for a
|
|
// given folder we make a note of it.
|
|
// This folder needs to be refreshed whenever we do the next
|
|
// refresh.
|
|
|
|
folder = ev.Data.(map[string]interface{})["folder"].(string)
|
|
}
|
|
|
|
c.foldersMut.Lock()
|
|
c.folders[folder] = struct{}{}
|
|
c.foldersMut.Unlock()
|
|
}
|
|
|
|
// calculateSummaries periodically recalculates folder summaries and
|
|
// completion percentage, and sends the results on the event bus.
|
|
func (c *folderSummaryService) calculateSummaries(ctx context.Context) error {
|
|
const pumpInterval = 2 * time.Second
|
|
pump := time.NewTimer(pumpInterval)
|
|
|
|
for {
|
|
select {
|
|
case <-pump.C:
|
|
t0 := time.Now()
|
|
for _, folder := range c.foldersToHandle() {
|
|
select {
|
|
case <-ctx.Done():
|
|
return ctx.Err()
|
|
default:
|
|
}
|
|
c.sendSummary(ctx, folder)
|
|
}
|
|
|
|
// We don't want to spend all our time calculating summaries. Lets
|
|
// set an arbitrary limit at not spending more than about 30% of
|
|
// our time here...
|
|
wait := 2*time.Since(t0) + pumpInterval
|
|
pump.Reset(wait)
|
|
|
|
case folder := <-c.immediate:
|
|
c.sendSummary(ctx, folder)
|
|
|
|
case <-ctx.Done():
|
|
return ctx.Err()
|
|
}
|
|
}
|
|
}
|
|
|
|
// foldersToHandle returns the list of folders needing a summary update, and
|
|
// clears the list.
|
|
func (c *folderSummaryService) foldersToHandle() []string {
|
|
// We only recalculate summaries if someone is listening to events
|
|
// (a request to /rest/events has been made within the last
|
|
// pingEventInterval).
|
|
|
|
c.lastEventReqMut.Lock()
|
|
last := c.lastEventReq
|
|
c.lastEventReqMut.Unlock()
|
|
if time.Since(last) > maxDurationSinceLastEventReq {
|
|
return nil
|
|
}
|
|
|
|
c.foldersMut.Lock()
|
|
res := make([]string, 0, len(c.folders))
|
|
for folder := range c.folders {
|
|
res = append(res, folder)
|
|
delete(c.folders, folder)
|
|
}
|
|
c.foldersMut.Unlock()
|
|
return res
|
|
}
|
|
|
|
type FolderSummaryEventData struct {
|
|
Folder string `json:"folder"`
|
|
Summary *FolderSummary `json:"summary"`
|
|
}
|
|
|
|
// sendSummary send the summary events for a single folder
|
|
func (c *folderSummaryService) sendSummary(ctx context.Context, folder string) {
|
|
// The folder summary contains how many bytes, files etc
|
|
// are in the folder and how in sync we are.
|
|
data, err := c.Summary(folder)
|
|
if err != nil {
|
|
return
|
|
}
|
|
c.evLogger.Log(events.FolderSummary, FolderSummaryEventData{
|
|
Folder: folder,
|
|
Summary: data,
|
|
})
|
|
|
|
metricFolderSummary.WithLabelValues(folder, metricScopeGlobal, metricTypeFiles).Set(float64(data.GlobalFiles))
|
|
metricFolderSummary.WithLabelValues(folder, metricScopeGlobal, metricTypeDirectories).Set(float64(data.GlobalDirectories))
|
|
metricFolderSummary.WithLabelValues(folder, metricScopeGlobal, metricTypeSymlinks).Set(float64(data.GlobalSymlinks))
|
|
metricFolderSummary.WithLabelValues(folder, metricScopeGlobal, metricTypeDeleted).Set(float64(data.GlobalDeleted))
|
|
metricFolderSummary.WithLabelValues(folder, metricScopeGlobal, metricTypeBytes).Set(float64(data.GlobalBytes))
|
|
|
|
metricFolderSummary.WithLabelValues(folder, metricScopeLocal, metricTypeFiles).Set(float64(data.LocalFiles))
|
|
metricFolderSummary.WithLabelValues(folder, metricScopeLocal, metricTypeDirectories).Set(float64(data.LocalDirectories))
|
|
metricFolderSummary.WithLabelValues(folder, metricScopeLocal, metricTypeSymlinks).Set(float64(data.LocalSymlinks))
|
|
metricFolderSummary.WithLabelValues(folder, metricScopeLocal, metricTypeDeleted).Set(float64(data.LocalDeleted))
|
|
metricFolderSummary.WithLabelValues(folder, metricScopeLocal, metricTypeBytes).Set(float64(data.LocalBytes))
|
|
|
|
metricFolderSummary.WithLabelValues(folder, metricScopeNeed, metricTypeFiles).Set(float64(data.NeedFiles))
|
|
metricFolderSummary.WithLabelValues(folder, metricScopeNeed, metricTypeDirectories).Set(float64(data.NeedDirectories))
|
|
metricFolderSummary.WithLabelValues(folder, metricScopeNeed, metricTypeSymlinks).Set(float64(data.NeedSymlinks))
|
|
metricFolderSummary.WithLabelValues(folder, metricScopeNeed, metricTypeDeleted).Set(float64(data.NeedDeletes))
|
|
metricFolderSummary.WithLabelValues(folder, metricScopeNeed, metricTypeBytes).Set(float64(data.NeedBytes))
|
|
|
|
for _, devCfg := range c.cfg.Folders()[folder].Devices {
|
|
select {
|
|
case <-ctx.Done():
|
|
return
|
|
default:
|
|
}
|
|
|
|
if devCfg.DeviceID.Equals(c.id) {
|
|
// We already know about ourselves.
|
|
continue
|
|
}
|
|
|
|
// Get completion percentage of this folder for the
|
|
// remote device.
|
|
comp, err := c.model.Completion(devCfg.DeviceID, folder)
|
|
if err != nil {
|
|
l.Debugf("Error getting completion for folder %v, device %v: %v", folder, devCfg.DeviceID, err)
|
|
continue
|
|
}
|
|
ev := comp.Map()
|
|
ev["folder"] = folder
|
|
ev["device"] = devCfg.DeviceID.String()
|
|
c.evLogger.Log(events.FolderCompletion, ev)
|
|
}
|
|
}
|