// Copyright (C) 2015 The Syncthing Authors. // // This Source Code Form is subject to the terms of the Mozilla Public // License, v. 2.0. If a copy of the MPL was not distributed with this file, // You can obtain one at https://mozilla.org/MPL/2.0/. package model import ( "context" "fmt" "strings" "time" "github.com/thejerf/suture/v4" "github.com/syncthing/syncthing/lib/config" "github.com/syncthing/syncthing/lib/db" "github.com/syncthing/syncthing/lib/events" "github.com/syncthing/syncthing/lib/protocol" "github.com/syncthing/syncthing/lib/svcutil" "github.com/syncthing/syncthing/lib/sync" ) const maxDurationSinceLastEventReq = time.Minute type FolderSummaryService interface { suture.Service Summary(folder string) (map[string]interface{}, error) OnEventRequest() } // The folderSummaryService adds summary information events (FolderSummary and // FolderCompletion) into the event stream at certain intervals. type folderSummaryService struct { *suture.Supervisor cfg config.Wrapper model Model id protocol.DeviceID evLogger events.Logger immediate chan string // For keeping track of folders to recalculate for foldersMut sync.Mutex folders map[string]struct{} // For keeping track of when the last event request on the API was lastEventReq time.Time lastEventReqMut sync.Mutex } func NewFolderSummaryService(cfg config.Wrapper, m Model, id protocol.DeviceID, evLogger events.Logger) FolderSummaryService { service := &folderSummaryService{ Supervisor: suture.New("folderSummaryService", svcutil.SpecWithDebugLogger(l)), cfg: cfg, model: m, id: id, evLogger: evLogger, immediate: make(chan string), folders: make(map[string]struct{}), foldersMut: sync.NewMutex(), lastEventReqMut: sync.NewMutex(), } service.Add(svcutil.AsService(service.listenForUpdates, fmt.Sprintf("%s/listenForUpdates", service))) service.Add(svcutil.AsService(service.calculateSummaries, fmt.Sprintf("%s/calculateSummaries", service))) return service } func (c *folderSummaryService) String() string { return fmt.Sprintf("FolderSummaryService@%p", c) } func (c *folderSummaryService) Summary(folder string) (map[string]interface{}, error) { var res = make(map[string]interface{}) var local, global, need, ro db.Counts var ourSeq, remoteSeq int64 errors, err := c.model.FolderErrors(folder) if err == nil { var snap *db.Snapshot if snap, err = c.model.DBSnapshot(folder); err == nil { global = snap.GlobalSize() local = snap.LocalSize() need = snap.NeedSize(protocol.LocalDeviceID) ro = snap.ReceiveOnlyChangedSize() ourSeq = snap.Sequence(protocol.LocalDeviceID) remoteSeq = snap.Sequence(protocol.GlobalDeviceID) snap.Release() } } // For API backwards compatibility (SyncTrayzor needs it) an empty folder // summary is returned for not running folders, an error might actually be // more appropriate if err != nil && err != ErrFolderPaused && err != errFolderNotRunning { return nil, err } res["errors"] = len(errors) res["pullErrors"] = len(errors) // deprecated res["invalid"] = "" // Deprecated, retains external API for now res["globalFiles"], res["globalDirectories"], res["globalSymlinks"], res["globalDeleted"], res["globalBytes"], res["globalTotalItems"] = global.Files, global.Directories, global.Symlinks, global.Deleted, global.Bytes, global.TotalItems() res["localFiles"], res["localDirectories"], res["localSymlinks"], res["localDeleted"], res["localBytes"], res["localTotalItems"] = local.Files, local.Directories, local.Symlinks, local.Deleted, local.Bytes, local.TotalItems() fcfg, haveFcfg := c.cfg.Folder(folder) if haveFcfg && fcfg.IgnoreDelete { need.Deleted = 0 } need.Bytes -= c.model.FolderProgressBytesCompleted(folder) // This may happen if we are in progress of pulling files that were // deleted globally after the pull started. if need.Bytes < 0 { need.Bytes = 0 } res["needFiles"], res["needDirectories"], res["needSymlinks"], res["needDeletes"], res["needBytes"], res["needTotalItems"] = need.Files, need.Directories, need.Symlinks, need.Deleted, need.Bytes, need.TotalItems() if haveFcfg && (fcfg.Type == config.FolderTypeReceiveOnly || fcfg.Type == config.FolderTypeReceiveEncrypted) { // Add statistics for things that have changed locally in a receive // only or receive encrypted folder. res["receiveOnlyChangedFiles"] = ro.Files res["receiveOnlyChangedDirectories"] = ro.Directories res["receiveOnlyChangedSymlinks"] = ro.Symlinks res["receiveOnlyChangedDeletes"] = ro.Deleted res["receiveOnlyChangedBytes"] = ro.Bytes res["receiveOnlyTotalItems"] = ro.TotalItems() } res["inSyncFiles"], res["inSyncBytes"] = global.Files-need.Files, global.Bytes-need.Bytes res["state"], res["stateChanged"], err = c.model.State(folder) if err != nil { res["error"] = err.Error() } res["version"] = ourSeq + remoteSeq // legacy res["sequence"] = ourSeq + remoteSeq // new name ignorePatterns, _, _ := c.model.CurrentIgnores(folder) res["ignorePatterns"] = false for _, line := range ignorePatterns { if len(line) > 0 && !strings.HasPrefix(line, "//") { res["ignorePatterns"] = true break } } err = c.model.WatchError(folder) if err != nil { res["watchError"] = err.Error() } return res, nil } func (c *folderSummaryService) OnEventRequest() { c.lastEventReqMut.Lock() c.lastEventReq = time.Now() c.lastEventReqMut.Unlock() } // listenForUpdates subscribes to the event bus and makes note of folders that // need their data recalculated. func (c *folderSummaryService) listenForUpdates(ctx context.Context) error { sub := c.evLogger.Subscribe(events.LocalIndexUpdated | events.RemoteIndexUpdated | events.StateChanged | events.RemoteDownloadProgress | events.DeviceConnected | events.FolderWatchStateChanged | events.DownloadProgress) defer sub.Unsubscribe() for { // This loop needs to be fast so we don't miss too many events. select { case ev := <-sub.C(): c.processUpdate(ev) case <-ctx.Done(): return ctx.Err() } } } func (c *folderSummaryService) processUpdate(ev events.Event) { var folder string switch ev.Type { case events.DeviceConnected: // When a device connects we schedule a refresh of all // folders shared with that device. data := ev.Data.(map[string]string) deviceID, _ := protocol.DeviceIDFromString(data["id"]) c.foldersMut.Lock() nextFolder: for _, folder := range c.cfg.Folders() { for _, dev := range folder.Devices { if dev.DeviceID == deviceID { c.folders[folder.ID] = struct{}{} continue nextFolder } } } c.foldersMut.Unlock() return case events.DownloadProgress: data := ev.Data.(map[string]map[string]*pullerProgress) c.foldersMut.Lock() for folder := range data { c.folders[folder] = struct{}{} } c.foldersMut.Unlock() return case events.StateChanged: data := ev.Data.(map[string]interface{}) if data["to"].(string) != "idle" { return } if from := data["from"].(string); from != "syncing" && from != "sync-preparing" { return } // The folder changed to idle from syncing. We should do an // immediate refresh to update the GUI. The send to // c.immediate must be nonblocking so that we can continue // handling events. folder = data["folder"].(string) select { case c.immediate <- folder: c.foldersMut.Lock() delete(c.folders, folder) c.foldersMut.Unlock() return default: // Refresh whenever we do the next summary. } default: // The other events all have a "folder" attribute that they // affect. Whenever the local or remote index is updated for a // given folder we make a note of it. // This folder needs to be refreshed whenever we do the next // refresh. folder = ev.Data.(map[string]interface{})["folder"].(string) } c.foldersMut.Lock() c.folders[folder] = struct{}{} c.foldersMut.Unlock() } // calculateSummaries periodically recalculates folder summaries and // completion percentage, and sends the results on the event bus. func (c *folderSummaryService) calculateSummaries(ctx context.Context) error { const pumpInterval = 2 * time.Second pump := time.NewTimer(pumpInterval) for { select { case <-pump.C: t0 := time.Now() for _, folder := range c.foldersToHandle() { select { case <-ctx.Done(): return ctx.Err() default: } c.sendSummary(ctx, folder) } // We don't want to spend all our time calculating summaries. Lets // set an arbitrary limit at not spending more than about 30% of // our time here... wait := 2*time.Since(t0) + pumpInterval pump.Reset(wait) case folder := <-c.immediate: c.sendSummary(ctx, folder) case <-ctx.Done(): return ctx.Err() } } } // foldersToHandle returns the list of folders needing a summary update, and // clears the list. func (c *folderSummaryService) foldersToHandle() []string { // We only recalculate summaries if someone is listening to events // (a request to /rest/events has been made within the last // pingEventInterval). c.lastEventReqMut.Lock() last := c.lastEventReq c.lastEventReqMut.Unlock() if time.Since(last) > maxDurationSinceLastEventReq { return nil } c.foldersMut.Lock() res := make([]string, 0, len(c.folders)) for folder := range c.folders { res = append(res, folder) delete(c.folders, folder) } c.foldersMut.Unlock() return res } // sendSummary send the summary events for a single folder func (c *folderSummaryService) sendSummary(ctx context.Context, folder string) { // The folder summary contains how many bytes, files etc // are in the folder and how in sync we are. data, err := c.Summary(folder) if err != nil { return } c.evLogger.Log(events.FolderSummary, map[string]interface{}{ "folder": folder, "summary": data, }) for _, devCfg := range c.cfg.Folders()[folder].Devices { select { case <-ctx.Done(): return default: } if devCfg.DeviceID.Equals(c.id) { // We already know about ourselves. continue } if _, ok := c.model.Connection(devCfg.DeviceID); !ok { // We're not interested in disconnected devices. continue } // Get completion percentage of this folder for the // remote device. comp := c.model.Completion(devCfg.DeviceID, folder).Map() comp["folder"] = folder comp["device"] = devCfg.DeviceID.String() c.evLogger.Log(events.FolderCompletion, comp) } }