mirror of
https://github.com/octoleo/syncthing.git
synced 2024-12-23 03:18:59 +00:00
This commit is contained in:
parent
e027175446
commit
a20c6ca868
@ -35,6 +35,7 @@ type fakeConnection struct {
|
|||||||
indexFn func(context.Context, string, []protocol.FileInfo)
|
indexFn func(context.Context, string, []protocol.FileInfo)
|
||||||
requestFn func(ctx context.Context, folder, name string, offset int64, size int, hash []byte, fromTemporary bool) ([]byte, error)
|
requestFn func(ctx context.Context, folder, name string, offset int64, size int, hash []byte, fromTemporary bool) ([]byte, error)
|
||||||
closeFn func(error)
|
closeFn func(error)
|
||||||
|
clusterConfigFn func(protocol.ClusterConfig)
|
||||||
mut sync.Mutex
|
mut sync.Mutex
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -91,7 +92,13 @@ func (f *fakeConnection) Request(ctx context.Context, folder, name string, offse
|
|||||||
return f.fileData[name], nil
|
return f.fileData[name], nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (f *fakeConnection) ClusterConfig(protocol.ClusterConfig) {}
|
func (f *fakeConnection) ClusterConfig(cc protocol.ClusterConfig) {
|
||||||
|
f.mut.Lock()
|
||||||
|
defer f.mut.Unlock()
|
||||||
|
if f.clusterConfigFn != nil {
|
||||||
|
f.clusterConfigFn(cc)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
func (f *fakeConnection) Ping() bool {
|
func (f *fakeConnection) Ping() bool {
|
||||||
f.mut.Lock()
|
f.mut.Lock()
|
||||||
|
@ -150,6 +150,7 @@ type model struct {
|
|||||||
helloMessages map[protocol.DeviceID]protocol.Hello
|
helloMessages map[protocol.DeviceID]protocol.Hello
|
||||||
deviceDownloads map[protocol.DeviceID]*deviceDownloadState
|
deviceDownloads map[protocol.DeviceID]*deviceDownloadState
|
||||||
remotePausedFolders map[protocol.DeviceID][]string // deviceID -> folders
|
remotePausedFolders map[protocol.DeviceID][]string // deviceID -> folders
|
||||||
|
indexSenderTokens map[protocol.DeviceID][]suture.ServiceToken
|
||||||
|
|
||||||
foldersRunning int32 // for testing only
|
foldersRunning int32 // for testing only
|
||||||
}
|
}
|
||||||
@ -222,6 +223,7 @@ func NewModel(cfg config.Wrapper, id protocol.DeviceID, clientName, clientVersio
|
|||||||
helloMessages: make(map[protocol.DeviceID]protocol.Hello),
|
helloMessages: make(map[protocol.DeviceID]protocol.Hello),
|
||||||
deviceDownloads: make(map[protocol.DeviceID]*deviceDownloadState),
|
deviceDownloads: make(map[protocol.DeviceID]*deviceDownloadState),
|
||||||
remotePausedFolders: make(map[protocol.DeviceID][]string),
|
remotePausedFolders: make(map[protocol.DeviceID][]string),
|
||||||
|
indexSenderTokens: make(map[protocol.DeviceID][]suture.ServiceToken),
|
||||||
}
|
}
|
||||||
for devID := range cfg.Devices() {
|
for devID := range cfg.Devices() {
|
||||||
m.deviceStatRefs[devID] = stats.NewDeviceStatisticsReference(m.db, devID.String())
|
m.deviceStatRefs[devID] = stats.NewDeviceStatisticsReference(m.db, devID.String())
|
||||||
@ -257,13 +259,16 @@ func (m *model) onServe() {
|
|||||||
func (m *model) Stop() {
|
func (m *model) Stop() {
|
||||||
m.cfg.Unsubscribe(m)
|
m.cfg.Unsubscribe(m)
|
||||||
m.Supervisor.Stop()
|
m.Supervisor.Stop()
|
||||||
devs := m.cfg.Devices()
|
m.pmut.RLock()
|
||||||
ids := make([]protocol.DeviceID, 0, len(devs))
|
closed := make([]chan struct{}, 0, len(m.conn))
|
||||||
for id := range devs {
|
for id, conn := range m.conn {
|
||||||
ids = append(ids, id)
|
closed = append(closed, m.closed[id])
|
||||||
|
go conn.Close(errStopped)
|
||||||
|
}
|
||||||
|
m.pmut.RUnlock()
|
||||||
|
for _, c := range closed {
|
||||||
|
<-c
|
||||||
}
|
}
|
||||||
w := m.closeConns(ids, errStopped)
|
|
||||||
w.Wait()
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// StartDeadlockDetector starts a deadlock detector on the models locks which
|
// StartDeadlockDetector starts a deadlock detector on the models locks which
|
||||||
@ -393,7 +398,12 @@ func (m *model) warnAboutOverwritingProtectedFiles(cfg config.FolderConfiguratio
|
|||||||
}
|
}
|
||||||
|
|
||||||
func (m *model) removeFolder(cfg config.FolderConfiguration) {
|
func (m *model) removeFolder(cfg config.FolderConfiguration) {
|
||||||
m.stopFolder(cfg, fmt.Errorf("removing folder %v", cfg.Description()))
|
m.fmut.RLock()
|
||||||
|
token, ok := m.folderRunnerToken[cfg.ID]
|
||||||
|
m.fmut.RUnlock()
|
||||||
|
if ok {
|
||||||
|
m.RemoveAndWait(token, 0)
|
||||||
|
}
|
||||||
|
|
||||||
m.fmut.Lock()
|
m.fmut.Lock()
|
||||||
|
|
||||||
@ -417,22 +427,6 @@ func (m *model) removeFolder(cfg config.FolderConfiguration) {
|
|||||||
db.DropFolder(m.db, cfg.ID)
|
db.DropFolder(m.db, cfg.ID)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (m *model) stopFolder(cfg config.FolderConfiguration, err error) {
|
|
||||||
// Stop the services running for this folder and wait for them to finish
|
|
||||||
// stopping to prevent races on restart.
|
|
||||||
m.fmut.RLock()
|
|
||||||
token, ok := m.folderRunnerToken[cfg.ID]
|
|
||||||
m.fmut.RUnlock()
|
|
||||||
|
|
||||||
if ok {
|
|
||||||
m.RemoveAndWait(token, 0)
|
|
||||||
}
|
|
||||||
|
|
||||||
// Wait for connections to stop to ensure that no more calls to methods
|
|
||||||
// expecting this folder to exist happen (e.g. .IndexUpdate).
|
|
||||||
m.closeConns(cfg.DeviceIDs(), err).Wait()
|
|
||||||
}
|
|
||||||
|
|
||||||
// Need to hold lock on m.fmut when calling this.
|
// Need to hold lock on m.fmut when calling this.
|
||||||
func (m *model) cleanupFolderLocked(cfg config.FolderConfiguration) {
|
func (m *model) cleanupFolderLocked(cfg config.FolderConfiguration) {
|
||||||
// clear up our config maps
|
// clear up our config maps
|
||||||
@ -464,25 +458,13 @@ func (m *model) restartFolder(from, to config.FolderConfiguration, cacheIgnoredF
|
|||||||
restartMut.Lock()
|
restartMut.Lock()
|
||||||
defer restartMut.Unlock()
|
defer restartMut.Unlock()
|
||||||
|
|
||||||
var infoMsg string
|
m.fmut.RLock()
|
||||||
var errMsg string
|
token, ok := m.folderRunnerToken[from.ID]
|
||||||
switch {
|
m.fmut.RUnlock()
|
||||||
case to.Paused:
|
if ok {
|
||||||
infoMsg = "Paused"
|
m.RemoveAndWait(token, 0)
|
||||||
errMsg = "pausing"
|
|
||||||
case from.Paused:
|
|
||||||
infoMsg = "Unpaused"
|
|
||||||
errMsg = "unpausing"
|
|
||||||
default:
|
|
||||||
infoMsg = "Restarted"
|
|
||||||
errMsg = "restarting"
|
|
||||||
}
|
}
|
||||||
|
|
||||||
err := fmt.Errorf("%v folder %v", errMsg, to.Description())
|
|
||||||
m.stopFolder(from, err)
|
|
||||||
// Need to send CC change to both from and to devices.
|
|
||||||
m.closeConns(to.DeviceIDs(), err)
|
|
||||||
|
|
||||||
m.fmut.Lock()
|
m.fmut.Lock()
|
||||||
defer m.fmut.Unlock()
|
defer m.fmut.Unlock()
|
||||||
|
|
||||||
@ -499,6 +481,16 @@ func (m *model) restartFolder(from, to config.FolderConfiguration, cacheIgnoredF
|
|||||||
}
|
}
|
||||||
m.addAndStartFolderLocked(to, fset, cacheIgnoredFiles)
|
m.addAndStartFolderLocked(to, fset, cacheIgnoredFiles)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
var infoMsg string
|
||||||
|
switch {
|
||||||
|
case to.Paused:
|
||||||
|
infoMsg = "Paused"
|
||||||
|
case from.Paused:
|
||||||
|
infoMsg = "Unpaused"
|
||||||
|
default:
|
||||||
|
infoMsg = "Restarted"
|
||||||
|
}
|
||||||
l.Infof("%v folder %v (%v)", infoMsg, to.Description(), to.Type)
|
l.Infof("%v folder %v (%v)", infoMsg, to.Description(), to.Type)
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -507,9 +499,6 @@ func (m *model) newFolder(cfg config.FolderConfiguration, cacheIgnoredFiles bool
|
|||||||
// we do it outside of the lock.
|
// we do it outside of the lock.
|
||||||
fset := db.NewFileSet(cfg.ID, cfg.Filesystem(), m.db)
|
fset := db.NewFileSet(cfg.ID, cfg.Filesystem(), m.db)
|
||||||
|
|
||||||
// Close connections to affected devices
|
|
||||||
m.closeConns(cfg.DeviceIDs(), fmt.Errorf("started folder %v", cfg.Description()))
|
|
||||||
|
|
||||||
m.fmut.Lock()
|
m.fmut.Lock()
|
||||||
defer m.fmut.Unlock()
|
defer m.fmut.Unlock()
|
||||||
m.addAndStartFolderLocked(cfg, fset, cacheIgnoredFiles)
|
m.addAndStartFolderLocked(cfg, fset, cacheIgnoredFiles)
|
||||||
@ -992,6 +981,9 @@ func (m *model) ClusterConfig(deviceID protocol.DeviceID, cm protocol.ClusterCon
|
|||||||
m.pmut.RLock()
|
m.pmut.RLock()
|
||||||
conn, ok := m.conn[deviceID]
|
conn, ok := m.conn[deviceID]
|
||||||
closed := m.closed[deviceID]
|
closed := m.closed[deviceID]
|
||||||
|
for _, token := range m.indexSenderTokens[deviceID] {
|
||||||
|
m.RemoveAndWait(token, 0)
|
||||||
|
}
|
||||||
m.pmut.RUnlock()
|
m.pmut.RUnlock()
|
||||||
if !ok {
|
if !ok {
|
||||||
panic("bug: ClusterConfig called on closed or nonexistent connection")
|
panic("bug: ClusterConfig called on closed or nonexistent connection")
|
||||||
@ -1024,6 +1016,7 @@ func (m *model) ClusterConfig(deviceID protocol.DeviceID, cm protocol.ClusterCon
|
|||||||
}
|
}
|
||||||
|
|
||||||
var paused []string
|
var paused []string
|
||||||
|
indexSenderTokens := make([]suture.ServiceToken, 0, len(cm.Folders))
|
||||||
for _, folder := range cm.Folders {
|
for _, folder := range cm.Folders {
|
||||||
cfg, ok := m.cfg.Folder(folder.ID)
|
cfg, ok := m.cfg.Folder(folder.ID)
|
||||||
if !ok || !cfg.SharedWith(deviceID) {
|
if !ok || !cfg.SharedWith(deviceID) {
|
||||||
@ -1142,14 +1135,12 @@ func (m *model) ClusterConfig(deviceID protocol.DeviceID, cm protocol.ClusterCon
|
|||||||
evLogger: m.evLogger,
|
evLogger: m.evLogger,
|
||||||
}
|
}
|
||||||
is.Service = util.AsService(is.serve, is.String())
|
is.Service = util.AsService(is.serve, is.String())
|
||||||
// The token isn't tracked as the service stops when the connection
|
indexSenderTokens = append(indexSenderTokens, m.Add(is))
|
||||||
// terminates and is automatically removed from supervisor (by
|
|
||||||
// implementing suture.IsCompletable).
|
|
||||||
m.Add(is)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
m.pmut.Lock()
|
m.pmut.Lock()
|
||||||
m.remotePausedFolders[deviceID] = paused
|
m.remotePausedFolders[deviceID] = paused
|
||||||
|
m.indexSenderTokens[deviceID] = indexSenderTokens
|
||||||
m.pmut.Unlock()
|
m.pmut.Unlock()
|
||||||
|
|
||||||
// This breaks if we send multiple CM messages during the same connection.
|
// This breaks if we send multiple CM messages during the same connection.
|
||||||
@ -1397,41 +1388,6 @@ func (m *model) Closed(conn protocol.Connection, err error) {
|
|||||||
close(closed)
|
close(closed)
|
||||||
}
|
}
|
||||||
|
|
||||||
// closeConns will close the underlying connection for given devices and return
|
|
||||||
// a waiter that will return once all the connections are finished closing.
|
|
||||||
func (m *model) closeConns(devs []protocol.DeviceID, err error) config.Waiter {
|
|
||||||
conns := make([]connections.Connection, 0, len(devs))
|
|
||||||
closed := make([]chan struct{}, 0, len(devs))
|
|
||||||
m.pmut.RLock()
|
|
||||||
for _, dev := range devs {
|
|
||||||
if conn, ok := m.conn[dev]; ok {
|
|
||||||
conns = append(conns, conn)
|
|
||||||
closed = append(closed, m.closed[dev])
|
|
||||||
}
|
|
||||||
}
|
|
||||||
m.pmut.RUnlock()
|
|
||||||
for _, conn := range conns {
|
|
||||||
conn.Close(err)
|
|
||||||
}
|
|
||||||
return &channelWaiter{chans: closed}
|
|
||||||
}
|
|
||||||
|
|
||||||
// closeConn closes the underlying connection for the given device and returns
|
|
||||||
// a waiter that will return once the connection is finished closing.
|
|
||||||
func (m *model) closeConn(dev protocol.DeviceID, err error) config.Waiter {
|
|
||||||
return m.closeConns([]protocol.DeviceID{dev}, err)
|
|
||||||
}
|
|
||||||
|
|
||||||
type channelWaiter struct {
|
|
||||||
chans []chan struct{}
|
|
||||||
}
|
|
||||||
|
|
||||||
func (w *channelWaiter) Wait() {
|
|
||||||
for _, c := range w.chans {
|
|
||||||
<-c
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// Implements protocol.RequestResponse
|
// Implements protocol.RequestResponse
|
||||||
type requestResponse struct {
|
type requestResponse struct {
|
||||||
data []byte
|
data []byte
|
||||||
@ -2467,6 +2423,9 @@ func (m *model) CommitConfiguration(from, to config.Configuration) bool {
|
|||||||
|
|
||||||
// Go through the folder configs and figure out if we need to restart or not.
|
// Go through the folder configs and figure out if we need to restart or not.
|
||||||
|
|
||||||
|
// Tracks devices affected by any configuration change to resend ClusterConfig.
|
||||||
|
clusterConfigDevices := make(map[protocol.DeviceID]struct{}, len(from.Devices)+len(to.Devices))
|
||||||
|
|
||||||
fromFolders := mapFolders(from.Folders)
|
fromFolders := mapFolders(from.Folders)
|
||||||
toFolders := mapFolders(to.Folders)
|
toFolders := mapFolders(to.Folders)
|
||||||
for folderID, cfg := range toFolders {
|
for folderID, cfg := range toFolders {
|
||||||
@ -2478,6 +2437,7 @@ func (m *model) CommitConfiguration(from, to config.Configuration) bool {
|
|||||||
l.Infoln("Adding folder", cfg.Description())
|
l.Infoln("Adding folder", cfg.Description())
|
||||||
m.newFolder(cfg, to.Options.CacheIgnoredFiles)
|
m.newFolder(cfg, to.Options.CacheIgnoredFiles)
|
||||||
}
|
}
|
||||||
|
clusterConfigDevices = addDeviceIDsToMap(clusterConfigDevices, cfg.DeviceIDs())
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -2486,6 +2446,7 @@ func (m *model) CommitConfiguration(from, to config.Configuration) bool {
|
|||||||
if !ok {
|
if !ok {
|
||||||
// The folder was removed.
|
// The folder was removed.
|
||||||
m.removeFolder(fromCfg)
|
m.removeFolder(fromCfg)
|
||||||
|
clusterConfigDevices = addDeviceIDsToMap(clusterConfigDevices, fromCfg.DeviceIDs())
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -2497,6 +2458,8 @@ func (m *model) CommitConfiguration(from, to config.Configuration) bool {
|
|||||||
// Check if anything differs that requires a restart.
|
// Check if anything differs that requires a restart.
|
||||||
if !reflect.DeepEqual(fromCfg.RequiresRestartOnly(), toCfg.RequiresRestartOnly()) || from.Options.CacheIgnoredFiles != to.Options.CacheIgnoredFiles {
|
if !reflect.DeepEqual(fromCfg.RequiresRestartOnly(), toCfg.RequiresRestartOnly()) || from.Options.CacheIgnoredFiles != to.Options.CacheIgnoredFiles {
|
||||||
m.restartFolder(fromCfg, toCfg, to.Options.CacheIgnoredFiles)
|
m.restartFolder(fromCfg, toCfg, to.Options.CacheIgnoredFiles)
|
||||||
|
clusterConfigDevices = addDeviceIDsToMap(clusterConfigDevices, fromCfg.DeviceIDs())
|
||||||
|
clusterConfigDevices = addDeviceIDsToMap(clusterConfigDevices, toCfg.DeviceIDs())
|
||||||
}
|
}
|
||||||
|
|
||||||
// Emit the folder pause/resume event
|
// Emit the folder pause/resume event
|
||||||
@ -2519,6 +2482,7 @@ func (m *model) CommitConfiguration(from, to config.Configuration) bool {
|
|||||||
// Pausing a device, unpausing is handled by the connection service.
|
// Pausing a device, unpausing is handled by the connection service.
|
||||||
fromDevices := from.DeviceMap()
|
fromDevices := from.DeviceMap()
|
||||||
toDevices := to.DeviceMap()
|
toDevices := to.DeviceMap()
|
||||||
|
closeDevices := make([]protocol.DeviceID, 0, len(to.Devices))
|
||||||
for deviceID, toCfg := range toDevices {
|
for deviceID, toCfg := range toDevices {
|
||||||
fromCfg, ok := fromDevices[deviceID]
|
fromCfg, ok := fromDevices[deviceID]
|
||||||
if !ok {
|
if !ok {
|
||||||
@ -2534,13 +2498,14 @@ func (m *model) CommitConfiguration(from, to config.Configuration) bool {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Ignored folder was removed, reconnect to retrigger the prompt.
|
// Ignored folder was removed, reconnect to retrigger the prompt.
|
||||||
if len(fromCfg.IgnoredFolders) > len(toCfg.IgnoredFolders) {
|
if !toCfg.Paused && len(fromCfg.IgnoredFolders) > len(toCfg.IgnoredFolders) {
|
||||||
m.closeConn(deviceID, errIgnoredFolderRemoved)
|
closeDevices = append(closeDevices, deviceID)
|
||||||
}
|
}
|
||||||
|
|
||||||
if toCfg.Paused {
|
if toCfg.Paused {
|
||||||
l.Infoln("Pausing", deviceID)
|
l.Infoln("Pausing", deviceID)
|
||||||
m.closeConn(deviceID, errDevicePaused)
|
closeDevices = append(closeDevices, deviceID)
|
||||||
|
delete(clusterConfigDevices, deviceID)
|
||||||
m.evLogger.Log(events.DevicePaused, map[string]string{"device": deviceID.String()})
|
m.evLogger.Log(events.DevicePaused, map[string]string{"device": deviceID.String()})
|
||||||
} else {
|
} else {
|
||||||
m.evLogger.Log(events.DeviceResumed, map[string]string{"device": deviceID.String()})
|
m.evLogger.Log(events.DeviceResumed, map[string]string{"device": deviceID.String()})
|
||||||
@ -2551,9 +2516,28 @@ func (m *model) CommitConfiguration(from, to config.Configuration) bool {
|
|||||||
for deviceID := range fromDevices {
|
for deviceID := range fromDevices {
|
||||||
delete(m.deviceStatRefs, deviceID)
|
delete(m.deviceStatRefs, deviceID)
|
||||||
removedDevices = append(removedDevices, deviceID)
|
removedDevices = append(removedDevices, deviceID)
|
||||||
|
delete(clusterConfigDevices, deviceID)
|
||||||
}
|
}
|
||||||
m.fmut.Unlock()
|
m.fmut.Unlock()
|
||||||
m.closeConns(removedDevices, errDeviceRemoved)
|
|
||||||
|
m.pmut.RLock()
|
||||||
|
for _, id := range closeDevices {
|
||||||
|
if conn, ok := m.conn[id]; ok {
|
||||||
|
go conn.Close(errDevicePaused)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
for _, id := range removedDevices {
|
||||||
|
if conn, ok := m.conn[id]; ok {
|
||||||
|
go conn.Close(errDeviceRemoved)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
for id := range clusterConfigDevices {
|
||||||
|
if conn, ok := m.conn[id]; ok {
|
||||||
|
cm := m.generateClusterConfig(conn.ID())
|
||||||
|
go conn.ClusterConfig(cm)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
m.pmut.RUnlock()
|
||||||
|
|
||||||
m.globalRequestLimiter.setCapacity(1024 * to.Options.MaxConcurrentIncomingRequestKiB())
|
m.globalRequestLimiter.setCapacity(1024 * to.Options.MaxConcurrentIncomingRequestKiB())
|
||||||
m.folderIOLimiter.setCapacity(to.Options.MaxFolderConcurrency())
|
m.folderIOLimiter.setCapacity(to.Options.MaxFolderConcurrency())
|
||||||
@ -2758,3 +2742,12 @@ func sanitizePath(path string) string {
|
|||||||
|
|
||||||
return strings.TrimSpace(b.String())
|
return strings.TrimSpace(b.String())
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func addDeviceIDsToMap(m map[protocol.DeviceID]struct{}, s []protocol.DeviceID) map[protocol.DeviceID]struct{} {
|
||||||
|
for _, id := range s {
|
||||||
|
if _, ok := m[id]; !ok {
|
||||||
|
m[id] = struct{}{}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return m
|
||||||
|
}
|
||||||
|
@ -3339,17 +3339,19 @@ func TestConnCloseOnRestart(t *testing.T) {
|
|||||||
closed := m.closed[device1]
|
closed := m.closed[device1]
|
||||||
m.pmut.RUnlock()
|
m.pmut.RUnlock()
|
||||||
|
|
||||||
newFcfg := fcfg.Copy()
|
waiter, err := w.RemoveDevice(device1)
|
||||||
newFcfg.Paused = true
|
if err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
done := make(chan struct{})
|
done := make(chan struct{})
|
||||||
go func() {
|
go func() {
|
||||||
m.restartFolder(fcfg, newFcfg, false)
|
waiter.Wait()
|
||||||
close(done)
|
close(done)
|
||||||
}()
|
}()
|
||||||
select {
|
select {
|
||||||
case <-done:
|
case <-done:
|
||||||
case <-time.After(5 * time.Second):
|
case <-time.After(5 * time.Second):
|
||||||
t.Fatal("Timed out before folder restart returned")
|
t.Fatal("Timed out before config took effect")
|
||||||
}
|
}
|
||||||
select {
|
select {
|
||||||
case <-closed:
|
case <-closed:
|
||||||
@ -3845,8 +3847,8 @@ func TestScanRenameCaseOnly(t *testing.T) {
|
|||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestConnectionTerminationOnFolderAdd(t *testing.T) {
|
func TestClusterConfigOnFolderAdd(t *testing.T) {
|
||||||
testConfigChangeClosesConnections(t, false, true, nil, func(cfg config.Wrapper) {
|
testConfigChangeTriggersClusterConfigs(t, false, true, nil, func(cfg config.Wrapper) {
|
||||||
fcfg := testFolderConfigTmp()
|
fcfg := testFolderConfigTmp()
|
||||||
fcfg.ID = "second"
|
fcfg.ID = "second"
|
||||||
fcfg.Label = "second"
|
fcfg.Label = "second"
|
||||||
@ -3859,8 +3861,8 @@ func TestConnectionTerminationOnFolderAdd(t *testing.T) {
|
|||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestConnectionTerminationOnFolderShare(t *testing.T) {
|
func TestClusterConfigOnFolderShare(t *testing.T) {
|
||||||
testConfigChangeClosesConnections(t, true, true, nil, func(cfg config.Wrapper) {
|
testConfigChangeTriggersClusterConfigs(t, true, true, nil, func(cfg config.Wrapper) {
|
||||||
fcfg := cfg.FolderList()[0]
|
fcfg := cfg.FolderList()[0]
|
||||||
fcfg.Devices = []config.FolderDeviceConfiguration{{device2, protocol.EmptyDeviceID}}
|
fcfg.Devices = []config.FolderDeviceConfiguration{{device2, protocol.EmptyDeviceID}}
|
||||||
if w, err := cfg.SetFolder(fcfg); err != nil {
|
if w, err := cfg.SetFolder(fcfg); err != nil {
|
||||||
@ -3871,8 +3873,8 @@ func TestConnectionTerminationOnFolderShare(t *testing.T) {
|
|||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestConnectionTerminationOnFolderUnshare(t *testing.T) {
|
func TestClusterConfigOnFolderUnshare(t *testing.T) {
|
||||||
testConfigChangeClosesConnections(t, true, false, nil, func(cfg config.Wrapper) {
|
testConfigChangeTriggersClusterConfigs(t, true, false, nil, func(cfg config.Wrapper) {
|
||||||
fcfg := cfg.FolderList()[0]
|
fcfg := cfg.FolderList()[0]
|
||||||
fcfg.Devices = nil
|
fcfg.Devices = nil
|
||||||
if w, err := cfg.SetFolder(fcfg); err != nil {
|
if w, err := cfg.SetFolder(fcfg); err != nil {
|
||||||
@ -3883,8 +3885,8 @@ func TestConnectionTerminationOnFolderUnshare(t *testing.T) {
|
|||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestConnectionTerminationOnFolderRemove(t *testing.T) {
|
func TestClusterConfigOnFolderRemove(t *testing.T) {
|
||||||
testConfigChangeClosesConnections(t, true, false, nil, func(cfg config.Wrapper) {
|
testConfigChangeTriggersClusterConfigs(t, true, false, nil, func(cfg config.Wrapper) {
|
||||||
rcfg := cfg.RawCopy()
|
rcfg := cfg.RawCopy()
|
||||||
rcfg.Folders = nil
|
rcfg.Folders = nil
|
||||||
if w, err := cfg.Replace(rcfg); err != nil {
|
if w, err := cfg.Replace(rcfg); err != nil {
|
||||||
@ -3895,8 +3897,8 @@ func TestConnectionTerminationOnFolderRemove(t *testing.T) {
|
|||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestConnectionTerminationOnFolderPause(t *testing.T) {
|
func TestClusterConfigOnFolderPause(t *testing.T) {
|
||||||
testConfigChangeClosesConnections(t, true, false, nil, func(cfg config.Wrapper) {
|
testConfigChangeTriggersClusterConfigs(t, true, false, nil, func(cfg config.Wrapper) {
|
||||||
fcfg := cfg.FolderList()[0]
|
fcfg := cfg.FolderList()[0]
|
||||||
fcfg.Paused = true
|
fcfg.Paused = true
|
||||||
if w, err := cfg.SetFolder(fcfg); err != nil {
|
if w, err := cfg.SetFolder(fcfg); err != nil {
|
||||||
@ -3907,8 +3909,8 @@ func TestConnectionTerminationOnFolderPause(t *testing.T) {
|
|||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestConnectionTerminationOnFolderUnpause(t *testing.T) {
|
func TestClusterConfigOnFolderUnpause(t *testing.T) {
|
||||||
testConfigChangeClosesConnections(t, true, false, func(cfg config.Wrapper) {
|
testConfigChangeTriggersClusterConfigs(t, true, false, func(cfg config.Wrapper) {
|
||||||
fcfg := cfg.FolderList()[0]
|
fcfg := cfg.FolderList()[0]
|
||||||
fcfg.Paused = true
|
fcfg.Paused = true
|
||||||
if w, err := cfg.SetFolder(fcfg); err != nil {
|
if w, err := cfg.SetFolder(fcfg); err != nil {
|
||||||
@ -3984,7 +3986,7 @@ func TestScanDeletedROChangedOnSR(t *testing.T) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func testConfigChangeClosesConnections(t *testing.T, expectFirstClosed, expectSecondClosed bool, pre func(config.Wrapper), fn func(config.Wrapper)) {
|
func testConfigChangeTriggersClusterConfigs(t *testing.T, expectFirst, expectSecond bool, pre func(config.Wrapper), fn func(config.Wrapper)) {
|
||||||
t.Helper()
|
t.Helper()
|
||||||
wcfg, _ := tmpDefaultWrapper()
|
wcfg, _ := tmpDefaultWrapper()
|
||||||
m := setupModel(wcfg)
|
m := setupModel(wcfg)
|
||||||
@ -3999,21 +4001,55 @@ func testConfigChangeClosesConnections(t *testing.T, expectFirstClosed, expectSe
|
|||||||
pre(wcfg)
|
pre(wcfg)
|
||||||
}
|
}
|
||||||
|
|
||||||
fc1 := &fakeConnection{id: device1, model: m}
|
cc1 := make(chan struct{}, 1)
|
||||||
fc2 := &fakeConnection{id: device2, model: m}
|
cc2 := make(chan struct{}, 1)
|
||||||
|
fc1 := &fakeConnection{
|
||||||
|
id: device1,
|
||||||
|
model: m,
|
||||||
|
clusterConfigFn: func(_ protocol.ClusterConfig) {
|
||||||
|
cc1 <- struct{}{}
|
||||||
|
},
|
||||||
|
}
|
||||||
|
fc2 := &fakeConnection{
|
||||||
|
id: device2,
|
||||||
|
model: m,
|
||||||
|
clusterConfigFn: func(_ protocol.ClusterConfig) {
|
||||||
|
cc2 <- struct{}{}
|
||||||
|
},
|
||||||
|
}
|
||||||
m.AddConnection(fc1, protocol.Hello{})
|
m.AddConnection(fc1, protocol.Hello{})
|
||||||
m.AddConnection(fc2, protocol.Hello{})
|
m.AddConnection(fc2, protocol.Hello{})
|
||||||
|
|
||||||
|
// Initial CCs
|
||||||
|
select {
|
||||||
|
case <-cc1:
|
||||||
|
default:
|
||||||
|
t.Fatal("missing initial CC from device1")
|
||||||
|
}
|
||||||
|
select {
|
||||||
|
case <-cc2:
|
||||||
|
default:
|
||||||
|
t.Fatal("missing initial CC from device2")
|
||||||
|
}
|
||||||
|
|
||||||
t.Log("Applying config change")
|
t.Log("Applying config change")
|
||||||
|
|
||||||
fn(wcfg)
|
fn(wcfg)
|
||||||
|
|
||||||
if expectFirstClosed != fc1.closed {
|
timeout := time.NewTimer(time.Second)
|
||||||
t.Errorf("first connection state mismatch: %t (expected) != %t", expectFirstClosed, fc1.closed)
|
if expectFirst {
|
||||||
|
select {
|
||||||
|
case <-cc1:
|
||||||
|
case <-timeout.C:
|
||||||
|
t.Errorf("timed out before receiving cluste rconfig for first device")
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
if expectSecond {
|
||||||
if expectSecondClosed != fc2.closed {
|
select {
|
||||||
t.Errorf("second connection state mismatch: %t (expected) != %t", expectSecondClosed, fc2.closed)
|
case <-cc2:
|
||||||
|
case <-timeout.C:
|
||||||
|
t.Errorf("timed out before receiving cluste rconfig for second device")
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -322,11 +322,9 @@ func (c *rawConnection) Request(ctx context.Context, folder string, name string,
|
|||||||
}
|
}
|
||||||
|
|
||||||
// ClusterConfig sends the cluster configuration message to the peer.
|
// ClusterConfig sends the cluster configuration message to the peer.
|
||||||
// It must be called just once (as per BEP), otherwise it will panic.
|
|
||||||
func (c *rawConnection) ClusterConfig(config ClusterConfig) {
|
func (c *rawConnection) ClusterConfig(config ClusterConfig) {
|
||||||
select {
|
select {
|
||||||
case c.clusterConfigBox <- &config:
|
case c.clusterConfigBox <- &config:
|
||||||
close(c.clusterConfigBox)
|
|
||||||
case <-c.closed:
|
case <-c.closed:
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -386,13 +384,12 @@ func (c *rawConnection) dispatcherLoop() (err error) {
|
|||||||
switch msg := msg.(type) {
|
switch msg := msg.(type) {
|
||||||
case *ClusterConfig:
|
case *ClusterConfig:
|
||||||
l.Debugln("read ClusterConfig message")
|
l.Debugln("read ClusterConfig message")
|
||||||
if state != stateInitial {
|
if state == stateInitial {
|
||||||
return fmt.Errorf("protocol error: cluster config message in state %d", state)
|
state = stateReady
|
||||||
}
|
}
|
||||||
if err := c.receiver.ClusterConfig(c.id, *msg); err != nil {
|
if err := c.receiver.ClusterConfig(c.id, *msg); err != nil {
|
||||||
return errors.Wrap(err, "receiver error")
|
return errors.Wrap(err, "receiver error")
|
||||||
}
|
}
|
||||||
state = stateReady
|
|
||||||
|
|
||||||
case *Index:
|
case *Index:
|
||||||
l.Debugln("read Index message")
|
l.Debugln("read Index message")
|
||||||
@ -683,6 +680,12 @@ func (c *rawConnection) writerLoop() {
|
|||||||
}
|
}
|
||||||
for {
|
for {
|
||||||
select {
|
select {
|
||||||
|
case cc := <-c.clusterConfigBox:
|
||||||
|
err := c.writeMessage(cc)
|
||||||
|
if err != nil {
|
||||||
|
c.internalClose(err)
|
||||||
|
return
|
||||||
|
}
|
||||||
case hm := <-c.outbox:
|
case hm := <-c.outbox:
|
||||||
err := c.writeMessage(hm.msg)
|
err := c.writeMessage(hm.msg)
|
||||||
if hm.done != nil {
|
if hm.done != nil {
|
||||||
|
Loading…
Reference in New Issue
Block a user