mirror of
https://github.com/octoleo/syncthing.git
synced 2024-12-23 03:18:59 +00:00
348 lines
9.9 KiB
Go
348 lines
9.9 KiB
Go
// Copyright (C) 2017 The Syncthing Authors.
|
|
//
|
|
// This Source Code Form is subject to the terms of the Mozilla Public
|
|
// License, v. 2.0. If a copy of the MPL was not distributed with this file,
|
|
// You can obtain one at https://mozilla.org/MPL/2.0/.
|
|
|
|
package connections
|
|
|
|
import (
|
|
"context"
|
|
"fmt"
|
|
"io"
|
|
"sync/atomic"
|
|
|
|
"github.com/syncthing/syncthing/lib/config"
|
|
"github.com/syncthing/syncthing/lib/protocol"
|
|
"github.com/syncthing/syncthing/lib/sync"
|
|
"golang.org/x/time/rate"
|
|
)
|
|
|
|
// limiter manages a read and write rate limit, reacting to config changes
|
|
// as appropriate.
|
|
type limiter struct {
|
|
myID protocol.DeviceID
|
|
mu sync.Mutex
|
|
write *rate.Limiter
|
|
read *rate.Limiter
|
|
limitsLAN atomic.Bool
|
|
deviceReadLimiters map[protocol.DeviceID]*rate.Limiter
|
|
deviceWriteLimiters map[protocol.DeviceID]*rate.Limiter
|
|
}
|
|
|
|
type waiter interface {
|
|
// This is the rate limiting operation
|
|
WaitN(ctx context.Context, n int) error
|
|
Limit() rate.Limit
|
|
}
|
|
|
|
const (
|
|
limiterBurstSize = 4 * 128 << 10
|
|
)
|
|
|
|
func newLimiter(myId protocol.DeviceID, cfg config.Wrapper) *limiter {
|
|
l := &limiter{
|
|
myID: myId,
|
|
write: rate.NewLimiter(rate.Inf, limiterBurstSize),
|
|
read: rate.NewLimiter(rate.Inf, limiterBurstSize),
|
|
mu: sync.NewMutex(),
|
|
deviceReadLimiters: make(map[protocol.DeviceID]*rate.Limiter),
|
|
deviceWriteLimiters: make(map[protocol.DeviceID]*rate.Limiter),
|
|
}
|
|
|
|
cfg.Subscribe(l)
|
|
prev := config.Configuration{Options: config.OptionsConfiguration{MaxRecvKbps: -1, MaxSendKbps: -1}}
|
|
|
|
l.CommitConfiguration(prev, cfg.RawCopy())
|
|
return l
|
|
}
|
|
|
|
// This function sets limiters according to corresponding DeviceConfiguration
|
|
func (lim *limiter) setLimitsLocked(device config.DeviceConfiguration) bool {
|
|
readLimiter := lim.getReadLimiterLocked(device.DeviceID)
|
|
writeLimiter := lim.getWriteLimiterLocked(device.DeviceID)
|
|
|
|
// limiters for this device are created so we can store previous rates for logging
|
|
previousReadLimit := readLimiter.Limit()
|
|
previousWriteLimit := writeLimiter.Limit()
|
|
currentReadLimit := rate.Limit(device.MaxRecvKbps) * 1024
|
|
currentWriteLimit := rate.Limit(device.MaxSendKbps) * 1024
|
|
if device.MaxSendKbps <= 0 {
|
|
currentWriteLimit = rate.Inf
|
|
}
|
|
if device.MaxRecvKbps <= 0 {
|
|
currentReadLimit = rate.Inf
|
|
}
|
|
// Nothing about this device has changed. Start processing next device
|
|
if previousWriteLimit == currentWriteLimit && previousReadLimit == currentReadLimit {
|
|
return false
|
|
}
|
|
|
|
readLimiter.SetLimit(currentReadLimit)
|
|
writeLimiter.SetLimit(currentWriteLimit)
|
|
|
|
return true
|
|
}
|
|
|
|
// This function handles removing, adding and updating of device limiters.
|
|
func (lim *limiter) processDevicesConfigurationLocked(from, to config.Configuration) {
|
|
seen := make(map[protocol.DeviceID]struct{})
|
|
|
|
// Mark devices which should not be removed, create new limiters if needed and assign new limiter rate
|
|
for _, dev := range to.Devices {
|
|
if dev.DeviceID == lim.myID {
|
|
// This limiter was created for local device. Should skip this device
|
|
continue
|
|
}
|
|
seen[dev.DeviceID] = struct{}{}
|
|
|
|
if lim.setLimitsLocked(dev) {
|
|
readLimitStr := "is unlimited"
|
|
if dev.MaxRecvKbps > 0 {
|
|
readLimitStr = fmt.Sprintf("limit is %d KiB/s", dev.MaxRecvKbps)
|
|
}
|
|
writeLimitStr := "is unlimited"
|
|
if dev.MaxSendKbps > 0 {
|
|
writeLimitStr = fmt.Sprintf("limit is %d KiB/s", dev.MaxSendKbps)
|
|
}
|
|
|
|
l.Infof("Device %s send rate %s, receive rate %s", dev.DeviceID, writeLimitStr, readLimitStr)
|
|
}
|
|
}
|
|
|
|
// Delete remote devices which were removed in new configuration
|
|
for _, dev := range from.Devices {
|
|
if _, ok := seen[dev.DeviceID]; !ok {
|
|
l.Debugf("deviceID: %s should be removed", dev.DeviceID)
|
|
|
|
delete(lim.deviceWriteLimiters, dev.DeviceID)
|
|
delete(lim.deviceReadLimiters, dev.DeviceID)
|
|
}
|
|
}
|
|
}
|
|
|
|
func (lim *limiter) CommitConfiguration(from, to config.Configuration) bool {
|
|
// to ensure atomic update of configuration
|
|
lim.mu.Lock()
|
|
defer lim.mu.Unlock()
|
|
|
|
// Delete, add or update limiters for devices
|
|
lim.processDevicesConfigurationLocked(from, to)
|
|
|
|
if from.Options.MaxRecvKbps == to.Options.MaxRecvKbps &&
|
|
from.Options.MaxSendKbps == to.Options.MaxSendKbps &&
|
|
from.Options.LimitBandwidthInLan == to.Options.LimitBandwidthInLan {
|
|
return true
|
|
}
|
|
|
|
limited := false
|
|
sendLimitStr := "is unlimited"
|
|
recvLimitStr := "is unlimited"
|
|
|
|
// The rate variables are in KiB/s in the config (despite the camel casing
|
|
// of the name). We multiply by 1024 to get bytes/s.
|
|
if to.Options.MaxRecvKbps <= 0 {
|
|
lim.read.SetLimit(rate.Inf)
|
|
} else {
|
|
lim.read.SetLimit(1024 * rate.Limit(to.Options.MaxRecvKbps))
|
|
recvLimitStr = fmt.Sprintf("limit is %d KiB/s", to.Options.MaxRecvKbps)
|
|
limited = true
|
|
}
|
|
|
|
if to.Options.MaxSendKbps <= 0 {
|
|
lim.write.SetLimit(rate.Inf)
|
|
} else {
|
|
lim.write.SetLimit(1024 * rate.Limit(to.Options.MaxSendKbps))
|
|
sendLimitStr = fmt.Sprintf("limit is %d KiB/s", to.Options.MaxSendKbps)
|
|
limited = true
|
|
}
|
|
|
|
lim.limitsLAN.Store(to.Options.LimitBandwidthInLan)
|
|
|
|
l.Infof("Overall send rate %s, receive rate %s", sendLimitStr, recvLimitStr)
|
|
|
|
if limited {
|
|
if to.Options.LimitBandwidthInLan {
|
|
l.Infoln("Rate limits apply to LAN connections")
|
|
} else {
|
|
l.Infoln("Rate limits do not apply to LAN connections")
|
|
}
|
|
}
|
|
|
|
return true
|
|
}
|
|
|
|
func (*limiter) String() string {
|
|
// required by config.Committer interface
|
|
return "connections.limiter"
|
|
}
|
|
|
|
func (lim *limiter) getLimiters(remoteID protocol.DeviceID, rw io.ReadWriter, isLAN bool) (io.Reader, io.Writer) {
|
|
lim.mu.Lock()
|
|
wr := lim.newLimitedWriterLocked(remoteID, rw, isLAN)
|
|
rd := lim.newLimitedReaderLocked(remoteID, rw, isLAN)
|
|
lim.mu.Unlock()
|
|
return rd, wr
|
|
}
|
|
|
|
func (lim *limiter) newLimitedReaderLocked(remoteID protocol.DeviceID, r io.Reader, isLAN bool) io.Reader {
|
|
return &limitedReader{
|
|
reader: r,
|
|
waiterHolder: waiterHolder{
|
|
waiter: totalWaiter{lim.getReadLimiterLocked(remoteID), lim.read},
|
|
limitsLAN: &lim.limitsLAN,
|
|
isLAN: isLAN,
|
|
},
|
|
}
|
|
}
|
|
|
|
func (lim *limiter) newLimitedWriterLocked(remoteID protocol.DeviceID, w io.Writer, isLAN bool) io.Writer {
|
|
return &limitedWriter{
|
|
writer: w,
|
|
waiterHolder: waiterHolder{
|
|
waiter: totalWaiter{lim.getWriteLimiterLocked(remoteID), lim.write},
|
|
limitsLAN: &lim.limitsLAN,
|
|
isLAN: isLAN,
|
|
},
|
|
}
|
|
}
|
|
|
|
func (lim *limiter) getReadLimiterLocked(deviceID protocol.DeviceID) *rate.Limiter {
|
|
return getRateLimiter(lim.deviceReadLimiters, deviceID)
|
|
}
|
|
|
|
func (lim *limiter) getWriteLimiterLocked(deviceID protocol.DeviceID) *rate.Limiter {
|
|
return getRateLimiter(lim.deviceWriteLimiters, deviceID)
|
|
}
|
|
|
|
func getRateLimiter(m map[protocol.DeviceID]*rate.Limiter, deviceID protocol.DeviceID) *rate.Limiter {
|
|
limiter, ok := m[deviceID]
|
|
if !ok {
|
|
limiter = rate.NewLimiter(rate.Inf, limiterBurstSize)
|
|
m[deviceID] = limiter
|
|
}
|
|
return limiter
|
|
}
|
|
|
|
// limitedReader is a rate limited io.Reader
|
|
type limitedReader struct {
|
|
reader io.Reader
|
|
waiterHolder
|
|
}
|
|
|
|
func (r *limitedReader) Read(buf []byte) (int, error) {
|
|
n, err := r.reader.Read(buf)
|
|
if !r.unlimited() {
|
|
r.take(n)
|
|
}
|
|
return n, err
|
|
}
|
|
|
|
// limitedWriter is a rate limited io.Writer
|
|
type limitedWriter struct {
|
|
writer io.Writer
|
|
waiterHolder
|
|
}
|
|
|
|
func (w *limitedWriter) Write(buf []byte) (int, error) {
|
|
if w.unlimited() {
|
|
return w.writer.Write(buf)
|
|
}
|
|
|
|
// This does (potentially) multiple smaller writes in order to be less
|
|
// bursty with large writes and slow rates. At the same time we don't
|
|
// want to do hilarious amounts of tiny writes when the rate is high, so
|
|
// try to be a bit adaptable. We range from the minimum write size of 1
|
|
// KiB up to the limiter burst size, aiming for about a write every
|
|
// 10ms.
|
|
singleWriteSize := int(w.waiter.Limit() / 100) // 10ms worth of data
|
|
singleWriteSize = ((singleWriteSize / 1024) + 1) * 1024 // round up to the next kibibyte
|
|
if singleWriteSize > limiterBurstSize {
|
|
singleWriteSize = limiterBurstSize
|
|
}
|
|
|
|
written := 0
|
|
for written < len(buf) {
|
|
toWrite := singleWriteSize
|
|
if toWrite > len(buf)-written {
|
|
toWrite = len(buf) - written
|
|
}
|
|
w.take(toWrite)
|
|
n, err := w.writer.Write(buf[written : written+toWrite])
|
|
written += n
|
|
if err != nil {
|
|
return written, err
|
|
}
|
|
}
|
|
|
|
return written, nil
|
|
}
|
|
|
|
// waiterHolder is the common functionality around having and evaluating a
|
|
// waiter, valid for both writers and readers
|
|
type waiterHolder struct {
|
|
waiter waiter
|
|
limitsLAN *atomic.Bool
|
|
isLAN bool
|
|
}
|
|
|
|
// unlimited returns true if the waiter is not limiting the rate
|
|
func (w waiterHolder) unlimited() bool {
|
|
if w.isLAN && !w.limitsLAN.Load() {
|
|
return true
|
|
}
|
|
return w.waiter.Limit() == rate.Inf
|
|
}
|
|
|
|
// take is a utility function to consume tokens, because no call to WaitN
|
|
// must be larger than the limiter burst size or it will hang.
|
|
func (w waiterHolder) take(tokens int) {
|
|
// For writes we already split the buffer into smaller operations so those
|
|
// will always end up in the fast path below. For reads, however, we don't
|
|
// control the size of the incoming buffer and don't split the calls
|
|
// into the lower level reads so we might get a large amount of data and
|
|
// end up in the loop further down.
|
|
|
|
if tokens <= limiterBurstSize {
|
|
// Fast path. We won't get an error from WaitN as we don't pass a
|
|
// context with a deadline.
|
|
_ = w.waiter.WaitN(context.TODO(), tokens)
|
|
return
|
|
}
|
|
|
|
for tokens > 0 {
|
|
// Consume limiterBurstSize tokens at a time until we're done.
|
|
if tokens > limiterBurstSize {
|
|
_ = w.waiter.WaitN(context.TODO(), limiterBurstSize)
|
|
tokens -= limiterBurstSize
|
|
} else {
|
|
_ = w.waiter.WaitN(context.TODO(), tokens)
|
|
tokens = 0
|
|
}
|
|
}
|
|
}
|
|
|
|
// totalWaiter waits for all of the waiters
|
|
type totalWaiter []waiter
|
|
|
|
func (tw totalWaiter) WaitN(ctx context.Context, n int) error {
|
|
for _, w := range tw {
|
|
if err := w.WaitN(ctx, n); err != nil {
|
|
// error here is context cancellation, most likely, so we abort
|
|
// early
|
|
return err
|
|
}
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (tw totalWaiter) Limit() rate.Limit {
|
|
min := rate.Inf
|
|
for _, w := range tw {
|
|
if l := w.Limit(); l < min {
|
|
min = l
|
|
}
|
|
}
|
|
return min
|
|
}
|