mirror of
https://github.com/octoleo/syncthing.git
synced 2024-11-14 01:04:14 +00:00
e52be3d83e
So there were some issues here. The main problem was that model.Close(deviceID) was overloaded to mean "the connection was closed by the protocol layer" and "i want to close this connection". That meant it could get called twice - once *to* close the connection and then once more when the connection *was* closed. After this refactor there is instead a Closed(conn) method that is the callback. I didn't need to change the parameter in the end, but I think it's clearer what it means when it takes the connection that was closed instead of a device ID. To close a connection, the new close(deviceID) method is used instead, which only closes the underlying connection and leaves the cleanup to the Closed() callback. I also changed how we do connection switching. Instead of the connection service calling close and then adding the connection, it just adds the new connection. The model knows that it already has a connection and makes sure to close and clean out that one before adding the new connection. To make sure to sequence this properly I added a new map of channels that get created on connection add and closed by Closed(), so that AddConnection() can do the close and wait for the cleanup to happen before proceeding. GitHub-Pull-Request: https://github.com/syncthing/syncthing/pull/3490
189 lines
5.0 KiB
Go
189 lines
5.0 KiB
Go
// Copyright (C) 2016 The Protocol Authors.
|
|
|
|
package protocol
|
|
|
|
import (
|
|
"crypto/tls"
|
|
"encoding/binary"
|
|
"net"
|
|
"testing"
|
|
|
|
"github.com/syncthing/syncthing/lib/dialer"
|
|
)
|
|
|
|
func BenchmarkRequestsRawTCP(b *testing.B) {
|
|
// Benchmarks the rate at which we can serve requests over a single,
|
|
// unencrypted TCP channel over the loopback interface.
|
|
|
|
// Get a connected TCP pair
|
|
conn0, conn1, err := getTCPConnectionPair()
|
|
if err != nil {
|
|
b.Fatal(err)
|
|
}
|
|
|
|
defer conn0.Close()
|
|
defer conn1.Close()
|
|
|
|
// Bench it
|
|
benchmarkRequestsConnPair(b, conn0, conn1)
|
|
}
|
|
|
|
func BenchmarkRequestsTLSoTCP(b *testing.B) {
|
|
// Benchmarks the rate at which we can serve requests over a single,
|
|
// TLS encrypted TCP channel over the loopback interface.
|
|
|
|
// Load a certificate, skipping this benchmark if it doesn't exist
|
|
cert, err := tls.LoadX509KeyPair("../../test/h1/cert.pem", "../../test/h1/key.pem")
|
|
if err != nil {
|
|
b.Skip(err)
|
|
return
|
|
}
|
|
|
|
// Get a connected TCP pair
|
|
conn0, conn1, err := getTCPConnectionPair()
|
|
if err != nil {
|
|
b.Fatal(err)
|
|
}
|
|
|
|
/// TLSify them
|
|
conn0, conn1 = negotiateTLS(cert, conn0, conn1)
|
|
|
|
defer conn0.Close()
|
|
defer conn1.Close()
|
|
|
|
// Bench it
|
|
benchmarkRequestsConnPair(b, conn0, conn1)
|
|
}
|
|
|
|
func benchmarkRequestsConnPair(b *testing.B, conn0, conn1 net.Conn) {
|
|
// Start up Connections on them
|
|
c0 := NewConnection(LocalDeviceID, conn0, conn0, new(fakeModel), "c0", CompressMetadata)
|
|
c0.Start()
|
|
c1 := NewConnection(LocalDeviceID, conn1, conn1, new(fakeModel), "c1", CompressMetadata)
|
|
c1.Start()
|
|
|
|
// Satisfy the assertions in the protocol by sending an initial cluster config
|
|
c0.ClusterConfig(ClusterConfig{})
|
|
c1.ClusterConfig(ClusterConfig{})
|
|
|
|
// Report some useful stats and reset the timer for the actual test
|
|
b.ReportAllocs()
|
|
b.SetBytes(128 << 10)
|
|
b.ResetTimer()
|
|
|
|
// Request 128 KiB blocks, which will be satisfied by zero copy from the
|
|
// other side (we'll get back a full block of zeroes).
|
|
var buf []byte
|
|
var err error
|
|
for i := 0; i < b.N; i++ {
|
|
// Use c0 and c1 for each alternating request, so we get as much
|
|
// data flowing in both directions.
|
|
if i%2 == 0 {
|
|
buf, err = c0.Request("folder", "file", int64(i), 128<<10, nil, false)
|
|
} else {
|
|
buf, err = c1.Request("folder", "file", int64(i), 128<<10, nil, false)
|
|
}
|
|
|
|
if err != nil {
|
|
b.Fatal(err)
|
|
}
|
|
if len(buf) != 128<<10 {
|
|
b.Fatal("Incorrect returned buf length", len(buf), "!=", 128<<10)
|
|
}
|
|
|
|
// The fake model is supposed to tag the end of the buffer with the
|
|
// requested offset, so we can verify that we get back data for this
|
|
// block correctly.
|
|
if binary.BigEndian.Uint64(buf[128<<10-8:]) != uint64(i) {
|
|
b.Fatal("Bad data returned")
|
|
}
|
|
}
|
|
}
|
|
|
|
// returns the two endpoints of a TCP connection over lo0
|
|
func getTCPConnectionPair() (net.Conn, net.Conn, error) {
|
|
lst, err := net.Listen("tcp", "127.0.0.1:0")
|
|
if err != nil {
|
|
return nil, nil, err
|
|
}
|
|
|
|
// We run the Accept in the background since it's blocking, and we use
|
|
// the channel to make the race thingies happy about writing vs reading
|
|
// conn0 and err0.
|
|
var conn0 net.Conn
|
|
var err0 error
|
|
done := make(chan struct{})
|
|
go func() {
|
|
conn0, err0 = lst.Accept()
|
|
close(done)
|
|
}()
|
|
|
|
// Dial the connection
|
|
conn1, err := net.Dial("tcp", lst.Addr().String())
|
|
if err != nil {
|
|
return nil, nil, err
|
|
}
|
|
|
|
// Check any error from accept
|
|
<-done
|
|
if err0 != nil {
|
|
return nil, nil, err0
|
|
}
|
|
|
|
// Set the buffer sizes etc as usual
|
|
dialer.SetTCPOptions(conn0)
|
|
dialer.SetTCPOptions(conn1)
|
|
|
|
return conn0, conn1, nil
|
|
}
|
|
|
|
func negotiateTLS(cert tls.Certificate, conn0, conn1 net.Conn) (net.Conn, net.Conn) {
|
|
cfg := &tls.Config{
|
|
Certificates: []tls.Certificate{cert},
|
|
NextProtos: []string{"bep/1.0"},
|
|
ClientAuth: tls.RequestClientCert,
|
|
SessionTicketsDisabled: true,
|
|
InsecureSkipVerify: true,
|
|
MinVersion: tls.VersionTLS12,
|
|
CipherSuites: []uint16{
|
|
tls.TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,
|
|
tls.TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,
|
|
tls.TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA,
|
|
tls.TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA,
|
|
tls.TLS_ECDHE_RSA_WITH_AES_256_CBC_SHA,
|
|
tls.TLS_ECDHE_ECDSA_WITH_AES_256_CBC_SHA,
|
|
},
|
|
}
|
|
|
|
tlsc0 := tls.Server(conn0, cfg)
|
|
tlsc1 := tls.Client(conn1, cfg)
|
|
return tlsc0, tlsc1
|
|
}
|
|
|
|
// The fake model does nothing much
|
|
|
|
type fakeModel struct{}
|
|
|
|
func (m *fakeModel) Index(deviceID DeviceID, folder string, files []FileInfo) {
|
|
}
|
|
|
|
func (m *fakeModel) IndexUpdate(deviceID DeviceID, folder string, files []FileInfo) {
|
|
}
|
|
|
|
func (m *fakeModel) Request(deviceID DeviceID, folder string, name string, offset int64, hash []byte, fromTemporary bool, buf []byte) error {
|
|
// We write the offset to the end of the buffer, so the receiver
|
|
// can verify that it did in fact get some data back over the
|
|
// connection.
|
|
binary.BigEndian.PutUint64(buf[len(buf)-8:], uint64(offset))
|
|
return nil
|
|
}
|
|
|
|
func (m *fakeModel) ClusterConfig(deviceID DeviceID, config ClusterConfig) {
|
|
}
|
|
|
|
func (m *fakeModel) Closed(conn Connection, err error) {
|
|
}
|
|
|
|
func (m *fakeModel) DownloadProgress(deviceID DeviceID, folder string, updates []FileDownloadProgressUpdate) {
|
|
}
|