mirror of https://github.com/cbeuw/Cloak
Change the frame distribution strategy so that each stream is assigned a fixed underlying TCP connection to send its frames to reduce framesorter ooo count
This commit is contained in:
parent
fac381dbca
commit
01036086c7
|
|
@ -79,7 +79,11 @@ func (sesh *Session) OpenStream() (*Stream, error) {
|
||||||
}
|
}
|
||||||
id := atomic.AddUint32(&sesh.nextStreamID, 1) - 1
|
id := atomic.AddUint32(&sesh.nextStreamID, 1) - 1
|
||||||
// Because atomic.AddUint32 returns the value after incrementation
|
// Because atomic.AddUint32 returns the value after incrementation
|
||||||
stream := makeStream(id, sesh)
|
connId, err := sesh.sb.assignRandomConn()
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
stream := makeStream(sesh, id, connId)
|
||||||
sesh.streamsM.Lock()
|
sesh.streamsM.Lock()
|
||||||
sesh.streams[id] = stream
|
sesh.streams[id] = stream
|
||||||
sesh.streamsM.Unlock()
|
sesh.streamsM.Unlock()
|
||||||
|
|
@ -123,7 +127,13 @@ func (sesh *Session) recvDataFromRemote(data []byte) {
|
||||||
// If the stream has been closed and the current frame is a closing frame, we do noop
|
// If the stream has been closed and the current frame is a closing frame, we do noop
|
||||||
return
|
return
|
||||||
} else {
|
} else {
|
||||||
stream = makeStream(frame.StreamID, sesh)
|
// it may be tempting to use the connId from which the frame was received. However it doesn't make
|
||||||
|
// any difference because we only care to send the data from the same stream through the same
|
||||||
|
// TCP connection. The remote may use a different connection to send the same stream than the one the client
|
||||||
|
// use to send.
|
||||||
|
connId, _ := sesh.sb.assignRandomConn()
|
||||||
|
// we ignore the error here. If the switchboard is broken, it will be reflected upon stream.Write
|
||||||
|
stream = makeStream(sesh, frame.StreamID, connId)
|
||||||
sesh.streams[frame.StreamID] = stream
|
sesh.streams[frame.StreamID] = stream
|
||||||
sesh.acceptCh <- stream
|
sesh.acceptCh <- stream
|
||||||
stream.writeFrame(frame)
|
stream.writeFrame(frame)
|
||||||
|
|
|
||||||
|
|
@ -28,27 +28,34 @@ type Stream struct {
|
||||||
|
|
||||||
writingM sync.RWMutex
|
writingM sync.RWMutex
|
||||||
|
|
||||||
// close(die) is used to notify different goroutines that this stream is closing
|
|
||||||
closed uint32
|
closed uint32
|
||||||
|
|
||||||
obfsBuf []byte
|
obfsBuf []byte
|
||||||
|
|
||||||
|
// we assign each stream a fixed underlying TCP connection to utilise order guarantee provided by TCP itself
|
||||||
|
// so that frameSorter should have few to none ooo frames to deal with
|
||||||
|
// overall the streams in a session should be uniformly distributed across all connections
|
||||||
|
assignedConnId uint32
|
||||||
}
|
}
|
||||||
|
|
||||||
func makeStream(id uint32, sesh *Session) *Stream {
|
func makeStream(sesh *Session, id uint32, assignedConnId uint32) *Stream {
|
||||||
buf := NewBufferedPipe()
|
buf := NewBufferedPipe()
|
||||||
|
|
||||||
stream := &Stream{
|
stream := &Stream{
|
||||||
id: id,
|
id: id,
|
||||||
session: sesh,
|
session: sesh,
|
||||||
sortedBuf: buf,
|
sortedBuf: buf,
|
||||||
obfsBuf: make([]byte, 17000),
|
obfsBuf: make([]byte, 17000),
|
||||||
sorter: NewFrameSorter(buf),
|
sorter: NewFrameSorter(buf),
|
||||||
|
assignedConnId: assignedConnId,
|
||||||
}
|
}
|
||||||
|
|
||||||
log.Tracef("stream %v opened", id)
|
log.Tracef("stream %v opened", id)
|
||||||
return stream
|
return stream
|
||||||
}
|
}
|
||||||
|
|
||||||
|
//func (s *Stream) reassignConnId(connId uint32) { atomic.StoreUint32(&s.assignedConnId,connId)}
|
||||||
|
|
||||||
func (s *Stream) isClosed() bool { return atomic.LoadUint32(&s.closed) == 1 }
|
func (s *Stream) isClosed() bool { return atomic.LoadUint32(&s.closed) == 1 }
|
||||||
|
|
||||||
func (s *Stream) writeFrame(frame *Frame) { s.sorter.writeNewFrame(frame) }
|
func (s *Stream) writeFrame(frame *Frame) { s.sorter.writeNewFrame(frame) }
|
||||||
|
|
@ -96,7 +103,7 @@ func (s *Stream) Write(in []byte) (n int, err error) {
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return i, err
|
return i, err
|
||||||
}
|
}
|
||||||
n, err = s.session.sb.Write(s.obfsBuf[:i])
|
n, err = s.session.sb.send(s.obfsBuf[:i], &s.assignedConnId)
|
||||||
return
|
return
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
@ -139,7 +146,7 @@ func (s *Stream) Close() error {
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
_, err = s.session.sb.Write(s.obfsBuf[:i])
|
_, err = s.session.sb.send(s.obfsBuf[:i], &s.assignedConnId)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
|
||||||
|
|
@ -99,6 +99,7 @@ func TestStream_Read(t *testing.T) {
|
||||||
i, _ := sesh.Obfs(f, obfsBuf)
|
i, _ := sesh.Obfs(f, obfsBuf)
|
||||||
streamID++
|
streamID++
|
||||||
ch <- obfsBuf[:i]
|
ch <- obfsBuf[:i]
|
||||||
|
time.Sleep(100 * time.Microsecond)
|
||||||
stream, err := sesh.Accept()
|
stream, err := sesh.Accept()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
t.Error("failed to accept stream", err)
|
t.Error("failed to accept stream", err)
|
||||||
|
|
@ -120,6 +121,7 @@ func TestStream_Read(t *testing.T) {
|
||||||
i, _ := sesh.Obfs(f, obfsBuf)
|
i, _ := sesh.Obfs(f, obfsBuf)
|
||||||
streamID++
|
streamID++
|
||||||
ch <- obfsBuf[:i]
|
ch <- obfsBuf[:i]
|
||||||
|
time.Sleep(100 * time.Microsecond)
|
||||||
stream, _ := sesh.Accept()
|
stream, _ := sesh.Accept()
|
||||||
i, err := stream.Read(nil)
|
i, err := stream.Read(nil)
|
||||||
if i != 0 || err != nil {
|
if i != 0 || err != nil {
|
||||||
|
|
@ -140,6 +142,7 @@ func TestStream_Read(t *testing.T) {
|
||||||
i, _ := sesh.Obfs(f, obfsBuf)
|
i, _ := sesh.Obfs(f, obfsBuf)
|
||||||
streamID++
|
streamID++
|
||||||
ch <- obfsBuf[:i]
|
ch <- obfsBuf[:i]
|
||||||
|
time.Sleep(100 * time.Microsecond)
|
||||||
stream, _ := sesh.Accept()
|
stream, _ := sesh.Accept()
|
||||||
stream.Close()
|
stream.Close()
|
||||||
i, err := stream.Read(buf)
|
i, err := stream.Read(buf)
|
||||||
|
|
@ -164,6 +167,7 @@ func TestStream_Read(t *testing.T) {
|
||||||
i, _ := sesh.Obfs(f, obfsBuf)
|
i, _ := sesh.Obfs(f, obfsBuf)
|
||||||
streamID++
|
streamID++
|
||||||
ch <- obfsBuf[:i]
|
ch <- obfsBuf[:i]
|
||||||
|
time.Sleep(100 * time.Microsecond)
|
||||||
stream, _ := sesh.Accept()
|
stream, _ := sesh.Accept()
|
||||||
sesh.Close()
|
sesh.Close()
|
||||||
i, err := stream.Read(buf)
|
i, err := stream.Read(buf)
|
||||||
|
|
|
||||||
|
|
@ -3,6 +3,7 @@ package multiplex
|
||||||
import (
|
import (
|
||||||
"errors"
|
"errors"
|
||||||
log "github.com/sirupsen/logrus"
|
log "github.com/sirupsen/logrus"
|
||||||
|
"math/rand"
|
||||||
"net"
|
"net"
|
||||||
"sync"
|
"sync"
|
||||||
"sync/atomic"
|
"sync/atomic"
|
||||||
|
|
@ -14,38 +15,20 @@ type switchboard struct {
|
||||||
|
|
||||||
*Valve
|
*Valve
|
||||||
|
|
||||||
// optimum is the connEnclave with the smallest sendQueue
|
connsM sync.RWMutex
|
||||||
optimum atomic.Value // *connEnclave
|
conns map[uint32]net.Conn
|
||||||
cesM sync.RWMutex
|
nextConnId uint32
|
||||||
ces []*connEnclave
|
|
||||||
|
|
||||||
broken uint32
|
broken uint32
|
||||||
}
|
}
|
||||||
|
|
||||||
func (sb *switchboard) getOptimum() *connEnclave {
|
|
||||||
if i := sb.optimum.Load(); i == nil {
|
|
||||||
return nil
|
|
||||||
} else {
|
|
||||||
return i.(*connEnclave)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// Some data comes from a Stream to be sent through one of the many
|
|
||||||
// remoteConn, but which remoteConn should we use to send the data?
|
|
||||||
//
|
|
||||||
// In this case, we pick the remoteConn that has about the smallest sendQueue.
|
|
||||||
type connEnclave struct {
|
|
||||||
remoteConn net.Conn
|
|
||||||
sendQueue uint32
|
|
||||||
}
|
|
||||||
|
|
||||||
func makeSwitchboard(sesh *Session, valve *Valve) *switchboard {
|
func makeSwitchboard(sesh *Session, valve *Valve) *switchboard {
|
||||||
// rates are uint64 because in the usermanager we want the bandwidth to be atomically
|
// rates are uint64 because in the usermanager we want the bandwidth to be atomically
|
||||||
// operated (so that the bandwidth can change on the fly).
|
// operated (so that the bandwidth can change on the fly).
|
||||||
sb := &switchboard{
|
sb := &switchboard{
|
||||||
session: sesh,
|
session: sesh,
|
||||||
Valve: valve,
|
Valve: valve,
|
||||||
ces: []*connEnclave{},
|
conns: make(map[uint32]net.Conn),
|
||||||
}
|
}
|
||||||
return sb
|
return sb
|
||||||
}
|
}
|
||||||
|
|
@ -53,65 +36,19 @@ func makeSwitchboard(sesh *Session, valve *Valve) *switchboard {
|
||||||
var errNilOptimum = errors.New("The optimal connection is nil")
|
var errNilOptimum = errors.New("The optimal connection is nil")
|
||||||
var errBrokenSwitchboard = errors.New("the switchboard is broken")
|
var errBrokenSwitchboard = errors.New("the switchboard is broken")
|
||||||
|
|
||||||
func (sb *switchboard) Write(data []byte) (int, error) {
|
|
||||||
if atomic.LoadUint32(&sb.broken) == 1 {
|
|
||||||
return 0, errBrokenSwitchboard
|
|
||||||
}
|
|
||||||
ce := sb.getOptimum()
|
|
||||||
if ce == nil {
|
|
||||||
return 0, errNilOptimum
|
|
||||||
}
|
|
||||||
atomic.AddUint32(&ce.sendQueue, uint32(len(data)))
|
|
||||||
go sb.updateOptimum()
|
|
||||||
n, err := ce.remoteConn.Write(data)
|
|
||||||
if err != nil {
|
|
||||||
return n, err
|
|
||||||
}
|
|
||||||
sb.txWait(n)
|
|
||||||
sb.Valve.AddTx(int64(n))
|
|
||||||
atomic.AddUint32(&ce.sendQueue, ^uint32(n-1))
|
|
||||||
go sb.updateOptimum()
|
|
||||||
return n, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (sb *switchboard) updateOptimum() {
|
|
||||||
currentOpti := sb.getOptimum()
|
|
||||||
currentOptiQ := atomic.LoadUint32(¤tOpti.sendQueue)
|
|
||||||
sb.cesM.RLock()
|
|
||||||
for _, ce := range sb.ces {
|
|
||||||
ceQ := atomic.LoadUint32(&ce.sendQueue)
|
|
||||||
if ceQ < currentOptiQ {
|
|
||||||
currentOpti = ce
|
|
||||||
currentOptiQ = ceQ
|
|
||||||
}
|
|
||||||
}
|
|
||||||
sb.cesM.RUnlock()
|
|
||||||
sb.optimum.Store(currentOpti)
|
|
||||||
}
|
|
||||||
|
|
||||||
func (sb *switchboard) addConn(conn net.Conn) {
|
func (sb *switchboard) addConn(conn net.Conn) {
|
||||||
var sendQueue uint32
|
connId := atomic.AddUint32(&sb.nextConnId, 1) - 1
|
||||||
newCe := &connEnclave{
|
sb.connsM.Lock()
|
||||||
remoteConn: conn,
|
sb.conns[connId] = conn
|
||||||
sendQueue: sendQueue,
|
sb.connsM.Unlock()
|
||||||
}
|
go sb.deplex(connId, conn)
|
||||||
sb.cesM.Lock()
|
|
||||||
sb.ces = append(sb.ces, newCe)
|
|
||||||
sb.cesM.Unlock()
|
|
||||||
sb.optimum.Store(newCe)
|
|
||||||
go sb.deplex(newCe)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func (sb *switchboard) removeConn(closing *connEnclave) {
|
func (sb *switchboard) removeConn(connId uint32) {
|
||||||
sb.cesM.Lock()
|
sb.connsM.Lock()
|
||||||
for i, ce := range sb.ces {
|
delete(sb.conns, connId)
|
||||||
if closing == ce {
|
remaining := len(sb.conns)
|
||||||
sb.ces = append(sb.ces[:i], sb.ces[i+1:]...)
|
sb.connsM.Unlock()
|
||||||
break
|
|
||||||
}
|
|
||||||
}
|
|
||||||
remaining := len(sb.ces)
|
|
||||||
sb.cesM.Unlock()
|
|
||||||
if remaining == 0 {
|
if remaining == 0 {
|
||||||
atomic.StoreUint32(&sb.broken, 1)
|
atomic.StoreUint32(&sb.broken, 1)
|
||||||
sb.session.SetTerminalMsg("no underlying connection left")
|
sb.session.SetTerminalMsg("no underlying connection left")
|
||||||
|
|
@ -119,29 +56,67 @@ func (sb *switchboard) removeConn(closing *connEnclave) {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// a pointer to connId is passed here so that the switchboard can reassign it
|
||||||
|
func (sb *switchboard) send(data []byte, connId *uint32) (int, error) {
|
||||||
|
var conn net.Conn
|
||||||
|
sb.connsM.RLock()
|
||||||
|
conn, ok := sb.conns[*connId]
|
||||||
|
sb.connsM.RUnlock()
|
||||||
|
if ok {
|
||||||
|
return conn.Write(data)
|
||||||
|
} else {
|
||||||
|
// do not call assignRandomConn() here.
|
||||||
|
// we'll have to do connsM.RLock() after we get a new connId from assignRandomConn, in order to
|
||||||
|
// get the new conn through conns[newConnId]
|
||||||
|
// however between connsM.RUnlock() in assignRandomConn and our call to connsM.RLock(), things may happen.
|
||||||
|
// in particular if newConnId is removed between the RUnlock and RLock, conns[newConnId] will return
|
||||||
|
// a nil pointer. To prevent this we must get newConnId and the reference to conn itself in one single mutex
|
||||||
|
// protection
|
||||||
|
if atomic.LoadUint32(&sb.broken) == 1 {
|
||||||
|
return 0, errBrokenSwitchboard
|
||||||
|
}
|
||||||
|
sb.connsM.RLock()
|
||||||
|
newConnId := rand.Intn(len(sb.conns))
|
||||||
|
conn = sb.conns[uint32(newConnId)]
|
||||||
|
sb.connsM.RUnlock()
|
||||||
|
return conn.Write(data)
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
func (sb *switchboard) assignRandomConn() (uint32, error) {
|
||||||
|
if atomic.LoadUint32(&sb.broken) == 1 {
|
||||||
|
return 0, errBrokenSwitchboard
|
||||||
|
}
|
||||||
|
sb.connsM.RLock()
|
||||||
|
connId := rand.Intn(len(sb.conns))
|
||||||
|
sb.connsM.RUnlock()
|
||||||
|
return uint32(connId), nil
|
||||||
|
}
|
||||||
|
|
||||||
// actively triggered by session.Close()
|
// actively triggered by session.Close()
|
||||||
func (sb *switchboard) closeAll() {
|
func (sb *switchboard) closeAll() {
|
||||||
if atomic.SwapUint32(&sb.broken, 1) == 1 {
|
if atomic.SwapUint32(&sb.broken, 1) == 1 {
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
sb.cesM.RLock()
|
sb.connsM.RLock()
|
||||||
for _, ce := range sb.ces {
|
for _, conn := range sb.conns {
|
||||||
ce.remoteConn.Close()
|
conn.Close()
|
||||||
}
|
}
|
||||||
sb.cesM.RUnlock()
|
sb.connsM.RUnlock()
|
||||||
}
|
}
|
||||||
|
|
||||||
// deplex function costantly reads from a TCP connection
|
// deplex function costantly reads from a TCP connection
|
||||||
func (sb *switchboard) deplex(ce *connEnclave) {
|
func (sb *switchboard) deplex(connId uint32, conn net.Conn) {
|
||||||
buf := make([]byte, 20480)
|
buf := make([]byte, 20480)
|
||||||
for {
|
for {
|
||||||
n, err := sb.session.unitRead(ce.remoteConn, buf)
|
n, err := sb.session.unitRead(conn, buf)
|
||||||
sb.rxWait(n)
|
sb.rxWait(n)
|
||||||
sb.Valve.AddRx(int64(n))
|
sb.Valve.AddRx(int64(n))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
log.Tracef("a connection for session %v has closed: %v", sb.session.id, err)
|
log.Tracef("a connection for session %v has closed: %v", sb.session.id, err)
|
||||||
go ce.remoteConn.Close()
|
go conn.Close()
|
||||||
sb.removeConn(ce)
|
sb.removeConn(connId)
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
||||||
Loading…
Reference in New Issue