perf: websocket grid connectivity for all internode communication (#18461)
This PR adds a WebSocket grid feature that allows servers to communicate via
a single two-way connection.
There are two request types:
* Single requests, which are `[]byte => ([]byte, error)`. This is for efficient small
roundtrips with small payloads.
* Streaming requests which are `[]byte, chan []byte => chan []byte (and error)`,
which allows for different combinations of full two-way streams with an initial payload.
Only a single stream is created between two machines - and there is, as such, no
server/client relation since both sides can initiate and handle requests. Which server
initiates the request is decided deterministically on the server names.
Requests are made through a mux client and server, which handles message
passing, congestion, cancelation, timeouts, etc.
If a connection is lost, all requests are canceled, and the calling server will try
to reconnect. Registered handlers can operate directly on byte
slices or use a higher-level generics abstraction.
There is no versioning of handlers/clients, and incompatible changes should
be handled by adding new handlers.
The request path can be changed to a new one for any protocol changes.
First, all servers create a "Manager." The manager must know its address
as well as all remote addresses. This will manage all connections.
To get a connection to any remote, ask the manager to provide it given
the remote address using.
```
func (m *Manager) Connection(host string) *Connection
```
All serverside handlers must also be registered on the manager. This will
make sure that all incoming requests are served. The number of in-flight
requests and responses must also be given for streaming requests.
The "Connection" returned manages the mux-clients. Requests issued
to the connection will be sent to the remote.
* `func (c *Connection) Request(ctx context.Context, h HandlerID, req []byte) ([]byte, error)`
performs a single request and returns the result. Any deadline provided on the request is
forwarded to the server, and canceling the context will make the function return at once.
* `func (c *Connection) NewStream(ctx context.Context, h HandlerID, payload []byte) (st *Stream, err error)`
will initiate a remote call and send the initial payload.
```Go
// A Stream is a two-way stream.
// All responses *must* be read by the caller.
// If the call is canceled through the context,
//The appropriate error will be returned.
type Stream struct {
// Responses from the remote server.
// Channel will be closed after an error or when the remote closes.
// All responses *must* be read by the caller until either an error is returned or the channel is closed.
// Canceling the context will cause the context cancellation error to be returned.
Responses <-chan Response
// Requests sent to the server.
// If the handler is defined with 0 incoming capacity this will be nil.
// Channel *must* be closed to signal the end of the stream.
// If the request context is canceled, the stream will no longer process requests.
Requests chan<- []byte
}
type Response struct {
Msg []byte
Err error
}
```
There are generic versions of the server/client handlers that allow the use of type
safe implementations for data types that support msgpack marshal/unmarshal.
2023-11-20 20:09:35 -05:00
|
|
|
// Copyright (c) 2015-2023 MinIO, Inc.
|
|
|
|
//
|
|
|
|
// This file is part of MinIO Object Storage stack
|
|
|
|
//
|
|
|
|
// This program is free software: you can redistribute it and/or modify
|
|
|
|
// it under the terms of the GNU Affero General Public License as published by
|
|
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
|
|
// (at your option) any later version.
|
|
|
|
//
|
|
|
|
// This program is distributed in the hope that it will be useful
|
|
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
|
|
// GNU Affero General Public License for more details.
|
|
|
|
//
|
|
|
|
// You should have received a copy of the GNU Affero General Public License
|
|
|
|
// along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
|
|
|
|
package grid
|
|
|
|
|
|
|
|
import (
|
|
|
|
"bytes"
|
|
|
|
"context"
|
|
|
|
"errors"
|
|
|
|
"fmt"
|
|
|
|
"os"
|
|
|
|
"runtime"
|
|
|
|
"strconv"
|
|
|
|
"strings"
|
|
|
|
"testing"
|
|
|
|
"time"
|
|
|
|
|
|
|
|
"github.com/minio/minio/internal/logger/target/testlogger"
|
|
|
|
)
|
|
|
|
|
|
|
|
func TestSingleRoundtrip(t *testing.T) {
|
|
|
|
defer testlogger.T.SetLogTB(t)()
|
|
|
|
errFatal := func(err error) {
|
|
|
|
t.Helper()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatal(err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
grid, err := SetupTestGrid(2)
|
|
|
|
errFatal(err)
|
|
|
|
remoteHost := grid.Hosts[1]
|
|
|
|
local := grid.Managers[0]
|
|
|
|
|
|
|
|
// 1: Echo
|
|
|
|
errFatal(local.RegisterSingleHandler(handlerTest, func(payload []byte) ([]byte, *RemoteErr) {
|
|
|
|
t.Log("1: server payload: ", len(payload), "bytes.")
|
|
|
|
return append([]byte{}, payload...), nil
|
|
|
|
}))
|
|
|
|
// 2: Return as error
|
|
|
|
errFatal(local.RegisterSingleHandler(handlerTest2, func(payload []byte) ([]byte, *RemoteErr) {
|
|
|
|
t.Log("2: server payload: ", len(payload), "bytes.")
|
|
|
|
err := RemoteErr(payload)
|
|
|
|
return nil, &err
|
|
|
|
}))
|
|
|
|
|
|
|
|
remote := grid.Managers[1]
|
|
|
|
|
|
|
|
// 1: Echo
|
|
|
|
errFatal(remote.RegisterSingleHandler(handlerTest, func(payload []byte) ([]byte, *RemoteErr) {
|
|
|
|
t.Log("1: server payload: ", len(payload), "bytes.")
|
|
|
|
return append([]byte{}, payload...), nil
|
|
|
|
}))
|
|
|
|
// 2: Return as error
|
|
|
|
errFatal(remote.RegisterSingleHandler(handlerTest2, func(payload []byte) ([]byte, *RemoteErr) {
|
|
|
|
t.Log("2: server payload: ", len(payload), "bytes.")
|
|
|
|
err := RemoteErr(payload)
|
|
|
|
return nil, &err
|
|
|
|
}))
|
|
|
|
|
|
|
|
// local to remote
|
|
|
|
remoteConn := local.Connection(remoteHost)
|
|
|
|
remoteConn.WaitForConnect(context.Background())
|
|
|
|
defer testlogger.T.SetErrorTB(t)()
|
|
|
|
|
|
|
|
t.Run("localToRemote", func(t *testing.T) {
|
|
|
|
const testPayload = "Hello Grid World!"
|
|
|
|
|
|
|
|
start := time.Now()
|
|
|
|
resp, err := remoteConn.Request(context.Background(), handlerTest, []byte(testPayload))
|
|
|
|
errFatal(err)
|
|
|
|
if string(resp) != testPayload {
|
|
|
|
t.Errorf("want %q, got %q", testPayload, string(resp))
|
|
|
|
}
|
|
|
|
t.Log("Roundtrip:", time.Since(start))
|
|
|
|
})
|
|
|
|
|
|
|
|
t.Run("localToRemoteErr", func(t *testing.T) {
|
|
|
|
const testPayload = "Hello Grid World!"
|
|
|
|
start := time.Now()
|
|
|
|
resp, err := remoteConn.Request(context.Background(), handlerTest2, []byte(testPayload))
|
|
|
|
t.Log("Roundtrip:", time.Since(start))
|
|
|
|
if len(resp) != 0 {
|
|
|
|
t.Errorf("want nil, got %q", string(resp))
|
|
|
|
}
|
|
|
|
if err != RemoteErr(testPayload) {
|
|
|
|
t.Errorf("want error %v(%T), got %v(%T)", RemoteErr(testPayload), RemoteErr(testPayload), err, err)
|
|
|
|
}
|
|
|
|
t.Log("Roundtrip:", time.Since(start))
|
|
|
|
})
|
|
|
|
|
|
|
|
t.Run("localToRemoteHuge", func(t *testing.T) {
|
|
|
|
testPayload := bytes.Repeat([]byte("?"), 1<<20)
|
|
|
|
|
|
|
|
start := time.Now()
|
|
|
|
resp, err := remoteConn.Request(context.Background(), handlerTest, testPayload)
|
|
|
|
errFatal(err)
|
|
|
|
if string(resp) != string(testPayload) {
|
|
|
|
t.Errorf("want %q, got %q", testPayload, string(resp))
|
|
|
|
}
|
|
|
|
t.Log("Roundtrip:", time.Since(start))
|
|
|
|
})
|
|
|
|
|
|
|
|
t.Run("localToRemoteErrHuge", func(t *testing.T) {
|
|
|
|
testPayload := bytes.Repeat([]byte("!"), 1<<10)
|
|
|
|
|
|
|
|
start := time.Now()
|
|
|
|
resp, err := remoteConn.Request(context.Background(), handlerTest2, testPayload)
|
|
|
|
if len(resp) != 0 {
|
|
|
|
t.Errorf("want nil, got %q", string(resp))
|
|
|
|
}
|
|
|
|
if err != RemoteErr(testPayload) {
|
|
|
|
t.Errorf("want error %v(%T), got %v(%T)", RemoteErr(testPayload), RemoteErr(testPayload), err, err)
|
|
|
|
}
|
|
|
|
t.Log("Roundtrip:", time.Since(start))
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
2023-12-01 03:18:04 -05:00
|
|
|
func TestSingleRoundtripNotReady(t *testing.T) {
|
|
|
|
defer testlogger.T.SetLogTB(t)()
|
|
|
|
errFatal := func(t testing.TB, err error) {
|
|
|
|
t.Helper()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatal(err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
grid, err := SetupTestGrid(2)
|
|
|
|
errFatal(t, err)
|
|
|
|
remoteHost := grid.Hosts[1]
|
|
|
|
local := grid.Managers[0]
|
|
|
|
|
|
|
|
// 1: Echo
|
|
|
|
errFatal(t, local.RegisterSingleHandler(handlerTest, func(payload []byte) ([]byte, *RemoteErr) {
|
|
|
|
t.Log("1: server payload: ", len(payload), "bytes.")
|
|
|
|
return append([]byte{}, payload...), nil
|
|
|
|
}))
|
|
|
|
// 2: Return as error
|
|
|
|
errFatal(t, local.RegisterSingleHandler(handlerTest2, func(payload []byte) ([]byte, *RemoteErr) {
|
|
|
|
t.Log("2: server payload: ", len(payload), "bytes.")
|
|
|
|
err := RemoteErr(payload)
|
|
|
|
return nil, &err
|
|
|
|
}))
|
|
|
|
|
|
|
|
// Do not register remote handlers
|
|
|
|
|
|
|
|
// local to remote
|
|
|
|
remoteConn := local.Connection(remoteHost)
|
|
|
|
remoteConn.WaitForConnect(context.Background())
|
|
|
|
defer testlogger.T.SetErrorTB(t)()
|
|
|
|
|
|
|
|
t.Run("localToRemote", func(t *testing.T) {
|
|
|
|
const testPayload = "Hello Grid World!"
|
|
|
|
// Single requests should have remote errors.
|
|
|
|
_, err := remoteConn.Request(context.Background(), handlerTest, []byte(testPayload))
|
|
|
|
if v, ok := err.(*RemoteErr); !ok || v.Error() != "Invalid Handler for type" {
|
|
|
|
t.Fatalf("Unexpected error: %v, %T", err, err)
|
|
|
|
}
|
|
|
|
// Streams should not be able to set up until registered.
|
|
|
|
// Thus, the error is a local error.
|
|
|
|
_, err = remoteConn.NewStream(context.Background(), handlerTest, []byte(testPayload))
|
|
|
|
if !errors.Is(err, ErrUnknownHandler) {
|
|
|
|
t.Fatalf("Unexpected error: %v, %T", err, err)
|
|
|
|
}
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
perf: websocket grid connectivity for all internode communication (#18461)
This PR adds a WebSocket grid feature that allows servers to communicate via
a single two-way connection.
There are two request types:
* Single requests, which are `[]byte => ([]byte, error)`. This is for efficient small
roundtrips with small payloads.
* Streaming requests which are `[]byte, chan []byte => chan []byte (and error)`,
which allows for different combinations of full two-way streams with an initial payload.
Only a single stream is created between two machines - and there is, as such, no
server/client relation since both sides can initiate and handle requests. Which server
initiates the request is decided deterministically on the server names.
Requests are made through a mux client and server, which handles message
passing, congestion, cancelation, timeouts, etc.
If a connection is lost, all requests are canceled, and the calling server will try
to reconnect. Registered handlers can operate directly on byte
slices or use a higher-level generics abstraction.
There is no versioning of handlers/clients, and incompatible changes should
be handled by adding new handlers.
The request path can be changed to a new one for any protocol changes.
First, all servers create a "Manager." The manager must know its address
as well as all remote addresses. This will manage all connections.
To get a connection to any remote, ask the manager to provide it given
the remote address using.
```
func (m *Manager) Connection(host string) *Connection
```
All serverside handlers must also be registered on the manager. This will
make sure that all incoming requests are served. The number of in-flight
requests and responses must also be given for streaming requests.
The "Connection" returned manages the mux-clients. Requests issued
to the connection will be sent to the remote.
* `func (c *Connection) Request(ctx context.Context, h HandlerID, req []byte) ([]byte, error)`
performs a single request and returns the result. Any deadline provided on the request is
forwarded to the server, and canceling the context will make the function return at once.
* `func (c *Connection) NewStream(ctx context.Context, h HandlerID, payload []byte) (st *Stream, err error)`
will initiate a remote call and send the initial payload.
```Go
// A Stream is a two-way stream.
// All responses *must* be read by the caller.
// If the call is canceled through the context,
//The appropriate error will be returned.
type Stream struct {
// Responses from the remote server.
// Channel will be closed after an error or when the remote closes.
// All responses *must* be read by the caller until either an error is returned or the channel is closed.
// Canceling the context will cause the context cancellation error to be returned.
Responses <-chan Response
// Requests sent to the server.
// If the handler is defined with 0 incoming capacity this will be nil.
// Channel *must* be closed to signal the end of the stream.
// If the request context is canceled, the stream will no longer process requests.
Requests chan<- []byte
}
type Response struct {
Msg []byte
Err error
}
```
There are generic versions of the server/client handlers that allow the use of type
safe implementations for data types that support msgpack marshal/unmarshal.
2023-11-20 20:09:35 -05:00
|
|
|
func TestSingleRoundtripGenerics(t *testing.T) {
|
|
|
|
defer testlogger.T.SetLogTB(t)()
|
|
|
|
errFatal := func(err error) {
|
|
|
|
t.Helper()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatal(err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
grid, err := SetupTestGrid(2)
|
|
|
|
errFatal(err)
|
|
|
|
remoteHost := grid.Hosts[1]
|
|
|
|
local := grid.Managers[0]
|
|
|
|
remote := grid.Managers[1]
|
|
|
|
|
|
|
|
// 1: Echo
|
|
|
|
h1 := NewSingleHandler[*testRequest, *testResponse](handlerTest, func() *testRequest {
|
|
|
|
return &testRequest{}
|
|
|
|
}, func() *testResponse {
|
|
|
|
return &testResponse{}
|
|
|
|
})
|
|
|
|
// Handles incoming requests, returns a response
|
|
|
|
handler1 := func(req *testRequest) (resp *testResponse, err *RemoteErr) {
|
|
|
|
resp = h1.NewResponse()
|
|
|
|
*resp = testResponse{
|
|
|
|
OrgNum: req.Num,
|
|
|
|
OrgString: req.String,
|
|
|
|
Embedded: *req,
|
|
|
|
}
|
|
|
|
return resp, nil
|
|
|
|
}
|
|
|
|
// Return error
|
|
|
|
h2 := NewSingleHandler[*testRequest, *testResponse](handlerTest2, newTestRequest, newTestResponse)
|
|
|
|
handler2 := func(req *testRequest) (resp *testResponse, err *RemoteErr) {
|
|
|
|
r := RemoteErr(req.String)
|
|
|
|
return nil, &r
|
|
|
|
}
|
|
|
|
errFatal(h1.Register(local, handler1))
|
|
|
|
errFatal(h2.Register(local, handler2))
|
|
|
|
|
|
|
|
errFatal(h1.Register(remote, handler1))
|
|
|
|
errFatal(h2.Register(remote, handler2))
|
|
|
|
|
|
|
|
// local to remote connection
|
|
|
|
remoteConn := local.Connection(remoteHost)
|
|
|
|
const testPayload = "Hello Grid World!"
|
|
|
|
|
|
|
|
start := time.Now()
|
|
|
|
req := testRequest{Num: 1, String: testPayload}
|
|
|
|
resp, err := h1.Call(context.Background(), remoteConn, &req)
|
|
|
|
errFatal(err)
|
|
|
|
if resp.OrgString != testPayload {
|
|
|
|
t.Errorf("want %q, got %q", testPayload, resp.OrgString)
|
|
|
|
}
|
|
|
|
t.Log("Roundtrip:", time.Since(start))
|
|
|
|
|
|
|
|
start = time.Now()
|
|
|
|
resp, err = h2.Call(context.Background(), remoteConn, &testRequest{Num: 1, String: testPayload})
|
|
|
|
t.Log("Roundtrip:", time.Since(start))
|
|
|
|
if err != RemoteErr(testPayload) {
|
|
|
|
t.Errorf("want error %v(%T), got %v(%T)", RemoteErr(testPayload), RemoteErr(testPayload), err, err)
|
|
|
|
}
|
|
|
|
if resp != nil {
|
|
|
|
t.Errorf("want nil, got %q", resp)
|
|
|
|
}
|
|
|
|
t.Log("Roundtrip:", time.Since(start))
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestStreamSuite(t *testing.T) {
|
|
|
|
defer testlogger.T.SetErrorTB(t)()
|
|
|
|
errFatal := func(err error) {
|
|
|
|
t.Helper()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatal(err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
grid, err := SetupTestGrid(2)
|
|
|
|
errFatal(err)
|
|
|
|
t.Cleanup(grid.Cleanup)
|
|
|
|
|
|
|
|
local := grid.Managers[0]
|
|
|
|
localHost := grid.Hosts[0]
|
|
|
|
remote := grid.Managers[1]
|
|
|
|
remoteHost := grid.Hosts[1]
|
|
|
|
|
|
|
|
connLocalToRemote := local.Connection(remoteHost)
|
|
|
|
connRemoteLocal := remote.Connection(localHost)
|
|
|
|
|
|
|
|
t.Run("testStreamRoundtrip", func(t *testing.T) {
|
|
|
|
defer timeout(5 * time.Second)()
|
|
|
|
testStreamRoundtrip(t, local, remote)
|
|
|
|
assertNoActive(t, connRemoteLocal)
|
|
|
|
assertNoActive(t, connLocalToRemote)
|
|
|
|
})
|
|
|
|
t.Run("testStreamCancel", func(t *testing.T) {
|
|
|
|
defer timeout(5 * time.Second)()
|
|
|
|
testStreamCancel(t, local, remote)
|
|
|
|
assertNoActive(t, connRemoteLocal)
|
|
|
|
assertNoActive(t, connLocalToRemote)
|
|
|
|
})
|
|
|
|
t.Run("testStreamDeadline", func(t *testing.T) {
|
|
|
|
defer timeout(5 * time.Second)()
|
|
|
|
testStreamDeadline(t, local, remote)
|
|
|
|
assertNoActive(t, connRemoteLocal)
|
|
|
|
assertNoActive(t, connLocalToRemote)
|
|
|
|
})
|
|
|
|
t.Run("testServerOutCongestion", func(t *testing.T) {
|
|
|
|
defer timeout(1 * time.Minute)()
|
|
|
|
testServerOutCongestion(t, local, remote)
|
|
|
|
assertNoActive(t, connRemoteLocal)
|
|
|
|
assertNoActive(t, connLocalToRemote)
|
|
|
|
})
|
|
|
|
t.Run("testServerInCongestion", func(t *testing.T) {
|
|
|
|
defer timeout(1 * time.Minute)()
|
|
|
|
testServerInCongestion(t, local, remote)
|
|
|
|
assertNoActive(t, connRemoteLocal)
|
|
|
|
assertNoActive(t, connLocalToRemote)
|
|
|
|
})
|
|
|
|
t.Run("testGenericsStreamRoundtrip", func(t *testing.T) {
|
|
|
|
defer timeout(1 * time.Minute)()
|
|
|
|
testGenericsStreamRoundtrip(t, local, remote)
|
|
|
|
assertNoActive(t, connRemoteLocal)
|
|
|
|
assertNoActive(t, connLocalToRemote)
|
|
|
|
})
|
|
|
|
t.Run("testGenericsStreamRoundtripSubroute", func(t *testing.T) {
|
|
|
|
defer timeout(1 * time.Minute)()
|
|
|
|
testGenericsStreamRoundtripSubroute(t, local, remote)
|
|
|
|
assertNoActive(t, connRemoteLocal)
|
|
|
|
assertNoActive(t, connLocalToRemote)
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
|
|
|
func testStreamRoundtrip(t *testing.T, local, remote *Manager) {
|
|
|
|
defer testlogger.T.SetErrorTB(t)()
|
|
|
|
defer timeout(5 * time.Second)()
|
|
|
|
errFatal := func(err error) {
|
|
|
|
t.Helper()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatal(err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// We fake a local and remote server.
|
|
|
|
remoteHost := remote.HostName()
|
|
|
|
|
|
|
|
// 1: Echo
|
|
|
|
register := func(manager *Manager) {
|
|
|
|
errFatal(manager.RegisterStreamingHandler(handlerTest, StreamHandler{
|
|
|
|
Handle: func(ctx context.Context, payload []byte, request <-chan []byte, resp chan<- []byte) *RemoteErr {
|
|
|
|
for in := range request {
|
|
|
|
b := append([]byte{}, payload...)
|
|
|
|
b = append(b, in...)
|
|
|
|
resp <- b
|
|
|
|
}
|
|
|
|
t.Log(GetCaller(ctx).Name, "Handler done")
|
|
|
|
return nil
|
|
|
|
},
|
|
|
|
OutCapacity: 1,
|
|
|
|
InCapacity: 1,
|
|
|
|
}))
|
|
|
|
// 2: Return as error
|
|
|
|
errFatal(manager.RegisterStreamingHandler(handlerTest2, StreamHandler{
|
|
|
|
Handle: func(ctx context.Context, payload []byte, request <-chan []byte, resp chan<- []byte) *RemoteErr {
|
|
|
|
for in := range request {
|
|
|
|
t.Log("2: Got err request", string(in))
|
|
|
|
err := RemoteErr(append(payload, in...))
|
|
|
|
return &err
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
},
|
|
|
|
OutCapacity: 1,
|
|
|
|
InCapacity: 1,
|
|
|
|
}))
|
|
|
|
}
|
|
|
|
register(local)
|
|
|
|
register(remote)
|
|
|
|
|
|
|
|
// local to remote
|
|
|
|
remoteConn := local.Connection(remoteHost)
|
|
|
|
const testPayload = "Hello Grid World!"
|
|
|
|
|
|
|
|
start := time.Now()
|
|
|
|
stream, err := remoteConn.NewStream(context.Background(), handlerTest, []byte(testPayload))
|
|
|
|
errFatal(err)
|
|
|
|
var n int
|
|
|
|
stream.Requests <- []byte(strconv.Itoa(n))
|
|
|
|
for resp := range stream.responses {
|
|
|
|
errFatal(resp.Err)
|
|
|
|
t.Logf("got resp: %+v", string(resp.Msg))
|
|
|
|
if string(resp.Msg) != testPayload+strconv.Itoa(n) {
|
|
|
|
t.Errorf("want %q, got %q", testPayload+strconv.Itoa(n), string(resp.Msg))
|
|
|
|
}
|
|
|
|
if n == 10 {
|
|
|
|
close(stream.Requests)
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
n++
|
|
|
|
t.Log("sending new client request")
|
|
|
|
stream.Requests <- []byte(strconv.Itoa(n))
|
|
|
|
}
|
|
|
|
t.Log("EOF. 10 Roundtrips:", time.Since(start))
|
|
|
|
}
|
|
|
|
|
|
|
|
func testStreamCancel(t *testing.T, local, remote *Manager) {
|
|
|
|
defer testlogger.T.SetErrorTB(t)()
|
|
|
|
errFatal := func(err error) {
|
|
|
|
t.Helper()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatal(err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// We fake a local and remote server.
|
|
|
|
remoteHost := remote.HostName()
|
|
|
|
|
|
|
|
// 1: Echo
|
|
|
|
serverCanceled := make(chan struct{})
|
|
|
|
register := func(manager *Manager) {
|
|
|
|
errFatal(manager.RegisterStreamingHandler(handlerTest, StreamHandler{
|
|
|
|
Handle: func(ctx context.Context, payload []byte, request <-chan []byte, resp chan<- []byte) *RemoteErr {
|
|
|
|
<-ctx.Done()
|
|
|
|
serverCanceled <- struct{}{}
|
|
|
|
t.Log(GetCaller(ctx).Name, "Server Context canceled")
|
|
|
|
return nil
|
|
|
|
},
|
|
|
|
OutCapacity: 1,
|
|
|
|
InCapacity: 0,
|
|
|
|
}))
|
|
|
|
errFatal(manager.RegisterStreamingHandler(handlerTest2, StreamHandler{
|
|
|
|
Handle: func(ctx context.Context, payload []byte, request <-chan []byte, resp chan<- []byte) *RemoteErr {
|
|
|
|
<-ctx.Done()
|
|
|
|
serverCanceled <- struct{}{}
|
|
|
|
t.Log(GetCaller(ctx).Name, "Server Context canceled")
|
|
|
|
return nil
|
|
|
|
},
|
|
|
|
OutCapacity: 1,
|
|
|
|
InCapacity: 1,
|
|
|
|
}))
|
|
|
|
}
|
|
|
|
register(local)
|
|
|
|
register(remote)
|
|
|
|
|
|
|
|
// local to remote
|
|
|
|
testHandler := func(t *testing.T, handler HandlerID) {
|
|
|
|
remoteConn := local.Connection(remoteHost)
|
|
|
|
const testPayload = "Hello Grid World!"
|
|
|
|
|
|
|
|
ctx, cancel := context.WithCancel(context.Background())
|
|
|
|
st, err := remoteConn.NewStream(ctx, handlerTest, []byte(testPayload))
|
|
|
|
errFatal(err)
|
|
|
|
clientCanceled := make(chan time.Time, 1)
|
|
|
|
err = nil
|
|
|
|
go func(t *testing.T) {
|
|
|
|
for resp := range st.responses {
|
|
|
|
t.Log("got resp:", string(resp.Msg), "err:", resp.Err)
|
|
|
|
if err != nil {
|
|
|
|
t.Log("ERROR: got second error:", resp.Err, "first:", err)
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
err = resp.Err
|
|
|
|
}
|
|
|
|
t.Log("Client Context canceled. err state:", err)
|
|
|
|
clientCanceled <- time.Now()
|
|
|
|
}(t)
|
|
|
|
start := time.Now()
|
|
|
|
cancel()
|
|
|
|
<-serverCanceled
|
|
|
|
t.Log("server cancel time:", time.Since(start))
|
|
|
|
clientEnd := <-clientCanceled
|
|
|
|
if !errors.Is(err, context.Canceled) {
|
|
|
|
t.Error("expected context.Canceled, got", err)
|
|
|
|
}
|
|
|
|
t.Log("client after", time.Since(clientEnd))
|
|
|
|
}
|
|
|
|
// local to remote, unbuffered
|
|
|
|
t.Run("unbuffered", func(t *testing.T) {
|
|
|
|
testHandler(t, handlerTest)
|
|
|
|
})
|
|
|
|
|
|
|
|
t.Run("buffered", func(t *testing.T) {
|
|
|
|
testHandler(t, handlerTest2)
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
|
|
|
// testStreamDeadline will test if server
|
|
|
|
func testStreamDeadline(t *testing.T, local, remote *Manager) {
|
|
|
|
defer testlogger.T.SetErrorTB(t)()
|
|
|
|
errFatal := func(err error) {
|
|
|
|
t.Helper()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatal(err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
const wantDL = 50 * time.Millisecond
|
|
|
|
// We fake a local and remote server.
|
|
|
|
remoteHost := remote.HostName()
|
|
|
|
|
|
|
|
// 1: Echo
|
|
|
|
serverCanceled := make(chan time.Duration, 1)
|
|
|
|
register := func(manager *Manager) {
|
|
|
|
errFatal(manager.RegisterStreamingHandler(handlerTest, StreamHandler{
|
|
|
|
Handle: func(ctx context.Context, payload []byte, request <-chan []byte, resp chan<- []byte) *RemoteErr {
|
|
|
|
started := time.Now()
|
|
|
|
dl, _ := ctx.Deadline()
|
|
|
|
if testing.Verbose() {
|
|
|
|
fmt.Println(GetCaller(ctx).Name, "Server deadline:", time.Until(dl))
|
|
|
|
}
|
|
|
|
<-ctx.Done()
|
|
|
|
serverCanceled <- time.Since(started)
|
|
|
|
if testing.Verbose() {
|
|
|
|
fmt.Println(GetCaller(ctx).Name, "Server Context canceled with", ctx.Err(), "after", time.Since(started))
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
},
|
|
|
|
OutCapacity: 1,
|
|
|
|
InCapacity: 0,
|
|
|
|
}))
|
|
|
|
errFatal(manager.RegisterStreamingHandler(handlerTest2, StreamHandler{
|
|
|
|
Handle: func(ctx context.Context, payload []byte, request <-chan []byte, resp chan<- []byte) *RemoteErr {
|
|
|
|
started := time.Now()
|
|
|
|
dl, _ := ctx.Deadline()
|
|
|
|
if testing.Verbose() {
|
|
|
|
fmt.Println(GetCaller(ctx).Name, "Server deadline:", time.Until(dl))
|
|
|
|
}
|
|
|
|
<-ctx.Done()
|
|
|
|
serverCanceled <- time.Since(started)
|
|
|
|
if testing.Verbose() {
|
|
|
|
fmt.Println(GetCaller(ctx).Name, "Server Context canceled with", ctx.Err(), "after", time.Since(started))
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
},
|
|
|
|
OutCapacity: 1,
|
|
|
|
InCapacity: 1,
|
|
|
|
}))
|
|
|
|
}
|
|
|
|
register(local)
|
|
|
|
register(remote)
|
|
|
|
// Double remote DL
|
|
|
|
local.debugMsg(debugAddToDeadline, wantDL)
|
|
|
|
defer local.debugMsg(debugAddToDeadline, time.Duration(0))
|
|
|
|
remote.debugMsg(debugAddToDeadline, wantDL)
|
|
|
|
defer remote.debugMsg(debugAddToDeadline, time.Duration(0))
|
|
|
|
|
|
|
|
testHandler := func(t *testing.T, handler HandlerID) {
|
|
|
|
remoteConn := local.Connection(remoteHost)
|
|
|
|
const testPayload = "Hello Grid World!"
|
|
|
|
|
|
|
|
ctx, cancel := context.WithTimeout(context.Background(), wantDL)
|
|
|
|
defer cancel()
|
|
|
|
st, err := remoteConn.NewStream(ctx, handler, []byte(testPayload))
|
|
|
|
errFatal(err)
|
|
|
|
clientCanceled := make(chan time.Duration, 1)
|
|
|
|
go func() {
|
|
|
|
started := time.Now()
|
|
|
|
for resp := range st.responses {
|
|
|
|
err = resp.Err
|
|
|
|
}
|
|
|
|
clientCanceled <- time.Since(started)
|
|
|
|
}()
|
|
|
|
serverEnd := <-serverCanceled
|
|
|
|
clientEnd := <-clientCanceled
|
|
|
|
t.Log("server cancel time:", serverEnd)
|
|
|
|
t.Log("client cancel time:", clientEnd)
|
|
|
|
if !errors.Is(err, context.DeadlineExceeded) {
|
|
|
|
t.Error("expected context.DeadlineExceeded, got", err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// local to remote, unbuffered
|
|
|
|
t.Run("unbuffered", func(t *testing.T) {
|
|
|
|
testHandler(t, handlerTest)
|
|
|
|
})
|
|
|
|
|
|
|
|
t.Run("buffered", func(t *testing.T) {
|
|
|
|
testHandler(t, handlerTest2)
|
|
|
|
})
|
|
|
|
}
|
|
|
|
|
|
|
|
func testServerOutCongestion(t *testing.T, local, remote *Manager) {
|
|
|
|
defer testlogger.T.SetErrorTB(t)()
|
|
|
|
errFatal := func(err error) {
|
|
|
|
t.Helper()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatal(err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// We fake a local and remote server.
|
|
|
|
remoteHost := remote.HostName()
|
|
|
|
|
|
|
|
// 1: Echo
|
|
|
|
serverSent := make(chan struct{})
|
|
|
|
register := func(manager *Manager) {
|
|
|
|
errFatal(manager.RegisterStreamingHandler(handlerTest, StreamHandler{
|
|
|
|
Handle: func(ctx context.Context, payload []byte, request <-chan []byte, resp chan<- []byte) *RemoteErr {
|
|
|
|
// Send many responses.
|
|
|
|
// Test that this doesn't block.
|
|
|
|
for i := byte(0); i < 100; i++ {
|
|
|
|
select {
|
|
|
|
case resp <- []byte{i}:
|
|
|
|
// ok
|
|
|
|
case <-ctx.Done():
|
|
|
|
return NewRemoteErr(ctx.Err())
|
|
|
|
}
|
|
|
|
if i == 0 {
|
|
|
|
close(serverSent)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
},
|
|
|
|
OutCapacity: 1,
|
|
|
|
InCapacity: 0,
|
|
|
|
}))
|
|
|
|
errFatal(manager.RegisterSingleHandler(handlerTest2, func(payload []byte) ([]byte, *RemoteErr) {
|
|
|
|
// Simple roundtrip
|
|
|
|
return append([]byte{}, payload...), nil
|
|
|
|
}))
|
|
|
|
}
|
|
|
|
register(local)
|
|
|
|
register(remote)
|
|
|
|
|
|
|
|
remoteConn := local.Connection(remoteHost)
|
|
|
|
const testPayload = "Hello Grid World!"
|
|
|
|
|
|
|
|
ctx, cancel := context.WithTimeout(context.Background(), time.Minute)
|
|
|
|
defer cancel()
|
|
|
|
st, err := remoteConn.NewStream(ctx, handlerTest, []byte(testPayload))
|
|
|
|
errFatal(err)
|
|
|
|
|
|
|
|
// Wait for the server to send the first response.
|
|
|
|
<-serverSent
|
|
|
|
|
|
|
|
// Now do 100 other requests to ensure that the server doesn't block.
|
|
|
|
for i := 0; i < 100; i++ {
|
|
|
|
_, err := remoteConn.Request(ctx, handlerTest2, []byte(testPayload))
|
|
|
|
errFatal(err)
|
|
|
|
}
|
|
|
|
// Drain responses
|
|
|
|
got := 0
|
|
|
|
for resp := range st.responses {
|
|
|
|
// t.Log("got response", resp)
|
|
|
|
errFatal(resp.Err)
|
|
|
|
if resp.Msg[0] != byte(got) {
|
|
|
|
t.Error("expected response", got, "got", resp.Msg[0])
|
|
|
|
}
|
|
|
|
got++
|
|
|
|
}
|
|
|
|
if got != 100 {
|
|
|
|
t.Error("expected 100 responses, got", got)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func testServerInCongestion(t *testing.T, local, remote *Manager) {
|
|
|
|
defer testlogger.T.SetErrorTB(t)()
|
|
|
|
errFatal := func(err error) {
|
|
|
|
t.Helper()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatal(err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// We fake a local and remote server.
|
|
|
|
remoteHost := remote.HostName()
|
|
|
|
|
|
|
|
// 1: Echo
|
|
|
|
processHandler := make(chan struct{})
|
|
|
|
register := func(manager *Manager) {
|
|
|
|
errFatal(manager.RegisterStreamingHandler(handlerTest, StreamHandler{
|
|
|
|
Handle: func(ctx context.Context, payload []byte, request <-chan []byte, resp chan<- []byte) *RemoteErr {
|
|
|
|
// Block incoming requests.
|
|
|
|
var n byte
|
|
|
|
<-processHandler
|
|
|
|
for {
|
|
|
|
select {
|
|
|
|
case in, ok := <-request:
|
|
|
|
if !ok {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
if in[0] != n {
|
|
|
|
return NewRemoteErrString(fmt.Sprintf("expected incoming %d, got %d", n, in[0]))
|
|
|
|
}
|
|
|
|
n++
|
|
|
|
resp <- append([]byte{}, in...)
|
|
|
|
case <-ctx.Done():
|
|
|
|
return NewRemoteErr(ctx.Err())
|
|
|
|
}
|
|
|
|
}
|
|
|
|
},
|
|
|
|
OutCapacity: 5,
|
|
|
|
InCapacity: 5,
|
|
|
|
}))
|
|
|
|
errFatal(manager.RegisterSingleHandler(handlerTest2, func(payload []byte) ([]byte, *RemoteErr) {
|
|
|
|
// Simple roundtrip
|
|
|
|
return append([]byte{}, payload...), nil
|
|
|
|
}))
|
|
|
|
}
|
|
|
|
register(local)
|
|
|
|
register(remote)
|
|
|
|
|
|
|
|
remoteConn := local.Connection(remoteHost)
|
|
|
|
const testPayload = "Hello Grid World!"
|
|
|
|
|
|
|
|
ctx, cancel := context.WithTimeout(context.Background(), time.Minute)
|
|
|
|
defer cancel()
|
|
|
|
st, err := remoteConn.NewStream(ctx, handlerTest, []byte(testPayload))
|
|
|
|
errFatal(err)
|
|
|
|
|
|
|
|
// Start sending requests.
|
|
|
|
go func() {
|
|
|
|
for i := byte(0); i < 100; i++ {
|
|
|
|
st.Requests <- []byte{i}
|
|
|
|
}
|
|
|
|
close(st.Requests)
|
|
|
|
}()
|
|
|
|
// Now do 100 other requests to ensure that the server doesn't block.
|
|
|
|
for i := 0; i < 100; i++ {
|
|
|
|
_, err := remoteConn.Request(ctx, handlerTest2, []byte(testPayload))
|
|
|
|
errFatal(err)
|
|
|
|
}
|
|
|
|
// Start processing requests.
|
|
|
|
close(processHandler)
|
|
|
|
|
|
|
|
// Drain responses
|
|
|
|
got := 0
|
|
|
|
for resp := range st.responses {
|
|
|
|
// t.Log("got response", resp)
|
|
|
|
errFatal(resp.Err)
|
|
|
|
if resp.Msg[0] != byte(got) {
|
|
|
|
t.Error("expected response", got, "got", resp.Msg[0])
|
|
|
|
}
|
|
|
|
got++
|
|
|
|
}
|
|
|
|
if got != 100 {
|
|
|
|
t.Error("expected 100 responses, got", got)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func testGenericsStreamRoundtrip(t *testing.T, local, remote *Manager) {
|
|
|
|
defer testlogger.T.SetErrorTB(t)()
|
|
|
|
defer timeout(5 * time.Second)()
|
|
|
|
errFatal := func(err error) {
|
|
|
|
t.Helper()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatal(err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// We fake a local and remote server.
|
|
|
|
remoteHost := remote.HostName()
|
|
|
|
handler := NewStream[*testRequest, *testRequest, *testResponse](handlerTest, newTestRequest, newTestRequest, newTestResponse)
|
|
|
|
handler.InCapacity = 1
|
|
|
|
handler.OutCapacity = 1
|
|
|
|
const payloads = 10
|
|
|
|
|
|
|
|
// 1: Echo
|
|
|
|
register := func(manager *Manager) {
|
|
|
|
errFatal(handler.Register(manager, func(ctx context.Context, pp *testRequest, in <-chan *testRequest, out chan<- *testResponse) *RemoteErr {
|
|
|
|
n := 0
|
|
|
|
for i := range in {
|
|
|
|
if n > payloads {
|
|
|
|
panic("too many requests")
|
|
|
|
}
|
|
|
|
|
|
|
|
// t.Log("Got request:", *i)
|
|
|
|
out <- &testResponse{
|
|
|
|
OrgNum: i.Num + pp.Num,
|
|
|
|
OrgString: pp.String + i.String,
|
|
|
|
Embedded: *i,
|
|
|
|
}
|
|
|
|
n++
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}))
|
|
|
|
}
|
|
|
|
register(local)
|
|
|
|
register(remote)
|
|
|
|
|
|
|
|
// local to remote
|
|
|
|
remoteConn := local.Connection(remoteHost)
|
|
|
|
const testPayload = "Hello Grid World!"
|
|
|
|
|
|
|
|
start := time.Now()
|
|
|
|
stream, err := handler.Call(context.Background(), remoteConn, &testRequest{Num: 1, String: testPayload})
|
|
|
|
errFatal(err)
|
|
|
|
go func() {
|
|
|
|
defer close(stream.Requests)
|
|
|
|
for i := 0; i < payloads; i++ {
|
|
|
|
// t.Log("sending new client request")
|
|
|
|
stream.Requests <- &testRequest{Num: i, String: testPayload}
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
var n int
|
|
|
|
err = stream.Results(func(resp *testResponse) error {
|
|
|
|
const wantString = testPayload + testPayload
|
|
|
|
if resp.OrgString != testPayload+testPayload {
|
|
|
|
t.Errorf("want %q, got %q", wantString, resp.OrgString)
|
|
|
|
}
|
|
|
|
if resp.OrgNum != n+1 {
|
|
|
|
t.Errorf("want %d, got %d", n+1, resp.OrgNum)
|
|
|
|
}
|
|
|
|
handler.PutResponse(resp)
|
|
|
|
n++
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
errFatal(err)
|
|
|
|
t.Log("EOF.", payloads, " Roundtrips:", time.Since(start))
|
|
|
|
}
|
|
|
|
|
|
|
|
func testGenericsStreamRoundtripSubroute(t *testing.T, local, remote *Manager) {
|
|
|
|
defer testlogger.T.SetErrorTB(t)()
|
|
|
|
defer timeout(5 * time.Second)()
|
|
|
|
errFatal := func(err error) {
|
|
|
|
t.Helper()
|
|
|
|
if err != nil {
|
|
|
|
t.Fatal(err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// We fake a local and remote server.
|
|
|
|
remoteHost := remote.HostName()
|
|
|
|
handler := NewStream[*testRequest, *testRequest, *testResponse](handlerTest, newTestRequest, newTestRequest, newTestResponse)
|
|
|
|
handler.InCapacity = 1
|
|
|
|
handler.OutCapacity = 1
|
|
|
|
const payloads = 10
|
|
|
|
|
|
|
|
// 1: Echo
|
|
|
|
register := func(manager *Manager) {
|
|
|
|
errFatal(handler.Register(manager, func(ctx context.Context, pp *testRequest, in <-chan *testRequest, out chan<- *testResponse) *RemoteErr {
|
|
|
|
sub := GetSubroute(ctx)
|
|
|
|
if sub != "subroute/1" {
|
|
|
|
t.Fatal("expected subroute/1, got", sub)
|
|
|
|
}
|
|
|
|
n := 0
|
|
|
|
for i := range in {
|
|
|
|
if n > payloads {
|
|
|
|
panic("too many requests")
|
|
|
|
}
|
|
|
|
|
|
|
|
// t.Log("Got request:", *i)
|
|
|
|
out <- &testResponse{
|
|
|
|
OrgNum: i.Num + pp.Num,
|
|
|
|
OrgString: pp.String + i.String,
|
|
|
|
Embedded: *i,
|
|
|
|
}
|
|
|
|
n++
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}, "subroute", "1"))
|
|
|
|
}
|
|
|
|
register(local)
|
|
|
|
register(remote)
|
|
|
|
|
|
|
|
// local to remote
|
|
|
|
remoteConn := local.Connection(remoteHost)
|
|
|
|
const testPayload = "Hello Grid World!"
|
|
|
|
// Add subroute
|
|
|
|
remoteSub := remoteConn.Subroute(strings.Join([]string{"subroute", "1"}, "/"))
|
|
|
|
|
|
|
|
start := time.Now()
|
|
|
|
stream, err := handler.Call(context.Background(), remoteSub, &testRequest{Num: 1, String: testPayload})
|
|
|
|
errFatal(err)
|
|
|
|
go func() {
|
|
|
|
defer close(stream.Requests)
|
|
|
|
for i := 0; i < payloads; i++ {
|
|
|
|
// t.Log("sending new client request")
|
|
|
|
stream.Requests <- &testRequest{Num: i, String: testPayload}
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
var n int
|
|
|
|
err = stream.Results(func(resp *testResponse) error {
|
|
|
|
// t.Logf("got resp: %+v", *resp.Msg)
|
|
|
|
const wantString = testPayload + testPayload
|
|
|
|
if resp.OrgString != testPayload+testPayload {
|
|
|
|
t.Errorf("want %q, got %q", wantString, resp.OrgString)
|
|
|
|
}
|
|
|
|
if resp.OrgNum != n+1 {
|
|
|
|
t.Errorf("want %d, got %d", n+1, resp.OrgNum)
|
|
|
|
}
|
|
|
|
handler.PutResponse(resp)
|
|
|
|
n++
|
|
|
|
return nil
|
|
|
|
})
|
|
|
|
|
|
|
|
errFatal(err)
|
|
|
|
t.Log("EOF.", payloads, " Roundtrips:", time.Since(start))
|
|
|
|
}
|
|
|
|
|
|
|
|
func timeout(after time.Duration) (cancel func()) {
|
|
|
|
c := time.After(after)
|
|
|
|
cc := make(chan struct{})
|
|
|
|
go func() {
|
|
|
|
select {
|
|
|
|
case <-cc:
|
|
|
|
return
|
|
|
|
case <-c:
|
|
|
|
buf := make([]byte, 1<<20)
|
|
|
|
stacklen := runtime.Stack(buf, true)
|
|
|
|
fmt.Printf("=== Timeout, assuming deadlock ===\n*** goroutine dump...\n%s\n*** end\n", string(buf[:stacklen]))
|
|
|
|
os.Exit(2)
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
return func() {
|
|
|
|
close(cc)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func assertNoActive(t *testing.T, c *Connection) {
|
|
|
|
t.Helper()
|
|
|
|
// Tiny bit racy for tests, but we try to play nice.
|
|
|
|
for i := 10; i >= 0; i-- {
|
|
|
|
runtime.Gosched()
|
|
|
|
stats := c.Stats()
|
|
|
|
if stats.IncomingStreams != 0 {
|
|
|
|
if i > 0 {
|
|
|
|
time.Sleep(100 * time.Millisecond)
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
var found []uint64
|
|
|
|
c.inStream.Range(func(key uint64, value *muxServer) bool {
|
|
|
|
found = append(found, key)
|
|
|
|
return true
|
|
|
|
})
|
|
|
|
t.Errorf("expected no active streams, got %d incoming: %v", stats.IncomingStreams, found)
|
|
|
|
}
|
|
|
|
if stats.OutgoingStreams != 0 {
|
|
|
|
if i > 0 {
|
|
|
|
time.Sleep(100 * time.Millisecond)
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
var found []uint64
|
|
|
|
c.outgoing.Range(func(key uint64, value *muxClient) bool {
|
|
|
|
found = append(found, key)
|
|
|
|
return true
|
|
|
|
})
|
|
|
|
t.Errorf("expected no active streams, got %d outgoing: %v", stats.OutgoingStreams, found)
|
|
|
|
}
|
|
|
|
return
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Inserted manually.
|
|
|
|
func _() {
|
|
|
|
// An "invalid array index" compiler error signifies that the constant values have changed.
|
|
|
|
// Re-run the stringer command to generate them again.
|
|
|
|
var x [1]struct{}
|
|
|
|
_ = x[StateUnconnected-0]
|
|
|
|
_ = x[StateConnecting-1]
|
|
|
|
_ = x[StateConnected-2]
|
|
|
|
_ = x[StateConnectionError-3]
|
|
|
|
_ = x[StateShutdown-4]
|
|
|
|
}
|
|
|
|
|
|
|
|
const stateName = "UnconnectedConnectingConnectedConnectionErrorShutdown"
|
|
|
|
|
|
|
|
var stateIndex = [...]uint8{0, 11, 21, 30, 45, 53}
|
|
|
|
|
|
|
|
func (i State) String() string {
|
|
|
|
if i >= State(len(stateIndex)-1) {
|
|
|
|
return "State(" + strconv.FormatInt(int64(i), 10) + ")"
|
|
|
|
}
|
|
|
|
return stateName[stateIndex[i]:stateIndex[i+1]]
|
|
|
|
}
|