mirror of
https://github.com/minio/minio.git
synced 2024-12-27 15:45:55 -05:00
7ec43bd177
We have observed cases where a blocked stream will block for cancellations. This happens when response channel is blocked and we want to push an error. This will have the response mutex locked, which will prevent all other operations until upstream is unblocked. Make this behavior non-blocking and if blocked spawn a goroutine that will send the response and close the output. Still a lot of "dancing". Added a test for this and reviewed.
1103 lines
30 KiB
Go
1103 lines
30 KiB
Go
// Copyright (c) 2015-2023 MinIO, Inc.
|
|
//
|
|
// This file is part of MinIO Object Storage stack
|
|
//
|
|
// This program is free software: you can redistribute it and/or modify
|
|
// it under the terms of the GNU Affero General Public License as published by
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
// (at your option) any later version.
|
|
//
|
|
// This program is distributed in the hope that it will be useful
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
// GNU Affero General Public License for more details.
|
|
//
|
|
// You should have received a copy of the GNU Affero General Public License
|
|
// along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
package grid
|
|
|
|
import (
|
|
"bytes"
|
|
"context"
|
|
"errors"
|
|
"fmt"
|
|
"os"
|
|
"runtime"
|
|
"strconv"
|
|
"strings"
|
|
"testing"
|
|
"time"
|
|
|
|
"github.com/minio/minio/internal/logger/target/testlogger"
|
|
)
|
|
|
|
func TestSingleRoundtrip(t *testing.T) {
|
|
defer testlogger.T.SetLogTB(t)()
|
|
errFatal := func(err error) {
|
|
t.Helper()
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
}
|
|
grid, err := SetupTestGrid(2)
|
|
errFatal(err)
|
|
remoteHost := grid.Hosts[1]
|
|
local := grid.Managers[0]
|
|
|
|
// 1: Echo
|
|
errFatal(local.RegisterSingleHandler(handlerTest, func(payload []byte) ([]byte, *RemoteErr) {
|
|
t.Log("1: server payload: ", len(payload), "bytes.")
|
|
return append([]byte{}, payload...), nil
|
|
}))
|
|
// 2: Return as error
|
|
errFatal(local.RegisterSingleHandler(handlerTest2, func(payload []byte) ([]byte, *RemoteErr) {
|
|
t.Log("2: server payload: ", len(payload), "bytes.")
|
|
err := RemoteErr(payload)
|
|
return nil, &err
|
|
}))
|
|
|
|
remote := grid.Managers[1]
|
|
|
|
// 1: Echo
|
|
errFatal(remote.RegisterSingleHandler(handlerTest, func(payload []byte) ([]byte, *RemoteErr) {
|
|
t.Log("1: server payload: ", len(payload), "bytes.")
|
|
return append([]byte{}, payload...), nil
|
|
}))
|
|
// 2: Return as error
|
|
errFatal(remote.RegisterSingleHandler(handlerTest2, func(payload []byte) ([]byte, *RemoteErr) {
|
|
t.Log("2: server payload: ", len(payload), "bytes.")
|
|
err := RemoteErr(payload)
|
|
return nil, &err
|
|
}))
|
|
|
|
// local to remote
|
|
remoteConn := local.Connection(remoteHost)
|
|
remoteConn.WaitForConnect(context.Background())
|
|
defer testlogger.T.SetErrorTB(t)()
|
|
|
|
t.Run("localToRemote", func(t *testing.T) {
|
|
const testPayload = "Hello Grid World!"
|
|
|
|
start := time.Now()
|
|
resp, err := remoteConn.Request(context.Background(), handlerTest, []byte(testPayload))
|
|
errFatal(err)
|
|
if string(resp) != testPayload {
|
|
t.Errorf("want %q, got %q", testPayload, string(resp))
|
|
}
|
|
t.Log("Roundtrip:", time.Since(start))
|
|
})
|
|
|
|
t.Run("localToRemoteErr", func(t *testing.T) {
|
|
const testPayload = "Hello Grid World!"
|
|
start := time.Now()
|
|
resp, err := remoteConn.Request(context.Background(), handlerTest2, []byte(testPayload))
|
|
t.Log("Roundtrip:", time.Since(start))
|
|
if len(resp) != 0 {
|
|
t.Errorf("want nil, got %q", string(resp))
|
|
}
|
|
if err != RemoteErr(testPayload) {
|
|
t.Errorf("want error %v(%T), got %v(%T)", RemoteErr(testPayload), RemoteErr(testPayload), err, err)
|
|
}
|
|
t.Log("Roundtrip:", time.Since(start))
|
|
})
|
|
|
|
t.Run("localToRemoteHuge", func(t *testing.T) {
|
|
testPayload := bytes.Repeat([]byte("?"), 1<<20)
|
|
|
|
start := time.Now()
|
|
resp, err := remoteConn.Request(context.Background(), handlerTest, testPayload)
|
|
errFatal(err)
|
|
if string(resp) != string(testPayload) {
|
|
t.Errorf("want %q, got %q", testPayload, string(resp))
|
|
}
|
|
t.Log("Roundtrip:", time.Since(start))
|
|
})
|
|
|
|
t.Run("localToRemoteErrHuge", func(t *testing.T) {
|
|
testPayload := bytes.Repeat([]byte("!"), 1<<10)
|
|
|
|
start := time.Now()
|
|
resp, err := remoteConn.Request(context.Background(), handlerTest2, testPayload)
|
|
if len(resp) != 0 {
|
|
t.Errorf("want nil, got %q", string(resp))
|
|
}
|
|
if err != RemoteErr(testPayload) {
|
|
t.Errorf("want error %v(%T), got %v(%T)", RemoteErr(testPayload), RemoteErr(testPayload), err, err)
|
|
}
|
|
t.Log("Roundtrip:", time.Since(start))
|
|
})
|
|
}
|
|
|
|
func TestSingleRoundtripNotReady(t *testing.T) {
|
|
defer testlogger.T.SetLogTB(t)()
|
|
errFatal := func(t testing.TB, err error) {
|
|
t.Helper()
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
}
|
|
grid, err := SetupTestGrid(2)
|
|
errFatal(t, err)
|
|
remoteHost := grid.Hosts[1]
|
|
local := grid.Managers[0]
|
|
|
|
// 1: Echo
|
|
errFatal(t, local.RegisterSingleHandler(handlerTest, func(payload []byte) ([]byte, *RemoteErr) {
|
|
t.Log("1: server payload: ", len(payload), "bytes.")
|
|
return append([]byte{}, payload...), nil
|
|
}))
|
|
// 2: Return as error
|
|
errFatal(t, local.RegisterSingleHandler(handlerTest2, func(payload []byte) ([]byte, *RemoteErr) {
|
|
t.Log("2: server payload: ", len(payload), "bytes.")
|
|
err := RemoteErr(payload)
|
|
return nil, &err
|
|
}))
|
|
|
|
// Do not register remote handlers
|
|
|
|
// local to remote
|
|
remoteConn := local.Connection(remoteHost)
|
|
remoteConn.WaitForConnect(context.Background())
|
|
defer testlogger.T.SetErrorTB(t)()
|
|
|
|
t.Run("localToRemote", func(t *testing.T) {
|
|
const testPayload = "Hello Grid World!"
|
|
// Single requests should have remote errors.
|
|
_, err := remoteConn.Request(context.Background(), handlerTest, []byte(testPayload))
|
|
if v, ok := err.(*RemoteErr); !ok || v.Error() != "Invalid Handler for type" {
|
|
t.Fatalf("Unexpected error: %v, %T", err, err)
|
|
}
|
|
// Streams should not be able to set up until registered.
|
|
// Thus, the error is a local error.
|
|
_, err = remoteConn.NewStream(context.Background(), handlerTest, []byte(testPayload))
|
|
if !errors.Is(err, ErrUnknownHandler) {
|
|
t.Fatalf("Unexpected error: %v, %T", err, err)
|
|
}
|
|
})
|
|
}
|
|
|
|
func TestSingleRoundtripGenerics(t *testing.T) {
|
|
defer testlogger.T.SetLogTB(t)()
|
|
errFatal := func(err error) {
|
|
t.Helper()
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
}
|
|
grid, err := SetupTestGrid(2)
|
|
errFatal(err)
|
|
remoteHost := grid.Hosts[1]
|
|
local := grid.Managers[0]
|
|
remote := grid.Managers[1]
|
|
|
|
// 1: Echo
|
|
h1 := NewSingleHandler[*testRequest, *testResponse](handlerTest, func() *testRequest {
|
|
return &testRequest{}
|
|
}, func() *testResponse {
|
|
return &testResponse{}
|
|
})
|
|
// Handles incoming requests, returns a response
|
|
handler1 := func(req *testRequest) (resp *testResponse, err *RemoteErr) {
|
|
resp = h1.NewResponse()
|
|
*resp = testResponse{
|
|
OrgNum: req.Num,
|
|
OrgString: req.String,
|
|
Embedded: *req,
|
|
}
|
|
return resp, nil
|
|
}
|
|
// Return error
|
|
h2 := NewSingleHandler[*testRequest, *testResponse](handlerTest2, newTestRequest, newTestResponse)
|
|
handler2 := func(req *testRequest) (resp *testResponse, err *RemoteErr) {
|
|
r := RemoteErr(req.String)
|
|
return nil, &r
|
|
}
|
|
errFatal(h1.Register(local, handler1))
|
|
errFatal(h2.Register(local, handler2))
|
|
|
|
errFatal(h1.Register(remote, handler1))
|
|
errFatal(h2.Register(remote, handler2))
|
|
|
|
// local to remote connection
|
|
remoteConn := local.Connection(remoteHost)
|
|
const testPayload = "Hello Grid World!"
|
|
|
|
start := time.Now()
|
|
req := testRequest{Num: 1, String: testPayload}
|
|
resp, err := h1.Call(context.Background(), remoteConn, &req)
|
|
errFatal(err)
|
|
if resp.OrgString != testPayload {
|
|
t.Errorf("want %q, got %q", testPayload, resp.OrgString)
|
|
}
|
|
t.Log("Roundtrip:", time.Since(start))
|
|
h1.PutResponse(resp)
|
|
|
|
start = time.Now()
|
|
resp, err = h2.Call(context.Background(), remoteConn, &testRequest{Num: 1, String: testPayload})
|
|
t.Log("Roundtrip:", time.Since(start))
|
|
if err != RemoteErr(testPayload) {
|
|
t.Errorf("want error %v(%T), got %v(%T)", RemoteErr(testPayload), RemoteErr(testPayload), err, err)
|
|
}
|
|
if resp != nil {
|
|
t.Errorf("want nil, got %q", resp)
|
|
}
|
|
h2.PutResponse(resp)
|
|
t.Log("Roundtrip:", time.Since(start))
|
|
}
|
|
|
|
func TestSingleRoundtripGenericsRecycle(t *testing.T) {
|
|
defer testlogger.T.SetLogTB(t)()
|
|
errFatal := func(err error) {
|
|
t.Helper()
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
}
|
|
grid, err := SetupTestGrid(2)
|
|
errFatal(err)
|
|
remoteHost := grid.Hosts[1]
|
|
local := grid.Managers[0]
|
|
remote := grid.Managers[1]
|
|
|
|
// 1: Echo
|
|
h1 := NewSingleHandler[*MSS, *MSS](handlerTest, NewMSS, NewMSS)
|
|
// Handles incoming requests, returns a response
|
|
handler1 := func(req *MSS) (resp *MSS, err *RemoteErr) {
|
|
resp = h1.NewResponse()
|
|
for k, v := range *req {
|
|
(*resp)[k] = v
|
|
}
|
|
return resp, nil
|
|
}
|
|
// Return error
|
|
h2 := NewSingleHandler[*MSS, *MSS](handlerTest2, NewMSS, NewMSS)
|
|
handler2 := func(req *MSS) (resp *MSS, err *RemoteErr) {
|
|
defer req.Recycle()
|
|
r := RemoteErr(req.Get("err"))
|
|
return nil, &r
|
|
}
|
|
errFatal(h1.Register(local, handler1))
|
|
errFatal(h2.Register(local, handler2))
|
|
|
|
errFatal(h1.Register(remote, handler1))
|
|
errFatal(h2.Register(remote, handler2))
|
|
|
|
// local to remote connection
|
|
remoteConn := local.Connection(remoteHost)
|
|
const testPayload = "Hello Grid World!"
|
|
|
|
start := time.Now()
|
|
req := NewMSSWith(map[string]string{"test": testPayload})
|
|
resp, err := h1.Call(context.Background(), remoteConn, req)
|
|
errFatal(err)
|
|
if resp.Get("test") != testPayload {
|
|
t.Errorf("want %q, got %q", testPayload, resp.Get("test"))
|
|
}
|
|
t.Log("Roundtrip:", time.Since(start))
|
|
h1.PutResponse(resp)
|
|
|
|
start = time.Now()
|
|
resp, err = h2.Call(context.Background(), remoteConn, NewMSSWith(map[string]string{"err": testPayload}))
|
|
t.Log("Roundtrip:", time.Since(start))
|
|
if err != RemoteErr(testPayload) {
|
|
t.Errorf("want error %v(%T), got %v(%T)", RemoteErr(testPayload), RemoteErr(testPayload), err, err)
|
|
}
|
|
if resp != nil {
|
|
t.Errorf("want nil, got %q", resp)
|
|
}
|
|
t.Log("Roundtrip:", time.Since(start))
|
|
h2.PutResponse(resp)
|
|
}
|
|
|
|
func TestStreamSuite(t *testing.T) {
|
|
defer testlogger.T.SetErrorTB(t)()
|
|
errFatal := func(err error) {
|
|
t.Helper()
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
}
|
|
grid, err := SetupTestGrid(2)
|
|
errFatal(err)
|
|
t.Cleanup(grid.Cleanup)
|
|
|
|
local := grid.Managers[0]
|
|
localHost := grid.Hosts[0]
|
|
remote := grid.Managers[1]
|
|
remoteHost := grid.Hosts[1]
|
|
|
|
connLocalToRemote := local.Connection(remoteHost)
|
|
connRemoteLocal := remote.Connection(localHost)
|
|
|
|
t.Run("testStreamRoundtrip", func(t *testing.T) {
|
|
defer timeout(5 * time.Second)()
|
|
testStreamRoundtrip(t, local, remote)
|
|
assertNoActive(t, connRemoteLocal)
|
|
assertNoActive(t, connLocalToRemote)
|
|
})
|
|
t.Run("testStreamCancel", func(t *testing.T) {
|
|
defer timeout(5 * time.Second)()
|
|
testStreamCancel(t, local, remote)
|
|
assertNoActive(t, connRemoteLocal)
|
|
assertNoActive(t, connLocalToRemote)
|
|
})
|
|
t.Run("testStreamDeadline", func(t *testing.T) {
|
|
defer timeout(5 * time.Second)()
|
|
testStreamDeadline(t, local, remote)
|
|
assertNoActive(t, connRemoteLocal)
|
|
assertNoActive(t, connLocalToRemote)
|
|
})
|
|
t.Run("testServerOutCongestion", func(t *testing.T) {
|
|
defer timeout(1 * time.Minute)()
|
|
testServerOutCongestion(t, local, remote)
|
|
assertNoActive(t, connRemoteLocal)
|
|
assertNoActive(t, connLocalToRemote)
|
|
})
|
|
t.Run("testServerInCongestion", func(t *testing.T) {
|
|
defer timeout(1 * time.Minute)()
|
|
testServerInCongestion(t, local, remote)
|
|
assertNoActive(t, connRemoteLocal)
|
|
assertNoActive(t, connLocalToRemote)
|
|
})
|
|
t.Run("testGenericsStreamRoundtrip", func(t *testing.T) {
|
|
defer timeout(1 * time.Minute)()
|
|
testGenericsStreamRoundtrip(t, local, remote)
|
|
assertNoActive(t, connRemoteLocal)
|
|
assertNoActive(t, connLocalToRemote)
|
|
})
|
|
t.Run("testGenericsStreamRoundtripSubroute", func(t *testing.T) {
|
|
defer timeout(1 * time.Minute)()
|
|
testGenericsStreamRoundtripSubroute(t, local, remote)
|
|
assertNoActive(t, connRemoteLocal)
|
|
assertNoActive(t, connLocalToRemote)
|
|
})
|
|
t.Run("testServerStreamResponseBlocked", func(t *testing.T) {
|
|
defer timeout(1 * time.Minute)()
|
|
testServerStreamResponseBlocked(t, local, remote)
|
|
assertNoActive(t, connRemoteLocal)
|
|
assertNoActive(t, connLocalToRemote)
|
|
})
|
|
}
|
|
|
|
func testStreamRoundtrip(t *testing.T, local, remote *Manager) {
|
|
defer testlogger.T.SetErrorTB(t)()
|
|
defer timeout(5 * time.Second)()
|
|
errFatal := func(err error) {
|
|
t.Helper()
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
}
|
|
|
|
// We fake a local and remote server.
|
|
remoteHost := remote.HostName()
|
|
|
|
// 1: Echo
|
|
register := func(manager *Manager) {
|
|
errFatal(manager.RegisterStreamingHandler(handlerTest, StreamHandler{
|
|
Handle: func(ctx context.Context, payload []byte, request <-chan []byte, resp chan<- []byte) *RemoteErr {
|
|
for in := range request {
|
|
b := append([]byte{}, payload...)
|
|
b = append(b, in...)
|
|
resp <- b
|
|
}
|
|
t.Log(GetCaller(ctx).Name, "Handler done")
|
|
return nil
|
|
},
|
|
OutCapacity: 1,
|
|
InCapacity: 1,
|
|
}))
|
|
// 2: Return as error
|
|
errFatal(manager.RegisterStreamingHandler(handlerTest2, StreamHandler{
|
|
Handle: func(ctx context.Context, payload []byte, request <-chan []byte, resp chan<- []byte) *RemoteErr {
|
|
for in := range request {
|
|
t.Log("2: Got err request", string(in))
|
|
err := RemoteErr(append(payload, in...))
|
|
return &err
|
|
}
|
|
return nil
|
|
},
|
|
OutCapacity: 1,
|
|
InCapacity: 1,
|
|
}))
|
|
}
|
|
register(local)
|
|
register(remote)
|
|
|
|
// local to remote
|
|
remoteConn := local.Connection(remoteHost)
|
|
const testPayload = "Hello Grid World!"
|
|
|
|
start := time.Now()
|
|
stream, err := remoteConn.NewStream(context.Background(), handlerTest, []byte(testPayload))
|
|
errFatal(err)
|
|
var n int
|
|
stream.Requests <- []byte(strconv.Itoa(n))
|
|
for resp := range stream.responses {
|
|
errFatal(resp.Err)
|
|
t.Logf("got resp: %+v", string(resp.Msg))
|
|
if string(resp.Msg) != testPayload+strconv.Itoa(n) {
|
|
t.Errorf("want %q, got %q", testPayload+strconv.Itoa(n), string(resp.Msg))
|
|
}
|
|
if n == 10 {
|
|
close(stream.Requests)
|
|
continue
|
|
}
|
|
n++
|
|
t.Log("sending new client request")
|
|
stream.Requests <- []byte(strconv.Itoa(n))
|
|
}
|
|
t.Log("EOF. 10 Roundtrips:", time.Since(start))
|
|
}
|
|
|
|
func testStreamCancel(t *testing.T, local, remote *Manager) {
|
|
defer testlogger.T.SetErrorTB(t)()
|
|
errFatal := func(err error) {
|
|
t.Helper()
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
}
|
|
|
|
// We fake a local and remote server.
|
|
remoteHost := remote.HostName()
|
|
|
|
// 1: Echo
|
|
serverCanceled := make(chan struct{})
|
|
register := func(manager *Manager) {
|
|
errFatal(manager.RegisterStreamingHandler(handlerTest, StreamHandler{
|
|
Handle: func(ctx context.Context, payload []byte, request <-chan []byte, resp chan<- []byte) *RemoteErr {
|
|
<-ctx.Done()
|
|
serverCanceled <- struct{}{}
|
|
t.Log(GetCaller(ctx).Name, "Server Context canceled")
|
|
return nil
|
|
},
|
|
OutCapacity: 1,
|
|
InCapacity: 0,
|
|
}))
|
|
errFatal(manager.RegisterStreamingHandler(handlerTest2, StreamHandler{
|
|
Handle: func(ctx context.Context, payload []byte, request <-chan []byte, resp chan<- []byte) *RemoteErr {
|
|
<-ctx.Done()
|
|
serverCanceled <- struct{}{}
|
|
t.Log(GetCaller(ctx).Name, "Server Context canceled")
|
|
return nil
|
|
},
|
|
OutCapacity: 1,
|
|
InCapacity: 1,
|
|
}))
|
|
}
|
|
register(local)
|
|
register(remote)
|
|
|
|
// local to remote
|
|
testHandler := func(t *testing.T, handler HandlerID) {
|
|
remoteConn := local.Connection(remoteHost)
|
|
const testPayload = "Hello Grid World!"
|
|
|
|
ctx, cancel := context.WithCancel(context.Background())
|
|
st, err := remoteConn.NewStream(ctx, handlerTest, []byte(testPayload))
|
|
errFatal(err)
|
|
clientCanceled := make(chan time.Time, 1)
|
|
err = nil
|
|
go func(t *testing.T) {
|
|
for resp := range st.responses {
|
|
t.Log("got resp:", string(resp.Msg), "err:", resp.Err)
|
|
if err != nil {
|
|
t.Log("ERROR: got second error:", resp.Err, "first:", err)
|
|
continue
|
|
}
|
|
err = resp.Err
|
|
}
|
|
t.Log("Client Context canceled. err state:", err)
|
|
clientCanceled <- time.Now()
|
|
}(t)
|
|
start := time.Now()
|
|
cancel()
|
|
<-serverCanceled
|
|
t.Log("server cancel time:", time.Since(start))
|
|
clientEnd := <-clientCanceled
|
|
if !errors.Is(err, context.Canceled) {
|
|
t.Error("expected context.Canceled, got", err)
|
|
}
|
|
t.Log("client after", time.Since(clientEnd))
|
|
}
|
|
// local to remote, unbuffered
|
|
t.Run("unbuffered", func(t *testing.T) {
|
|
testHandler(t, handlerTest)
|
|
})
|
|
|
|
t.Run("buffered", func(t *testing.T) {
|
|
testHandler(t, handlerTest2)
|
|
})
|
|
}
|
|
|
|
// testStreamDeadline will test if server
|
|
func testStreamDeadline(t *testing.T, local, remote *Manager) {
|
|
defer testlogger.T.SetErrorTB(t)()
|
|
errFatal := func(err error) {
|
|
t.Helper()
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
}
|
|
|
|
const wantDL = 50 * time.Millisecond
|
|
// We fake a local and remote server.
|
|
remoteHost := remote.HostName()
|
|
|
|
// 1: Echo
|
|
serverCanceled := make(chan time.Duration, 1)
|
|
register := func(manager *Manager) {
|
|
errFatal(manager.RegisterStreamingHandler(handlerTest, StreamHandler{
|
|
Handle: func(ctx context.Context, payload []byte, request <-chan []byte, resp chan<- []byte) *RemoteErr {
|
|
started := time.Now()
|
|
dl, _ := ctx.Deadline()
|
|
if testing.Verbose() {
|
|
fmt.Println(GetCaller(ctx).Name, "Server deadline:", time.Until(dl))
|
|
}
|
|
<-ctx.Done()
|
|
serverCanceled <- time.Since(started)
|
|
if testing.Verbose() {
|
|
fmt.Println(GetCaller(ctx).Name, "Server Context canceled with", ctx.Err(), "after", time.Since(started))
|
|
}
|
|
return nil
|
|
},
|
|
OutCapacity: 1,
|
|
InCapacity: 0,
|
|
}))
|
|
errFatal(manager.RegisterStreamingHandler(handlerTest2, StreamHandler{
|
|
Handle: func(ctx context.Context, payload []byte, request <-chan []byte, resp chan<- []byte) *RemoteErr {
|
|
started := time.Now()
|
|
dl, _ := ctx.Deadline()
|
|
if testing.Verbose() {
|
|
fmt.Println(GetCaller(ctx).Name, "Server deadline:", time.Until(dl))
|
|
}
|
|
<-ctx.Done()
|
|
serverCanceled <- time.Since(started)
|
|
if testing.Verbose() {
|
|
fmt.Println(GetCaller(ctx).Name, "Server Context canceled with", ctx.Err(), "after", time.Since(started))
|
|
}
|
|
return nil
|
|
},
|
|
OutCapacity: 1,
|
|
InCapacity: 1,
|
|
}))
|
|
}
|
|
register(local)
|
|
register(remote)
|
|
// Double remote DL
|
|
local.debugMsg(debugAddToDeadline, wantDL)
|
|
defer local.debugMsg(debugAddToDeadline, time.Duration(0))
|
|
remote.debugMsg(debugAddToDeadline, wantDL)
|
|
defer remote.debugMsg(debugAddToDeadline, time.Duration(0))
|
|
|
|
testHandler := func(t *testing.T, handler HandlerID) {
|
|
remoteConn := local.Connection(remoteHost)
|
|
const testPayload = "Hello Grid World!"
|
|
|
|
ctx, cancel := context.WithTimeout(context.Background(), wantDL)
|
|
defer cancel()
|
|
st, err := remoteConn.NewStream(ctx, handler, []byte(testPayload))
|
|
errFatal(err)
|
|
clientCanceled := make(chan time.Duration, 1)
|
|
go func() {
|
|
started := time.Now()
|
|
for resp := range st.responses {
|
|
err = resp.Err
|
|
}
|
|
clientCanceled <- time.Since(started)
|
|
}()
|
|
serverEnd := <-serverCanceled
|
|
clientEnd := <-clientCanceled
|
|
t.Log("server cancel time:", serverEnd)
|
|
t.Log("client cancel time:", clientEnd)
|
|
if !errors.Is(err, context.DeadlineExceeded) {
|
|
t.Error("expected context.DeadlineExceeded, got", err)
|
|
}
|
|
}
|
|
// local to remote, unbuffered
|
|
t.Run("unbuffered", func(t *testing.T) {
|
|
testHandler(t, handlerTest)
|
|
})
|
|
|
|
t.Run("buffered", func(t *testing.T) {
|
|
testHandler(t, handlerTest2)
|
|
})
|
|
}
|
|
|
|
func testServerOutCongestion(t *testing.T, local, remote *Manager) {
|
|
defer testlogger.T.SetErrorTB(t)()
|
|
errFatal := func(err error) {
|
|
t.Helper()
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
}
|
|
|
|
// We fake a local and remote server.
|
|
remoteHost := remote.HostName()
|
|
|
|
// 1: Echo
|
|
serverSent := make(chan struct{})
|
|
register := func(manager *Manager) {
|
|
errFatal(manager.RegisterStreamingHandler(handlerTest, StreamHandler{
|
|
Handle: func(ctx context.Context, payload []byte, request <-chan []byte, resp chan<- []byte) *RemoteErr {
|
|
// Send many responses.
|
|
// Test that this doesn't block.
|
|
for i := byte(0); i < 100; i++ {
|
|
select {
|
|
case resp <- []byte{i}:
|
|
// ok
|
|
case <-ctx.Done():
|
|
return NewRemoteErr(ctx.Err())
|
|
}
|
|
if i == 0 {
|
|
close(serverSent)
|
|
}
|
|
}
|
|
return nil
|
|
},
|
|
OutCapacity: 1,
|
|
InCapacity: 0,
|
|
}))
|
|
errFatal(manager.RegisterSingleHandler(handlerTest2, func(payload []byte) ([]byte, *RemoteErr) {
|
|
// Simple roundtrip
|
|
return append([]byte{}, payload...), nil
|
|
}))
|
|
}
|
|
register(local)
|
|
register(remote)
|
|
|
|
remoteConn := local.Connection(remoteHost)
|
|
const testPayload = "Hello Grid World!"
|
|
|
|
ctx, cancel := context.WithTimeout(context.Background(), time.Minute)
|
|
defer cancel()
|
|
st, err := remoteConn.NewStream(ctx, handlerTest, []byte(testPayload))
|
|
errFatal(err)
|
|
|
|
// Wait for the server to send the first response.
|
|
<-serverSent
|
|
|
|
// Now do 100 other requests to ensure that the server doesn't block.
|
|
for i := 0; i < 100; i++ {
|
|
_, err := remoteConn.Request(ctx, handlerTest2, []byte(testPayload))
|
|
errFatal(err)
|
|
}
|
|
// Drain responses
|
|
got := 0
|
|
for resp := range st.responses {
|
|
// t.Log("got response", resp)
|
|
errFatal(resp.Err)
|
|
if resp.Msg[0] != byte(got) {
|
|
t.Error("expected response", got, "got", resp.Msg[0])
|
|
}
|
|
got++
|
|
}
|
|
if got != 100 {
|
|
t.Error("expected 100 responses, got", got)
|
|
}
|
|
}
|
|
|
|
func testServerInCongestion(t *testing.T, local, remote *Manager) {
|
|
defer testlogger.T.SetErrorTB(t)()
|
|
errFatal := func(err error) {
|
|
t.Helper()
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
}
|
|
|
|
// We fake a local and remote server.
|
|
remoteHost := remote.HostName()
|
|
|
|
// 1: Echo
|
|
processHandler := make(chan struct{})
|
|
register := func(manager *Manager) {
|
|
errFatal(manager.RegisterStreamingHandler(handlerTest, StreamHandler{
|
|
Handle: func(ctx context.Context, payload []byte, request <-chan []byte, resp chan<- []byte) *RemoteErr {
|
|
// Block incoming requests.
|
|
var n byte
|
|
<-processHandler
|
|
for {
|
|
select {
|
|
case in, ok := <-request:
|
|
if !ok {
|
|
return nil
|
|
}
|
|
if in[0] != n {
|
|
return NewRemoteErrString(fmt.Sprintf("expected incoming %d, got %d", n, in[0]))
|
|
}
|
|
n++
|
|
resp <- append([]byte{}, in...)
|
|
case <-ctx.Done():
|
|
return NewRemoteErr(ctx.Err())
|
|
}
|
|
}
|
|
},
|
|
OutCapacity: 5,
|
|
InCapacity: 5,
|
|
}))
|
|
errFatal(manager.RegisterSingleHandler(handlerTest2, func(payload []byte) ([]byte, *RemoteErr) {
|
|
// Simple roundtrip
|
|
return append([]byte{}, payload...), nil
|
|
}))
|
|
}
|
|
register(local)
|
|
register(remote)
|
|
|
|
remoteConn := local.Connection(remoteHost)
|
|
const testPayload = "Hello Grid World!"
|
|
|
|
ctx, cancel := context.WithTimeout(context.Background(), time.Minute)
|
|
defer cancel()
|
|
st, err := remoteConn.NewStream(ctx, handlerTest, []byte(testPayload))
|
|
errFatal(err)
|
|
|
|
// Start sending requests.
|
|
go func() {
|
|
for i := byte(0); i < 100; i++ {
|
|
st.Requests <- []byte{i}
|
|
}
|
|
close(st.Requests)
|
|
}()
|
|
// Now do 100 other requests to ensure that the server doesn't block.
|
|
for i := 0; i < 100; i++ {
|
|
_, err := remoteConn.Request(ctx, handlerTest2, []byte(testPayload))
|
|
errFatal(err)
|
|
}
|
|
// Start processing requests.
|
|
close(processHandler)
|
|
|
|
// Drain responses
|
|
got := 0
|
|
for resp := range st.responses {
|
|
// t.Log("got response", resp)
|
|
errFatal(resp.Err)
|
|
if resp.Msg[0] != byte(got) {
|
|
t.Error("expected response", got, "got", resp.Msg[0])
|
|
}
|
|
got++
|
|
}
|
|
if got != 100 {
|
|
t.Error("expected 100 responses, got", got)
|
|
}
|
|
}
|
|
|
|
func testGenericsStreamRoundtrip(t *testing.T, local, remote *Manager) {
|
|
defer testlogger.T.SetErrorTB(t)()
|
|
defer timeout(5 * time.Second)()
|
|
errFatal := func(err error) {
|
|
t.Helper()
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
}
|
|
|
|
// We fake a local and remote server.
|
|
remoteHost := remote.HostName()
|
|
handler := NewStream[*testRequest, *testRequest, *testResponse](handlerTest, newTestRequest, newTestRequest, newTestResponse)
|
|
handler.InCapacity = 1
|
|
handler.OutCapacity = 1
|
|
const payloads = 10
|
|
|
|
// 1: Echo
|
|
register := func(manager *Manager) {
|
|
errFatal(handler.Register(manager, func(ctx context.Context, pp *testRequest, in <-chan *testRequest, out chan<- *testResponse) *RemoteErr {
|
|
n := 0
|
|
for i := range in {
|
|
if n > payloads {
|
|
panic("too many requests")
|
|
}
|
|
|
|
// t.Log("Got request:", *i)
|
|
out <- &testResponse{
|
|
OrgNum: i.Num + pp.Num,
|
|
OrgString: pp.String + i.String,
|
|
Embedded: *i,
|
|
}
|
|
n++
|
|
}
|
|
return nil
|
|
}))
|
|
}
|
|
register(local)
|
|
register(remote)
|
|
|
|
// local to remote
|
|
remoteConn := local.Connection(remoteHost)
|
|
const testPayload = "Hello Grid World!"
|
|
|
|
start := time.Now()
|
|
stream, err := handler.Call(context.Background(), remoteConn, &testRequest{Num: 1, String: testPayload})
|
|
errFatal(err)
|
|
go func() {
|
|
defer close(stream.Requests)
|
|
for i := 0; i < payloads; i++ {
|
|
// t.Log("sending new client request")
|
|
stream.Requests <- &testRequest{Num: i, String: testPayload}
|
|
}
|
|
}()
|
|
var n int
|
|
err = stream.Results(func(resp *testResponse) error {
|
|
const wantString = testPayload + testPayload
|
|
if resp.OrgString != testPayload+testPayload {
|
|
t.Errorf("want %q, got %q", wantString, resp.OrgString)
|
|
}
|
|
if resp.OrgNum != n+1 {
|
|
t.Errorf("want %d, got %d", n+1, resp.OrgNum)
|
|
}
|
|
handler.PutResponse(resp)
|
|
n++
|
|
return nil
|
|
})
|
|
errFatal(err)
|
|
t.Log("EOF.", payloads, " Roundtrips:", time.Since(start))
|
|
}
|
|
|
|
func testGenericsStreamRoundtripSubroute(t *testing.T, local, remote *Manager) {
|
|
defer testlogger.T.SetErrorTB(t)()
|
|
defer timeout(5 * time.Second)()
|
|
errFatal := func(err error) {
|
|
t.Helper()
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
}
|
|
|
|
// We fake a local and remote server.
|
|
remoteHost := remote.HostName()
|
|
handler := NewStream[*testRequest, *testRequest, *testResponse](handlerTest, newTestRequest, newTestRequest, newTestResponse)
|
|
handler.InCapacity = 1
|
|
handler.OutCapacity = 1
|
|
const payloads = 10
|
|
|
|
// 1: Echo
|
|
register := func(manager *Manager) {
|
|
errFatal(handler.Register(manager, func(ctx context.Context, pp *testRequest, in <-chan *testRequest, out chan<- *testResponse) *RemoteErr {
|
|
sub := GetSubroute(ctx)
|
|
if sub != "subroute/1" {
|
|
t.Fatal("expected subroute/1, got", sub)
|
|
}
|
|
n := 0
|
|
for i := range in {
|
|
if n > payloads {
|
|
panic("too many requests")
|
|
}
|
|
|
|
// t.Log("Got request:", *i)
|
|
out <- &testResponse{
|
|
OrgNum: i.Num + pp.Num,
|
|
OrgString: pp.String + i.String,
|
|
Embedded: *i,
|
|
}
|
|
n++
|
|
}
|
|
return nil
|
|
}, "subroute", "1"))
|
|
}
|
|
register(local)
|
|
register(remote)
|
|
|
|
// local to remote
|
|
remoteConn := local.Connection(remoteHost)
|
|
const testPayload = "Hello Grid World!"
|
|
// Add subroute
|
|
remoteSub := remoteConn.Subroute(strings.Join([]string{"subroute", "1"}, "/"))
|
|
|
|
start := time.Now()
|
|
stream, err := handler.Call(context.Background(), remoteSub, &testRequest{Num: 1, String: testPayload})
|
|
errFatal(err)
|
|
go func() {
|
|
defer close(stream.Requests)
|
|
for i := 0; i < payloads; i++ {
|
|
// t.Log("sending new client request")
|
|
stream.Requests <- &testRequest{Num: i, String: testPayload}
|
|
}
|
|
}()
|
|
var n int
|
|
err = stream.Results(func(resp *testResponse) error {
|
|
// t.Logf("got resp: %+v", *resp.Msg)
|
|
const wantString = testPayload + testPayload
|
|
if resp.OrgString != testPayload+testPayload {
|
|
t.Errorf("want %q, got %q", wantString, resp.OrgString)
|
|
}
|
|
if resp.OrgNum != n+1 {
|
|
t.Errorf("want %d, got %d", n+1, resp.OrgNum)
|
|
}
|
|
handler.PutResponse(resp)
|
|
n++
|
|
return nil
|
|
})
|
|
|
|
errFatal(err)
|
|
t.Log("EOF.", payloads, " Roundtrips:", time.Since(start))
|
|
}
|
|
|
|
// testServerStreamResponseBlocked will test if server can handle a blocked response stream
|
|
func testServerStreamResponseBlocked(t *testing.T, local, remote *Manager) {
|
|
defer testlogger.T.SetErrorTB(t)()
|
|
errFatal := func(err error) {
|
|
t.Helper()
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
}
|
|
|
|
// We fake a local and remote server.
|
|
remoteHost := remote.HostName()
|
|
|
|
// 1: Echo
|
|
serverSent := make(chan struct{})
|
|
serverCanceled := make(chan struct{})
|
|
register := func(manager *Manager) {
|
|
errFatal(manager.RegisterStreamingHandler(handlerTest, StreamHandler{
|
|
Handle: func(ctx context.Context, payload []byte, _ <-chan []byte, resp chan<- []byte) *RemoteErr {
|
|
// Send many responses.
|
|
// Test that this doesn't block.
|
|
for i := byte(0); i < 100; i++ {
|
|
select {
|
|
case resp <- []byte{i}:
|
|
// ok
|
|
case <-ctx.Done():
|
|
close(serverCanceled)
|
|
return NewRemoteErr(ctx.Err())
|
|
}
|
|
if i == 1 {
|
|
close(serverSent)
|
|
}
|
|
}
|
|
return nil
|
|
},
|
|
OutCapacity: 1,
|
|
InCapacity: 0,
|
|
}))
|
|
}
|
|
register(local)
|
|
register(remote)
|
|
|
|
remoteConn := local.Connection(remoteHost)
|
|
const testPayload = "Hello Grid World!"
|
|
|
|
ctx, cancel := context.WithTimeout(context.Background(), 5*time.Second)
|
|
|
|
st, err := remoteConn.NewStream(ctx, handlerTest, []byte(testPayload))
|
|
errFatal(err)
|
|
|
|
// Wait for the server to send the first response.
|
|
<-serverSent
|
|
|
|
// Read back from the stream and block.
|
|
nowBlocking := make(chan struct{})
|
|
stopBlocking := make(chan struct{})
|
|
defer close(stopBlocking)
|
|
go func() {
|
|
st.Results(func(b []byte) error {
|
|
close(nowBlocking)
|
|
// Block until test is done.
|
|
<-stopBlocking
|
|
return nil
|
|
})
|
|
}()
|
|
|
|
<-nowBlocking
|
|
// Wait for the receiver channel to fill.
|
|
for len(st.responses) != cap(st.responses) {
|
|
time.Sleep(time.Millisecond)
|
|
}
|
|
cancel()
|
|
<-serverCanceled
|
|
local.debugMsg(debugIsOutgoingClosed, st.muxID, func(closed bool) {
|
|
if !closed {
|
|
t.Error("expected outgoing closed")
|
|
} else {
|
|
t.Log("outgoing was closed")
|
|
}
|
|
})
|
|
|
|
// Drain responses and check if error propagated.
|
|
err = st.Results(func(b []byte) error {
|
|
return nil
|
|
})
|
|
if !errors.Is(err, context.Canceled) {
|
|
t.Error("expected context.Canceled, got", err)
|
|
}
|
|
}
|
|
|
|
func timeout(after time.Duration) (cancel func()) {
|
|
c := time.After(after)
|
|
cc := make(chan struct{})
|
|
go func() {
|
|
select {
|
|
case <-cc:
|
|
return
|
|
case <-c:
|
|
buf := make([]byte, 1<<20)
|
|
stacklen := runtime.Stack(buf, true)
|
|
fmt.Printf("=== Timeout, assuming deadlock ===\n*** goroutine dump...\n%s\n*** end\n", string(buf[:stacklen]))
|
|
os.Exit(2)
|
|
}
|
|
}()
|
|
return func() {
|
|
close(cc)
|
|
}
|
|
}
|
|
|
|
func assertNoActive(t *testing.T, c *Connection) {
|
|
t.Helper()
|
|
// Tiny bit racy for tests, but we try to play nice.
|
|
for i := 10; i >= 0; i-- {
|
|
runtime.Gosched()
|
|
stats := c.Stats()
|
|
if stats.IncomingStreams != 0 {
|
|
if i > 0 {
|
|
time.Sleep(100 * time.Millisecond)
|
|
continue
|
|
}
|
|
var found []uint64
|
|
c.inStream.Range(func(key uint64, value *muxServer) bool {
|
|
found = append(found, key)
|
|
return true
|
|
})
|
|
t.Errorf("expected no active streams, got %d incoming: %v", stats.IncomingStreams, found)
|
|
}
|
|
if stats.OutgoingStreams != 0 {
|
|
if i > 0 {
|
|
time.Sleep(100 * time.Millisecond)
|
|
continue
|
|
}
|
|
var found []uint64
|
|
c.outgoing.Range(func(key uint64, value *muxClient) bool {
|
|
found = append(found, key)
|
|
return true
|
|
})
|
|
t.Errorf("expected no active streams, got %d outgoing: %v", stats.OutgoingStreams, found)
|
|
}
|
|
return
|
|
}
|
|
}
|
|
|
|
// Inserted manually.
|
|
func _() {
|
|
// An "invalid array index" compiler error signifies that the constant values have changed.
|
|
// Re-run the stringer command to generate them again.
|
|
var x [1]struct{}
|
|
_ = x[StateUnconnected-0]
|
|
_ = x[StateConnecting-1]
|
|
_ = x[StateConnected-2]
|
|
_ = x[StateConnectionError-3]
|
|
_ = x[StateShutdown-4]
|
|
}
|
|
|
|
const stateName = "UnconnectedConnectingConnectedConnectionErrorShutdown"
|
|
|
|
var stateIndex = [...]uint8{0, 11, 21, 30, 45, 53}
|
|
|
|
func (i State) String() string {
|
|
if i >= State(len(stateIndex)-1) {
|
|
return "State(" + strconv.FormatInt(int64(i), 10) + ")"
|
|
}
|
|
return stateName[stateIndex[i]:stateIndex[i+1]]
|
|
}
|