mirror of
https://source.quilibrium.com/quilibrium/ceremonyclient.git
synced 2024-12-26 16:45:18 +00:00
571 lines
16 KiB
Go
571 lines
16 KiB
Go
package swarm_test
|
|
|
|
import (
|
|
"bytes"
|
|
"context"
|
|
"errors"
|
|
"fmt"
|
|
"io"
|
|
"strings"
|
|
"sync"
|
|
"testing"
|
|
"time"
|
|
|
|
"github.com/libp2p/go-libp2p/core/control"
|
|
"github.com/libp2p/go-libp2p/core/network"
|
|
mocknetwork "github.com/libp2p/go-libp2p/core/network/mocks"
|
|
"github.com/libp2p/go-libp2p/core/peer"
|
|
"github.com/libp2p/go-libp2p/core/peerstore"
|
|
"github.com/libp2p/go-libp2p/core/protocol"
|
|
"github.com/libp2p/go-libp2p/core/test"
|
|
"github.com/libp2p/go-libp2p/p2p/net/swarm"
|
|
. "github.com/libp2p/go-libp2p/p2p/net/swarm/testing"
|
|
|
|
logging "github.com/ipfs/go-log/v2"
|
|
ma "github.com/multiformats/go-multiaddr"
|
|
manet "github.com/multiformats/go-multiaddr/net"
|
|
"github.com/stretchr/testify/require"
|
|
"go.uber.org/mock/gomock"
|
|
)
|
|
|
|
var log = logging.Logger("swarm_test")
|
|
|
|
func EchoStreamHandler(stream network.Stream) {
|
|
go func() {
|
|
defer stream.Close()
|
|
|
|
// pull out the ipfs conn
|
|
c := stream.Conn()
|
|
log.Infof("%s ponging to %s", c.LocalPeer(), c.RemotePeer())
|
|
|
|
buf := make([]byte, 4)
|
|
|
|
for {
|
|
if _, err := stream.Read(buf); err != nil {
|
|
if err != io.EOF {
|
|
log.Error("ping receive error:", err)
|
|
}
|
|
return
|
|
}
|
|
|
|
if !bytes.Equal(buf, []byte("ping")) {
|
|
log.Errorf("ping receive error: ping != %s %v", buf, buf)
|
|
return
|
|
}
|
|
|
|
if _, err := stream.Write([]byte("pong")); err != nil {
|
|
log.Error("pond send error:", err)
|
|
return
|
|
}
|
|
}
|
|
}()
|
|
}
|
|
|
|
func makeDialOnlySwarm(t *testing.T) *swarm.Swarm {
|
|
swarm := GenSwarm(t, OptDialOnly)
|
|
swarm.SetStreamHandler(EchoStreamHandler)
|
|
return swarm
|
|
}
|
|
|
|
func makeSwarms(t *testing.T, num int, opts ...Option) []*swarm.Swarm {
|
|
swarms := make([]*swarm.Swarm, 0, num)
|
|
for i := 0; i < num; i++ {
|
|
swarm := GenSwarm(t, opts...)
|
|
swarm.SetStreamHandler(EchoStreamHandler)
|
|
swarms = append(swarms, swarm)
|
|
}
|
|
return swarms
|
|
}
|
|
|
|
func connectSwarms(t *testing.T, ctx context.Context, swarms []*swarm.Swarm) {
|
|
var wg sync.WaitGroup
|
|
connect := func(s *swarm.Swarm, dst peer.ID, addr ma.Multiaddr) {
|
|
// TODO: make a DialAddr func.
|
|
s.Peerstore().AddAddr(dst, addr, peerstore.PermanentAddrTTL)
|
|
if _, err := s.DialPeer(ctx, dst); err != nil {
|
|
t.Fatal("error swarm dialing to peer", err)
|
|
}
|
|
wg.Done()
|
|
}
|
|
|
|
log.Info("Connecting swarms simultaneously.")
|
|
for i, s1 := range swarms {
|
|
for _, s2 := range swarms[i+1:] {
|
|
wg.Add(1)
|
|
connect(s1, s2.LocalPeer(), s2.ListenAddresses()[0]) // try the first.
|
|
}
|
|
}
|
|
wg.Wait()
|
|
|
|
for _, s := range swarms {
|
|
log.Infof("%s swarm routing table: %s", s.LocalPeer(), s.Peers())
|
|
}
|
|
}
|
|
|
|
func subtestSwarm(t *testing.T, SwarmNum int, MsgNum int) {
|
|
swarms := makeSwarms(t, SwarmNum, OptDisableReuseport)
|
|
|
|
// connect everyone
|
|
connectSwarms(t, context.Background(), swarms)
|
|
|
|
// ping/pong
|
|
for _, s1 := range swarms {
|
|
log.Debugf("-------------------------------------------------------")
|
|
log.Debugf("%s ping pong round", s1.LocalPeer())
|
|
log.Debugf("-------------------------------------------------------")
|
|
|
|
_, cancel := context.WithCancel(context.Background())
|
|
got := map[peer.ID]int{}
|
|
errChan := make(chan error, MsgNum*len(swarms))
|
|
streamChan := make(chan network.Stream, MsgNum)
|
|
|
|
// send out "ping" x MsgNum to every peer
|
|
go func() {
|
|
defer close(streamChan)
|
|
|
|
var wg sync.WaitGroup
|
|
send := func(p peer.ID) {
|
|
defer wg.Done()
|
|
|
|
// first, one stream per peer (nice)
|
|
stream, err := s1.NewStream(context.Background(), p)
|
|
if err != nil {
|
|
errChan <- err
|
|
return
|
|
}
|
|
|
|
// send out ping!
|
|
for k := 0; k < MsgNum; k++ { // with k messages
|
|
msg := "ping"
|
|
log.Debugf("%s %s %s (%d)", s1.LocalPeer(), msg, p, k)
|
|
if _, err := stream.Write([]byte(msg)); err != nil {
|
|
errChan <- err
|
|
continue
|
|
}
|
|
}
|
|
|
|
// read it later
|
|
streamChan <- stream
|
|
}
|
|
|
|
for _, s2 := range swarms {
|
|
if s2.LocalPeer() == s1.LocalPeer() {
|
|
continue // dont send to self...
|
|
}
|
|
|
|
wg.Add(1)
|
|
go send(s2.LocalPeer())
|
|
}
|
|
wg.Wait()
|
|
}()
|
|
|
|
// receive "pong" x MsgNum from every peer
|
|
go func() {
|
|
defer close(errChan)
|
|
count := 0
|
|
countShouldBe := MsgNum * (len(swarms) - 1)
|
|
for stream := range streamChan { // one per peer
|
|
// get peer on the other side
|
|
p := stream.Conn().RemotePeer()
|
|
|
|
// receive pings
|
|
msgCount := 0
|
|
msg := make([]byte, 4)
|
|
for k := 0; k < MsgNum; k++ { // with k messages
|
|
|
|
// read from the stream
|
|
if _, err := stream.Read(msg); err != nil {
|
|
errChan <- err
|
|
continue
|
|
}
|
|
|
|
if string(msg) != "pong" {
|
|
errChan <- fmt.Errorf("unexpected message: %s", msg)
|
|
continue
|
|
}
|
|
|
|
log.Debugf("%s %s %s (%d)", s1.LocalPeer(), msg, p, k)
|
|
msgCount++
|
|
}
|
|
|
|
got[p] = msgCount
|
|
count += msgCount
|
|
stream.Close()
|
|
}
|
|
|
|
if count != countShouldBe {
|
|
errChan <- fmt.Errorf("count mismatch: %d != %d", count, countShouldBe)
|
|
}
|
|
}()
|
|
|
|
// check any errors (blocks till consumer is done)
|
|
for err := range errChan {
|
|
if err != nil {
|
|
t.Error(err.Error())
|
|
}
|
|
}
|
|
|
|
log.Debugf("%s got pongs", s1.LocalPeer())
|
|
if (len(swarms) - 1) != len(got) {
|
|
t.Errorf("got (%d) less messages than sent (%d).", len(got), len(swarms))
|
|
}
|
|
|
|
for p, n := range got {
|
|
if n != MsgNum {
|
|
t.Error("peer did not get all msgs", p, n, "/", MsgNum)
|
|
}
|
|
}
|
|
|
|
cancel()
|
|
<-time.After(10 * time.Millisecond)
|
|
}
|
|
}
|
|
|
|
func TestSwarm(t *testing.T) {
|
|
t.Parallel()
|
|
subtestSwarm(t, 5, 100)
|
|
}
|
|
|
|
func TestBasicSwarm(t *testing.T) {
|
|
// t.Skip("skipping for another test")
|
|
t.Parallel()
|
|
subtestSwarm(t, 2, 1)
|
|
}
|
|
|
|
func TestConnectionGating(t *testing.T) {
|
|
ctx := context.Background()
|
|
tcs := map[string]struct {
|
|
p1Gater func(gater *MockConnectionGater) *MockConnectionGater
|
|
p2Gater func(gater *MockConnectionGater) *MockConnectionGater
|
|
|
|
p1ConnectednessToP2 network.Connectedness
|
|
p2ConnectednessToP1 network.Connectedness
|
|
isP1OutboundErr bool
|
|
disableOnQUIC bool
|
|
}{
|
|
"no gating": {
|
|
p1ConnectednessToP2: network.Connected,
|
|
p2ConnectednessToP1: network.Connected,
|
|
isP1OutboundErr: false,
|
|
},
|
|
"p1 gates outbound peer dial": {
|
|
p1Gater: func(c *MockConnectionGater) *MockConnectionGater {
|
|
c.PeerDial = func(p peer.ID) bool { return false }
|
|
return c
|
|
},
|
|
p1ConnectednessToP2: network.NotConnected,
|
|
p2ConnectednessToP1: network.NotConnected,
|
|
isP1OutboundErr: true,
|
|
},
|
|
"p1 gates outbound addr dialing": {
|
|
p1Gater: func(c *MockConnectionGater) *MockConnectionGater {
|
|
c.Dial = func(p peer.ID, addr ma.Multiaddr) bool { return false }
|
|
return c
|
|
},
|
|
p1ConnectednessToP2: network.NotConnected,
|
|
p2ConnectednessToP1: network.NotConnected,
|
|
isP1OutboundErr: true,
|
|
},
|
|
"p2 accepts inbound peer dial if outgoing dial is gated": {
|
|
p2Gater: func(c *MockConnectionGater) *MockConnectionGater {
|
|
c.Dial = func(peer.ID, ma.Multiaddr) bool { return false }
|
|
return c
|
|
},
|
|
p1ConnectednessToP2: network.Connected,
|
|
p2ConnectednessToP1: network.Connected,
|
|
isP1OutboundErr: false,
|
|
},
|
|
"p2 gates inbound peer dial before securing": {
|
|
p2Gater: func(c *MockConnectionGater) *MockConnectionGater {
|
|
c.Accept = func(c network.ConnMultiaddrs) bool { return false }
|
|
return c
|
|
},
|
|
p1ConnectednessToP2: network.NotConnected,
|
|
p2ConnectednessToP1: network.NotConnected,
|
|
isP1OutboundErr: true,
|
|
// QUIC gates the connection after completion of the handshake
|
|
disableOnQUIC: true,
|
|
},
|
|
"p2 gates inbound peer dial before multiplexing": {
|
|
p1Gater: func(c *MockConnectionGater) *MockConnectionGater {
|
|
c.Secured = func(network.Direction, peer.ID, network.ConnMultiaddrs) bool { return false }
|
|
return c
|
|
},
|
|
p1ConnectednessToP2: network.NotConnected,
|
|
p2ConnectednessToP1: network.NotConnected,
|
|
isP1OutboundErr: true,
|
|
},
|
|
"p2 gates inbound peer dial after upgrading": {
|
|
p1Gater: func(c *MockConnectionGater) *MockConnectionGater {
|
|
c.Upgraded = func(c network.Conn) (bool, control.DisconnectReason) { return false, 0 }
|
|
return c
|
|
},
|
|
p1ConnectednessToP2: network.NotConnected,
|
|
p2ConnectednessToP1: network.NotConnected,
|
|
isP1OutboundErr: true,
|
|
},
|
|
"p2 gates outbound dials": {
|
|
p2Gater: func(c *MockConnectionGater) *MockConnectionGater {
|
|
c.PeerDial = func(p peer.ID) bool { return false }
|
|
return c
|
|
},
|
|
p1ConnectednessToP2: network.Connected,
|
|
p2ConnectednessToP1: network.Connected,
|
|
isP1OutboundErr: false,
|
|
},
|
|
}
|
|
|
|
for n, tc := range tcs {
|
|
for _, useQuic := range []bool{false, true} {
|
|
trString := "TCP"
|
|
optTransport := OptDisableQUIC
|
|
if useQuic {
|
|
if tc.disableOnQUIC {
|
|
continue
|
|
}
|
|
trString = "QUIC"
|
|
optTransport = OptDisableTCP
|
|
}
|
|
t.Run(fmt.Sprintf("%s %s", n, trString), func(t *testing.T) {
|
|
p1Gater := DefaultMockConnectionGater()
|
|
p2Gater := DefaultMockConnectionGater()
|
|
if tc.p1Gater != nil {
|
|
p1Gater = tc.p1Gater(p1Gater)
|
|
}
|
|
if tc.p2Gater != nil {
|
|
p2Gater = tc.p2Gater(p2Gater)
|
|
}
|
|
|
|
sw1 := GenSwarm(t, OptConnGater(p1Gater), optTransport)
|
|
sw2 := GenSwarm(t, OptConnGater(p2Gater), optTransport)
|
|
|
|
p1 := sw1.LocalPeer()
|
|
p2 := sw2.LocalPeer()
|
|
sw1.Peerstore().AddAddr(p2, sw2.ListenAddresses()[0], peerstore.PermanentAddrTTL)
|
|
// 1 -> 2
|
|
_, err := sw1.DialPeer(ctx, p2)
|
|
|
|
require.Equal(t, tc.isP1OutboundErr, err != nil, n)
|
|
require.Equal(t, tc.p1ConnectednessToP2, sw1.Connectedness(p2), n)
|
|
|
|
require.Eventually(t, func() bool {
|
|
return tc.p2ConnectednessToP1 == sw2.Connectedness(p1)
|
|
}, 2*time.Second, 100*time.Millisecond, n)
|
|
})
|
|
}
|
|
}
|
|
}
|
|
|
|
func TestNoDial(t *testing.T) {
|
|
swarms := makeSwarms(t, 2)
|
|
|
|
_, err := swarms[0].NewStream(network.WithNoDial(context.Background(), "swarm test"), swarms[1].LocalPeer())
|
|
if err != network.ErrNoConn {
|
|
t.Fatal("should have failed with ErrNoConn")
|
|
}
|
|
}
|
|
|
|
func TestCloseWithOpenStreams(t *testing.T) {
|
|
ctx := context.Background()
|
|
swarms := makeSwarms(t, 2)
|
|
connectSwarms(t, ctx, swarms)
|
|
|
|
s, err := swarms[0].NewStream(ctx, swarms[1].LocalPeer())
|
|
require.NoError(t, err)
|
|
defer s.Close()
|
|
// close swarm before stream.
|
|
require.NoError(t, swarms[0].Close())
|
|
}
|
|
|
|
func TestTypedNilConn(t *testing.T) {
|
|
s := GenSwarm(t)
|
|
defer s.Close()
|
|
|
|
// We can't dial ourselves.
|
|
c, err := s.DialPeer(context.Background(), s.LocalPeer())
|
|
require.Error(t, err)
|
|
// If we fail to dial, the connection should be nil.
|
|
require.Nil(t, c)
|
|
}
|
|
|
|
func TestPreventDialListenAddr(t *testing.T) {
|
|
s := GenSwarm(t, OptDialOnly)
|
|
if err := s.Listen(ma.StringCast("/ip4/0.0.0.0/udp/0/quic-v1")); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
addrs, err := s.InterfaceListenAddresses()
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
var addr ma.Multiaddr
|
|
for _, a := range addrs {
|
|
_, s, err := manet.DialArgs(a)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if strings.Split(s, ":")[0] == "127.0.0.1" {
|
|
addr = a
|
|
break
|
|
}
|
|
}
|
|
remote := test.RandPeerIDFatal(t)
|
|
s.Peerstore().AddAddr(remote, addr, time.Hour)
|
|
_, err = s.DialPeer(context.Background(), remote)
|
|
if !errors.Is(err, swarm.ErrNoGoodAddresses) {
|
|
t.Fatal("expected dial to fail: %w", err)
|
|
}
|
|
}
|
|
|
|
func TestStreamCount(t *testing.T) {
|
|
s1 := GenSwarm(t)
|
|
s2 := GenSwarm(t)
|
|
connectSwarms(t, context.Background(), []*swarm.Swarm{s2, s1})
|
|
|
|
countStreams := func() (n int) {
|
|
var num int
|
|
for _, c := range s1.ConnsToPeer(s2.LocalPeer()) {
|
|
n += c.Stat().NumStreams
|
|
num += len(c.GetStreams())
|
|
}
|
|
require.Equal(t, n, num, "inconsistent stream count")
|
|
return
|
|
}
|
|
|
|
streams := make(chan network.Stream, 20)
|
|
streamAccepted := make(chan struct{}, 1)
|
|
s1.SetStreamHandler(func(str network.Stream) {
|
|
streams <- str
|
|
streamAccepted <- struct{}{}
|
|
})
|
|
|
|
for i := 0; i < 10; i++ {
|
|
str, err := s2.NewStream(context.Background(), s1.LocalPeer())
|
|
require.NoError(t, err)
|
|
str.Write([]byte("foobar"))
|
|
<-streamAccepted
|
|
}
|
|
require.Eventually(t, func() bool { return len(streams) == 10 }, 5*time.Second, 10*time.Millisecond)
|
|
require.Equal(t, 10, countStreams())
|
|
(<-streams).Reset()
|
|
(<-streams).Close()
|
|
require.Equal(t, 8, countStreams())
|
|
|
|
str, err := s1.NewStream(context.Background(), s2.LocalPeer())
|
|
require.NoError(t, err)
|
|
require.Equal(t, 9, countStreams())
|
|
str.Close()
|
|
require.Equal(t, 8, countStreams())
|
|
}
|
|
|
|
func TestResourceManager(t *testing.T) {
|
|
ctrl := gomock.NewController(t)
|
|
defer ctrl.Finish()
|
|
|
|
rcmgr1 := mocknetwork.NewMockResourceManager(ctrl)
|
|
s1 := GenSwarm(t, WithSwarmOpts(swarm.WithResourceManager(rcmgr1)))
|
|
defer s1.Close()
|
|
|
|
rcmgr2 := mocknetwork.NewMockResourceManager(ctrl)
|
|
s2 := GenSwarm(t, WithSwarmOpts(swarm.WithResourceManager(rcmgr2)))
|
|
defer s2.Close()
|
|
connectSwarms(t, context.Background(), []*swarm.Swarm{s1, s2})
|
|
|
|
strChan := make(chan network.Stream)
|
|
s2.SetStreamHandler(func(str network.Stream) { strChan <- str })
|
|
|
|
streamScope1 := mocknetwork.NewMockStreamManagementScope(ctrl)
|
|
rcmgr1.EXPECT().OpenStream(s2.LocalPeer(), network.DirOutbound).Return(streamScope1, nil)
|
|
streamScope2 := mocknetwork.NewMockStreamManagementScope(ctrl)
|
|
rcmgr2.EXPECT().OpenStream(s1.LocalPeer(), network.DirInbound).Return(streamScope2, nil)
|
|
str, err := s1.NewStream(context.Background(), s2.LocalPeer())
|
|
require.NoError(t, err)
|
|
defer str.Close()
|
|
str.Write([]byte("foobar"))
|
|
|
|
p := protocol.ID("proto")
|
|
streamScope1.EXPECT().SetProtocol(p)
|
|
require.NoError(t, str.SetProtocol(p))
|
|
|
|
sstr := <-strChan
|
|
streamScope2.EXPECT().Done()
|
|
require.NoError(t, sstr.Close())
|
|
streamScope1.EXPECT().Done()
|
|
}
|
|
|
|
func TestResourceManagerNewStream(t *testing.T) {
|
|
ctrl := gomock.NewController(t)
|
|
defer ctrl.Finish()
|
|
|
|
rcmgr1 := mocknetwork.NewMockResourceManager(ctrl)
|
|
s1 := GenSwarm(t, WithSwarmOpts(swarm.WithResourceManager(rcmgr1)))
|
|
defer s1.Close()
|
|
|
|
s2 := GenSwarm(t)
|
|
defer s2.Close()
|
|
|
|
connectSwarms(t, context.Background(), []*swarm.Swarm{s1, s2})
|
|
|
|
rerr := errors.New("denied")
|
|
rcmgr1.EXPECT().OpenStream(s2.LocalPeer(), network.DirOutbound).Return(nil, rerr)
|
|
_, err := s1.NewStream(context.Background(), s2.LocalPeer())
|
|
require.ErrorIs(t, err, rerr)
|
|
}
|
|
|
|
func TestResourceManagerAcceptStream(t *testing.T) {
|
|
ctrl := gomock.NewController(t)
|
|
defer ctrl.Finish()
|
|
|
|
rcmgr1 := mocknetwork.NewMockResourceManager(ctrl)
|
|
s1 := GenSwarm(t, WithSwarmOpts(swarm.WithResourceManager(rcmgr1)))
|
|
defer s1.Close()
|
|
|
|
rcmgr2 := mocknetwork.NewMockResourceManager(ctrl)
|
|
s2 := GenSwarm(t, WithSwarmOpts(swarm.WithResourceManager(rcmgr2)))
|
|
defer s2.Close()
|
|
s2.SetStreamHandler(func(str network.Stream) { t.Fatal("didn't expect to accept a stream") })
|
|
|
|
connectSwarms(t, context.Background(), []*swarm.Swarm{s1, s2})
|
|
|
|
streamScope := mocknetwork.NewMockStreamManagementScope(ctrl)
|
|
rcmgr1.EXPECT().OpenStream(s2.LocalPeer(), network.DirOutbound).Return(streamScope, nil)
|
|
streamScope.EXPECT().Done()
|
|
rcmgr2.EXPECT().OpenStream(s1.LocalPeer(), network.DirInbound).Return(nil, errors.New("nope"))
|
|
str, err := s1.NewStream(context.Background(), s2.LocalPeer())
|
|
require.NoError(t, err)
|
|
// The peer's resource manager is blocking any new stream.
|
|
// Depending on how quickly we receive the stream reset, it surfaces either during the write or the read call.
|
|
_, err = str.Write([]byte("foobar"))
|
|
if err == nil {
|
|
_, err = str.Read([]byte{0})
|
|
}
|
|
require.EqualError(t, err, "stream reset")
|
|
}
|
|
|
|
func TestListenCloseCount(t *testing.T) {
|
|
s := GenSwarm(t, OptDialOnly)
|
|
addrsToListen := []ma.Multiaddr{
|
|
ma.StringCast("/ip4/0.0.0.0/tcp/0"),
|
|
ma.StringCast("/ip4/0.0.0.0/udp/0/quic-v1"),
|
|
}
|
|
|
|
if err := s.Listen(addrsToListen...); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
listenedAddrs := s.ListenAddresses()
|
|
require.Len(t, listenedAddrs, 2)
|
|
var addrToClose ma.Multiaddr
|
|
for _, addr := range listenedAddrs {
|
|
if _, err := addr.ValueForProtocol(ma.P_QUIC_V1); err == nil {
|
|
// make a copy of the address to make sure the multiaddr comparison actually works
|
|
addrToClose = ma.StringCast(addr.String())
|
|
}
|
|
}
|
|
|
|
s.ListenClose(addrToClose)
|
|
|
|
remainingAddrs := s.ListenAddresses()
|
|
require.Len(t, remainingAddrs, 1)
|
|
_, err := remainingAddrs[0].ValueForProtocol(ma.P_TCP)
|
|
require.NoError(t, err, "expected the TCP address to still be present")
|
|
}
|