//nolint:unused package pex_test import ( "context" "errors" "strings" "testing" "time" "github.com/stretchr/testify/require" dbm "github.com/tendermint/tm-db" "github.com/tendermint/tendermint/crypto/ed25519" "github.com/tendermint/tendermint/internal/p2p" "github.com/tendermint/tendermint/internal/p2p/p2ptest" "github.com/tendermint/tendermint/internal/p2p/pex" "github.com/tendermint/tendermint/libs/log" p2pproto "github.com/tendermint/tendermint/proto/tendermint/p2p" "github.com/tendermint/tendermint/types" ) const ( checkFrequency = 500 * time.Millisecond defaultBufferSize = 2 shortWait = 10 * time.Second longWait = 60 * time.Second firstNode = 0 secondNode = 1 thirdNode = 2 ) func TestReactorBasic(t *testing.T) { ctx, cancel := context.WithCancel(context.Background()) defer cancel() // start a network with one mock reactor and one "real" reactor testNet := setupNetwork(ctx, t, testOptions{ MockNodes: 1, TotalNodes: 2, }) testNet.connectAll(ctx, t) testNet.start(ctx, t) // assert that the mock node receives a request from the real node testNet.listenForRequest(ctx, t, secondNode, firstNode, shortWait) // assert that when a mock node sends a request it receives a response (and // the correct one) testNet.sendRequest(ctx, t, firstNode, secondNode) testNet.listenForResponse(ctx, t, secondNode, firstNode, shortWait, []p2pproto.PexAddress(nil)) } func TestReactorConnectFullNetwork(t *testing.T) { ctx, cancel := context.WithCancel(context.Background()) defer cancel() testNet := setupNetwork(ctx, t, testOptions{ TotalNodes: 4, }) // make every node be only connected with one other node (it actually ends up // being two because of two way connections but oh well) testNet.connectN(ctx, t, 1) testNet.start(ctx, t) // assert that all nodes add each other in the network for idx := 0; idx < len(testNet.nodes); idx++ { testNet.requireNumberOfPeers(t, idx, len(testNet.nodes)-1, longWait) } } func TestReactorSendsRequestsTooOften(t *testing.T) { ctx, cancel := context.WithCancel(context.Background()) defer cancel() r := setupSingle(ctx, t) badNode := newNodeID(t, "b") r.pexInCh <- p2p.Envelope{ From: badNode, Message: &p2pproto.PexRequest{}, } resp := <-r.pexOutCh msg, ok := resp.Message.(*p2pproto.PexResponse) require.True(t, ok) require.Empty(t, msg.Addresses) r.pexInCh <- p2p.Envelope{ From: badNode, Message: &p2pproto.PexRequest{}, } peerErr := <-r.pexErrCh require.Error(t, peerErr.Err) require.Empty(t, r.pexOutCh) require.Contains(t, peerErr.Err.Error(), "peer sent a request too close after a prior one") require.Equal(t, badNode, peerErr.NodeID) } func TestReactorSendsResponseWithoutRequest(t *testing.T) { t.Skip("This test needs updated https://github.com/tendermint/tendermint/issue/7634") ctx, cancel := context.WithCancel(context.Background()) defer cancel() testNet := setupNetwork(ctx, t, testOptions{ MockNodes: 1, TotalNodes: 3, }) testNet.connectAll(ctx, t) testNet.start(ctx, t) // firstNode sends the secondNode an unrequested response // NOTE: secondNode will send a request by default during startup so we send // two responses to counter that. testNet.sendResponse(ctx, t, firstNode, secondNode, []int{thirdNode}) testNet.sendResponse(ctx, t, firstNode, secondNode, []int{thirdNode}) // secondNode should evict the firstNode testNet.listenForPeerUpdate(ctx, t, secondNode, firstNode, p2p.PeerStatusDown, shortWait) } func TestReactorNeverSendsTooManyPeers(t *testing.T) { t.Skip("This test needs updated https://github.com/tendermint/tendermint/issue/7634") ctx, cancel := context.WithCancel(context.Background()) defer cancel() testNet := setupNetwork(ctx, t, testOptions{ MockNodes: 1, TotalNodes: 2, }) testNet.connectAll(ctx, t) testNet.start(ctx, t) testNet.addNodes(ctx, t, 110) nodes := make([]int, 110) for i := 0; i < len(nodes); i++ { nodes[i] = i + 2 } testNet.addAddresses(t, secondNode, nodes) // first we check that even although we have 110 peers, honest pex reactors // only send 100 (test if secondNode sends firstNode 100 addresses) testNet.pingAndlistenForNAddresses(ctx, t, secondNode, firstNode, shortWait, 100) } func TestReactorErrorsOnReceivingTooManyPeers(t *testing.T) { ctx, cancel := context.WithCancel(context.Background()) defer cancel() r := setupSingle(ctx, t) peer := p2p.NodeAddress{Protocol: p2p.MemoryProtocol, NodeID: randomNodeID(t)} added, err := r.manager.Add(peer) require.NoError(t, err) require.True(t, added) addresses := make([]p2pproto.PexAddress, 101) for i := 0; i < len(addresses); i++ { nodeAddress := p2p.NodeAddress{Protocol: p2p.MemoryProtocol, NodeID: randomNodeID(t)} addresses[i] = p2pproto.PexAddress{ URL: nodeAddress.String(), } } r.peerCh <- p2p.PeerUpdate{ NodeID: peer.NodeID, Status: p2p.PeerStatusUp, } select { // wait for a request and then send a response with too many addresses case req := <-r.pexOutCh: if _, ok := req.Message.(*p2pproto.PexRequest); !ok { t.Fatal("expected v2 pex request") } r.pexInCh <- p2p.Envelope{ From: peer.NodeID, Message: &p2pproto.PexResponse{ Addresses: addresses, }, } case <-time.After(10 * time.Second): t.Fatal("pex failed to send a request within 10 seconds") } peerErr := <-r.pexErrCh require.Error(t, peerErr.Err) require.Empty(t, r.pexOutCh) require.Contains(t, peerErr.Err.Error(), "peer sent too many addresses") require.Equal(t, peer.NodeID, peerErr.NodeID) } func TestReactorSmallPeerStoreInALargeNetwork(t *testing.T) { ctx, cancel := context.WithCancel(context.Background()) defer cancel() testNet := setupNetwork(ctx, t, testOptions{ TotalNodes: 8, MaxPeers: 4, MaxConnected: 3, BufferSize: 8, }) testNet.connectN(ctx, t, 1) testNet.start(ctx, t) // test that all nodes reach full capacity for _, nodeID := range testNet.nodes { require.Eventually(t, func() bool { // nolint:scopelint return testNet.network.Nodes[nodeID].PeerManager.PeerRatio() >= 0.9 }, longWait, checkFrequency) } } func TestReactorLargePeerStoreInASmallNetwork(t *testing.T) { ctx, cancel := context.WithCancel(context.Background()) defer cancel() testNet := setupNetwork(ctx, t, testOptions{ TotalNodes: 3, MaxPeers: 25, MaxConnected: 25, BufferSize: 5, }) testNet.connectN(ctx, t, 1) testNet.start(ctx, t) // assert that all nodes add each other in the network for idx := 0; idx < len(testNet.nodes); idx++ { testNet.requireNumberOfPeers(t, idx, len(testNet.nodes)-1, longWait) } } func TestReactorWithNetworkGrowth(t *testing.T) { t.Skip("This test needs updated https://github.com/tendermint/tendermint/issue/7634") ctx, cancel := context.WithCancel(context.Background()) defer cancel() testNet := setupNetwork(ctx, t, testOptions{ TotalNodes: 5, BufferSize: 5, }) testNet.connectAll(ctx, t) testNet.start(ctx, t) // assert that all nodes add each other in the network for idx := 0; idx < len(testNet.nodes); idx++ { testNet.requireNumberOfPeers(t, idx, len(testNet.nodes)-1, shortWait) } // now we inject 10 more nodes testNet.addNodes(ctx, t, 10) for i := 5; i < testNet.total; i++ { node := testNet.nodes[i] require.NoError(t, testNet.reactors[node].Start(ctx)) require.True(t, testNet.reactors[node].IsRunning()) // we connect all new nodes to a single entry point and check that the // node can distribute the addresses to all the others testNet.connectPeers(ctx, t, 0, i) } require.Len(t, testNet.reactors, 15) // assert that all nodes add each other in the network for idx := 0; idx < len(testNet.nodes); idx++ { testNet.requireNumberOfPeers(t, idx, len(testNet.nodes)-1, longWait) } } type singleTestReactor struct { reactor *pex.Reactor pexInCh chan p2p.Envelope pexOutCh chan p2p.Envelope pexErrCh chan p2p.PeerError pexCh *p2p.Channel peerCh chan p2p.PeerUpdate manager *p2p.PeerManager } func setupSingle(ctx context.Context, t *testing.T) *singleTestReactor { t.Helper() nodeID := newNodeID(t, "a") chBuf := 2 pexInCh := make(chan p2p.Envelope, chBuf) pexOutCh := make(chan p2p.Envelope, chBuf) pexErrCh := make(chan p2p.PeerError, chBuf) pexCh := p2p.NewChannel( p2p.ChannelID(pex.PexChannel), new(p2pproto.PexMessage), pexInCh, pexOutCh, pexErrCh, ) peerCh := make(chan p2p.PeerUpdate, chBuf) peerUpdates := p2p.NewPeerUpdates(peerCh, chBuf) peerManager, err := p2p.NewPeerManager(nodeID, dbm.NewMemDB(), p2p.PeerManagerOptions{}) require.NoError(t, err) chCreator := func(context.Context, *p2p.ChannelDescriptor) (*p2p.Channel, error) { return pexCh, nil } reactor, err := pex.NewReactor(ctx, log.TestingLogger(), peerManager, chCreator, peerUpdates) require.NoError(t, err) require.NoError(t, reactor.Start(ctx)) t.Cleanup(reactor.Wait) return &singleTestReactor{ reactor: reactor, pexInCh: pexInCh, pexOutCh: pexOutCh, pexErrCh: pexErrCh, pexCh: pexCh, peerCh: peerCh, manager: peerManager, } } type reactorTestSuite struct { network *p2ptest.Network logger log.Logger reactors map[types.NodeID]*pex.Reactor pexChannels map[types.NodeID]*p2p.Channel peerChans map[types.NodeID]chan p2p.PeerUpdate peerUpdates map[types.NodeID]*p2p.PeerUpdates nodes []types.NodeID mocks []types.NodeID total int opts testOptions } type testOptions struct { MockNodes int TotalNodes int BufferSize int MaxPeers uint16 MaxConnected uint16 } // setup setups a test suite with a network of nodes. Mocknodes represent the // hollow nodes that the test can listen and send on func setupNetwork(ctx context.Context, t *testing.T, opts testOptions) *reactorTestSuite { t.Helper() require.Greater(t, opts.TotalNodes, opts.MockNodes) if opts.BufferSize == 0 { opts.BufferSize = defaultBufferSize } networkOpts := p2ptest.NetworkOptions{ NumNodes: opts.TotalNodes, BufferSize: opts.BufferSize, NodeOpts: p2ptest.NodeOptions{ MaxPeers: opts.MaxPeers, MaxConnected: opts.MaxConnected, }, } chBuf := opts.BufferSize realNodes := opts.TotalNodes - opts.MockNodes rts := &reactorTestSuite{ logger: log.TestingLogger().With("testCase", t.Name()), network: p2ptest.MakeNetwork(ctx, t, networkOpts), reactors: make(map[types.NodeID]*pex.Reactor, realNodes), pexChannels: make(map[types.NodeID]*p2p.Channel, opts.TotalNodes), peerChans: make(map[types.NodeID]chan p2p.PeerUpdate, opts.TotalNodes), peerUpdates: make(map[types.NodeID]*p2p.PeerUpdates, opts.TotalNodes), total: opts.TotalNodes, opts: opts, } // NOTE: we don't assert that the channels get drained after stopping the // reactor rts.pexChannels = rts.network.MakeChannelsNoCleanup(ctx, t, pex.ChannelDescriptor()) idx := 0 for nodeID := range rts.network.Nodes { rts.peerChans[nodeID] = make(chan p2p.PeerUpdate, chBuf) rts.peerUpdates[nodeID] = p2p.NewPeerUpdates(rts.peerChans[nodeID], chBuf) rts.network.Nodes[nodeID].PeerManager.Register(ctx, rts.peerUpdates[nodeID]) chCreator := func(context.Context, *p2p.ChannelDescriptor) (*p2p.Channel, error) { return rts.pexChannels[nodeID], nil } // the first nodes in the array are always mock nodes if idx < opts.MockNodes { rts.mocks = append(rts.mocks, nodeID) } else { var err error rts.reactors[nodeID], err = pex.NewReactor( ctx, rts.logger.With("nodeID", nodeID), rts.network.Nodes[nodeID].PeerManager, chCreator, rts.peerUpdates[nodeID], ) require.NoError(t, err) } rts.nodes = append(rts.nodes, nodeID) idx++ } require.Len(t, rts.reactors, realNodes) t.Cleanup(func() { for _, reactor := range rts.reactors { if reactor.IsRunning() { reactor.Wait() require.False(t, reactor.IsRunning()) } } }) return rts } // starts up the pex reactors for each node func (r *reactorTestSuite) start(ctx context.Context, t *testing.T) { t.Helper() for _, reactor := range r.reactors { require.NoError(t, reactor.Start(ctx)) require.True(t, reactor.IsRunning()) } } func (r *reactorTestSuite) addNodes(ctx context.Context, t *testing.T, nodes int) { t.Helper() for i := 0; i < nodes; i++ { node := r.network.MakeNode(ctx, t, p2ptest.NodeOptions{ MaxPeers: r.opts.MaxPeers, MaxConnected: r.opts.MaxConnected, }) r.network.Nodes[node.NodeID] = node nodeID := node.NodeID r.pexChannels[nodeID] = node.MakeChannelNoCleanup(ctx, t, pex.ChannelDescriptor()) r.peerChans[nodeID] = make(chan p2p.PeerUpdate, r.opts.BufferSize) r.peerUpdates[nodeID] = p2p.NewPeerUpdates(r.peerChans[nodeID], r.opts.BufferSize) r.network.Nodes[nodeID].PeerManager.Register(ctx, r.peerUpdates[nodeID]) chCreator := func(context.Context, *p2p.ChannelDescriptor) (*p2p.Channel, error) { return r.pexChannels[nodeID], nil } var err error r.reactors[nodeID], err = pex.NewReactor( ctx, r.logger.With("nodeID", nodeID), r.network.Nodes[nodeID].PeerManager, chCreator, r.peerUpdates[nodeID], ) require.NoError(t, err) r.nodes = append(r.nodes, nodeID) r.total++ } } func (r *reactorTestSuite) listenFor( ctx context.Context, t *testing.T, node types.NodeID, conditional func(msg *p2p.Envelope) bool, assertion func(t *testing.T, msg *p2p.Envelope) bool, waitPeriod time.Duration, ) { ctx, cancel := context.WithTimeout(ctx, waitPeriod) defer cancel() iter := r.pexChannels[node].Receive(ctx) for iter.Next(ctx) { envelope := iter.Envelope() if conditional(envelope) && assertion(t, envelope) { return } } if errors.Is(ctx.Err(), context.DeadlineExceeded) { require.Fail(t, "timed out waiting for message", "node=%v, waitPeriod=%s", node, waitPeriod) } } func (r *reactorTestSuite) listenForRequest(ctx context.Context, t *testing.T, fromNode, toNode int, waitPeriod time.Duration) { to, from := r.checkNodePair(t, toNode, fromNode) conditional := func(msg *p2p.Envelope) bool { _, ok := msg.Message.(*p2pproto.PexRequest) return ok && msg.From == from } assertion := func(t *testing.T, msg *p2p.Envelope) bool { require.Equal(t, &p2pproto.PexRequest{}, msg.Message) return true } r.listenFor(ctx, t, to, conditional, assertion, waitPeriod) } func (r *reactorTestSuite) pingAndlistenForNAddresses( ctx context.Context, t *testing.T, fromNode, toNode int, waitPeriod time.Duration, addresses int, ) { t.Helper() to, from := r.checkNodePair(t, toNode, fromNode) conditional := func(msg *p2p.Envelope) bool { _, ok := msg.Message.(*p2pproto.PexResponse) return ok && msg.From == from } assertion := func(t *testing.T, msg *p2p.Envelope) bool { m, ok := msg.Message.(*p2pproto.PexResponse) if !ok { require.Fail(t, "expected pex response v2") return true } // assert the same amount of addresses if len(m.Addresses) == addresses { return true } // if we didn't get the right length, we wait and send the // request again time.Sleep(300 * time.Millisecond) r.sendRequest(ctx, t, toNode, fromNode) return false } r.sendRequest(ctx, t, toNode, fromNode) r.listenFor(ctx, t, to, conditional, assertion, waitPeriod) } func (r *reactorTestSuite) listenForResponse( ctx context.Context, t *testing.T, fromNode, toNode int, waitPeriod time.Duration, addresses []p2pproto.PexAddress, ) { to, from := r.checkNodePair(t, toNode, fromNode) conditional := func(msg *p2p.Envelope) bool { _, ok := msg.Message.(*p2pproto.PexResponse) return ok && msg.From == from } assertion := func(t *testing.T, msg *p2p.Envelope) bool { require.Equal(t, &p2pproto.PexResponse{Addresses: addresses}, msg.Message) return true } r.listenFor(ctx, t, to, conditional, assertion, waitPeriod) } func (r *reactorTestSuite) listenForPeerUpdate( ctx context.Context, t *testing.T, onNode, withNode int, status p2p.PeerStatus, waitPeriod time.Duration, ) { on, with := r.checkNodePair(t, onNode, withNode) sub := r.network.Nodes[on].PeerManager.Subscribe(ctx) timesUp := time.After(waitPeriod) for { select { case <-ctx.Done(): require.Fail(t, "operation canceled") return case peerUpdate := <-sub.Updates(): if peerUpdate.NodeID == with { require.Equal(t, status, peerUpdate.Status) return } case <-timesUp: require.Fail(t, "timed out waiting for peer status", "%v with status %v", with, status) return } } } func (r *reactorTestSuite) getAddressesFor(nodes []int) []p2pproto.PexAddress { addresses := make([]p2pproto.PexAddress, len(nodes)) for idx, node := range nodes { nodeID := r.nodes[node] addresses[idx] = p2pproto.PexAddress{ URL: r.network.Nodes[nodeID].NodeAddress.String(), } } return addresses } func (r *reactorTestSuite) sendRequest(ctx context.Context, t *testing.T, fromNode, toNode int) { t.Helper() to, from := r.checkNodePair(t, toNode, fromNode) require.NoError(t, r.pexChannels[from].Send(ctx, p2p.Envelope{ To: to, Message: &p2pproto.PexRequest{}, })) } func (r *reactorTestSuite) sendResponse( ctx context.Context, t *testing.T, fromNode, toNode int, withNodes []int, ) { t.Helper() from, to := r.checkNodePair(t, fromNode, toNode) addrs := r.getAddressesFor(withNodes) require.NoError(t, r.pexChannels[from].Send(ctx, p2p.Envelope{ To: to, Message: &p2pproto.PexResponse{ Addresses: addrs, }, })) } func (r *reactorTestSuite) requireNumberOfPeers( t *testing.T, nodeIndex, numPeers int, waitPeriod time.Duration, ) { t.Helper() require.Eventuallyf(t, func() bool { actualNumPeers := len(r.network.Nodes[r.nodes[nodeIndex]].PeerManager.Peers()) return actualNumPeers >= numPeers }, waitPeriod, checkFrequency, "peer failed to connect with the asserted amount of peers "+ "index=%d, node=%q, waitPeriod=%s expected=%d actual=%d", nodeIndex, r.nodes[nodeIndex], waitPeriod, numPeers, len(r.network.Nodes[r.nodes[nodeIndex]].PeerManager.Peers()), ) } func (r *reactorTestSuite) connectAll(ctx context.Context, t *testing.T) { r.connectN(ctx, t, r.total-1) } // connects all nodes with n other nodes func (r *reactorTestSuite) connectN(ctx context.Context, t *testing.T, n int) { if n >= r.total { require.Fail(t, "connectN: n must be less than the size of the network - 1") } for i := 0; i < r.total; i++ { for j := 0; j < n; j++ { r.connectPeers(ctx, t, i, (i+j+1)%r.total) } } } // connects node1 to node2 func (r *reactorTestSuite) connectPeers(ctx context.Context, t *testing.T, sourceNode, targetNode int) { t.Helper() node1, node2 := r.checkNodePair(t, sourceNode, targetNode) n1 := r.network.Nodes[node1] if n1 == nil { require.Fail(t, "connectPeers: source node %v is not part of the testnet", node1) return } n2 := r.network.Nodes[node2] if n2 == nil { require.Fail(t, "connectPeers: target node %v is not part of the testnet", node2) return } sourceSub := n1.PeerManager.Subscribe(ctx) targetSub := n2.PeerManager.Subscribe(ctx) sourceAddress := n1.NodeAddress targetAddress := n2.NodeAddress added, err := n1.PeerManager.Add(targetAddress) require.NoError(t, err) if !added { return } select { case peerUpdate := <-targetSub.Updates(): require.Equal(t, peerUpdate.NodeID, node1) require.Equal(t, peerUpdate.Status, p2p.PeerStatusUp) case <-time.After(2 * time.Second): require.Fail(t, "timed out waiting for peer", "%v accepting %v", targetNode, sourceNode) } select { case peerUpdate := <-sourceSub.Updates(): require.Equal(t, peerUpdate.NodeID, node2) require.Equal(t, peerUpdate.Status, p2p.PeerStatusUp) case <-time.After(2 * time.Second): require.Fail(t, "timed out waiting for peer", "%v dialing %v", sourceNode, targetNode) } added, err = n2.PeerManager.Add(sourceAddress) require.NoError(t, err) require.True(t, added) } func (r *reactorTestSuite) checkNodePair(t *testing.T, first, second int) (types.NodeID, types.NodeID) { require.NotEqual(t, first, second) require.Less(t, first, r.total) require.Less(t, second, r.total) return r.nodes[first], r.nodes[second] } func (r *reactorTestSuite) addAddresses(t *testing.T, node int, addrs []int) { peerManager := r.network.Nodes[r.nodes[node]].PeerManager for _, addr := range addrs { require.Less(t, addr, r.total) address := r.network.Nodes[r.nodes[addr]].NodeAddress added, err := peerManager.Add(address) require.NoError(t, err) require.True(t, added) } } func newNodeID(t *testing.T, id string) types.NodeID { nodeID, err := types.NewNodeID(strings.Repeat(id, 2*types.NodeIDByteLength)) require.NoError(t, err) return nodeID } func randomNodeID(t *testing.T) types.NodeID { return types.NodeIDFromPubKey(ed25519.GenPrivKey().PubKey()) }