package evidence import ( "fmt" "sync" "testing" "time" "github.com/go-kit/kit/log/term" "github.com/stretchr/testify/assert" "github.com/stretchr/testify/require" dbm "github.com/tendermint/tm-db" cfg "github.com/tendermint/tendermint/config" "github.com/tendermint/tendermint/libs/log" "github.com/tendermint/tendermint/p2p" sm "github.com/tendermint/tendermint/state" "github.com/tendermint/tendermint/types" ) // evidenceLogger is a TestingLogger which uses a different // color for each validator ("validator" key must exist). func evidenceLogger() log.Logger { return log.TestingLoggerWithColorFn(func(keyvals ...interface{}) term.FgBgColor { for i := 0; i < len(keyvals)-1; i += 2 { if keyvals[i] == "validator" { return term.FgBgColor{Fg: term.Color(uint8(keyvals[i+1].(int) + 1))} } } return term.FgBgColor{} }) } // connect N evidence reactors through N switches func makeAndConnectReactors(config *cfg.Config, stateDBs []dbm.DB) []*Reactor { N := len(stateDBs) reactors := make([]*Reactor, N) logger := evidenceLogger() for i := 0; i < N; i++ { evidenceDB := dbm.NewMemDB() blockStoreDB := dbm.NewMemDB() blockStore := initializeBlockStore(blockStoreDB, sm.LoadState(stateDBs[i]), []byte("myval")) pool, err := NewPool(stateDBs[i], evidenceDB, blockStore) if err != nil { panic(err) } reactors[i] = NewReactor(pool) reactors[i].SetLogger(logger.With("validator", i)) } p2p.MakeConnectedSwitches(config.P2P, N, func(i int, s *p2p.Switch) *p2p.Switch { s.AddReactor("EVIDENCE", reactors[i]) return s }, p2p.Connect2Switches) return reactors } // wait for all evidence on all reactors func waitForEvidence(t *testing.T, evs types.EvidenceList, reactors []*Reactor) { // wait for the evidence in all evpools wg := new(sync.WaitGroup) for i := 0; i < len(reactors); i++ { wg.Add(1) go _waitForEvidence(t, wg, evs, i, reactors) } done := make(chan struct{}) go func() { wg.Wait() close(done) }() timer := time.After(timeout) select { case <-timer: t.Fatal("Timed out waiting for evidence") case <-done: } } // wait for all evidence on a single evpool func _waitForEvidence( t *testing.T, wg *sync.WaitGroup, evs types.EvidenceList, reactorIdx int, reactors []*Reactor, ) { evpool := reactors[reactorIdx].evpool for len(evpool.AllPendingEvidence()) != len(evs) { time.Sleep(time.Millisecond * 100) } reapedEv := evpool.AllPendingEvidence() // put the reaped evidence in a map so we can quickly check we got everything evMap := make(map[string]types.Evidence) for _, e := range reapedEv { evMap[string(e.Hash())] = e } for i, expectedEv := range evs { gotEv := evMap[string(expectedEv.Hash())] assert.Equal(t, expectedEv, gotEv, fmt.Sprintf("evidence at index %d on reactor %d don't match: %v vs %v", i, reactorIdx, expectedEv, gotEv)) } wg.Done() } func sendEvidence(t *testing.T, evpool *Pool, val types.PrivValidator, n int) types.EvidenceList { evList := make([]types.Evidence, n) for i := 0; i < n; i++ { ev := types.NewMockDuplicateVoteEvidenceWithValidator(int64(i+1), time.Now().UTC(), val, evidenceChainID) err := evpool.AddEvidence(ev) require.NoError(t, err) evList[i] = ev } return evList } var ( numEvidence = 10 timeout = 120 * time.Second // ridiculously high because CircleCI is slow ) func TestReactorBroadcastEvidence(t *testing.T) { config := cfg.TestConfig() N := 7 // create statedb for everyone stateDBs := make([]dbm.DB, N) val := types.NewMockPV() // we need validators saved for heights at least as high as we have evidence for height := int64(numEvidence) + 10 for i := 0; i < N; i++ { stateDBs[i] = initializeValidatorState(val, height) } // make reactors from statedb reactors := makeAndConnectReactors(config, stateDBs) // set the peer height on each reactor for _, r := range reactors { for _, peer := range r.Switch.Peers().List() { ps := peerState{height} peer.Set(types.PeerStateKey, ps) } } // send a bunch of valid evidence to the first reactor's evpool // and wait for them all to be received in the others evList := sendEvidence(t, reactors[0].evpool, val, numEvidence) waitForEvidence(t, evList, reactors) } type peerState struct { height int64 } func (ps peerState) GetHeight() int64 { return ps.height } func TestReactorSelectiveBroadcast(t *testing.T) { config := cfg.TestConfig() val := types.NewMockPV() height1 := int64(numEvidence) + 10 height2 := int64(numEvidence) / 2 // DB1 is ahead of DB2 stateDB1 := initializeValidatorState(val, height1) stateDB2 := initializeValidatorState(val, height2) // make reactors from statedb reactors := makeAndConnectReactors(config, []dbm.DB{stateDB1, stateDB2}) // set the peer height on each reactor for _, r := range reactors { for _, peer := range r.Switch.Peers().List() { ps := peerState{height1} peer.Set(types.PeerStateKey, ps) } } // update the first reactor peer's height to be very small peer := reactors[0].Switch.Peers().List()[0] ps := peerState{height2} peer.Set(types.PeerStateKey, ps) // send a bunch of valid evidence to the first reactor's evpool evList := sendEvidence(t, reactors[0].evpool, val, numEvidence) // only ones less than the peers height should make it through waitForEvidence(t, evList[:numEvidence/2], reactors[1:2]) // peers should still be connected peers := reactors[1].Switch.Peers().List() assert.Equal(t, 1, len(peers)) }