- package v0
-
- import (
- "context"
- "crypto/rand"
- "encoding/binary"
- "fmt"
- mrand "math/rand"
- "os"
- "testing"
- "time"
-
- "github.com/gogo/protobuf/proto"
- gogotypes "github.com/gogo/protobuf/types"
- "github.com/stretchr/testify/assert"
- "github.com/stretchr/testify/require"
-
- "github.com/tendermint/tendermint/abci/example/counter"
- "github.com/tendermint/tendermint/abci/example/kvstore"
- abciserver "github.com/tendermint/tendermint/abci/server"
- abci "github.com/tendermint/tendermint/abci/types"
- cfg "github.com/tendermint/tendermint/config"
- "github.com/tendermint/tendermint/internal/mempool"
- "github.com/tendermint/tendermint/libs/log"
- tmrand "github.com/tendermint/tendermint/libs/rand"
- "github.com/tendermint/tendermint/libs/service"
- "github.com/tendermint/tendermint/proxy"
- "github.com/tendermint/tendermint/types"
- )
-
- // A cleanupFunc cleans up any config / test files created for a particular
- // test.
- type cleanupFunc func()
-
- func newMempoolWithApp(cc proxy.ClientCreator) (*CListMempool, cleanupFunc) {
- return newMempoolWithAppAndConfig(cc, cfg.ResetTestRoot("mempool_test"))
- }
-
- func newMempoolWithAppAndConfig(cc proxy.ClientCreator, config *cfg.Config) (*CListMempool, cleanupFunc) {
- appConnMem, _ := cc.NewABCIClient()
- appConnMem.SetLogger(log.TestingLogger().With("module", "abci-client", "connection", "mempool"))
- err := appConnMem.Start()
- if err != nil {
- panic(err)
- }
-
- mp := NewCListMempool(config.Mempool, appConnMem, 0)
- mp.SetLogger(log.TestingLogger())
-
- return mp, func() { os.RemoveAll(config.RootDir) }
- }
-
- func ensureNoFire(t *testing.T, ch <-chan struct{}, timeoutMS int) {
- timer := time.NewTimer(time.Duration(timeoutMS) * time.Millisecond)
- select {
- case <-ch:
- t.Fatal("Expected not to fire")
- case <-timer.C:
- }
- }
-
- func ensureFire(t *testing.T, ch <-chan struct{}, timeoutMS int) {
- timer := time.NewTimer(time.Duration(timeoutMS) * time.Millisecond)
- select {
- case <-ch:
- case <-timer.C:
- t.Fatal("Expected to fire")
- }
- }
-
- func checkTxs(t *testing.T, mp mempool.Mempool, count int, peerID uint16) types.Txs {
- txs := make(types.Txs, count)
- txInfo := mempool.TxInfo{SenderID: peerID}
- for i := 0; i < count; i++ {
- txBytes := make([]byte, 20)
- txs[i] = txBytes
- _, err := rand.Read(txBytes)
- if err != nil {
- t.Error(err)
- }
- if err := mp.CheckTx(context.Background(), txBytes, nil, txInfo); err != nil {
- // Skip invalid txs.
- // TestMempoolFilters will fail otherwise. It asserts a number of txs
- // returned.
- if mempool.IsPreCheckError(err) {
- continue
- }
- t.Fatalf("CheckTx failed: %v while checking #%d tx", err, i)
- }
- }
- return txs
- }
-
- func TestReapMaxBytesMaxGas(t *testing.T) {
- app := kvstore.NewApplication()
- cc := proxy.NewLocalClientCreator(app)
- mp, cleanup := newMempoolWithApp(cc)
- defer cleanup()
-
- // Ensure gas calculation behaves as expected
- checkTxs(t, mp, 1, mempool.UnknownPeerID)
- tx0 := mp.TxsFront().Value.(*mempoolTx)
- // assert that kv store has gas wanted = 1.
- require.Equal(t, app.CheckTx(abci.RequestCheckTx{Tx: tx0.tx}).GasWanted, int64(1), "KVStore had a gas value neq to 1")
- require.Equal(t, tx0.gasWanted, int64(1), "transactions gas was set incorrectly")
- // ensure each tx is 20 bytes long
- require.Equal(t, len(tx0.tx), 20, "Tx is longer than 20 bytes")
- mp.Flush()
-
- // each table driven test creates numTxsToCreate txs with checkTx, and at the end clears all remaining txs.
- // each tx has 20 bytes
- tests := []struct {
- numTxsToCreate int
- maxBytes int64
- maxGas int64
- expectedNumTxs int
- }{
- {20, -1, -1, 20},
- {20, -1, 0, 0},
- {20, -1, 10, 10},
- {20, -1, 30, 20},
- {20, 0, -1, 0},
- {20, 0, 10, 0},
- {20, 10, 10, 0},
- {20, 24, 10, 1},
- {20, 240, 5, 5},
- {20, 240, -1, 10},
- {20, 240, 10, 10},
- {20, 240, 15, 10},
- {20, 20000, -1, 20},
- {20, 20000, 5, 5},
- {20, 20000, 30, 20},
- }
- for tcIndex, tt := range tests {
- checkTxs(t, mp, tt.numTxsToCreate, mempool.UnknownPeerID)
- got := mp.ReapMaxBytesMaxGas(tt.maxBytes, tt.maxGas)
- assert.Equal(t, tt.expectedNumTxs, len(got), "Got %d txs, expected %d, tc #%d",
- len(got), tt.expectedNumTxs, tcIndex)
- mp.Flush()
- }
- }
-
- func TestMempoolFilters(t *testing.T) {
- app := kvstore.NewApplication()
- cc := proxy.NewLocalClientCreator(app)
- mp, cleanup := newMempoolWithApp(cc)
- defer cleanup()
- emptyTxArr := []types.Tx{[]byte{}}
-
- nopPreFilter := func(tx types.Tx) error { return nil }
- nopPostFilter := func(tx types.Tx, res *abci.ResponseCheckTx) error { return nil }
-
- // each table driven test creates numTxsToCreate txs with checkTx, and at the end clears all remaining txs.
- // each tx has 20 bytes
- tests := []struct {
- numTxsToCreate int
- preFilter mempool.PreCheckFunc
- postFilter mempool.PostCheckFunc
- expectedNumTxs int
- }{
- {10, nopPreFilter, nopPostFilter, 10},
- {10, mempool.PreCheckMaxBytes(10), nopPostFilter, 0},
- {10, mempool.PreCheckMaxBytes(22), nopPostFilter, 10},
- {10, nopPreFilter, mempool.PostCheckMaxGas(-1), 10},
- {10, nopPreFilter, mempool.PostCheckMaxGas(0), 0},
- {10, nopPreFilter, mempool.PostCheckMaxGas(1), 10},
- {10, nopPreFilter, mempool.PostCheckMaxGas(3000), 10},
- {10, mempool.PreCheckMaxBytes(10), mempool.PostCheckMaxGas(20), 0},
- {10, mempool.PreCheckMaxBytes(30), mempool.PostCheckMaxGas(20), 10},
- {10, mempool.PreCheckMaxBytes(22), mempool.PostCheckMaxGas(1), 10},
- {10, mempool.PreCheckMaxBytes(22), mempool.PostCheckMaxGas(0), 0},
- }
- for tcIndex, tt := range tests {
- err := mp.Update(1, emptyTxArr, abciResponses(len(emptyTxArr), abci.CodeTypeOK), tt.preFilter, tt.postFilter)
- require.NoError(t, err)
- checkTxs(t, mp, tt.numTxsToCreate, mempool.UnknownPeerID)
- require.Equal(t, tt.expectedNumTxs, mp.Size(), "mempool had the incorrect size, on test case %d", tcIndex)
- mp.Flush()
- }
- }
-
- func TestMempoolUpdate(t *testing.T) {
- app := kvstore.NewApplication()
- cc := proxy.NewLocalClientCreator(app)
- mp, cleanup := newMempoolWithApp(cc)
- defer cleanup()
-
- // 1. Adds valid txs to the cache
- {
- err := mp.Update(1, []types.Tx{[]byte{0x01}}, abciResponses(1, abci.CodeTypeOK), nil, nil)
- require.NoError(t, err)
- err = mp.CheckTx(context.Background(), []byte{0x01}, nil, mempool.TxInfo{})
- require.NoError(t, err)
- }
-
- // 2. Removes valid txs from the mempool
- {
- err := mp.CheckTx(context.Background(), []byte{0x02}, nil, mempool.TxInfo{})
- require.NoError(t, err)
- err = mp.Update(1, []types.Tx{[]byte{0x02}}, abciResponses(1, abci.CodeTypeOK), nil, nil)
- require.NoError(t, err)
- assert.Zero(t, mp.Size())
- }
-
- // 3. Removes invalid transactions from the cache and the mempool (if present)
- {
- err := mp.CheckTx(context.Background(), []byte{0x03}, nil, mempool.TxInfo{})
- require.NoError(t, err)
- err = mp.Update(1, []types.Tx{[]byte{0x03}}, abciResponses(1, 1), nil, nil)
- require.NoError(t, err)
- assert.Zero(t, mp.Size())
-
- err = mp.CheckTx(context.Background(), []byte{0x03}, nil, mempool.TxInfo{})
- require.NoError(t, err)
- }
- }
-
- func TestMempool_KeepInvalidTxsInCache(t *testing.T) {
- app := counter.NewApplication(true)
- cc := proxy.NewLocalClientCreator(app)
- wcfg := cfg.DefaultConfig()
- wcfg.Mempool.KeepInvalidTxsInCache = true
- mp, cleanup := newMempoolWithAppAndConfig(cc, wcfg)
- defer cleanup()
-
- // 1. An invalid transaction must remain in the cache after Update
- {
- a := make([]byte, 8)
- binary.BigEndian.PutUint64(a, 0)
-
- b := make([]byte, 8)
- binary.BigEndian.PutUint64(b, 1)
-
- err := mp.CheckTx(context.Background(), b, nil, mempool.TxInfo{})
- require.NoError(t, err)
-
- // simulate new block
- _ = app.DeliverTx(abci.RequestDeliverTx{Tx: a})
- _ = app.DeliverTx(abci.RequestDeliverTx{Tx: b})
- err = mp.Update(1, []types.Tx{a, b},
- []*abci.ResponseDeliverTx{{Code: abci.CodeTypeOK}, {Code: 2}}, nil, nil)
- require.NoError(t, err)
-
- // a must be added to the cache
- err = mp.CheckTx(context.Background(), a, nil, mempool.TxInfo{})
- require.NoError(t, err)
-
- // b must remain in the cache
- err = mp.CheckTx(context.Background(), b, nil, mempool.TxInfo{})
- require.NoError(t, err)
- }
-
- // 2. An invalid transaction must remain in the cache
- {
- a := make([]byte, 8)
- binary.BigEndian.PutUint64(a, 0)
-
- // remove a from the cache to test (2)
- mp.cache.Remove(a)
-
- err := mp.CheckTx(context.Background(), a, nil, mempool.TxInfo{})
- require.NoError(t, err)
- }
- }
-
- func TestTxsAvailable(t *testing.T) {
- app := kvstore.NewApplication()
- cc := proxy.NewLocalClientCreator(app)
- mp, cleanup := newMempoolWithApp(cc)
- defer cleanup()
- mp.EnableTxsAvailable()
-
- timeoutMS := 500
-
- // with no txs, it shouldnt fire
- ensureNoFire(t, mp.TxsAvailable(), timeoutMS)
-
- // send a bunch of txs, it should only fire once
- txs := checkTxs(t, mp, 100, mempool.UnknownPeerID)
- ensureFire(t, mp.TxsAvailable(), timeoutMS)
- ensureNoFire(t, mp.TxsAvailable(), timeoutMS)
-
- // call update with half the txs.
- // it should fire once now for the new height
- // since there are still txs left
- committedTxs, txs := txs[:50], txs[50:]
- if err := mp.Update(1, committedTxs, abciResponses(len(committedTxs), abci.CodeTypeOK), nil, nil); err != nil {
- t.Error(err)
- }
- ensureFire(t, mp.TxsAvailable(), timeoutMS)
- ensureNoFire(t, mp.TxsAvailable(), timeoutMS)
-
- // send a bunch more txs. we already fired for this height so it shouldnt fire again
- moreTxs := checkTxs(t, mp, 50, mempool.UnknownPeerID)
- ensureNoFire(t, mp.TxsAvailable(), timeoutMS)
-
- // now call update with all the txs. it should not fire as there are no txs left
- committedTxs = append(txs, moreTxs...) //nolint: gocritic
- if err := mp.Update(2, committedTxs, abciResponses(len(committedTxs), abci.CodeTypeOK), nil, nil); err != nil {
- t.Error(err)
- }
- ensureNoFire(t, mp.TxsAvailable(), timeoutMS)
-
- // send a bunch more txs, it should only fire once
- checkTxs(t, mp, 100, mempool.UnknownPeerID)
- ensureFire(t, mp.TxsAvailable(), timeoutMS)
- ensureNoFire(t, mp.TxsAvailable(), timeoutMS)
- }
-
- func TestSerialReap(t *testing.T) {
- app := counter.NewApplication(true)
- cc := proxy.NewLocalClientCreator(app)
-
- mp, cleanup := newMempoolWithApp(cc)
- defer cleanup()
-
- appConnCon, _ := cc.NewABCIClient()
- appConnCon.SetLogger(log.TestingLogger().With("module", "abci-client", "connection", "consensus"))
- err := appConnCon.Start()
- require.Nil(t, err)
-
- cacheMap := make(map[string]struct{})
- deliverTxsRange := func(start, end int) {
- // Deliver some txs.
- for i := start; i < end; i++ {
-
- // This will succeed
- txBytes := make([]byte, 8)
- binary.BigEndian.PutUint64(txBytes, uint64(i))
- err := mp.CheckTx(context.Background(), txBytes, nil, mempool.TxInfo{})
- _, cached := cacheMap[string(txBytes)]
- if cached {
- require.NotNil(t, err, "expected error for cached tx")
- } else {
- require.Nil(t, err, "expected no err for uncached tx")
- }
- cacheMap[string(txBytes)] = struct{}{}
-
- // Duplicates are cached and should return error
- err = mp.CheckTx(context.Background(), txBytes, nil, mempool.TxInfo{})
- require.NotNil(t, err, "Expected error after CheckTx on duplicated tx")
- }
- }
-
- reapCheck := func(exp int) {
- txs := mp.ReapMaxBytesMaxGas(-1, -1)
- require.Equal(t, len(txs), exp, fmt.Sprintf("Expected to reap %v txs but got %v", exp, len(txs)))
- }
-
- updateRange := func(start, end int) {
- txs := make([]types.Tx, 0)
- for i := start; i < end; i++ {
- txBytes := make([]byte, 8)
- binary.BigEndian.PutUint64(txBytes, uint64(i))
- txs = append(txs, txBytes)
- }
- if err := mp.Update(0, txs, abciResponses(len(txs), abci.CodeTypeOK), nil, nil); err != nil {
- t.Error(err)
- }
- }
-
- commitRange := func(start, end int) {
- ctx := context.Background()
- // Deliver some txs.
- for i := start; i < end; i++ {
- txBytes := make([]byte, 8)
- binary.BigEndian.PutUint64(txBytes, uint64(i))
- res, err := appConnCon.DeliverTxSync(ctx, abci.RequestDeliverTx{Tx: txBytes})
- if err != nil {
- t.Errorf("client error committing tx: %v", err)
- }
- if res.IsErr() {
- t.Errorf("error committing tx. Code:%v result:%X log:%v",
- res.Code, res.Data, res.Log)
- }
- }
- res, err := appConnCon.CommitSync(ctx)
- if err != nil {
- t.Errorf("client error committing: %v", err)
- }
- if len(res.Data) != 8 {
- t.Errorf("error committing. Hash:%X", res.Data)
- }
- }
-
- //----------------------------------------
-
- // Deliver some txs.
- deliverTxsRange(0, 100)
-
- // Reap the txs.
- reapCheck(100)
-
- // Reap again. We should get the same amount
- reapCheck(100)
-
- // Deliver 0 to 999, we should reap 900 new txs
- // because 100 were already counted.
- deliverTxsRange(0, 1000)
-
- // Reap the txs.
- reapCheck(1000)
-
- // Reap again. We should get the same amount
- reapCheck(1000)
-
- // Commit from the conensus AppConn
- commitRange(0, 500)
- updateRange(0, 500)
-
- // We should have 500 left.
- reapCheck(500)
-
- // Deliver 100 invalid txs and 100 valid txs
- deliverTxsRange(900, 1100)
-
- // We should have 600 now.
- reapCheck(600)
- }
-
- func TestMempool_CheckTxChecksTxSize(t *testing.T) {
- app := kvstore.NewApplication()
- cc := proxy.NewLocalClientCreator(app)
- mempl, cleanup := newMempoolWithApp(cc)
- defer cleanup()
-
- maxTxSize := mempl.config.MaxTxBytes
-
- testCases := []struct {
- len int
- err bool
- }{
- // check small txs. no error
- 0: {10, false},
- 1: {1000, false},
- 2: {1000000, false},
-
- // check around maxTxSize
- 3: {maxTxSize - 1, false},
- 4: {maxTxSize, false},
- 5: {maxTxSize + 1, true},
- }
-
- for i, testCase := range testCases {
- caseString := fmt.Sprintf("case %d, len %d", i, testCase.len)
-
- tx := tmrand.Bytes(testCase.len)
-
- err := mempl.CheckTx(context.Background(), tx, nil, mempool.TxInfo{})
- bv := gogotypes.BytesValue{Value: tx}
- bz, err2 := bv.Marshal()
- require.NoError(t, err2)
- require.Equal(t, len(bz), proto.Size(&bv), caseString)
-
- if !testCase.err {
- require.NoError(t, err, caseString)
- } else {
- require.Equal(t, err, mempool.ErrTxTooLarge{
- Max: maxTxSize,
- Actual: testCase.len,
- }, caseString)
- }
- }
- }
-
- func TestMempoolTxsBytes(t *testing.T) {
- app := kvstore.NewApplication()
- cc := proxy.NewLocalClientCreator(app)
- config := cfg.ResetTestRoot("mempool_test")
- config.Mempool.MaxTxsBytes = 10
- mp, cleanup := newMempoolWithAppAndConfig(cc, config)
- defer cleanup()
-
- // 1. zero by default
- assert.EqualValues(t, 0, mp.SizeBytes())
-
- // 2. len(tx) after CheckTx
- err := mp.CheckTx(context.Background(), []byte{0x01}, nil, mempool.TxInfo{})
- require.NoError(t, err)
- assert.EqualValues(t, 1, mp.SizeBytes())
-
- // 3. zero again after tx is removed by Update
- err = mp.Update(1, []types.Tx{[]byte{0x01}}, abciResponses(1, abci.CodeTypeOK), nil, nil)
- require.NoError(t, err)
- assert.EqualValues(t, 0, mp.SizeBytes())
-
- // 4. zero after Flush
- err = mp.CheckTx(context.Background(), []byte{0x02, 0x03}, nil, mempool.TxInfo{})
- require.NoError(t, err)
- assert.EqualValues(t, 2, mp.SizeBytes())
-
- mp.Flush()
- assert.EqualValues(t, 0, mp.SizeBytes())
-
- // 5. ErrMempoolIsFull is returned when/if MaxTxsBytes limit is reached.
- err = mp.CheckTx(
- context.Background(),
- []byte{0x04, 0x04, 0x04, 0x04, 0x04, 0x04, 0x04, 0x04, 0x04, 0x04},
- nil,
- mempool.TxInfo{},
- )
- require.NoError(t, err)
-
- err = mp.CheckTx(context.Background(), []byte{0x05}, nil, mempool.TxInfo{})
- if assert.Error(t, err) {
- assert.IsType(t, mempool.ErrMempoolIsFull{}, err)
- }
-
- // 6. zero after tx is rechecked and removed due to not being valid anymore
- app2 := counter.NewApplication(true)
- cc = proxy.NewLocalClientCreator(app2)
- mp, cleanup = newMempoolWithApp(cc)
- defer cleanup()
-
- txBytes := make([]byte, 8)
- binary.BigEndian.PutUint64(txBytes, uint64(0))
-
- err = mp.CheckTx(context.Background(), txBytes, nil, mempool.TxInfo{})
- require.NoError(t, err)
- assert.EqualValues(t, 8, mp.SizeBytes())
-
- appConnCon, _ := cc.NewABCIClient()
- appConnCon.SetLogger(log.TestingLogger().With("module", "abci-client", "connection", "consensus"))
- err = appConnCon.Start()
- require.Nil(t, err)
- t.Cleanup(func() {
- if err := appConnCon.Stop(); err != nil {
- t.Error(err)
- }
- })
- ctx := context.Background()
- res, err := appConnCon.DeliverTxSync(ctx, abci.RequestDeliverTx{Tx: txBytes})
- require.NoError(t, err)
- require.EqualValues(t, 0, res.Code)
- res2, err := appConnCon.CommitSync(ctx)
- require.NoError(t, err)
- require.NotEmpty(t, res2.Data)
-
- // Pretend like we committed nothing so txBytes gets rechecked and removed.
- err = mp.Update(1, []types.Tx{}, abciResponses(0, abci.CodeTypeOK), nil, nil)
- require.NoError(t, err)
- assert.EqualValues(t, 0, mp.SizeBytes())
-
- // 7. Test RemoveTxByKey function
- err = mp.CheckTx(context.Background(), []byte{0x06}, nil, mempool.TxInfo{})
- require.NoError(t, err)
- assert.EqualValues(t, 1, mp.SizeBytes())
- mp.RemoveTxByKey(mempool.TxKey([]byte{0x07}), true)
- assert.EqualValues(t, 1, mp.SizeBytes())
- mp.RemoveTxByKey(mempool.TxKey([]byte{0x06}), true)
- assert.EqualValues(t, 0, mp.SizeBytes())
-
- }
-
- // This will non-deterministically catch some concurrency failures like
- // https://github.com/tendermint/tendermint/issues/3509
- // TODO: all of the tests should probably also run using the remote proxy app
- // since otherwise we're not actually testing the concurrency of the mempool here!
- func TestMempoolRemoteAppConcurrency(t *testing.T) {
- sockPath := fmt.Sprintf("unix:///tmp/echo_%v.sock", tmrand.Str(6))
- app := kvstore.NewApplication()
- cc, server := newRemoteApp(t, sockPath, app)
- t.Cleanup(func() {
- if err := server.Stop(); err != nil {
- t.Error(err)
- }
- })
- config := cfg.ResetTestRoot("mempool_test")
- mp, cleanup := newMempoolWithAppAndConfig(cc, config)
- defer cleanup()
-
- // generate small number of txs
- nTxs := 10
- txLen := 200
- txs := make([]types.Tx, nTxs)
- for i := 0; i < nTxs; i++ {
- txs[i] = tmrand.Bytes(txLen)
- }
-
- // simulate a group of peers sending them over and over
- N := config.Mempool.Size
- maxPeers := 5
- for i := 0; i < N; i++ {
- peerID := mrand.Intn(maxPeers)
- txNum := mrand.Intn(nTxs)
- tx := txs[txNum]
-
- // this will err with ErrTxInCache many times ...
- mp.CheckTx(context.Background(), tx, nil, mempool.TxInfo{SenderID: uint16(peerID)}) //nolint: errcheck // will error
- }
- err := mp.FlushAppConn()
- require.NoError(t, err)
- }
-
- // caller must close server
- func newRemoteApp(
- t *testing.T,
- addr string,
- app abci.Application,
- ) (
- clientCreator proxy.ClientCreator,
- server service.Service,
- ) {
- clientCreator = proxy.NewRemoteClientCreator(addr, "socket", true)
-
- // Start server
- server = abciserver.NewSocketServer(addr, app)
- server.SetLogger(log.TestingLogger().With("module", "abci-server"))
- if err := server.Start(); err != nil {
- t.Fatalf("Error starting socket server: %v", err.Error())
- }
- return clientCreator, server
- }
-
- func abciResponses(n int, code uint32) []*abci.ResponseDeliverTx {
- responses := make([]*abci.ResponseDeliverTx, 0, n)
- for i := 0; i < n; i++ {
- responses = append(responses, &abci.ResponseDeliverTx{Code: code})
- }
- return responses
- }
|