- package statesync
-
- import (
- "context"
- "errors"
- "sync"
- "testing"
- "time"
-
- "github.com/stretchr/testify/assert"
- "github.com/stretchr/testify/mock"
- "github.com/stretchr/testify/require"
-
- abci "github.com/tendermint/tendermint/abci/types"
- "github.com/tendermint/tendermint/internal/proxy"
- proxymocks "github.com/tendermint/tendermint/internal/proxy/mocks"
- sm "github.com/tendermint/tendermint/internal/state"
- "github.com/tendermint/tendermint/internal/statesync/mocks"
- ssproto "github.com/tendermint/tendermint/proto/tendermint/statesync"
- "github.com/tendermint/tendermint/types"
- "github.com/tendermint/tendermint/version"
- )
-
- func TestSyncer_SyncAny(t *testing.T) {
- ctx, cancel := context.WithCancel(context.Background())
- defer cancel()
-
- state := sm.State{
- ChainID: "chain",
- Version: sm.Version{
- Consensus: version.Consensus{
- Block: version.BlockProtocol,
- App: 0,
- },
- Software: version.TMVersion,
- },
-
- LastBlockHeight: 1,
- LastBlockID: types.BlockID{Hash: []byte("blockhash")},
- LastBlockTime: time.Now(),
- LastResultsHash: []byte("last_results_hash"),
- AppHash: []byte("app_hash"),
-
- LastValidators: &types.ValidatorSet{Proposer: &types.Validator{Address: []byte("val1")}},
- Validators: &types.ValidatorSet{Proposer: &types.Validator{Address: []byte("val2")}},
- NextValidators: &types.ValidatorSet{Proposer: &types.Validator{Address: []byte("val3")}},
-
- ConsensusParams: *types.DefaultConsensusParams(),
- LastHeightConsensusParamsChanged: 1,
- }
- commit := &types.Commit{BlockID: types.BlockID{Hash: []byte("blockhash")}}
-
- chunks := []*chunk{
- {Height: 1, Format: 1, Index: 0, Chunk: []byte{1, 1, 0}},
- {Height: 1, Format: 1, Index: 1, Chunk: []byte{1, 1, 1}},
- {Height: 1, Format: 1, Index: 2, Chunk: []byte{1, 1, 2}},
- }
- s := &snapshot{Height: 1, Format: 1, Chunks: 3, Hash: []byte{1, 2, 3}}
-
- stateProvider := &mocks.StateProvider{}
- stateProvider.On("AppHash", mock.Anything, uint64(1)).Return(state.AppHash, nil)
- stateProvider.On("AppHash", mock.Anything, uint64(2)).Return([]byte("app_hash_2"), nil)
- stateProvider.On("Commit", mock.Anything, uint64(1)).Return(commit, nil)
- stateProvider.On("State", mock.Anything, uint64(1)).Return(state, nil)
- connSnapshot := &proxymocks.AppConnSnapshot{}
- connQuery := &proxymocks.AppConnQuery{}
-
- peerAID := types.NodeID("aa")
- peerBID := types.NodeID("bb")
- peerCID := types.NodeID("cc")
- rts := setup(ctx, t, connSnapshot, connQuery, stateProvider, 4)
-
- rts.reactor.syncer = rts.syncer
-
- // Adding a chunk should error when no sync is in progress
- _, err := rts.syncer.AddChunk(&chunk{Height: 1, Format: 1, Index: 0, Chunk: []byte{1}})
- require.Error(t, err)
-
- // Adding a couple of peers should trigger snapshot discovery messages
- err = rts.syncer.AddPeer(ctx, peerAID)
- require.NoError(t, err)
- e := <-rts.snapshotOutCh
- require.Equal(t, &ssproto.SnapshotsRequest{}, e.Message)
- require.Equal(t, peerAID, e.To)
-
- err = rts.syncer.AddPeer(ctx, peerBID)
- require.NoError(t, err)
- e = <-rts.snapshotOutCh
- require.Equal(t, &ssproto.SnapshotsRequest{}, e.Message)
- require.Equal(t, peerBID, e.To)
-
- // Both peers report back with snapshots. One of them also returns a snapshot we don't want, in
- // format 2, which will be rejected by the ABCI application.
- new, err := rts.syncer.AddSnapshot(peerAID, s)
- require.NoError(t, err)
- require.True(t, new)
-
- new, err = rts.syncer.AddSnapshot(peerBID, s)
- require.NoError(t, err)
- require.False(t, new)
-
- s2 := &snapshot{Height: 2, Format: 2, Chunks: 3, Hash: []byte{1}}
- new, err = rts.syncer.AddSnapshot(peerBID, s2)
- require.NoError(t, err)
- require.True(t, new)
-
- new, err = rts.syncer.AddSnapshot(peerCID, s2)
- require.NoError(t, err)
- require.False(t, new)
-
- // We start a sync, with peers sending back chunks when requested. We first reject the snapshot
- // with height 2 format 2, and accept the snapshot at height 1.
- connSnapshot.On("OfferSnapshot", mock.Anything, abci.RequestOfferSnapshot{
- Snapshot: &abci.Snapshot{
- Height: 2,
- Format: 2,
- Chunks: 3,
- Hash: []byte{1},
- },
- AppHash: []byte("app_hash_2"),
- }).Return(&abci.ResponseOfferSnapshot{Result: abci.ResponseOfferSnapshot_REJECT_FORMAT}, nil)
- connSnapshot.On("OfferSnapshot", mock.Anything, abci.RequestOfferSnapshot{
- Snapshot: &abci.Snapshot{
- Height: s.Height,
- Format: s.Format,
- Chunks: s.Chunks,
- Hash: s.Hash,
- Metadata: s.Metadata,
- },
- AppHash: []byte("app_hash"),
- }).Times(2).Return(&abci.ResponseOfferSnapshot{Result: abci.ResponseOfferSnapshot_ACCEPT}, nil)
-
- chunkRequests := make(map[uint32]int)
- chunkRequestsMtx := sync.Mutex{}
-
- chunkProcessDone := make(chan struct{})
-
- go func() {
- defer close(chunkProcessDone)
- var seen int
- for {
- if seen >= 4 {
- return
- }
-
- select {
- case <-ctx.Done():
- t.Logf("sent %d chunks", seen)
- return
- case e := <-rts.chunkOutCh:
- msg, ok := e.Message.(*ssproto.ChunkRequest)
- assert.True(t, ok)
-
- assert.EqualValues(t, 1, msg.Height)
- assert.EqualValues(t, 1, msg.Format)
- assert.LessOrEqual(t, msg.Index, uint32(len(chunks)))
-
- added, err := rts.syncer.AddChunk(chunks[msg.Index])
- assert.NoError(t, err)
- assert.True(t, added)
-
- chunkRequestsMtx.Lock()
- chunkRequests[msg.Index]++
- chunkRequestsMtx.Unlock()
- seen++
- t.Logf("added chunk (%d of 4): %d", seen, msg.Index)
- }
- }
- }()
-
- // The first time we're applying chunk 2 we tell it to retry the snapshot and discard chunk 1,
- // which should cause it to keep the existing chunk 0 and 2, and restart restoration from
- // beginning. We also wait for a little while, to exercise the retry logic in fetchChunks().
- connSnapshot.On("ApplySnapshotChunk", mock.Anything, abci.RequestApplySnapshotChunk{
- Index: 2, Chunk: []byte{1, 1, 2},
- }).Once().Run(func(args mock.Arguments) { time.Sleep(2 * time.Second) }).Return(
- &abci.ResponseApplySnapshotChunk{
- Result: abci.ResponseApplySnapshotChunk_RETRY_SNAPSHOT,
- RefetchChunks: []uint32{1},
- }, nil)
-
- connSnapshot.On("ApplySnapshotChunk", mock.Anything, abci.RequestApplySnapshotChunk{
- Index: 0, Chunk: []byte{1, 1, 0},
- }).Times(2).Return(&abci.ResponseApplySnapshotChunk{Result: abci.ResponseApplySnapshotChunk_ACCEPT}, nil)
- connSnapshot.On("ApplySnapshotChunk", mock.Anything, abci.RequestApplySnapshotChunk{
- Index: 1, Chunk: []byte{1, 1, 1},
- }).Times(2).Return(&abci.ResponseApplySnapshotChunk{Result: abci.ResponseApplySnapshotChunk_ACCEPT}, nil)
- connSnapshot.On("ApplySnapshotChunk", mock.Anything, abci.RequestApplySnapshotChunk{
- Index: 2, Chunk: []byte{1, 1, 2},
- }).Once().Return(&abci.ResponseApplySnapshotChunk{Result: abci.ResponseApplySnapshotChunk_ACCEPT}, nil)
- connQuery.On("Info", mock.Anything, proxy.RequestInfo).Return(&abci.ResponseInfo{
- AppVersion: 9,
- LastBlockHeight: 1,
- LastBlockAppHash: []byte("app_hash"),
- }, nil)
-
- newState, lastCommit, err := rts.syncer.SyncAny(ctx, 0, func() error { return nil })
- require.NoError(t, err)
-
- <-chunkProcessDone
-
- chunkRequestsMtx.Lock()
- require.Equal(t, map[uint32]int{0: 1, 1: 2, 2: 1}, chunkRequests)
- chunkRequestsMtx.Unlock()
-
- // The syncer should have updated the state app version from the ABCI info response.
- expectState := state
- expectState.Version.Consensus.App = 9
-
- require.Equal(t, expectState, newState)
- require.Equal(t, commit, lastCommit)
-
- require.Equal(t, len(chunks), int(rts.syncer.processingSnapshot.Chunks))
- require.Equal(t, expectState.LastBlockHeight, rts.syncer.lastSyncedSnapshotHeight)
- require.True(t, rts.syncer.avgChunkTime > 0)
-
- require.Equal(t, int64(rts.syncer.processingSnapshot.Chunks), rts.reactor.SnapshotChunksTotal())
- require.Equal(t, rts.syncer.lastSyncedSnapshotHeight, rts.reactor.SnapshotHeight())
- require.Equal(t, time.Duration(rts.syncer.avgChunkTime), rts.reactor.ChunkProcessAvgTime())
- require.Equal(t, int64(len(rts.syncer.snapshots.snapshots)), rts.reactor.TotalSnapshots())
- require.Equal(t, int64(0), rts.reactor.SnapshotChunksCount())
-
- connSnapshot.AssertExpectations(t)
- connQuery.AssertExpectations(t)
- }
-
- func TestSyncer_SyncAny_noSnapshots(t *testing.T) {
- stateProvider := &mocks.StateProvider{}
- stateProvider.On("AppHash", mock.Anything, mock.Anything).Return([]byte("app_hash"), nil)
-
- ctx, cancel := context.WithCancel(context.Background())
- defer cancel()
-
- rts := setup(ctx, t, nil, nil, stateProvider, 2)
-
- _, _, err := rts.syncer.SyncAny(ctx, 0, func() error { return nil })
- require.Equal(t, errNoSnapshots, err)
- }
-
- func TestSyncer_SyncAny_abort(t *testing.T) {
- stateProvider := &mocks.StateProvider{}
- stateProvider.On("AppHash", mock.Anything, mock.Anything).Return([]byte("app_hash"), nil)
-
- ctx, cancel := context.WithCancel(context.Background())
- defer cancel()
-
- rts := setup(ctx, t, nil, nil, stateProvider, 2)
-
- s := &snapshot{Height: 1, Format: 1, Chunks: 3, Hash: []byte{1, 2, 3}}
- peerID := types.NodeID("aa")
-
- _, err := rts.syncer.AddSnapshot(peerID, s)
- require.NoError(t, err)
-
- rts.conn.On("OfferSnapshot", mock.Anything, abci.RequestOfferSnapshot{
- Snapshot: toABCI(s), AppHash: []byte("app_hash"),
- }).Once().Return(&abci.ResponseOfferSnapshot{Result: abci.ResponseOfferSnapshot_ABORT}, nil)
-
- _, _, err = rts.syncer.SyncAny(ctx, 0, func() error { return nil })
- require.Equal(t, errAbort, err)
- rts.conn.AssertExpectations(t)
- }
-
- func TestSyncer_SyncAny_reject(t *testing.T) {
- stateProvider := &mocks.StateProvider{}
- stateProvider.On("AppHash", mock.Anything, mock.Anything).Return([]byte("app_hash"), nil)
-
- ctx, cancel := context.WithCancel(context.Background())
- defer cancel()
-
- rts := setup(ctx, t, nil, nil, stateProvider, 2)
-
- // s22 is tried first, then s12, then s11, then errNoSnapshots
- s22 := &snapshot{Height: 2, Format: 2, Chunks: 3, Hash: []byte{1, 2, 3}}
- s12 := &snapshot{Height: 1, Format: 2, Chunks: 3, Hash: []byte{1, 2, 3}}
- s11 := &snapshot{Height: 1, Format: 1, Chunks: 3, Hash: []byte{1, 2, 3}}
-
- peerID := types.NodeID("aa")
-
- _, err := rts.syncer.AddSnapshot(peerID, s22)
- require.NoError(t, err)
-
- _, err = rts.syncer.AddSnapshot(peerID, s12)
- require.NoError(t, err)
-
- _, err = rts.syncer.AddSnapshot(peerID, s11)
- require.NoError(t, err)
-
- rts.conn.On("OfferSnapshot", mock.Anything, abci.RequestOfferSnapshot{
- Snapshot: toABCI(s22), AppHash: []byte("app_hash"),
- }).Once().Return(&abci.ResponseOfferSnapshot{Result: abci.ResponseOfferSnapshot_REJECT}, nil)
-
- rts.conn.On("OfferSnapshot", mock.Anything, abci.RequestOfferSnapshot{
- Snapshot: toABCI(s12), AppHash: []byte("app_hash"),
- }).Once().Return(&abci.ResponseOfferSnapshot{Result: abci.ResponseOfferSnapshot_REJECT}, nil)
-
- rts.conn.On("OfferSnapshot", mock.Anything, abci.RequestOfferSnapshot{
- Snapshot: toABCI(s11), AppHash: []byte("app_hash"),
- }).Once().Return(&abci.ResponseOfferSnapshot{Result: abci.ResponseOfferSnapshot_REJECT}, nil)
-
- _, _, err = rts.syncer.SyncAny(ctx, 0, func() error { return nil })
- require.Equal(t, errNoSnapshots, err)
- rts.conn.AssertExpectations(t)
- }
-
- func TestSyncer_SyncAny_reject_format(t *testing.T) {
- stateProvider := &mocks.StateProvider{}
- stateProvider.On("AppHash", mock.Anything, mock.Anything).Return([]byte("app_hash"), nil)
-
- ctx, cancel := context.WithCancel(context.Background())
- defer cancel()
-
- rts := setup(ctx, t, nil, nil, stateProvider, 2)
-
- // s22 is tried first, which reject s22 and s12, then s11 will abort.
- s22 := &snapshot{Height: 2, Format: 2, Chunks: 3, Hash: []byte{1, 2, 3}}
- s12 := &snapshot{Height: 1, Format: 2, Chunks: 3, Hash: []byte{1, 2, 3}}
- s11 := &snapshot{Height: 1, Format: 1, Chunks: 3, Hash: []byte{1, 2, 3}}
-
- peerID := types.NodeID("aa")
-
- _, err := rts.syncer.AddSnapshot(peerID, s22)
- require.NoError(t, err)
-
- _, err = rts.syncer.AddSnapshot(peerID, s12)
- require.NoError(t, err)
-
- _, err = rts.syncer.AddSnapshot(peerID, s11)
- require.NoError(t, err)
-
- rts.conn.On("OfferSnapshot", mock.Anything, abci.RequestOfferSnapshot{
- Snapshot: toABCI(s22), AppHash: []byte("app_hash"),
- }).Once().Return(&abci.ResponseOfferSnapshot{Result: abci.ResponseOfferSnapshot_REJECT_FORMAT}, nil)
-
- rts.conn.On("OfferSnapshot", mock.Anything, abci.RequestOfferSnapshot{
- Snapshot: toABCI(s11), AppHash: []byte("app_hash"),
- }).Once().Return(&abci.ResponseOfferSnapshot{Result: abci.ResponseOfferSnapshot_ABORT}, nil)
-
- _, _, err = rts.syncer.SyncAny(ctx, 0, func() error { return nil })
- require.Equal(t, errAbort, err)
- rts.conn.AssertExpectations(t)
- }
-
- func TestSyncer_SyncAny_reject_sender(t *testing.T) {
- stateProvider := &mocks.StateProvider{}
- stateProvider.On("AppHash", mock.Anything, mock.Anything).Return([]byte("app_hash"), nil)
-
- ctx, cancel := context.WithCancel(context.Background())
- defer cancel()
-
- rts := setup(ctx, t, nil, nil, stateProvider, 2)
-
- peerAID := types.NodeID("aa")
- peerBID := types.NodeID("bb")
- peerCID := types.NodeID("cc")
-
- // sbc will be offered first, which will be rejected with reject_sender, causing all snapshots
- // submitted by both b and c (i.e. sb, sc, sbc) to be rejected. Finally, sa will reject and
- // errNoSnapshots is returned.
- sa := &snapshot{Height: 1, Format: 1, Chunks: 3, Hash: []byte{1, 2, 3}}
- sb := &snapshot{Height: 2, Format: 1, Chunks: 3, Hash: []byte{1, 2, 3}}
- sc := &snapshot{Height: 3, Format: 1, Chunks: 3, Hash: []byte{1, 2, 3}}
- sbc := &snapshot{Height: 4, Format: 1, Chunks: 3, Hash: []byte{1, 2, 3}}
-
- _, err := rts.syncer.AddSnapshot(peerAID, sa)
- require.NoError(t, err)
-
- _, err = rts.syncer.AddSnapshot(peerBID, sb)
- require.NoError(t, err)
-
- _, err = rts.syncer.AddSnapshot(peerCID, sc)
- require.NoError(t, err)
-
- _, err = rts.syncer.AddSnapshot(peerBID, sbc)
- require.NoError(t, err)
-
- _, err = rts.syncer.AddSnapshot(peerCID, sbc)
- require.NoError(t, err)
-
- rts.conn.On("OfferSnapshot", mock.Anything, abci.RequestOfferSnapshot{
- Snapshot: toABCI(sbc), AppHash: []byte("app_hash"),
- }).Once().Return(&abci.ResponseOfferSnapshot{Result: abci.ResponseOfferSnapshot_REJECT_SENDER}, nil)
-
- rts.conn.On("OfferSnapshot", mock.Anything, abci.RequestOfferSnapshot{
- Snapshot: toABCI(sa), AppHash: []byte("app_hash"),
- }).Once().Return(&abci.ResponseOfferSnapshot{Result: abci.ResponseOfferSnapshot_REJECT}, nil)
-
- _, _, err = rts.syncer.SyncAny(ctx, 0, func() error { return nil })
- require.Equal(t, errNoSnapshots, err)
- rts.conn.AssertExpectations(t)
- }
-
- func TestSyncer_SyncAny_abciError(t *testing.T) {
- stateProvider := &mocks.StateProvider{}
- stateProvider.On("AppHash", mock.Anything, mock.Anything).Return([]byte("app_hash"), nil)
-
- ctx, cancel := context.WithCancel(context.Background())
- defer cancel()
-
- rts := setup(ctx, t, nil, nil, stateProvider, 2)
-
- errBoom := errors.New("boom")
- s := &snapshot{Height: 1, Format: 1, Chunks: 3, Hash: []byte{1, 2, 3}}
-
- peerID := types.NodeID("aa")
-
- _, err := rts.syncer.AddSnapshot(peerID, s)
- require.NoError(t, err)
-
- rts.conn.On("OfferSnapshot", mock.Anything, abci.RequestOfferSnapshot{
- Snapshot: toABCI(s), AppHash: []byte("app_hash"),
- }).Once().Return(nil, errBoom)
-
- _, _, err = rts.syncer.SyncAny(ctx, 0, func() error { return nil })
- require.True(t, errors.Is(err, errBoom))
- rts.conn.AssertExpectations(t)
- }
-
- func TestSyncer_offerSnapshot(t *testing.T) {
- unknownErr := errors.New("unknown error")
- boom := errors.New("boom")
-
- testcases := map[string]struct {
- result abci.ResponseOfferSnapshot_Result
- err error
- expectErr error
- }{
- "accept": {abci.ResponseOfferSnapshot_ACCEPT, nil, nil},
- "abort": {abci.ResponseOfferSnapshot_ABORT, nil, errAbort},
- "reject": {abci.ResponseOfferSnapshot_REJECT, nil, errRejectSnapshot},
- "reject_format": {abci.ResponseOfferSnapshot_REJECT_FORMAT, nil, errRejectFormat},
- "reject_sender": {abci.ResponseOfferSnapshot_REJECT_SENDER, nil, errRejectSender},
- "unknown": {abci.ResponseOfferSnapshot_UNKNOWN, nil, unknownErr},
- "error": {0, boom, boom},
- "unknown non-zero": {9, nil, unknownErr},
- }
-
- ctx, cancel := context.WithCancel(context.Background())
- defer cancel()
-
- for name, tc := range testcases {
- tc := tc
- t.Run(name, func(t *testing.T) {
- ctx, cancel := context.WithCancel(ctx)
- defer cancel()
-
- stateProvider := &mocks.StateProvider{}
- stateProvider.On("AppHash", mock.Anything, mock.Anything).Return([]byte("app_hash"), nil)
-
- rts := setup(ctx, t, nil, nil, stateProvider, 2)
-
- s := &snapshot{Height: 1, Format: 1, Chunks: 3, Hash: []byte{1, 2, 3}, trustedAppHash: []byte("app_hash")}
- rts.conn.On("OfferSnapshot", mock.Anything, abci.RequestOfferSnapshot{
- Snapshot: toABCI(s),
- AppHash: []byte("app_hash"),
- }).Return(&abci.ResponseOfferSnapshot{Result: tc.result}, tc.err)
-
- err := rts.syncer.offerSnapshot(ctx, s)
- if tc.expectErr == unknownErr {
- require.Error(t, err)
- } else {
- unwrapped := errors.Unwrap(err)
- if unwrapped != nil {
- err = unwrapped
- }
- require.Equal(t, tc.expectErr, err)
- }
- })
- }
- }
-
- func TestSyncer_applyChunks_Results(t *testing.T) {
- unknownErr := errors.New("unknown error")
- boom := errors.New("boom")
-
- testcases := map[string]struct {
- result abci.ResponseApplySnapshotChunk_Result
- err error
- expectErr error
- }{
- "accept": {abci.ResponseApplySnapshotChunk_ACCEPT, nil, nil},
- "abort": {abci.ResponseApplySnapshotChunk_ABORT, nil, errAbort},
- "retry": {abci.ResponseApplySnapshotChunk_RETRY, nil, nil},
- "retry_snapshot": {abci.ResponseApplySnapshotChunk_RETRY_SNAPSHOT, nil, errRetrySnapshot},
- "reject_snapshot": {abci.ResponseApplySnapshotChunk_REJECT_SNAPSHOT, nil, errRejectSnapshot},
- "unknown": {abci.ResponseApplySnapshotChunk_UNKNOWN, nil, unknownErr},
- "error": {0, boom, boom},
- "unknown non-zero": {9, nil, unknownErr},
- }
-
- ctx, cancel := context.WithCancel(context.Background())
- defer cancel()
-
- for name, tc := range testcases {
- tc := tc
- t.Run(name, func(t *testing.T) {
- ctx, cancel := context.WithCancel(ctx)
- defer cancel()
-
- stateProvider := &mocks.StateProvider{}
- stateProvider.On("AppHash", mock.Anything, mock.Anything).Return([]byte("app_hash"), nil)
-
- rts := setup(ctx, t, nil, nil, stateProvider, 2)
-
- body := []byte{1, 2, 3}
- chunks, err := newChunkQueue(&snapshot{Height: 1, Format: 1, Chunks: 1}, t.TempDir())
- require.NoError(t, err)
-
- fetchStartTime := time.Now()
-
- _, err = chunks.Add(&chunk{Height: 1, Format: 1, Index: 0, Chunk: body})
- require.NoError(t, err)
-
- rts.conn.On("ApplySnapshotChunk", mock.Anything, abci.RequestApplySnapshotChunk{
- Index: 0, Chunk: body,
- }).Once().Return(&abci.ResponseApplySnapshotChunk{Result: tc.result}, tc.err)
- if tc.result == abci.ResponseApplySnapshotChunk_RETRY {
- rts.conn.On("ApplySnapshotChunk", mock.Anything, abci.RequestApplySnapshotChunk{
- Index: 0, Chunk: body,
- }).Once().Return(&abci.ResponseApplySnapshotChunk{
- Result: abci.ResponseApplySnapshotChunk_ACCEPT}, nil)
- }
-
- err = rts.syncer.applyChunks(ctx, chunks, fetchStartTime)
- if tc.expectErr == unknownErr {
- require.Error(t, err)
- } else {
- unwrapped := errors.Unwrap(err)
- if unwrapped != nil {
- err = unwrapped
- }
- require.Equal(t, tc.expectErr, err)
- }
-
- rts.conn.AssertExpectations(t)
- })
- }
- }
-
- func TestSyncer_applyChunks_RefetchChunks(t *testing.T) {
- // Discarding chunks via refetch_chunks should work the same for all results
- testcases := map[string]struct {
- result abci.ResponseApplySnapshotChunk_Result
- }{
- "accept": {abci.ResponseApplySnapshotChunk_ACCEPT},
- "abort": {abci.ResponseApplySnapshotChunk_ABORT},
- "retry": {abci.ResponseApplySnapshotChunk_RETRY},
- "retry_snapshot": {abci.ResponseApplySnapshotChunk_RETRY_SNAPSHOT},
- "reject_snapshot": {abci.ResponseApplySnapshotChunk_REJECT_SNAPSHOT},
- }
- ctx, cancel := context.WithCancel(context.Background())
- defer cancel()
-
- for name, tc := range testcases {
- tc := tc
- t.Run(name, func(t *testing.T) {
- ctx, cancel := context.WithCancel(ctx)
- defer cancel()
-
- stateProvider := &mocks.StateProvider{}
- stateProvider.On("AppHash", mock.Anything, mock.Anything).Return([]byte("app_hash"), nil)
-
- rts := setup(ctx, t, nil, nil, stateProvider, 2)
-
- chunks, err := newChunkQueue(&snapshot{Height: 1, Format: 1, Chunks: 3}, t.TempDir())
- require.NoError(t, err)
-
- fetchStartTime := time.Now()
-
- added, err := chunks.Add(&chunk{Height: 1, Format: 1, Index: 0, Chunk: []byte{0}})
- require.True(t, added)
- require.NoError(t, err)
- added, err = chunks.Add(&chunk{Height: 1, Format: 1, Index: 1, Chunk: []byte{1}})
- require.True(t, added)
- require.NoError(t, err)
- added, err = chunks.Add(&chunk{Height: 1, Format: 1, Index: 2, Chunk: []byte{2}})
- require.True(t, added)
- require.NoError(t, err)
-
- // The first two chunks are accepted, before the last one asks for 1 to be refetched
- rts.conn.On("ApplySnapshotChunk", mock.Anything, abci.RequestApplySnapshotChunk{
- Index: 0, Chunk: []byte{0},
- }).Once().Return(&abci.ResponseApplySnapshotChunk{Result: abci.ResponseApplySnapshotChunk_ACCEPT}, nil)
- rts.conn.On("ApplySnapshotChunk", mock.Anything, abci.RequestApplySnapshotChunk{
- Index: 1, Chunk: []byte{1},
- }).Once().Return(&abci.ResponseApplySnapshotChunk{Result: abci.ResponseApplySnapshotChunk_ACCEPT}, nil)
- rts.conn.On("ApplySnapshotChunk", mock.Anything, abci.RequestApplySnapshotChunk{
- Index: 2, Chunk: []byte{2},
- }).Once().Return(&abci.ResponseApplySnapshotChunk{
- Result: tc.result,
- RefetchChunks: []uint32{1},
- }, nil)
-
- // Since removing the chunk will cause Next() to block, we spawn a goroutine, then
- // check the queue contents, and finally close the queue to end the goroutine.
- // We don't really care about the result of applyChunks, since it has separate test.
- go func() {
- rts.syncer.applyChunks(ctx, chunks, fetchStartTime) //nolint:errcheck // purposefully ignore error
- }()
-
- time.Sleep(50 * time.Millisecond)
- require.True(t, chunks.Has(0))
- require.False(t, chunks.Has(1))
- require.True(t, chunks.Has(2))
-
- require.NoError(t, chunks.Close())
- })
- }
- }
-
- func TestSyncer_applyChunks_RejectSenders(t *testing.T) {
- // Banning chunks senders via ban_chunk_senders should work the same for all results
- testcases := map[string]struct {
- result abci.ResponseApplySnapshotChunk_Result
- }{
- "accept": {abci.ResponseApplySnapshotChunk_ACCEPT},
- "abort": {abci.ResponseApplySnapshotChunk_ABORT},
- "retry": {abci.ResponseApplySnapshotChunk_RETRY},
- "retry_snapshot": {abci.ResponseApplySnapshotChunk_RETRY_SNAPSHOT},
- "reject_snapshot": {abci.ResponseApplySnapshotChunk_REJECT_SNAPSHOT},
- }
- ctx, cancel := context.WithCancel(context.Background())
- defer cancel()
-
- for name, tc := range testcases {
- tc := tc
- t.Run(name, func(t *testing.T) {
- ctx, cancel := context.WithCancel(ctx)
- defer cancel()
-
- stateProvider := &mocks.StateProvider{}
- stateProvider.On("AppHash", mock.Anything, mock.Anything).Return([]byte("app_hash"), nil)
-
- rts := setup(ctx, t, nil, nil, stateProvider, 2)
-
- // Set up three peers across two snapshots, and ask for one of them to be banned.
- // It should be banned from all snapshots.
- peerAID := types.NodeID("aa")
- peerBID := types.NodeID("bb")
- peerCID := types.NodeID("cc")
-
- s1 := &snapshot{Height: 1, Format: 1, Chunks: 3}
- s2 := &snapshot{Height: 2, Format: 1, Chunks: 3}
-
- _, err := rts.syncer.AddSnapshot(peerAID, s1)
- require.NoError(t, err)
-
- _, err = rts.syncer.AddSnapshot(peerAID, s2)
- require.NoError(t, err)
-
- _, err = rts.syncer.AddSnapshot(peerBID, s1)
- require.NoError(t, err)
-
- _, err = rts.syncer.AddSnapshot(peerBID, s2)
- require.NoError(t, err)
-
- _, err = rts.syncer.AddSnapshot(peerCID, s1)
- require.NoError(t, err)
-
- _, err = rts.syncer.AddSnapshot(peerCID, s2)
- require.NoError(t, err)
-
- chunks, err := newChunkQueue(s1, t.TempDir())
- require.NoError(t, err)
-
- fetchStartTime := time.Now()
-
- added, err := chunks.Add(&chunk{Height: 1, Format: 1, Index: 0, Chunk: []byte{0}, Sender: peerAID})
- require.True(t, added)
- require.NoError(t, err)
-
- added, err = chunks.Add(&chunk{Height: 1, Format: 1, Index: 1, Chunk: []byte{1}, Sender: peerBID})
- require.True(t, added)
- require.NoError(t, err)
-
- added, err = chunks.Add(&chunk{Height: 1, Format: 1, Index: 2, Chunk: []byte{2}, Sender: peerCID})
- require.True(t, added)
- require.NoError(t, err)
-
- // The first two chunks are accepted, before the last one asks for b sender to be rejected
- rts.conn.On("ApplySnapshotChunk", mock.Anything, abci.RequestApplySnapshotChunk{
- Index: 0, Chunk: []byte{0}, Sender: "aa",
- }).Once().Return(&abci.ResponseApplySnapshotChunk{Result: abci.ResponseApplySnapshotChunk_ACCEPT}, nil)
- rts.conn.On("ApplySnapshotChunk", mock.Anything, abci.RequestApplySnapshotChunk{
- Index: 1, Chunk: []byte{1}, Sender: "bb",
- }).Once().Return(&abci.ResponseApplySnapshotChunk{Result: abci.ResponseApplySnapshotChunk_ACCEPT}, nil)
- rts.conn.On("ApplySnapshotChunk", mock.Anything, abci.RequestApplySnapshotChunk{
- Index: 2, Chunk: []byte{2}, Sender: "cc",
- }).Once().Return(&abci.ResponseApplySnapshotChunk{
- Result: tc.result,
- RejectSenders: []string{string(peerBID)},
- }, nil)
-
- // On retry, the last chunk will be tried again, so we just accept it then.
- if tc.result == abci.ResponseApplySnapshotChunk_RETRY {
- rts.conn.On("ApplySnapshotChunk", mock.Anything, abci.RequestApplySnapshotChunk{
- Index: 2, Chunk: []byte{2}, Sender: "cc",
- }).Once().Return(&abci.ResponseApplySnapshotChunk{Result: abci.ResponseApplySnapshotChunk_ACCEPT}, nil)
- }
-
- // We don't really care about the result of applyChunks, since it has separate test.
- // However, it will block on e.g. retry result, so we spawn a goroutine that will
- // be shut down when the chunk queue closes.
- go func() {
- rts.syncer.applyChunks(ctx, chunks, fetchStartTime) //nolint:errcheck // purposefully ignore error
- }()
-
- time.Sleep(50 * time.Millisecond)
-
- s1peers := rts.syncer.snapshots.GetPeers(s1)
- require.Len(t, s1peers, 2)
- require.EqualValues(t, "aa", s1peers[0])
- require.EqualValues(t, "cc", s1peers[1])
-
- rts.syncer.snapshots.GetPeers(s1)
- require.Len(t, s1peers, 2)
- require.EqualValues(t, "aa", s1peers[0])
- require.EqualValues(t, "cc", s1peers[1])
-
- require.NoError(t, chunks.Close())
- })
- }
- }
-
- func TestSyncer_verifyApp(t *testing.T) {
- boom := errors.New("boom")
- s := &snapshot{Height: 3, Format: 1, Chunks: 5, Hash: []byte{1, 2, 3}, trustedAppHash: []byte("app_hash")}
-
- testcases := map[string]struct {
- response *abci.ResponseInfo
- err error
- expectErr error
- }{
- "verified": {&abci.ResponseInfo{
- LastBlockHeight: 3,
- LastBlockAppHash: []byte("app_hash"),
- AppVersion: 9,
- }, nil, nil},
- "invalid height": {&abci.ResponseInfo{
- LastBlockHeight: 5,
- LastBlockAppHash: []byte("app_hash"),
- AppVersion: 9,
- }, nil, errVerifyFailed},
- "invalid hash": {&abci.ResponseInfo{
- LastBlockHeight: 3,
- LastBlockAppHash: []byte("xxx"),
- AppVersion: 9,
- }, nil, errVerifyFailed},
- "error": {nil, boom, boom},
- }
- ctx, cancel := context.WithCancel(context.Background())
- defer cancel()
-
- for name, tc := range testcases {
- tc := tc
- t.Run(name, func(t *testing.T) {
- ctx, cancel := context.WithCancel(ctx)
- defer cancel()
-
- rts := setup(ctx, t, nil, nil, nil, 2)
-
- rts.connQuery.On("Info", mock.Anything, proxy.RequestInfo).Return(tc.response, tc.err)
- version, err := rts.syncer.verifyApp(ctx, s)
- unwrapped := errors.Unwrap(err)
- if unwrapped != nil {
- err = unwrapped
- }
-
- require.Equal(t, tc.expectErr, err)
- if err == nil {
- require.Equal(t, tc.response.AppVersion, version)
- }
- })
- }
- }
-
- func toABCI(s *snapshot) *abci.Snapshot {
- return &abci.Snapshot{
- Height: s.Height,
- Format: s.Format,
- Chunks: s.Chunks,
- Hash: s.Hash,
- Metadata: s.Metadata,
- }
- }
|