You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

1749 lines
60 KiB

10 years ago
10 years ago
10 years ago
9 years ago
9 years ago
9 years ago
9 years ago
9 years ago
10 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
8 years ago
8 years ago
10 years ago
8 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
10 years ago
9 years ago
10 years ago
8 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
fix data race Closes #1442 ``` WARNING: DATA RACE Write at 0x00c4209de7c8 by goroutine 23: github.com/tendermint/tendermint/types.(*Block).fillHeader() /home/vagrant/go/src/github.com/tendermint/tendermint/types/block.go:88 +0x157 github.com/tendermint/tendermint/types.(*Block).Hash() /home/vagrant/go/src/github.com/tendermint/tendermint/types/block.go:104 +0x121 github.com/tendermint/tendermint/types.(*Block).HashesTo() /home/vagrant/go/src/github.com/tendermint/tendermint/types/block.go:135 +0x4f github.com/tendermint/tendermint/consensus.(*ConsensusState).enterPrecommit() /home/vagrant/go/src/github.com/tendermint/tendermint/consensus/state.go:1037 +0x182d github.com/tendermint/tendermint/consensus.(*ConsensusState).addVote() /home/vagrant/go/src/github.com/tendermint/tendermint/consensus/state.go:1425 +0x1a6c github.com/tendermint/tendermint/consensus.(*ConsensusState).tryAddVote() /home/vagrant/go/src/github.com/tendermint/tendermint/consensus/state.go:1318 +0x77 github.com/tendermint/tendermint/consensus.(*ConsensusState).handleMsg() /home/vagrant/go/src/github.com/tendermint/tendermint/consensus/state.go:581 +0x7a9 github.com/tendermint/tendermint/consensus.(*ConsensusState).receiveRoutine() /home/vagrant/go/src/github.com/tendermint/tendermint/consensus/state.go:539 +0x6c3 Previous read at 0x00c4209de7c8 by goroutine 47: github.com/tendermint/tendermint/vendor/github.com/tendermint/tmlibs/common.(*HexBytes).MarshalJSON() <autogenerated>:1 +0x52 github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.invokeMarshalJSON() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/json-encode.go:433 +0x88 github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.(*Codec)._encodeReflectJSON() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/json-encode.go:82 +0x8d2 github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.(*Codec).encodeReflectJSON() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/json-encode.go:50 +0x10e github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.(*Codec).encodeReflectJSONStruct() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/json-encode.go:348 +0x539 github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.(*Codec)._encodeReflectJSON() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/json-encode.go:119 +0x83f github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.(*Codec).encodeReflectJSON() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/json-encode.go:50 +0x10e github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.(*Codec).encodeReflectJSONStruct() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/json-encode.go:348 +0x539 github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.(*Codec)._encodeReflectJSON() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/json-encode.go:119 +0x83f github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.(*Codec).encodeReflectJSON() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/json-encode.go:50 +0x10e github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.(*Codec).encodeReflectJSONStruct() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/json-encode.go:348 +0x539 github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.(*Codec)._encodeReflectJSON() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/json-encode.go:119 +0x83f github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.(*Codec).encodeReflectJSON() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/json-encode.go:50 +0x10e github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.(*Codec).encodeReflectJSONStruct() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/json-encode.go:348 +0x539 github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.(*Codec)._encodeReflectJSON() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/json-encode.go:119 +0x83f github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.(*Codec).encodeReflectJSON() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/json-encode.go:50 +0x10e github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.(*Codec).MarshalJSON() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/amino.go:296 +0x182 github.com/tendermint/tendermint/rpc/lib/types.NewRPCSuccessResponse() /home/vagrant/go/src/github.com/tendermint/tendermint/rpc/lib/types/types.go:100 +0x12c github.com/tendermint/tendermint/rpc/lib/server.makeJSONRPCHandler.func1() /home/vagrant/go/src/github.com/tendermint/tendermint/rpc/lib/server/handlers.go:152 +0xab7 net/http.HandlerFunc.ServeHTTP() /usr/lib/go-1.9/src/net/http/server.go:1918 +0x51 net/http.(*ServeMux).ServeHTTP() /usr/lib/go-1.9/src/net/http/server.go:2254 +0xa2 github.com/tendermint/tendermint/rpc/lib/server.RecoverAndLogHandler.func1() /home/vagrant/go/src/github.com/tendermint/tendermint/rpc/lib/server/http_server.go:138 +0x4fa net/http.HandlerFunc.ServeHTTP() /usr/lib/go-1.9/src/net/http/server.go:1918 +0x51 net/http.serverHandler.ServeHTTP() /usr/lib/go-1.9/src/net/http/server.go:2619 +0xbc net/http.(*conn).serve() /usr/lib/go-1.9/src/net/http/server.go:1801 +0x83b Goroutine 23 (running) created at: github.com/tendermint/tendermint/consensus.(*ConsensusState).OnStart() /home/vagrant/go/src/github.com/tendermint/tendermint/consensus/state.go:250 +0x35b github.com/tendermint/tendermint/vendor/github.com/tendermint/tmlibs/common.(*BaseService).Start() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/tmlibs/common/service.go:130 +0x5fc github.com/tendermint/tendermint/consensus.(*ConsensusReactor).OnStart() /home/vagrant/go/src/github.com/tendermint/tendermint/consensus/reactor.go:69 +0x1b4 github.com/tendermint/tendermint/vendor/github.com/tendermint/tmlibs/common.(*BaseService).Start() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/tmlibs/common/service.go:130 +0x5fc github.com/tendermint/tendermint/consensus.(*ConsensusReactor).Start() <autogenerated>:1 +0x43 github.com/tendermint/tendermint/p2p.(*Switch).OnStart() /home/vagrant/go/src/github.com/tendermint/tendermint/p2p/switch.go:177 +0x124 github.com/tendermint/tendermint/vendor/github.com/tendermint/tmlibs/common.(*BaseService).Start() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/tmlibs/common/service.go:130 +0x5fc github.com/tendermint/tendermint/node.(*Node).OnStart() /home/vagrant/go/src/github.com/tendermint/tendermint/node/node.go:416 +0xa1b github.com/tendermint/tendermint/vendor/github.com/tendermint/tmlibs/common.(*BaseService).Start() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/tmlibs/common/service.go:130 +0x5fc github.com/tendermint/tendermint/rpc/test.StartTendermint() /home/vagrant/go/src/github.com/tendermint/tendermint/rpc/test/helpers.go:100 +0x5b github.com/tendermint/tendermint/rpc/client_test.TestMain() /home/vagrant/go/src/github.com/tendermint/tendermint/rpc/client/main_test.go:17 +0x4c main.main() github.com/tendermint/tendermint/rpc/client/_test/_testmain.go:76 +0x1cd Goroutine 47 (running) created at: net/http.(*Server).Serve() /usr/lib/go-1.9/src/net/http/server.go:2720 +0x37c net/http.Serve() /usr/lib/go-1.9/src/net/http/server.go:2323 +0xe2 github.com/tendermint/tendermint/rpc/lib/server.StartHTTPServer.func1() /home/vagrant/go/src/github.com/tendermint/tendermint/rpc/lib/server/http_server.go:35 +0xb3 ```
7 years ago
10 years ago
10 years ago
fix data race Closes #1442 ``` WARNING: DATA RACE Write at 0x00c4209de7c8 by goroutine 23: github.com/tendermint/tendermint/types.(*Block).fillHeader() /home/vagrant/go/src/github.com/tendermint/tendermint/types/block.go:88 +0x157 github.com/tendermint/tendermint/types.(*Block).Hash() /home/vagrant/go/src/github.com/tendermint/tendermint/types/block.go:104 +0x121 github.com/tendermint/tendermint/types.(*Block).HashesTo() /home/vagrant/go/src/github.com/tendermint/tendermint/types/block.go:135 +0x4f github.com/tendermint/tendermint/consensus.(*ConsensusState).enterPrecommit() /home/vagrant/go/src/github.com/tendermint/tendermint/consensus/state.go:1037 +0x182d github.com/tendermint/tendermint/consensus.(*ConsensusState).addVote() /home/vagrant/go/src/github.com/tendermint/tendermint/consensus/state.go:1425 +0x1a6c github.com/tendermint/tendermint/consensus.(*ConsensusState).tryAddVote() /home/vagrant/go/src/github.com/tendermint/tendermint/consensus/state.go:1318 +0x77 github.com/tendermint/tendermint/consensus.(*ConsensusState).handleMsg() /home/vagrant/go/src/github.com/tendermint/tendermint/consensus/state.go:581 +0x7a9 github.com/tendermint/tendermint/consensus.(*ConsensusState).receiveRoutine() /home/vagrant/go/src/github.com/tendermint/tendermint/consensus/state.go:539 +0x6c3 Previous read at 0x00c4209de7c8 by goroutine 47: github.com/tendermint/tendermint/vendor/github.com/tendermint/tmlibs/common.(*HexBytes).MarshalJSON() <autogenerated>:1 +0x52 github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.invokeMarshalJSON() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/json-encode.go:433 +0x88 github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.(*Codec)._encodeReflectJSON() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/json-encode.go:82 +0x8d2 github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.(*Codec).encodeReflectJSON() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/json-encode.go:50 +0x10e github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.(*Codec).encodeReflectJSONStruct() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/json-encode.go:348 +0x539 github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.(*Codec)._encodeReflectJSON() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/json-encode.go:119 +0x83f github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.(*Codec).encodeReflectJSON() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/json-encode.go:50 +0x10e github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.(*Codec).encodeReflectJSONStruct() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/json-encode.go:348 +0x539 github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.(*Codec)._encodeReflectJSON() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/json-encode.go:119 +0x83f github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.(*Codec).encodeReflectJSON() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/json-encode.go:50 +0x10e github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.(*Codec).encodeReflectJSONStruct() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/json-encode.go:348 +0x539 github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.(*Codec)._encodeReflectJSON() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/json-encode.go:119 +0x83f github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.(*Codec).encodeReflectJSON() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/json-encode.go:50 +0x10e github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.(*Codec).encodeReflectJSONStruct() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/json-encode.go:348 +0x539 github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.(*Codec)._encodeReflectJSON() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/json-encode.go:119 +0x83f github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.(*Codec).encodeReflectJSON() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/json-encode.go:50 +0x10e github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.(*Codec).MarshalJSON() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/amino.go:296 +0x182 github.com/tendermint/tendermint/rpc/lib/types.NewRPCSuccessResponse() /home/vagrant/go/src/github.com/tendermint/tendermint/rpc/lib/types/types.go:100 +0x12c github.com/tendermint/tendermint/rpc/lib/server.makeJSONRPCHandler.func1() /home/vagrant/go/src/github.com/tendermint/tendermint/rpc/lib/server/handlers.go:152 +0xab7 net/http.HandlerFunc.ServeHTTP() /usr/lib/go-1.9/src/net/http/server.go:1918 +0x51 net/http.(*ServeMux).ServeHTTP() /usr/lib/go-1.9/src/net/http/server.go:2254 +0xa2 github.com/tendermint/tendermint/rpc/lib/server.RecoverAndLogHandler.func1() /home/vagrant/go/src/github.com/tendermint/tendermint/rpc/lib/server/http_server.go:138 +0x4fa net/http.HandlerFunc.ServeHTTP() /usr/lib/go-1.9/src/net/http/server.go:1918 +0x51 net/http.serverHandler.ServeHTTP() /usr/lib/go-1.9/src/net/http/server.go:2619 +0xbc net/http.(*conn).serve() /usr/lib/go-1.9/src/net/http/server.go:1801 +0x83b Goroutine 23 (running) created at: github.com/tendermint/tendermint/consensus.(*ConsensusState).OnStart() /home/vagrant/go/src/github.com/tendermint/tendermint/consensus/state.go:250 +0x35b github.com/tendermint/tendermint/vendor/github.com/tendermint/tmlibs/common.(*BaseService).Start() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/tmlibs/common/service.go:130 +0x5fc github.com/tendermint/tendermint/consensus.(*ConsensusReactor).OnStart() /home/vagrant/go/src/github.com/tendermint/tendermint/consensus/reactor.go:69 +0x1b4 github.com/tendermint/tendermint/vendor/github.com/tendermint/tmlibs/common.(*BaseService).Start() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/tmlibs/common/service.go:130 +0x5fc github.com/tendermint/tendermint/consensus.(*ConsensusReactor).Start() <autogenerated>:1 +0x43 github.com/tendermint/tendermint/p2p.(*Switch).OnStart() /home/vagrant/go/src/github.com/tendermint/tendermint/p2p/switch.go:177 +0x124 github.com/tendermint/tendermint/vendor/github.com/tendermint/tmlibs/common.(*BaseService).Start() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/tmlibs/common/service.go:130 +0x5fc github.com/tendermint/tendermint/node.(*Node).OnStart() /home/vagrant/go/src/github.com/tendermint/tendermint/node/node.go:416 +0xa1b github.com/tendermint/tendermint/vendor/github.com/tendermint/tmlibs/common.(*BaseService).Start() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/tmlibs/common/service.go:130 +0x5fc github.com/tendermint/tendermint/rpc/test.StartTendermint() /home/vagrant/go/src/github.com/tendermint/tendermint/rpc/test/helpers.go:100 +0x5b github.com/tendermint/tendermint/rpc/client_test.TestMain() /home/vagrant/go/src/github.com/tendermint/tendermint/rpc/client/main_test.go:17 +0x4c main.main() github.com/tendermint/tendermint/rpc/client/_test/_testmain.go:76 +0x1cd Goroutine 47 (running) created at: net/http.(*Server).Serve() /usr/lib/go-1.9/src/net/http/server.go:2720 +0x37c net/http.Serve() /usr/lib/go-1.9/src/net/http/server.go:2323 +0xe2 github.com/tendermint/tendermint/rpc/lib/server.StartHTTPServer.func1() /home/vagrant/go/src/github.com/tendermint/tendermint/rpc/lib/server/http_server.go:35 +0xb3 ```
7 years ago
fix data race Closes #1442 ``` WARNING: DATA RACE Write at 0x00c4209de7c8 by goroutine 23: github.com/tendermint/tendermint/types.(*Block).fillHeader() /home/vagrant/go/src/github.com/tendermint/tendermint/types/block.go:88 +0x157 github.com/tendermint/tendermint/types.(*Block).Hash() /home/vagrant/go/src/github.com/tendermint/tendermint/types/block.go:104 +0x121 github.com/tendermint/tendermint/types.(*Block).HashesTo() /home/vagrant/go/src/github.com/tendermint/tendermint/types/block.go:135 +0x4f github.com/tendermint/tendermint/consensus.(*ConsensusState).enterPrecommit() /home/vagrant/go/src/github.com/tendermint/tendermint/consensus/state.go:1037 +0x182d github.com/tendermint/tendermint/consensus.(*ConsensusState).addVote() /home/vagrant/go/src/github.com/tendermint/tendermint/consensus/state.go:1425 +0x1a6c github.com/tendermint/tendermint/consensus.(*ConsensusState).tryAddVote() /home/vagrant/go/src/github.com/tendermint/tendermint/consensus/state.go:1318 +0x77 github.com/tendermint/tendermint/consensus.(*ConsensusState).handleMsg() /home/vagrant/go/src/github.com/tendermint/tendermint/consensus/state.go:581 +0x7a9 github.com/tendermint/tendermint/consensus.(*ConsensusState).receiveRoutine() /home/vagrant/go/src/github.com/tendermint/tendermint/consensus/state.go:539 +0x6c3 Previous read at 0x00c4209de7c8 by goroutine 47: github.com/tendermint/tendermint/vendor/github.com/tendermint/tmlibs/common.(*HexBytes).MarshalJSON() <autogenerated>:1 +0x52 github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.invokeMarshalJSON() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/json-encode.go:433 +0x88 github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.(*Codec)._encodeReflectJSON() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/json-encode.go:82 +0x8d2 github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.(*Codec).encodeReflectJSON() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/json-encode.go:50 +0x10e github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.(*Codec).encodeReflectJSONStruct() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/json-encode.go:348 +0x539 github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.(*Codec)._encodeReflectJSON() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/json-encode.go:119 +0x83f github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.(*Codec).encodeReflectJSON() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/json-encode.go:50 +0x10e github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.(*Codec).encodeReflectJSONStruct() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/json-encode.go:348 +0x539 github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.(*Codec)._encodeReflectJSON() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/json-encode.go:119 +0x83f github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.(*Codec).encodeReflectJSON() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/json-encode.go:50 +0x10e github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.(*Codec).encodeReflectJSONStruct() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/json-encode.go:348 +0x539 github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.(*Codec)._encodeReflectJSON() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/json-encode.go:119 +0x83f github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.(*Codec).encodeReflectJSON() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/json-encode.go:50 +0x10e github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.(*Codec).encodeReflectJSONStruct() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/json-encode.go:348 +0x539 github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.(*Codec)._encodeReflectJSON() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/json-encode.go:119 +0x83f github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.(*Codec).encodeReflectJSON() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/json-encode.go:50 +0x10e github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino.(*Codec).MarshalJSON() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/go-amino/amino.go:296 +0x182 github.com/tendermint/tendermint/rpc/lib/types.NewRPCSuccessResponse() /home/vagrant/go/src/github.com/tendermint/tendermint/rpc/lib/types/types.go:100 +0x12c github.com/tendermint/tendermint/rpc/lib/server.makeJSONRPCHandler.func1() /home/vagrant/go/src/github.com/tendermint/tendermint/rpc/lib/server/handlers.go:152 +0xab7 net/http.HandlerFunc.ServeHTTP() /usr/lib/go-1.9/src/net/http/server.go:1918 +0x51 net/http.(*ServeMux).ServeHTTP() /usr/lib/go-1.9/src/net/http/server.go:2254 +0xa2 github.com/tendermint/tendermint/rpc/lib/server.RecoverAndLogHandler.func1() /home/vagrant/go/src/github.com/tendermint/tendermint/rpc/lib/server/http_server.go:138 +0x4fa net/http.HandlerFunc.ServeHTTP() /usr/lib/go-1.9/src/net/http/server.go:1918 +0x51 net/http.serverHandler.ServeHTTP() /usr/lib/go-1.9/src/net/http/server.go:2619 +0xbc net/http.(*conn).serve() /usr/lib/go-1.9/src/net/http/server.go:1801 +0x83b Goroutine 23 (running) created at: github.com/tendermint/tendermint/consensus.(*ConsensusState).OnStart() /home/vagrant/go/src/github.com/tendermint/tendermint/consensus/state.go:250 +0x35b github.com/tendermint/tendermint/vendor/github.com/tendermint/tmlibs/common.(*BaseService).Start() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/tmlibs/common/service.go:130 +0x5fc github.com/tendermint/tendermint/consensus.(*ConsensusReactor).OnStart() /home/vagrant/go/src/github.com/tendermint/tendermint/consensus/reactor.go:69 +0x1b4 github.com/tendermint/tendermint/vendor/github.com/tendermint/tmlibs/common.(*BaseService).Start() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/tmlibs/common/service.go:130 +0x5fc github.com/tendermint/tendermint/consensus.(*ConsensusReactor).Start() <autogenerated>:1 +0x43 github.com/tendermint/tendermint/p2p.(*Switch).OnStart() /home/vagrant/go/src/github.com/tendermint/tendermint/p2p/switch.go:177 +0x124 github.com/tendermint/tendermint/vendor/github.com/tendermint/tmlibs/common.(*BaseService).Start() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/tmlibs/common/service.go:130 +0x5fc github.com/tendermint/tendermint/node.(*Node).OnStart() /home/vagrant/go/src/github.com/tendermint/tendermint/node/node.go:416 +0xa1b github.com/tendermint/tendermint/vendor/github.com/tendermint/tmlibs/common.(*BaseService).Start() /home/vagrant/go/src/github.com/tendermint/tendermint/vendor/github.com/tendermint/tmlibs/common/service.go:130 +0x5fc github.com/tendermint/tendermint/rpc/test.StartTendermint() /home/vagrant/go/src/github.com/tendermint/tendermint/rpc/test/helpers.go:100 +0x5b github.com/tendermint/tendermint/rpc/client_test.TestMain() /home/vagrant/go/src/github.com/tendermint/tendermint/rpc/client/main_test.go:17 +0x4c main.main() github.com/tendermint/tendermint/rpc/client/_test/_testmain.go:76 +0x1cd Goroutine 47 (running) created at: net/http.(*Server).Serve() /usr/lib/go-1.9/src/net/http/server.go:2720 +0x37c net/http.Serve() /usr/lib/go-1.9/src/net/http/server.go:2323 +0xe2 github.com/tendermint/tendermint/rpc/lib/server.StartHTTPServer.func1() /home/vagrant/go/src/github.com/tendermint/tendermint/rpc/lib/server/http_server.go:35 +0xb3 ```
7 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
10 years ago
8 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
9 years ago
9 years ago
9 years ago
8 years ago
8 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
8 years ago
8 years ago
8 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
10 years ago
10 years ago
10 years ago
10 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
7 years ago
7 years ago
7 years ago
7 years ago
7 years ago
7 years ago
7 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
7 years ago
10 years ago
10 years ago
10 years ago
10 years ago
8 years ago
10 years ago
10 years ago
8 years ago
10 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
10 years ago
10 years ago
10 years ago
10 years ago
10 years ago
10 years ago
10 years ago
7 years ago
10 years ago
10 years ago
10 years ago
10 years ago
10 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
10 years ago
10 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
10 years ago
10 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
10 years ago
10 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
9 years ago
7 years ago
10 years ago
9 years ago
9 years ago
10 years ago
10 years ago
7 years ago
10 years ago
7 years ago
10 years ago
10 years ago
10 years ago
10 years ago
10 years ago
10 years ago
10 years ago
10 years ago
8 years ago
8 years ago
7 years ago
7 years ago
7 years ago
7 years ago
10 years ago
10 years ago
7 years ago
7 years ago
10 years ago
10 years ago
10 years ago
10 years ago
9 years ago
8 years ago
9 years ago
7 years ago
10 years ago
10 years ago
  1. package consensus
  2. import (
  3. "bytes"
  4. "errors"
  5. "fmt"
  6. "reflect"
  7. "runtime/debug"
  8. "sync"
  9. "time"
  10. fail "github.com/ebuchman/fail-test"
  11. cmn "github.com/tendermint/tendermint/libs/common"
  12. "github.com/tendermint/tendermint/libs/log"
  13. tmtime "github.com/tendermint/tendermint/types/time"
  14. cfg "github.com/tendermint/tendermint/config"
  15. cstypes "github.com/tendermint/tendermint/consensus/types"
  16. tmevents "github.com/tendermint/tendermint/libs/events"
  17. "github.com/tendermint/tendermint/p2p"
  18. sm "github.com/tendermint/tendermint/state"
  19. "github.com/tendermint/tendermint/types"
  20. )
  21. //-----------------------------------------------------------------------------
  22. // Config
  23. const (
  24. proposalHeartbeatIntervalSeconds = 2
  25. )
  26. //-----------------------------------------------------------------------------
  27. // Errors
  28. var (
  29. ErrInvalidProposalSignature = errors.New("Error invalid proposal signature")
  30. ErrInvalidProposalPOLRound = errors.New("Error invalid proposal POL round")
  31. ErrAddingVote = errors.New("Error adding vote")
  32. ErrVoteHeightMismatch = errors.New("Error vote height mismatch")
  33. )
  34. //-----------------------------------------------------------------------------
  35. var (
  36. msgQueueSize = 1000
  37. )
  38. // msgs from the reactor which may update the state
  39. type msgInfo struct {
  40. Msg ConsensusMessage `json:"msg"`
  41. PeerID p2p.ID `json:"peer_key"`
  42. }
  43. // internally generated messages which may update the state
  44. type timeoutInfo struct {
  45. Duration time.Duration `json:"duration"`
  46. Height int64 `json:"height"`
  47. Round int `json:"round"`
  48. Step cstypes.RoundStepType `json:"step"`
  49. }
  50. func (ti *timeoutInfo) String() string {
  51. return fmt.Sprintf("%v ; %d/%d %v", ti.Duration, ti.Height, ti.Round, ti.Step)
  52. }
  53. // ConsensusState handles execution of the consensus algorithm.
  54. // It processes votes and proposals, and upon reaching agreement,
  55. // commits blocks to the chain and executes them against the application.
  56. // The internal state machine receives input from peers, the internal validator, and from a timer.
  57. type ConsensusState struct {
  58. cmn.BaseService
  59. // config details
  60. config *cfg.ConsensusConfig
  61. privValidator types.PrivValidator // for signing votes
  62. // services for creating and executing blocks
  63. blockExec *sm.BlockExecutor
  64. blockStore sm.BlockStore
  65. mempool sm.Mempool
  66. evpool sm.EvidencePool
  67. // internal state
  68. mtx sync.RWMutex
  69. cstypes.RoundState
  70. triggeredTimeoutPrecommit bool
  71. state sm.State // State until height-1.
  72. // state changes may be triggered by: msgs from peers,
  73. // msgs from ourself, or by timeouts
  74. peerMsgQueue chan msgInfo
  75. internalMsgQueue chan msgInfo
  76. timeoutTicker TimeoutTicker
  77. // information about about added votes and block parts are written on this channel
  78. // so statistics can be computed by reactor
  79. statsMsgQueue chan msgInfo
  80. // we use eventBus to trigger msg broadcasts in the reactor,
  81. // and to notify external subscribers, eg. through a websocket
  82. eventBus *types.EventBus
  83. // a Write-Ahead Log ensures we can recover from any kind of crash
  84. // and helps us avoid signing conflicting votes
  85. wal WAL
  86. replayMode bool // so we don't log signing errors during replay
  87. doWALCatchup bool // determines if we even try to do the catchup
  88. // for tests where we want to limit the number of transitions the state makes
  89. nSteps int
  90. // some functions can be overwritten for testing
  91. decideProposal func(height int64, round int)
  92. doPrevote func(height int64, round int)
  93. setProposal func(proposal *types.Proposal) error
  94. // closed when we finish shutting down
  95. done chan struct{}
  96. // synchronous pubsub between consensus state and reactor.
  97. // state only emits EventNewRoundStep, EventVote and EventProposalHeartbeat
  98. evsw tmevents.EventSwitch
  99. // for reporting metrics
  100. metrics *Metrics
  101. }
  102. // StateOption sets an optional parameter on the ConsensusState.
  103. type StateOption func(*ConsensusState)
  104. // NewConsensusState returns a new ConsensusState.
  105. func NewConsensusState(
  106. config *cfg.ConsensusConfig,
  107. state sm.State,
  108. blockExec *sm.BlockExecutor,
  109. blockStore sm.BlockStore,
  110. mempool sm.Mempool,
  111. evpool sm.EvidencePool,
  112. options ...StateOption,
  113. ) *ConsensusState {
  114. cs := &ConsensusState{
  115. config: config,
  116. blockExec: blockExec,
  117. blockStore: blockStore,
  118. mempool: mempool,
  119. peerMsgQueue: make(chan msgInfo, msgQueueSize),
  120. internalMsgQueue: make(chan msgInfo, msgQueueSize),
  121. timeoutTicker: NewTimeoutTicker(),
  122. statsMsgQueue: make(chan msgInfo, msgQueueSize),
  123. done: make(chan struct{}),
  124. doWALCatchup: true,
  125. wal: nilWAL{},
  126. evpool: evpool,
  127. evsw: tmevents.NewEventSwitch(),
  128. metrics: NopMetrics(),
  129. }
  130. // set function defaults (may be overwritten before calling Start)
  131. cs.decideProposal = cs.defaultDecideProposal
  132. cs.doPrevote = cs.defaultDoPrevote
  133. cs.setProposal = cs.defaultSetProposal
  134. cs.updateToState(state)
  135. // Don't call scheduleRound0 yet.
  136. // We do that upon Start().
  137. cs.reconstructLastCommit(state)
  138. cs.BaseService = *cmn.NewBaseService(nil, "ConsensusState", cs)
  139. for _, option := range options {
  140. option(cs)
  141. }
  142. return cs
  143. }
  144. //----------------------------------------
  145. // Public interface
  146. // SetLogger implements Service.
  147. func (cs *ConsensusState) SetLogger(l log.Logger) {
  148. cs.BaseService.Logger = l
  149. cs.timeoutTicker.SetLogger(l)
  150. }
  151. // SetEventBus sets event bus.
  152. func (cs *ConsensusState) SetEventBus(b *types.EventBus) {
  153. cs.eventBus = b
  154. cs.blockExec.SetEventBus(b)
  155. }
  156. // StateMetrics sets the metrics.
  157. func StateMetrics(metrics *Metrics) StateOption {
  158. return func(cs *ConsensusState) { cs.metrics = metrics }
  159. }
  160. // String returns a string.
  161. func (cs *ConsensusState) String() string {
  162. // better not to access shared variables
  163. return fmt.Sprintf("ConsensusState") //(H:%v R:%v S:%v", cs.Height, cs.Round, cs.Step)
  164. }
  165. // GetState returns a copy of the chain state.
  166. func (cs *ConsensusState) GetState() sm.State {
  167. cs.mtx.RLock()
  168. defer cs.mtx.RUnlock()
  169. return cs.state.Copy()
  170. }
  171. // GetLastHeight returns the last height committed.
  172. // If there were no blocks, returns 0.
  173. func (cs *ConsensusState) GetLastHeight() int64 {
  174. cs.mtx.Lock()
  175. defer cs.mtx.Unlock()
  176. return cs.RoundState.Height - 1
  177. }
  178. // GetRoundState returns a shallow copy of the internal consensus state.
  179. func (cs *ConsensusState) GetRoundState() *cstypes.RoundState {
  180. cs.mtx.RLock()
  181. defer cs.mtx.RUnlock()
  182. rs := cs.RoundState // copy
  183. return &rs
  184. }
  185. // GetRoundStateJSON returns a json of RoundState, marshalled using go-amino.
  186. func (cs *ConsensusState) GetRoundStateJSON() ([]byte, error) {
  187. cs.mtx.RLock()
  188. defer cs.mtx.RUnlock()
  189. return cdc.MarshalJSON(cs.RoundState)
  190. }
  191. // GetRoundStateSimpleJSON returns a json of RoundStateSimple, marshalled using go-amino.
  192. func (cs *ConsensusState) GetRoundStateSimpleJSON() ([]byte, error) {
  193. cs.mtx.RLock()
  194. defer cs.mtx.RUnlock()
  195. return cdc.MarshalJSON(cs.RoundState.RoundStateSimple())
  196. }
  197. // GetValidators returns a copy of the current validators.
  198. func (cs *ConsensusState) GetValidators() (int64, []*types.Validator) {
  199. cs.mtx.RLock()
  200. defer cs.mtx.RUnlock()
  201. return cs.state.LastBlockHeight, cs.state.Validators.Copy().Validators
  202. }
  203. // SetPrivValidator sets the private validator account for signing votes.
  204. func (cs *ConsensusState) SetPrivValidator(priv types.PrivValidator) {
  205. cs.mtx.Lock()
  206. defer cs.mtx.Unlock()
  207. cs.privValidator = priv
  208. }
  209. // SetTimeoutTicker sets the local timer. It may be useful to overwrite for testing.
  210. func (cs *ConsensusState) SetTimeoutTicker(timeoutTicker TimeoutTicker) {
  211. cs.mtx.Lock()
  212. defer cs.mtx.Unlock()
  213. cs.timeoutTicker = timeoutTicker
  214. }
  215. // LoadCommit loads the commit for a given height.
  216. func (cs *ConsensusState) LoadCommit(height int64) *types.Commit {
  217. cs.mtx.RLock()
  218. defer cs.mtx.RUnlock()
  219. if height == cs.blockStore.Height() {
  220. return cs.blockStore.LoadSeenCommit(height)
  221. }
  222. return cs.blockStore.LoadBlockCommit(height)
  223. }
  224. // OnStart implements cmn.Service.
  225. // It loads the latest state via the WAL, and starts the timeout and receive routines.
  226. func (cs *ConsensusState) OnStart() error {
  227. if err := cs.evsw.Start(); err != nil {
  228. return err
  229. }
  230. // we may set the WAL in testing before calling Start,
  231. // so only OpenWAL if its still the nilWAL
  232. if _, ok := cs.wal.(nilWAL); ok {
  233. walFile := cs.config.WalFile()
  234. wal, err := cs.OpenWAL(walFile)
  235. if err != nil {
  236. cs.Logger.Error("Error loading ConsensusState wal", "err", err.Error())
  237. return err
  238. }
  239. cs.wal = wal
  240. }
  241. // we need the timeoutRoutine for replay so
  242. // we don't block on the tick chan.
  243. // NOTE: we will get a build up of garbage go routines
  244. // firing on the tockChan until the receiveRoutine is started
  245. // to deal with them (by that point, at most one will be valid)
  246. if err := cs.timeoutTicker.Start(); err != nil {
  247. return err
  248. }
  249. // we may have lost some votes if the process crashed
  250. // reload from consensus log to catchup
  251. if cs.doWALCatchup {
  252. if err := cs.catchupReplay(cs.Height); err != nil {
  253. cs.Logger.Error("Error on catchup replay. Proceeding to start ConsensusState anyway", "err", err.Error())
  254. // NOTE: if we ever do return an error here,
  255. // make sure to stop the timeoutTicker
  256. }
  257. }
  258. // now start the receiveRoutine
  259. go cs.receiveRoutine(0)
  260. // schedule the first round!
  261. // use GetRoundState so we don't race the receiveRoutine for access
  262. cs.scheduleRound0(cs.GetRoundState())
  263. return nil
  264. }
  265. // timeoutRoutine: receive requests for timeouts on tickChan and fire timeouts on tockChan
  266. // receiveRoutine: serializes processing of proposoals, block parts, votes; coordinates state transitions
  267. func (cs *ConsensusState) startRoutines(maxSteps int) {
  268. err := cs.timeoutTicker.Start()
  269. if err != nil {
  270. cs.Logger.Error("Error starting timeout ticker", "err", err)
  271. return
  272. }
  273. go cs.receiveRoutine(maxSteps)
  274. }
  275. // OnStop implements cmn.Service. It stops all routines and waits for the WAL to finish.
  276. func (cs *ConsensusState) OnStop() {
  277. cs.evsw.Stop()
  278. cs.timeoutTicker.Stop()
  279. }
  280. // Wait waits for the the main routine to return.
  281. // NOTE: be sure to Stop() the event switch and drain
  282. // any event channels or this may deadlock
  283. func (cs *ConsensusState) Wait() {
  284. <-cs.done
  285. }
  286. // OpenWAL opens a file to log all consensus messages and timeouts for deterministic accountability
  287. func (cs *ConsensusState) OpenWAL(walFile string) (WAL, error) {
  288. wal, err := NewWAL(walFile)
  289. if err != nil {
  290. cs.Logger.Error("Failed to open WAL for consensus state", "wal", walFile, "err", err)
  291. return nil, err
  292. }
  293. wal.SetLogger(cs.Logger.With("wal", walFile))
  294. if err := wal.Start(); err != nil {
  295. return nil, err
  296. }
  297. return wal, nil
  298. }
  299. //------------------------------------------------------------
  300. // Public interface for passing messages into the consensus state, possibly causing a state transition.
  301. // If peerID == "", the msg is considered internal.
  302. // Messages are added to the appropriate queue (peer or internal).
  303. // If the queue is full, the function may block.
  304. // TODO: should these return anything or let callers just use events?
  305. // AddVote inputs a vote.
  306. func (cs *ConsensusState) AddVote(vote *types.Vote, peerID p2p.ID) (added bool, err error) {
  307. if peerID == "" {
  308. cs.internalMsgQueue <- msgInfo{&VoteMessage{vote}, ""}
  309. } else {
  310. cs.peerMsgQueue <- msgInfo{&VoteMessage{vote}, peerID}
  311. }
  312. // TODO: wait for event?!
  313. return false, nil
  314. }
  315. // SetProposal inputs a proposal.
  316. func (cs *ConsensusState) SetProposal(proposal *types.Proposal, peerID p2p.ID) error {
  317. if peerID == "" {
  318. cs.internalMsgQueue <- msgInfo{&ProposalMessage{proposal}, ""}
  319. } else {
  320. cs.peerMsgQueue <- msgInfo{&ProposalMessage{proposal}, peerID}
  321. }
  322. // TODO: wait for event?!
  323. return nil
  324. }
  325. // AddProposalBlockPart inputs a part of the proposal block.
  326. func (cs *ConsensusState) AddProposalBlockPart(height int64, round int, part *types.Part, peerID p2p.ID) error {
  327. if peerID == "" {
  328. cs.internalMsgQueue <- msgInfo{&BlockPartMessage{height, round, part}, ""}
  329. } else {
  330. cs.peerMsgQueue <- msgInfo{&BlockPartMessage{height, round, part}, peerID}
  331. }
  332. // TODO: wait for event?!
  333. return nil
  334. }
  335. // SetProposalAndBlock inputs the proposal and all block parts.
  336. func (cs *ConsensusState) SetProposalAndBlock(proposal *types.Proposal, block *types.Block, parts *types.PartSet, peerID p2p.ID) error {
  337. if err := cs.SetProposal(proposal, peerID); err != nil {
  338. return err
  339. }
  340. for i := 0; i < parts.Total(); i++ {
  341. part := parts.GetPart(i)
  342. if err := cs.AddProposalBlockPart(proposal.Height, proposal.Round, part, peerID); err != nil {
  343. return err
  344. }
  345. }
  346. return nil
  347. }
  348. //------------------------------------------------------------
  349. // internal functions for managing the state
  350. func (cs *ConsensusState) updateHeight(height int64) {
  351. cs.metrics.Height.Set(float64(height))
  352. cs.Height = height
  353. }
  354. func (cs *ConsensusState) updateRoundStep(round int, step cstypes.RoundStepType) {
  355. cs.Round = round
  356. cs.Step = step
  357. }
  358. // enterNewRound(height, 0) at cs.StartTime.
  359. func (cs *ConsensusState) scheduleRound0(rs *cstypes.RoundState) {
  360. //cs.Logger.Info("scheduleRound0", "now", tmtime.Now(), "startTime", cs.StartTime)
  361. sleepDuration := rs.StartTime.Sub(tmtime.Now()) // nolint: gotype, gosimple
  362. cs.scheduleTimeout(sleepDuration, rs.Height, 0, cstypes.RoundStepNewHeight)
  363. }
  364. // Attempt to schedule a timeout (by sending timeoutInfo on the tickChan)
  365. func (cs *ConsensusState) scheduleTimeout(duration time.Duration, height int64, round int, step cstypes.RoundStepType) {
  366. cs.timeoutTicker.ScheduleTimeout(timeoutInfo{duration, height, round, step})
  367. }
  368. // send a msg into the receiveRoutine regarding our own proposal, block part, or vote
  369. func (cs *ConsensusState) sendInternalMessage(mi msgInfo) {
  370. select {
  371. case cs.internalMsgQueue <- mi:
  372. default:
  373. // NOTE: using the go-routine means our votes can
  374. // be processed out of order.
  375. // TODO: use CList here for strict determinism and
  376. // attempt push to internalMsgQueue in receiveRoutine
  377. cs.Logger.Info("Internal msg queue is full. Using a go-routine")
  378. go func() { cs.internalMsgQueue <- mi }()
  379. }
  380. }
  381. // Reconstruct LastCommit from SeenCommit, which we saved along with the block,
  382. // (which happens even before saving the state)
  383. func (cs *ConsensusState) reconstructLastCommit(state sm.State) {
  384. if state.LastBlockHeight == 0 {
  385. return
  386. }
  387. seenCommit := cs.blockStore.LoadSeenCommit(state.LastBlockHeight)
  388. lastPrecommits := types.NewVoteSet(state.ChainID, state.LastBlockHeight, seenCommit.Round(), types.PrecommitType, state.LastValidators)
  389. for _, precommit := range seenCommit.Precommits {
  390. if precommit == nil {
  391. continue
  392. }
  393. added, err := lastPrecommits.AddVote(precommit)
  394. if !added || err != nil {
  395. cmn.PanicCrisis(fmt.Sprintf("Failed to reconstruct LastCommit: %v", err))
  396. }
  397. }
  398. if !lastPrecommits.HasTwoThirdsMajority() {
  399. cmn.PanicSanity("Failed to reconstruct LastCommit: Does not have +2/3 maj")
  400. }
  401. cs.LastCommit = lastPrecommits
  402. }
  403. // Updates ConsensusState and increments height to match that of state.
  404. // The round becomes 0 and cs.Step becomes cstypes.RoundStepNewHeight.
  405. func (cs *ConsensusState) updateToState(state sm.State) {
  406. if cs.CommitRound > -1 && 0 < cs.Height && cs.Height != state.LastBlockHeight {
  407. cmn.PanicSanity(fmt.Sprintf("updateToState() expected state height of %v but found %v",
  408. cs.Height, state.LastBlockHeight))
  409. }
  410. if !cs.state.IsEmpty() && cs.state.LastBlockHeight+1 != cs.Height {
  411. // This might happen when someone else is mutating cs.state.
  412. // Someone forgot to pass in state.Copy() somewhere?!
  413. cmn.PanicSanity(fmt.Sprintf("Inconsistent cs.state.LastBlockHeight+1 %v vs cs.Height %v",
  414. cs.state.LastBlockHeight+1, cs.Height))
  415. }
  416. // If state isn't further out than cs.state, just ignore.
  417. // This happens when SwitchToConsensus() is called in the reactor.
  418. // We don't want to reset e.g. the Votes, but we still want to
  419. // signal the new round step, because other services (eg. mempool)
  420. // depend on having an up-to-date peer state!
  421. if !cs.state.IsEmpty() && (state.LastBlockHeight <= cs.state.LastBlockHeight) {
  422. cs.Logger.Info("Ignoring updateToState()", "newHeight", state.LastBlockHeight+1, "oldHeight", cs.state.LastBlockHeight+1)
  423. cs.newStep()
  424. return
  425. }
  426. // Reset fields based on state.
  427. validators := state.Validators
  428. lastPrecommits := (*types.VoteSet)(nil)
  429. if cs.CommitRound > -1 && cs.Votes != nil {
  430. if !cs.Votes.Precommits(cs.CommitRound).HasTwoThirdsMajority() {
  431. cmn.PanicSanity("updateToState(state) called but last Precommit round didn't have +2/3")
  432. }
  433. lastPrecommits = cs.Votes.Precommits(cs.CommitRound)
  434. }
  435. // Next desired block height
  436. height := state.LastBlockHeight + 1
  437. // RoundState fields
  438. cs.updateHeight(height)
  439. cs.updateRoundStep(0, cstypes.RoundStepNewHeight)
  440. if cs.CommitTime.IsZero() {
  441. // "Now" makes it easier to sync up dev nodes.
  442. // We add timeoutCommit to allow transactions
  443. // to be gathered for the first block.
  444. // And alternative solution that relies on clocks:
  445. // cs.StartTime = state.LastBlockTime.Add(timeoutCommit)
  446. cs.StartTime = cs.config.Commit(tmtime.Now())
  447. } else {
  448. cs.StartTime = cs.config.Commit(cs.CommitTime)
  449. }
  450. cs.Validators = validators
  451. cs.Proposal = nil
  452. cs.ProposalBlock = nil
  453. cs.ProposalBlockParts = nil
  454. cs.LockedRound = -1
  455. cs.LockedBlock = nil
  456. cs.LockedBlockParts = nil
  457. cs.ValidRound = -1
  458. cs.ValidBlock = nil
  459. cs.ValidBlockParts = nil
  460. cs.Votes = cstypes.NewHeightVoteSet(state.ChainID, height, validators)
  461. cs.CommitRound = -1
  462. cs.LastCommit = lastPrecommits
  463. cs.LastValidators = state.LastValidators
  464. cs.state = state
  465. // Finally, broadcast RoundState
  466. cs.newStep()
  467. }
  468. func (cs *ConsensusState) newStep() {
  469. rs := cs.RoundStateEvent()
  470. cs.wal.Write(rs)
  471. cs.nSteps++
  472. // newStep is called by updateToState in NewConsensusState before the eventBus is set!
  473. if cs.eventBus != nil {
  474. cs.eventBus.PublishEventNewRoundStep(rs)
  475. cs.evsw.FireEvent(types.EventNewRoundStep, &cs.RoundState)
  476. }
  477. }
  478. //-----------------------------------------
  479. // the main go routines
  480. // receiveRoutine handles messages which may cause state transitions.
  481. // it's argument (n) is the number of messages to process before exiting - use 0 to run forever
  482. // It keeps the RoundState and is the only thing that updates it.
  483. // Updates (state transitions) happen on timeouts, complete proposals, and 2/3 majorities.
  484. // ConsensusState must be locked before any internal state is updated.
  485. func (cs *ConsensusState) receiveRoutine(maxSteps int) {
  486. onExit := func(cs *ConsensusState) {
  487. // NOTE: the internalMsgQueue may have signed messages from our
  488. // priv_val that haven't hit the WAL, but its ok because
  489. // priv_val tracks LastSig
  490. // close wal now that we're done writing to it
  491. cs.wal.Stop()
  492. cs.wal.Wait()
  493. close(cs.done)
  494. }
  495. defer func() {
  496. if r := recover(); r != nil {
  497. cs.Logger.Error("CONSENSUS FAILURE!!!", "err", r, "stack", string(debug.Stack()))
  498. // stop gracefully
  499. //
  500. // NOTE: We most probably shouldn't be running any further when there is
  501. // some unexpected panic. Some unknown error happened, and so we don't
  502. // know if that will result in the validator signing an invalid thing. It
  503. // might be worthwhile to explore a mechanism for manual resuming via
  504. // some console or secure RPC system, but for now, halting the chain upon
  505. // unexpected consensus bugs sounds like the better option.
  506. onExit(cs)
  507. }
  508. }()
  509. for {
  510. if maxSteps > 0 {
  511. if cs.nSteps >= maxSteps {
  512. cs.Logger.Info("reached max steps. exiting receive routine")
  513. cs.nSteps = 0
  514. return
  515. }
  516. }
  517. rs := cs.RoundState
  518. var mi msgInfo
  519. select {
  520. case <-cs.mempool.TxsAvailable():
  521. cs.handleTxsAvailable()
  522. case mi = <-cs.peerMsgQueue:
  523. cs.wal.Write(mi)
  524. // handles proposals, block parts, votes
  525. // may generate internal events (votes, complete proposals, 2/3 majorities)
  526. cs.handleMsg(mi)
  527. case mi = <-cs.internalMsgQueue:
  528. cs.wal.WriteSync(mi) // NOTE: fsync
  529. // handles proposals, block parts, votes
  530. cs.handleMsg(mi)
  531. case ti := <-cs.timeoutTicker.Chan(): // tockChan:
  532. cs.wal.Write(ti)
  533. // if the timeout is relevant to the rs
  534. // go to the next step
  535. cs.handleTimeout(ti, rs)
  536. case <-cs.Quit():
  537. onExit(cs)
  538. return
  539. }
  540. }
  541. }
  542. // state transitions on complete-proposal, 2/3-any, 2/3-one
  543. func (cs *ConsensusState) handleMsg(mi msgInfo) {
  544. cs.mtx.Lock()
  545. defer cs.mtx.Unlock()
  546. var err error
  547. msg, peerID := mi.Msg, mi.PeerID
  548. switch msg := msg.(type) {
  549. case *ProposalMessage:
  550. // will not cause transition.
  551. // once proposal is set, we can receive block parts
  552. err = cs.setProposal(msg.Proposal)
  553. case *BlockPartMessage:
  554. // if the proposal is complete, we'll enterPrevote or tryFinalizeCommit
  555. added, err := cs.addProposalBlockPart(msg, peerID)
  556. if added {
  557. cs.statsMsgQueue <- mi
  558. }
  559. if err != nil && msg.Round != cs.Round {
  560. cs.Logger.Debug("Received block part from wrong round", "height", cs.Height, "csRound", cs.Round, "blockRound", msg.Round)
  561. err = nil
  562. }
  563. case *VoteMessage:
  564. // attempt to add the vote and dupeout the validator if its a duplicate signature
  565. // if the vote gives us a 2/3-any or 2/3-one, we transition
  566. added, err := cs.tryAddVote(msg.Vote, peerID)
  567. if added {
  568. cs.statsMsgQueue <- mi
  569. }
  570. if err == ErrAddingVote {
  571. // TODO: punish peer
  572. // We probably don't want to stop the peer here. The vote does not
  573. // necessarily comes from a malicious peer but can be just broadcasted by
  574. // a typical peer.
  575. // https://github.com/tendermint/tendermint/issues/1281
  576. }
  577. // NOTE: the vote is broadcast to peers by the reactor listening
  578. // for vote events
  579. // TODO: If rs.Height == vote.Height && rs.Round < vote.Round,
  580. // the peer is sending us CatchupCommit precommits.
  581. // We could make note of this and help filter in broadcastHasVoteMessage().
  582. default:
  583. cs.Logger.Error("Unknown msg type", reflect.TypeOf(msg))
  584. }
  585. if err != nil {
  586. cs.Logger.Error("Error with msg", "height", cs.Height, "round", cs.Round, "type", reflect.TypeOf(msg), "peer", peerID, "err", err, "msg", msg)
  587. }
  588. }
  589. func (cs *ConsensusState) handleTimeout(ti timeoutInfo, rs cstypes.RoundState) {
  590. cs.Logger.Debug("Received tock", "timeout", ti.Duration, "height", ti.Height, "round", ti.Round, "step", ti.Step)
  591. // timeouts must be for current height, round, step
  592. if ti.Height != rs.Height || ti.Round < rs.Round || (ti.Round == rs.Round && ti.Step < rs.Step) {
  593. cs.Logger.Debug("Ignoring tock because we're ahead", "height", rs.Height, "round", rs.Round, "step", rs.Step)
  594. return
  595. }
  596. // the timeout will now cause a state transition
  597. cs.mtx.Lock()
  598. defer cs.mtx.Unlock()
  599. switch ti.Step {
  600. case cstypes.RoundStepNewHeight:
  601. // NewRound event fired from enterNewRound.
  602. // XXX: should we fire timeout here (for timeout commit)?
  603. cs.enterNewRound(ti.Height, 0)
  604. case cstypes.RoundStepNewRound:
  605. cs.enterPropose(ti.Height, 0)
  606. case cstypes.RoundStepPropose:
  607. cs.eventBus.PublishEventTimeoutPropose(cs.RoundStateEvent())
  608. cs.enterPrevote(ti.Height, ti.Round)
  609. case cstypes.RoundStepPrevoteWait:
  610. cs.eventBus.PublishEventTimeoutWait(cs.RoundStateEvent())
  611. cs.enterPrecommit(ti.Height, ti.Round)
  612. case cstypes.RoundStepPrecommitWait:
  613. cs.eventBus.PublishEventTimeoutWait(cs.RoundStateEvent())
  614. cs.enterPrecommit(ti.Height, ti.Round)
  615. cs.enterNewRound(ti.Height, ti.Round+1)
  616. default:
  617. panic(fmt.Sprintf("Invalid timeout step: %v", ti.Step))
  618. }
  619. }
  620. func (cs *ConsensusState) handleTxsAvailable() {
  621. cs.mtx.Lock()
  622. defer cs.mtx.Unlock()
  623. // we only need to do this for round 0
  624. cs.enterPropose(cs.Height, 0)
  625. }
  626. //-----------------------------------------------------------------------------
  627. // State functions
  628. // Used internally by handleTimeout and handleMsg to make state transitions
  629. // Enter: `timeoutNewHeight` by startTime (commitTime+timeoutCommit),
  630. // or, if SkipTimeout==true, after receiving all precommits from (height,round-1)
  631. // Enter: `timeoutPrecommits` after any +2/3 precommits from (height,round-1)
  632. // Enter: +2/3 precommits for nil at (height,round-1)
  633. // Enter: +2/3 prevotes any or +2/3 precommits for block or any from (height, round)
  634. // NOTE: cs.StartTime was already set for height.
  635. func (cs *ConsensusState) enterNewRound(height int64, round int) {
  636. logger := cs.Logger.With("height", height, "round", round)
  637. if cs.Height != height || round < cs.Round || (cs.Round == round && cs.Step != cstypes.RoundStepNewHeight) {
  638. logger.Debug(fmt.Sprintf("enterNewRound(%v/%v): Invalid args. Current step: %v/%v/%v", height, round, cs.Height, cs.Round, cs.Step))
  639. return
  640. }
  641. if now := tmtime.Now(); cs.StartTime.After(now) {
  642. logger.Info("Need to set a buffer and log message here for sanity.", "startTime", cs.StartTime, "now", now)
  643. }
  644. logger.Info(fmt.Sprintf("enterNewRound(%v/%v). Current: %v/%v/%v", height, round, cs.Height, cs.Round, cs.Step))
  645. // Increment validators if necessary
  646. validators := cs.Validators
  647. if cs.Round < round {
  648. validators = validators.Copy()
  649. validators.IncrementAccum(round - cs.Round)
  650. }
  651. // Setup new round
  652. // we don't fire newStep for this step,
  653. // but we fire an event, so update the round step first
  654. cs.updateRoundStep(round, cstypes.RoundStepNewRound)
  655. cs.Validators = validators
  656. if round == 0 {
  657. // We've already reset these upon new height,
  658. // and meanwhile we might have received a proposal
  659. // for round 0.
  660. } else {
  661. logger.Info("Resetting Proposal info")
  662. cs.Proposal = nil
  663. cs.ProposalBlock = nil
  664. cs.ProposalBlockParts = nil
  665. }
  666. cs.Votes.SetRound(round + 1) // also track next round (round+1) to allow round-skipping
  667. cs.triggeredTimeoutPrecommit = false
  668. cs.eventBus.PublishEventNewRound(cs.RoundStateEvent())
  669. cs.metrics.Rounds.Set(float64(round))
  670. // Wait for txs to be available in the mempool
  671. // before we enterPropose in round 0. If the last block changed the app hash,
  672. // we may need an empty "proof" block, and enterPropose immediately.
  673. waitForTxs := cs.config.WaitForTxs() && round == 0 && !cs.needProofBlock(height)
  674. if waitForTxs {
  675. if cs.config.CreateEmptyBlocksInterval > 0 {
  676. cs.scheduleTimeout(cs.config.CreateEmptyBlocksInterval, height, round,
  677. cstypes.RoundStepNewRound)
  678. }
  679. go cs.proposalHeartbeat(height, round)
  680. } else {
  681. cs.enterPropose(height, round)
  682. }
  683. }
  684. // needProofBlock returns true on the first height (so the genesis app hash is signed right away)
  685. // and where the last block (height-1) caused the app hash to change
  686. func (cs *ConsensusState) needProofBlock(height int64) bool {
  687. if height == 1 {
  688. return true
  689. }
  690. lastBlockMeta := cs.blockStore.LoadBlockMeta(height - 1)
  691. return !bytes.Equal(cs.state.AppHash, lastBlockMeta.Header.AppHash)
  692. }
  693. func (cs *ConsensusState) proposalHeartbeat(height int64, round int) {
  694. counter := 0
  695. addr := cs.privValidator.GetAddress()
  696. valIndex, _ := cs.Validators.GetByAddress(addr)
  697. chainID := cs.state.ChainID
  698. for {
  699. rs := cs.GetRoundState()
  700. // if we've already moved on, no need to send more heartbeats
  701. if rs.Step > cstypes.RoundStepNewRound || rs.Round > round || rs.Height > height {
  702. return
  703. }
  704. heartbeat := &types.Heartbeat{
  705. Height: rs.Height,
  706. Round: rs.Round,
  707. Sequence: counter,
  708. ValidatorAddress: addr,
  709. ValidatorIndex: valIndex,
  710. }
  711. cs.privValidator.SignHeartbeat(chainID, heartbeat)
  712. cs.eventBus.PublishEventProposalHeartbeat(types.EventDataProposalHeartbeat{heartbeat})
  713. cs.evsw.FireEvent(types.EventProposalHeartbeat, heartbeat)
  714. counter++
  715. time.Sleep(proposalHeartbeatIntervalSeconds * time.Second)
  716. }
  717. }
  718. // Enter (CreateEmptyBlocks): from enterNewRound(height,round)
  719. // Enter (CreateEmptyBlocks, CreateEmptyBlocksInterval > 0 ): after enterNewRound(height,round), after timeout of CreateEmptyBlocksInterval
  720. // Enter (!CreateEmptyBlocks) : after enterNewRound(height,round), once txs are in the mempool
  721. func (cs *ConsensusState) enterPropose(height int64, round int) {
  722. logger := cs.Logger.With("height", height, "round", round)
  723. if cs.Height != height || round < cs.Round || (cs.Round == round && cstypes.RoundStepPropose <= cs.Step) {
  724. logger.Debug(fmt.Sprintf("enterPropose(%v/%v): Invalid args. Current step: %v/%v/%v", height, round, cs.Height, cs.Round, cs.Step))
  725. return
  726. }
  727. logger.Info(fmt.Sprintf("enterPropose(%v/%v). Current: %v/%v/%v", height, round, cs.Height, cs.Round, cs.Step))
  728. defer func() {
  729. // Done enterPropose:
  730. cs.updateRoundStep(round, cstypes.RoundStepPropose)
  731. cs.newStep()
  732. // If we have the whole proposal + POL, then goto Prevote now.
  733. // else, we'll enterPrevote when the rest of the proposal is received (in AddProposalBlockPart),
  734. // or else after timeoutPropose
  735. if cs.isProposalComplete() {
  736. cs.enterPrevote(height, cs.Round)
  737. }
  738. }()
  739. // If we don't get the proposal and all block parts quick enough, enterPrevote
  740. cs.scheduleTimeout(cs.config.Propose(round), height, round, cstypes.RoundStepPropose)
  741. // Nothing more to do if we're not a validator
  742. if cs.privValidator == nil {
  743. logger.Debug("This node is not a validator")
  744. return
  745. }
  746. // if not a validator, we're done
  747. if !cs.Validators.HasAddress(cs.privValidator.GetAddress()) {
  748. logger.Debug("This node is not a validator", "addr", cs.privValidator.GetAddress(), "vals", cs.Validators)
  749. return
  750. }
  751. logger.Debug("This node is a validator")
  752. if cs.isProposer() {
  753. logger.Info("enterPropose: Our turn to propose", "proposer", cs.Validators.GetProposer().Address, "privValidator", cs.privValidator)
  754. cs.decideProposal(height, round)
  755. } else {
  756. logger.Info("enterPropose: Not our turn to propose", "proposer", cs.Validators.GetProposer().Address, "privValidator", cs.privValidator)
  757. }
  758. }
  759. func (cs *ConsensusState) isProposer() bool {
  760. return bytes.Equal(cs.Validators.GetProposer().Address, cs.privValidator.GetAddress())
  761. }
  762. func (cs *ConsensusState) defaultDecideProposal(height int64, round int) {
  763. var block *types.Block
  764. var blockParts *types.PartSet
  765. // Decide on block
  766. if cs.ValidBlock != nil {
  767. // If there is valid block, choose that.
  768. block, blockParts = cs.ValidBlock, cs.ValidBlockParts
  769. } else {
  770. // Create a new proposal block from state/txs from the mempool.
  771. block, blockParts = cs.createProposalBlock()
  772. if block == nil { // on error
  773. return
  774. }
  775. }
  776. // Make proposal
  777. polRound, polBlockID := cs.Votes.POLInfo()
  778. proposal := types.NewProposal(height, round, blockParts.Header(), polRound, polBlockID)
  779. if err := cs.privValidator.SignProposal(cs.state.ChainID, proposal); err == nil {
  780. // Set fields
  781. /* fields set by setProposal and addBlockPart
  782. cs.Proposal = proposal
  783. cs.ProposalBlock = block
  784. cs.ProposalBlockParts = blockParts
  785. */
  786. // send proposal and block parts on internal msg queue
  787. cs.sendInternalMessage(msgInfo{&ProposalMessage{proposal}, ""})
  788. for i := 0; i < blockParts.Total(); i++ {
  789. part := blockParts.GetPart(i)
  790. cs.sendInternalMessage(msgInfo{&BlockPartMessage{cs.Height, cs.Round, part}, ""})
  791. }
  792. cs.Logger.Info("Signed proposal", "height", height, "round", round, "proposal", proposal)
  793. cs.Logger.Debug(fmt.Sprintf("Signed proposal block: %v", block))
  794. } else {
  795. if !cs.replayMode {
  796. cs.Logger.Error("enterPropose: Error signing proposal", "height", height, "round", round, "err", err)
  797. }
  798. }
  799. }
  800. // Returns true if the proposal block is complete &&
  801. // (if POLRound was proposed, we have +2/3 prevotes from there).
  802. func (cs *ConsensusState) isProposalComplete() bool {
  803. if cs.Proposal == nil || cs.ProposalBlock == nil {
  804. return false
  805. }
  806. // we have the proposal. if there's a POLRound,
  807. // make sure we have the prevotes from it too
  808. if cs.Proposal.POLRound < 0 {
  809. return true
  810. }
  811. // if this is false the proposer is lying or we haven't received the POL yet
  812. return cs.Votes.Prevotes(cs.Proposal.POLRound).HasTwoThirdsMajority()
  813. }
  814. // Create the next block to propose and return it.
  815. // We really only need to return the parts, but the block
  816. // is returned for convenience so we can log the proposal block.
  817. // Returns nil block upon error.
  818. // NOTE: keep it side-effect free for clarity.
  819. func (cs *ConsensusState) createProposalBlock() (block *types.Block, blockParts *types.PartSet) {
  820. var commit *types.Commit
  821. if cs.Height == 1 {
  822. // We're creating a proposal for the first block.
  823. // The commit is empty, but not nil.
  824. commit = &types.Commit{}
  825. } else if cs.LastCommit.HasTwoThirdsMajority() {
  826. // Make the commit from LastCommit
  827. commit = cs.LastCommit.MakeCommit()
  828. } else {
  829. // This shouldn't happen.
  830. cs.Logger.Error("enterPropose: Cannot propose anything: No commit for the previous block.")
  831. return
  832. }
  833. maxBytes := cs.state.ConsensusParams.BlockSize.MaxBytes
  834. maxGas := cs.state.ConsensusParams.BlockSize.MaxGas
  835. // bound evidence to 1/10th of the block
  836. evidence := cs.evpool.PendingEvidence(types.MaxEvidenceBytesPerBlock(maxBytes))
  837. // Mempool validated transactions
  838. txs := cs.mempool.ReapMaxBytesMaxGas(types.MaxDataBytes(
  839. maxBytes,
  840. cs.state.Validators.Size(),
  841. len(evidence),
  842. ), maxGas)
  843. proposerAddr := cs.privValidator.GetAddress()
  844. block, parts := cs.state.MakeBlock(cs.Height, txs, commit, evidence, proposerAddr)
  845. return block, parts
  846. }
  847. // Enter: `timeoutPropose` after entering Propose.
  848. // Enter: proposal block and POL is ready.
  849. // Prevote for LockedBlock if we're locked, or ProposalBlock if valid.
  850. // Otherwise vote nil.
  851. func (cs *ConsensusState) enterPrevote(height int64, round int) {
  852. if cs.Height != height || round < cs.Round || (cs.Round == round && cstypes.RoundStepPrevote <= cs.Step) {
  853. cs.Logger.Debug(fmt.Sprintf("enterPrevote(%v/%v): Invalid args. Current step: %v/%v/%v", height, round, cs.Height, cs.Round, cs.Step))
  854. return
  855. }
  856. defer func() {
  857. // Done enterPrevote:
  858. cs.updateRoundStep(round, cstypes.RoundStepPrevote)
  859. cs.newStep()
  860. }()
  861. // fire event for how we got here
  862. if cs.isProposalComplete() {
  863. cs.eventBus.PublishEventCompleteProposal(cs.RoundStateEvent())
  864. } else {
  865. // we received +2/3 prevotes for a future round
  866. // TODO: catchup event?
  867. }
  868. cs.Logger.Info(fmt.Sprintf("enterPrevote(%v/%v). Current: %v/%v/%v", height, round, cs.Height, cs.Round, cs.Step))
  869. // Sign and broadcast vote as necessary
  870. cs.doPrevote(height, round)
  871. // Once `addVote` hits any +2/3 prevotes, we will go to PrevoteWait
  872. // (so we have more time to try and collect +2/3 prevotes for a single block)
  873. }
  874. func (cs *ConsensusState) defaultDoPrevote(height int64, round int) {
  875. logger := cs.Logger.With("height", height, "round", round)
  876. // If a block is locked, prevote that.
  877. if cs.LockedBlock != nil {
  878. logger.Info("enterPrevote: Block was locked")
  879. cs.signAddVote(types.PrevoteType, cs.LockedBlock.Hash(), cs.LockedBlockParts.Header())
  880. return
  881. }
  882. // If ProposalBlock is nil, prevote nil.
  883. if cs.ProposalBlock == nil {
  884. logger.Info("enterPrevote: ProposalBlock is nil")
  885. cs.signAddVote(types.PrevoteType, nil, types.PartSetHeader{})
  886. return
  887. }
  888. // Validate proposal block
  889. err := cs.blockExec.ValidateBlock(cs.state, cs.ProposalBlock)
  890. if err != nil {
  891. // ProposalBlock is invalid, prevote nil.
  892. logger.Error("enterPrevote: ProposalBlock is invalid", "err", err)
  893. cs.signAddVote(types.PrevoteType, nil, types.PartSetHeader{})
  894. return
  895. }
  896. // Prevote cs.ProposalBlock
  897. // NOTE: the proposal signature is validated when it is received,
  898. // and the proposal block parts are validated as they are received (against the merkle hash in the proposal)
  899. logger.Info("enterPrevote: ProposalBlock is valid")
  900. cs.signAddVote(types.PrevoteType, cs.ProposalBlock.Hash(), cs.ProposalBlockParts.Header())
  901. }
  902. // Enter: any +2/3 prevotes at next round.
  903. func (cs *ConsensusState) enterPrevoteWait(height int64, round int) {
  904. logger := cs.Logger.With("height", height, "round", round)
  905. if cs.Height != height || round < cs.Round || (cs.Round == round && cstypes.RoundStepPrevoteWait <= cs.Step) {
  906. logger.Debug(fmt.Sprintf("enterPrevoteWait(%v/%v): Invalid args. Current step: %v/%v/%v", height, round, cs.Height, cs.Round, cs.Step))
  907. return
  908. }
  909. if !cs.Votes.Prevotes(round).HasTwoThirdsAny() {
  910. cmn.PanicSanity(fmt.Sprintf("enterPrevoteWait(%v/%v), but Prevotes does not have any +2/3 votes", height, round))
  911. }
  912. logger.Info(fmt.Sprintf("enterPrevoteWait(%v/%v). Current: %v/%v/%v", height, round, cs.Height, cs.Round, cs.Step))
  913. defer func() {
  914. // Done enterPrevoteWait:
  915. cs.updateRoundStep(round, cstypes.RoundStepPrevoteWait)
  916. cs.newStep()
  917. }()
  918. // Wait for some more prevotes; enterPrecommit
  919. cs.scheduleTimeout(cs.config.Prevote(round), height, round, cstypes.RoundStepPrevoteWait)
  920. }
  921. // Enter: `timeoutPrevote` after any +2/3 prevotes.
  922. // Enter: `timeoutPrecommit` after any +2/3 precommits.
  923. // Enter: +2/3 precomits for block or nil.
  924. // Lock & precommit the ProposalBlock if we have enough prevotes for it (a POL in this round)
  925. // else, unlock an existing lock and precommit nil if +2/3 of prevotes were nil,
  926. // else, precommit nil otherwise.
  927. func (cs *ConsensusState) enterPrecommit(height int64, round int) {
  928. logger := cs.Logger.With("height", height, "round", round)
  929. if cs.Height != height || round < cs.Round || (cs.Round == round && cstypes.RoundStepPrecommit <= cs.Step) {
  930. logger.Debug(fmt.Sprintf("enterPrecommit(%v/%v): Invalid args. Current step: %v/%v/%v", height, round, cs.Height, cs.Round, cs.Step))
  931. return
  932. }
  933. logger.Info(fmt.Sprintf("enterPrecommit(%v/%v). Current: %v/%v/%v", height, round, cs.Height, cs.Round, cs.Step))
  934. defer func() {
  935. // Done enterPrecommit:
  936. cs.updateRoundStep(round, cstypes.RoundStepPrecommit)
  937. cs.newStep()
  938. }()
  939. // check for a polka
  940. blockID, ok := cs.Votes.Prevotes(round).TwoThirdsMajority()
  941. // If we don't have a polka, we must precommit nil.
  942. if !ok {
  943. if cs.LockedBlock != nil {
  944. logger.Info("enterPrecommit: No +2/3 prevotes during enterPrecommit while we're locked. Precommitting nil")
  945. } else {
  946. logger.Info("enterPrecommit: No +2/3 prevotes during enterPrecommit. Precommitting nil.")
  947. }
  948. cs.signAddVote(types.PrecommitType, nil, types.PartSetHeader{})
  949. return
  950. }
  951. // At this point +2/3 prevoted for a particular block or nil.
  952. cs.eventBus.PublishEventPolka(cs.RoundStateEvent())
  953. // the latest POLRound should be this round.
  954. polRound, _ := cs.Votes.POLInfo()
  955. if polRound < round {
  956. cmn.PanicSanity(fmt.Sprintf("This POLRound should be %v but got %v", round, polRound))
  957. }
  958. // +2/3 prevoted nil. Unlock and precommit nil.
  959. if len(blockID.Hash) == 0 {
  960. if cs.LockedBlock == nil {
  961. logger.Info("enterPrecommit: +2/3 prevoted for nil.")
  962. } else {
  963. logger.Info("enterPrecommit: +2/3 prevoted for nil. Unlocking")
  964. cs.LockedRound = -1
  965. cs.LockedBlock = nil
  966. cs.LockedBlockParts = nil
  967. cs.eventBus.PublishEventUnlock(cs.RoundStateEvent())
  968. }
  969. cs.signAddVote(types.PrecommitType, nil, types.PartSetHeader{})
  970. return
  971. }
  972. // At this point, +2/3 prevoted for a particular block.
  973. // If we're already locked on that block, precommit it, and update the LockedRound
  974. if cs.LockedBlock.HashesTo(blockID.Hash) {
  975. logger.Info("enterPrecommit: +2/3 prevoted locked block. Relocking")
  976. cs.LockedRound = round
  977. cs.eventBus.PublishEventRelock(cs.RoundStateEvent())
  978. cs.signAddVote(types.PrecommitType, blockID.Hash, blockID.PartsHeader)
  979. return
  980. }
  981. // If +2/3 prevoted for proposal block, stage and precommit it
  982. if cs.ProposalBlock.HashesTo(blockID.Hash) {
  983. logger.Info("enterPrecommit: +2/3 prevoted proposal block. Locking", "hash", blockID.Hash)
  984. // Validate the block.
  985. if err := cs.blockExec.ValidateBlock(cs.state, cs.ProposalBlock); err != nil {
  986. cmn.PanicConsensus(fmt.Sprintf("enterPrecommit: +2/3 prevoted for an invalid block: %v", err))
  987. }
  988. cs.LockedRound = round
  989. cs.LockedBlock = cs.ProposalBlock
  990. cs.LockedBlockParts = cs.ProposalBlockParts
  991. cs.eventBus.PublishEventLock(cs.RoundStateEvent())
  992. cs.signAddVote(types.PrecommitType, blockID.Hash, blockID.PartsHeader)
  993. return
  994. }
  995. // There was a polka in this round for a block we don't have.
  996. // Fetch that block, unlock, and precommit nil.
  997. // The +2/3 prevotes for this round is the POL for our unlock.
  998. // TODO: In the future save the POL prevotes for justification.
  999. cs.LockedRound = -1
  1000. cs.LockedBlock = nil
  1001. cs.LockedBlockParts = nil
  1002. if !cs.ProposalBlockParts.HasHeader(blockID.PartsHeader) {
  1003. cs.ProposalBlock = nil
  1004. cs.ProposalBlockParts = types.NewPartSetFromHeader(blockID.PartsHeader)
  1005. }
  1006. cs.eventBus.PublishEventUnlock(cs.RoundStateEvent())
  1007. cs.signAddVote(types.PrecommitType, nil, types.PartSetHeader{})
  1008. }
  1009. // Enter: any +2/3 precommits for next round.
  1010. func (cs *ConsensusState) enterPrecommitWait(height int64, round int) {
  1011. logger := cs.Logger.With("height", height, "round", round)
  1012. if cs.Height != height || round < cs.Round || (cs.Round == round && cs.triggeredTimeoutPrecommit) {
  1013. logger.Debug(
  1014. fmt.Sprintf(
  1015. "enterPrecommitWait(%v/%v): Invalid args. "+
  1016. "Current state is Height/Round: %v/%v/, triggeredTimeoutPrecommit:%v",
  1017. height, round, cs.Height, cs.Round, cs.triggeredTimeoutPrecommit))
  1018. return
  1019. }
  1020. if !cs.Votes.Precommits(round).HasTwoThirdsAny() {
  1021. cmn.PanicSanity(fmt.Sprintf("enterPrecommitWait(%v/%v), but Precommits does not have any +2/3 votes", height, round))
  1022. }
  1023. logger.Info(fmt.Sprintf("enterPrecommitWait(%v/%v). Current: %v/%v/%v", height, round, cs.Height, cs.Round, cs.Step))
  1024. defer func() {
  1025. // Done enterPrecommitWait:
  1026. cs.triggeredTimeoutPrecommit = true
  1027. cs.newStep()
  1028. }()
  1029. // Wait for some more precommits; enterNewRound
  1030. cs.scheduleTimeout(cs.config.Precommit(round), height, round, cstypes.RoundStepPrecommitWait)
  1031. }
  1032. // Enter: +2/3 precommits for block
  1033. func (cs *ConsensusState) enterCommit(height int64, commitRound int) {
  1034. logger := cs.Logger.With("height", height, "commitRound", commitRound)
  1035. if cs.Height != height || cstypes.RoundStepCommit <= cs.Step {
  1036. logger.Debug(fmt.Sprintf("enterCommit(%v/%v): Invalid args. Current step: %v/%v/%v", height, commitRound, cs.Height, cs.Round, cs.Step))
  1037. return
  1038. }
  1039. logger.Info(fmt.Sprintf("enterCommit(%v/%v). Current: %v/%v/%v", height, commitRound, cs.Height, cs.Round, cs.Step))
  1040. defer func() {
  1041. // Done enterCommit:
  1042. // keep cs.Round the same, commitRound points to the right Precommits set.
  1043. cs.updateRoundStep(cs.Round, cstypes.RoundStepCommit)
  1044. cs.CommitRound = commitRound
  1045. cs.CommitTime = tmtime.Now()
  1046. cs.newStep()
  1047. // Maybe finalize immediately.
  1048. cs.tryFinalizeCommit(height)
  1049. }()
  1050. blockID, ok := cs.Votes.Precommits(commitRound).TwoThirdsMajority()
  1051. if !ok {
  1052. cmn.PanicSanity("RunActionCommit() expects +2/3 precommits")
  1053. }
  1054. // The Locked* fields no longer matter.
  1055. // Move them over to ProposalBlock if they match the commit hash,
  1056. // otherwise they'll be cleared in updateToState.
  1057. if cs.LockedBlock.HashesTo(blockID.Hash) {
  1058. logger.Info("Commit is for locked block. Set ProposalBlock=LockedBlock", "blockHash", blockID.Hash)
  1059. cs.ProposalBlock = cs.LockedBlock
  1060. cs.ProposalBlockParts = cs.LockedBlockParts
  1061. }
  1062. // If we don't have the block being committed, set up to get it.
  1063. if !cs.ProposalBlock.HashesTo(blockID.Hash) {
  1064. if !cs.ProposalBlockParts.HasHeader(blockID.PartsHeader) {
  1065. logger.Info("Commit is for a block we don't know about. Set ProposalBlock=nil", "proposal", cs.ProposalBlock.Hash(), "commit", blockID.Hash)
  1066. // We're getting the wrong block.
  1067. // Set up ProposalBlockParts and keep waiting.
  1068. cs.ProposalBlock = nil
  1069. cs.ProposalBlockParts = types.NewPartSetFromHeader(blockID.PartsHeader)
  1070. } else {
  1071. // We just need to keep waiting.
  1072. }
  1073. }
  1074. }
  1075. // If we have the block AND +2/3 commits for it, finalize.
  1076. func (cs *ConsensusState) tryFinalizeCommit(height int64) {
  1077. logger := cs.Logger.With("height", height)
  1078. if cs.Height != height {
  1079. cmn.PanicSanity(fmt.Sprintf("tryFinalizeCommit() cs.Height: %v vs height: %v", cs.Height, height))
  1080. }
  1081. blockID, ok := cs.Votes.Precommits(cs.CommitRound).TwoThirdsMajority()
  1082. if !ok || len(blockID.Hash) == 0 {
  1083. logger.Error("Attempt to finalize failed. There was no +2/3 majority, or +2/3 was for <nil>.")
  1084. return
  1085. }
  1086. if !cs.ProposalBlock.HashesTo(blockID.Hash) {
  1087. // TODO: this happens every time if we're not a validator (ugly logs)
  1088. // TODO: ^^ wait, why does it matter that we're a validator?
  1089. logger.Info("Attempt to finalize failed. We don't have the commit block.", "proposal-block", cs.ProposalBlock.Hash(), "commit-block", blockID.Hash)
  1090. return
  1091. }
  1092. // go
  1093. cs.finalizeCommit(height)
  1094. }
  1095. // Increment height and goto cstypes.RoundStepNewHeight
  1096. func (cs *ConsensusState) finalizeCommit(height int64) {
  1097. if cs.Height != height || cs.Step != cstypes.RoundStepCommit {
  1098. cs.Logger.Debug(fmt.Sprintf("finalizeCommit(%v): Invalid args. Current step: %v/%v/%v", height, cs.Height, cs.Round, cs.Step))
  1099. return
  1100. }
  1101. blockID, ok := cs.Votes.Precommits(cs.CommitRound).TwoThirdsMajority()
  1102. block, blockParts := cs.ProposalBlock, cs.ProposalBlockParts
  1103. if !ok {
  1104. cmn.PanicSanity(fmt.Sprintf("Cannot finalizeCommit, commit does not have two thirds majority"))
  1105. }
  1106. if !blockParts.HasHeader(blockID.PartsHeader) {
  1107. cmn.PanicSanity(fmt.Sprintf("Expected ProposalBlockParts header to be commit header"))
  1108. }
  1109. if !block.HashesTo(blockID.Hash) {
  1110. cmn.PanicSanity(fmt.Sprintf("Cannot finalizeCommit, ProposalBlock does not hash to commit hash"))
  1111. }
  1112. if err := cs.blockExec.ValidateBlock(cs.state, block); err != nil {
  1113. cmn.PanicConsensus(fmt.Sprintf("+2/3 committed an invalid block: %v", err))
  1114. }
  1115. cs.Logger.Info(fmt.Sprintf("Finalizing commit of block with %d txs", block.NumTxs),
  1116. "height", block.Height, "hash", block.Hash(), "root", block.AppHash)
  1117. cs.Logger.Info(fmt.Sprintf("%v", block))
  1118. fail.Fail() // XXX
  1119. // Save to blockStore.
  1120. if cs.blockStore.Height() < block.Height {
  1121. // NOTE: the seenCommit is local justification to commit this block,
  1122. // but may differ from the LastCommit included in the next block
  1123. precommits := cs.Votes.Precommits(cs.CommitRound)
  1124. seenCommit := precommits.MakeCommit()
  1125. cs.blockStore.SaveBlock(block, blockParts, seenCommit)
  1126. } else {
  1127. // Happens during replay if we already saved the block but didn't commit
  1128. cs.Logger.Info("Calling finalizeCommit on already stored block", "height", block.Height)
  1129. }
  1130. fail.Fail() // XXX
  1131. // Write EndHeightMessage{} for this height, implying that the blockstore
  1132. // has saved the block.
  1133. //
  1134. // If we crash before writing this EndHeightMessage{}, we will recover by
  1135. // running ApplyBlock during the ABCI handshake when we restart. If we
  1136. // didn't save the block to the blockstore before writing
  1137. // EndHeightMessage{}, we'd have to change WAL replay -- currently it
  1138. // complains about replaying for heights where an #ENDHEIGHT entry already
  1139. // exists.
  1140. //
  1141. // Either way, the ConsensusState should not be resumed until we
  1142. // successfully call ApplyBlock (ie. later here, or in Handshake after
  1143. // restart).
  1144. cs.wal.WriteSync(EndHeightMessage{height}) // NOTE: fsync
  1145. fail.Fail() // XXX
  1146. // Create a copy of the state for staging and an event cache for txs.
  1147. stateCopy := cs.state.Copy()
  1148. // Execute and commit the block, update and save the state, and update the mempool.
  1149. // NOTE The block.AppHash wont reflect these txs until the next block.
  1150. var err error
  1151. stateCopy, err = cs.blockExec.ApplyBlock(stateCopy, types.BlockID{block.Hash(), blockParts.Header()}, block)
  1152. if err != nil {
  1153. cs.Logger.Error("Error on ApplyBlock. Did the application crash? Please restart tendermint", "err", err)
  1154. err := cmn.Kill()
  1155. if err != nil {
  1156. cs.Logger.Error("Failed to kill this process - please do so manually", "err", err)
  1157. }
  1158. return
  1159. }
  1160. fail.Fail() // XXX
  1161. // must be called before we update state
  1162. cs.recordMetrics(height, block)
  1163. // NewHeightStep!
  1164. cs.updateToState(stateCopy)
  1165. fail.Fail() // XXX
  1166. // cs.StartTime is already set.
  1167. // Schedule Round0 to start soon.
  1168. cs.scheduleRound0(&cs.RoundState)
  1169. // By here,
  1170. // * cs.Height has been increment to height+1
  1171. // * cs.Step is now cstypes.RoundStepNewHeight
  1172. // * cs.StartTime is set to when we will start round0.
  1173. }
  1174. func (cs *ConsensusState) recordMetrics(height int64, block *types.Block) {
  1175. cs.metrics.Validators.Set(float64(cs.Validators.Size()))
  1176. cs.metrics.ValidatorsPower.Set(float64(cs.Validators.TotalVotingPower()))
  1177. missingValidators := 0
  1178. missingValidatorsPower := int64(0)
  1179. for i, val := range cs.Validators.Validators {
  1180. var vote *types.Vote
  1181. if i < len(block.LastCommit.Precommits) {
  1182. vote = block.LastCommit.Precommits[i]
  1183. }
  1184. if vote == nil {
  1185. missingValidators++
  1186. missingValidatorsPower += val.VotingPower
  1187. }
  1188. }
  1189. cs.metrics.MissingValidators.Set(float64(missingValidators))
  1190. cs.metrics.MissingValidatorsPower.Set(float64(missingValidatorsPower))
  1191. cs.metrics.ByzantineValidators.Set(float64(len(block.Evidence.Evidence)))
  1192. byzantineValidatorsPower := int64(0)
  1193. for _, ev := range block.Evidence.Evidence {
  1194. if _, val := cs.Validators.GetByAddress(ev.Address()); val != nil {
  1195. byzantineValidatorsPower += val.VotingPower
  1196. }
  1197. }
  1198. cs.metrics.ByzantineValidatorsPower.Set(float64(byzantineValidatorsPower))
  1199. if height > 1 {
  1200. lastBlockMeta := cs.blockStore.LoadBlockMeta(height - 1)
  1201. cs.metrics.BlockIntervalSeconds.Set(
  1202. block.Time.Sub(lastBlockMeta.Header.Time).Seconds(),
  1203. )
  1204. }
  1205. cs.metrics.NumTxs.Set(float64(block.NumTxs))
  1206. cs.metrics.BlockSizeBytes.Set(float64(block.Size()))
  1207. cs.metrics.TotalTxs.Set(float64(block.TotalTxs))
  1208. cs.metrics.CommittedHeight.Set(float64(block.Height))
  1209. }
  1210. //-----------------------------------------------------------------------------
  1211. func (cs *ConsensusState) defaultSetProposal(proposal *types.Proposal) error {
  1212. // Already have one
  1213. // TODO: possibly catch double proposals
  1214. if cs.Proposal != nil {
  1215. return nil
  1216. }
  1217. // Does not apply
  1218. if proposal.Height != cs.Height || proposal.Round != cs.Round {
  1219. return nil
  1220. }
  1221. // We don't care about the proposal if we're already in cstypes.RoundStepCommit.
  1222. if cstypes.RoundStepCommit <= cs.Step {
  1223. return nil
  1224. }
  1225. // Verify POLRound, which must be -1 or between 0 and proposal.Round exclusive.
  1226. if proposal.POLRound != -1 &&
  1227. (proposal.POLRound < 0 || proposal.Round <= proposal.POLRound) {
  1228. return ErrInvalidProposalPOLRound
  1229. }
  1230. // Verify signature
  1231. if !cs.Validators.GetProposer().PubKey.VerifyBytes(proposal.SignBytes(cs.state.ChainID), proposal.Signature) {
  1232. return ErrInvalidProposalSignature
  1233. }
  1234. cs.Proposal = proposal
  1235. cs.ProposalBlockParts = types.NewPartSetFromHeader(proposal.BlockPartsHeader)
  1236. cs.Logger.Info("Received proposal", "proposal", proposal)
  1237. return nil
  1238. }
  1239. // NOTE: block is not necessarily valid.
  1240. // Asynchronously triggers either enterPrevote (before we timeout of propose) or tryFinalizeCommit, once we have the full block.
  1241. func (cs *ConsensusState) addProposalBlockPart(msg *BlockPartMessage, peerID p2p.ID) (added bool, err error) {
  1242. height, round, part := msg.Height, msg.Round, msg.Part
  1243. // Blocks might be reused, so round mismatch is OK
  1244. if cs.Height != height {
  1245. cs.Logger.Debug("Received block part from wrong height", "height", height, "round", round)
  1246. return false, nil
  1247. }
  1248. // We're not expecting a block part.
  1249. if cs.ProposalBlockParts == nil {
  1250. // NOTE: this can happen when we've gone to a higher round and
  1251. // then receive parts from the previous round - not necessarily a bad peer.
  1252. cs.Logger.Info("Received a block part when we're not expecting any",
  1253. "height", height, "round", round, "index", part.Index, "peer", peerID)
  1254. return false, nil
  1255. }
  1256. added, err = cs.ProposalBlockParts.AddPart(part)
  1257. if err != nil {
  1258. return added, err
  1259. }
  1260. if added && cs.ProposalBlockParts.IsComplete() {
  1261. // Added and completed!
  1262. _, err = cdc.UnmarshalBinaryReader(
  1263. cs.ProposalBlockParts.GetReader(),
  1264. &cs.ProposalBlock,
  1265. int64(cs.state.ConsensusParams.BlockSize.MaxBytes),
  1266. )
  1267. if err != nil {
  1268. return added, err
  1269. }
  1270. // NOTE: it's possible to receive complete proposal blocks for future rounds without having the proposal
  1271. cs.Logger.Info("Received complete proposal block", "height", cs.ProposalBlock.Height, "hash", cs.ProposalBlock.Hash())
  1272. // Update Valid* if we can.
  1273. prevotes := cs.Votes.Prevotes(cs.Round)
  1274. blockID, hasTwoThirds := prevotes.TwoThirdsMajority()
  1275. if hasTwoThirds && !blockID.IsZero() && (cs.ValidRound < cs.Round) {
  1276. if cs.ProposalBlock.HashesTo(blockID.Hash) {
  1277. cs.Logger.Info("Updating valid block to new proposal block",
  1278. "valid-round", cs.Round, "valid-block-hash", cs.ProposalBlock.Hash())
  1279. cs.ValidRound = cs.Round
  1280. cs.ValidBlock = cs.ProposalBlock
  1281. cs.ValidBlockParts = cs.ProposalBlockParts
  1282. }
  1283. // TODO: In case there is +2/3 majority in Prevotes set for some
  1284. // block and cs.ProposalBlock contains different block, either
  1285. // proposer is faulty or voting power of faulty processes is more
  1286. // than 1/3. We should trigger in the future accountability
  1287. // procedure at this point.
  1288. }
  1289. if cs.Step <= cstypes.RoundStepPropose && cs.isProposalComplete() {
  1290. // Move onto the next step
  1291. cs.enterPrevote(height, cs.Round)
  1292. if hasTwoThirds { // this is optimisation as this will be triggered when prevote is added
  1293. cs.enterPrecommit(height, cs.Round)
  1294. }
  1295. } else if cs.Step == cstypes.RoundStepCommit {
  1296. // If we're waiting on the proposal block...
  1297. cs.tryFinalizeCommit(height)
  1298. }
  1299. return added, nil
  1300. }
  1301. return added, nil
  1302. }
  1303. // Attempt to add the vote. if its a duplicate signature, dupeout the validator
  1304. func (cs *ConsensusState) tryAddVote(vote *types.Vote, peerID p2p.ID) (bool, error) {
  1305. added, err := cs.addVote(vote, peerID)
  1306. if err != nil {
  1307. // If the vote height is off, we'll just ignore it,
  1308. // But if it's a conflicting sig, add it to the cs.evpool.
  1309. // If it's otherwise invalid, punish peer.
  1310. if err == ErrVoteHeightMismatch {
  1311. return added, err
  1312. } else if voteErr, ok := err.(*types.ErrVoteConflictingVotes); ok {
  1313. if bytes.Equal(vote.ValidatorAddress, cs.privValidator.GetAddress()) {
  1314. cs.Logger.Error("Found conflicting vote from ourselves. Did you unsafe_reset a validator?", "height", vote.Height, "round", vote.Round, "type", vote.Type)
  1315. return added, err
  1316. }
  1317. cs.evpool.AddEvidence(voteErr.DuplicateVoteEvidence)
  1318. return added, err
  1319. } else {
  1320. // Probably an invalid signature / Bad peer.
  1321. // Seems this can also err sometimes with "Unexpected step" - perhaps not from a bad peer ?
  1322. cs.Logger.Error("Error attempting to add vote", "err", err)
  1323. return added, ErrAddingVote
  1324. }
  1325. }
  1326. return added, nil
  1327. }
  1328. //-----------------------------------------------------------------------------
  1329. func (cs *ConsensusState) addVote(vote *types.Vote, peerID p2p.ID) (added bool, err error) {
  1330. cs.Logger.Debug("addVote", "voteHeight", vote.Height, "voteType", vote.Type, "valIndex", vote.ValidatorIndex, "csHeight", cs.Height)
  1331. // A precommit for the previous height?
  1332. // These come in while we wait timeoutCommit
  1333. if vote.Height+1 == cs.Height {
  1334. if !(cs.Step == cstypes.RoundStepNewHeight && vote.Type == types.PrecommitType) {
  1335. // TODO: give the reason ..
  1336. // fmt.Errorf("tryAddVote: Wrong height, not a LastCommit straggler commit.")
  1337. return added, ErrVoteHeightMismatch
  1338. }
  1339. added, err = cs.LastCommit.AddVote(vote)
  1340. if !added {
  1341. return added, err
  1342. }
  1343. cs.Logger.Info(fmt.Sprintf("Added to lastPrecommits: %v", cs.LastCommit.StringShort()))
  1344. cs.eventBus.PublishEventVote(types.EventDataVote{vote})
  1345. cs.evsw.FireEvent(types.EventVote, vote)
  1346. // if we can skip timeoutCommit and have all the votes now,
  1347. if cs.config.SkipTimeoutCommit && cs.LastCommit.HasAll() {
  1348. // go straight to new round (skip timeout commit)
  1349. // cs.scheduleTimeout(time.Duration(0), cs.Height, 0, cstypes.RoundStepNewHeight)
  1350. cs.enterNewRound(cs.Height, 0)
  1351. }
  1352. return
  1353. }
  1354. // Height mismatch is ignored.
  1355. // Not necessarily a bad peer, but not favourable behaviour.
  1356. if vote.Height != cs.Height {
  1357. err = ErrVoteHeightMismatch
  1358. cs.Logger.Info("Vote ignored and not added", "voteHeight", vote.Height, "csHeight", cs.Height, "err", err)
  1359. return
  1360. }
  1361. height := cs.Height
  1362. added, err = cs.Votes.AddVote(vote, peerID)
  1363. if !added {
  1364. // Either duplicate, or error upon cs.Votes.AddByIndex()
  1365. return
  1366. }
  1367. cs.eventBus.PublishEventVote(types.EventDataVote{vote})
  1368. cs.evsw.FireEvent(types.EventVote, vote)
  1369. switch vote.Type {
  1370. case types.PrevoteType:
  1371. prevotes := cs.Votes.Prevotes(vote.Round)
  1372. cs.Logger.Info("Added to prevote", "vote", vote, "prevotes", prevotes.StringShort())
  1373. // If +2/3 prevotes for a block or nil for *any* round:
  1374. if blockID, ok := prevotes.TwoThirdsMajority(); ok {
  1375. // There was a polka!
  1376. // If we're locked but this is a recent polka, unlock.
  1377. // If it matches our ProposalBlock, update the ValidBlock
  1378. // Unlock if `cs.LockedRound < vote.Round <= cs.Round`
  1379. // NOTE: If vote.Round > cs.Round, we'll deal with it when we get to vote.Round
  1380. if (cs.LockedBlock != nil) &&
  1381. (cs.LockedRound < vote.Round) &&
  1382. (vote.Round <= cs.Round) &&
  1383. !cs.LockedBlock.HashesTo(blockID.Hash) {
  1384. cs.Logger.Info("Unlocking because of POL.", "lockedRound", cs.LockedRound, "POLRound", vote.Round)
  1385. cs.LockedRound = -1
  1386. cs.LockedBlock = nil
  1387. cs.LockedBlockParts = nil
  1388. cs.eventBus.PublishEventUnlock(cs.RoundStateEvent())
  1389. }
  1390. // Update Valid* if we can.
  1391. // NOTE: our proposal block may be nil or not what received a polka..
  1392. // TODO: we may want to still update the ValidBlock and obtain it via gossipping
  1393. if len(blockID.Hash) != 0 &&
  1394. (cs.ValidRound < vote.Round) &&
  1395. (vote.Round <= cs.Round) &&
  1396. cs.ProposalBlock.HashesTo(blockID.Hash) {
  1397. cs.Logger.Info("Updating ValidBlock because of POL.", "validRound", cs.ValidRound, "POLRound", vote.Round)
  1398. cs.ValidRound = vote.Round
  1399. cs.ValidBlock = cs.ProposalBlock
  1400. cs.ValidBlockParts = cs.ProposalBlockParts
  1401. }
  1402. }
  1403. // If +2/3 prevotes for *anything* for future round:
  1404. if cs.Round < vote.Round && prevotes.HasTwoThirdsAny() {
  1405. // Round-skip if there is any 2/3+ of votes ahead of us
  1406. cs.enterNewRound(height, vote.Round)
  1407. } else if cs.Round == vote.Round && cstypes.RoundStepPrevote <= cs.Step { // current round
  1408. if prevotes.HasTwoThirdsMajority() {
  1409. cs.enterPrecommit(height, vote.Round)
  1410. } else if prevotes.HasTwoThirdsAny() {
  1411. cs.enterPrevoteWait(height, vote.Round)
  1412. }
  1413. } else if cs.Proposal != nil && 0 <= cs.Proposal.POLRound && cs.Proposal.POLRound == vote.Round {
  1414. // If the proposal is now complete, enter prevote of cs.Round.
  1415. if cs.isProposalComplete() {
  1416. cs.enterPrevote(height, cs.Round)
  1417. }
  1418. }
  1419. case types.PrecommitType:
  1420. precommits := cs.Votes.Precommits(vote.Round)
  1421. cs.Logger.Info("Added to precommit", "vote", vote, "precommits", precommits.StringShort())
  1422. blockID, ok := precommits.TwoThirdsMajority()
  1423. if ok {
  1424. // Executed as TwoThirdsMajority could be from a higher round
  1425. cs.enterNewRound(height, vote.Round)
  1426. cs.enterPrecommit(height, vote.Round)
  1427. if len(blockID.Hash) != 0 {
  1428. cs.enterCommit(height, vote.Round)
  1429. if cs.config.SkipTimeoutCommit && precommits.HasAll() {
  1430. cs.enterNewRound(cs.Height, 0)
  1431. }
  1432. } else {
  1433. cs.enterPrecommitWait(height, vote.Round)
  1434. }
  1435. } else if cs.Round <= vote.Round && precommits.HasTwoThirdsAny() {
  1436. cs.enterNewRound(height, vote.Round)
  1437. cs.enterPrecommitWait(height, vote.Round)
  1438. }
  1439. default:
  1440. panic(fmt.Sprintf("Unexpected vote type %X", vote.Type)) // go-wire should prevent this.
  1441. }
  1442. return
  1443. }
  1444. func (cs *ConsensusState) signVote(type_ types.SignedMsgType, hash []byte, header types.PartSetHeader) (*types.Vote, error) {
  1445. addr := cs.privValidator.GetAddress()
  1446. valIndex, _ := cs.Validators.GetByAddress(addr)
  1447. vote := &types.Vote{
  1448. ValidatorAddress: addr,
  1449. ValidatorIndex: valIndex,
  1450. Height: cs.Height,
  1451. Round: cs.Round,
  1452. Timestamp: cs.voteTime(),
  1453. Type: type_,
  1454. BlockID: types.BlockID{hash, header},
  1455. }
  1456. err := cs.privValidator.SignVote(cs.state.ChainID, vote)
  1457. return vote, err
  1458. }
  1459. func (cs *ConsensusState) voteTime() time.Time {
  1460. now := tmtime.Now()
  1461. minVoteTime := now
  1462. // TODO: We should remove next line in case we don't vote for v in case cs.ProposalBlock == nil,
  1463. // even if cs.LockedBlock != nil. See https://github.com/tendermint/spec.
  1464. if cs.LockedBlock != nil {
  1465. minVoteTime = cs.config.MinValidVoteTime(cs.LockedBlock.Time)
  1466. } else if cs.ProposalBlock != nil {
  1467. minVoteTime = cs.config.MinValidVoteTime(cs.ProposalBlock.Time)
  1468. }
  1469. if now.After(minVoteTime) {
  1470. return now
  1471. }
  1472. return minVoteTime
  1473. }
  1474. // sign the vote and publish on internalMsgQueue
  1475. func (cs *ConsensusState) signAddVote(type_ types.SignedMsgType, hash []byte, header types.PartSetHeader) *types.Vote {
  1476. // if we don't have a key or we're not in the validator set, do nothing
  1477. if cs.privValidator == nil || !cs.Validators.HasAddress(cs.privValidator.GetAddress()) {
  1478. return nil
  1479. }
  1480. vote, err := cs.signVote(type_, hash, header)
  1481. if err == nil {
  1482. cs.sendInternalMessage(msgInfo{&VoteMessage{vote}, ""})
  1483. cs.Logger.Info("Signed and pushed vote", "height", cs.Height, "round", cs.Round, "vote", vote, "err", err)
  1484. return vote
  1485. }
  1486. //if !cs.replayMode {
  1487. cs.Logger.Error("Error signing vote", "height", cs.Height, "round", cs.Round, "vote", vote, "err", err)
  1488. //}
  1489. return nil
  1490. }
  1491. //---------------------------------------------------------
  1492. func CompareHRS(h1 int64, r1 int, s1 cstypes.RoundStepType, h2 int64, r2 int, s2 cstypes.RoundStepType) int {
  1493. if h1 < h2 {
  1494. return -1
  1495. } else if h1 > h2 {
  1496. return 1
  1497. }
  1498. if r1 < r2 {
  1499. return -1
  1500. } else if r1 > r2 {
  1501. return 1
  1502. }
  1503. if s1 < s2 {
  1504. return -1
  1505. } else if s1 > s2 {
  1506. return 1
  1507. }
  1508. return 0
  1509. }