You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

1516 lines
52 KiB

10 years ago
8 years ago
8 years ago
8 years ago
10 years ago
10 years ago
9 years ago
9 years ago
9 years ago
9 years ago
9 years ago
10 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
8 years ago
8 years ago
10 years ago
10 years ago
8 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
10 years ago
9 years ago
10 years ago
8 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
10 years ago
10 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
10 years ago
8 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
9 years ago
9 years ago
8 years ago
9 years ago
8 years ago
8 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
8 years ago
8 years ago
8 years ago
8 years ago
8 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
8 years ago
10 years ago
8 years ago
10 years ago
8 years ago
10 years ago
10 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
8 years ago
8 years ago
8 years ago
8 years ago
8 years ago
8 years ago
8 years ago
8 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
8 years ago
8 years ago
10 years ago
8 years ago
8 years ago
10 years ago
8 years ago
8 years ago
10 years ago
10 years ago
10 years ago
8 years ago
10 years ago
10 years ago
10 years ago
8 years ago
10 years ago
8 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
8 years ago
10 years ago
10 years ago
10 years ago
10 years ago
10 years ago
10 years ago
10 years ago
8 years ago
8 years ago
8 years ago
8 years ago
10 years ago
8 years ago
10 years ago
10 years ago
8 years ago
8 years ago
10 years ago
10 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
8 years ago
8 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
10 years ago
10 years ago
8 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
10 years ago
10 years ago
8 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
10 years ago
10 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
10 years ago
8 years ago
8 years ago
9 years ago
8 years ago
10 years ago
8 years ago
9 years ago
9 years ago
10 years ago
10 years ago
10 years ago
10 years ago
10 years ago
9 years ago
10 years ago
10 years ago
8 years ago
10 years ago
10 years ago
8 years ago
10 years ago
10 years ago
8 years ago
8 years ago
8 years ago
8 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
10 years ago
10 years ago
10 years ago
10 years ago
10 years ago
10 years ago
9 years ago
8 years ago
9 years ago
8 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
10 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
8 years ago
10 years ago
10 years ago
8 years ago
new pubsub package comment out failing consensus tests for now rewrite rpc httpclient to use new pubsub package import pubsub as tmpubsub, query as tmquery make event IDs constants EventKey -> EventTypeKey rename EventsPubsub to PubSub mempool does not use pubsub rename eventsSub to pubsub new subscribe API fix channel size issues and consensus tests bugs refactor rpc client add missing discardFromChan method add mutex rename pubsub to eventBus remove IsRunning from WSRPCConnection interface (not needed) add a comment in broadcastNewRoundStepsAndVotes rename registerEventCallbacks to broadcastNewRoundStepsAndVotes See https://dave.cheney.net/2014/03/19/channel-axioms stop eventBuses after reactor tests remove unnecessary Unsubscribe return subscribe helper function move discardFromChan to where it is used subscribe now returns an err this gives us ability to refuse to subscribe if pubsub is at its max capacity. use context for control overflow cache queries handle err when subscribing in replay_test rename testClientID to testSubscriber extract var set channel buffer capacity to 1 in replay_file fix byzantine_test unsubscribe from single event, not all events refactor httpclient to return events to appropriate channels return failing testReplayCrashBeforeWriteVote test fix TestValidatorSetChanges refactor code a bit fix testReplayCrashBeforeWriteVote add comment fix TestValidatorSetChanges fixes from Bucky's review update comment [ci skip] test TxEventBuffer update changelog fix TestValidatorSetChanges (2nd attempt) only do wg.Done when no errors benchmark event bus create pubsub server inside NewEventBus only expose config params (later if needed) set buffer capacity to 0 so we are not testing cache new tx event format: key = "Tx" plus a tag {"tx.hash": XYZ} This should allow to subscribe to all transactions! or a specific one using a query: "tm.events.type = Tx and tx.hash = '013ABF99434...'" use TimeoutCommit instead of afterPublishEventNewBlockTimeout TimeoutCommit is the time a node waits after committing a block, before it goes into the next height. So it will finish everything from the last block, but then wait a bit. The idea is this gives it time to hear more votes from other validators, to strengthen the commit it includes in the next block. But it also gives it time to hear about new transactions. waitForBlockWithUpdatedVals rewrite WAL crash tests Task: test that we can recover from any WAL crash. Solution: the old tests were relying on event hub being run in the same thread (we were injecting the private validator's last signature). when considering a rewrite, we considered two possible solutions: write a "fuzzy" testing system where WAL is crashing upon receiving a new message, or inject failures and trigger them in tests using something like https://github.com/coreos/gofail. remove sleep no cs.Lock around wal.Save test different cases (empty block, non-empty block, ...) comments add comments test 4 cases: empty block, non-empty block, non-empty block with smaller part size, many blocks fixes as per Bucky's last review reset subscriptions on UnsubscribeAll use a simple counter to track message for which we panicked also, set a smaller part size for all test cases
8 years ago
8 years ago
10 years ago
8 years ago
10 years ago
  1. package consensus
  2. import (
  3. "bytes"
  4. "errors"
  5. "fmt"
  6. "reflect"
  7. "runtime/debug"
  8. "sync"
  9. "time"
  10. fail "github.com/ebuchman/fail-test"
  11. wire "github.com/tendermint/go-wire"
  12. cmn "github.com/tendermint/tmlibs/common"
  13. "github.com/tendermint/tmlibs/log"
  14. cfg "github.com/tendermint/tendermint/config"
  15. cstypes "github.com/tendermint/tendermint/consensus/types"
  16. "github.com/tendermint/tendermint/proxy"
  17. sm "github.com/tendermint/tendermint/state"
  18. "github.com/tendermint/tendermint/types"
  19. )
  20. //-----------------------------------------------------------------------------
  21. // Config
  22. const (
  23. proposalHeartbeatIntervalSeconds = 2
  24. )
  25. //-----------------------------------------------------------------------------
  26. // Errors
  27. var (
  28. ErrInvalidProposalSignature = errors.New("Error invalid proposal signature")
  29. ErrInvalidProposalPOLRound = errors.New("Error invalid proposal POL round")
  30. ErrAddingVote = errors.New("Error adding vote")
  31. ErrVoteHeightMismatch = errors.New("Error vote height mismatch")
  32. )
  33. //-----------------------------------------------------------------------------
  34. var (
  35. msgQueueSize = 1000
  36. )
  37. // msgs from the reactor which may update the state
  38. type msgInfo struct {
  39. Msg ConsensusMessage `json:"msg"`
  40. PeerKey string `json:"peer_key"`
  41. }
  42. // internally generated messages which may update the state
  43. type timeoutInfo struct {
  44. Duration time.Duration `json:"duration"`
  45. Height int `json:"height"`
  46. Round int `json:"round"`
  47. Step cstypes.RoundStepType `json:"step"`
  48. }
  49. func (ti *timeoutInfo) String() string {
  50. return fmt.Sprintf("%v ; %d/%d %v", ti.Duration, ti.Height, ti.Round, ti.Step)
  51. }
  52. // ConsensusState handles execution of the consensus algorithm.
  53. // It processes votes and proposals, and upon reaching agreement,
  54. // commits blocks to the chain and executes them against the application.
  55. // The internal state machine receives input from peers, the internal validator, and from a timer.
  56. type ConsensusState struct {
  57. cmn.BaseService
  58. // config details
  59. config *cfg.ConsensusConfig
  60. privValidator types.PrivValidator // for signing votes
  61. // services for creating and executing blocks
  62. proxyAppConn proxy.AppConnConsensus
  63. blockStore types.BlockStore
  64. mempool types.Mempool
  65. // internal state
  66. mtx sync.Mutex
  67. cstypes.RoundState
  68. state *sm.State // State until height-1.
  69. // state changes may be triggered by msgs from peers,
  70. // msgs from ourself, or by timeouts
  71. peerMsgQueue chan msgInfo
  72. internalMsgQueue chan msgInfo
  73. timeoutTicker TimeoutTicker
  74. // we use eventBus to trigger msg broadcasts in the reactor,
  75. // and to notify external subscribers, eg. through a websocket
  76. eventBus *types.EventBus
  77. // a Write-Ahead Log ensures we can recover from any kind of crash
  78. // and helps us avoid signing conflicting votes
  79. wal WAL
  80. replayMode bool // so we don't log signing errors during replay
  81. doWALCatchup bool // determines if we even try to do the catchup
  82. // for tests where we want to limit the number of transitions the state makes
  83. nSteps int
  84. // some functions can be overwritten for testing
  85. decideProposal func(height, round int)
  86. doPrevote func(height, round int)
  87. setProposal func(proposal *types.Proposal) error
  88. // closed when we finish shutting down
  89. done chan struct{}
  90. }
  91. // NewConsensusState returns a new ConsensusState.
  92. func NewConsensusState(config *cfg.ConsensusConfig, state *sm.State, proxyAppConn proxy.AppConnConsensus, blockStore types.BlockStore, mempool types.Mempool) *ConsensusState {
  93. cs := &ConsensusState{
  94. config: config,
  95. proxyAppConn: proxyAppConn,
  96. blockStore: blockStore,
  97. mempool: mempool,
  98. peerMsgQueue: make(chan msgInfo, msgQueueSize),
  99. internalMsgQueue: make(chan msgInfo, msgQueueSize),
  100. timeoutTicker: NewTimeoutTicker(),
  101. done: make(chan struct{}),
  102. doWALCatchup: true,
  103. wal: nilWAL{},
  104. }
  105. // set function defaults (may be overwritten before calling Start)
  106. cs.decideProposal = cs.defaultDecideProposal
  107. cs.doPrevote = cs.defaultDoPrevote
  108. cs.setProposal = cs.defaultSetProposal
  109. cs.updateToState(state)
  110. // Don't call scheduleRound0 yet.
  111. // We do that upon Start().
  112. cs.reconstructLastCommit(state)
  113. cs.BaseService = *cmn.NewBaseService(nil, "ConsensusState", cs)
  114. return cs
  115. }
  116. //----------------------------------------
  117. // Public interface
  118. // SetLogger implements Service.
  119. func (cs *ConsensusState) SetLogger(l log.Logger) {
  120. cs.BaseService.Logger = l
  121. cs.timeoutTicker.SetLogger(l)
  122. }
  123. // SetEventBus sets event bus.
  124. func (cs *ConsensusState) SetEventBus(b *types.EventBus) {
  125. cs.eventBus = b
  126. }
  127. // String returns a string.
  128. func (cs *ConsensusState) String() string {
  129. // better not to access shared variables
  130. return cmn.Fmt("ConsensusState") //(H:%v R:%v S:%v", cs.Height, cs.Round, cs.Step)
  131. }
  132. // GetState returns a copy of the chain state.
  133. func (cs *ConsensusState) GetState() *sm.State {
  134. cs.mtx.Lock()
  135. defer cs.mtx.Unlock()
  136. return cs.state.Copy()
  137. }
  138. // GetRoundState returns a copy of the internal consensus state.
  139. func (cs *ConsensusState) GetRoundState() *cstypes.RoundState {
  140. cs.mtx.Lock()
  141. defer cs.mtx.Unlock()
  142. return cs.getRoundState()
  143. }
  144. func (cs *ConsensusState) getRoundState() *cstypes.RoundState {
  145. rs := cs.RoundState // copy
  146. return &rs
  147. }
  148. // GetValidators returns a copy of the current validators.
  149. func (cs *ConsensusState) GetValidators() (int, []*types.Validator) {
  150. cs.mtx.Lock()
  151. defer cs.mtx.Unlock()
  152. return cs.state.LastBlockHeight, cs.state.Validators.Copy().Validators
  153. }
  154. // SetPrivValidator sets the private validator account for signing votes.
  155. func (cs *ConsensusState) SetPrivValidator(priv types.PrivValidator) {
  156. cs.mtx.Lock()
  157. defer cs.mtx.Unlock()
  158. cs.privValidator = priv
  159. }
  160. // SetTimeoutTicker sets the local timer. It may be useful to overwrite for testing.
  161. func (cs *ConsensusState) SetTimeoutTicker(timeoutTicker TimeoutTicker) {
  162. cs.mtx.Lock()
  163. defer cs.mtx.Unlock()
  164. cs.timeoutTicker = timeoutTicker
  165. }
  166. // LoadCommit loads the commit for a given height.
  167. func (cs *ConsensusState) LoadCommit(height int) *types.Commit {
  168. cs.mtx.Lock()
  169. defer cs.mtx.Unlock()
  170. if height == cs.blockStore.Height() {
  171. return cs.blockStore.LoadSeenCommit(height)
  172. }
  173. return cs.blockStore.LoadBlockCommit(height)
  174. }
  175. // OnStart implements cmn.Service.
  176. // It loads the latest state via the WAL, and starts the timeout and receive routines.
  177. func (cs *ConsensusState) OnStart() error {
  178. // we may set the WAL in testing before calling Start,
  179. // so only OpenWAL if its still the nilWAL
  180. if _, ok := cs.wal.(nilWAL); ok {
  181. walFile := cs.config.WalFile()
  182. wal, err := cs.OpenWAL(walFile)
  183. if err != nil {
  184. cs.Logger.Error("Error loading ConsensusState wal", "err", err.Error())
  185. return err
  186. }
  187. cs.wal = wal
  188. }
  189. // we need the timeoutRoutine for replay so
  190. // we don't block on the tick chan.
  191. // NOTE: we will get a build up of garbage go routines
  192. // firing on the tockChan until the receiveRoutine is started
  193. // to deal with them (by that point, at most one will be valid)
  194. _, err := cs.timeoutTicker.Start()
  195. if err != nil {
  196. return err
  197. }
  198. // we may have lost some votes if the process crashed
  199. // reload from consensus log to catchup
  200. if cs.doWALCatchup {
  201. if err := cs.catchupReplay(cs.Height); err != nil {
  202. cs.Logger.Error("Error on catchup replay. Proceeding to start ConsensusState anyway", "err", err.Error())
  203. // NOTE: if we ever do return an error here,
  204. // make sure to stop the timeoutTicker
  205. }
  206. }
  207. // now start the receiveRoutine
  208. go cs.receiveRoutine(0)
  209. // schedule the first round!
  210. // use GetRoundState so we don't race the receiveRoutine for access
  211. cs.scheduleRound0(cs.GetRoundState())
  212. return nil
  213. }
  214. // timeoutRoutine: receive requests for timeouts on tickChan and fire timeouts on tockChan
  215. // receiveRoutine: serializes processing of proposoals, block parts, votes; coordinates state transitions
  216. func (cs *ConsensusState) startRoutines(maxSteps int) {
  217. _, err := cs.timeoutTicker.Start()
  218. if err != nil {
  219. panic(err)
  220. }
  221. go cs.receiveRoutine(maxSteps)
  222. }
  223. // OnStop implements cmn.Service. It stops all routines and waits for the WAL to finish.
  224. func (cs *ConsensusState) OnStop() {
  225. cs.BaseService.OnStop()
  226. cs.timeoutTicker.Stop()
  227. // Make BaseService.Wait() wait until cs.wal.Wait()
  228. if cs.IsRunning() {
  229. cs.wal.Wait()
  230. }
  231. }
  232. // Wait waits for the the main routine to return.
  233. // NOTE: be sure to Stop() the event switch and drain
  234. // any event channels or this may deadlock
  235. func (cs *ConsensusState) Wait() {
  236. <-cs.done
  237. }
  238. // OpenWAL opens a file to log all consensus messages and timeouts for deterministic accountability
  239. func (cs *ConsensusState) OpenWAL(walFile string) (WAL, error) {
  240. wal, err := NewWAL(walFile, cs.config.WalLight)
  241. if err != nil {
  242. cs.Logger.Error("Failed to open WAL for consensus state", "wal", walFile, "err", err)
  243. return nil, err
  244. }
  245. wal.SetLogger(cs.Logger.With("wal", walFile))
  246. if _, err := wal.Start(); err != nil {
  247. return nil, err
  248. }
  249. return wal, nil
  250. }
  251. //------------------------------------------------------------
  252. // Public interface for passing messages into the consensus state, possibly causing a state transition.
  253. // If peerKey == "", the msg is considered internal.
  254. // Messages are added to the appropriate queue (peer or internal).
  255. // If the queue is full, the function may block.
  256. // TODO: should these return anything or let callers just use events?
  257. // AddVote inputs a vote.
  258. func (cs *ConsensusState) AddVote(vote *types.Vote, peerKey string) (added bool, err error) {
  259. if peerKey == "" {
  260. cs.internalMsgQueue <- msgInfo{&VoteMessage{vote}, ""}
  261. } else {
  262. cs.peerMsgQueue <- msgInfo{&VoteMessage{vote}, peerKey}
  263. }
  264. // TODO: wait for event?!
  265. return false, nil
  266. }
  267. // SetProposal inputs a proposal.
  268. func (cs *ConsensusState) SetProposal(proposal *types.Proposal, peerKey string) error {
  269. if peerKey == "" {
  270. cs.internalMsgQueue <- msgInfo{&ProposalMessage{proposal}, ""}
  271. } else {
  272. cs.peerMsgQueue <- msgInfo{&ProposalMessage{proposal}, peerKey}
  273. }
  274. // TODO: wait for event?!
  275. return nil
  276. }
  277. // AddProposalBlockPart inputs a part of the proposal block.
  278. func (cs *ConsensusState) AddProposalBlockPart(height, round int, part *types.Part, peerKey string) error {
  279. if peerKey == "" {
  280. cs.internalMsgQueue <- msgInfo{&BlockPartMessage{height, round, part}, ""}
  281. } else {
  282. cs.peerMsgQueue <- msgInfo{&BlockPartMessage{height, round, part}, peerKey}
  283. }
  284. // TODO: wait for event?!
  285. return nil
  286. }
  287. // SetProposalAndBlock inputs the proposal and all block parts.
  288. func (cs *ConsensusState) SetProposalAndBlock(proposal *types.Proposal, block *types.Block, parts *types.PartSet, peerKey string) error {
  289. if err := cs.SetProposal(proposal, peerKey); err != nil {
  290. return err
  291. }
  292. for i := 0; i < parts.Total(); i++ {
  293. part := parts.GetPart(i)
  294. if err := cs.AddProposalBlockPart(proposal.Height, proposal.Round, part, peerKey); err != nil {
  295. return err
  296. }
  297. }
  298. return nil
  299. }
  300. //------------------------------------------------------------
  301. // internal functions for managing the state
  302. func (cs *ConsensusState) updateHeight(height int) {
  303. cs.Height = height
  304. }
  305. func (cs *ConsensusState) updateRoundStep(round int, step cstypes.RoundStepType) {
  306. cs.Round = round
  307. cs.Step = step
  308. }
  309. // enterNewRound(height, 0) at cs.StartTime.
  310. func (cs *ConsensusState) scheduleRound0(rs *cstypes.RoundState) {
  311. //cs.Logger.Info("scheduleRound0", "now", time.Now(), "startTime", cs.StartTime)
  312. sleepDuration := rs.StartTime.Sub(time.Now())
  313. cs.scheduleTimeout(sleepDuration, rs.Height, 0, cstypes.RoundStepNewHeight)
  314. }
  315. // Attempt to schedule a timeout (by sending timeoutInfo on the tickChan)
  316. func (cs *ConsensusState) scheduleTimeout(duration time.Duration, height, round int, step cstypes.RoundStepType) {
  317. cs.timeoutTicker.ScheduleTimeout(timeoutInfo{duration, height, round, step})
  318. }
  319. // send a msg into the receiveRoutine regarding our own proposal, block part, or vote
  320. func (cs *ConsensusState) sendInternalMessage(mi msgInfo) {
  321. select {
  322. case cs.internalMsgQueue <- mi:
  323. default:
  324. // NOTE: using the go-routine means our votes can
  325. // be processed out of order.
  326. // TODO: use CList here for strict determinism and
  327. // attempt push to internalMsgQueue in receiveRoutine
  328. cs.Logger.Info("Internal msg queue is full. Using a go-routine")
  329. go func() { cs.internalMsgQueue <- mi }()
  330. }
  331. }
  332. // Reconstruct LastCommit from SeenCommit, which we saved along with the block,
  333. // (which happens even before saving the state)
  334. func (cs *ConsensusState) reconstructLastCommit(state *sm.State) {
  335. if state.LastBlockHeight == 0 {
  336. return
  337. }
  338. seenCommit := cs.blockStore.LoadSeenCommit(state.LastBlockHeight)
  339. lastPrecommits := types.NewVoteSet(state.ChainID, state.LastBlockHeight, seenCommit.Round(), types.VoteTypePrecommit, state.LastValidators)
  340. for _, precommit := range seenCommit.Precommits {
  341. if precommit == nil {
  342. continue
  343. }
  344. added, err := lastPrecommits.AddVote(precommit)
  345. if !added || err != nil {
  346. cmn.PanicCrisis(cmn.Fmt("Failed to reconstruct LastCommit: %v", err))
  347. }
  348. }
  349. if !lastPrecommits.HasTwoThirdsMajority() {
  350. cmn.PanicSanity("Failed to reconstruct LastCommit: Does not have +2/3 maj")
  351. }
  352. cs.LastCommit = lastPrecommits
  353. }
  354. // Updates ConsensusState and increments height to match that of state.
  355. // The round becomes 0 and cs.Step becomes cstypes.RoundStepNewHeight.
  356. func (cs *ConsensusState) updateToState(state *sm.State) {
  357. if cs.CommitRound > -1 && 0 < cs.Height && cs.Height != state.LastBlockHeight {
  358. cmn.PanicSanity(cmn.Fmt("updateToState() expected state height of %v but found %v",
  359. cs.Height, state.LastBlockHeight))
  360. }
  361. if cs.state != nil && cs.state.LastBlockHeight+1 != cs.Height {
  362. // This might happen when someone else is mutating cs.state.
  363. // Someone forgot to pass in state.Copy() somewhere?!
  364. cmn.PanicSanity(cmn.Fmt("Inconsistent cs.state.LastBlockHeight+1 %v vs cs.Height %v",
  365. cs.state.LastBlockHeight+1, cs.Height))
  366. }
  367. // If state isn't further out than cs.state, just ignore.
  368. // This happens when SwitchToConsensus() is called in the reactor.
  369. // We don't want to reset e.g. the Votes.
  370. if cs.state != nil && (state.LastBlockHeight <= cs.state.LastBlockHeight) {
  371. cs.Logger.Info("Ignoring updateToState()", "newHeight", state.LastBlockHeight+1, "oldHeight", cs.state.LastBlockHeight+1)
  372. return
  373. }
  374. // Reset fields based on state.
  375. validators := state.Validators
  376. lastPrecommits := (*types.VoteSet)(nil)
  377. if cs.CommitRound > -1 && cs.Votes != nil {
  378. if !cs.Votes.Precommits(cs.CommitRound).HasTwoThirdsMajority() {
  379. cmn.PanicSanity("updateToState(state) called but last Precommit round didn't have +2/3")
  380. }
  381. lastPrecommits = cs.Votes.Precommits(cs.CommitRound)
  382. }
  383. // Next desired block height
  384. height := state.LastBlockHeight + 1
  385. // RoundState fields
  386. cs.updateHeight(height)
  387. cs.updateRoundStep(0, cstypes.RoundStepNewHeight)
  388. if cs.CommitTime.IsZero() {
  389. // "Now" makes it easier to sync up dev nodes.
  390. // We add timeoutCommit to allow transactions
  391. // to be gathered for the first block.
  392. // And alternative solution that relies on clocks:
  393. // cs.StartTime = state.LastBlockTime.Add(timeoutCommit)
  394. cs.StartTime = cs.config.Commit(time.Now())
  395. } else {
  396. cs.StartTime = cs.config.Commit(cs.CommitTime)
  397. }
  398. cs.Validators = validators
  399. cs.Proposal = nil
  400. cs.ProposalBlock = nil
  401. cs.ProposalBlockParts = nil
  402. cs.LockedRound = 0
  403. cs.LockedBlock = nil
  404. cs.LockedBlockParts = nil
  405. cs.Votes = cstypes.NewHeightVoteSet(state.ChainID, height, validators)
  406. cs.CommitRound = -1
  407. cs.LastCommit = lastPrecommits
  408. cs.LastValidators = state.LastValidators
  409. cs.state = state
  410. // Finally, broadcast RoundState
  411. cs.newStep()
  412. }
  413. func (cs *ConsensusState) newStep() {
  414. rs := cs.RoundStateEvent()
  415. cs.wal.Save(rs)
  416. cs.nSteps += 1
  417. // newStep is called by updateToStep in NewConsensusState before the eventBus is set!
  418. if cs.eventBus != nil {
  419. cs.eventBus.PublishEventNewRoundStep(rs)
  420. }
  421. }
  422. //-----------------------------------------
  423. // the main go routines
  424. // receiveRoutine handles messages which may cause state transitions.
  425. // it's argument (n) is the number of messages to process before exiting - use 0 to run forever
  426. // It keeps the RoundState and is the only thing that updates it.
  427. // Updates (state transitions) happen on timeouts, complete proposals, and 2/3 majorities.
  428. // ConsensusState must be locked before any internal state is updated.
  429. func (cs *ConsensusState) receiveRoutine(maxSteps int) {
  430. defer func() {
  431. if r := recover(); r != nil {
  432. cs.Logger.Error("CONSENSUS FAILURE!!!", "err", r, "stack", string(debug.Stack()))
  433. }
  434. }()
  435. for {
  436. if maxSteps > 0 {
  437. if cs.nSteps >= maxSteps {
  438. cs.Logger.Info("reached max steps. exiting receive routine")
  439. cs.nSteps = 0
  440. return
  441. }
  442. }
  443. rs := cs.RoundState
  444. var mi msgInfo
  445. select {
  446. case height := <-cs.mempool.TxsAvailable():
  447. cs.handleTxsAvailable(height)
  448. case mi = <-cs.peerMsgQueue:
  449. cs.wal.Save(mi)
  450. // handles proposals, block parts, votes
  451. // may generate internal events (votes, complete proposals, 2/3 majorities)
  452. cs.handleMsg(mi)
  453. case mi = <-cs.internalMsgQueue:
  454. cs.wal.Save(mi)
  455. // handles proposals, block parts, votes
  456. cs.handleMsg(mi)
  457. case ti := <-cs.timeoutTicker.Chan(): // tockChan:
  458. cs.wal.Save(ti)
  459. // if the timeout is relevant to the rs
  460. // go to the next step
  461. cs.handleTimeout(ti, rs)
  462. case <-cs.Quit:
  463. // NOTE: the internalMsgQueue may have signed messages from our
  464. // priv_val that haven't hit the WAL, but its ok because
  465. // priv_val tracks LastSig
  466. // close wal now that we're done writing to it
  467. cs.wal.Stop()
  468. close(cs.done)
  469. return
  470. }
  471. }
  472. }
  473. // state transitions on complete-proposal, 2/3-any, 2/3-one
  474. func (cs *ConsensusState) handleMsg(mi msgInfo) {
  475. cs.mtx.Lock()
  476. defer cs.mtx.Unlock()
  477. var err error
  478. msg, peerKey := mi.Msg, mi.PeerKey
  479. switch msg := msg.(type) {
  480. case *ProposalMessage:
  481. // will not cause transition.
  482. // once proposal is set, we can receive block parts
  483. err = cs.setProposal(msg.Proposal)
  484. case *BlockPartMessage:
  485. // if the proposal is complete, we'll enterPrevote or tryFinalizeCommit
  486. _, err = cs.addProposalBlockPart(msg.Height, msg.Part, peerKey != "")
  487. if err != nil && msg.Round != cs.Round {
  488. err = nil
  489. }
  490. case *VoteMessage:
  491. // attempt to add the vote and dupeout the validator if its a duplicate signature
  492. // if the vote gives us a 2/3-any or 2/3-one, we transition
  493. err := cs.tryAddVote(msg.Vote, peerKey)
  494. if err == ErrAddingVote {
  495. // TODO: punish peer
  496. }
  497. // NOTE: the vote is broadcast to peers by the reactor listening
  498. // for vote events
  499. // TODO: If rs.Height == vote.Height && rs.Round < vote.Round,
  500. // the peer is sending us CatchupCommit precommits.
  501. // We could make note of this and help filter in broadcastHasVoteMessage().
  502. default:
  503. cs.Logger.Error("Unknown msg type", reflect.TypeOf(msg))
  504. }
  505. if err != nil {
  506. cs.Logger.Error("Error with msg", "type", reflect.TypeOf(msg), "peer", peerKey, "err", err, "msg", msg)
  507. }
  508. }
  509. func (cs *ConsensusState) handleTimeout(ti timeoutInfo, rs cstypes.RoundState) {
  510. cs.Logger.Debug("Received tock", "timeout", ti.Duration, "height", ti.Height, "round", ti.Round, "step", ti.Step)
  511. // timeouts must be for current height, round, step
  512. if ti.Height != rs.Height || ti.Round < rs.Round || (ti.Round == rs.Round && ti.Step < rs.Step) {
  513. cs.Logger.Debug("Ignoring tock because we're ahead", "height", rs.Height, "round", rs.Round, "step", rs.Step)
  514. return
  515. }
  516. // the timeout will now cause a state transition
  517. cs.mtx.Lock()
  518. defer cs.mtx.Unlock()
  519. switch ti.Step {
  520. case cstypes.RoundStepNewHeight:
  521. // NewRound event fired from enterNewRound.
  522. // XXX: should we fire timeout here (for timeout commit)?
  523. cs.enterNewRound(ti.Height, 0)
  524. case cstypes.RoundStepNewRound:
  525. cs.enterPropose(ti.Height, 0)
  526. case cstypes.RoundStepPropose:
  527. cs.eventBus.PublishEventTimeoutPropose(cs.RoundStateEvent())
  528. cs.enterPrevote(ti.Height, ti.Round)
  529. case cstypes.RoundStepPrevoteWait:
  530. cs.eventBus.PublishEventTimeoutWait(cs.RoundStateEvent())
  531. cs.enterPrecommit(ti.Height, ti.Round)
  532. case cstypes.RoundStepPrecommitWait:
  533. cs.eventBus.PublishEventTimeoutWait(cs.RoundStateEvent())
  534. cs.enterNewRound(ti.Height, ti.Round+1)
  535. default:
  536. panic(cmn.Fmt("Invalid timeout step: %v", ti.Step))
  537. }
  538. }
  539. func (cs *ConsensusState) handleTxsAvailable(height int) {
  540. cs.mtx.Lock()
  541. defer cs.mtx.Unlock()
  542. // we only need to do this for round 0
  543. cs.enterPropose(height, 0)
  544. }
  545. //-----------------------------------------------------------------------------
  546. // State functions
  547. // Used internally by handleTimeout and handleMsg to make state transitions
  548. // Enter: `timeoutNewHeight` by startTime (commitTime+timeoutCommit),
  549. // or, if SkipTimeout==true, after receiving all precommits from (height,round-1)
  550. // Enter: `timeoutPrecommits` after any +2/3 precommits from (height,round-1)
  551. // Enter: +2/3 precommits for nil at (height,round-1)
  552. // Enter: +2/3 prevotes any or +2/3 precommits for block or any from (height, round)
  553. // NOTE: cs.StartTime was already set for height.
  554. func (cs *ConsensusState) enterNewRound(height int, round int) {
  555. if cs.Height != height || round < cs.Round || (cs.Round == round && cs.Step != cstypes.RoundStepNewHeight) {
  556. cs.Logger.Debug(cmn.Fmt("enterNewRound(%v/%v): Invalid args. Current step: %v/%v/%v", height, round, cs.Height, cs.Round, cs.Step))
  557. return
  558. }
  559. if now := time.Now(); cs.StartTime.After(now) {
  560. cs.Logger.Info("Need to set a buffer and log message here for sanity.", "startTime", cs.StartTime, "now", now)
  561. }
  562. cs.Logger.Info(cmn.Fmt("enterNewRound(%v/%v). Current: %v/%v/%v", height, round, cs.Height, cs.Round, cs.Step))
  563. // Increment validators if necessary
  564. validators := cs.Validators
  565. if cs.Round < round {
  566. validators = validators.Copy()
  567. validators.IncrementAccum(round - cs.Round)
  568. }
  569. // Setup new round
  570. // we don't fire newStep for this step,
  571. // but we fire an event, so update the round step first
  572. cs.updateRoundStep(round, cstypes.RoundStepNewRound)
  573. cs.Validators = validators
  574. if round == 0 {
  575. // We've already reset these upon new height,
  576. // and meanwhile we might have received a proposal
  577. // for round 0.
  578. } else {
  579. cs.Proposal = nil
  580. cs.ProposalBlock = nil
  581. cs.ProposalBlockParts = nil
  582. }
  583. cs.Votes.SetRound(round + 1) // also track next round (round+1) to allow round-skipping
  584. cs.eventBus.PublishEventNewRound(cs.RoundStateEvent())
  585. // Wait for txs to be available in the mempool
  586. // before we enterPropose in round 0. If the last block changed the app hash,
  587. // we may need an empty "proof" block, and enterPropose immediately.
  588. waitForTxs := cs.config.WaitForTxs() && round == 0 && !cs.needProofBlock(height)
  589. if waitForTxs {
  590. if cs.config.CreateEmptyBlocksInterval > 0 {
  591. cs.scheduleTimeout(cs.config.EmptyBlocksInterval(), height, round, cstypes.RoundStepNewRound)
  592. }
  593. go cs.proposalHeartbeat(height, round)
  594. } else {
  595. cs.enterPropose(height, round)
  596. }
  597. }
  598. // needProofBlock returns true on the first height (so the genesis app hash is signed right away)
  599. // and where the last block (height-1) caused the app hash to change
  600. func (cs *ConsensusState) needProofBlock(height int) bool {
  601. if height == 1 {
  602. return true
  603. }
  604. lastBlockMeta := cs.blockStore.LoadBlockMeta(height - 1)
  605. if !bytes.Equal(cs.state.AppHash, lastBlockMeta.Header.AppHash) {
  606. return true
  607. }
  608. return false
  609. }
  610. func (cs *ConsensusState) proposalHeartbeat(height, round int) {
  611. counter := 0
  612. addr := cs.privValidator.GetAddress()
  613. valIndex, v := cs.Validators.GetByAddress(addr)
  614. if v == nil {
  615. // not a validator
  616. valIndex = -1
  617. }
  618. chainID := cs.state.ChainID
  619. for {
  620. rs := cs.GetRoundState()
  621. // if we've already moved on, no need to send more heartbeats
  622. if rs.Step > cstypes.RoundStepNewRound || rs.Round > round || rs.Height > height {
  623. return
  624. }
  625. heartbeat := &types.Heartbeat{
  626. Height: rs.Height,
  627. Round: rs.Round,
  628. Sequence: counter,
  629. ValidatorAddress: addr,
  630. ValidatorIndex: valIndex,
  631. }
  632. cs.privValidator.SignHeartbeat(chainID, heartbeat)
  633. cs.eventBus.PublishEventProposalHeartbeat(types.EventDataProposalHeartbeat{heartbeat})
  634. counter += 1
  635. time.Sleep(proposalHeartbeatIntervalSeconds * time.Second)
  636. }
  637. }
  638. // Enter (CreateEmptyBlocks): from enterNewRound(height,round)
  639. // Enter (CreateEmptyBlocks, CreateEmptyBlocksInterval > 0 ): after enterNewRound(height,round), after timeout of CreateEmptyBlocksInterval
  640. // Enter (!CreateEmptyBlocks) : after enterNewRound(height,round), once txs are in the mempool
  641. func (cs *ConsensusState) enterPropose(height int, round int) {
  642. if cs.Height != height || round < cs.Round || (cs.Round == round && cstypes.RoundStepPropose <= cs.Step) {
  643. cs.Logger.Debug(cmn.Fmt("enterPropose(%v/%v): Invalid args. Current step: %v/%v/%v", height, round, cs.Height, cs.Round, cs.Step))
  644. return
  645. }
  646. cs.Logger.Info(cmn.Fmt("enterPropose(%v/%v). Current: %v/%v/%v", height, round, cs.Height, cs.Round, cs.Step))
  647. defer func() {
  648. // Done enterPropose:
  649. cs.updateRoundStep(round, cstypes.RoundStepPropose)
  650. cs.newStep()
  651. // If we have the whole proposal + POL, then goto Prevote now.
  652. // else, we'll enterPrevote when the rest of the proposal is received (in AddProposalBlockPart),
  653. // or else after timeoutPropose
  654. if cs.isProposalComplete() {
  655. cs.enterPrevote(height, cs.Round)
  656. }
  657. }()
  658. // If we don't get the proposal and all block parts quick enough, enterPrevote
  659. cs.scheduleTimeout(cs.config.Propose(round), height, round, cstypes.RoundStepPropose)
  660. // Nothing more to do if we're not a validator
  661. if cs.privValidator == nil {
  662. cs.Logger.Debug("This node is not a validator")
  663. return
  664. }
  665. if !cs.isProposer() {
  666. cs.Logger.Info("enterPropose: Not our turn to propose", "proposer", cs.Validators.GetProposer().Address, "privValidator", cs.privValidator)
  667. if cs.Validators.HasAddress(cs.privValidator.GetAddress()) {
  668. cs.Logger.Debug("This node is a validator")
  669. } else {
  670. cs.Logger.Debug("This node is not a validator")
  671. }
  672. } else {
  673. cs.Logger.Info("enterPropose: Our turn to propose", "proposer", cs.Validators.GetProposer().Address, "privValidator", cs.privValidator)
  674. cs.Logger.Debug("This node is a validator")
  675. cs.decideProposal(height, round)
  676. }
  677. }
  678. func (cs *ConsensusState) isProposer() bool {
  679. return bytes.Equal(cs.Validators.GetProposer().Address, cs.privValidator.GetAddress())
  680. }
  681. func (cs *ConsensusState) defaultDecideProposal(height, round int) {
  682. var block *types.Block
  683. var blockParts *types.PartSet
  684. // Decide on block
  685. if cs.LockedBlock != nil {
  686. // If we're locked onto a block, just choose that.
  687. block, blockParts = cs.LockedBlock, cs.LockedBlockParts
  688. } else {
  689. // Create a new proposal block from state/txs from the mempool.
  690. block, blockParts = cs.createProposalBlock()
  691. if block == nil { // on error
  692. return
  693. }
  694. }
  695. // Make proposal
  696. polRound, polBlockID := cs.Votes.POLInfo()
  697. proposal := types.NewProposal(height, round, blockParts.Header(), polRound, polBlockID)
  698. if err := cs.privValidator.SignProposal(cs.state.ChainID, proposal); err == nil {
  699. // Set fields
  700. /* fields set by setProposal and addBlockPart
  701. cs.Proposal = proposal
  702. cs.ProposalBlock = block
  703. cs.ProposalBlockParts = blockParts
  704. */
  705. // send proposal and block parts on internal msg queue
  706. cs.sendInternalMessage(msgInfo{&ProposalMessage{proposal}, ""})
  707. for i := 0; i < blockParts.Total(); i++ {
  708. part := blockParts.GetPart(i)
  709. cs.sendInternalMessage(msgInfo{&BlockPartMessage{cs.Height, cs.Round, part}, ""})
  710. }
  711. cs.Logger.Info("Signed proposal", "height", height, "round", round, "proposal", proposal)
  712. cs.Logger.Debug(cmn.Fmt("Signed proposal block: %v", block))
  713. } else {
  714. if !cs.replayMode {
  715. cs.Logger.Error("enterPropose: Error signing proposal", "height", height, "round", round, "err", err)
  716. }
  717. }
  718. }
  719. // Returns true if the proposal block is complete &&
  720. // (if POLRound was proposed, we have +2/3 prevotes from there).
  721. func (cs *ConsensusState) isProposalComplete() bool {
  722. if cs.Proposal == nil || cs.ProposalBlock == nil {
  723. return false
  724. }
  725. // we have the proposal. if there's a POLRound,
  726. // make sure we have the prevotes from it too
  727. if cs.Proposal.POLRound < 0 {
  728. return true
  729. } else {
  730. // if this is false the proposer is lying or we haven't received the POL yet
  731. return cs.Votes.Prevotes(cs.Proposal.POLRound).HasTwoThirdsMajority()
  732. }
  733. }
  734. // Create the next block to propose and return it.
  735. // Returns nil block upon error.
  736. // NOTE: keep it side-effect free for clarity.
  737. func (cs *ConsensusState) createProposalBlock() (block *types.Block, blockParts *types.PartSet) {
  738. var commit *types.Commit
  739. if cs.Height == 1 {
  740. // We're creating a proposal for the first block.
  741. // The commit is empty, but not nil.
  742. commit = &types.Commit{}
  743. } else if cs.LastCommit.HasTwoThirdsMajority() {
  744. // Make the commit from LastCommit
  745. commit = cs.LastCommit.MakeCommit()
  746. } else {
  747. // This shouldn't happen.
  748. cs.Logger.Error("enterPropose: Cannot propose anything: No commit for the previous block.")
  749. return
  750. }
  751. // Mempool validated transactions
  752. txs := cs.mempool.Reap(cs.config.MaxBlockSizeTxs)
  753. return types.MakeBlock(cs.Height, cs.state.ChainID, txs, commit,
  754. cs.state.LastBlockID, cs.state.Validators.Hash(),
  755. cs.state.AppHash, cs.state.Params.BlockPartSizeBytes)
  756. }
  757. // Enter: `timeoutPropose` after entering Propose.
  758. // Enter: proposal block and POL is ready.
  759. // Enter: any +2/3 prevotes for future round.
  760. // Prevote for LockedBlock if we're locked, or ProposalBlock if valid.
  761. // Otherwise vote nil.
  762. func (cs *ConsensusState) enterPrevote(height int, round int) {
  763. if cs.Height != height || round < cs.Round || (cs.Round == round && cstypes.RoundStepPrevote <= cs.Step) {
  764. cs.Logger.Debug(cmn.Fmt("enterPrevote(%v/%v): Invalid args. Current step: %v/%v/%v", height, round, cs.Height, cs.Round, cs.Step))
  765. return
  766. }
  767. defer func() {
  768. // Done enterPrevote:
  769. cs.updateRoundStep(round, cstypes.RoundStepPrevote)
  770. cs.newStep()
  771. }()
  772. // fire event for how we got here
  773. if cs.isProposalComplete() {
  774. cs.eventBus.PublishEventCompleteProposal(cs.RoundStateEvent())
  775. } else {
  776. // we received +2/3 prevotes for a future round
  777. // TODO: catchup event?
  778. }
  779. cs.Logger.Info(cmn.Fmt("enterPrevote(%v/%v). Current: %v/%v/%v", height, round, cs.Height, cs.Round, cs.Step))
  780. // Sign and broadcast vote as necessary
  781. cs.doPrevote(height, round)
  782. // Once `addVote` hits any +2/3 prevotes, we will go to PrevoteWait
  783. // (so we have more time to try and collect +2/3 prevotes for a single block)
  784. }
  785. func (cs *ConsensusState) defaultDoPrevote(height int, round int) {
  786. logger := cs.Logger.With("height", height, "round", round)
  787. // If a block is locked, prevote that.
  788. if cs.LockedBlock != nil {
  789. logger.Info("enterPrevote: Block was locked")
  790. cs.signAddVote(types.VoteTypePrevote, cs.LockedBlock.Hash(), cs.LockedBlockParts.Header())
  791. return
  792. }
  793. // If ProposalBlock is nil, prevote nil.
  794. if cs.ProposalBlock == nil {
  795. logger.Info("enterPrevote: ProposalBlock is nil")
  796. cs.signAddVote(types.VoteTypePrevote, nil, types.PartSetHeader{})
  797. return
  798. }
  799. // Validate proposal block
  800. err := cs.state.ValidateBlock(cs.ProposalBlock)
  801. if err != nil {
  802. // ProposalBlock is invalid, prevote nil.
  803. logger.Error("enterPrevote: ProposalBlock is invalid", "err", err)
  804. cs.signAddVote(types.VoteTypePrevote, nil, types.PartSetHeader{})
  805. return
  806. }
  807. // Prevote cs.ProposalBlock
  808. // NOTE: the proposal signature is validated when it is received,
  809. // and the proposal block parts are validated as they are received (against the merkle hash in the proposal)
  810. logger.Info("enterPrevote: ProposalBlock is valid")
  811. cs.signAddVote(types.VoteTypePrevote, cs.ProposalBlock.Hash(), cs.ProposalBlockParts.Header())
  812. }
  813. // Enter: any +2/3 prevotes at next round.
  814. func (cs *ConsensusState) enterPrevoteWait(height int, round int) {
  815. if cs.Height != height || round < cs.Round || (cs.Round == round && cstypes.RoundStepPrevoteWait <= cs.Step) {
  816. cs.Logger.Debug(cmn.Fmt("enterPrevoteWait(%v/%v): Invalid args. Current step: %v/%v/%v", height, round, cs.Height, cs.Round, cs.Step))
  817. return
  818. }
  819. if !cs.Votes.Prevotes(round).HasTwoThirdsAny() {
  820. cmn.PanicSanity(cmn.Fmt("enterPrevoteWait(%v/%v), but Prevotes does not have any +2/3 votes", height, round))
  821. }
  822. cs.Logger.Info(cmn.Fmt("enterPrevoteWait(%v/%v). Current: %v/%v/%v", height, round, cs.Height, cs.Round, cs.Step))
  823. defer func() {
  824. // Done enterPrevoteWait:
  825. cs.updateRoundStep(round, cstypes.RoundStepPrevoteWait)
  826. cs.newStep()
  827. }()
  828. // Wait for some more prevotes; enterPrecommit
  829. cs.scheduleTimeout(cs.config.Prevote(round), height, round, cstypes.RoundStepPrevoteWait)
  830. }
  831. // Enter: `timeoutPrevote` after any +2/3 prevotes.
  832. // Enter: +2/3 precomits for block or nil.
  833. // Enter: any +2/3 precommits for next round.
  834. // Lock & precommit the ProposalBlock if we have enough prevotes for it (a POL in this round)
  835. // else, unlock an existing lock and precommit nil if +2/3 of prevotes were nil,
  836. // else, precommit nil otherwise.
  837. func (cs *ConsensusState) enterPrecommit(height int, round int) {
  838. if cs.Height != height || round < cs.Round || (cs.Round == round && cstypes.RoundStepPrecommit <= cs.Step) {
  839. cs.Logger.Debug(cmn.Fmt("enterPrecommit(%v/%v): Invalid args. Current step: %v/%v/%v", height, round, cs.Height, cs.Round, cs.Step))
  840. return
  841. }
  842. cs.Logger.Info(cmn.Fmt("enterPrecommit(%v/%v). Current: %v/%v/%v", height, round, cs.Height, cs.Round, cs.Step))
  843. defer func() {
  844. // Done enterPrecommit:
  845. cs.updateRoundStep(round, cstypes.RoundStepPrecommit)
  846. cs.newStep()
  847. }()
  848. blockID, ok := cs.Votes.Prevotes(round).TwoThirdsMajority()
  849. // If we don't have a polka, we must precommit nil
  850. if !ok {
  851. if cs.LockedBlock != nil {
  852. cs.Logger.Info("enterPrecommit: No +2/3 prevotes during enterPrecommit while we're locked. Precommitting nil")
  853. } else {
  854. cs.Logger.Info("enterPrecommit: No +2/3 prevotes during enterPrecommit. Precommitting nil.")
  855. }
  856. cs.signAddVote(types.VoteTypePrecommit, nil, types.PartSetHeader{})
  857. return
  858. }
  859. // At this point +2/3 prevoted for a particular block or nil
  860. cs.eventBus.PublishEventPolka(cs.RoundStateEvent())
  861. // the latest POLRound should be this round
  862. polRound, _ := cs.Votes.POLInfo()
  863. if polRound < round {
  864. cmn.PanicSanity(cmn.Fmt("This POLRound should be %v but got %", round, polRound))
  865. }
  866. // +2/3 prevoted nil. Unlock and precommit nil.
  867. if len(blockID.Hash) == 0 {
  868. if cs.LockedBlock == nil {
  869. cs.Logger.Info("enterPrecommit: +2/3 prevoted for nil.")
  870. } else {
  871. cs.Logger.Info("enterPrecommit: +2/3 prevoted for nil. Unlocking")
  872. cs.LockedRound = 0
  873. cs.LockedBlock = nil
  874. cs.LockedBlockParts = nil
  875. cs.eventBus.PublishEventUnlock(cs.RoundStateEvent())
  876. }
  877. cs.signAddVote(types.VoteTypePrecommit, nil, types.PartSetHeader{})
  878. return
  879. }
  880. // At this point, +2/3 prevoted for a particular block.
  881. // If we're already locked on that block, precommit it, and update the LockedRound
  882. if cs.LockedBlock.HashesTo(blockID.Hash) {
  883. cs.Logger.Info("enterPrecommit: +2/3 prevoted locked block. Relocking")
  884. cs.LockedRound = round
  885. cs.eventBus.PublishEventRelock(cs.RoundStateEvent())
  886. cs.signAddVote(types.VoteTypePrecommit, blockID.Hash, blockID.PartsHeader)
  887. return
  888. }
  889. // If +2/3 prevoted for proposal block, stage and precommit it
  890. if cs.ProposalBlock.HashesTo(blockID.Hash) {
  891. cs.Logger.Info("enterPrecommit: +2/3 prevoted proposal block. Locking", "hash", blockID.Hash)
  892. // Validate the block.
  893. if err := cs.state.ValidateBlock(cs.ProposalBlock); err != nil {
  894. cmn.PanicConsensus(cmn.Fmt("enterPrecommit: +2/3 prevoted for an invalid block: %v", err))
  895. }
  896. cs.LockedRound = round
  897. cs.LockedBlock = cs.ProposalBlock
  898. cs.LockedBlockParts = cs.ProposalBlockParts
  899. cs.eventBus.PublishEventLock(cs.RoundStateEvent())
  900. cs.signAddVote(types.VoteTypePrecommit, blockID.Hash, blockID.PartsHeader)
  901. return
  902. }
  903. // There was a polka in this round for a block we don't have.
  904. // Fetch that block, unlock, and precommit nil.
  905. // The +2/3 prevotes for this round is the POL for our unlock.
  906. // TODO: In the future save the POL prevotes for justification.
  907. cs.LockedRound = 0
  908. cs.LockedBlock = nil
  909. cs.LockedBlockParts = nil
  910. if !cs.ProposalBlockParts.HasHeader(blockID.PartsHeader) {
  911. cs.ProposalBlock = nil
  912. cs.ProposalBlockParts = types.NewPartSetFromHeader(blockID.PartsHeader)
  913. }
  914. cs.eventBus.PublishEventUnlock(cs.RoundStateEvent())
  915. cs.signAddVote(types.VoteTypePrecommit, nil, types.PartSetHeader{})
  916. }
  917. // Enter: any +2/3 precommits for next round.
  918. func (cs *ConsensusState) enterPrecommitWait(height int, round int) {
  919. if cs.Height != height || round < cs.Round || (cs.Round == round && cstypes.RoundStepPrecommitWait <= cs.Step) {
  920. cs.Logger.Debug(cmn.Fmt("enterPrecommitWait(%v/%v): Invalid args. Current step: %v/%v/%v", height, round, cs.Height, cs.Round, cs.Step))
  921. return
  922. }
  923. if !cs.Votes.Precommits(round).HasTwoThirdsAny() {
  924. cmn.PanicSanity(cmn.Fmt("enterPrecommitWait(%v/%v), but Precommits does not have any +2/3 votes", height, round))
  925. }
  926. cs.Logger.Info(cmn.Fmt("enterPrecommitWait(%v/%v). Current: %v/%v/%v", height, round, cs.Height, cs.Round, cs.Step))
  927. defer func() {
  928. // Done enterPrecommitWait:
  929. cs.updateRoundStep(round, cstypes.RoundStepPrecommitWait)
  930. cs.newStep()
  931. }()
  932. // Wait for some more precommits; enterNewRound
  933. cs.scheduleTimeout(cs.config.Precommit(round), height, round, cstypes.RoundStepPrecommitWait)
  934. }
  935. // Enter: +2/3 precommits for block
  936. func (cs *ConsensusState) enterCommit(height int, commitRound int) {
  937. if cs.Height != height || cstypes.RoundStepCommit <= cs.Step {
  938. cs.Logger.Debug(cmn.Fmt("enterCommit(%v/%v): Invalid args. Current step: %v/%v/%v", height, commitRound, cs.Height, cs.Round, cs.Step))
  939. return
  940. }
  941. cs.Logger.Info(cmn.Fmt("enterCommit(%v/%v). Current: %v/%v/%v", height, commitRound, cs.Height, cs.Round, cs.Step))
  942. defer func() {
  943. // Done enterCommit:
  944. // keep cs.Round the same, commitRound points to the right Precommits set.
  945. cs.updateRoundStep(cs.Round, cstypes.RoundStepCommit)
  946. cs.CommitRound = commitRound
  947. cs.CommitTime = time.Now()
  948. cs.newStep()
  949. // Maybe finalize immediately.
  950. cs.tryFinalizeCommit(height)
  951. }()
  952. blockID, ok := cs.Votes.Precommits(commitRound).TwoThirdsMajority()
  953. if !ok {
  954. cmn.PanicSanity("RunActionCommit() expects +2/3 precommits")
  955. }
  956. // The Locked* fields no longer matter.
  957. // Move them over to ProposalBlock if they match the commit hash,
  958. // otherwise they'll be cleared in updateToState.
  959. if cs.LockedBlock.HashesTo(blockID.Hash) {
  960. cs.ProposalBlock = cs.LockedBlock
  961. cs.ProposalBlockParts = cs.LockedBlockParts
  962. }
  963. // If we don't have the block being committed, set up to get it.
  964. if !cs.ProposalBlock.HashesTo(blockID.Hash) {
  965. if !cs.ProposalBlockParts.HasHeader(blockID.PartsHeader) {
  966. // We're getting the wrong block.
  967. // Set up ProposalBlockParts and keep waiting.
  968. cs.ProposalBlock = nil
  969. cs.ProposalBlockParts = types.NewPartSetFromHeader(blockID.PartsHeader)
  970. } else {
  971. // We just need to keep waiting.
  972. }
  973. }
  974. }
  975. // If we have the block AND +2/3 commits for it, finalize.
  976. func (cs *ConsensusState) tryFinalizeCommit(height int) {
  977. if cs.Height != height {
  978. cmn.PanicSanity(cmn.Fmt("tryFinalizeCommit() cs.Height: %v vs height: %v", cs.Height, height))
  979. }
  980. blockID, ok := cs.Votes.Precommits(cs.CommitRound).TwoThirdsMajority()
  981. if !ok || len(blockID.Hash) == 0 {
  982. cs.Logger.Error("Attempt to finalize failed. There was no +2/3 majority, or +2/3 was for <nil>.", "height", height)
  983. return
  984. }
  985. if !cs.ProposalBlock.HashesTo(blockID.Hash) {
  986. // TODO: this happens every time if we're not a validator (ugly logs)
  987. // TODO: ^^ wait, why does it matter that we're a validator?
  988. cs.Logger.Info("Attempt to finalize failed. We don't have the commit block.", "height", height, "proposal-block", cs.ProposalBlock.Hash(), "commit-block", blockID.Hash)
  989. return
  990. }
  991. // go
  992. cs.finalizeCommit(height)
  993. }
  994. // Increment height and goto cstypes.RoundStepNewHeight
  995. func (cs *ConsensusState) finalizeCommit(height int) {
  996. if cs.Height != height || cs.Step != cstypes.RoundStepCommit {
  997. cs.Logger.Debug(cmn.Fmt("finalizeCommit(%v): Invalid args. Current step: %v/%v/%v", height, cs.Height, cs.Round, cs.Step))
  998. return
  999. }
  1000. blockID, ok := cs.Votes.Precommits(cs.CommitRound).TwoThirdsMajority()
  1001. block, blockParts := cs.ProposalBlock, cs.ProposalBlockParts
  1002. if !ok {
  1003. cmn.PanicSanity(cmn.Fmt("Cannot finalizeCommit, commit does not have two thirds majority"))
  1004. }
  1005. if !blockParts.HasHeader(blockID.PartsHeader) {
  1006. cmn.PanicSanity(cmn.Fmt("Expected ProposalBlockParts header to be commit header"))
  1007. }
  1008. if !block.HashesTo(blockID.Hash) {
  1009. cmn.PanicSanity(cmn.Fmt("Cannot finalizeCommit, ProposalBlock does not hash to commit hash"))
  1010. }
  1011. if err := cs.state.ValidateBlock(block); err != nil {
  1012. cmn.PanicConsensus(cmn.Fmt("+2/3 committed an invalid block: %v", err))
  1013. }
  1014. cs.Logger.Info(cmn.Fmt("Finalizing commit of block with %d txs", block.NumTxs),
  1015. "height", block.Height, "hash", block.Hash(), "root", block.AppHash)
  1016. cs.Logger.Info(cmn.Fmt("%v", block))
  1017. fail.Fail() // XXX
  1018. // Save to blockStore.
  1019. if cs.blockStore.Height() < block.Height {
  1020. // NOTE: the seenCommit is local justification to commit this block,
  1021. // but may differ from the LastCommit included in the next block
  1022. precommits := cs.Votes.Precommits(cs.CommitRound)
  1023. seenCommit := precommits.MakeCommit()
  1024. cs.blockStore.SaveBlock(block, blockParts, seenCommit)
  1025. } else {
  1026. // Happens during replay if we already saved the block but didn't commit
  1027. cs.Logger.Info("Calling finalizeCommit on already stored block", "height", block.Height)
  1028. }
  1029. fail.Fail() // XXX
  1030. // Finish writing to the WAL for this height.
  1031. // NOTE: If we fail before writing this, we'll never write it,
  1032. // and just recover by running ApplyBlock in the Handshake.
  1033. // If we moved it before persisting the block, we'd have to allow
  1034. // WAL replay for blocks with an #ENDHEIGHT
  1035. // As is, ConsensusState should not be started again
  1036. // until we successfully call ApplyBlock (ie. here or in Handshake after restart)
  1037. cs.wal.Save(EndHeightMessage{uint64(height)})
  1038. fail.Fail() // XXX
  1039. // Create a copy of the state for staging
  1040. // and an event cache for txs
  1041. stateCopy := cs.state.Copy()
  1042. txEventBuffer := types.NewTxEventBuffer(cs.eventBus, block.NumTxs)
  1043. // Execute and commit the block, update and save the state, and update the mempool.
  1044. // All calls to the proxyAppConn come here.
  1045. // NOTE: the block.AppHash wont reflect these txs until the next block
  1046. err := stateCopy.ApplyBlock(txEventBuffer, cs.proxyAppConn, block, blockParts.Header(), cs.mempool)
  1047. if err != nil {
  1048. cs.Logger.Error("Error on ApplyBlock. Did the application crash? Please restart tendermint", "err", err)
  1049. err := cmn.Kill()
  1050. if err != nil {
  1051. cs.Logger.Error("Failed to kill this process - please do so manually", "err", err)
  1052. }
  1053. return
  1054. }
  1055. fail.Fail() // XXX
  1056. // Fire event for new block.
  1057. // NOTE: If we fail before firing, these events will never fire
  1058. //
  1059. // TODO: Either
  1060. // * Fire before persisting state, in ApplyBlock
  1061. // * Fire on start up if we haven't written any new WAL msgs
  1062. // Both options mean we may fire more than once. Is that fine ?
  1063. cs.eventBus.PublishEventNewBlock(types.EventDataNewBlock{block})
  1064. cs.eventBus.PublishEventNewBlockHeader(types.EventDataNewBlockHeader{block.Header})
  1065. err = txEventBuffer.Flush()
  1066. if err != nil {
  1067. cs.Logger.Error("Failed to flush event buffer", "err", err)
  1068. }
  1069. fail.Fail() // XXX
  1070. // NewHeightStep!
  1071. cs.updateToState(stateCopy)
  1072. fail.Fail() // XXX
  1073. // cs.StartTime is already set.
  1074. // Schedule Round0 to start soon.
  1075. cs.scheduleRound0(&cs.RoundState)
  1076. // By here,
  1077. // * cs.Height has been increment to height+1
  1078. // * cs.Step is now cstypes.RoundStepNewHeight
  1079. // * cs.StartTime is set to when we will start round0.
  1080. }
  1081. //-----------------------------------------------------------------------------
  1082. func (cs *ConsensusState) defaultSetProposal(proposal *types.Proposal) error {
  1083. // Already have one
  1084. // TODO: possibly catch double proposals
  1085. if cs.Proposal != nil {
  1086. return nil
  1087. }
  1088. // Does not apply
  1089. if proposal.Height != cs.Height || proposal.Round != cs.Round {
  1090. return nil
  1091. }
  1092. // We don't care about the proposal if we're already in cstypes.RoundStepCommit.
  1093. if cstypes.RoundStepCommit <= cs.Step {
  1094. return nil
  1095. }
  1096. // Verify POLRound, which must be -1 or between 0 and proposal.Round exclusive.
  1097. if proposal.POLRound != -1 &&
  1098. (proposal.POLRound < 0 || proposal.Round <= proposal.POLRound) {
  1099. return ErrInvalidProposalPOLRound
  1100. }
  1101. // Verify signature
  1102. if !cs.Validators.GetProposer().PubKey.VerifyBytes(types.SignBytes(cs.state.ChainID, proposal), proposal.Signature) {
  1103. return ErrInvalidProposalSignature
  1104. }
  1105. cs.Proposal = proposal
  1106. cs.ProposalBlockParts = types.NewPartSetFromHeader(proposal.BlockPartsHeader)
  1107. return nil
  1108. }
  1109. // NOTE: block is not necessarily valid.
  1110. // Asynchronously triggers either enterPrevote (before we timeout of propose) or tryFinalizeCommit, once we have the full block.
  1111. func (cs *ConsensusState) addProposalBlockPart(height int, part *types.Part, verify bool) (added bool, err error) {
  1112. // Blocks might be reused, so round mismatch is OK
  1113. if cs.Height != height {
  1114. return false, nil
  1115. }
  1116. // We're not expecting a block part.
  1117. if cs.ProposalBlockParts == nil {
  1118. return false, nil // TODO: bad peer? Return error?
  1119. }
  1120. added, err = cs.ProposalBlockParts.AddPart(part, verify)
  1121. if err != nil {
  1122. return added, err
  1123. }
  1124. if added && cs.ProposalBlockParts.IsComplete() {
  1125. // Added and completed!
  1126. var n int
  1127. var err error
  1128. cs.ProposalBlock = wire.ReadBinary(&types.Block{}, cs.ProposalBlockParts.GetReader(),
  1129. cs.state.Params.BlockSizeParams.MaxBytes, &n, &err).(*types.Block)
  1130. // NOTE: it's possible to receive complete proposal blocks for future rounds without having the proposal
  1131. cs.Logger.Info("Received complete proposal block", "height", cs.ProposalBlock.Height, "hash", cs.ProposalBlock.Hash())
  1132. if cs.Step == cstypes.RoundStepPropose && cs.isProposalComplete() {
  1133. // Move onto the next step
  1134. cs.enterPrevote(height, cs.Round)
  1135. } else if cs.Step == cstypes.RoundStepCommit {
  1136. // If we're waiting on the proposal block...
  1137. cs.tryFinalizeCommit(height)
  1138. }
  1139. return true, err
  1140. }
  1141. return added, nil
  1142. }
  1143. // Attempt to add the vote. if its a duplicate signature, dupeout the validator
  1144. func (cs *ConsensusState) tryAddVote(vote *types.Vote, peerKey string) error {
  1145. _, err := cs.addVote(vote, peerKey)
  1146. if err != nil {
  1147. // If the vote height is off, we'll just ignore it,
  1148. // But if it's a conflicting sig, broadcast evidence tx for slashing.
  1149. // If it's otherwise invalid, punish peer.
  1150. if err == ErrVoteHeightMismatch {
  1151. return err
  1152. } else if _, ok := err.(*types.ErrVoteConflictingVotes); ok {
  1153. if bytes.Equal(vote.ValidatorAddress, cs.privValidator.GetAddress()) {
  1154. cs.Logger.Error("Found conflicting vote from ourselves. Did you unsafe_reset a validator?", "height", vote.Height, "round", vote.Round, "type", vote.Type)
  1155. return err
  1156. }
  1157. cs.Logger.Error("Found conflicting vote. Publish evidence (TODO)", "height", vote.Height, "round", vote.Round, "type", vote.Type, "valAddr", vote.ValidatorAddress, "valIndex", vote.ValidatorIndex)
  1158. // TODO: track evidence for inclusion in a block
  1159. return err
  1160. } else {
  1161. // Probably an invalid signature. Bad peer.
  1162. cs.Logger.Error("Error attempting to add vote", "err", err)
  1163. return ErrAddingVote
  1164. }
  1165. }
  1166. return nil
  1167. }
  1168. //-----------------------------------------------------------------------------
  1169. func (cs *ConsensusState) addVote(vote *types.Vote, peerKey string) (added bool, err error) {
  1170. cs.Logger.Debug("addVote", "voteHeight", vote.Height, "voteType", vote.Type, "valIndex", vote.ValidatorIndex, "csHeight", cs.Height)
  1171. // A precommit for the previous height?
  1172. // These come in while we wait timeoutCommit
  1173. if vote.Height+1 == cs.Height {
  1174. if !(cs.Step == cstypes.RoundStepNewHeight && vote.Type == types.VoteTypePrecommit) {
  1175. // TODO: give the reason ..
  1176. // fmt.Errorf("tryAddVote: Wrong height, not a LastCommit straggler commit.")
  1177. return added, ErrVoteHeightMismatch
  1178. }
  1179. added, err = cs.LastCommit.AddVote(vote)
  1180. if added {
  1181. cs.Logger.Info(cmn.Fmt("Added to lastPrecommits: %v", cs.LastCommit.StringShort()))
  1182. cs.eventBus.PublishEventVote(types.EventDataVote{vote})
  1183. // if we can skip timeoutCommit and have all the votes now,
  1184. if cs.config.SkipTimeoutCommit && cs.LastCommit.HasAll() {
  1185. // go straight to new round (skip timeout commit)
  1186. // cs.scheduleTimeout(time.Duration(0), cs.Height, 0, cstypes.RoundStepNewHeight)
  1187. cs.enterNewRound(cs.Height, 0)
  1188. }
  1189. }
  1190. return
  1191. }
  1192. // A prevote/precommit for this height?
  1193. if vote.Height == cs.Height {
  1194. height := cs.Height
  1195. added, err = cs.Votes.AddVote(vote, peerKey)
  1196. if added {
  1197. cs.eventBus.PublishEventVote(types.EventDataVote{vote})
  1198. switch vote.Type {
  1199. case types.VoteTypePrevote:
  1200. prevotes := cs.Votes.Prevotes(vote.Round)
  1201. cs.Logger.Info("Added to prevote", "vote", vote, "prevotes", prevotes.StringShort())
  1202. // First, unlock if prevotes is a valid POL.
  1203. // >> lockRound < POLRound <= unlockOrChangeLockRound (see spec)
  1204. // NOTE: If (lockRound < POLRound) but !(POLRound <= unlockOrChangeLockRound),
  1205. // we'll still enterNewRound(H,vote.R) and enterPrecommit(H,vote.R) to process it
  1206. // there.
  1207. if (cs.LockedBlock != nil) && (cs.LockedRound < vote.Round) && (vote.Round <= cs.Round) {
  1208. blockID, ok := prevotes.TwoThirdsMajority()
  1209. if ok && !cs.LockedBlock.HashesTo(blockID.Hash) {
  1210. cs.Logger.Info("Unlocking because of POL.", "lockedRound", cs.LockedRound, "POLRound", vote.Round)
  1211. cs.LockedRound = 0
  1212. cs.LockedBlock = nil
  1213. cs.LockedBlockParts = nil
  1214. cs.eventBus.PublishEventUnlock(cs.RoundStateEvent())
  1215. }
  1216. }
  1217. if cs.Round <= vote.Round && prevotes.HasTwoThirdsAny() {
  1218. // Round-skip over to PrevoteWait or goto Precommit.
  1219. cs.enterNewRound(height, vote.Round) // if the vote is ahead of us
  1220. if prevotes.HasTwoThirdsMajority() {
  1221. cs.enterPrecommit(height, vote.Round)
  1222. } else {
  1223. cs.enterPrevote(height, vote.Round) // if the vote is ahead of us
  1224. cs.enterPrevoteWait(height, vote.Round)
  1225. }
  1226. } else if cs.Proposal != nil && 0 <= cs.Proposal.POLRound && cs.Proposal.POLRound == vote.Round {
  1227. // If the proposal is now complete, enter prevote of cs.Round.
  1228. if cs.isProposalComplete() {
  1229. cs.enterPrevote(height, cs.Round)
  1230. }
  1231. }
  1232. case types.VoteTypePrecommit:
  1233. precommits := cs.Votes.Precommits(vote.Round)
  1234. cs.Logger.Info("Added to precommit", "vote", vote, "precommits", precommits.StringShort())
  1235. blockID, ok := precommits.TwoThirdsMajority()
  1236. if ok {
  1237. if len(blockID.Hash) == 0 {
  1238. cs.enterNewRound(height, vote.Round+1)
  1239. } else {
  1240. cs.enterNewRound(height, vote.Round)
  1241. cs.enterPrecommit(height, vote.Round)
  1242. cs.enterCommit(height, vote.Round)
  1243. if cs.config.SkipTimeoutCommit && precommits.HasAll() {
  1244. // if we have all the votes now,
  1245. // go straight to new round (skip timeout commit)
  1246. // cs.scheduleTimeout(time.Duration(0), cs.Height, 0, cstypes.RoundStepNewHeight)
  1247. cs.enterNewRound(cs.Height, 0)
  1248. }
  1249. }
  1250. } else if cs.Round <= vote.Round && precommits.HasTwoThirdsAny() {
  1251. cs.enterNewRound(height, vote.Round)
  1252. cs.enterPrecommit(height, vote.Round)
  1253. cs.enterPrecommitWait(height, vote.Round)
  1254. }
  1255. default:
  1256. cmn.PanicSanity(cmn.Fmt("Unexpected vote type %X", vote.Type)) // Should not happen.
  1257. }
  1258. }
  1259. // Either duplicate, or error upon cs.Votes.AddByIndex()
  1260. return
  1261. } else {
  1262. err = ErrVoteHeightMismatch
  1263. }
  1264. // Height mismatch, bad peer?
  1265. cs.Logger.Info("Vote ignored and not added", "voteHeight", vote.Height, "csHeight", cs.Height, "err", err)
  1266. return
  1267. }
  1268. func (cs *ConsensusState) signVote(type_ byte, hash []byte, header types.PartSetHeader) (*types.Vote, error) {
  1269. addr := cs.privValidator.GetAddress()
  1270. valIndex, _ := cs.Validators.GetByAddress(addr)
  1271. vote := &types.Vote{
  1272. ValidatorAddress: addr,
  1273. ValidatorIndex: valIndex,
  1274. Height: cs.Height,
  1275. Round: cs.Round,
  1276. Type: type_,
  1277. BlockID: types.BlockID{hash, header},
  1278. }
  1279. err := cs.privValidator.SignVote(cs.state.ChainID, vote)
  1280. return vote, err
  1281. }
  1282. // sign the vote and publish on internalMsgQueue
  1283. func (cs *ConsensusState) signAddVote(type_ byte, hash []byte, header types.PartSetHeader) *types.Vote {
  1284. // if we don't have a key or we're not in the validator set, do nothing
  1285. if cs.privValidator == nil || !cs.Validators.HasAddress(cs.privValidator.GetAddress()) {
  1286. return nil
  1287. }
  1288. vote, err := cs.signVote(type_, hash, header)
  1289. if err == nil {
  1290. cs.sendInternalMessage(msgInfo{&VoteMessage{vote}, ""})
  1291. cs.Logger.Info("Signed and pushed vote", "height", cs.Height, "round", cs.Round, "vote", vote, "err", err)
  1292. return vote
  1293. } else {
  1294. //if !cs.replayMode {
  1295. cs.Logger.Error("Error signing vote", "height", cs.Height, "round", cs.Round, "vote", vote, "err", err)
  1296. //}
  1297. return nil
  1298. }
  1299. }
  1300. //---------------------------------------------------------
  1301. func CompareHRS(h1, r1 int, s1 cstypes.RoundStepType, h2, r2 int, s2 cstypes.RoundStepType) int {
  1302. if h1 < h2 {
  1303. return -1
  1304. } else if h1 > h2 {
  1305. return 1
  1306. }
  1307. if r1 < r2 {
  1308. return -1
  1309. } else if r1 > r2 {
  1310. return 1
  1311. }
  1312. if s1 < s2 {
  1313. return -1
  1314. } else if s1 > s2 {
  1315. return 1
  1316. }
  1317. return 0
  1318. }