|
package state
|
|
|
|
import (
|
|
"bytes"
|
|
"errors"
|
|
"fmt"
|
|
|
|
"github.com/gogo/protobuf/proto"
|
|
"github.com/google/orderedcode"
|
|
dbm "github.com/tendermint/tm-db"
|
|
|
|
abci "github.com/tendermint/tendermint/abci/types"
|
|
tmmath "github.com/tendermint/tendermint/libs/math"
|
|
tmos "github.com/tendermint/tendermint/libs/os"
|
|
tmstate "github.com/tendermint/tendermint/proto/tendermint/state"
|
|
tmproto "github.com/tendermint/tendermint/proto/tendermint/types"
|
|
"github.com/tendermint/tendermint/types"
|
|
)
|
|
|
|
const (
|
|
// persist validators every valSetCheckpointInterval blocks to avoid
|
|
// LoadValidators taking too much time.
|
|
// https://github.com/tendermint/tendermint/pull/3438
|
|
// 100000 results in ~ 100ms to get 100 validators (see BenchmarkLoadValidators)
|
|
valSetCheckpointInterval = 100000
|
|
)
|
|
|
|
//------------------------------------------------------------------------
|
|
|
|
const (
|
|
// prefixes are unique across all tm db's
|
|
prefixValidators = int64(5)
|
|
prefixConsensusParams = int64(6)
|
|
prefixABCIResponses = int64(7)
|
|
prefixState = int64(8)
|
|
)
|
|
|
|
func encodeKey(prefix int64, height int64) []byte {
|
|
res, err := orderedcode.Append(nil, prefix, height)
|
|
if err != nil {
|
|
panic(err)
|
|
}
|
|
return res
|
|
}
|
|
|
|
func validatorsKey(height int64) []byte {
|
|
return encodeKey(prefixValidators, height)
|
|
}
|
|
|
|
func consensusParamsKey(height int64) []byte {
|
|
return encodeKey(prefixConsensusParams, height)
|
|
}
|
|
|
|
func abciResponsesKey(height int64) []byte {
|
|
return encodeKey(prefixABCIResponses, height)
|
|
}
|
|
|
|
// stateKey should never change after being set in init()
|
|
var stateKey []byte
|
|
|
|
func init() {
|
|
var err error
|
|
stateKey, err = orderedcode.Append(nil, prefixState)
|
|
if err != nil {
|
|
panic(err)
|
|
}
|
|
}
|
|
|
|
//----------------------
|
|
|
|
//go:generate ../scripts/mockery_generate.sh Store
|
|
|
|
// Store defines the state store interface
|
|
//
|
|
// It is used to retrieve current state and save and load ABCI responses,
|
|
// validators and consensus parameters
|
|
type Store interface {
|
|
// Load loads the current state of the blockchain
|
|
Load() (State, error)
|
|
// LoadValidators loads the validator set at a given height
|
|
LoadValidators(int64) (*types.ValidatorSet, error)
|
|
// LoadABCIResponses loads the abciResponse for a given height
|
|
LoadABCIResponses(int64) (*tmstate.ABCIResponses, error)
|
|
// LoadConsensusParams loads the consensus params for a given height
|
|
LoadConsensusParams(int64) (types.ConsensusParams, error)
|
|
// Save overwrites the previous state with the updated one
|
|
Save(State) error
|
|
// SaveABCIResponses saves ABCIResponses for a given height
|
|
SaveABCIResponses(int64, *tmstate.ABCIResponses) error
|
|
// SaveValidatorSet saves the validator set at a given height
|
|
SaveValidatorSets(int64, int64, *types.ValidatorSet) error
|
|
// Bootstrap is used for bootstrapping state when not starting from a initial height.
|
|
Bootstrap(State) error
|
|
// PruneStates takes the height from which to prune up to (exclusive)
|
|
PruneStates(int64) error
|
|
}
|
|
|
|
// dbStore wraps a db (github.com/tendermint/tm-db)
|
|
type dbStore struct {
|
|
db dbm.DB
|
|
}
|
|
|
|
var _ Store = (*dbStore)(nil)
|
|
|
|
// NewStore creates the dbStore of the state pkg.
|
|
func NewStore(db dbm.DB) Store {
|
|
return dbStore{db}
|
|
}
|
|
|
|
// LoadState loads the State from the database.
|
|
func (store dbStore) Load() (State, error) {
|
|
return store.loadState(stateKey)
|
|
}
|
|
|
|
func (store dbStore) loadState(key []byte) (state State, err error) {
|
|
buf, err := store.db.Get(key)
|
|
if err != nil {
|
|
return state, err
|
|
}
|
|
if len(buf) == 0 {
|
|
return state, nil
|
|
}
|
|
|
|
sp := new(tmstate.State)
|
|
|
|
err = proto.Unmarshal(buf, sp)
|
|
if err != nil {
|
|
// DATA HAS BEEN CORRUPTED OR THE SPEC HAS CHANGED
|
|
tmos.Exit(fmt.Sprintf(`LoadState: Data has been corrupted or its spec has changed:
|
|
%v\n`, err))
|
|
}
|
|
|
|
sm, err := StateFromProto(sp)
|
|
if err != nil {
|
|
return state, err
|
|
}
|
|
|
|
return *sm, nil
|
|
}
|
|
|
|
// Save persists the State, the ValidatorsInfo, and the ConsensusParamsInfo to the database.
|
|
// This flushes the writes (e.g. calls SetSync).
|
|
func (store dbStore) Save(state State) error {
|
|
return store.save(state, stateKey)
|
|
}
|
|
|
|
func (store dbStore) save(state State, key []byte) error {
|
|
batch := store.db.NewBatch()
|
|
defer batch.Close()
|
|
|
|
nextHeight := state.LastBlockHeight + 1
|
|
// If first block, save validators for the block.
|
|
if nextHeight == 1 {
|
|
nextHeight = state.InitialHeight
|
|
// This extra logic due to Tendermint validator set changes being delayed 1 block.
|
|
// It may get overwritten due to InitChain validator updates.
|
|
if err := store.saveValidatorsInfo(nextHeight, nextHeight, state.Validators, batch); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
// Save next validators.
|
|
err := store.saveValidatorsInfo(nextHeight+1, state.LastHeightValidatorsChanged, state.NextValidators, batch)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// Save next consensus params.
|
|
if err := store.saveConsensusParamsInfo(nextHeight,
|
|
state.LastHeightConsensusParamsChanged, state.ConsensusParams, batch); err != nil {
|
|
return err
|
|
}
|
|
|
|
if err := batch.Set(key, state.Bytes()); err != nil {
|
|
return err
|
|
}
|
|
|
|
return batch.WriteSync()
|
|
}
|
|
|
|
// BootstrapState saves a new state, used e.g. by state sync when starting from non-zero height.
|
|
func (store dbStore) Bootstrap(state State) error {
|
|
height := state.LastBlockHeight + 1
|
|
if height == 1 {
|
|
height = state.InitialHeight
|
|
}
|
|
|
|
batch := store.db.NewBatch()
|
|
defer batch.Close()
|
|
|
|
if height > 1 && !state.LastValidators.IsNilOrEmpty() {
|
|
if err := store.saveValidatorsInfo(height-1, height-1, state.LastValidators, batch); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
|
|
if err := store.saveValidatorsInfo(height, height, state.Validators, batch); err != nil {
|
|
return err
|
|
}
|
|
|
|
if err := store.saveValidatorsInfo(height+1, height+1, state.NextValidators, batch); err != nil {
|
|
return err
|
|
}
|
|
|
|
if err := store.saveConsensusParamsInfo(height,
|
|
state.LastHeightConsensusParamsChanged, state.ConsensusParams, batch); err != nil {
|
|
return err
|
|
}
|
|
|
|
if err := batch.Set(stateKey, state.Bytes()); err != nil {
|
|
return err
|
|
}
|
|
|
|
return batch.WriteSync()
|
|
}
|
|
|
|
// PruneStates deletes states up to the height specified (exclusive). It is not
|
|
// guaranteed to delete all states, since the last checkpointed state and states being pointed to by
|
|
// e.g. `LastHeightChanged` must remain. The state at retain height must also exist.
|
|
// Pruning is done in descending order.
|
|
func (store dbStore) PruneStates(retainHeight int64) error {
|
|
if retainHeight <= 0 {
|
|
return fmt.Errorf("height %v must be greater than 0", retainHeight)
|
|
}
|
|
|
|
// NOTE: We need to prune consensus params first because the validator
|
|
// sets have always one extra height. If validator sets were pruned first
|
|
// we could get a situation where we prune up to the last validator set
|
|
// yet don't have the respective consensus params at that height and thus
|
|
// return an error
|
|
if err := store.pruneConsensusParams(retainHeight); err != nil {
|
|
return err
|
|
}
|
|
|
|
if err := store.pruneValidatorSets(retainHeight); err != nil {
|
|
return err
|
|
}
|
|
|
|
if err := store.pruneABCIResponses(retainHeight); err != nil {
|
|
return err
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
// pruneValidatorSets calls a reverse iterator from base height to retain height (exclusive), deleting
|
|
// all validator sets in between. Due to the fact that most validator sets stored reference an earlier
|
|
// validator set, it is likely that there will remain one validator set left after pruning.
|
|
func (store dbStore) pruneValidatorSets(retainHeight int64) error {
|
|
valInfo, err := loadValidatorsInfo(store.db, retainHeight)
|
|
if err != nil {
|
|
return fmt.Errorf("validators at height %v not found: %w", retainHeight, err)
|
|
}
|
|
|
|
// We will prune up to the validator set at the given "height". As we don't save validator sets every
|
|
// height but only when they change or at a check point, it is likely that the validator set at the height
|
|
// we prune to is empty and thus dependent on the validator set saved at a previous height. We must find
|
|
// that validator set and make sure it is not pruned.
|
|
lastRecordedValSetHeight := lastStoredHeightFor(retainHeight, valInfo.LastHeightChanged)
|
|
lastRecordedValSet, err := loadValidatorsInfo(store.db, lastRecordedValSetHeight)
|
|
if err != nil || lastRecordedValSet.ValidatorSet == nil {
|
|
return fmt.Errorf("couldn't find validators at height %d (height %d was originally requested): %w",
|
|
lastStoredHeightFor(retainHeight, valInfo.LastHeightChanged),
|
|
retainHeight,
|
|
err,
|
|
)
|
|
}
|
|
|
|
// if this is not equal to the retain height, prune from the retain height to the height above
|
|
// the last saved validator set. This way we can skip over the dependent validator set.
|
|
if lastRecordedValSetHeight < retainHeight {
|
|
err := store.pruneRange(
|
|
validatorsKey(lastRecordedValSetHeight+1),
|
|
validatorsKey(retainHeight),
|
|
)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
}
|
|
|
|
// prune all the validators sets up to last saved validator set
|
|
return store.pruneRange(
|
|
validatorsKey(1),
|
|
validatorsKey(lastRecordedValSetHeight),
|
|
)
|
|
}
|
|
|
|
// pruneConsensusParams calls a reverse iterator from base height to retain height batch deleting
|
|
// all consensus params in between. If the consensus params at the new base height is dependent
|
|
// on a prior height then this will keep that lower height too.
|
|
func (store dbStore) pruneConsensusParams(retainHeight int64) error {
|
|
paramsInfo, err := store.loadConsensusParamsInfo(retainHeight)
|
|
if err != nil {
|
|
return fmt.Errorf("consensus params at height %v not found: %w", retainHeight, err)
|
|
}
|
|
|
|
// As we don't save the consensus params at every height, only when there is a consensus params change,
|
|
// we must not prune (or save) the last consensus params that the consensus params info at height
|
|
// is dependent on.
|
|
if paramsInfo.ConsensusParams.Equal(&tmproto.ConsensusParams{}) {
|
|
// sanity check that the consensus params at the last height it was changed is there
|
|
lastRecordedConsensusParams, err := store.loadConsensusParamsInfo(paramsInfo.LastHeightChanged)
|
|
if err != nil || lastRecordedConsensusParams.ConsensusParams.Equal(&tmproto.ConsensusParams{}) {
|
|
return fmt.Errorf(
|
|
"couldn't find consensus params at height %d (height %d was originally requested): %w",
|
|
paramsInfo.LastHeightChanged,
|
|
retainHeight,
|
|
err,
|
|
)
|
|
}
|
|
|
|
// prune the params above the height with which it last changed and below the retain height.
|
|
err = store.pruneRange(
|
|
consensusParamsKey(paramsInfo.LastHeightChanged+1),
|
|
consensusParamsKey(retainHeight),
|
|
)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
}
|
|
|
|
// prune all the consensus params up to either the last height the params changed or if the params
|
|
// last changed at the retain height, then up to the retain height.
|
|
return store.pruneRange(
|
|
consensusParamsKey(1),
|
|
consensusParamsKey(paramsInfo.LastHeightChanged),
|
|
)
|
|
}
|
|
|
|
// pruneABCIResponses calls a reverse iterator from base height to retain height batch deleting
|
|
// all abci responses in between
|
|
func (store dbStore) pruneABCIResponses(height int64) error {
|
|
return store.pruneRange(abciResponsesKey(1), abciResponsesKey(height))
|
|
}
|
|
|
|
// pruneRange is a generic function for deleting a range of keys in reverse order.
|
|
// we keep filling up batches of at most 1000 keys, perform a deletion and continue until
|
|
// we have gone through all of keys in the range. This avoids doing any writes whilst
|
|
// iterating.
|
|
func (store dbStore) pruneRange(start []byte, end []byte) error {
|
|
var err error
|
|
batch := store.db.NewBatch()
|
|
defer batch.Close()
|
|
|
|
end, err = store.reverseBatchDelete(batch, start, end)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// iterate until the last batch of the pruning range in which case we will perform a
|
|
// write sync
|
|
for !bytes.Equal(start, end) {
|
|
if err := batch.Write(); err != nil {
|
|
return err
|
|
}
|
|
|
|
if err := batch.Close(); err != nil {
|
|
return err
|
|
}
|
|
|
|
batch = store.db.NewBatch()
|
|
|
|
// fill a new batch of keys for deletion over the remainding range
|
|
end, err = store.reverseBatchDelete(batch, start, end)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
}
|
|
|
|
return batch.WriteSync()
|
|
}
|
|
|
|
// reverseBatchDelete runs a reverse iterator (from end to start) filling up a batch until either
|
|
// (a) the iterator reaches the start or (b) the iterator has added a 1000 keys (this avoids the
|
|
// batch from growing too large)
|
|
func (store dbStore) reverseBatchDelete(batch dbm.Batch, start, end []byte) ([]byte, error) {
|
|
iter, err := store.db.ReverseIterator(start, end)
|
|
if err != nil {
|
|
return end, fmt.Errorf("iterator error: %w", err)
|
|
}
|
|
defer iter.Close()
|
|
|
|
size := 0
|
|
for ; iter.Valid(); iter.Next() {
|
|
if err := batch.Delete(iter.Key()); err != nil {
|
|
return end, fmt.Errorf("pruning error at key %X: %w", iter.Key(), err)
|
|
}
|
|
|
|
// avoid batches growing too large by capping them
|
|
size++
|
|
if size == 1000 {
|
|
return iter.Key(), iter.Error()
|
|
}
|
|
}
|
|
return start, iter.Error()
|
|
}
|
|
|
|
//------------------------------------------------------------------------
|
|
|
|
// ABCIResponsesResultsHash returns the root hash of a Merkle tree of
|
|
// ResponseDeliverTx responses (see ABCIResults.Hash)
|
|
//
|
|
// See merkle.SimpleHashFromByteSlices
|
|
func ABCIResponsesResultsHash(ar *tmstate.ABCIResponses) []byte {
|
|
return types.NewResults(ar.DeliverTxs).Hash()
|
|
}
|
|
|
|
// LoadABCIResponses loads the ABCIResponses for the given height from the
|
|
// database. If not found, ErrNoABCIResponsesForHeight is returned.
|
|
//
|
|
// This is useful for recovering from crashes where we called app.Commit and
|
|
// before we called s.Save(). It can also be used to produce Merkle proofs of
|
|
// the result of txs.
|
|
func (store dbStore) LoadABCIResponses(height int64) (*tmstate.ABCIResponses, error) {
|
|
buf, err := store.db.Get(abciResponsesKey(height))
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
if len(buf) == 0 {
|
|
|
|
return nil, ErrNoABCIResponsesForHeight{height}
|
|
}
|
|
|
|
abciResponses := new(tmstate.ABCIResponses)
|
|
err = abciResponses.Unmarshal(buf)
|
|
if err != nil {
|
|
// DATA HAS BEEN CORRUPTED OR THE SPEC HAS CHANGED
|
|
tmos.Exit(fmt.Sprintf(`LoadABCIResponses: Data has been corrupted or its spec has
|
|
changed: %v\n`, err))
|
|
}
|
|
// TODO: ensure that buf is completely read.
|
|
|
|
return abciResponses, nil
|
|
}
|
|
|
|
// SaveABCIResponses persists the ABCIResponses to the database.
|
|
// This is useful in case we crash after app.Commit and before s.Save().
|
|
// Responses are indexed by height so they can also be loaded later to produce
|
|
// Merkle proofs.
|
|
//
|
|
// Exposed for testing.
|
|
func (store dbStore) SaveABCIResponses(height int64, abciResponses *tmstate.ABCIResponses) error {
|
|
return store.saveABCIResponses(height, abciResponses)
|
|
}
|
|
|
|
func (store dbStore) saveABCIResponses(height int64, abciResponses *tmstate.ABCIResponses) error {
|
|
var dtxs []*abci.ResponseDeliverTx
|
|
// strip nil values,
|
|
for _, tx := range abciResponses.DeliverTxs {
|
|
if tx != nil {
|
|
dtxs = append(dtxs, tx)
|
|
}
|
|
}
|
|
|
|
abciResponses.DeliverTxs = dtxs
|
|
|
|
bz, err := abciResponses.Marshal()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
return store.db.SetSync(abciResponsesKey(height), bz)
|
|
}
|
|
|
|
// SaveValidatorSets is used to save the validator set over multiple heights.
|
|
// It is exposed so that a backfill operation during state sync can populate
|
|
// the store with the necessary amount of validator sets to verify any evidence
|
|
// it may encounter.
|
|
func (store dbStore) SaveValidatorSets(lowerHeight, upperHeight int64, vals *types.ValidatorSet) error {
|
|
batch := store.db.NewBatch()
|
|
defer batch.Close()
|
|
|
|
// batch together all the validator sets from lowerHeight to upperHeight
|
|
for height := lowerHeight; height <= upperHeight; height++ {
|
|
if err := store.saveValidatorsInfo(height, lowerHeight, vals, batch); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
|
|
return batch.WriteSync()
|
|
}
|
|
|
|
//-----------------------------------------------------------------------------
|
|
|
|
// LoadValidators loads the ValidatorSet for a given height.
|
|
// Returns ErrNoValSetForHeight if the validator set can't be found for this height.
|
|
func (store dbStore) LoadValidators(height int64) (*types.ValidatorSet, error) {
|
|
|
|
valInfo, err := loadValidatorsInfo(store.db, height)
|
|
if err != nil {
|
|
return nil, ErrNoValSetForHeight{height}
|
|
}
|
|
if valInfo.ValidatorSet == nil {
|
|
lastStoredHeight := lastStoredHeightFor(height, valInfo.LastHeightChanged)
|
|
valInfo2, err := loadValidatorsInfo(store.db, lastStoredHeight)
|
|
if err != nil || valInfo2.ValidatorSet == nil {
|
|
return nil,
|
|
fmt.Errorf("couldn't find validators at height %d (height %d was originally requested): %w",
|
|
lastStoredHeight,
|
|
height,
|
|
err,
|
|
)
|
|
}
|
|
|
|
vs, err := types.ValidatorSetFromProto(valInfo2.ValidatorSet)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
vs.IncrementProposerPriority(tmmath.SafeConvertInt32(height - lastStoredHeight)) // mutate
|
|
vi2, err := vs.ToProto()
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
valInfo2.ValidatorSet = vi2
|
|
valInfo = valInfo2
|
|
}
|
|
|
|
vip, err := types.ValidatorSetFromProto(valInfo.ValidatorSet)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
return vip, nil
|
|
}
|
|
|
|
func lastStoredHeightFor(height, lastHeightChanged int64) int64 {
|
|
checkpointHeight := height - height%valSetCheckpointInterval
|
|
return tmmath.MaxInt64(checkpointHeight, lastHeightChanged)
|
|
}
|
|
|
|
// CONTRACT: Returned ValidatorsInfo can be mutated.
|
|
func loadValidatorsInfo(db dbm.DB, height int64) (*tmstate.ValidatorsInfo, error) {
|
|
buf, err := db.Get(validatorsKey(height))
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
if len(buf) == 0 {
|
|
return nil, errors.New("value retrieved from db is empty")
|
|
}
|
|
|
|
v := new(tmstate.ValidatorsInfo)
|
|
err = v.Unmarshal(buf)
|
|
if err != nil {
|
|
// DATA HAS BEEN CORRUPTED OR THE SPEC HAS CHANGED
|
|
tmos.Exit(fmt.Sprintf(`LoadValidators: Data has been corrupted or its spec has changed:
|
|
%v\n`, err))
|
|
}
|
|
// TODO: ensure that buf is completely read.
|
|
|
|
return v, nil
|
|
}
|
|
|
|
// saveValidatorsInfo persists the validator set.
|
|
//
|
|
// `height` is the effective height for which the validator is responsible for
|
|
// signing. It should be called from s.Save(), right before the state itself is
|
|
// persisted.
|
|
func (store dbStore) saveValidatorsInfo(
|
|
height, lastHeightChanged int64,
|
|
valSet *types.ValidatorSet,
|
|
batch dbm.Batch,
|
|
) error {
|
|
if lastHeightChanged > height {
|
|
return errors.New("lastHeightChanged cannot be greater than ValidatorsInfo height")
|
|
}
|
|
valInfo := &tmstate.ValidatorsInfo{
|
|
LastHeightChanged: lastHeightChanged,
|
|
}
|
|
// Only persist validator set if it was updated or checkpoint height (see
|
|
// valSetCheckpointInterval) is reached.
|
|
if height == lastHeightChanged || height%valSetCheckpointInterval == 0 {
|
|
pv, err := valSet.ToProto()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
valInfo.ValidatorSet = pv
|
|
}
|
|
|
|
bz, err := valInfo.Marshal()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
return batch.Set(validatorsKey(height), bz)
|
|
}
|
|
|
|
//-----------------------------------------------------------------------------
|
|
|
|
// ConsensusParamsInfo represents the latest consensus params, or the last height it changed
|
|
|
|
// Allocate empty Consensus params at compile time to avoid multiple allocations during runtime
|
|
var (
|
|
empty = types.ConsensusParams{}
|
|
emptypb = tmproto.ConsensusParams{}
|
|
)
|
|
|
|
// LoadConsensusParams loads the ConsensusParams for a given height.
|
|
func (store dbStore) LoadConsensusParams(height int64) (types.ConsensusParams, error) {
|
|
paramsInfo, err := store.loadConsensusParamsInfo(height)
|
|
if err != nil {
|
|
return empty, fmt.Errorf("could not find consensus params for height #%d: %w", height, err)
|
|
}
|
|
|
|
if paramsInfo.ConsensusParams.Equal(&emptypb) {
|
|
paramsInfo2, err := store.loadConsensusParamsInfo(paramsInfo.LastHeightChanged)
|
|
if err != nil {
|
|
return empty, fmt.Errorf(
|
|
"couldn't find consensus params at height %d (height %d was originally requested): %w",
|
|
paramsInfo.LastHeightChanged,
|
|
height,
|
|
err,
|
|
)
|
|
}
|
|
|
|
paramsInfo = paramsInfo2
|
|
}
|
|
|
|
return types.ConsensusParamsFromProto(paramsInfo.ConsensusParams), nil
|
|
}
|
|
|
|
func (store dbStore) loadConsensusParamsInfo(height int64) (*tmstate.ConsensusParamsInfo, error) {
|
|
buf, err := store.db.Get(consensusParamsKey(height))
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
if len(buf) == 0 {
|
|
return nil, errors.New("value retrieved from db is empty")
|
|
}
|
|
|
|
paramsInfo := new(tmstate.ConsensusParamsInfo)
|
|
if err = paramsInfo.Unmarshal(buf); err != nil {
|
|
// DATA HAS BEEN CORRUPTED OR THE SPEC HAS CHANGED
|
|
tmos.Exit(fmt.Sprintf(`LoadConsensusParams: Data has been corrupted or its spec has changed:
|
|
%v\n`, err))
|
|
}
|
|
// TODO: ensure that buf is completely read.
|
|
|
|
return paramsInfo, nil
|
|
}
|
|
|
|
// saveConsensusParamsInfo persists the consensus params for the next block to disk.
|
|
// It should be called from s.Save(), right before the state itself is persisted.
|
|
// If the consensus params did not change after processing the latest block,
|
|
// only the last height for which they changed is persisted.
|
|
func (store dbStore) saveConsensusParamsInfo(
|
|
nextHeight, changeHeight int64,
|
|
params types.ConsensusParams,
|
|
batch dbm.Batch,
|
|
) error {
|
|
paramsInfo := &tmstate.ConsensusParamsInfo{
|
|
LastHeightChanged: changeHeight,
|
|
}
|
|
|
|
if changeHeight == nextHeight {
|
|
paramsInfo.ConsensusParams = params.ToProto()
|
|
}
|
|
bz, err := paramsInfo.Marshal()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
return batch.Set(consensusParamsKey(nextHeight), bz)
|
|
}
|