// Package keymigrate translates all legacy formatted keys to their
|
|
// new components.
|
|
//
|
|
// The key migration operation as implemented provides a potential
|
|
// model for database migration operations. Crucially, the migration
|
|
// as implemented does not depend on any tendermint code.
|
|
package keymigrate
|
|
|
|
import (
|
|
"bytes"
|
|
"context"
|
|
"encoding/binary"
|
|
"encoding/hex"
|
|
"fmt"
|
|
"math/rand"
|
|
"runtime"
|
|
"strconv"
|
|
"sync"
|
|
|
|
"github.com/google/orderedcode"
|
|
"github.com/pkg/errors"
|
|
dbm "github.com/tendermint/tm-db"
|
|
)
|
|
|
|
type (
|
|
keyID []byte
|
|
migrateFunc func(keyID) (keyID, error)
|
|
)
|
|
|
|
func getAllLegacyKeys(db dbm.DB) ([]keyID, error) {
|
|
out := []keyID{}
|
|
|
|
iter, err := db.Iterator(nil, nil)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
for ; iter.Valid(); iter.Next() {
|
|
k := iter.Key()
|
|
|
|
// make sure it's a key with a legacy format, and skip
|
|
// all other keys, to make it safe to resume the migration.
|
|
if !keyIsLegacy(k) {
|
|
continue
|
|
}
|
|
|
|
// there's inconsistency around tm-db's handling of
|
|
// key copies.
|
|
nk := make([]byte, len(k))
|
|
copy(nk, k)
|
|
out = append(out, nk)
|
|
}
|
|
|
|
if err = iter.Error(); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
if err = iter.Close(); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
return out, nil
|
|
}
|
|
|
|
func makeKeyChan(keys []keyID) <-chan keyID {
|
|
out := make(chan keyID, len(keys))
|
|
defer close(out)
|
|
|
|
for _, key := range keys {
|
|
out <- key
|
|
}
|
|
|
|
return out
|
|
}
|
|
|
|
func keyIsLegacy(key keyID) bool {
|
|
for _, prefix := range []keyID{
|
|
// core "store"
|
|
keyID("consensusParamsKey:"),
|
|
keyID("abciResponsesKey:"),
|
|
keyID("validatorsKey:"),
|
|
keyID("stateKey"),
|
|
keyID("H:"),
|
|
keyID("P:"),
|
|
keyID("C:"),
|
|
keyID("SC:"),
|
|
keyID("BH:"),
|
|
// light
|
|
keyID("size"),
|
|
keyID("lb/"),
|
|
// evidence
|
|
keyID([]byte{0x00}),
|
|
keyID([]byte{0x01}),
|
|
// tx index
|
|
keyID("tx.height/"),
|
|
keyID("tx.hash/"),
|
|
} {
|
|
if bytes.HasPrefix(key, prefix) {
|
|
return true
|
|
}
|
|
}
|
|
|
|
// this means it's a tx index...
|
|
if bytes.Count(key, []byte("/")) >= 3 {
|
|
return true
|
|
}
|
|
|
|
return keyIsHash(key)
|
|
}
|
|
|
|
func keyIsHash(key keyID) bool {
|
|
return len(key) == 32 && !bytes.Contains(key, []byte("/"))
|
|
}
|
|
|
|
func migarateKey(key keyID) (keyID, error) {
|
|
switch {
|
|
case bytes.HasPrefix(key, keyID("H:")):
|
|
val, err := strconv.Atoi(string(key[2:]))
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
return orderedcode.Append(nil, int64(0), int64(val))
|
|
case bytes.HasPrefix(key, keyID("P:")):
|
|
parts := bytes.Split(key[2:], []byte(":"))
|
|
if len(parts) != 2 {
|
|
return nil, fmt.Errorf("block parts key has %d rather than 2 components",
|
|
len(parts))
|
|
}
|
|
valOne, err := strconv.Atoi(string(parts[0]))
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
valTwo, err := strconv.Atoi(string(parts[1]))
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
return orderedcode.Append(nil, int64(1), int64(valOne), int64(valTwo))
|
|
case bytes.HasPrefix(key, keyID("C:")):
|
|
val, err := strconv.Atoi(string(key[2:]))
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
return orderedcode.Append(nil, int64(2), int64(val))
|
|
case bytes.HasPrefix(key, keyID("SC:")):
|
|
val, err := strconv.Atoi(string(key[3:]))
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
return orderedcode.Append(nil, int64(3), int64(val))
|
|
case bytes.HasPrefix(key, keyID("BH:")):
|
|
val, err := strconv.Atoi(string(key[3:]))
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
return orderedcode.Append(nil, int64(4), int64(val))
|
|
case bytes.HasPrefix(key, keyID("validatorsKey:")):
|
|
val, err := strconv.Atoi(string(key[14:]))
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
return orderedcode.Append(nil, int64(5), int64(val))
|
|
case bytes.HasPrefix(key, keyID("consensusParamsKey:")):
|
|
val, err := strconv.Atoi(string(key[19:]))
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
return orderedcode.Append(nil, int64(6), int64(val))
|
|
case bytes.HasPrefix(key, keyID("abciResponsesKey:")):
|
|
val, err := strconv.Atoi(string(key[17:]))
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
return orderedcode.Append(nil, int64(7), int64(val))
|
|
case bytes.HasPrefix(key, keyID("stateKey")):
|
|
return orderedcode.Append(nil, int64(8))
|
|
case bytes.HasPrefix(key, []byte{0x00}): // committed evidence
|
|
return convertEvidence(key, 9)
|
|
case bytes.HasPrefix(key, []byte{0x01}): // pending evidence
|
|
return convertEvidence(key, 10)
|
|
case bytes.HasPrefix(key, keyID("lb/")):
|
|
if len(key) < 24 {
|
|
return nil, fmt.Errorf("light block evidence %q in invalid format", string(key))
|
|
}
|
|
|
|
val, err := strconv.Atoi(string(key[len(key)-20:]))
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
return orderedcode.Append(nil, int64(11), int64(val))
|
|
case bytes.HasPrefix(key, keyID("size")):
|
|
return orderedcode.Append(nil, int64(12))
|
|
case bytes.HasPrefix(key, keyID("tx.height")):
|
|
parts := bytes.Split(key, []byte("/"))
|
|
if len(parts) != 4 {
|
|
return nil, fmt.Errorf("key has %d parts rather than 4", len(parts))
|
|
}
|
|
parts = parts[1:] // drop prefix
|
|
|
|
elems := make([]interface{}, 0, len(parts)+1)
|
|
elems = append(elems, "tx.height")
|
|
|
|
for idx, pt := range parts {
|
|
val, err := strconv.Atoi(string(pt))
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
if idx == 0 {
|
|
elems = append(elems, fmt.Sprintf("%d", val))
|
|
} else {
|
|
elems = append(elems, int64(val))
|
|
}
|
|
}
|
|
|
|
return orderedcode.Append(nil, elems...)
|
|
case bytes.Count(key, []byte("/")) >= 3: // tx indexer
|
|
parts := bytes.Split(key, []byte("/"))
|
|
|
|
elems := make([]interface{}, 0, 4)
|
|
if len(parts) == 4 {
|
|
elems = append(elems, string(parts[0]), string(parts[1]))
|
|
|
|
val, err := strconv.Atoi(string(parts[2]))
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
elems = append(elems, int64(val))
|
|
|
|
val2, err := strconv.Atoi(string(parts[3]))
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
elems = append(elems, int64(val2))
|
|
} else {
|
|
elems = append(elems, string(parts[0]))
|
|
parts = parts[1:]
|
|
|
|
val, err := strconv.Atoi(string(parts[len(parts)-1]))
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
val2, err := strconv.Atoi(string(parts[len(parts)-2]))
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
appKey := bytes.Join(parts[:len(parts)-3], []byte("/"))
|
|
elems = append(elems, string(appKey), int64(val), int64(val2))
|
|
}
|
|
return orderedcode.Append(nil, elems...)
|
|
case keyIsHash(key):
|
|
return orderedcode.Append(nil, "tx.hash", string(key))
|
|
default:
|
|
return nil, fmt.Errorf("key %q is in the wrong format", string(key))
|
|
}
|
|
}
|
|
|
|
func convertEvidence(key keyID, newPrefix int64) ([]byte, error) {
|
|
parts := bytes.Split(key[1:], []byte("/"))
|
|
if len(parts) != 2 {
|
|
return nil, fmt.Errorf("evidence key is malformed with %d parts not 2",
|
|
len(parts))
|
|
}
|
|
|
|
hb, err := hex.DecodeString(string(parts[0]))
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
evidenceHash, err := hex.DecodeString(string(parts[1]))
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
return orderedcode.Append(nil, newPrefix, binary.BigEndian.Uint64(hb), string(evidenceHash))
|
|
}
|
|
|
|
func replaceKey(db dbm.DB, key keyID, gooseFn migrateFunc) error {
|
|
exists, err := db.Has(key)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
if !exists {
|
|
return nil
|
|
}
|
|
|
|
newKey, err := gooseFn(key)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
val, err := db.Get(key)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
batch := db.NewBatch()
|
|
|
|
if err = batch.Set(newKey, val); err != nil {
|
|
return err
|
|
}
|
|
if err = batch.Delete(key); err != nil {
|
|
return err
|
|
}
|
|
|
|
// 10% of the time, force a write to disk, but mostly don't,
|
|
// because it's faster.
|
|
if rand.Intn(100)%10 == 0 { // nolint:gosec
|
|
if err = batch.WriteSync(); err != nil {
|
|
return err
|
|
}
|
|
} else {
|
|
if err = batch.Write(); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
|
|
if err = batch.Close(); err != nil {
|
|
return err
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
// Migrate converts all legacy key formats to new key formats. The
|
|
// operation is idempotent, so it's safe to resume a failed
|
|
// operation. The operation is somewhat parallelized, relying on the
|
|
// concurrency safety of the underlying databases.
|
|
//
|
|
// Migrate has "continue on error" semantics and will iterate through
|
|
// all legacy keys attempt to migrate them, and will collect all
|
|
// errors and will return only at the end of the operation.
|
|
//
|
|
// The context allows for a safe termination of the operation
|
|
// (e.g connected to a singal handler,) to abort the operation
|
|
// in-between migration operations.
|
|
func Migrate(ctx context.Context, db dbm.DB) error {
|
|
keys, err := getAllLegacyKeys(db)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
numWorkers := runtime.NumCPU()
|
|
wg := &sync.WaitGroup{}
|
|
|
|
errs := make(chan error, numWorkers)
|
|
|
|
keyCh := makeKeyChan(keys)
|
|
|
|
// run migrations.
|
|
for i := 0; i < numWorkers; i++ {
|
|
wg.Add(1)
|
|
go func() {
|
|
defer wg.Done()
|
|
for key := range keyCh {
|
|
err := replaceKey(db, key, migarateKey)
|
|
if err != nil {
|
|
errs <- err
|
|
}
|
|
|
|
if ctx.Err() != nil {
|
|
return
|
|
}
|
|
}
|
|
}()
|
|
}
|
|
|
|
// collect and process the errors.
|
|
errStrs := []string{}
|
|
signal := make(chan struct{})
|
|
go func() {
|
|
defer close(signal)
|
|
for err := range errs {
|
|
if err == nil {
|
|
continue
|
|
}
|
|
errStrs = append(errStrs, err.Error())
|
|
}
|
|
}()
|
|
|
|
// Wait for everything to be done.
|
|
wg.Wait()
|
|
close(errs)
|
|
<-signal
|
|
|
|
// check the error results
|
|
if len(errs) != 0 {
|
|
return errors.Errorf("encountered errors during migration: %v", errStrs)
|
|
}
|
|
|
|
return nil
|
|
}
|