package main import ( "fmt" "math/rand" "sort" "strings" e2e "github.com/tendermint/tendermint/test/e2e/pkg" ) var ( // testnetCombinations defines global testnet options, where we generate a // separate testnet for each combination (Cartesian product) of options. testnetCombinations = map[string][]interface{}{ "topology": {"single", "quad", "large"}, "ipv6": {false, true}, "initialHeight": {0, 1000}, "initialState": { map[string]string{}, map[string]string{"initial01": "a", "initial02": "b", "initial03": "c"}, }, "validators": {"genesis", "initchain"}, } // The following specify randomly chosen values for testnet nodes. nodeDatabases = uniformChoice{"goleveldb", "cleveldb", "rocksdb", "boltdb", "badgerdb"} // FIXME disabled grpc due to https://github.com/tendermint/tendermint/issues/5439 nodeABCIProtocols = uniformChoice{"unix", "tcp", "builtin"} // "grpc" nodePrivvalProtocols = uniformChoice{"file", "unix", "tcp"} // FIXME disabled v1 due to https://github.com/tendermint/tendermint/issues/5444 nodeFastSyncs = uniformChoice{"", "v0", "v2"} // "v1" nodeStateSyncs = uniformChoice{false, true} nodePersistIntervals = uniformChoice{0, 1, 5} nodeSnapshotIntervals = uniformChoice{0, 3} nodeRetainBlocks = uniformChoice{0, 1, 5} nodePerturbations = probSetChoice{ "disconnect": 0.1, "pause": 0.1, // FIXME disabled due to https://github.com/tendermint/tendermint/issues/5422 // "kill": 0.1, // "restart": 0.1, } ) // Generate generates random testnets using the given RNG. func Generate(r *rand.Rand) ([]e2e.Manifest, error) { manifests := []e2e.Manifest{} for _, opt := range combinations(testnetCombinations) { manifest, err := generateTestnet(r, opt) if err != nil { return nil, err } manifests = append(manifests, manifest) } return manifests, nil } // generateTestnet generates a single testnet with the given options. func generateTestnet(r *rand.Rand, opt map[string]interface{}) (e2e.Manifest, error) { manifest := e2e.Manifest{ IPv6: opt["ipv6"].(bool), InitialHeight: int64(opt["initialHeight"].(int)), InitialState: opt["initialState"].(map[string]string), Validators: &map[string]int64{}, ValidatorUpdates: map[string]map[string]int64{}, Nodes: map[string]*e2e.ManifestNode{}, } var numSeeds, numValidators, numFulls int switch opt["topology"].(string) { case "single": numValidators = 1 case "quad": numValidators = 4 case "large": // FIXME Networks are kept small since large ones use too much CPU. numSeeds = r.Intn(4) numValidators = 4 + r.Intn(7) numFulls = r.Intn(5) default: return manifest, fmt.Errorf("unknown topology %q", opt["topology"]) } // First we generate seed nodes, starting at the initial height. for i := 1; i <= numSeeds; i++ { manifest.Nodes[fmt.Sprintf("seed%02d", i)] = generateNode(r, e2e.ModeSeed, 0, false) } // Next, we generate validators. We make sure a BFT quorum of validators start // at the initial height, and that we have two archive nodes. We also set up // the initial validator set, and validator set updates for delayed nodes. nextStartAt := manifest.InitialHeight + 5 quorum := numValidators*2/3 + 1 for i := 1; i <= numValidators; i++ { startAt := int64(0) if i > quorum { startAt = nextStartAt nextStartAt += 5 } name := fmt.Sprintf("validator%02d", i) manifest.Nodes[name] = generateNode(r, e2e.ModeValidator, startAt, i <= 2) if startAt == 0 { (*manifest.Validators)[name] = int64(30 + r.Intn(71)) } else { manifest.ValidatorUpdates[fmt.Sprint(startAt+5)] = map[string]int64{ name: int64(30 + r.Intn(71)), } } } // Move validators to InitChain if specified. switch opt["validators"].(string) { case "genesis": case "initchain": manifest.ValidatorUpdates["0"] = *manifest.Validators manifest.Validators = &map[string]int64{} default: return manifest, fmt.Errorf("invalid validators option %q", opt["validators"]) } // Finally, we generate random full nodes. for i := 1; i <= numFulls; i++ { startAt := int64(0) if r.Float64() >= 0.5 { startAt = nextStartAt nextStartAt += 5 } manifest.Nodes[fmt.Sprintf("full%02d", i)] = generateNode(r, e2e.ModeFull, startAt, false) } // We now set up peer discovery for nodes. Seed nodes are fully meshed with // each other, while non-seed nodes either use a set of random seeds or a // set of random peers that start before themselves. var seedNames, peerNames []string for name, node := range manifest.Nodes { if node.Mode == string(e2e.ModeSeed) { seedNames = append(seedNames, name) } else { peerNames = append(peerNames, name) } } for _, name := range seedNames { for _, otherName := range seedNames { if name != otherName { manifest.Nodes[name].Seeds = append(manifest.Nodes[name].Seeds, otherName) } } } sort.Slice(peerNames, func(i, j int) bool { iName, jName := peerNames[i], peerNames[j] switch { case manifest.Nodes[iName].StartAt < manifest.Nodes[jName].StartAt: return true case manifest.Nodes[iName].StartAt > manifest.Nodes[jName].StartAt: return false default: return strings.Compare(iName, jName) == -1 } }) for i, name := range peerNames { if len(seedNames) > 0 && (i == 0 || r.Float64() >= 0.5) { manifest.Nodes[name].Seeds = uniformSetChoice(seedNames).Choose(r) } else if i > 0 { manifest.Nodes[name].PersistentPeers = uniformSetChoice(peerNames[:i]).Choose(r) } } return manifest, nil } // generateNode randomly generates a node, with some constraints to avoid // generating invalid configurations. We do not set Seeds or PersistentPeers // here, since we need to know the overall network topology and startup // sequencing. func generateNode(r *rand.Rand, mode e2e.Mode, startAt int64, forceArchive bool) *e2e.ManifestNode { node := e2e.ManifestNode{ Mode: string(mode), StartAt: startAt, Database: nodeDatabases.Choose(r).(string), ABCIProtocol: nodeABCIProtocols.Choose(r).(string), PrivvalProtocol: nodePrivvalProtocols.Choose(r).(string), FastSync: nodeFastSyncs.Choose(r).(string), StateSync: nodeStateSyncs.Choose(r).(bool) && startAt > 0, PersistInterval: ptrUint64(uint64(nodePersistIntervals.Choose(r).(int))), SnapshotInterval: uint64(nodeSnapshotIntervals.Choose(r).(int)), RetainBlocks: uint64(nodeRetainBlocks.Choose(r).(int)), Perturb: nodePerturbations.Choose(r), } // If this node is forced to be an archive node, retain all blocks and // enable state sync snapshotting. if forceArchive { node.RetainBlocks = 0 node.SnapshotInterval = 3 } // If a node which does not persist state also does not retain blocks, randomly // choose to either persist state or retain all blocks. if node.PersistInterval != nil && *node.PersistInterval == 0 && node.RetainBlocks > 0 { if r.Float64() > 0.5 { node.RetainBlocks = 0 } else { node.PersistInterval = ptrUint64(node.RetainBlocks) } } // If either PersistInterval or SnapshotInterval are greater than RetainBlocks, // expand the block retention time. if node.RetainBlocks > 0 { if node.PersistInterval != nil && node.RetainBlocks < *node.PersistInterval { node.RetainBlocks = *node.PersistInterval } if node.RetainBlocks < node.SnapshotInterval { node.RetainBlocks = node.SnapshotInterval } } return &node } func ptrUint64(i uint64) *uint64 { return &i }