- package v0
-
- import (
- "errors"
- "fmt"
- "math"
- "sync/atomic"
- "time"
-
- flow "github.com/tendermint/tendermint/libs/flowrate"
- "github.com/tendermint/tendermint/libs/log"
- "github.com/tendermint/tendermint/libs/service"
- tmsync "github.com/tendermint/tendermint/libs/sync"
- "github.com/tendermint/tendermint/p2p"
- "github.com/tendermint/tendermint/types"
- )
-
- /*
- eg, L = latency = 0.1s
- P = num peers = 10
- FN = num full nodes
- BS = 1kB block size
- CB = 1 Mbit/s = 128 kB/s
- CB/P = 12.8 kB
- B/S = CB/P/BS = 12.8 blocks/s
-
- 12.8 * 0.1 = 1.28 blocks on conn
- */
-
- const (
- requestIntervalMS = 2
- maxTotalRequesters = 600
- maxPendingRequests = maxTotalRequesters
- maxPendingRequestsPerPeer = 20
-
- // Minimum recv rate to ensure we're receiving blocks from a peer fast
- // enough. If a peer is not sending us data at at least that rate, we
- // consider them to have timedout and we disconnect.
- //
- // Assuming a DSL connection (not a good choice) 128 Kbps (upload) ~ 15 KB/s,
- // sending data across atlantic ~ 7.5 KB/s.
- minRecvRate = 7680
-
- // Maximum difference between current and new block's height.
- maxDiffBetweenCurrentAndReceivedBlockHeight = 100
- )
-
- var peerTimeout = 15 * time.Second // not const so we can override with tests
-
- /*
- Peers self report their heights when we join the block pool.
- Starting from our latest pool.height, we request blocks
- in sequence from peers that reported higher heights than ours.
- Every so often we ask peers what height they're on so we can keep going.
-
- Requests are continuously made for blocks of higher heights until
- the limit is reached. If most of the requests have no available peers, and we
- are not at peer limits, we can probably switch to consensus reactor
- */
-
- // BlockRequest stores a block request identified by the block Height and the
- // PeerID responsible for delivering the block.
- type BlockRequest struct {
- Height int64
- PeerID p2p.ID
- }
-
- // BlockPool keeps track of the fast sync peers, block requests and block responses.
- type BlockPool struct {
- service.BaseService
- lastAdvance time.Time
-
- mtx tmsync.RWMutex
- // block requests
- requesters map[int64]*bpRequester
- height int64 // the lowest key in requesters.
- // peers
- peers map[p2p.ID]*bpPeer
- maxPeerHeight int64 // the biggest reported height
-
- // atomic
- numPending int32 // number of requests pending assignment or block response
-
- requestsCh chan<- BlockRequest
- errorsCh chan<- peerError
- }
-
- // NewBlockPool returns a new BlockPool with the height equal to start. Block
- // requests and errors will be sent to requestsCh and errorsCh accordingly.
- func NewBlockPool(start int64, requestsCh chan<- BlockRequest, errorsCh chan<- peerError) *BlockPool {
- bp := &BlockPool{
- peers: make(map[p2p.ID]*bpPeer),
-
- requesters: make(map[int64]*bpRequester),
- height: start,
- numPending: 0,
-
- requestsCh: requestsCh,
- errorsCh: errorsCh,
- }
- bp.BaseService = *service.NewBaseService(nil, "BlockPool", bp)
- return bp
- }
-
- // OnStart implements service.Service by spawning requesters routine and recording
- // pool's start time.
- func (pool *BlockPool) OnStart() error {
- pool.lastAdvance = time.Now()
- go pool.makeRequestersRoutine()
- return nil
- }
-
- // spawns requesters as needed
- func (pool *BlockPool) makeRequestersRoutine() {
- for {
- if !pool.IsRunning() {
- break
- }
-
- _, numPending, lenRequesters := pool.GetStatus()
- switch {
- case numPending >= maxPendingRequests:
- // sleep for a bit.
- time.Sleep(requestIntervalMS * time.Millisecond)
- // check for timed out peers
- pool.removeTimedoutPeers()
- case lenRequesters >= maxTotalRequesters:
- // sleep for a bit.
- time.Sleep(requestIntervalMS * time.Millisecond)
- // check for timed out peers
- pool.removeTimedoutPeers()
- default:
- // request for more blocks.
- pool.makeNextRequester()
- }
- }
- }
-
- func (pool *BlockPool) removeTimedoutPeers() {
- pool.mtx.Lock()
- defer pool.mtx.Unlock()
-
- for _, peer := range pool.peers {
- // check if peer timed out
- if !peer.didTimeout && peer.numPending > 0 {
- curRate := peer.recvMonitor.Status().CurRate
- // curRate can be 0 on start
- if curRate != 0 && curRate < minRecvRate {
- err := errors.New("peer is not sending us data fast enough")
- pool.sendError(err, peer.id)
- pool.Logger.Error("SendTimeout", "peer", peer.id,
- "reason", err,
- "curRate", fmt.Sprintf("%d KB/s", curRate/1024),
- "minRate", fmt.Sprintf("%d KB/s", minRecvRate/1024))
- peer.didTimeout = true
- }
- }
-
- if peer.didTimeout {
- pool.removePeer(peer.id)
- }
- }
- }
-
- // GetStatus returns pool's height, numPending requests and the number of
- // requesters.
- func (pool *BlockPool) GetStatus() (height int64, numPending int32, lenRequesters int) {
- pool.mtx.RLock()
- defer pool.mtx.RUnlock()
-
- return pool.height, atomic.LoadInt32(&pool.numPending), len(pool.requesters)
- }
-
- // IsCaughtUp returns true if this node is caught up, false - otherwise.
- func (pool *BlockPool) IsCaughtUp() bool {
- pool.mtx.RLock()
- defer pool.mtx.RUnlock()
-
- // Need at least 1 peer to be considered caught up.
- if len(pool.peers) == 0 {
- return false
- }
- // NOTE: we use maxPeerHeight - 1 because to sync block H requires block H+1
- // to verify the LastCommit.
- return pool.height >= (pool.maxPeerHeight - 1)
- }
-
- // PeekTwoBlocks returns blocks at pool.height and pool.height+1.
- // We need to see the second block's Commit to validate the first block.
- // So we peek two blocks at a time.
- // The caller will verify the commit.
- func (pool *BlockPool) PeekTwoBlocks() (first *types.Block, second *types.Block) {
- pool.mtx.RLock()
- defer pool.mtx.RUnlock()
-
- if r := pool.requesters[pool.height]; r != nil {
- first = r.getBlock()
- }
- if r := pool.requesters[pool.height+1]; r != nil {
- second = r.getBlock()
- }
- return
- }
-
- // PopRequest pops the first block at pool.height.
- // It must have been validated by 'second'.Commit from PeekTwoBlocks().
- func (pool *BlockPool) PopRequest() {
- pool.mtx.Lock()
- defer pool.mtx.Unlock()
-
- if r := pool.requesters[pool.height]; r != nil {
- if err := r.Stop(); err != nil {
- pool.Logger.Error("Error stopping requester", "err", err)
- }
- delete(pool.requesters, pool.height)
- pool.height++
- pool.lastAdvance = time.Now()
- } else {
- panic(fmt.Sprintf("Expected requester to pop, got nothing at height %v", pool.height))
- }
- }
-
- // RedoRequest invalidates the block at pool.height,
- // Remove the peer and redo request from others.
- // Returns the ID of the removed peer.
- func (pool *BlockPool) RedoRequest(height int64) p2p.ID {
- pool.mtx.Lock()
- defer pool.mtx.Unlock()
-
- request := pool.requesters[height]
- peerID := request.getPeerID()
- if peerID != p2p.ID("") {
- // RemovePeer will redo all requesters associated with this peer.
- pool.removePeer(peerID)
- }
- return peerID
- }
-
- // AddBlock validates that the block comes from the peer it was expected from and calls the requester to store it.
- // TODO: ensure that blocks come in order for each peer.
- func (pool *BlockPool) AddBlock(peerID p2p.ID, block *types.Block, blockSize int) {
- pool.mtx.Lock()
- defer pool.mtx.Unlock()
-
- requester := pool.requesters[block.Height]
- if requester == nil {
- pool.Logger.Error("peer sent us a block we didn't expect",
- "peer", peerID, "curHeight", pool.height, "blockHeight", block.Height)
- diff := pool.height - block.Height
- if diff < 0 {
- diff *= -1
- }
- if diff > maxDiffBetweenCurrentAndReceivedBlockHeight {
- pool.sendError(errors.New("peer sent us a block we didn't expect with a height too far ahead/behind"), peerID)
- }
- return
- }
-
- if requester.setBlock(block, peerID) {
- atomic.AddInt32(&pool.numPending, -1)
- peer := pool.peers[peerID]
- if peer != nil {
- peer.decrPending(blockSize)
- }
- } else {
- err := errors.New("requester is different or block already exists")
- pool.Logger.Error(err.Error(), "peer", peerID, "requester", requester.getPeerID(), "blockHeight", block.Height)
- pool.sendError(err, peerID)
- }
- }
-
- // MaxPeerHeight returns the highest reported height.
- func (pool *BlockPool) MaxPeerHeight() int64 {
- pool.mtx.RLock()
- defer pool.mtx.RUnlock()
- return pool.maxPeerHeight
- }
-
- // LastAdvance returns the time when the last block was processed (or start
- // time if no blocks were processed).
- func (pool *BlockPool) LastAdvance() time.Time {
- pool.mtx.RLock()
- defer pool.mtx.RUnlock()
- return pool.lastAdvance
- }
-
- // SetPeerRange sets the peer's alleged blockchain base and height.
- func (pool *BlockPool) SetPeerRange(peerID p2p.ID, base int64, height int64) {
- pool.mtx.Lock()
- defer pool.mtx.Unlock()
-
- peer := pool.peers[peerID]
- if peer != nil {
- peer.base = base
- peer.height = height
- } else {
- peer = newBPPeer(pool, peerID, base, height)
- peer.setLogger(pool.Logger.With("peer", peerID))
- pool.peers[peerID] = peer
- }
-
- if height > pool.maxPeerHeight {
- pool.maxPeerHeight = height
- }
- }
-
- // RemovePeer removes the peer with peerID from the pool. If there's no peer
- // with peerID, function is a no-op.
- func (pool *BlockPool) RemovePeer(peerID p2p.ID) {
- pool.mtx.Lock()
- defer pool.mtx.Unlock()
-
- pool.removePeer(peerID)
- }
-
- func (pool *BlockPool) removePeer(peerID p2p.ID) {
- for _, requester := range pool.requesters {
- if requester.getPeerID() == peerID {
- requester.redo(peerID)
- }
- }
-
- peer, ok := pool.peers[peerID]
- if ok {
- if peer.timeout != nil {
- peer.timeout.Stop()
- }
-
- delete(pool.peers, peerID)
-
- // Find a new peer with the biggest height and update maxPeerHeight if the
- // peer's height was the biggest.
- if peer.height == pool.maxPeerHeight {
- pool.updateMaxPeerHeight()
- }
- }
- }
-
- // If no peers are left, maxPeerHeight is set to 0.
- func (pool *BlockPool) updateMaxPeerHeight() {
- var max int64
- for _, peer := range pool.peers {
- if peer.height > max {
- max = peer.height
- }
- }
- pool.maxPeerHeight = max
- }
-
- // Pick an available peer with the given height available.
- // If no peers are available, returns nil.
- func (pool *BlockPool) pickIncrAvailablePeer(height int64) *bpPeer {
- pool.mtx.Lock()
- defer pool.mtx.Unlock()
-
- for _, peer := range pool.peers {
- if peer.didTimeout {
- pool.removePeer(peer.id)
- continue
- }
- if peer.numPending >= maxPendingRequestsPerPeer {
- continue
- }
- if height < peer.base || height > peer.height {
- continue
- }
- peer.incrPending()
- return peer
- }
- return nil
- }
-
- func (pool *BlockPool) makeNextRequester() {
- pool.mtx.Lock()
- defer pool.mtx.Unlock()
-
- nextHeight := pool.height + pool.requestersLen()
- if nextHeight > pool.maxPeerHeight {
- return
- }
-
- request := newBPRequester(pool, nextHeight)
-
- pool.requesters[nextHeight] = request
- atomic.AddInt32(&pool.numPending, 1)
-
- err := request.Start()
- if err != nil {
- request.Logger.Error("Error starting request", "err", err)
- }
- }
-
- func (pool *BlockPool) requestersLen() int64 {
- return int64(len(pool.requesters))
- }
-
- func (pool *BlockPool) sendRequest(height int64, peerID p2p.ID) {
- if !pool.IsRunning() {
- return
- }
- pool.requestsCh <- BlockRequest{height, peerID}
- }
-
- func (pool *BlockPool) sendError(err error, peerID p2p.ID) {
- if !pool.IsRunning() {
- return
- }
- pool.errorsCh <- peerError{err, peerID}
- }
-
- // for debugging purposes
- //nolint:unused
- func (pool *BlockPool) debug() string {
- pool.mtx.Lock()
- defer pool.mtx.Unlock()
-
- str := ""
- nextHeight := pool.height + pool.requestersLen()
- for h := pool.height; h < nextHeight; h++ {
- if pool.requesters[h] == nil {
- str += fmt.Sprintf("H(%v):X ", h)
- } else {
- str += fmt.Sprintf("H(%v):", h)
- str += fmt.Sprintf("B?(%v) ", pool.requesters[h].block != nil)
- }
- }
- return str
- }
-
- //-------------------------------------
-
- type bpPeer struct {
- didTimeout bool
- numPending int32
- height int64
- base int64
- pool *BlockPool
- id p2p.ID
- recvMonitor *flow.Monitor
-
- timeout *time.Timer
-
- logger log.Logger
- }
-
- func newBPPeer(pool *BlockPool, peerID p2p.ID, base int64, height int64) *bpPeer {
- peer := &bpPeer{
- pool: pool,
- id: peerID,
- base: base,
- height: height,
- numPending: 0,
- logger: log.NewNopLogger(),
- }
- return peer
- }
-
- func (peer *bpPeer) setLogger(l log.Logger) {
- peer.logger = l
- }
-
- func (peer *bpPeer) resetMonitor() {
- peer.recvMonitor = flow.New(time.Second, time.Second*40)
- initialValue := float64(minRecvRate) * math.E
- peer.recvMonitor.SetREMA(initialValue)
- }
-
- func (peer *bpPeer) resetTimeout() {
- if peer.timeout == nil {
- peer.timeout = time.AfterFunc(peerTimeout, peer.onTimeout)
- } else {
- peer.timeout.Reset(peerTimeout)
- }
- }
-
- func (peer *bpPeer) incrPending() {
- if peer.numPending == 0 {
- peer.resetMonitor()
- peer.resetTimeout()
- }
- peer.numPending++
- }
-
- func (peer *bpPeer) decrPending(recvSize int) {
- peer.numPending--
- if peer.numPending == 0 {
- peer.timeout.Stop()
- } else {
- peer.recvMonitor.Update(recvSize)
- peer.resetTimeout()
- }
- }
-
- func (peer *bpPeer) onTimeout() {
- peer.pool.mtx.Lock()
- defer peer.pool.mtx.Unlock()
-
- err := errors.New("peer did not send us anything")
- peer.pool.sendError(err, peer.id)
- peer.logger.Error("SendTimeout", "reason", err, "timeout", peerTimeout)
- peer.didTimeout = true
- }
-
- //-------------------------------------
-
- type bpRequester struct {
- service.BaseService
- pool *BlockPool
- height int64
- gotBlockCh chan struct{}
- redoCh chan p2p.ID // redo may send multitime, add peerId to identify repeat
-
- mtx tmsync.Mutex
- peerID p2p.ID
- block *types.Block
- }
-
- func newBPRequester(pool *BlockPool, height int64) *bpRequester {
- bpr := &bpRequester{
- pool: pool,
- height: height,
- gotBlockCh: make(chan struct{}, 1),
- redoCh: make(chan p2p.ID, 1),
-
- peerID: "",
- block: nil,
- }
- bpr.BaseService = *service.NewBaseService(nil, "bpRequester", bpr)
- return bpr
- }
-
- func (bpr *bpRequester) OnStart() error {
- go bpr.requestRoutine()
- return nil
- }
-
- // Returns true if the peer matches and block doesn't already exist.
- func (bpr *bpRequester) setBlock(block *types.Block, peerID p2p.ID) bool {
- bpr.mtx.Lock()
- if bpr.block != nil || bpr.peerID != peerID {
- bpr.mtx.Unlock()
- return false
- }
- bpr.block = block
- bpr.mtx.Unlock()
-
- select {
- case bpr.gotBlockCh <- struct{}{}:
- default:
- }
- return true
- }
-
- func (bpr *bpRequester) getBlock() *types.Block {
- bpr.mtx.Lock()
- defer bpr.mtx.Unlock()
- return bpr.block
- }
-
- func (bpr *bpRequester) getPeerID() p2p.ID {
- bpr.mtx.Lock()
- defer bpr.mtx.Unlock()
- return bpr.peerID
- }
-
- // This is called from the requestRoutine, upon redo().
- func (bpr *bpRequester) reset() {
- bpr.mtx.Lock()
- defer bpr.mtx.Unlock()
-
- if bpr.block != nil {
- atomic.AddInt32(&bpr.pool.numPending, 1)
- }
-
- bpr.peerID = ""
- bpr.block = nil
- }
-
- // Tells bpRequester to pick another peer and try again.
- // NOTE: Nonblocking, and does nothing if another redo
- // was already requested.
- func (bpr *bpRequester) redo(peerID p2p.ID) {
- select {
- case bpr.redoCh <- peerID:
- default:
- }
- }
-
- // Responsible for making more requests as necessary
- // Returns only when a block is found (e.g. AddBlock() is called)
- func (bpr *bpRequester) requestRoutine() {
- OUTER_LOOP:
- for {
- // Pick a peer to send request to.
- var peer *bpPeer
- PICK_PEER_LOOP:
- for {
- if !bpr.IsRunning() || !bpr.pool.IsRunning() {
- return
- }
- peer = bpr.pool.pickIncrAvailablePeer(bpr.height)
- if peer == nil {
- time.Sleep(requestIntervalMS * time.Millisecond)
- continue PICK_PEER_LOOP
- }
- break PICK_PEER_LOOP
- }
- bpr.mtx.Lock()
- bpr.peerID = peer.id
- bpr.mtx.Unlock()
-
- // Send request and wait.
- bpr.pool.sendRequest(bpr.height, peer.id)
- WAIT_LOOP:
- for {
- select {
- case <-bpr.pool.Quit():
- if err := bpr.Stop(); err != nil {
- bpr.Logger.Error("Error stopped requester", "err", err)
- }
- return
- case <-bpr.Quit():
- return
- case peerID := <-bpr.redoCh:
- if peerID == bpr.peerID {
- bpr.reset()
- continue OUTER_LOOP
- } else {
- continue WAIT_LOOP
- }
- case <-bpr.gotBlockCh:
- // We got a block!
- // Continue the for-loop and wait til Quit.
- continue WAIT_LOOP
- }
- }
- }
- }
|