mirror of
https://gitlab.com/pulsechaincom/prysm-pulse.git
synced 2024-12-25 12:57:18 +00:00
a069738c20
* update shared/params * update eth2-types deps * update protobufs * update shared/* * fix testutil/state * update beacon-chain/state * update beacon-chain/db * update tests * fix test * update beacon-chain/core * update beacon-chain/blockchain * update beacon-chain/cache * beacon-chain/forkchoice * update beacon-chain/operations * update beacon-chain/p2p * update beacon-chain/rpc * update sync/initial-sync * update deps * update deps * go fmt * update beacon-chain/sync * update endtoend/ * bazel build //beacon-chain - works w/o issues * update slasher code * udpate tools/ * update validator/ * update fastssz * fix build * fix test building * update tests * update ethereumapis deps * fix tests * update state/stategen * fix build * fix test * add FarFutureSlot * go imports * Radek's suggestions * Ivan's suggestions * type conversions * Nishant's suggestions * add more tests to rpc_send_request * fix test * clean up * fix conflicts Co-authored-by: prylabs-bulldozer[bot] <58059840+prylabs-bulldozer[bot]@users.noreply.github.com> Co-authored-by: nisdas <nishdas93@gmail.com>
385 lines
11 KiB
Go
385 lines
11 KiB
Go
package sync
|
|
|
|
import (
|
|
"context"
|
|
"encoding/hex"
|
|
"sort"
|
|
"sync"
|
|
"time"
|
|
|
|
"github.com/pkg/errors"
|
|
types "github.com/prysmaticlabs/eth2-types"
|
|
ethpb "github.com/prysmaticlabs/ethereumapis/eth/v1alpha1"
|
|
"github.com/prysmaticlabs/prysm/beacon-chain/core/helpers"
|
|
p2ptypes "github.com/prysmaticlabs/prysm/beacon-chain/p2p/types"
|
|
"github.com/prysmaticlabs/prysm/shared/bytesutil"
|
|
"github.com/prysmaticlabs/prysm/shared/params"
|
|
"github.com/prysmaticlabs/prysm/shared/rand"
|
|
"github.com/prysmaticlabs/prysm/shared/runutil"
|
|
"github.com/prysmaticlabs/prysm/shared/slotutil"
|
|
"github.com/prysmaticlabs/prysm/shared/sszutil"
|
|
"github.com/prysmaticlabs/prysm/shared/traceutil"
|
|
"github.com/sirupsen/logrus"
|
|
"github.com/trailofbits/go-mutexasserts"
|
|
"go.opencensus.io/trace"
|
|
)
|
|
|
|
var processPendingBlocksPeriod = slotutil.DivideSlotBy(3 /* times per slot */)
|
|
|
|
const maxPeerRequest = 50
|
|
const numOfTries = 5
|
|
const maxBlocksPerSlot = 3
|
|
|
|
// processes pending blocks queue on every processPendingBlocksPeriod
|
|
func (s *Service) processPendingBlocksQueue() {
|
|
// Prevents multiple queue processing goroutines (invoked by RunEvery) from contending for data.
|
|
locker := new(sync.Mutex)
|
|
runutil.RunEvery(s.ctx, processPendingBlocksPeriod, func() {
|
|
locker.Lock()
|
|
if err := s.processPendingBlocks(s.ctx); err != nil {
|
|
log.WithError(err).Debug("Could not process pending blocks")
|
|
}
|
|
locker.Unlock()
|
|
})
|
|
}
|
|
|
|
// processes the block tree inside the queue
|
|
func (s *Service) processPendingBlocks(ctx context.Context) error {
|
|
ctx, span := trace.StartSpan(ctx, "processPendingBlocks")
|
|
defer span.End()
|
|
|
|
pids := s.p2p.Peers().Connected()
|
|
if err := s.validatePendingSlots(); err != nil {
|
|
return errors.Wrap(err, "could not validate pending slots")
|
|
}
|
|
slots := s.sortedPendingSlots()
|
|
var parentRoots [][32]byte
|
|
|
|
span.AddAttributes(
|
|
trace.Int64Attribute("numSlots", int64(len(slots))),
|
|
trace.Int64Attribute("numPeers", int64(len(pids))),
|
|
)
|
|
|
|
randGen := rand.NewGenerator()
|
|
for _, slot := range slots {
|
|
ctx, span := trace.StartSpan(ctx, "processPendingBlocks.InnerLoop")
|
|
span.AddAttributes(trace.Int64Attribute("slot", int64(slot)))
|
|
|
|
s.pendingQueueLock.RLock()
|
|
bs := s.pendingBlocksInCache(slot)
|
|
// Skip if there's no block in the queue.
|
|
if len(bs) == 0 {
|
|
s.pendingQueueLock.RUnlock()
|
|
span.End()
|
|
continue
|
|
}
|
|
s.pendingQueueLock.RUnlock()
|
|
|
|
// Loop through the pending queue and mark the potential parent blocks as seen.
|
|
for _, b := range bs {
|
|
if b == nil || b.Block == nil {
|
|
span.End()
|
|
continue
|
|
}
|
|
|
|
s.pendingQueueLock.RLock()
|
|
inPendingQueue := s.seenPendingBlocks[bytesutil.ToBytes32(b.Block.ParentRoot)]
|
|
s.pendingQueueLock.RUnlock()
|
|
|
|
blkRoot, err := b.Block.HashTreeRoot()
|
|
if err != nil {
|
|
traceutil.AnnotateError(span, err)
|
|
span.End()
|
|
return err
|
|
}
|
|
parentIsBad := s.hasBadBlock(bytesutil.ToBytes32(b.Block.ParentRoot))
|
|
blockIsBad := s.hasBadBlock(blkRoot)
|
|
// Check if parent is a bad block.
|
|
if parentIsBad || blockIsBad {
|
|
// Set block as bad if its parent block is bad too.
|
|
if parentIsBad {
|
|
s.setBadBlock(ctx, blkRoot)
|
|
}
|
|
// Remove block from queue.
|
|
s.pendingQueueLock.Lock()
|
|
if err := s.deleteBlockFromPendingQueue(slot, b, blkRoot); err != nil {
|
|
s.pendingQueueLock.Unlock()
|
|
return err
|
|
}
|
|
s.pendingQueueLock.Unlock()
|
|
span.End()
|
|
continue
|
|
}
|
|
|
|
inDB := s.db.HasBlock(ctx, bytesutil.ToBytes32(b.Block.ParentRoot))
|
|
hasPeer := len(pids) != 0
|
|
|
|
// Only request for missing parent block if it's not in DB, not in pending cache
|
|
// and has peer in the peer list.
|
|
if !inPendingQueue && !inDB && hasPeer {
|
|
log.WithFields(logrus.Fields{
|
|
"currentSlot": b.Block.Slot,
|
|
"parentRoot": hex.EncodeToString(bytesutil.Trunc(b.Block.ParentRoot)),
|
|
}).Debug("Requesting parent block")
|
|
parentRoots = append(parentRoots, bytesutil.ToBytes32(b.Block.ParentRoot))
|
|
|
|
span.End()
|
|
continue
|
|
}
|
|
|
|
if !inDB {
|
|
span.End()
|
|
continue
|
|
}
|
|
|
|
if err := s.validateBeaconBlock(ctx, b, blkRoot); err != nil {
|
|
log.Debugf("Could not validate block from slot %d: %v", b.Block.Slot, err)
|
|
s.setBadBlock(ctx, blkRoot)
|
|
traceutil.AnnotateError(span, err)
|
|
// In the next iteration of the queue, this block will be removed from
|
|
// the pending queue as it has been marked as a 'bad' block.
|
|
span.End()
|
|
continue
|
|
}
|
|
|
|
if err := s.chain.ReceiveBlock(ctx, b, blkRoot); err != nil {
|
|
log.Debugf("Could not process block from slot %d: %v", b.Block.Slot, err)
|
|
s.setBadBlock(ctx, blkRoot)
|
|
traceutil.AnnotateError(span, err)
|
|
// In the next iteration of the queue, this block will be removed from
|
|
// the pending queue as it has been marked as a 'bad' block.
|
|
span.End()
|
|
continue
|
|
}
|
|
|
|
s.setSeenBlockIndexSlot(b.Block.Slot, b.Block.ProposerIndex)
|
|
|
|
// Broadcasting the block again once a node is able to process it.
|
|
if err := s.p2p.Broadcast(ctx, b); err != nil {
|
|
log.WithError(err).Debug("Could not broadcast block")
|
|
}
|
|
|
|
s.pendingQueueLock.Lock()
|
|
if err := s.deleteBlockFromPendingQueue(slot, b, blkRoot); err != nil {
|
|
return err
|
|
}
|
|
s.pendingQueueLock.Unlock()
|
|
|
|
log.WithFields(logrus.Fields{
|
|
"slot": slot,
|
|
"blockRoot": hex.EncodeToString(bytesutil.Trunc(blkRoot[:])),
|
|
}).Debug("Processed pending block and cleared it in cache")
|
|
|
|
span.End()
|
|
}
|
|
}
|
|
|
|
return s.sendBatchRootRequest(ctx, parentRoots, randGen)
|
|
}
|
|
|
|
func (s *Service) sendBatchRootRequest(ctx context.Context, roots [][32]byte, randGen *rand.Rand) error {
|
|
ctx, span := trace.StartSpan(ctx, "sendBatchRootRequest")
|
|
defer span.End()
|
|
|
|
if len(roots) == 0 {
|
|
return nil
|
|
}
|
|
|
|
_, bestPeers := s.p2p.Peers().BestFinalized(maxPeerRequest, s.chain.FinalizedCheckpt().Epoch)
|
|
if len(bestPeers) == 0 {
|
|
return nil
|
|
}
|
|
roots = s.dedupRoots(roots)
|
|
// Randomly choose a peer to query from our best peers. If that peer cannot return
|
|
// all the requested blocks, we randomly select another peer.
|
|
pid := bestPeers[randGen.Int()%len(bestPeers)]
|
|
for i := 0; i < numOfTries; i++ {
|
|
req := p2ptypes.BeaconBlockByRootsReq(roots)
|
|
if len(roots) > int(params.BeaconNetworkConfig().MaxRequestBlocks) {
|
|
req = roots[:params.BeaconNetworkConfig().MaxRequestBlocks]
|
|
}
|
|
if err := s.sendRecentBeaconBlocksRequest(ctx, &req, pid); err != nil {
|
|
traceutil.AnnotateError(span, err)
|
|
log.Debugf("Could not send recent block request: %v", err)
|
|
}
|
|
newRoots := make([][32]byte, 0, len(roots))
|
|
s.pendingQueueLock.RLock()
|
|
for _, rt := range roots {
|
|
if !s.seenPendingBlocks[rt] {
|
|
newRoots = append(newRoots, rt)
|
|
}
|
|
}
|
|
s.pendingQueueLock.RUnlock()
|
|
if len(newRoots) == 0 {
|
|
break
|
|
}
|
|
// Choosing a new peer with the leftover set of
|
|
// roots to request.
|
|
roots = newRoots
|
|
pid = bestPeers[randGen.Int()%len(bestPeers)]
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (s *Service) sortedPendingSlots() []types.Slot {
|
|
s.pendingQueueLock.RLock()
|
|
defer s.pendingQueueLock.RUnlock()
|
|
|
|
items := s.slotToPendingBlocks.Items()
|
|
|
|
slots := make([]types.Slot, 0, len(items))
|
|
for k := range items {
|
|
slot := cacheKeyToSlot(k)
|
|
slots = append(slots, slot)
|
|
}
|
|
sort.Slice(slots, func(i, j int) bool {
|
|
return slots[i] < slots[j]
|
|
})
|
|
return slots
|
|
}
|
|
|
|
// validatePendingSlots validates the pending blocks
|
|
// by their slot. If they are before the current finalized
|
|
// checkpoint, these blocks are removed from the queue.
|
|
func (s *Service) validatePendingSlots() error {
|
|
s.pendingQueueLock.Lock()
|
|
defer s.pendingQueueLock.Unlock()
|
|
oldBlockRoots := make(map[[32]byte]bool)
|
|
|
|
finalizedEpoch := s.chain.FinalizedCheckpt().Epoch
|
|
if s.slotToPendingBlocks == nil {
|
|
return errors.New("slotToPendingBlocks cache can't be nil")
|
|
}
|
|
items := s.slotToPendingBlocks.Items()
|
|
for k := range items {
|
|
slot := cacheKeyToSlot(k)
|
|
blks := s.pendingBlocksInCache(slot)
|
|
for _, b := range blks {
|
|
epoch := helpers.SlotToEpoch(slot)
|
|
// remove all descendant blocks of old blocks
|
|
if oldBlockRoots[bytesutil.ToBytes32(b.Block.ParentRoot)] {
|
|
root, err := b.Block.HashTreeRoot()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
oldBlockRoots[root] = true
|
|
if err := s.deleteBlockFromPendingQueue(slot, b, root); err != nil {
|
|
return err
|
|
}
|
|
continue
|
|
}
|
|
// don't process old blocks
|
|
if finalizedEpoch > 0 && epoch <= finalizedEpoch {
|
|
blkRoot, err := b.Block.HashTreeRoot()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
oldBlockRoots[blkRoot] = true
|
|
if err := s.deleteBlockFromPendingQueue(slot, b, blkRoot); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
}
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (s *Service) clearPendingSlots() {
|
|
s.pendingQueueLock.Lock()
|
|
defer s.pendingQueueLock.Unlock()
|
|
s.slotToPendingBlocks.Flush()
|
|
s.seenPendingBlocks = make(map[[32]byte]bool)
|
|
}
|
|
|
|
// Delete block from the list from the pending queue using the slot as key.
|
|
// Note: this helper is not thread safe.
|
|
func (s *Service) deleteBlockFromPendingQueue(slot types.Slot, b *ethpb.SignedBeaconBlock, r [32]byte) error {
|
|
mutexasserts.AssertRWMutexLocked(&s.pendingQueueLock)
|
|
|
|
blks := s.pendingBlocksInCache(slot)
|
|
if len(blks) == 0 {
|
|
return nil
|
|
}
|
|
|
|
newBlks := make([]*ethpb.SignedBeaconBlock, 0, len(blks))
|
|
for _, blk := range blks {
|
|
if sszutil.DeepEqual(blk, b) {
|
|
continue
|
|
}
|
|
newBlks = append(newBlks, blk)
|
|
}
|
|
if len(newBlks) == 0 {
|
|
s.slotToPendingBlocks.Delete(slotToCacheKey(slot))
|
|
return nil
|
|
}
|
|
|
|
// Decrease exp time in proportion to how many blocks are still in the cache for slot key.
|
|
d := pendingBlockExpTime / time.Duration(len(newBlks))
|
|
if err := s.slotToPendingBlocks.Replace(slotToCacheKey(slot), newBlks, d); err != nil {
|
|
return err
|
|
}
|
|
delete(s.seenPendingBlocks, r)
|
|
return nil
|
|
}
|
|
|
|
// Insert block to the list in the pending queue using the slot as key.
|
|
// Note: this helper is not thread safe.
|
|
func (s *Service) insertBlockToPendingQueue(slot types.Slot, b *ethpb.SignedBeaconBlock, r [32]byte) error {
|
|
mutexasserts.AssertRWMutexLocked(&s.pendingQueueLock)
|
|
|
|
if s.seenPendingBlocks[r] {
|
|
return nil
|
|
}
|
|
|
|
if err := s.addPendingBlockToCache(b); err != nil {
|
|
return err
|
|
}
|
|
|
|
s.seenPendingBlocks[r] = true
|
|
return nil
|
|
}
|
|
|
|
// This returns signed beacon blocks given input key from slotToPendingBlocks.
|
|
func (s *Service) pendingBlocksInCache(slot types.Slot) []*ethpb.SignedBeaconBlock {
|
|
k := slotToCacheKey(slot)
|
|
value, ok := s.slotToPendingBlocks.Get(k)
|
|
if !ok {
|
|
return []*ethpb.SignedBeaconBlock{}
|
|
}
|
|
blks, ok := value.([]*ethpb.SignedBeaconBlock)
|
|
if !ok {
|
|
return []*ethpb.SignedBeaconBlock{}
|
|
}
|
|
return blks
|
|
}
|
|
|
|
// This adds input signed beacon block to slotToPendingBlocks cache.
|
|
func (s *Service) addPendingBlockToCache(b *ethpb.SignedBeaconBlock) error {
|
|
if err := helpers.VerifyNilBeaconBlock(b); err != nil {
|
|
return err
|
|
}
|
|
|
|
blks := s.pendingBlocksInCache(b.Block.Slot)
|
|
|
|
if len(blks) >= maxBlocksPerSlot {
|
|
return nil
|
|
}
|
|
|
|
blks = append(blks, b)
|
|
k := slotToCacheKey(b.Block.Slot)
|
|
s.slotToPendingBlocks.Set(k, blks, pendingBlockExpTime)
|
|
return nil
|
|
}
|
|
|
|
// This converts input string to slot.
|
|
func cacheKeyToSlot(s string) types.Slot {
|
|
b := []byte(s)
|
|
return bytesutil.BytesToSlotBigEndian(b)
|
|
}
|
|
|
|
// This converts input slot to a key to be used for slotToPendingBlocks cache.
|
|
func slotToCacheKey(s types.Slot) string {
|
|
b := bytesutil.SlotToBytesBigEndian(s)
|
|
return string(b)
|
|
}
|