package blockchain import ( "context" "fmt" "time" "github.com/pkg/errors" "github.com/prysmaticlabs/prysm/v4/async/event" "github.com/prysmaticlabs/prysm/v4/beacon-chain/core/blocks" "github.com/prysmaticlabs/prysm/v4/beacon-chain/core/feed" statefeed "github.com/prysmaticlabs/prysm/v4/beacon-chain/core/feed/state" "github.com/prysmaticlabs/prysm/v4/beacon-chain/core/helpers" coreTime "github.com/prysmaticlabs/prysm/v4/beacon-chain/core/time" "github.com/prysmaticlabs/prysm/v4/beacon-chain/core/transition" forkchoicetypes "github.com/prysmaticlabs/prysm/v4/beacon-chain/forkchoice/types" "github.com/prysmaticlabs/prysm/v4/beacon-chain/state" "github.com/prysmaticlabs/prysm/v4/config/features" "github.com/prysmaticlabs/prysm/v4/config/params" consensusblocks "github.com/prysmaticlabs/prysm/v4/consensus-types/blocks" "github.com/prysmaticlabs/prysm/v4/consensus-types/interfaces" "github.com/prysmaticlabs/prysm/v4/consensus-types/primitives" "github.com/prysmaticlabs/prysm/v4/crypto/bls" "github.com/prysmaticlabs/prysm/v4/encoding/bytesutil" "github.com/prysmaticlabs/prysm/v4/monitoring/tracing" ethpbv1 "github.com/prysmaticlabs/prysm/v4/proto/eth/v1" ethpb "github.com/prysmaticlabs/prysm/v4/proto/prysm/v1alpha1" "github.com/prysmaticlabs/prysm/v4/proto/prysm/v1alpha1/attestation" "github.com/prysmaticlabs/prysm/v4/runtime/version" "github.com/prysmaticlabs/prysm/v4/time/slots" "github.com/sirupsen/logrus" "go.opencensus.io/trace" ) // A custom slot deadline for processing state slots in our cache. const slotDeadline = 5 * time.Second // A custom deadline for deposit trie insertion. const depositDeadline = 20 * time.Second // This defines size of the upper bound for initial sync block cache. var initialSyncBlockCacheSize = uint64(2 * params.BeaconConfig().SlotsPerEpoch) // onBlock is called when a gossip block is received. It runs regular state transition on the block. // The block's signing root should be computed before calling this method to avoid redundant // computation in this method and methods it calls into. // // Spec pseudocode definition: // // def on_block(store: Store, signed_block: ReadOnlySignedBeaconBlock) -> None: // block = signed_block.message // # Parent block must be known // assert block.parent_root in store.block_states // # Make a copy of the state to avoid mutability issues // pre_state = copy(store.block_states[block.parent_root]) // # Blocks cannot be in the future. If they are, their consideration must be delayed until the are in the past. // assert get_current_slot(store) >= block.slot // // # Check that block is later than the finalized epoch slot (optimization to reduce calls to get_ancestor) // finalized_slot = compute_start_slot_at_epoch(store.finalized_checkpoint.epoch) // assert block.slot > finalized_slot // # Check block is a descendant of the finalized block at the checkpoint finalized slot // assert get_ancestor(store, block.parent_root, finalized_slot) == store.finalized_checkpoint.root // // # Check the block is valid and compute the post-state // state = pre_state.copy() // state_transition(state, signed_block, True) // # Add new block to the store // store.blocks[hash_tree_root(block)] = block // # Add new state for this block to the store // store.block_states[hash_tree_root(block)] = state // // # Update justified checkpoint // if state.current_justified_checkpoint.epoch > store.justified_checkpoint.epoch: // if state.current_justified_checkpoint.epoch > store.best_justified_checkpoint.epoch: // store.best_justified_checkpoint = state.current_justified_checkpoint // if should_update_justified_checkpoint(store, state.current_justified_checkpoint): // store.justified_checkpoint = state.current_justified_checkpoint // // # Update finalized checkpoint // if state.finalized_checkpoint.epoch > store.finalized_checkpoint.epoch: // store.finalized_checkpoint = state.finalized_checkpoint // // # Potentially update justified if different from store // if store.justified_checkpoint != state.current_justified_checkpoint: // # Update justified if new justified is later than store justified // if state.current_justified_checkpoint.epoch > store.justified_checkpoint.epoch: // store.justified_checkpoint = state.current_justified_checkpoint // return // // # Update justified if store justified is not in chain with finalized checkpoint // finalized_slot = compute_start_slot_at_epoch(store.finalized_checkpoint.epoch) // ancestor_at_finalized_slot = get_ancestor(store, store.justified_checkpoint.root, finalized_slot) // if ancestor_at_finalized_slot != store.finalized_checkpoint.root: // store.justified_checkpoint = state.current_justified_checkpoint func (s *Service) onBlock(ctx context.Context, signed interfaces.ReadOnlySignedBeaconBlock, blockRoot [32]byte) error { ctx, span := trace.StartSpan(ctx, "blockChain.onBlock") defer span.End() if err := consensusblocks.BeaconBlockIsNil(signed); err != nil { return invalidBlock{error: err} } startTime := time.Now() b := signed.Block() preState, err := s.getBlockPreState(ctx, b) if err != nil { return err } // Verify that the parent block is in forkchoice if !s.cfg.ForkChoiceStore.HasNode(b.ParentRoot()) { return ErrNotDescendantOfFinalized } // Save current justified and finalized epochs for future use. currStoreJustifiedEpoch := s.cfg.ForkChoiceStore.JustifiedCheckpoint().Epoch currStoreFinalizedEpoch := s.cfg.ForkChoiceStore.FinalizedCheckpoint().Epoch preStateFinalizedEpoch := preState.FinalizedCheckpoint().Epoch preStateJustifiedEpoch := preState.CurrentJustifiedCheckpoint().Epoch preStateVersion, preStateHeader, err := getStateVersionAndPayload(preState) if err != nil { return err } stateTransitionStartTime := time.Now() postState, err := transition.ExecuteStateTransition(ctx, preState, signed) if err != nil { return invalidBlock{error: err} } stateTransitionProcessingTime.Observe(float64(time.Since(stateTransitionStartTime).Milliseconds())) postStateVersion, postStateHeader, err := getStateVersionAndPayload(postState) if err != nil { return err } isValidPayload, err := s.notifyNewPayload(ctx, postStateVersion, postStateHeader, signed) if err != nil { return errors.Wrap(err, "could not validate new payload") } if isValidPayload { if err := s.validateMergeTransitionBlock(ctx, preStateVersion, preStateHeader, signed); err != nil { return err } } if err := s.savePostStateInfo(ctx, blockRoot, signed, postState); err != nil { return err } if err := s.insertBlockToForkchoiceStore(ctx, signed.Block(), blockRoot, postState); err != nil { return errors.Wrapf(err, "could not insert block %d to fork choice store", signed.Block().Slot()) } if err := s.handleBlockAttestations(ctx, signed.Block(), postState); err != nil { return errors.Wrap(err, "could not handle block's attestations") } s.InsertSlashingsToForkChoiceStore(ctx, signed.Block().Body().AttesterSlashings()) if isValidPayload { if err := s.cfg.ForkChoiceStore.SetOptimisticToValid(ctx, blockRoot); err != nil { return errors.Wrap(err, "could not set optimistic block to valid") } } // If slasher is configured, forward the attestations in the block via // an event feed for processing. if features.Get().EnableSlasher { // Feed the indexed attestation to slasher if enabled. This action // is done in the background to avoid adding more load to this critical code path. go func() { // Using a different context to prevent timeouts as this operation can be expensive // and we want to avoid affecting the critical code path. ctx := context.TODO() for _, att := range signed.Block().Body().Attestations() { committee, err := helpers.BeaconCommitteeFromState(ctx, preState, att.Data.Slot, att.Data.CommitteeIndex) if err != nil { log.WithError(err).Error("Could not get attestation committee") tracing.AnnotateError(span, err) return } indexedAtt, err := attestation.ConvertToIndexed(ctx, att, committee) if err != nil { log.WithError(err).Error("Could not convert to indexed attestation") tracing.AnnotateError(span, err) return } s.cfg.SlasherAttestationsFeed.Send(indexedAtt) } }() } justified := s.cfg.ForkChoiceStore.JustifiedCheckpoint() start := time.Now() headRoot, err := s.cfg.ForkChoiceStore.Head(ctx) if err != nil { log.WithError(err).Warn("Could not update head") } if blockRoot != headRoot { receivedWeight, err := s.cfg.ForkChoiceStore.Weight(blockRoot) if err != nil { log.WithField("root", fmt.Sprintf("%#x", blockRoot)).Warn("could not determine node weight") } headWeight, err := s.cfg.ForkChoiceStore.Weight(headRoot) if err != nil { log.WithField("root", fmt.Sprintf("%#x", headRoot)).Warn("could not determine node weight") } log.WithFields(logrus.Fields{ "receivedRoot": fmt.Sprintf("%#x", blockRoot), "receivedWeight": receivedWeight, "headRoot": fmt.Sprintf("%#x", headRoot), "headWeight": headWeight, }).Debug("Head block is not the received block") } newBlockHeadElapsedTime.Observe(float64(time.Since(start).Milliseconds())) // verify conditions for FCU, notifies FCU, and saves the new head. // This function also prunes attestations, other similar operations happen in prunePostBlockOperationPools. if err := s.forkchoiceUpdateWithExecution(ctx, headRoot, s.CurrentSlot()+1); err != nil { return err } // Send notification of the processed block to the state feed. s.cfg.StateNotifier.StateFeed().Send(&feed.Event{ Type: statefeed.BlockProcessed, Data: &statefeed.BlockProcessedData{ Slot: signed.Block().Slot(), BlockRoot: blockRoot, SignedBlock: signed, Verified: true, }, }) // Updating next slot state cache can happen in the background. It shouldn't block rest of the process. go func() { // Use a custom deadline here, since this method runs asynchronously. // We ignore the parent method's context and instead create a new one // with a custom deadline, therefore using the background context instead. slotCtx, cancel := context.WithTimeout(context.Background(), slotDeadline) defer cancel() if err := transition.UpdateNextSlotCache(slotCtx, blockRoot[:], postState); err != nil { log.WithError(err).Debug("could not update next slot state cache") } }() // Save justified check point to db. postStateJustifiedEpoch := postState.CurrentJustifiedCheckpoint().Epoch if justified.Epoch > currStoreJustifiedEpoch || (justified.Epoch == postStateJustifiedEpoch && justified.Epoch > preStateJustifiedEpoch) { if err := s.cfg.BeaconDB.SaveJustifiedCheckpoint(ctx, ðpb.Checkpoint{ Epoch: justified.Epoch, Root: justified.Root[:], }); err != nil { return err } } // Save finalized check point to db and more. postStateFinalizedEpoch := postState.FinalizedCheckpoint().Epoch finalized := s.cfg.ForkChoiceStore.FinalizedCheckpoint() if finalized.Epoch > currStoreFinalizedEpoch || (finalized.Epoch == postStateFinalizedEpoch && finalized.Epoch > preStateFinalizedEpoch) { if err := s.updateFinalized(ctx, ðpb.Checkpoint{Epoch: finalized.Epoch, Root: finalized.Root[:]}); err != nil { return err } isOptimistic, err := s.cfg.ForkChoiceStore.IsOptimistic(finalized.Root) if err != nil { return errors.Wrap(err, "could not check if node is optimistically synced") } go func() { // Send an event regarding the new finalized checkpoint over a common event feed. stateRoot := signed.Block().StateRoot() s.cfg.StateNotifier.StateFeed().Send(&feed.Event{ Type: statefeed.FinalizedCheckpoint, Data: ðpbv1.EventFinalizedCheckpoint{ Epoch: postState.FinalizedCheckpoint().Epoch, Block: postState.FinalizedCheckpoint().Root, State: stateRoot[:], ExecutionOptimistic: isOptimistic, }, }) // Use a custom deadline here, since this method runs asynchronously. // We ignore the parent method's context and instead create a new one // with a custom deadline, therefore using the background context instead. depCtx, cancel := context.WithTimeout(context.Background(), depositDeadline) defer cancel() if err := s.insertFinalizedDeposits(depCtx, finalized.Root); err != nil { log.WithError(err).Error("Could not insert finalized deposits.") } }() } defer reportAttestationInclusion(b) if err := s.handleEpochBoundary(ctx, postState); err != nil { return err } onBlockProcessingTime.Observe(float64(time.Since(startTime).Milliseconds())) return nil } func getStateVersionAndPayload(st state.BeaconState) (int, interfaces.ExecutionData, error) { if st == nil { return 0, nil, errors.New("nil state") } var preStateHeader interfaces.ExecutionData var err error preStateVersion := st.Version() switch preStateVersion { case version.Phase0, version.Altair: default: preStateHeader, err = st.LatestExecutionPayloadHeader() if err != nil { return 0, nil, err } } return preStateVersion, preStateHeader, nil } func (s *Service) onBlockBatch(ctx context.Context, blks []interfaces.ReadOnlySignedBeaconBlock, blockRoots [][32]byte) error { ctx, span := trace.StartSpan(ctx, "blockChain.onBlockBatch") defer span.End() if len(blks) == 0 || len(blockRoots) == 0 { return errors.New("no blocks provided") } if len(blks) != len(blockRoots) { return errWrongBlockCount } if err := consensusblocks.BeaconBlockIsNil(blks[0]); err != nil { return invalidBlock{error: err} } b := blks[0].Block() // Retrieve incoming block's pre state. if err := s.verifyBlkPreState(ctx, b); err != nil { return err } preState, err := s.cfg.StateGen.StateByRootInitialSync(ctx, b.ParentRoot()) if err != nil { return err } if preState == nil || preState.IsNil() { return fmt.Errorf("nil pre state for slot %d", b.Slot()) } // Fill in missing blocks if err := s.fillInForkChoiceMissingBlocks(ctx, blks[0].Block(), preState.CurrentJustifiedCheckpoint(), preState.FinalizedCheckpoint()); err != nil { return errors.Wrap(err, "could not fill in missing blocks to forkchoice") } jCheckpoints := make([]*ethpb.Checkpoint, len(blks)) fCheckpoints := make([]*ethpb.Checkpoint, len(blks)) sigSet := bls.NewSet() type versionAndHeader struct { version int header interfaces.ExecutionData } preVersionAndHeaders := make([]*versionAndHeader, len(blks)) postVersionAndHeaders := make([]*versionAndHeader, len(blks)) var set *bls.SignatureBatch boundaries := make(map[[32]byte]state.BeaconState) for i, b := range blks { v, h, err := getStateVersionAndPayload(preState) if err != nil { return err } preVersionAndHeaders[i] = &versionAndHeader{ version: v, header: h, } set, preState, err = transition.ExecuteStateTransitionNoVerifyAnySig(ctx, preState, b) if err != nil { return invalidBlock{error: err} } // Save potential boundary states. if slots.IsEpochStart(preState.Slot()) { boundaries[blockRoots[i]] = preState.Copy() } jCheckpoints[i] = preState.CurrentJustifiedCheckpoint() fCheckpoints[i] = preState.FinalizedCheckpoint() v, h, err = getStateVersionAndPayload(preState) if err != nil { return err } postVersionAndHeaders[i] = &versionAndHeader{ version: v, header: h, } sigSet.Join(set) } var verify bool if features.Get().EnableVerboseSigVerification { verify, err = sigSet.VerifyVerbosely() } else { verify, err = sigSet.Verify() } if err != nil { return invalidBlock{error: err} } if !verify { return errors.New("batch block signature verification failed") } // blocks have been verified, save them and call the engine pendingNodes := make([]*forkchoicetypes.BlockAndCheckpoints, len(blks)) var isValidPayload bool for i, b := range blks { isValidPayload, err = s.notifyNewPayload(ctx, postVersionAndHeaders[i].version, postVersionAndHeaders[i].header, b) if err != nil { return err } if isValidPayload { if err := s.validateMergeTransitionBlock(ctx, preVersionAndHeaders[i].version, preVersionAndHeaders[i].header, b); err != nil { return err } } args := &forkchoicetypes.BlockAndCheckpoints{Block: b.Block(), JustifiedCheckpoint: jCheckpoints[i], FinalizedCheckpoint: fCheckpoints[i]} pendingNodes[len(blks)-i-1] = args if err := s.saveInitSyncBlock(ctx, blockRoots[i], b); err != nil { tracing.AnnotateError(span, err) return err } if err := s.cfg.BeaconDB.SaveStateSummary(ctx, ðpb.StateSummary{ Slot: b.Block().Slot(), Root: blockRoots[i][:], }); err != nil { tracing.AnnotateError(span, err) return err } if i > 0 && jCheckpoints[i].Epoch > jCheckpoints[i-1].Epoch { if err := s.cfg.BeaconDB.SaveJustifiedCheckpoint(ctx, jCheckpoints[i]); err != nil { tracing.AnnotateError(span, err) return err } } if i > 0 && fCheckpoints[i].Epoch > fCheckpoints[i-1].Epoch { if err := s.updateFinalized(ctx, fCheckpoints[i]); err != nil { tracing.AnnotateError(span, err) return err } } } // Save boundary states that will be useful for forkchoice for r, st := range boundaries { if err := s.cfg.StateGen.SaveState(ctx, r, st); err != nil { return err } } // Also saves the last post state which to be used as pre state for the next batch. lastBR := blockRoots[len(blks)-1] if err := s.cfg.StateGen.SaveState(ctx, lastBR, preState); err != nil { return err } // Insert all nodes but the last one to forkchoice if err := s.cfg.ForkChoiceStore.InsertChain(ctx, pendingNodes); err != nil { return errors.Wrap(err, "could not insert batch to forkchoice") } // Insert the last block to forkchoice if err := s.cfg.ForkChoiceStore.InsertNode(ctx, preState, lastBR); err != nil { return errors.Wrap(err, "could not insert last block in batch to forkchoice") } // Set their optimistic status if isValidPayload { if err := s.cfg.ForkChoiceStore.SetOptimisticToValid(ctx, lastBR); err != nil { return errors.Wrap(err, "could not set optimistic block to valid") } } lastB := blks[len(blks)-1] arg := ¬ifyForkchoiceUpdateArg{ headState: preState, headRoot: lastBR, headBlock: lastB.Block(), } if _, err := s.notifyForkchoiceUpdate(ctx, arg); err != nil { return err } return s.saveHeadNoDB(ctx, lastB, lastBR, preState) } // Epoch boundary bookkeeping such as logging epoch summaries. func (s *Service) handleEpochBoundary(ctx context.Context, postState state.BeaconState) error { ctx, span := trace.StartSpan(ctx, "blockChain.handleEpochBoundary") defer span.End() var err error if postState.Slot()+1 == s.nextEpochBoundarySlot { copied := postState.Copy() copied, err := transition.ProcessSlots(ctx, copied, copied.Slot()+1) if err != nil { return err } // Update caches for the next epoch at epoch boundary slot - 1. if err := helpers.UpdateCommitteeCache(ctx, copied, coreTime.CurrentEpoch(copied)); err != nil { return err } if err := helpers.UpdateProposerIndicesInCache(ctx, copied); err != nil { return err } } else if postState.Slot() >= s.nextEpochBoundarySlot { s.nextEpochBoundarySlot, err = slots.EpochStart(coreTime.NextEpoch(postState)) if err != nil { return err } // Update caches at epoch boundary slot. // The following updates have shortcut to return nil cheaply if fulfilled during boundary slot - 1. if err := helpers.UpdateCommitteeCache(ctx, postState, coreTime.CurrentEpoch(postState)); err != nil { return err } if err := helpers.UpdateProposerIndicesInCache(ctx, postState); err != nil { return err } headSt, err := s.HeadState(ctx) if err != nil { return err } if err := reportEpochMetrics(ctx, postState, headSt); err != nil { return err } } return nil } // This feeds in the block to fork choice store. It's allows fork choice store // to gain information on the most current chain. func (s *Service) insertBlockToForkchoiceStore(ctx context.Context, blk interfaces.ReadOnlyBeaconBlock, root [32]byte, st state.BeaconState) error { ctx, span := trace.StartSpan(ctx, "blockChain.insertBlockToForkchoiceStore") defer span.End() if !s.cfg.ForkChoiceStore.HasNode(blk.ParentRoot()) { fCheckpoint := st.FinalizedCheckpoint() jCheckpoint := st.CurrentJustifiedCheckpoint() if err := s.fillInForkChoiceMissingBlocks(ctx, blk, fCheckpoint, jCheckpoint); err != nil { return err } } return s.cfg.ForkChoiceStore.InsertNode(ctx, st, root) } // This feeds in the attestations included in the block to fork choice store. It's allows fork choice store // to gain information on the most current chain. func (s *Service) handleBlockAttestations(ctx context.Context, blk interfaces.ReadOnlyBeaconBlock, st state.BeaconState) error { // Feed in block's attestations to fork choice store. for _, a := range blk.Body().Attestations() { committee, err := helpers.BeaconCommitteeFromState(ctx, st, a.Data.Slot, a.Data.CommitteeIndex) if err != nil { return err } indices, err := attestation.AttestingIndices(a.AggregationBits, committee) if err != nil { return err } r := bytesutil.ToBytes32(a.Data.BeaconBlockRoot) if s.cfg.ForkChoiceStore.HasNode(r) { s.cfg.ForkChoiceStore.ProcessAttestation(ctx, indices, r, a.Data.Target.Epoch) } else if err := s.cfg.AttPool.SaveBlockAttestation(a); err != nil { return err } } return nil } // InsertSlashingsToForkChoiceStore inserts attester slashing indices to fork choice store. // To call this function, it's caller's responsibility to ensure the slashing object is valid. // This function requires a write lock on forkchoice. func (s *Service) InsertSlashingsToForkChoiceStore(ctx context.Context, slashings []*ethpb.AttesterSlashing) { for _, slashing := range slashings { indices := blocks.SlashableAttesterIndices(slashing) for _, index := range indices { s.cfg.ForkChoiceStore.InsertSlashedIndex(ctx, primitives.ValidatorIndex(index)) } } } // This saves post state info to DB or cache. This also saves post state info to fork choice store. // Post state info consists of processed block and state. Do not call this method unless the block and state are verified. func (s *Service) savePostStateInfo(ctx context.Context, r [32]byte, b interfaces.ReadOnlySignedBeaconBlock, st state.BeaconState) error { ctx, span := trace.StartSpan(ctx, "blockChain.savePostStateInfo") defer span.End() if err := s.cfg.BeaconDB.SaveBlock(ctx, b); err != nil { return errors.Wrapf(err, "could not save block from slot %d", b.Block().Slot()) } if err := s.cfg.StateGen.SaveState(ctx, r, st); err != nil { return errors.Wrap(err, "could not save state") } return nil } // This removes the attestations in block `b` from the attestation mem pool. func (s *Service) pruneAttsFromPool(headBlock interfaces.ReadOnlySignedBeaconBlock) error { atts := headBlock.Block().Body().Attestations() for _, att := range atts { if helpers.IsAggregated(att) { if err := s.cfg.AttPool.DeleteAggregatedAttestation(att); err != nil { return err } } else { if err := s.cfg.AttPool.DeleteUnaggregatedAttestation(att); err != nil { return err } } } return nil } // validateMergeTransitionBlock validates the merge transition block. func (s *Service) validateMergeTransitionBlock(ctx context.Context, stateVersion int, stateHeader interfaces.ExecutionData, blk interfaces.ReadOnlySignedBeaconBlock) error { // Skip validation if block is older than Bellatrix. if blocks.IsPreBellatrixVersion(blk.Block().Version()) { return nil } // Skip validation if block has an empty payload. payload, err := blk.Block().Body().Execution() if err != nil { return invalidBlock{error: err} } isEmpty, err := consensusblocks.IsEmptyExecutionData(payload) if err != nil { return err } if isEmpty { return nil } // Handle case where pre-state is Altair but block contains payload. // To reach here, the block must have contained a valid payload. if blocks.IsPreBellatrixVersion(stateVersion) { return s.validateMergeBlock(ctx, blk) } // Skip validation if the block is not a merge transition block. // To reach here. The payload must be non-empty. If the state header is empty then it's at transition. empty, err := consensusblocks.IsEmptyExecutionData(stateHeader) if err != nil { return err } if !empty { return nil } return s.validateMergeBlock(ctx, blk) } // This routine checks if there is a cached proposer payload ID available for the next slot proposer. // If there is not, it will call forkchoice updated with the correct payload attribute then cache the payload ID. func (s *Service) fillMissingPayloadIDRoutine(ctx context.Context, stateFeed *event.Feed) { // Wait for state to be initialized. stateChannel := make(chan *feed.Event, 1) stateSub := stateFeed.Subscribe(stateChannel) go func() { select { case <-s.ctx.Done(): stateSub.Unsubscribe() return case <-stateChannel: stateSub.Unsubscribe() break } attThreshold := params.BeaconConfig().SecondsPerSlot / 3 ticker := slots.NewSlotTickerWithOffset(s.genesisTime, time.Duration(attThreshold)*time.Second, params.BeaconConfig().SecondsPerSlot) for { select { case <-ticker.C(): s.lateBlockTasks(ctx) case <-ctx.Done(): log.Debug("Context closed, exiting routine") return } } }() } // lateBlockTasks is called 4 seconds into the slot and performs tasks // related to late blocks. It emits a MissedSlot state feed event. // It calls FCU and sets the right attributes if we are proposing next slot // it also updates the next slot cache to deal with skipped slots. func (s *Service) lateBlockTasks(ctx context.Context) { if s.CurrentSlot() == s.HeadSlot() { return } s.cfg.StateNotifier.StateFeed().Send(&feed.Event{ Type: statefeed.MissedSlot, }) // Head root should be empty when retrieving proposer index for the next slot. _, id, has := s.cfg.ProposerSlotIndexCache.GetProposerPayloadIDs(s.CurrentSlot()+1, [32]byte{} /* head root */) // There exists proposer for next slot, but we haven't called fcu w/ payload attribute yet. if !has || id != [8]byte{} { return } s.headLock.RLock() headBlock, err := s.headBlock() if err != nil { s.headLock.RUnlock() log.WithError(err).Debug("could not perform late block tasks: failed to retrieve head block") return } headRoot := s.headRoot() headState := s.headState(ctx) s.headLock.RUnlock() _, err = s.notifyForkchoiceUpdate(ctx, ¬ifyForkchoiceUpdateArg{ headState: headState, headRoot: headRoot, headBlock: headBlock.Block(), }) if err != nil { log.WithError(err).Debug("could not perform late block tasks: failed to update forkchoice with engine") } lastRoot, lastState := transition.LastCachedState() if lastState == nil { lastRoot, lastState = headRoot[:], headState } if err = transition.UpdateNextSlotCache(ctx, lastRoot, lastState); err != nil { log.WithError(err).Debug("could not update next slot state cache") } }