package blockchain import ( "bytes" "context" "fmt" "github.com/pkg/errors" "github.com/prysmaticlabs/prysm/v4/beacon-chain/core/feed" statefeed "github.com/prysmaticlabs/prysm/v4/beacon-chain/core/feed/state" "github.com/prysmaticlabs/prysm/v4/beacon-chain/core/helpers" "github.com/prysmaticlabs/prysm/v4/beacon-chain/forkchoice" "github.com/prysmaticlabs/prysm/v4/beacon-chain/state" fieldparams "github.com/prysmaticlabs/prysm/v4/config/fieldparams" "github.com/prysmaticlabs/prysm/v4/config/params" "github.com/prysmaticlabs/prysm/v4/consensus-types/blocks" "github.com/prysmaticlabs/prysm/v4/consensus-types/interfaces" "github.com/prysmaticlabs/prysm/v4/consensus-types/primitives" "github.com/prysmaticlabs/prysm/v4/encoding/bytesutil" "github.com/prysmaticlabs/prysm/v4/math" ethpbv1 "github.com/prysmaticlabs/prysm/v4/proto/eth/v1" "github.com/prysmaticlabs/prysm/v4/runtime/version" "github.com/prysmaticlabs/prysm/v4/time/slots" "github.com/sirupsen/logrus" "go.opencensus.io/trace" ) // UpdateAndSaveHeadWithBalances updates the beacon state head after getting justified balanced from cache. // This function is only used in spec-tests, it does save the head after updating it. func (s *Service) UpdateAndSaveHeadWithBalances(ctx context.Context) error { s.cfg.ForkChoiceStore.Lock() defer s.cfg.ForkChoiceStore.Unlock() headRoot, err := s.cfg.ForkChoiceStore.Head(ctx) if err != nil { return errors.Wrap(err, "could not update head") } headBlock, err := s.getBlock(ctx, headRoot) if err != nil { return err } headState, err := s.cfg.StateGen.StateByRoot(ctx, headRoot) if err != nil { return errors.Wrap(err, "could not retrieve head state in DB") } return s.saveHead(ctx, headRoot, headBlock, headState) } // This defines the current chain service's view of head. type head struct { root [32]byte // current head root. block interfaces.ReadOnlySignedBeaconBlock // current head block. state state.BeaconState // current head state. slot primitives.Slot // the head block slot number optimistic bool // optimistic status when saved head } // This saves head info to the local service cache, it also saves the // new head root to the DB. // Caller of the method MUST acquire a lock on forkchoice. func (s *Service) saveHead(ctx context.Context, newHeadRoot [32]byte, headBlock interfaces.ReadOnlySignedBeaconBlock, headState state.BeaconState) error { ctx, span := trace.StartSpan(ctx, "blockChain.saveHead") defer span.End() // Do nothing if head hasn't changed. if !s.isNewHead(newHeadRoot) { return nil } if err := blocks.BeaconBlockIsNil(headBlock); err != nil { return err } if headState == nil || headState.IsNil() { return errors.New("cannot save nil head state") } // If the head state is not available, just return nil. // There's nothing to cache if !s.cfg.BeaconDB.HasStateSummary(ctx, newHeadRoot) { return nil } s.headLock.RLock() oldHeadBlock, err := s.headBlock() if err != nil { s.headLock.RUnlock() return errors.Wrap(err, "could not get old head block") } oldStateRoot := oldHeadBlock.Block().StateRoot() s.headLock.RUnlock() headSlot := s.HeadSlot() newHeadSlot := headBlock.Block().Slot() newStateRoot := headBlock.Block().StateRoot() r, err := s.HeadRoot(ctx) if err != nil { return errors.Wrap(err, "could not get old head root") } oldHeadRoot := bytesutil.ToBytes32(r) isOptimistic, err := s.cfg.ForkChoiceStore.IsOptimistic(newHeadRoot) if err != nil { log.WithError(err).Error("could not check if node is optimistically synced") } if headBlock.Block().ParentRoot() != oldHeadRoot { // A chain re-org occurred, so we fire an event notifying the rest of the services. commonRoot, forkSlot, err := s.cfg.ForkChoiceStore.CommonAncestor(ctx, oldHeadRoot, newHeadRoot) if err != nil { log.WithError(err).Error("Could not find common ancestor root") commonRoot = params.BeaconConfig().ZeroHash } dis := headSlot + newHeadSlot - 2*forkSlot dep := math.Max(uint64(headSlot-forkSlot), uint64(newHeadSlot-forkSlot)) oldWeight, err := s.cfg.ForkChoiceStore.Weight(oldHeadRoot) if err != nil { log.WithField("root", fmt.Sprintf("%#x", oldHeadRoot)).Warn("could not determine node weight") } newWeight, err := s.cfg.ForkChoiceStore.Weight(newHeadRoot) if err != nil { log.WithField("root", fmt.Sprintf("%#x", newHeadRoot)).Warn("could not determine node weight") } log.WithFields(logrus.Fields{ "newSlot": fmt.Sprintf("%d", newHeadSlot), "newRoot": fmt.Sprintf("%#x", newHeadRoot), "newWeight": newWeight, "oldSlot": fmt.Sprintf("%d", headSlot), "oldRoot": fmt.Sprintf("%#x", oldHeadRoot), "oldWeight": oldWeight, "commonAncestorRoot": fmt.Sprintf("%#x", commonRoot), "distance": dis, "depth": dep, }).Info("Chain reorg occurred") reorgDistance.Observe(float64(dis)) reorgDepth.Observe(float64(dep)) s.cfg.StateNotifier.StateFeed().Send(&feed.Event{ Type: statefeed.Reorg, Data: ðpbv1.EventChainReorg{ Slot: newHeadSlot, Depth: math.Max(uint64(headSlot-forkSlot), uint64(newHeadSlot-forkSlot)), OldHeadBlock: oldHeadRoot[:], NewHeadBlock: newHeadRoot[:], OldHeadState: oldStateRoot[:], NewHeadState: newStateRoot[:], Epoch: slots.ToEpoch(newHeadSlot), ExecutionOptimistic: isOptimistic, }, }) if err := s.saveOrphanedOperations(ctx, oldHeadRoot, newHeadRoot); err != nil { return err } reorgCount.Inc() } // Cache the new head info. newHead := &head{ root: newHeadRoot, block: headBlock, state: headState, optimistic: isOptimistic, slot: headBlock.Block().Slot(), } if err := s.setHead(newHead); err != nil { return errors.Wrap(err, "could not set head") } // Save the new head root to DB. if err := s.cfg.BeaconDB.SaveHeadBlockRoot(ctx, newHeadRoot); err != nil { return errors.Wrap(err, "could not save head root in DB") } // Forward an event capturing a new chain head over a common event feed // done in a goroutine to avoid blocking the critical runtime main routine. go func() { if err := s.notifyNewHeadEvent(ctx, newHeadSlot, headState, newStateRoot[:], newHeadRoot[:]); err != nil { log.WithError(err).Error("Could not notify event feed of new chain head") } }() return nil } // This gets called to update canonical root mapping. It does not save head block // root in DB. With the inception of initial-sync-cache-state flag, it uses finalized // check point as anchors to resume sync therefore head is no longer needed to be saved on per slot basis. func (s *Service) saveHeadNoDB(ctx context.Context, b interfaces.ReadOnlySignedBeaconBlock, r [32]byte, hs state.BeaconState, optimistic bool) error { if err := blocks.BeaconBlockIsNil(b); err != nil { return err } cachedHeadRoot, err := s.HeadRoot(ctx) if err != nil { return errors.Wrap(err, "could not get head root from cache") } if bytes.Equal(r[:], cachedHeadRoot) { return nil } bCp, err := b.Copy() if err != nil { return err } if err := s.setHeadInitialSync(r, bCp, hs, optimistic); err != nil { return errors.Wrap(err, "could not set head") } return nil } // This sets head view object which is used to track the head slot, root, block, state and optimistic status func (s *Service) setHead(newHead *head) error { s.headLock.Lock() defer s.headLock.Unlock() // This does a full copy of the block and state. bCp, err := newHead.block.Copy() if err != nil { return err } s.head = &head{ root: newHead.root, block: bCp, state: newHead.state.Copy(), optimistic: newHead.optimistic, slot: newHead.slot, } return nil } // This sets head view object which is used to track the head slot, root, block and state. The method // assumes that state being passed into the method will not be modified by any other alternate // caller which holds the state's reference. func (s *Service) setHeadInitialSync(root [32]byte, block interfaces.ReadOnlySignedBeaconBlock, state state.BeaconState, optimistic bool) error { s.headLock.Lock() defer s.headLock.Unlock() // This does a full copy of the block only. bCp, err := block.Copy() if err != nil { return err } s.head = &head{ root: root, block: bCp, state: state, optimistic: optimistic, } return nil } // This returns the head slot. // This is a lock free version. func (s *Service) headSlot() primitives.Slot { if s.head == nil || s.head.block == nil || s.head.block.Block() == nil { return 0 } return s.head.block.Block().Slot() } // This returns the head root. // It does a full copy on head root for immutability. // This is a lock free version. func (s *Service) headRoot() [32]byte { if s.head == nil { return params.BeaconConfig().ZeroHash } return s.head.root } // This returns the head block. // It does a full copy on head block for immutability. // This is a lock free version. func (s *Service) headBlock() (interfaces.ReadOnlySignedBeaconBlock, error) { return s.head.block.Copy() } // This returns the head state. // It does a full copy on head state for immutability. // This is a lock free version. func (s *Service) headState(ctx context.Context) state.BeaconState { _, span := trace.StartSpan(ctx, "blockChain.headState") defer span.End() return s.head.state.Copy() } // This returns a read only version of the head state. // It does not perform a copy of the head state. // This is a lock free version. func (s *Service) headStateReadOnly(ctx context.Context) state.ReadOnlyBeaconState { _, span := trace.StartSpan(ctx, "blockChain.headStateReadOnly") defer span.End() return s.head.state } // This returns the genesis validators root of the head state. // This is a lock free version. func (s *Service) headGenesisValidatorsRoot() [32]byte { return bytesutil.ToBytes32(s.head.state.GenesisValidatorsRoot()) } // This returns the validator referenced by the provided index in // the head state. // This is a lock free version. func (s *Service) headValidatorAtIndex(index primitives.ValidatorIndex) (state.ReadOnlyValidator, error) { return s.head.state.ValidatorAtIndexReadOnly(index) } // This returns the validator index referenced by the provided pubkey in // the head state. // This is a lock free version. func (s *Service) headValidatorIndexAtPubkey(pubKey [fieldparams.BLSPubkeyLength]byte) (primitives.ValidatorIndex, bool) { return s.head.state.ValidatorIndexByPubkey(pubKey) } // Returns true if head state exists. // This is the lock free version. func (s *Service) hasHeadState() bool { return s.head != nil && s.head.state != nil } // Notifies a common event feed of a new chain head event. Called right after a new // chain head is determined, set, and saved to disk. func (s *Service) notifyNewHeadEvent( ctx context.Context, newHeadSlot primitives.Slot, newHeadState state.BeaconState, newHeadStateRoot, newHeadRoot []byte, ) error { previousDutyDependentRoot := s.originBlockRoot[:] currentDutyDependentRoot := s.originBlockRoot[:] var previousDutyEpoch primitives.Epoch currentDutyEpoch := slots.ToEpoch(newHeadSlot) if currentDutyEpoch > 0 { previousDutyEpoch = currentDutyEpoch.Sub(1) } currentDutySlot, err := slots.EpochStart(currentDutyEpoch) if err != nil { return errors.Wrap(err, "could not get duty slot") } previousDutySlot, err := slots.EpochStart(previousDutyEpoch) if err != nil { return errors.Wrap(err, "could not get duty slot") } if currentDutySlot > 0 { currentDutyDependentRoot, err = helpers.BlockRootAtSlot(newHeadState, currentDutySlot-1) if err != nil { return errors.Wrap(err, "could not get duty dependent root") } } if previousDutySlot > 0 { previousDutyDependentRoot, err = helpers.BlockRootAtSlot(newHeadState, previousDutySlot-1) if err != nil { return errors.Wrap(err, "could not get duty dependent root") } } isOptimistic, err := s.IsOptimistic(ctx) if err != nil { return errors.Wrap(err, "could not check if node is optimistically synced") } s.cfg.StateNotifier.StateFeed().Send(&feed.Event{ Type: statefeed.NewHead, Data: ðpbv1.EventHead{ Slot: newHeadSlot, Block: newHeadRoot, State: newHeadStateRoot, EpochTransition: slots.IsEpochStart(newHeadSlot), PreviousDutyDependentRoot: previousDutyDependentRoot, CurrentDutyDependentRoot: currentDutyDependentRoot, ExecutionOptimistic: isOptimistic, }, }) return nil } // This saves the Attestations and BLSToExecChanges between `orphanedRoot` and the common ancestor root that is derived using `newHeadRoot`. // It also filters out the attestations that is one epoch older as a defense so invalid attestations don't flow into the attestation pool. func (s *Service) saveOrphanedOperations(ctx context.Context, orphanedRoot [32]byte, newHeadRoot [32]byte) error { commonAncestorRoot, _, err := s.cfg.ForkChoiceStore.CommonAncestor(ctx, newHeadRoot, orphanedRoot) switch { // Exit early if there's no common ancestor and root doesn't exist, there would be nothing to save. case errors.Is(err, forkchoice.ErrUnknownCommonAncestor): return nil case err != nil: return err } for orphanedRoot != commonAncestorRoot { if ctx.Err() != nil { return ctx.Err() } orphanedBlk, err := s.getBlock(ctx, orphanedRoot) if err != nil { return err } // If the block is an epoch older, break out of the loop since we can't include atts anyway. // This prevents stuck within this for loop longer than necessary. if orphanedBlk.Block().Slot()+params.BeaconConfig().SlotsPerEpoch <= s.CurrentSlot() { break } for _, a := range orphanedBlk.Block().Body().Attestations() { // if the attestation is one epoch older, it wouldn't been useful to save it. if a.Data.Slot+params.BeaconConfig().SlotsPerEpoch < s.CurrentSlot() { continue } if helpers.IsAggregated(a) { if err := s.cfg.AttPool.SaveAggregatedAttestation(a); err != nil { return err } } else { if err := s.cfg.AttPool.SaveUnaggregatedAttestation(a); err != nil { return err } } saveOrphanedAttCount.Inc() } for _, as := range orphanedBlk.Block().Body().AttesterSlashings() { if err := s.cfg.SlashingPool.InsertAttesterSlashing(ctx, s.headStateReadOnly(ctx), as); err != nil { log.WithError(err).Error("Could not insert reorg attester slashing") } } for _, vs := range orphanedBlk.Block().Body().ProposerSlashings() { if err := s.cfg.SlashingPool.InsertProposerSlashing(ctx, s.headStateReadOnly(ctx), vs); err != nil { log.WithError(err).Error("Could not insert reorg proposer slashing") } } for _, v := range orphanedBlk.Block().Body().VoluntaryExits() { s.cfg.ExitPool.InsertVoluntaryExit(v) } if orphanedBlk.Version() >= version.Capella { changes, err := orphanedBlk.Block().Body().BLSToExecutionChanges() if err != nil { return errors.Wrap(err, "could not get BLSToExecutionChanges") } for _, c := range changes { s.cfg.BLSToExecPool.InsertBLSToExecChange(c) } } parentRoot := orphanedBlk.Block().ParentRoot() orphanedRoot = bytesutil.ToBytes32(parentRoot[:]) } return nil }