mirror of
https://gitlab.com/pulsechaincom/prysm-pulse.git
synced 2024-12-25 21:07:18 +00:00
16b04699d0
* polling interval * adding proto message * changing proto messages * changing naming * adding slot functionality * initial sync working * new changes * more sync fixes * its working now * finally working * add tests * fix tests * tests * adding tests * lint * log checks * making changes to simulator * update logs * fix tests * get sync to work with crystallized state * fixing race * making requested changes * unexport * documentation * gazelle and fix merge conflicts * adding repeated requests * fix lint * adding new clock , db methods, and util func * revert change to test * gazelle * add in test * gazelle * finally working * save slot * fix lint and constant
351 lines
11 KiB
Go
351 lines
11 KiB
Go
// Package sync defines the utilities for the beacon-chain to sync with the network.
|
|
package sync
|
|
|
|
import (
|
|
"context"
|
|
"fmt"
|
|
|
|
"github.com/golang/protobuf/proto"
|
|
v "github.com/prysmaticlabs/prysm/beacon-chain/core/validators"
|
|
"github.com/prysmaticlabs/prysm/beacon-chain/types"
|
|
pb "github.com/prysmaticlabs/prysm/proto/beacon/p2p/v1"
|
|
"github.com/prysmaticlabs/prysm/shared/event"
|
|
"github.com/prysmaticlabs/prysm/shared/p2p"
|
|
"github.com/sirupsen/logrus"
|
|
"go.opencensus.io/trace"
|
|
)
|
|
|
|
var log = logrus.WithField("prefix", "sync")
|
|
|
|
type chainService interface {
|
|
IncomingBlockFeed() *event.Feed
|
|
}
|
|
|
|
type attestationService interface {
|
|
IncomingAttestationFeed() *event.Feed
|
|
}
|
|
|
|
type beaconDB interface {
|
|
GetCrystallizedState() (*types.CrystallizedState, error)
|
|
GetBlock([32]byte) (*types.Block, error)
|
|
HasBlock([32]byte) bool
|
|
GetChainHead() (*types.Block, error)
|
|
GetAttestation([32]byte) (*types.Attestation, error)
|
|
GetBlockBySlot(uint64) (*types.Block, error)
|
|
}
|
|
|
|
type queryService interface {
|
|
IsSynced() (bool, error)
|
|
}
|
|
|
|
type p2pAPI interface {
|
|
Subscribe(msg proto.Message, channel chan p2p.Message) event.Subscription
|
|
Send(msg proto.Message, peer p2p.Peer)
|
|
Broadcast(msg proto.Message)
|
|
}
|
|
|
|
// Service is the gateway and the bridge between the p2p network and the local beacon chain.
|
|
// In broad terms, a new block is synced in 4 steps:
|
|
// 1. Receive a block hash from a peer
|
|
// 2. Request the block for the hash from the network
|
|
// 3. Receive the block
|
|
// 4. Forward block to the beacon service for full validation
|
|
//
|
|
// In addition, Service will handle the following responsibilities:
|
|
// * Decide which messages are forwarded to other peers
|
|
// * Filter redundant data and unwanted data
|
|
// * Drop peers that send invalid data
|
|
// * Throttle incoming requests
|
|
type Service struct {
|
|
ctx context.Context
|
|
cancel context.CancelFunc
|
|
p2p p2pAPI
|
|
chainService chainService
|
|
attestationService attestationService
|
|
queryService queryService
|
|
db beaconDB
|
|
blockAnnouncementFeed *event.Feed
|
|
announceBlockBuf chan p2p.Message
|
|
blockBuf chan p2p.Message
|
|
blockRequestBySlot chan p2p.Message
|
|
chainHeadReqBuf chan p2p.Message
|
|
attestationBuf chan p2p.Message
|
|
}
|
|
|
|
// Config allows the channel's buffer sizes to be changed.
|
|
type Config struct {
|
|
BlockAnnounceBufferSize int
|
|
BlockBufferSize int
|
|
BlockRequestBufferSize int
|
|
AttestationBufferSize int
|
|
ChainHeadReqBufferSize int
|
|
ChainService chainService
|
|
AttestService attestationService
|
|
QueryService queryService
|
|
BeaconDB beaconDB
|
|
P2P p2pAPI
|
|
}
|
|
|
|
// DefaultConfig provides the default configuration for a sync service.
|
|
func DefaultConfig() Config {
|
|
return Config{
|
|
BlockAnnounceBufferSize: 100,
|
|
BlockBufferSize: 100,
|
|
BlockRequestBufferSize: 100,
|
|
ChainHeadReqBufferSize: 100,
|
|
AttestationBufferSize: 100,
|
|
}
|
|
}
|
|
|
|
// NewSyncService accepts a context and returns a new Service.
|
|
func NewSyncService(ctx context.Context, cfg Config) *Service {
|
|
ctx, cancel := context.WithCancel(ctx)
|
|
return &Service{
|
|
ctx: ctx,
|
|
cancel: cancel,
|
|
p2p: cfg.P2P,
|
|
chainService: cfg.ChainService,
|
|
db: cfg.BeaconDB,
|
|
attestationService: cfg.AttestService,
|
|
queryService: cfg.QueryService,
|
|
blockAnnouncementFeed: new(event.Feed),
|
|
announceBlockBuf: make(chan p2p.Message, cfg.BlockAnnounceBufferSize),
|
|
blockBuf: make(chan p2p.Message, cfg.BlockBufferSize),
|
|
blockRequestBySlot: make(chan p2p.Message, cfg.BlockRequestBufferSize),
|
|
attestationBuf: make(chan p2p.Message, cfg.AttestationBufferSize),
|
|
chainHeadReqBuf: make(chan p2p.Message, cfg.ChainHeadReqBufferSize),
|
|
}
|
|
}
|
|
|
|
// IsSyncedWithNetwork queries other nodes in the network
|
|
// to determine whether or not the local chain is synced
|
|
// with the rest of the network.
|
|
// TODO(#661): Implement this method.
|
|
func (ss *Service) IsSyncedWithNetwork() bool {
|
|
return false
|
|
}
|
|
|
|
// Start begins the block processing goroutine.
|
|
func (ss *Service) Start() {
|
|
synced, err := ss.queryService.IsSynced()
|
|
if err != nil {
|
|
log.Error(err)
|
|
}
|
|
|
|
if !synced {
|
|
log.Info("Chain state not detected starting initial sync")
|
|
return
|
|
}
|
|
|
|
go ss.run()
|
|
}
|
|
|
|
// ResumeSync resumes normal sync after initial sync is complete.
|
|
func (ss *Service) ResumeSync() {
|
|
go ss.run()
|
|
}
|
|
|
|
// Stop kills the block processing goroutine, but does not wait until the goroutine exits.
|
|
func (ss *Service) Stop() error {
|
|
log.Info("Stopping service")
|
|
ss.cancel()
|
|
return nil
|
|
}
|
|
|
|
// BlockAnnouncementFeed returns an event feed processes can subscribe to for
|
|
// newly received, incoming p2p blocks.
|
|
func (ss *Service) BlockAnnouncementFeed() *event.Feed {
|
|
return ss.blockAnnouncementFeed
|
|
}
|
|
|
|
// run handles incoming block sync.
|
|
func (ss *Service) run() {
|
|
announceBlockSub := ss.p2p.Subscribe(&pb.BeaconBlockAnnounce{}, ss.announceBlockBuf)
|
|
blockSub := ss.p2p.Subscribe(&pb.BeaconBlockResponse{}, ss.blockBuf)
|
|
blockRequestSub := ss.p2p.Subscribe(&pb.BeaconBlockRequestBySlotNumber{}, ss.blockRequestBySlot)
|
|
attestationSub := ss.p2p.Subscribe(&pb.AggregatedAttestation{}, ss.attestationBuf)
|
|
chainHeadReqSub := ss.p2p.Subscribe(&pb.ChainHeadRequest{}, ss.chainHeadReqBuf)
|
|
|
|
defer announceBlockSub.Unsubscribe()
|
|
defer blockSub.Unsubscribe()
|
|
defer blockRequestSub.Unsubscribe()
|
|
defer chainHeadReqSub.Unsubscribe()
|
|
defer attestationSub.Unsubscribe()
|
|
|
|
for {
|
|
select {
|
|
case <-ss.ctx.Done():
|
|
log.Debug("Exiting goroutine")
|
|
return
|
|
case msg := <-ss.announceBlockBuf:
|
|
ss.receiveBlockAnnounce(msg)
|
|
case msg := <-ss.attestationBuf:
|
|
ss.receiveAttestation(msg)
|
|
case msg := <-ss.blockBuf:
|
|
ss.receiveBlock(msg)
|
|
case msg := <-ss.blockRequestBySlot:
|
|
ss.handleBlockRequestBySlot(msg)
|
|
case msg := <-ss.chainHeadReqBuf:
|
|
ss.handleChainHeadRequest(msg)
|
|
}
|
|
}
|
|
}
|
|
|
|
// receiveBlockHash accepts a block hash.
|
|
// New hashes are forwarded to other peers in the network (unimplemented), and
|
|
// the contents of the block are requested if the local chain doesn't have the block.
|
|
func (ss *Service) receiveBlockAnnounce(msg p2p.Message) {
|
|
ctx, receiveBlockSpan := trace.StartSpan(msg.Ctx, "receiveBlockHash")
|
|
defer receiveBlockSpan.End()
|
|
|
|
data := msg.Data.(*pb.BeaconBlockAnnounce)
|
|
var h [32]byte
|
|
copy(h[:], data.Hash[:32])
|
|
|
|
if ss.db.HasBlock(h) {
|
|
log.Debugf("Received a hash for a block that has already been processed: %#x", h)
|
|
return
|
|
}
|
|
|
|
log.WithField("blockHash", fmt.Sprintf("%#x", h)).Debug("Received incoming block hash, requesting full block data from sender")
|
|
// Request the full block data from peer that sent the block hash.
|
|
_, sendBlockRequestSpan := trace.StartSpan(ctx, "sendBlockRequest")
|
|
ss.p2p.Send(&pb.BeaconBlockRequest{Hash: h[:]}, msg.Peer)
|
|
sendBlockRequestSpan.End()
|
|
}
|
|
|
|
// receiveBlock processes a block from the p2p layer.
|
|
func (ss *Service) receiveBlock(msg p2p.Message) {
|
|
ctx, receiveBlockSpan := trace.StartSpan(msg.Ctx, "receiveBlock")
|
|
defer receiveBlockSpan.End()
|
|
|
|
response := msg.Data.(*pb.BeaconBlockResponse)
|
|
block := types.NewBlock(response.Block)
|
|
blockHash, err := block.Hash()
|
|
if err != nil {
|
|
log.Errorf("Could not hash received block: %v", err)
|
|
}
|
|
|
|
log.Debugf("Processing response to block request: %#x", blockHash)
|
|
|
|
if ss.db.HasBlock(blockHash) {
|
|
log.Debug("Received a block that already exists. Exiting...")
|
|
return
|
|
}
|
|
|
|
cState, err := ss.db.GetCrystallizedState()
|
|
if err != nil {
|
|
log.Errorf("Failed to get crystallized state: %v", err)
|
|
return
|
|
}
|
|
|
|
if block.SlotNumber() < cState.LastFinalizedSlot() {
|
|
log.Debug("Discarding received block with a slot number smaller than the last finalized slot")
|
|
return
|
|
}
|
|
|
|
// Verify attestation coming from proposer then forward block to the subscribers.
|
|
attestation := types.NewAttestation(response.Attestation)
|
|
|
|
proposerShardID, _, err := v.ProposerShardAndIndex(cState.ShardAndCommitteesForSlots(), cState.LastStateRecalculationSlot(), block.SlotNumber())
|
|
if err != nil {
|
|
log.Errorf("Failed to get proposer shard ID: %v", err)
|
|
return
|
|
}
|
|
|
|
// TODO(#258): stubbing public key with empty 32 bytes.
|
|
if err := attestation.VerifyProposerAttestation([32]byte{}, proposerShardID); err != nil {
|
|
log.Errorf("Failed to verify proposer attestation: %v", err)
|
|
return
|
|
}
|
|
|
|
_, sendAttestationSpan := trace.StartSpan(ctx, "sendAttestation")
|
|
log.WithField("attestationHash", fmt.Sprintf("%#x", attestation.Key())).Debug("Sending newly received attestation to subscribers")
|
|
ss.attestationService.IncomingAttestationFeed().Send(attestation)
|
|
sendAttestationSpan.End()
|
|
|
|
_, sendBlockSpan := trace.StartSpan(ctx, "sendBlock")
|
|
log.WithField("blockHash", fmt.Sprintf("%#x", blockHash)).Debug("Sending newly received block to subscribers")
|
|
ss.chainService.IncomingBlockFeed().Send(block)
|
|
sendBlockSpan.End()
|
|
}
|
|
|
|
// handleBlockRequestBySlot processes a block request from the p2p layer.
|
|
// if found, the block is sent to the requesting peer.
|
|
func (ss *Service) handleBlockRequestBySlot(msg p2p.Message) {
|
|
ctx, blockRequestSpan := trace.StartSpan(msg.Ctx, "blockRequestBySlot")
|
|
defer blockRequestSpan.End()
|
|
|
|
request, ok := msg.Data.(*pb.BeaconBlockRequestBySlotNumber)
|
|
if !ok {
|
|
log.Error("Received malformed beacon block request p2p message")
|
|
return
|
|
}
|
|
|
|
ctx, getBlockSpan := trace.StartSpan(ctx, "getBlockBySlot")
|
|
block, err := ss.db.GetBlockBySlot(request.GetSlotNumber())
|
|
getBlockSpan.End()
|
|
if err != nil || block == nil {
|
|
log.Errorf("Error retrieving block from db: %v", err)
|
|
return
|
|
}
|
|
|
|
_, sendBlockSpan := trace.StartSpan(ctx, "sendBlock")
|
|
log.WithField("slotNumber", fmt.Sprintf("%d", request.GetSlotNumber())).Debug("Sending requested block to peer")
|
|
ss.p2p.Send(block.Proto(), msg.Peer)
|
|
sendBlockSpan.End()
|
|
}
|
|
|
|
func (ss *Service) handleChainHeadRequest(msg p2p.Message) {
|
|
if _, ok := msg.Data.(*pb.ChainHeadRequest); !ok {
|
|
log.Errorf("message is of the incorrect type")
|
|
return
|
|
}
|
|
|
|
block, err := ss.db.GetChainHead()
|
|
if err != nil {
|
|
log.Errorf("Could not retrieve chain head %v", err)
|
|
return
|
|
}
|
|
|
|
hash, err := block.Hash()
|
|
if err != nil {
|
|
log.Errorf("Could not hash block %v", err)
|
|
return
|
|
}
|
|
|
|
req := &pb.ChainHeadResponse{
|
|
Slot: block.SlotNumber(),
|
|
Hash: hash[:],
|
|
Block: block.Proto(),
|
|
}
|
|
|
|
ss.p2p.Send(req, msg.Peer)
|
|
}
|
|
|
|
// receiveAttestation accepts an broadcasted attestation from the p2p layer,
|
|
// discard the attestation if we have gotten before, send it to attestation
|
|
// service if we have not.
|
|
func (ss *Service) receiveAttestation(msg p2p.Message) {
|
|
data := msg.Data.(*pb.AggregatedAttestation)
|
|
a := types.NewAttestation(data)
|
|
h := a.Key()
|
|
|
|
attestation, err := ss.db.GetAttestation(h)
|
|
if err != nil {
|
|
log.Errorf("Could not check for attestation in DB: %v", err)
|
|
return
|
|
}
|
|
if attestation != nil {
|
|
validatorExists := attestation.ContainsValidator(a.AttesterBitfield())
|
|
if validatorExists {
|
|
log.Debugf("Received attestation %#x already", h)
|
|
return
|
|
}
|
|
}
|
|
|
|
log.WithField("attestationHash", fmt.Sprintf("%#x", h)).Debug("Forwarding attestation to subscribed services")
|
|
// Request the full block data from peer that sent the block hash.
|
|
ss.attestationService.IncomingAttestationFeed().Send(a)
|
|
}
|