mirror of
https://gitlab.com/pulsechaincom/prysm-pulse.git
synced 2024-12-22 11:32:09 +00:00
6d3c6a6331
* moving route registration out of the start function and into registration * moving where grpc is set
1101 lines
36 KiB
Go
1101 lines
36 KiB
Go
// Package node is the main service which launches a beacon node and manages
|
|
// the lifecycle of all its associated services at runtime, such as p2p, RPC, sync,
|
|
// gracefully closing them if the process ends.
|
|
package node
|
|
|
|
import (
|
|
"bytes"
|
|
"context"
|
|
"fmt"
|
|
"os"
|
|
"os/signal"
|
|
"path/filepath"
|
|
"strings"
|
|
"sync"
|
|
"syscall"
|
|
|
|
"github.com/ethereum/go-ethereum/common"
|
|
"github.com/gorilla/mux"
|
|
"github.com/pkg/errors"
|
|
apigateway "github.com/prysmaticlabs/prysm/v5/api/gateway"
|
|
"github.com/prysmaticlabs/prysm/v5/api/server"
|
|
"github.com/prysmaticlabs/prysm/v5/async/event"
|
|
"github.com/prysmaticlabs/prysm/v5/beacon-chain/blockchain"
|
|
"github.com/prysmaticlabs/prysm/v5/beacon-chain/builder"
|
|
"github.com/prysmaticlabs/prysm/v5/beacon-chain/cache"
|
|
"github.com/prysmaticlabs/prysm/v5/beacon-chain/cache/depositcache"
|
|
"github.com/prysmaticlabs/prysm/v5/beacon-chain/cache/depositsnapshot"
|
|
"github.com/prysmaticlabs/prysm/v5/beacon-chain/db"
|
|
"github.com/prysmaticlabs/prysm/v5/beacon-chain/db/filesystem"
|
|
"github.com/prysmaticlabs/prysm/v5/beacon-chain/db/kv"
|
|
"github.com/prysmaticlabs/prysm/v5/beacon-chain/db/slasherkv"
|
|
interopcoldstart "github.com/prysmaticlabs/prysm/v5/beacon-chain/deterministic-genesis"
|
|
"github.com/prysmaticlabs/prysm/v5/beacon-chain/execution"
|
|
"github.com/prysmaticlabs/prysm/v5/beacon-chain/forkchoice"
|
|
doublylinkedtree "github.com/prysmaticlabs/prysm/v5/beacon-chain/forkchoice/doubly-linked-tree"
|
|
"github.com/prysmaticlabs/prysm/v5/beacon-chain/gateway"
|
|
"github.com/prysmaticlabs/prysm/v5/beacon-chain/monitor"
|
|
"github.com/prysmaticlabs/prysm/v5/beacon-chain/node/registration"
|
|
"github.com/prysmaticlabs/prysm/v5/beacon-chain/operations/attestations"
|
|
"github.com/prysmaticlabs/prysm/v5/beacon-chain/operations/blstoexec"
|
|
"github.com/prysmaticlabs/prysm/v5/beacon-chain/operations/slashings"
|
|
"github.com/prysmaticlabs/prysm/v5/beacon-chain/operations/synccommittee"
|
|
"github.com/prysmaticlabs/prysm/v5/beacon-chain/operations/voluntaryexits"
|
|
"github.com/prysmaticlabs/prysm/v5/beacon-chain/p2p"
|
|
"github.com/prysmaticlabs/prysm/v5/beacon-chain/p2p/peers"
|
|
"github.com/prysmaticlabs/prysm/v5/beacon-chain/rpc"
|
|
"github.com/prysmaticlabs/prysm/v5/beacon-chain/slasher"
|
|
"github.com/prysmaticlabs/prysm/v5/beacon-chain/startup"
|
|
"github.com/prysmaticlabs/prysm/v5/beacon-chain/state"
|
|
"github.com/prysmaticlabs/prysm/v5/beacon-chain/state/stategen"
|
|
regularsync "github.com/prysmaticlabs/prysm/v5/beacon-chain/sync"
|
|
"github.com/prysmaticlabs/prysm/v5/beacon-chain/sync/backfill"
|
|
"github.com/prysmaticlabs/prysm/v5/beacon-chain/sync/backfill/coverage"
|
|
"github.com/prysmaticlabs/prysm/v5/beacon-chain/sync/checkpoint"
|
|
"github.com/prysmaticlabs/prysm/v5/beacon-chain/sync/genesis"
|
|
initialsync "github.com/prysmaticlabs/prysm/v5/beacon-chain/sync/initial-sync"
|
|
"github.com/prysmaticlabs/prysm/v5/beacon-chain/verification"
|
|
"github.com/prysmaticlabs/prysm/v5/cmd"
|
|
"github.com/prysmaticlabs/prysm/v5/cmd/beacon-chain/flags"
|
|
"github.com/prysmaticlabs/prysm/v5/config/features"
|
|
"github.com/prysmaticlabs/prysm/v5/config/params"
|
|
"github.com/prysmaticlabs/prysm/v5/consensus-types/primitives"
|
|
"github.com/prysmaticlabs/prysm/v5/container/slice"
|
|
"github.com/prysmaticlabs/prysm/v5/encoding/bytesutil"
|
|
"github.com/prysmaticlabs/prysm/v5/monitoring/prometheus"
|
|
"github.com/prysmaticlabs/prysm/v5/runtime"
|
|
"github.com/prysmaticlabs/prysm/v5/runtime/debug"
|
|
"github.com/prysmaticlabs/prysm/v5/runtime/prereqs"
|
|
"github.com/prysmaticlabs/prysm/v5/runtime/version"
|
|
"github.com/sirupsen/logrus"
|
|
"github.com/urfave/cli/v2"
|
|
)
|
|
|
|
const testSkipPowFlag = "test-skip-pow"
|
|
|
|
// Used as a struct to keep cli flag options for configuring services
|
|
// for the beacon node. We keep this as a separate struct to not pollute the actual BeaconNode
|
|
// struct, as it is merely used to pass down configuration options into the appropriate services.
|
|
type serviceFlagOpts struct {
|
|
blockchainFlagOpts []blockchain.Option
|
|
executionChainFlagOpts []execution.Option
|
|
builderOpts []builder.Option
|
|
}
|
|
|
|
// BeaconNode defines a struct that handles the services running a random beacon chain
|
|
// full PoS node. It handles the lifecycle of the entire system and registers
|
|
// services to a service registry.
|
|
type BeaconNode struct {
|
|
cliCtx *cli.Context
|
|
ctx context.Context
|
|
cancel context.CancelFunc
|
|
services *runtime.ServiceRegistry
|
|
lock sync.RWMutex
|
|
stop chan struct{} // Channel to wait for termination notifications.
|
|
db db.Database
|
|
slasherDB db.SlasherDatabase
|
|
attestationPool attestations.Pool
|
|
exitPool voluntaryexits.PoolManager
|
|
slashingsPool slashings.PoolManager
|
|
syncCommitteePool synccommittee.Pool
|
|
blsToExecPool blstoexec.PoolManager
|
|
depositCache cache.DepositCache
|
|
trackedValidatorsCache *cache.TrackedValidatorsCache
|
|
payloadIDCache *cache.PayloadIDCache
|
|
stateFeed *event.Feed
|
|
blockFeed *event.Feed
|
|
opFeed *event.Feed
|
|
stateGen *stategen.State
|
|
collector *bcnodeCollector
|
|
slasherBlockHeadersFeed *event.Feed
|
|
slasherAttestationsFeed *event.Feed
|
|
finalizedStateAtStartUp state.BeaconState
|
|
serviceFlagOpts *serviceFlagOpts
|
|
GenesisInitializer genesis.Initializer
|
|
CheckpointInitializer checkpoint.Initializer
|
|
forkChoicer forkchoice.ForkChoicer
|
|
clockWaiter startup.ClockWaiter
|
|
BackfillOpts []backfill.ServiceOption
|
|
initialSyncComplete chan struct{}
|
|
BlobStorage *filesystem.BlobStorage
|
|
BlobStorageOptions []filesystem.BlobStorageOption
|
|
blobRetentionEpochs primitives.Epoch
|
|
verifyInitWaiter *verification.InitializerWaiter
|
|
syncChecker *initialsync.SyncChecker
|
|
}
|
|
|
|
// New creates a new node instance, sets up configuration options, and registers
|
|
// every required service to the node.
|
|
func New(cliCtx *cli.Context, cancel context.CancelFunc, opts ...Option) (*BeaconNode, error) {
|
|
if err := configureTracing(cliCtx); err != nil {
|
|
return nil, err
|
|
}
|
|
prereqs.WarnIfPlatformNotSupported(cliCtx.Context)
|
|
if hasNetworkFlag(cliCtx) && cliCtx.IsSet(cmd.ChainConfigFileFlag.Name) {
|
|
return nil, fmt.Errorf("%s cannot be passed concurrently with network flag", cmd.ChainConfigFileFlag.Name)
|
|
}
|
|
if err := features.ConfigureBeaconChain(cliCtx); err != nil {
|
|
return nil, err
|
|
}
|
|
if err := cmd.ConfigureBeaconChain(cliCtx); err != nil {
|
|
return nil, err
|
|
}
|
|
flags.ConfigureGlobalFlags(cliCtx)
|
|
if err := configureChainConfig(cliCtx); err != nil {
|
|
return nil, err
|
|
}
|
|
if err := configureHistoricalSlasher(cliCtx); err != nil {
|
|
return nil, err
|
|
}
|
|
err := configureBuilderCircuitBreaker(cliCtx)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
if err := configureSlotsPerArchivedPoint(cliCtx); err != nil {
|
|
return nil, err
|
|
}
|
|
if err := configureEth1Config(cliCtx); err != nil {
|
|
return nil, err
|
|
}
|
|
configureNetwork(cliCtx)
|
|
if err := configureInteropConfig(cliCtx); err != nil {
|
|
return nil, err
|
|
}
|
|
if err := configureExecutionSetting(cliCtx); err != nil {
|
|
return nil, err
|
|
}
|
|
configureFastSSZHashingAlgorithm()
|
|
|
|
// Initializes any forks here.
|
|
params.BeaconConfig().InitializeForkSchedule()
|
|
|
|
registry := runtime.NewServiceRegistry()
|
|
|
|
ctx := cliCtx.Context
|
|
beacon := &BeaconNode{
|
|
cliCtx: cliCtx,
|
|
ctx: ctx,
|
|
cancel: cancel,
|
|
services: registry,
|
|
stop: make(chan struct{}),
|
|
stateFeed: new(event.Feed),
|
|
blockFeed: new(event.Feed),
|
|
opFeed: new(event.Feed),
|
|
attestationPool: attestations.NewPool(),
|
|
exitPool: voluntaryexits.NewPool(),
|
|
slashingsPool: slashings.NewPool(),
|
|
syncCommitteePool: synccommittee.NewPool(),
|
|
blsToExecPool: blstoexec.NewPool(),
|
|
trackedValidatorsCache: cache.NewTrackedValidatorsCache(),
|
|
payloadIDCache: cache.NewPayloadIDCache(),
|
|
slasherBlockHeadersFeed: new(event.Feed),
|
|
slasherAttestationsFeed: new(event.Feed),
|
|
serviceFlagOpts: &serviceFlagOpts{},
|
|
}
|
|
|
|
beacon.initialSyncComplete = make(chan struct{})
|
|
beacon.syncChecker = &initialsync.SyncChecker{}
|
|
for _, opt := range opts {
|
|
if err := opt(beacon); err != nil {
|
|
return nil, err
|
|
}
|
|
}
|
|
|
|
synchronizer := startup.NewClockSynchronizer()
|
|
beacon.clockWaiter = synchronizer
|
|
|
|
beacon.forkChoicer = doublylinkedtree.New()
|
|
depositAddress, err := execution.DepositContractAddress()
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
// Allow tests to set it as an opt.
|
|
if beacon.BlobStorage == nil {
|
|
beacon.BlobStorageOptions = append(beacon.BlobStorageOptions, filesystem.WithSaveFsync(features.Get().BlobSaveFsync))
|
|
blobs, err := filesystem.NewBlobStorage(beacon.BlobStorageOptions...)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
beacon.BlobStorage = blobs
|
|
}
|
|
|
|
log.Debugln("Starting DB")
|
|
if err := beacon.startDB(cliCtx, depositAddress); err != nil {
|
|
return nil, err
|
|
}
|
|
beacon.BlobStorage.WarmCache()
|
|
|
|
log.Debugln("Starting Slashing DB")
|
|
if err := beacon.startSlasherDB(cliCtx); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
log.Debugln("Registering P2P Service")
|
|
if err := beacon.registerP2P(cliCtx); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
bfs, err := backfill.NewUpdater(ctx, beacon.db)
|
|
if err != nil {
|
|
return nil, errors.Wrap(err, "backfill status initialization error")
|
|
}
|
|
|
|
log.Debugln("Starting State Gen")
|
|
if err := beacon.startStateGen(ctx, bfs, beacon.forkChoicer); err != nil {
|
|
if errors.Is(err, stategen.ErrNoGenesisBlock) {
|
|
log.Errorf("No genesis block/state is found. Prysm only provides a mainnet genesis "+
|
|
"state bundled in the application. You must provide the --%s or --%s flag to load "+
|
|
"a genesis block/state for this network.", "genesis-state", "genesis-beacon-api-url")
|
|
}
|
|
return nil, err
|
|
}
|
|
|
|
beacon.verifyInitWaiter = verification.NewInitializerWaiter(
|
|
beacon.clockWaiter, forkchoice.NewROForkChoice(beacon.forkChoicer), beacon.stateGen)
|
|
|
|
pa := peers.NewAssigner(beacon.fetchP2P().Peers(), beacon.forkChoicer)
|
|
beacon.BackfillOpts = append(beacon.BackfillOpts, backfill.WithVerifierWaiter(beacon.verifyInitWaiter),
|
|
backfill.WithInitSyncWaiter(initSyncWaiter(ctx, beacon.initialSyncComplete)))
|
|
bf, err := backfill.NewService(ctx, bfs, beacon.BlobStorage, beacon.clockWaiter, beacon.fetchP2P(), pa, beacon.BackfillOpts...)
|
|
if err != nil {
|
|
return nil, errors.Wrap(err, "error initializing backfill service")
|
|
}
|
|
if err := beacon.services.RegisterService(bf); err != nil {
|
|
return nil, errors.Wrap(err, "error registering backfill service")
|
|
}
|
|
|
|
log.Debugln("Registering POW Chain Service")
|
|
if err := beacon.registerPOWChainService(); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
log.Debugln("Registering Attestation Pool Service")
|
|
if err := beacon.registerAttestationPool(); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
log.Debugln("Registering Deterministic Genesis Service")
|
|
if err := beacon.registerDeterministicGenesisService(); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
log.Debugln("Registering Blockchain Service")
|
|
if err := beacon.registerBlockchainService(beacon.forkChoicer, synchronizer, beacon.initialSyncComplete); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
log.Debugln("Registering Initial Sync Service")
|
|
if err := beacon.registerInitialSyncService(beacon.initialSyncComplete); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
log.Debugln("Registering Sync Service")
|
|
if err := beacon.registerSyncService(beacon.initialSyncComplete, bfs); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
log.Debugln("Registering Slasher Service")
|
|
if err := beacon.registerSlasherService(); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
log.Debugln("Registering builder service")
|
|
if err := beacon.registerBuilderService(cliCtx); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
log.Debugln("Registering RPC Service")
|
|
router := newRouter(cliCtx)
|
|
if err := beacon.registerRPCService(router); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
log.Debugln("Registering GRPC Gateway Service")
|
|
if err := beacon.registerGRPCGateway(router); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
log.Debugln("Registering Validator Monitoring Service")
|
|
if err := beacon.registerValidatorMonitorService(beacon.initialSyncComplete); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
if !cliCtx.Bool(cmd.DisableMonitoringFlag.Name) {
|
|
log.Debugln("Registering Prometheus Service")
|
|
if err := beacon.registerPrometheusService(cliCtx); err != nil {
|
|
return nil, err
|
|
}
|
|
}
|
|
|
|
// db.DatabasePath is the path to the containing directory
|
|
// db.NewFileName expands that to the canonical full path using
|
|
// the same construction as NewDB()
|
|
c, err := newBeaconNodePromCollector(db.NewFileName(beacon.db.DatabasePath()))
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
beacon.collector = c
|
|
|
|
// Do not store the finalized state as it has been provided to the respective services during
|
|
// their initialization.
|
|
beacon.finalizedStateAtStartUp = nil
|
|
|
|
return beacon, nil
|
|
}
|
|
func initSyncWaiter(ctx context.Context, complete chan struct{}) func() error {
|
|
return func() error {
|
|
select {
|
|
case <-complete:
|
|
return nil
|
|
case <-ctx.Done():
|
|
return ctx.Err()
|
|
}
|
|
}
|
|
}
|
|
|
|
func newRouter(cliCtx *cli.Context) *mux.Router {
|
|
var allowedOrigins []string
|
|
if cliCtx.IsSet(flags.GPRCGatewayCorsDomain.Name) {
|
|
allowedOrigins = strings.Split(cliCtx.String(flags.GPRCGatewayCorsDomain.Name), ",")
|
|
} else {
|
|
allowedOrigins = strings.Split(flags.GPRCGatewayCorsDomain.Value, ",")
|
|
}
|
|
r := mux.NewRouter()
|
|
r.Use(server.NormalizeQueryValuesHandler)
|
|
r.Use(server.CorsHandler(allowedOrigins))
|
|
return r
|
|
}
|
|
|
|
// StateFeed implements statefeed.Notifier.
|
|
func (b *BeaconNode) StateFeed() *event.Feed {
|
|
return b.stateFeed
|
|
}
|
|
|
|
// BlockFeed implements blockfeed.Notifier.
|
|
func (b *BeaconNode) BlockFeed() *event.Feed {
|
|
return b.blockFeed
|
|
}
|
|
|
|
// OperationFeed implements opfeed.Notifier.
|
|
func (b *BeaconNode) OperationFeed() *event.Feed {
|
|
return b.opFeed
|
|
}
|
|
|
|
// Start the BeaconNode and kicks off every registered service.
|
|
func (b *BeaconNode) Start() {
|
|
b.lock.Lock()
|
|
|
|
log.WithFields(logrus.Fields{
|
|
"version": version.Version(),
|
|
}).Info("Starting beacon node")
|
|
|
|
b.services.StartAll()
|
|
|
|
stop := b.stop
|
|
b.lock.Unlock()
|
|
|
|
go func() {
|
|
sigc := make(chan os.Signal, 1)
|
|
signal.Notify(sigc, syscall.SIGINT, syscall.SIGTERM)
|
|
defer signal.Stop(sigc)
|
|
<-sigc
|
|
log.Info("Got interrupt, shutting down...")
|
|
debug.Exit(b.cliCtx) // Ensure trace and CPU profile data are flushed.
|
|
go b.Close()
|
|
for i := 10; i > 0; i-- {
|
|
<-sigc
|
|
if i > 1 {
|
|
log.WithField("times", i-1).Info("Already shutting down, interrupt more to panic")
|
|
}
|
|
}
|
|
panic("Panic closing the beacon node")
|
|
}()
|
|
|
|
// Wait for stop channel to be closed.
|
|
<-stop
|
|
}
|
|
|
|
// Close handles graceful shutdown of the system.
|
|
func (b *BeaconNode) Close() {
|
|
b.lock.Lock()
|
|
defer b.lock.Unlock()
|
|
|
|
log.Info("Stopping beacon node")
|
|
b.services.StopAll()
|
|
if err := b.db.Close(); err != nil {
|
|
log.WithError(err).Error("Failed to close database")
|
|
}
|
|
b.collector.unregister()
|
|
b.cancel()
|
|
close(b.stop)
|
|
}
|
|
|
|
func (b *BeaconNode) startDB(cliCtx *cli.Context, depositAddress string) error {
|
|
baseDir := cliCtx.String(cmd.DataDirFlag.Name)
|
|
dbPath := filepath.Join(baseDir, kv.BeaconNodeDbDirName)
|
|
clearDB := cliCtx.Bool(cmd.ClearDB.Name)
|
|
forceClearDB := cliCtx.Bool(cmd.ForceClearDB.Name)
|
|
|
|
log.WithField("databasePath", dbPath).Info("Checking DB")
|
|
|
|
d, err := kv.NewKVStore(b.ctx, dbPath)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
clearDBConfirmed := false
|
|
if clearDB && !forceClearDB {
|
|
actionText := "This will delete your beacon chain database stored in your data directory. " +
|
|
"Your database backups will not be removed - do you want to proceed? (Y/N)"
|
|
deniedText := "Database will not be deleted. No changes have been made."
|
|
clearDBConfirmed, err = cmd.ConfirmAction(actionText, deniedText)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
}
|
|
if clearDBConfirmed || forceClearDB {
|
|
log.Warning("Removing database")
|
|
if err := d.ClearDB(); err != nil {
|
|
return errors.Wrap(err, "could not clear database")
|
|
}
|
|
if err := b.BlobStorage.Clear(); err != nil {
|
|
return errors.Wrap(err, "could not clear blob storage")
|
|
}
|
|
d, err = kv.NewKVStore(b.ctx, dbPath)
|
|
if err != nil {
|
|
return errors.Wrap(err, "could not create new database")
|
|
}
|
|
}
|
|
|
|
if err := d.RunMigrations(b.ctx); err != nil {
|
|
return err
|
|
}
|
|
|
|
b.db = d
|
|
|
|
var depositCache cache.DepositCache
|
|
if features.Get().EnableEIP4881 {
|
|
depositCache, err = depositsnapshot.New()
|
|
} else {
|
|
depositCache, err = depositcache.New()
|
|
}
|
|
if err != nil {
|
|
return errors.Wrap(err, "could not create deposit cache")
|
|
}
|
|
|
|
b.depositCache = depositCache
|
|
|
|
if b.GenesisInitializer != nil {
|
|
if err := b.GenesisInitializer.Initialize(b.ctx, d); err != nil {
|
|
if err == db.ErrExistingGenesisState {
|
|
return errors.New("Genesis state flag specified but a genesis state " +
|
|
"exists already. Run again with --clear-db and/or ensure you are using the " +
|
|
"appropriate testnet flag to load the given genesis state.")
|
|
}
|
|
return errors.Wrap(err, "could not load genesis from file")
|
|
}
|
|
}
|
|
|
|
if err := b.db.EnsureEmbeddedGenesis(b.ctx); err != nil {
|
|
return err
|
|
}
|
|
|
|
if b.CheckpointInitializer != nil {
|
|
if err := b.CheckpointInitializer.Initialize(b.ctx, d); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
|
|
knownContract, err := b.db.DepositContractAddress(b.ctx)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
addr := common.HexToAddress(depositAddress)
|
|
if len(knownContract) == 0 {
|
|
if err := b.db.SaveDepositContractAddress(b.ctx, addr); err != nil {
|
|
return errors.Wrap(err, "could not save deposit contract")
|
|
}
|
|
}
|
|
if len(knownContract) > 0 && !bytes.Equal(addr.Bytes(), knownContract) {
|
|
return fmt.Errorf("database contract is %#x but tried to run with %#x. This likely means "+
|
|
"you are trying to run on a different network than what the database contains. You can run once with "+
|
|
"'--clear-db' to wipe the old database or use an alternative data directory with '--datadir'",
|
|
knownContract, addr.Bytes())
|
|
}
|
|
log.Infof("Deposit contract: %#x", addr.Bytes())
|
|
return nil
|
|
}
|
|
|
|
func (b *BeaconNode) startSlasherDB(cliCtx *cli.Context) error {
|
|
if !features.Get().EnableSlasher {
|
|
return nil
|
|
}
|
|
baseDir := cliCtx.String(cmd.DataDirFlag.Name)
|
|
|
|
if cliCtx.IsSet(flags.SlasherDirFlag.Name) {
|
|
baseDir = cliCtx.String(flags.SlasherDirFlag.Name)
|
|
}
|
|
|
|
dbPath := filepath.Join(baseDir, kv.BeaconNodeDbDirName)
|
|
clearDB := cliCtx.Bool(cmd.ClearDB.Name)
|
|
forceClearDB := cliCtx.Bool(cmd.ForceClearDB.Name)
|
|
|
|
log.WithField("databasePath", dbPath).Info("Checking DB")
|
|
|
|
d, err := slasherkv.NewKVStore(b.ctx, dbPath)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
clearDBConfirmed := false
|
|
if clearDB && !forceClearDB {
|
|
actionText := "This will delete your beacon chain database stored in your data directory. " +
|
|
"Your database backups will not be removed - do you want to proceed? (Y/N)"
|
|
deniedText := "Database will not be deleted. No changes have been made."
|
|
clearDBConfirmed, err = cmd.ConfirmAction(actionText, deniedText)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
}
|
|
if clearDBConfirmed || forceClearDB {
|
|
log.Warning("Removing database")
|
|
if err := d.ClearDB(); err != nil {
|
|
return errors.Wrap(err, "could not clear database")
|
|
}
|
|
if err := b.BlobStorage.Clear(); err != nil {
|
|
return errors.Wrap(err, "could not clear blob storage")
|
|
}
|
|
d, err = slasherkv.NewKVStore(b.ctx, dbPath)
|
|
if err != nil {
|
|
return errors.Wrap(err, "could not create new database")
|
|
}
|
|
}
|
|
|
|
b.slasherDB = d
|
|
return nil
|
|
}
|
|
|
|
func (b *BeaconNode) startStateGen(ctx context.Context, bfs coverage.AvailableBlocker, fc forkchoice.ForkChoicer) error {
|
|
opts := []stategen.Option{stategen.WithAvailableBlocker(bfs)}
|
|
sg := stategen.New(b.db, fc, opts...)
|
|
|
|
cp, err := b.db.FinalizedCheckpoint(ctx)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
r := bytesutil.ToBytes32(cp.Root)
|
|
// Consider edge case where finalized root are zeros instead of genesis root hash.
|
|
if r == params.BeaconConfig().ZeroHash {
|
|
genesisBlock, err := b.db.GenesisBlock(ctx)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
if genesisBlock != nil && !genesisBlock.IsNil() {
|
|
r, err = genesisBlock.Block().HashTreeRoot()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
}
|
|
}
|
|
|
|
b.finalizedStateAtStartUp, err = sg.StateByRoot(ctx, r)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
b.stateGen = sg
|
|
return nil
|
|
}
|
|
|
|
func (b *BeaconNode) registerP2P(cliCtx *cli.Context) error {
|
|
bootstrapNodeAddrs, dataDir, err := registration.P2PPreregistration(cliCtx)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
svc, err := p2p.NewService(b.ctx, &p2p.Config{
|
|
NoDiscovery: cliCtx.Bool(cmd.NoDiscovery.Name),
|
|
StaticPeers: slice.SplitCommaSeparated(cliCtx.StringSlice(cmd.StaticPeers.Name)),
|
|
BootstrapNodeAddr: bootstrapNodeAddrs,
|
|
RelayNodeAddr: cliCtx.String(cmd.RelayNode.Name),
|
|
DataDir: dataDir,
|
|
LocalIP: cliCtx.String(cmd.P2PIP.Name),
|
|
HostAddress: cliCtx.String(cmd.P2PHost.Name),
|
|
HostDNS: cliCtx.String(cmd.P2PHostDNS.Name),
|
|
PrivateKey: cliCtx.String(cmd.P2PPrivKey.Name),
|
|
StaticPeerID: cliCtx.Bool(cmd.P2PStaticID.Name),
|
|
MetaDataDir: cliCtx.String(cmd.P2PMetadata.Name),
|
|
TCPPort: cliCtx.Uint(cmd.P2PTCPPort.Name),
|
|
UDPPort: cliCtx.Uint(cmd.P2PUDPPort.Name),
|
|
MaxPeers: cliCtx.Uint(cmd.P2PMaxPeers.Name),
|
|
QueueSize: cliCtx.Uint(cmd.PubsubQueueSize.Name),
|
|
AllowListCIDR: cliCtx.String(cmd.P2PAllowList.Name),
|
|
DenyListCIDR: slice.SplitCommaSeparated(cliCtx.StringSlice(cmd.P2PDenyList.Name)),
|
|
EnableUPnP: cliCtx.Bool(cmd.EnableUPnPFlag.Name),
|
|
StateNotifier: b,
|
|
DB: b.db,
|
|
ClockWaiter: b.clockWaiter,
|
|
})
|
|
if err != nil {
|
|
return err
|
|
}
|
|
return b.services.RegisterService(svc)
|
|
}
|
|
|
|
func (b *BeaconNode) fetchP2P() p2p.P2P {
|
|
var p *p2p.Service
|
|
if err := b.services.FetchService(&p); err != nil {
|
|
panic(err)
|
|
}
|
|
return p
|
|
}
|
|
|
|
func (b *BeaconNode) fetchBuilderService() *builder.Service {
|
|
var s *builder.Service
|
|
if err := b.services.FetchService(&s); err != nil {
|
|
panic(err)
|
|
}
|
|
return s
|
|
}
|
|
|
|
func (b *BeaconNode) registerAttestationPool() error {
|
|
s, err := attestations.NewService(b.ctx, &attestations.Config{
|
|
Pool: b.attestationPool,
|
|
InitialSyncComplete: b.initialSyncComplete,
|
|
})
|
|
if err != nil {
|
|
return errors.Wrap(err, "could not register atts pool service")
|
|
}
|
|
return b.services.RegisterService(s)
|
|
}
|
|
|
|
func (b *BeaconNode) registerBlockchainService(fc forkchoice.ForkChoicer, gs *startup.ClockSynchronizer, syncComplete chan struct{}) error {
|
|
var web3Service *execution.Service
|
|
if err := b.services.FetchService(&web3Service); err != nil {
|
|
return err
|
|
}
|
|
|
|
var attService *attestations.Service
|
|
if err := b.services.FetchService(&attService); err != nil {
|
|
return err
|
|
}
|
|
|
|
// skipcq: CRT-D0001
|
|
opts := append(
|
|
b.serviceFlagOpts.blockchainFlagOpts,
|
|
blockchain.WithForkChoiceStore(fc),
|
|
blockchain.WithDatabase(b.db),
|
|
blockchain.WithDepositCache(b.depositCache),
|
|
blockchain.WithChainStartFetcher(web3Service),
|
|
blockchain.WithExecutionEngineCaller(web3Service),
|
|
blockchain.WithAttestationPool(b.attestationPool),
|
|
blockchain.WithExitPool(b.exitPool),
|
|
blockchain.WithSlashingPool(b.slashingsPool),
|
|
blockchain.WithBLSToExecPool(b.blsToExecPool),
|
|
blockchain.WithP2PBroadcaster(b.fetchP2P()),
|
|
blockchain.WithStateNotifier(b),
|
|
blockchain.WithAttestationService(attService),
|
|
blockchain.WithStateGen(b.stateGen),
|
|
blockchain.WithSlasherAttestationsFeed(b.slasherAttestationsFeed),
|
|
blockchain.WithFinalizedStateAtStartUp(b.finalizedStateAtStartUp),
|
|
blockchain.WithClockSynchronizer(gs),
|
|
blockchain.WithSyncComplete(syncComplete),
|
|
blockchain.WithBlobStorage(b.BlobStorage),
|
|
blockchain.WithTrackedValidatorsCache(b.trackedValidatorsCache),
|
|
blockchain.WithPayloadIDCache(b.payloadIDCache),
|
|
blockchain.WithSyncChecker(b.syncChecker),
|
|
)
|
|
|
|
blockchainService, err := blockchain.NewService(b.ctx, opts...)
|
|
if err != nil {
|
|
return errors.Wrap(err, "could not register blockchain service")
|
|
}
|
|
return b.services.RegisterService(blockchainService)
|
|
}
|
|
|
|
func (b *BeaconNode) registerPOWChainService() error {
|
|
if b.cliCtx.Bool(testSkipPowFlag) {
|
|
return b.services.RegisterService(&execution.Service{})
|
|
}
|
|
bs, err := execution.NewPowchainCollector(b.ctx)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
depositContractAddr, err := execution.DepositContractAddress()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// skipcq: CRT-D0001
|
|
opts := append(
|
|
b.serviceFlagOpts.executionChainFlagOpts,
|
|
execution.WithDepositContractAddress(common.HexToAddress(depositContractAddr)),
|
|
execution.WithDatabase(b.db),
|
|
execution.WithDepositCache(b.depositCache),
|
|
execution.WithStateNotifier(b),
|
|
execution.WithStateGen(b.stateGen),
|
|
execution.WithBeaconNodeStatsUpdater(bs),
|
|
execution.WithFinalizedStateAtStartup(b.finalizedStateAtStartUp),
|
|
execution.WithJwtId(b.cliCtx.String(flags.JwtId.Name)),
|
|
)
|
|
web3Service, err := execution.NewService(b.ctx, opts...)
|
|
if err != nil {
|
|
return errors.Wrap(err, "could not register proof-of-work chain web3Service")
|
|
}
|
|
|
|
return b.services.RegisterService(web3Service)
|
|
}
|
|
|
|
func (b *BeaconNode) registerSyncService(initialSyncComplete chan struct{}, bFillStore *backfill.Store) error {
|
|
var web3Service *execution.Service
|
|
if err := b.services.FetchService(&web3Service); err != nil {
|
|
return err
|
|
}
|
|
|
|
var chainService *blockchain.Service
|
|
if err := b.services.FetchService(&chainService); err != nil {
|
|
return err
|
|
}
|
|
|
|
var initSync *initialsync.Service
|
|
if err := b.services.FetchService(&initSync); err != nil {
|
|
return err
|
|
}
|
|
|
|
rs := regularsync.NewService(
|
|
b.ctx,
|
|
regularsync.WithDatabase(b.db),
|
|
regularsync.WithP2P(b.fetchP2P()),
|
|
regularsync.WithChainService(chainService),
|
|
regularsync.WithInitialSync(initSync),
|
|
regularsync.WithBlockNotifier(b),
|
|
regularsync.WithAttestationNotifier(b),
|
|
regularsync.WithOperationNotifier(b),
|
|
regularsync.WithAttestationPool(b.attestationPool),
|
|
regularsync.WithExitPool(b.exitPool),
|
|
regularsync.WithSlashingPool(b.slashingsPool),
|
|
regularsync.WithSyncCommsPool(b.syncCommitteePool),
|
|
regularsync.WithBlsToExecPool(b.blsToExecPool),
|
|
regularsync.WithStateGen(b.stateGen),
|
|
regularsync.WithSlasherAttestationsFeed(b.slasherAttestationsFeed),
|
|
regularsync.WithSlasherBlockHeadersFeed(b.slasherBlockHeadersFeed),
|
|
regularsync.WithPayloadReconstructor(web3Service),
|
|
regularsync.WithClockWaiter(b.clockWaiter),
|
|
regularsync.WithInitialSyncComplete(initialSyncComplete),
|
|
regularsync.WithStateNotifier(b),
|
|
regularsync.WithBlobStorage(b.BlobStorage),
|
|
regularsync.WithVerifierWaiter(b.verifyInitWaiter),
|
|
regularsync.WithAvailableBlocker(bFillStore),
|
|
)
|
|
return b.services.RegisterService(rs)
|
|
}
|
|
|
|
func (b *BeaconNode) registerInitialSyncService(complete chan struct{}) error {
|
|
var chainService *blockchain.Service
|
|
if err := b.services.FetchService(&chainService); err != nil {
|
|
return err
|
|
}
|
|
|
|
opts := []initialsync.Option{
|
|
initialsync.WithVerifierWaiter(b.verifyInitWaiter),
|
|
initialsync.WithSyncChecker(b.syncChecker),
|
|
}
|
|
is := initialsync.NewService(b.ctx, &initialsync.Config{
|
|
DB: b.db,
|
|
Chain: chainService,
|
|
P2P: b.fetchP2P(),
|
|
StateNotifier: b,
|
|
BlockNotifier: b,
|
|
ClockWaiter: b.clockWaiter,
|
|
InitialSyncComplete: complete,
|
|
BlobStorage: b.BlobStorage,
|
|
}, opts...)
|
|
return b.services.RegisterService(is)
|
|
}
|
|
|
|
func (b *BeaconNode) registerSlasherService() error {
|
|
if !features.Get().EnableSlasher {
|
|
return nil
|
|
}
|
|
var chainService *blockchain.Service
|
|
if err := b.services.FetchService(&chainService); err != nil {
|
|
return err
|
|
}
|
|
var syncService *initialsync.Service
|
|
if err := b.services.FetchService(&syncService); err != nil {
|
|
return err
|
|
}
|
|
|
|
slasherSrv, err := slasher.New(b.ctx, &slasher.ServiceConfig{
|
|
IndexedAttestationsFeed: b.slasherAttestationsFeed,
|
|
BeaconBlockHeadersFeed: b.slasherBlockHeadersFeed,
|
|
Database: b.slasherDB,
|
|
StateNotifier: b,
|
|
AttestationStateFetcher: chainService,
|
|
StateGen: b.stateGen,
|
|
SlashingPoolInserter: b.slashingsPool,
|
|
SyncChecker: syncService,
|
|
HeadStateFetcher: chainService,
|
|
ClockWaiter: b.clockWaiter,
|
|
})
|
|
if err != nil {
|
|
return err
|
|
}
|
|
return b.services.RegisterService(slasherSrv)
|
|
}
|
|
|
|
func (b *BeaconNode) registerRPCService(router *mux.Router) error {
|
|
var chainService *blockchain.Service
|
|
if err := b.services.FetchService(&chainService); err != nil {
|
|
return err
|
|
}
|
|
|
|
var web3Service *execution.Service
|
|
if err := b.services.FetchService(&web3Service); err != nil {
|
|
return err
|
|
}
|
|
|
|
var syncService *initialsync.Service
|
|
if err := b.services.FetchService(&syncService); err != nil {
|
|
return err
|
|
}
|
|
|
|
var slasherService *slasher.Service
|
|
if features.Get().EnableSlasher {
|
|
if err := b.services.FetchService(&slasherService); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
|
|
genesisValidators := b.cliCtx.Uint64(flags.InteropNumValidatorsFlag.Name)
|
|
var depositFetcher cache.DepositFetcher
|
|
var chainStartFetcher execution.ChainStartFetcher
|
|
if genesisValidators > 0 {
|
|
var interopService *interopcoldstart.Service
|
|
if err := b.services.FetchService(&interopService); err != nil {
|
|
return err
|
|
}
|
|
depositFetcher = interopService
|
|
chainStartFetcher = interopService
|
|
} else {
|
|
depositFetcher = b.depositCache
|
|
chainStartFetcher = web3Service
|
|
}
|
|
|
|
host := b.cliCtx.String(flags.RPCHost.Name)
|
|
port := b.cliCtx.String(flags.RPCPort.Name)
|
|
beaconMonitoringHost := b.cliCtx.String(cmd.MonitoringHostFlag.Name)
|
|
beaconMonitoringPort := b.cliCtx.Int(flags.MonitoringPortFlag.Name)
|
|
cert := b.cliCtx.String(flags.CertFlag.Name)
|
|
key := b.cliCtx.String(flags.KeyFlag.Name)
|
|
mockEth1DataVotes := b.cliCtx.Bool(flags.InteropMockEth1DataVotesFlag.Name)
|
|
|
|
maxMsgSize := b.cliCtx.Int(cmd.GrpcMaxCallRecvMsgSizeFlag.Name)
|
|
enableDebugRPCEndpoints := b.cliCtx.Bool(flags.EnableDebugRPCEndpoints.Name)
|
|
|
|
p2pService := b.fetchP2P()
|
|
rpcService := rpc.NewService(b.ctx, &rpc.Config{
|
|
ExecutionEngineCaller: web3Service,
|
|
ExecutionPayloadReconstructor: web3Service,
|
|
Host: host,
|
|
Port: port,
|
|
BeaconMonitoringHost: beaconMonitoringHost,
|
|
BeaconMonitoringPort: beaconMonitoringPort,
|
|
CertFlag: cert,
|
|
KeyFlag: key,
|
|
BeaconDB: b.db,
|
|
Broadcaster: p2pService,
|
|
PeersFetcher: p2pService,
|
|
PeerManager: p2pService,
|
|
MetadataProvider: p2pService,
|
|
ChainInfoFetcher: chainService,
|
|
HeadFetcher: chainService,
|
|
CanonicalFetcher: chainService,
|
|
ForkFetcher: chainService,
|
|
ForkchoiceFetcher: chainService,
|
|
FinalizationFetcher: chainService,
|
|
BlockReceiver: chainService,
|
|
BlobReceiver: chainService,
|
|
AttestationReceiver: chainService,
|
|
GenesisTimeFetcher: chainService,
|
|
GenesisFetcher: chainService,
|
|
OptimisticModeFetcher: chainService,
|
|
AttestationsPool: b.attestationPool,
|
|
ExitPool: b.exitPool,
|
|
SlashingsPool: b.slashingsPool,
|
|
BLSChangesPool: b.blsToExecPool,
|
|
SyncCommitteeObjectPool: b.syncCommitteePool,
|
|
ExecutionChainService: web3Service,
|
|
ExecutionChainInfoFetcher: web3Service,
|
|
ChainStartFetcher: chainStartFetcher,
|
|
MockEth1Votes: mockEth1DataVotes,
|
|
SyncService: syncService,
|
|
DepositFetcher: depositFetcher,
|
|
PendingDepositFetcher: b.depositCache,
|
|
BlockNotifier: b,
|
|
StateNotifier: b,
|
|
OperationNotifier: b,
|
|
StateGen: b.stateGen,
|
|
EnableDebugRPCEndpoints: enableDebugRPCEndpoints,
|
|
MaxMsgSize: maxMsgSize,
|
|
BlockBuilder: b.fetchBuilderService(),
|
|
Router: router,
|
|
ClockWaiter: b.clockWaiter,
|
|
BlobStorage: b.BlobStorage,
|
|
TrackedValidatorsCache: b.trackedValidatorsCache,
|
|
PayloadIDCache: b.payloadIDCache,
|
|
})
|
|
|
|
return b.services.RegisterService(rpcService)
|
|
}
|
|
|
|
func (b *BeaconNode) registerPrometheusService(_ *cli.Context) error {
|
|
var additionalHandlers []prometheus.Handler
|
|
var p *p2p.Service
|
|
if err := b.services.FetchService(&p); err != nil {
|
|
panic(err)
|
|
}
|
|
additionalHandlers = append(additionalHandlers, prometheus.Handler{Path: "/p2p", Handler: p.InfoHandler})
|
|
|
|
var c *blockchain.Service
|
|
if err := b.services.FetchService(&c); err != nil {
|
|
panic(err)
|
|
}
|
|
|
|
service := prometheus.NewService(
|
|
fmt.Sprintf("%s:%d", b.cliCtx.String(cmd.MonitoringHostFlag.Name), b.cliCtx.Int(flags.MonitoringPortFlag.Name)),
|
|
b.services,
|
|
additionalHandlers...,
|
|
)
|
|
hook := prometheus.NewLogrusCollector()
|
|
logrus.AddHook(hook)
|
|
return b.services.RegisterService(service)
|
|
}
|
|
|
|
func (b *BeaconNode) registerGRPCGateway(router *mux.Router) error {
|
|
if b.cliCtx.Bool(flags.DisableGRPCGateway.Name) {
|
|
return nil
|
|
}
|
|
gatewayPort := b.cliCtx.Int(flags.GRPCGatewayPort.Name)
|
|
gatewayHost := b.cliCtx.String(flags.GRPCGatewayHost.Name)
|
|
rpcHost := b.cliCtx.String(flags.RPCHost.Name)
|
|
selfAddress := fmt.Sprintf("%s:%d", rpcHost, b.cliCtx.Int(flags.RPCPort.Name))
|
|
gatewayAddress := fmt.Sprintf("%s:%d", gatewayHost, gatewayPort)
|
|
allowedOrigins := strings.Split(b.cliCtx.String(flags.GPRCGatewayCorsDomain.Name), ",")
|
|
enableDebugRPCEndpoints := b.cliCtx.Bool(flags.EnableDebugRPCEndpoints.Name)
|
|
selfCert := b.cliCtx.String(flags.CertFlag.Name)
|
|
maxCallSize := b.cliCtx.Uint64(cmd.GrpcMaxCallRecvMsgSizeFlag.Name)
|
|
httpModules := b.cliCtx.String(flags.HTTPModules.Name)
|
|
timeout := b.cliCtx.Int(cmd.ApiTimeoutFlag.Name)
|
|
|
|
gatewayConfig := gateway.DefaultConfig(enableDebugRPCEndpoints, httpModules)
|
|
muxs := make([]*apigateway.PbMux, 0)
|
|
if gatewayConfig.V1AlphaPbMux != nil {
|
|
muxs = append(muxs, gatewayConfig.V1AlphaPbMux)
|
|
}
|
|
if gatewayConfig.EthPbMux != nil {
|
|
muxs = append(muxs, gatewayConfig.EthPbMux)
|
|
}
|
|
|
|
opts := []apigateway.Option{
|
|
apigateway.WithRouter(router),
|
|
apigateway.WithGatewayAddr(gatewayAddress),
|
|
apigateway.WithRemoteAddr(selfAddress),
|
|
apigateway.WithPbHandlers(muxs),
|
|
apigateway.WithRemoteCert(selfCert),
|
|
apigateway.WithMaxCallRecvMsgSize(maxCallSize),
|
|
apigateway.WithAllowedOrigins(allowedOrigins),
|
|
apigateway.WithTimeout(uint64(timeout)),
|
|
}
|
|
g, err := apigateway.New(b.ctx, opts...)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
return b.services.RegisterService(g)
|
|
}
|
|
|
|
func (b *BeaconNode) registerDeterministicGenesisService() error {
|
|
genesisTime := b.cliCtx.Uint64(flags.InteropGenesisTimeFlag.Name)
|
|
genesisValidators := b.cliCtx.Uint64(flags.InteropNumValidatorsFlag.Name)
|
|
|
|
if genesisValidators > 0 {
|
|
svc := interopcoldstart.NewService(b.ctx, &interopcoldstart.Config{
|
|
GenesisTime: genesisTime,
|
|
NumValidators: genesisValidators,
|
|
BeaconDB: b.db,
|
|
DepositCache: b.depositCache,
|
|
})
|
|
svc.Start()
|
|
|
|
// Register genesis state as start-up state when interop mode.
|
|
// The start-up state gets reused across services.
|
|
st, err := b.db.GenesisState(b.ctx)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
b.finalizedStateAtStartUp = st
|
|
|
|
return b.services.RegisterService(svc)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (b *BeaconNode) registerValidatorMonitorService(initialSyncComplete chan struct{}) error {
|
|
cliSlice := b.cliCtx.IntSlice(cmd.ValidatorMonitorIndicesFlag.Name)
|
|
if cliSlice == nil {
|
|
return nil
|
|
}
|
|
tracked := make([]primitives.ValidatorIndex, len(cliSlice))
|
|
for i := range tracked {
|
|
tracked[i] = primitives.ValidatorIndex(cliSlice[i])
|
|
}
|
|
|
|
var chainService *blockchain.Service
|
|
if err := b.services.FetchService(&chainService); err != nil {
|
|
return err
|
|
}
|
|
monitorConfig := &monitor.ValidatorMonitorConfig{
|
|
StateNotifier: b,
|
|
AttestationNotifier: b,
|
|
StateGen: b.stateGen,
|
|
HeadFetcher: chainService,
|
|
InitialSyncComplete: initialSyncComplete,
|
|
}
|
|
svc, err := monitor.NewService(b.ctx, monitorConfig, tracked)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
return b.services.RegisterService(svc)
|
|
}
|
|
|
|
func (b *BeaconNode) registerBuilderService(cliCtx *cli.Context) error {
|
|
var chainService *blockchain.Service
|
|
if err := b.services.FetchService(&chainService); err != nil {
|
|
return err
|
|
}
|
|
|
|
opts := append(b.serviceFlagOpts.builderOpts,
|
|
builder.WithHeadFetcher(chainService),
|
|
builder.WithDatabase(b.db))
|
|
// make cache the default.
|
|
if !cliCtx.Bool(features.DisableRegistrationCache.Name) {
|
|
opts = append(opts, builder.WithRegistrationCache())
|
|
}
|
|
svc, err := builder.NewService(b.ctx, opts...)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
return b.services.RegisterService(svc)
|
|
}
|
|
|
|
func hasNetworkFlag(cliCtx *cli.Context) bool {
|
|
for _, flag := range features.NetworkFlags {
|
|
for _, name := range flag.Names() {
|
|
if cliCtx.IsSet(name) {
|
|
return true
|
|
}
|
|
}
|
|
}
|
|
return false
|
|
}
|