mirror of
https://gitlab.com/pulsechaincom/prysm-pulse.git
synced 2024-12-22 03:30:35 +00:00
d6ae838bbf
* WIP * event stream wip * returning nil * temp removing some tests * wip health checks * fixing conficts * updating fields based on linting * fixing more errors * fixing mocks * fixing more mocks * fixing more linting * removing white space for lint * fixing log format * gaz * reverting changes on grpc * fixing unit tests * adding in tests for health tracker and event stream * adding more tests for streaming slot * gaz * Update api/client/event/event_stream.go Co-authored-by: Radosław Kapka <rkapka@wp.pl> * review comments * Update validator/client/runner.go Co-authored-by: Radosław Kapka <rkapka@wp.pl> * Update validator/client/validator.go Co-authored-by: Radosław Kapka <rkapka@wp.pl> * Update validator/client/validator.go Co-authored-by: Radosław Kapka <rkapka@wp.pl> * Update validator/client/validator.go Co-authored-by: Radosław Kapka <rkapka@wp.pl> * Update validator/client/validator.go Co-authored-by: Radosław Kapka <rkapka@wp.pl> * Update validator/client/beacon-api/beacon_api_validator_client.go Co-authored-by: Radosław Kapka <rkapka@wp.pl> * Update validator/client/validator.go Co-authored-by: Radosław Kapka <rkapka@wp.pl> * Update validator/client/validator.go Co-authored-by: Radosław Kapka <rkapka@wp.pl> * addressing radek comments * Update validator/client/validator.go Co-authored-by: Radosław Kapka <rkapka@wp.pl> * addressing review feedback * moving things to below next slot ticker * fixing tests * update naming * adding TODO comment * Update api/client/beacon/health.go Co-authored-by: Radosław Kapka <rkapka@wp.pl> * addressing comments * fixing broken linting * fixing more import issues * fixing more import issues * linting * updating based on radek's comments * addressing more comments * fixing nogo error * fixing duplicate import * gaz * adding radek's review suggestion * Update proto/prysm/v1alpha1/node.proto Co-authored-by: Preston Van Loon <pvanloon@offchainlabs.com> * preston review comments * Update api/client/event/event_stream.go Co-authored-by: Preston Van Loon <pvanloon@offchainlabs.com> * Update validator/client/validator.go Co-authored-by: Preston Van Loon <pvanloon@offchainlabs.com> * addressing some more preston review items * fixing tests for linting * fixing missed linting * updating based on feedback to simplify * adding interface check at the top * reverting some comments * cleaning up intatiations * reworking the health tracker * fixing linting * fixing more linting to adhear to interface * adding interface check at the the top of the file * fixing unit tests * attempting to fix dependency cycle * addressing radek's comment * Update validator/client/beacon-api/beacon_api_validator_client.go Co-authored-by: Preston Van Loon <pvanloon@offchainlabs.com> * adding more tests and feedback items * fixing TODO comment --------- Co-authored-by: Radosław Kapka <rkapka@wp.pl> Co-authored-by: Preston Van Loon <pvanloon@offchainlabs.com>
316 lines
10 KiB
Go
316 lines
10 KiB
Go
package client
|
|
|
|
import (
|
|
"context"
|
|
"fmt"
|
|
"sync"
|
|
"time"
|
|
|
|
"github.com/pkg/errors"
|
|
"github.com/prysmaticlabs/prysm/v5/api/client"
|
|
"github.com/prysmaticlabs/prysm/v5/api/client/event"
|
|
fieldparams "github.com/prysmaticlabs/prysm/v5/config/fieldparams"
|
|
"github.com/prysmaticlabs/prysm/v5/config/params"
|
|
"github.com/prysmaticlabs/prysm/v5/consensus-types/primitives"
|
|
"github.com/prysmaticlabs/prysm/v5/encoding/bytesutil"
|
|
"github.com/prysmaticlabs/prysm/v5/time/slots"
|
|
"github.com/prysmaticlabs/prysm/v5/validator/client/iface"
|
|
"go.opencensus.io/trace"
|
|
"google.golang.org/grpc/codes"
|
|
"google.golang.org/grpc/status"
|
|
)
|
|
|
|
// Time to wait before trying to reconnect with beacon node.
|
|
var backOffPeriod = 10 * time.Second
|
|
|
|
// Run the main validator routine. This routine exits if the context is
|
|
// canceled.
|
|
//
|
|
// Order of operations:
|
|
// 1 - Initialize validator data
|
|
// 2 - Wait for validator activation
|
|
// 3 - Wait for the next slot start
|
|
// 4 - Update assignments
|
|
// 5 - Determine role at current slot
|
|
// 6 - Perform assigned role, if any
|
|
func run(ctx context.Context, v iface.Validator) {
|
|
cleanup := v.Done
|
|
defer cleanup()
|
|
|
|
headSlot, err := initializeValidatorAndGetHeadSlot(ctx, v)
|
|
if err != nil {
|
|
return // Exit if context is canceled.
|
|
}
|
|
if err := v.UpdateDuties(ctx, headSlot); err != nil {
|
|
handleAssignmentError(err, headSlot)
|
|
}
|
|
eventsChan := make(chan *event.Event, 1)
|
|
healthTracker := v.HealthTracker()
|
|
runHealthCheckRoutine(ctx, v, eventsChan)
|
|
|
|
accountsChangedChan := make(chan [][fieldparams.BLSPubkeyLength]byte, 1)
|
|
km, err := v.Keymanager()
|
|
if err != nil {
|
|
log.WithError(err).Fatal("Could not get keymanager")
|
|
}
|
|
sub := km.SubscribeAccountChanges(accountsChangedChan)
|
|
// check if proposer settings is still nil
|
|
// Set properties on the beacon node like the fee recipient for validators that are being used & active.
|
|
if v.ProposerSettings() == nil {
|
|
log.Warn("Validator client started without proposer settings such as fee recipient" +
|
|
" and will continue to use settings provided in the beacon node.")
|
|
}
|
|
deadline := time.Now().Add(5 * time.Minute)
|
|
if err := v.PushProposerSettings(ctx, km, headSlot, deadline); err != nil {
|
|
if errors.Is(err, ErrBuilderValidatorRegistration) {
|
|
log.WithError(err).Warn("Push proposer settings error")
|
|
} else {
|
|
log.WithError(err).Fatal("Failed to update proposer settings") // allow fatal. skipcq
|
|
}
|
|
}
|
|
for {
|
|
ctx, span := trace.StartSpan(ctx, "validator.processSlot")
|
|
select {
|
|
case <-ctx.Done():
|
|
log.Info("Context canceled, stopping validator")
|
|
span.End()
|
|
sub.Unsubscribe()
|
|
close(accountsChangedChan)
|
|
return // Exit if context is canceled.
|
|
case slot := <-v.NextSlot():
|
|
if !healthTracker.IsHealthy() {
|
|
continue
|
|
}
|
|
span.AddAttributes(trace.Int64Attribute("slot", int64(slot))) // lint:ignore uintcast -- This conversion is OK for tracing.
|
|
|
|
deadline := v.SlotDeadline(slot)
|
|
slotCtx, cancel := context.WithDeadline(ctx, deadline)
|
|
log := log.WithField("slot", slot)
|
|
log.WithField("deadline", deadline).Debug("Set deadline for proposals and attestations")
|
|
|
|
// Keep trying to update assignments if they are nil or if we are past an
|
|
// epoch transition in the beacon node's state.
|
|
if err := v.UpdateDuties(ctx, slot); err != nil {
|
|
handleAssignmentError(err, slot)
|
|
cancel()
|
|
span.End()
|
|
continue
|
|
}
|
|
|
|
// call push proposer setting at the start of each epoch to account for the following edge case:
|
|
// proposer is activated at the start of epoch and tries to propose immediately
|
|
if slots.IsEpochStart(slot) {
|
|
go func() {
|
|
// deadline set for 1 epoch from call to not overlap.
|
|
epochDeadline := v.SlotDeadline(slot + params.BeaconConfig().SlotsPerEpoch - 1)
|
|
if err := v.PushProposerSettings(ctx, km, slot, epochDeadline); err != nil {
|
|
log.WithError(err).Warn("Failed to update proposer settings")
|
|
}
|
|
}()
|
|
}
|
|
|
|
// Start fetching domain data for the next epoch.
|
|
if slots.IsEpochEnd(slot) {
|
|
go v.UpdateDomainDataCaches(ctx, slot+1)
|
|
}
|
|
|
|
var wg sync.WaitGroup
|
|
|
|
allRoles, err := v.RolesAt(ctx, slot)
|
|
if err != nil {
|
|
log.WithError(err).Error("Could not get validator roles")
|
|
cancel()
|
|
span.End()
|
|
continue
|
|
}
|
|
performRoles(slotCtx, allRoles, v, slot, &wg, span)
|
|
case isHealthyAgain := <-healthTracker.HealthUpdates():
|
|
if isHealthyAgain {
|
|
headSlot, err = initializeValidatorAndGetHeadSlot(ctx, v)
|
|
if err != nil {
|
|
log.WithError(err).Error("Failed to re initialize validator and get head slot")
|
|
continue
|
|
}
|
|
if err := v.UpdateDuties(ctx, headSlot); err != nil {
|
|
handleAssignmentError(err, headSlot)
|
|
continue
|
|
}
|
|
}
|
|
case e := <-eventsChan:
|
|
v.ProcessEvent(e)
|
|
case currentKeys := <-accountsChangedChan: // should be less of a priority than next slot
|
|
onAccountsChanged(ctx, v, currentKeys, accountsChangedChan)
|
|
}
|
|
}
|
|
}
|
|
|
|
func onAccountsChanged(ctx context.Context, v iface.Validator, current [][48]byte, ac chan [][fieldparams.BLSPubkeyLength]byte) {
|
|
anyActive, err := v.HandleKeyReload(ctx, current)
|
|
if err != nil {
|
|
log.WithError(err).Error("Could not properly handle reloaded keys")
|
|
}
|
|
if !anyActive {
|
|
log.Warn("No active keys found. Waiting for activation...")
|
|
err := v.WaitForActivation(ctx, ac)
|
|
if err != nil {
|
|
log.WithError(err).Warn("Could not wait for validator activation")
|
|
}
|
|
}
|
|
}
|
|
|
|
func initializeValidatorAndGetHeadSlot(ctx context.Context, v iface.Validator) (primitives.Slot, error) {
|
|
ticker := time.NewTicker(backOffPeriod)
|
|
defer ticker.Stop()
|
|
|
|
firstTime := true
|
|
|
|
var (
|
|
headSlot primitives.Slot
|
|
err error
|
|
)
|
|
|
|
for {
|
|
if !firstTime {
|
|
if ctx.Err() != nil {
|
|
log.Info("Context canceled, stopping validator")
|
|
return headSlot, errors.New("context canceled")
|
|
}
|
|
<-ticker.C
|
|
}
|
|
|
|
firstTime = false
|
|
|
|
if err := v.WaitForChainStart(ctx); err != nil {
|
|
if isConnectionError(err) {
|
|
log.WithError(err).Warn("Could not determine if beacon chain started")
|
|
continue
|
|
}
|
|
|
|
log.WithError(err).Fatal("Could not determine if beacon chain started")
|
|
}
|
|
|
|
if err := v.WaitForKeymanagerInitialization(ctx); err != nil {
|
|
// log.Fatal will prevent defer from being called
|
|
v.Done()
|
|
log.WithError(err).Fatal("Wallet is not ready")
|
|
}
|
|
|
|
if err := v.WaitForSync(ctx); err != nil {
|
|
if isConnectionError(err) {
|
|
log.WithError(err).Warn("Could not determine if beacon chain started")
|
|
continue
|
|
}
|
|
|
|
log.WithError(err).Fatal("Could not determine if beacon node synced")
|
|
}
|
|
|
|
if err := v.WaitForActivation(ctx, nil /* accountsChangedChan */); err != nil {
|
|
log.WithError(err).Fatal("Could not wait for validator activation")
|
|
}
|
|
|
|
headSlot, err = v.CanonicalHeadSlot(ctx)
|
|
if isConnectionError(err) {
|
|
log.WithError(err).Warn("Could not get current canonical head slot")
|
|
continue
|
|
}
|
|
|
|
if err != nil {
|
|
log.WithError(err).Fatal("Could not get current canonical head slot")
|
|
}
|
|
|
|
if err := v.CheckDoppelGanger(ctx); err != nil {
|
|
if isConnectionError(err) {
|
|
log.WithError(err).Warn("Could not wait for checking doppelganger")
|
|
continue
|
|
}
|
|
|
|
log.WithError(err).Fatal("Could not succeed with doppelganger check")
|
|
}
|
|
break
|
|
}
|
|
return headSlot, nil
|
|
}
|
|
|
|
func performRoles(slotCtx context.Context, allRoles map[[48]byte][]iface.ValidatorRole, v iface.Validator, slot primitives.Slot, wg *sync.WaitGroup, span *trace.Span) {
|
|
for pubKey, roles := range allRoles {
|
|
wg.Add(len(roles))
|
|
for _, role := range roles {
|
|
go func(role iface.ValidatorRole, pubKey [fieldparams.BLSPubkeyLength]byte) {
|
|
defer wg.Done()
|
|
switch role {
|
|
case iface.RoleAttester:
|
|
v.SubmitAttestation(slotCtx, slot, pubKey)
|
|
case iface.RoleProposer:
|
|
v.ProposeBlock(slotCtx, slot, pubKey)
|
|
case iface.RoleAggregator:
|
|
v.SubmitAggregateAndProof(slotCtx, slot, pubKey)
|
|
case iface.RoleSyncCommittee:
|
|
v.SubmitSyncCommitteeMessage(slotCtx, slot, pubKey)
|
|
case iface.RoleSyncCommitteeAggregator:
|
|
v.SubmitSignedContributionAndProof(slotCtx, slot, pubKey)
|
|
case iface.RoleUnknown:
|
|
log.WithField("pubkey", fmt.Sprintf("%#x", bytesutil.Trunc(pubKey[:]))).Trace("No active roles, doing nothing")
|
|
default:
|
|
log.Warnf("Unhandled role %v", role)
|
|
}
|
|
}(role, pubKey)
|
|
}
|
|
}
|
|
|
|
// Wait for all processes to complete, then report span complete.
|
|
go func() {
|
|
wg.Wait()
|
|
defer span.End()
|
|
defer func() {
|
|
if err := recover(); err != nil { // catch any panic in logging
|
|
log.WithField("error", err).
|
|
Error("Panic occurred when logging validator report. This" +
|
|
" should never happen! Please file a report at github.com/prysmaticlabs/prysm/issues/new")
|
|
}
|
|
}()
|
|
// Log performance in the previous slot
|
|
v.LogSubmittedAtts(slot)
|
|
v.LogSubmittedSyncCommitteeMessages()
|
|
if err := v.LogValidatorGainsAndLosses(slotCtx, slot); err != nil {
|
|
log.WithError(err).Error("Could not report validator's rewards/penalties")
|
|
}
|
|
}()
|
|
}
|
|
|
|
func isConnectionError(err error) bool {
|
|
return err != nil && errors.Is(err, client.ErrConnectionIssue)
|
|
}
|
|
|
|
func handleAssignmentError(err error, slot primitives.Slot) {
|
|
if errors.Is(err, ErrValidatorsAllExited) {
|
|
log.Warn(ErrValidatorsAllExited)
|
|
} else if errCode, ok := status.FromError(err); ok && errCode.Code() == codes.NotFound {
|
|
log.WithField(
|
|
"epoch", slot/params.BeaconConfig().SlotsPerEpoch,
|
|
).Warn("Validator not yet assigned to epoch")
|
|
} else {
|
|
log.WithError(err).Error("Failed to update assignments")
|
|
}
|
|
}
|
|
|
|
func runHealthCheckRoutine(ctx context.Context, v iface.Validator, eventsChan chan<- *event.Event) {
|
|
log.Info("Starting health check routine for beacon node apis")
|
|
healthCheckTicker := time.NewTicker(time.Duration(params.BeaconConfig().SecondsPerSlot) * time.Second)
|
|
tracker := v.HealthTracker()
|
|
go func() {
|
|
// trigger the healthcheck immediately the first time
|
|
for ; true; <-healthCheckTicker.C {
|
|
if ctx.Err() != nil {
|
|
log.WithError(ctx.Err()).Error("Context cancelled")
|
|
return
|
|
}
|
|
isHealthy := tracker.CheckHealth(ctx)
|
|
// in case of node returning healthy but event stream died
|
|
if isHealthy && !v.EventStreamIsRunning() {
|
|
log.Info("Event stream reconnecting...")
|
|
go v.StartEventStream(ctx, event.DefaultEventTopics, eventsChan)
|
|
}
|
|
}
|
|
}()
|
|
}
|