mirror of
https://gitlab.com/pulsechaincom/prysm-pulse.git
synced 2024-12-26 05:17:22 +00:00
Archive Data Even Through Skip Slots (#4054)
* red test first * does not archive through skip slot * test out at runtime * underflow check * fix tests * rem info log
This commit is contained in:
parent
d071a0a90a
commit
ae07dc7962
@ -6,7 +6,7 @@ import (
|
|||||||
|
|
||||||
"github.com/pkg/errors"
|
"github.com/pkg/errors"
|
||||||
"github.com/prysmaticlabs/prysm/beacon-chain/blockchain"
|
"github.com/prysmaticlabs/prysm/beacon-chain/blockchain"
|
||||||
"github.com/prysmaticlabs/prysm/beacon-chain/core/epoch"
|
epochProcessing "github.com/prysmaticlabs/prysm/beacon-chain/core/epoch"
|
||||||
"github.com/prysmaticlabs/prysm/beacon-chain/core/helpers"
|
"github.com/prysmaticlabs/prysm/beacon-chain/core/helpers"
|
||||||
"github.com/prysmaticlabs/prysm/beacon-chain/core/statefeed"
|
"github.com/prysmaticlabs/prysm/beacon-chain/core/statefeed"
|
||||||
"github.com/prysmaticlabs/prysm/beacon-chain/core/validators"
|
"github.com/prysmaticlabs/prysm/beacon-chain/core/validators"
|
||||||
@ -27,6 +27,7 @@ type Service struct {
|
|||||||
beaconDB db.Database
|
beaconDB db.Database
|
||||||
headFetcher blockchain.HeadFetcher
|
headFetcher blockchain.HeadFetcher
|
||||||
stateNotifier statefeed.Notifier
|
stateNotifier statefeed.Notifier
|
||||||
|
lastArchivedEpoch uint64
|
||||||
}
|
}
|
||||||
|
|
||||||
// Config options for the archiver service.
|
// Config options for the archiver service.
|
||||||
@ -66,13 +67,12 @@ func (s *Service) Status() error {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// We archive committee information pertaining to the head state's epoch.
|
// We archive committee information pertaining to the head state's epoch.
|
||||||
func (s *Service) archiveCommitteeInfo(ctx context.Context, headState *pb.BeaconState) error {
|
func (s *Service) archiveCommitteeInfo(ctx context.Context, headState *pb.BeaconState, epoch uint64) error {
|
||||||
currentEpoch := helpers.SlotToEpoch(headState.Slot)
|
proposerSeed, err := helpers.Seed(headState, epoch, params.BeaconConfig().DomainBeaconProposer)
|
||||||
proposerSeed, err := helpers.Seed(headState, currentEpoch, params.BeaconConfig().DomainBeaconProposer)
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return errors.Wrap(err, "could not generate seed")
|
return errors.Wrap(err, "could not generate seed")
|
||||||
}
|
}
|
||||||
attesterSeed, err := helpers.Seed(headState, currentEpoch, params.BeaconConfig().DomainBeaconAttester)
|
attesterSeed, err := helpers.Seed(headState, epoch, params.BeaconConfig().DomainBeaconAttester)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return errors.Wrap(err, "could not generate seed")
|
return errors.Wrap(err, "could not generate seed")
|
||||||
}
|
}
|
||||||
@ -81,17 +81,18 @@ func (s *Service) archiveCommitteeInfo(ctx context.Context, headState *pb.Beacon
|
|||||||
ProposerSeed: proposerSeed[:],
|
ProposerSeed: proposerSeed[:],
|
||||||
AttesterSeed: attesterSeed[:],
|
AttesterSeed: attesterSeed[:],
|
||||||
}
|
}
|
||||||
if err := s.beaconDB.SaveArchivedCommitteeInfo(ctx, currentEpoch, info); err != nil {
|
if err := s.beaconDB.SaveArchivedCommitteeInfo(ctx, epoch, info); err != nil {
|
||||||
return errors.Wrap(err, "could not archive committee info")
|
return errors.Wrap(err, "could not archive committee info")
|
||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// We archive active validator set changes that happened during the previous epoch.
|
// We archive active validator set changes that happened during the previous epoch.
|
||||||
func (s *Service) archiveActiveSetChanges(ctx context.Context, headState *pb.BeaconState) error {
|
func (s *Service) archiveActiveSetChanges(ctx context.Context, headState *pb.BeaconState, epoch uint64) error {
|
||||||
activations := validators.ActivatedValidatorIndices(helpers.PrevEpoch(headState), headState.Validators)
|
prevEpoch := epoch - 1
|
||||||
slashings := validators.SlashedValidatorIndices(helpers.PrevEpoch(headState), headState.Validators)
|
activations := validators.ActivatedValidatorIndices(prevEpoch, headState.Validators)
|
||||||
activeValidatorCount, err := helpers.ActiveValidatorCount(headState, helpers.PrevEpoch(headState))
|
slashings := validators.SlashedValidatorIndices(prevEpoch, headState.Validators)
|
||||||
|
activeValidatorCount, err := helpers.ActiveValidatorCount(headState, prevEpoch)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return errors.Wrap(err, "could not get active validator count")
|
return errors.Wrap(err, "could not get active validator count")
|
||||||
}
|
}
|
||||||
@ -104,7 +105,7 @@ func (s *Service) archiveActiveSetChanges(ctx context.Context, headState *pb.Bea
|
|||||||
Exited: exited,
|
Exited: exited,
|
||||||
Slashed: slashings,
|
Slashed: slashings,
|
||||||
}
|
}
|
||||||
if err := s.beaconDB.SaveArchivedActiveValidatorChanges(ctx, helpers.PrevEpoch(headState), activeSetChanges); err != nil {
|
if err := s.beaconDB.SaveArchivedActiveValidatorChanges(ctx, prevEpoch, activeSetChanges); err != nil {
|
||||||
return errors.Wrap(err, "could not archive active validator set changes")
|
return errors.Wrap(err, "could not archive active validator set changes")
|
||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
@ -112,19 +113,18 @@ func (s *Service) archiveActiveSetChanges(ctx context.Context, headState *pb.Bea
|
|||||||
|
|
||||||
// We compute participation metrics by first retrieving the head state and
|
// We compute participation metrics by first retrieving the head state and
|
||||||
// matching validator attestations during the epoch.
|
// matching validator attestations during the epoch.
|
||||||
func (s *Service) archiveParticipation(ctx context.Context, headState *pb.BeaconState) error {
|
func (s *Service) archiveParticipation(ctx context.Context, headState *pb.BeaconState, epoch uint64) error {
|
||||||
participation, err := epoch.ComputeValidatorParticipation(headState, helpers.SlotToEpoch(headState.Slot))
|
participation, err := epochProcessing.ComputeValidatorParticipation(headState, epoch)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return errors.Wrap(err, "could not compute participation")
|
return errors.Wrap(err, "could not compute participation")
|
||||||
}
|
}
|
||||||
return s.beaconDB.SaveArchivedValidatorParticipation(ctx, helpers.SlotToEpoch(headState.Slot), participation)
|
return s.beaconDB.SaveArchivedValidatorParticipation(ctx, epoch, participation)
|
||||||
}
|
}
|
||||||
|
|
||||||
// We archive validator balances and active indices.
|
// We archive validator balances and active indices.
|
||||||
func (s *Service) archiveBalances(ctx context.Context, headState *pb.BeaconState) error {
|
func (s *Service) archiveBalances(ctx context.Context, headState *pb.BeaconState, epoch uint64) error {
|
||||||
balances := headState.Balances
|
balances := headState.Balances
|
||||||
currentEpoch := helpers.CurrentEpoch(headState)
|
if err := s.beaconDB.SaveArchivedBalances(ctx, epoch, balances); err != nil {
|
||||||
if err := s.beaconDB.SaveArchivedBalances(ctx, currentEpoch, balances); err != nil {
|
|
||||||
return errors.Wrap(err, "could not archive balances")
|
return errors.Wrap(err, "could not archive balances")
|
||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
@ -146,29 +146,35 @@ func (s *Service) run(ctx context.Context) {
|
|||||||
log.WithError(err).Error("Head state is not available")
|
log.WithError(err).Error("Head state is not available")
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
if !helpers.IsEpochEnd(headState.Slot) {
|
currentEpoch := helpers.CurrentEpoch(headState)
|
||||||
|
if !helpers.IsEpochEnd(headState.Slot) && currentEpoch <= s.lastArchivedEpoch {
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
if err := s.archiveCommitteeInfo(ctx, headState); err != nil {
|
epochToArchive := currentEpoch
|
||||||
|
if !helpers.IsEpochEnd(headState.Slot) {
|
||||||
|
epochToArchive--
|
||||||
|
}
|
||||||
|
if err := s.archiveCommitteeInfo(ctx, headState, epochToArchive); err != nil {
|
||||||
log.WithError(err).Error("Could not archive committee info")
|
log.WithError(err).Error("Could not archive committee info")
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
if err := s.archiveActiveSetChanges(ctx, headState); err != nil {
|
if err := s.archiveActiveSetChanges(ctx, headState, epochToArchive); err != nil {
|
||||||
log.WithError(err).Error("Could not archive active validator set changes")
|
log.WithError(err).Error("Could not archive active validator set changes")
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
if err := s.archiveParticipation(ctx, headState); err != nil {
|
if err := s.archiveParticipation(ctx, headState, epochToArchive); err != nil {
|
||||||
log.WithError(err).Error("Could not archive validator participation")
|
log.WithError(err).Error("Could not archive validator participation")
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
if err := s.archiveBalances(ctx, headState); err != nil {
|
if err := s.archiveBalances(ctx, headState, epochToArchive); err != nil {
|
||||||
log.WithError(err).Error("Could not archive validator balances and active indices")
|
log.WithError(err).Error("Could not archive validator balances and active indices")
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
log.WithField(
|
log.WithField(
|
||||||
"epoch",
|
"epoch",
|
||||||
helpers.CurrentEpoch(headState),
|
epochToArchive,
|
||||||
).Debug("Successfully archived beacon chain data during epoch")
|
).Debug("Successfully archived beacon chain data during epoch")
|
||||||
|
s.lastArchivedEpoch = epochToArchive
|
||||||
}
|
}
|
||||||
case <-s.ctx.Done():
|
case <-s.ctx.Done():
|
||||||
log.Debug("Context closed, exiting goroutine")
|
log.Debug("Context closed, exiting goroutine")
|
||||||
|
@ -74,6 +74,56 @@ func TestArchiverService_OnlyArchiveAtEpochEnd(t *testing.T) {
|
|||||||
testutil.AssertLogsDoNotContain(t, hook, "Successfully archived")
|
testutil.AssertLogsDoNotContain(t, hook, "Successfully archived")
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func TestArchiverService_ArchivesEvenThroughSkipSlot(t *testing.T) {
|
||||||
|
hook := logTest.NewGlobal()
|
||||||
|
svc, beaconDB := setupService(t)
|
||||||
|
validatorCount := uint64(100)
|
||||||
|
headState := setupState(t, validatorCount)
|
||||||
|
defer dbutil.TeardownDB(t, beaconDB)
|
||||||
|
event := &statefeed.Event{
|
||||||
|
Type: statefeed.BlockProcessed,
|
||||||
|
Data: &statefeed.BlockProcessedData{
|
||||||
|
BlockRoot: [32]byte{1, 2, 3},
|
||||||
|
Verified: true,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
exitRoutine := make(chan bool)
|
||||||
|
go func() {
|
||||||
|
svc.run(svc.ctx)
|
||||||
|
<-exitRoutine
|
||||||
|
}()
|
||||||
|
|
||||||
|
// Send out an event every slot, skipping the end slot of the epoch.
|
||||||
|
for i := uint64(0); i < params.BeaconConfig().SlotsPerEpoch+1; i++ {
|
||||||
|
headState.Slot = i
|
||||||
|
svc.headFetcher = &mock.ChainService{
|
||||||
|
State: headState,
|
||||||
|
}
|
||||||
|
if helpers.IsEpochEnd(i) {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
// Send in a loop to ensure it is delivered (busy wait for the service to subscribe to the state feed).
|
||||||
|
for sent := 0; sent == 0; {
|
||||||
|
sent = svc.stateNotifier.StateFeed().Send(event)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if err := svc.Stop(); err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
exitRoutine <- true
|
||||||
|
|
||||||
|
// The context should have been canceled.
|
||||||
|
if svc.ctx.Err() != context.Canceled {
|
||||||
|
t.Error("context was not canceled")
|
||||||
|
}
|
||||||
|
|
||||||
|
testutil.AssertLogsContain(t, hook, "Received block processed event")
|
||||||
|
// Even though there was a skip slot, we should still be able to archive
|
||||||
|
// upon the next block event afterwards.
|
||||||
|
testutil.AssertLogsContain(t, hook, "Successfully archived")
|
||||||
|
}
|
||||||
|
|
||||||
func TestArchiverService_ComputesAndSavesParticipation(t *testing.T) {
|
func TestArchiverService_ComputesAndSavesParticipation(t *testing.T) {
|
||||||
hook := logTest.NewGlobal()
|
hook := logTest.NewGlobal()
|
||||||
validatorCount := uint64(100)
|
validatorCount := uint64(100)
|
||||||
@ -106,7 +156,7 @@ func TestArchiverService_ComputesAndSavesParticipation(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
if !proto.Equal(wanted, retrieved) {
|
if !proto.Equal(wanted, retrieved) {
|
||||||
t.Errorf("Wanted participation for epoch %d %v, retrieved %v", currentEpoch, wanted, retrieved)
|
t.Errorf("Wanted participation for epoch %d %v, retrieved %v", currentEpoch-1, wanted, retrieved)
|
||||||
}
|
}
|
||||||
testutil.AssertLogsContain(t, hook, "Successfully archived")
|
testutil.AssertLogsContain(t, hook, "Successfully archived")
|
||||||
}
|
}
|
||||||
@ -280,7 +330,7 @@ func TestArchiverService_SavesExitedValidatorChanges(t *testing.T) {
|
|||||||
},
|
},
|
||||||
}
|
}
|
||||||
triggerStateEvent(t, svc, event)
|
triggerStateEvent(t, svc, event)
|
||||||
|
testutil.AssertLogsContain(t, hook, "Successfully archived")
|
||||||
retrieved, err := beaconDB.ArchivedActiveValidatorChanges(svc.ctx, prevEpoch)
|
retrieved, err := beaconDB.ArchivedActiveValidatorChanges(svc.ctx, prevEpoch)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
t.Fatal(err)
|
t.Fatal(err)
|
||||||
@ -291,7 +341,6 @@ func TestArchiverService_SavesExitedValidatorChanges(t *testing.T) {
|
|||||||
if !reflect.DeepEqual(retrieved.Exited, []uint64{95}) {
|
if !reflect.DeepEqual(retrieved.Exited, []uint64{95}) {
|
||||||
t.Errorf("Wanted indices 95 exited, received %v", retrieved.Exited)
|
t.Errorf("Wanted indices 95 exited, received %v", retrieved.Exited)
|
||||||
}
|
}
|
||||||
testutil.AssertLogsContain(t, hook, "Successfully archived")
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func setupState(t *testing.T, validatorCount uint64) *pb.BeaconState {
|
func setupState(t *testing.T, validatorCount uint64) *pb.BeaconState {
|
||||||
|
Loading…
Reference in New Issue
Block a user