mirror of
https://gitlab.com/pulsechaincom/prysm-pulse.git
synced 2025-01-06 01:32:18 +00:00
cc741ed8af
* begin state service * begin on the state trie idea * created beacon state structure * add in the full clone getter * return by value instead * add all setters * new state setters are being completed * arrays roots exposed * close to finishing all these headerssss * functionality complete * added in proto benchmark test * test for compatibility * add test for compat * comments fixed * add clone * add clone * remove underlying copies * make it immutable * integrate it into chainservice * revert * wrap up comments for package * address all comments and godocs * address all comments * clone the pending attestation properly * properly clone remaining items * tests pass fixed bug * begin using it instead of head state * prevent nil pointer exceptions * begin using new struct in db * integrated new type into db package * add proper nil checks * using new state in archiver * refactored much of core * editing all the precompute functions * done with most core refactor * fixed up some bugs in the clone comparisons * append current epoch atts * add missing setters * add new setters * fix other core methods * fix up transition * main service and forkchoice * fix rpc * integrated to powchain * some more changes * fix build * improve processing of deposits * fix error * prevent panic * comment * fix process att * gaz * fix up att process * resolve existing review comments * resolve another batch of gh comments * resolve broken cpt state * revise testutil to use the new state * begin updating the state transition func to pass in more compartmentalized args * finish editing transition function to return errors * block operations pretty much done with refactor * state transition fully refactored * got epoch processing completed * fix build in fork choice * fixing more of the build * fix up broken sync package * it builds nowww it buildssss * revert registry changes * Recompute on Read (#4627) * compute on read * fix up eth1 data votes * looking into slashings bug introduced in core/ * able to advance more slots * add logging * can now sync with testnet yay * remove the leaves algorithm and other merkle imports * expose initialize unsafe funcs * Update beacon-chain/db/kv/state.go * lint Co-authored-by: Raul Jordan <raul@prysmaticlabs.com> * More Optimizations for New State (#4641) * map optimization * more optimizations * use a custom hasher * comment * block operations optimizations * Update beacon-chain/state/types.go Co-Authored-By: Raul Jordan <raul@prysmaticlabs.com> * fixed up various operations to use the validator index map access Co-authored-by: Raul Jordan <raul@prysmaticlabs.com> * archiver tests pass * fixing cache tests * cache tests passing * edited validator tests * powchain tests passing * halfway thru sync tests * more sync test fixes * add in tests for state/ * working through rpc tests * assignments tests passed * almost done with rpc/beacon tests * resolved painful validator test * fixed up even more tests * resolve tests * fix build * reduce a randao mixes copy * fixes under //beacon-chain/blockchain/... * build //beacon-chain/core/... * fixes * Runtime Optimizations (#4648) * parallelize shuffling * clean up * lint * fix build * use callback to read from registry * fix array roots and size map * new improvements * reduce hash allocs * improved shuffling * terence's review * use different method * raul's comment * new array roots * remove clone in pre-compute * Update beacon-chain/state/types.go Co-Authored-By: Raul Jordan <raul@prysmaticlabs.com> * raul's review * lint * fix build issues * fix visibility Co-authored-by: Raul Jordan <raul@prysmaticlabs.com> * fix visibility * build works for all * fix blockchain test * fix a few tests * fix more tests * update validator in slashing * archiver passing * fixed rpc/validator * progress on core tests * resolve broken rpc tests * blockchain tests passed * fix up some tests in core * fix message diff * remove unnecessary save * Save validator after slashing * Update validators one by one * another update * fix everything * fix more precompute tests * fix blocks tests * more elegant fix * more helper fixes * change back ? * fix test * fix skip slot * fix test * reset caches * fix testutil * raceoff fixed * passing * Retrieve cached state in the beginning * lint * Fixed tests part 1 * Fixed rest of the tests * Minor changes to avoid copying, small refactor to reduce deplicated code * Handle att req for slot 0 * New beacon state: Only populate merkle layers as needed, copy merkle layers on copy/clone. (#4689) * Only populate merkle layers as needed, copy merkle layers on copy/clone. * use custom copy * Make maps of correct size * slightly fast, doesn't wait for lock Co-authored-by: prylabs-bulldozer[bot] <58059840+prylabs-bulldozer[bot]@users.noreply.github.com> * Target root can't be 0x00 * Don't use cache for current slot (may not be the right fix) * fixed up tests * Remove some copy for init sync. Not sure if it is safe enough for runtime though... testing... * Align with prev logic for process slots cachedState.Slot() < slot * Fix Initial Sync Flag (#4692) * fixes * fix up some test failures due to lack of nil checks * fix up some test failures due to lack of nil checks * fix up imports * revert some changes * imports Co-authored-by: Raul Jordan <raul@prysmaticlabs.com> * resolving further conflicts * Better skip slot cache (#4694) * Return copy of skip slot cache state, disable skip slot cache on sync * fix * Fix pruning * fix up issues with broken tests Co-authored-by: Nishant Das <nish1993@hotmail.com> Co-authored-by: Preston Van Loon <preston@prysmaticlabs.com> Co-authored-by: shayzluf <thezluf@gmail.com> Co-authored-by: terence tsao <terence@prysmaticlabs.com> Co-authored-by: prylabs-bulldozer[bot] <58059840+prylabs-bulldozer[bot]@users.noreply.github.com>
487 lines
14 KiB
Go
487 lines
14 KiB
Go
package archiver
|
|
|
|
import (
|
|
"context"
|
|
"fmt"
|
|
"io/ioutil"
|
|
"reflect"
|
|
"testing"
|
|
|
|
"github.com/gogo/protobuf/proto"
|
|
ethpb "github.com/prysmaticlabs/ethereumapis/eth/v1alpha1"
|
|
"github.com/prysmaticlabs/go-bitfield"
|
|
mock "github.com/prysmaticlabs/prysm/beacon-chain/blockchain/testing"
|
|
"github.com/prysmaticlabs/prysm/beacon-chain/core/epoch/precompute"
|
|
"github.com/prysmaticlabs/prysm/beacon-chain/core/feed"
|
|
statefeed "github.com/prysmaticlabs/prysm/beacon-chain/core/feed/state"
|
|
"github.com/prysmaticlabs/prysm/beacon-chain/core/helpers"
|
|
"github.com/prysmaticlabs/prysm/beacon-chain/db"
|
|
dbutil "github.com/prysmaticlabs/prysm/beacon-chain/db/testing"
|
|
stateTrie "github.com/prysmaticlabs/prysm/beacon-chain/state"
|
|
pb "github.com/prysmaticlabs/prysm/proto/beacon/p2p/v1"
|
|
"github.com/prysmaticlabs/prysm/shared/params"
|
|
"github.com/prysmaticlabs/prysm/shared/testutil"
|
|
"github.com/sirupsen/logrus"
|
|
logTest "github.com/sirupsen/logrus/hooks/test"
|
|
)
|
|
|
|
func init() {
|
|
logrus.SetLevel(logrus.DebugLevel)
|
|
logrus.SetOutput(ioutil.Discard)
|
|
}
|
|
|
|
func TestArchiverService_ReceivesBlockProcessedEvent(t *testing.T) {
|
|
hook := logTest.NewGlobal()
|
|
svc, beaconDB := setupService(t)
|
|
defer dbutil.TeardownDB(t, beaconDB)
|
|
st, err := stateTrie.InitializeFromProto(&pb.BeaconState{
|
|
Slot: 1,
|
|
})
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
svc.headFetcher = &mock.ChainService{
|
|
State: st,
|
|
}
|
|
|
|
event := &feed.Event{
|
|
Type: statefeed.BlockProcessed,
|
|
Data: &statefeed.BlockProcessedData{
|
|
BlockRoot: [32]byte{1, 2, 3},
|
|
Verified: true,
|
|
},
|
|
}
|
|
triggerStateEvent(t, svc, event)
|
|
testutil.AssertLogsContain(t, hook, fmt.Sprintf("%#x", event.Data.(*statefeed.BlockProcessedData).BlockRoot))
|
|
testutil.AssertLogsContain(t, hook, "Received block processed event")
|
|
}
|
|
|
|
func TestArchiverService_OnlyArchiveAtEpochEnd(t *testing.T) {
|
|
hook := logTest.NewGlobal()
|
|
svc, beaconDB := setupService(t)
|
|
defer dbutil.TeardownDB(t, beaconDB)
|
|
// The head state is NOT an epoch end.
|
|
st, err := stateTrie.InitializeFromProto(&pb.BeaconState{
|
|
Slot: params.BeaconConfig().SlotsPerEpoch - 2,
|
|
})
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
svc.headFetcher = &mock.ChainService{
|
|
State: st,
|
|
}
|
|
event := &feed.Event{
|
|
Type: statefeed.BlockProcessed,
|
|
Data: &statefeed.BlockProcessedData{
|
|
BlockRoot: [32]byte{1, 2, 3},
|
|
Verified: true,
|
|
},
|
|
}
|
|
triggerStateEvent(t, svc, event)
|
|
|
|
// The context should have been canceled.
|
|
if svc.ctx.Err() != context.Canceled {
|
|
t.Error("context was not canceled")
|
|
}
|
|
testutil.AssertLogsContain(t, hook, "Received block processed event")
|
|
// The service should ONLY log any archival logs if we receive a
|
|
// head slot that is an epoch end.
|
|
testutil.AssertLogsDoNotContain(t, hook, "Successfully archived")
|
|
}
|
|
|
|
func TestArchiverService_ArchivesEvenThroughSkipSlot(t *testing.T) {
|
|
hook := logTest.NewGlobal()
|
|
svc, beaconDB := setupService(t)
|
|
validatorCount := uint64(100)
|
|
headState, err := setupState(validatorCount)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
defer dbutil.TeardownDB(t, beaconDB)
|
|
event := &feed.Event{
|
|
Type: statefeed.BlockProcessed,
|
|
Data: &statefeed.BlockProcessedData{
|
|
BlockRoot: [32]byte{1, 2, 3},
|
|
Verified: true,
|
|
},
|
|
}
|
|
|
|
exitRoutine := make(chan bool)
|
|
go func() {
|
|
svc.run(svc.ctx)
|
|
<-exitRoutine
|
|
}()
|
|
|
|
// Send out an event every slot, skipping the end slot of the epoch.
|
|
for i := uint64(0); i < params.BeaconConfig().SlotsPerEpoch+1; i++ {
|
|
if err := headState.SetSlot(i); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
svc.headFetcher = &mock.ChainService{
|
|
State: headState,
|
|
}
|
|
if helpers.IsEpochEnd(i) {
|
|
continue
|
|
}
|
|
// Send in a loop to ensure it is delivered (busy wait for the service to subscribe to the state feed).
|
|
for sent := 0; sent == 0; {
|
|
sent = svc.stateNotifier.StateFeed().Send(event)
|
|
}
|
|
}
|
|
if err := svc.Stop(); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
exitRoutine <- true
|
|
|
|
// The context should have been canceled.
|
|
if svc.ctx.Err() != context.Canceled {
|
|
t.Error("context was not canceled")
|
|
}
|
|
|
|
testutil.AssertLogsContain(t, hook, "Received block processed event")
|
|
// Even though there was a skip slot, we should still be able to archive
|
|
// upon the next block event afterwards.
|
|
testutil.AssertLogsContain(t, hook, "Successfully archived")
|
|
}
|
|
|
|
func TestArchiverService_ComputesAndSavesParticipation(t *testing.T) {
|
|
hook := logTest.NewGlobal()
|
|
validatorCount := uint64(100)
|
|
headState, err := setupState(validatorCount)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
svc, beaconDB := setupService(t)
|
|
defer dbutil.TeardownDB(t, beaconDB)
|
|
svc.headFetcher = &mock.ChainService{
|
|
State: headState,
|
|
}
|
|
event := &feed.Event{
|
|
Type: statefeed.BlockProcessed,
|
|
Data: &statefeed.BlockProcessedData{
|
|
BlockRoot: [32]byte{1, 2, 3},
|
|
Verified: true,
|
|
},
|
|
}
|
|
triggerStateEvent(t, svc, event)
|
|
|
|
attestedBalance := uint64(1)
|
|
|
|
currentEpoch := helpers.CurrentEpoch(headState)
|
|
wanted := ðpb.ValidatorParticipation{
|
|
VotedEther: attestedBalance,
|
|
EligibleEther: validatorCount * params.BeaconConfig().MaxEffectiveBalance,
|
|
GlobalParticipationRate: float32(attestedBalance) / float32(validatorCount*params.BeaconConfig().MaxEffectiveBalance),
|
|
}
|
|
|
|
retrieved, err := svc.beaconDB.ArchivedValidatorParticipation(svc.ctx, currentEpoch)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
if !proto.Equal(wanted, retrieved) {
|
|
t.Errorf("Wanted participation for epoch %d %v, retrieved %v", currentEpoch-1, wanted, retrieved)
|
|
}
|
|
testutil.AssertLogsContain(t, hook, "Successfully archived")
|
|
}
|
|
|
|
func TestArchiverService_SavesIndicesAndBalances(t *testing.T) {
|
|
hook := logTest.NewGlobal()
|
|
validatorCount := uint64(100)
|
|
headState, err := setupState(validatorCount)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
svc, beaconDB := setupService(t)
|
|
defer dbutil.TeardownDB(t, beaconDB)
|
|
svc.headFetcher = &mock.ChainService{
|
|
State: headState,
|
|
}
|
|
event := &feed.Event{
|
|
Type: statefeed.BlockProcessed,
|
|
Data: &statefeed.BlockProcessedData{
|
|
BlockRoot: [32]byte{1, 2, 3},
|
|
Verified: true,
|
|
},
|
|
}
|
|
triggerStateEvent(t, svc, event)
|
|
|
|
retrieved, err := svc.beaconDB.ArchivedBalances(svc.ctx, helpers.CurrentEpoch(headState))
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if !reflect.DeepEqual(headState.Balances(), retrieved) {
|
|
t.Errorf(
|
|
"Wanted balances for epoch %d %v, retrieved %v",
|
|
helpers.CurrentEpoch(headState),
|
|
headState.Balances(),
|
|
retrieved,
|
|
)
|
|
}
|
|
testutil.AssertLogsContain(t, hook, "Successfully archived")
|
|
}
|
|
|
|
func TestArchiverService_SavesCommitteeInfo(t *testing.T) {
|
|
hook := logTest.NewGlobal()
|
|
validatorCount := uint64(100)
|
|
headState, err := setupState(validatorCount)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
svc, beaconDB := setupService(t)
|
|
defer dbutil.TeardownDB(t, beaconDB)
|
|
svc.headFetcher = &mock.ChainService{
|
|
State: headState,
|
|
}
|
|
event := &feed.Event{
|
|
Type: statefeed.BlockProcessed,
|
|
Data: &statefeed.BlockProcessedData{
|
|
BlockRoot: [32]byte{1, 2, 3},
|
|
Verified: true,
|
|
},
|
|
}
|
|
triggerStateEvent(t, svc, event)
|
|
|
|
currentEpoch := helpers.CurrentEpoch(headState)
|
|
proposerSeed, err := helpers.Seed(headState, currentEpoch, params.BeaconConfig().DomainBeaconProposer)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
attesterSeed, err := helpers.Seed(headState, currentEpoch, params.BeaconConfig().DomainBeaconAttester)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
wanted := &pb.ArchivedCommitteeInfo{
|
|
ProposerSeed: proposerSeed[:],
|
|
AttesterSeed: attesterSeed[:],
|
|
}
|
|
|
|
retrieved, err := svc.beaconDB.ArchivedCommitteeInfo(svc.ctx, helpers.CurrentEpoch(headState))
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if !proto.Equal(wanted, retrieved) {
|
|
t.Errorf(
|
|
"Wanted committee info for epoch %d %v, retrieved %v",
|
|
helpers.CurrentEpoch(headState),
|
|
wanted,
|
|
retrieved,
|
|
)
|
|
}
|
|
testutil.AssertLogsContain(t, hook, "Successfully archived")
|
|
}
|
|
|
|
func TestArchiverService_SavesActivatedValidatorChanges(t *testing.T) {
|
|
hook := logTest.NewGlobal()
|
|
validatorCount := uint64(100)
|
|
headState, err := setupState(validatorCount)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
svc, beaconDB := setupService(t)
|
|
defer dbutil.TeardownDB(t, beaconDB)
|
|
svc.headFetcher = &mock.ChainService{
|
|
State: headState,
|
|
}
|
|
prevEpoch := helpers.PrevEpoch(headState)
|
|
delayedActEpoch := helpers.DelayedActivationExitEpoch(prevEpoch)
|
|
val1, err := headState.ValidatorAtIndex(4)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
val1.ActivationEpoch = delayedActEpoch
|
|
val2, err := headState.ValidatorAtIndex(5)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
val2.ActivationEpoch = delayedActEpoch
|
|
if err := headState.UpdateValidatorAtIndex(4, val1); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if err := headState.UpdateValidatorAtIndex(5, val1); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
event := &feed.Event{
|
|
Type: statefeed.BlockProcessed,
|
|
Data: &statefeed.BlockProcessedData{
|
|
BlockRoot: [32]byte{1, 2, 3},
|
|
Verified: true,
|
|
},
|
|
}
|
|
triggerStateEvent(t, svc, event)
|
|
|
|
retrieved, err := beaconDB.ArchivedActiveValidatorChanges(svc.ctx, prevEpoch)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if retrieved == nil {
|
|
t.Fatal("Retrieved indices are nil")
|
|
}
|
|
if !reflect.DeepEqual(retrieved.Activated, []uint64{4, 5}) {
|
|
t.Errorf("Wanted indices 4 5 activated, received %v", retrieved.Activated)
|
|
}
|
|
testutil.AssertLogsContain(t, hook, "Successfully archived")
|
|
}
|
|
|
|
func TestArchiverService_SavesSlashedValidatorChanges(t *testing.T) {
|
|
hook := logTest.NewGlobal()
|
|
validatorCount := uint64(100)
|
|
headState, err := setupState(validatorCount)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
svc, beaconDB := setupService(t)
|
|
defer dbutil.TeardownDB(t, beaconDB)
|
|
svc.headFetcher = &mock.ChainService{
|
|
State: headState,
|
|
}
|
|
prevEpoch := helpers.PrevEpoch(headState)
|
|
val1, err := headState.ValidatorAtIndex(95)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
val1.Slashed = true
|
|
val2, err := headState.ValidatorAtIndex(96)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
val2.Slashed = true
|
|
if err := headState.UpdateValidatorAtIndex(95, val1); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if err := headState.UpdateValidatorAtIndex(96, val1); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
event := &feed.Event{
|
|
Type: statefeed.BlockProcessed,
|
|
Data: &statefeed.BlockProcessedData{
|
|
BlockRoot: [32]byte{1, 2, 3},
|
|
Verified: true,
|
|
},
|
|
}
|
|
triggerStateEvent(t, svc, event)
|
|
|
|
retrieved, err := beaconDB.ArchivedActiveValidatorChanges(svc.ctx, prevEpoch)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if retrieved == nil {
|
|
t.Fatal("Retrieved indices are nil")
|
|
}
|
|
if !reflect.DeepEqual(retrieved.Slashed, []uint64{95, 96}) {
|
|
t.Errorf("Wanted indices 95, 96 slashed, received %v", retrieved.Slashed)
|
|
}
|
|
testutil.AssertLogsContain(t, hook, "Successfully archived")
|
|
}
|
|
|
|
func TestArchiverService_SavesExitedValidatorChanges(t *testing.T) {
|
|
hook := logTest.NewGlobal()
|
|
validatorCount := uint64(100)
|
|
headState, err := setupState(validatorCount)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
svc, beaconDB := setupService(t)
|
|
defer dbutil.TeardownDB(t, beaconDB)
|
|
svc.headFetcher = &mock.ChainService{
|
|
State: headState,
|
|
}
|
|
prevEpoch := helpers.PrevEpoch(headState)
|
|
val, err := headState.ValidatorAtIndex(95)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
val.ExitEpoch = prevEpoch
|
|
val.WithdrawableEpoch = prevEpoch + params.BeaconConfig().MinValidatorWithdrawabilityDelay
|
|
if err := headState.UpdateValidatorAtIndex(95, val); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
event := &feed.Event{
|
|
Type: statefeed.BlockProcessed,
|
|
Data: &statefeed.BlockProcessedData{
|
|
BlockRoot: [32]byte{1, 2, 3},
|
|
Verified: true,
|
|
},
|
|
}
|
|
triggerStateEvent(t, svc, event)
|
|
testutil.AssertLogsContain(t, hook, "Successfully archived")
|
|
retrieved, err := beaconDB.ArchivedActiveValidatorChanges(svc.ctx, prevEpoch)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if retrieved == nil {
|
|
t.Fatal("Retrieved indices are nil")
|
|
}
|
|
if !reflect.DeepEqual(retrieved.Exited, []uint64{95}) {
|
|
t.Errorf("Wanted indices 95 exited, received %v", retrieved.Exited)
|
|
}
|
|
}
|
|
|
|
func setupState(validatorCount uint64) (*stateTrie.BeaconState, error) {
|
|
validators := make([]*ethpb.Validator, validatorCount)
|
|
balances := make([]uint64, validatorCount)
|
|
for i := 0; i < len(validators); i++ {
|
|
validators[i] = ðpb.Validator{
|
|
ExitEpoch: params.BeaconConfig().FarFutureEpoch,
|
|
WithdrawableEpoch: params.BeaconConfig().FarFutureEpoch,
|
|
EffectiveBalance: params.BeaconConfig().MaxEffectiveBalance,
|
|
}
|
|
balances[i] = params.BeaconConfig().MaxEffectiveBalance
|
|
}
|
|
|
|
atts := []*pb.PendingAttestation{{Data: ðpb.AttestationData{Target: ðpb.Checkpoint{}}}}
|
|
|
|
// We initialize a head state that has attestations from participated
|
|
// validators in a simulated fashion.
|
|
return stateTrie.InitializeFromProto(&pb.BeaconState{
|
|
Slot: (2 * params.BeaconConfig().SlotsPerEpoch) - 1,
|
|
Validators: validators,
|
|
Balances: balances,
|
|
BlockRoots: make([][]byte, params.BeaconConfig().SlotsPerHistoricalRoot),
|
|
Slashings: []uint64{0, 1e9, 1e9},
|
|
RandaoMixes: make([][]byte, params.BeaconConfig().EpochsPerHistoricalVector),
|
|
CurrentEpochAttestations: atts,
|
|
FinalizedCheckpoint: ðpb.Checkpoint{},
|
|
JustificationBits: bitfield.Bitvector4{0x00},
|
|
CurrentJustifiedCheckpoint: ðpb.Checkpoint{},
|
|
})
|
|
}
|
|
|
|
func setupService(t *testing.T) (*Service, db.Database) {
|
|
beaconDB := dbutil.SetupDB(t)
|
|
ctx, cancel := context.WithCancel(context.Background())
|
|
validatorCount := uint64(100)
|
|
totalBalance := validatorCount * params.BeaconConfig().MaxEffectiveBalance
|
|
mockChainService := &mock.ChainService{}
|
|
return &Service{
|
|
beaconDB: beaconDB,
|
|
ctx: ctx,
|
|
cancel: cancel,
|
|
stateNotifier: mockChainService.StateNotifier(),
|
|
participationFetcher: &mock.ChainService{
|
|
Balance: &precompute.Balance{PrevEpoch: totalBalance, PrevEpochTargetAttesters: 1}},
|
|
}, beaconDB
|
|
}
|
|
|
|
func triggerStateEvent(t *testing.T, svc *Service, event *feed.Event) {
|
|
exitRoutine := make(chan bool)
|
|
go func() {
|
|
svc.run(svc.ctx)
|
|
<-exitRoutine
|
|
}()
|
|
|
|
// Send in a loop to ensure it is delivered (busy wait for the service to subscribe to the state feed).
|
|
for sent := 0; sent == 0; {
|
|
sent = svc.stateNotifier.StateFeed().Send(event)
|
|
}
|
|
if err := svc.Stop(); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
exitRoutine <- true
|
|
|
|
// The context should have been canceled.
|
|
if svc.ctx.Err() != context.Canceled {
|
|
t.Error("context was not canceled")
|
|
}
|
|
}
|