mirror of
https://gitlab.com/pulsechaincom/prysm-pulse.git
synced 2024-12-26 05:17:22 +00:00
cc741ed8af
* begin state service * begin on the state trie idea * created beacon state structure * add in the full clone getter * return by value instead * add all setters * new state setters are being completed * arrays roots exposed * close to finishing all these headerssss * functionality complete * added in proto benchmark test * test for compatibility * add test for compat * comments fixed * add clone * add clone * remove underlying copies * make it immutable * integrate it into chainservice * revert * wrap up comments for package * address all comments and godocs * address all comments * clone the pending attestation properly * properly clone remaining items * tests pass fixed bug * begin using it instead of head state * prevent nil pointer exceptions * begin using new struct in db * integrated new type into db package * add proper nil checks * using new state in archiver * refactored much of core * editing all the precompute functions * done with most core refactor * fixed up some bugs in the clone comparisons * append current epoch atts * add missing setters * add new setters * fix other core methods * fix up transition * main service and forkchoice * fix rpc * integrated to powchain * some more changes * fix build * improve processing of deposits * fix error * prevent panic * comment * fix process att * gaz * fix up att process * resolve existing review comments * resolve another batch of gh comments * resolve broken cpt state * revise testutil to use the new state * begin updating the state transition func to pass in more compartmentalized args * finish editing transition function to return errors * block operations pretty much done with refactor * state transition fully refactored * got epoch processing completed * fix build in fork choice * fixing more of the build * fix up broken sync package * it builds nowww it buildssss * revert registry changes * Recompute on Read (#4627) * compute on read * fix up eth1 data votes * looking into slashings bug introduced in core/ * able to advance more slots * add logging * can now sync with testnet yay * remove the leaves algorithm and other merkle imports * expose initialize unsafe funcs * Update beacon-chain/db/kv/state.go * lint Co-authored-by: Raul Jordan <raul@prysmaticlabs.com> * More Optimizations for New State (#4641) * map optimization * more optimizations * use a custom hasher * comment * block operations optimizations * Update beacon-chain/state/types.go Co-Authored-By: Raul Jordan <raul@prysmaticlabs.com> * fixed up various operations to use the validator index map access Co-authored-by: Raul Jordan <raul@prysmaticlabs.com> * archiver tests pass * fixing cache tests * cache tests passing * edited validator tests * powchain tests passing * halfway thru sync tests * more sync test fixes * add in tests for state/ * working through rpc tests * assignments tests passed * almost done with rpc/beacon tests * resolved painful validator test * fixed up even more tests * resolve tests * fix build * reduce a randao mixes copy * fixes under //beacon-chain/blockchain/... * build //beacon-chain/core/... * fixes * Runtime Optimizations (#4648) * parallelize shuffling * clean up * lint * fix build * use callback to read from registry * fix array roots and size map * new improvements * reduce hash allocs * improved shuffling * terence's review * use different method * raul's comment * new array roots * remove clone in pre-compute * Update beacon-chain/state/types.go Co-Authored-By: Raul Jordan <raul@prysmaticlabs.com> * raul's review * lint * fix build issues * fix visibility Co-authored-by: Raul Jordan <raul@prysmaticlabs.com> * fix visibility * build works for all * fix blockchain test * fix a few tests * fix more tests * update validator in slashing * archiver passing * fixed rpc/validator * progress on core tests * resolve broken rpc tests * blockchain tests passed * fix up some tests in core * fix message diff * remove unnecessary save * Save validator after slashing * Update validators one by one * another update * fix everything * fix more precompute tests * fix blocks tests * more elegant fix * more helper fixes * change back ? * fix test * fix skip slot * fix test * reset caches * fix testutil * raceoff fixed * passing * Retrieve cached state in the beginning * lint * Fixed tests part 1 * Fixed rest of the tests * Minor changes to avoid copying, small refactor to reduce deplicated code * Handle att req for slot 0 * New beacon state: Only populate merkle layers as needed, copy merkle layers on copy/clone. (#4689) * Only populate merkle layers as needed, copy merkle layers on copy/clone. * use custom copy * Make maps of correct size * slightly fast, doesn't wait for lock Co-authored-by: prylabs-bulldozer[bot] <58059840+prylabs-bulldozer[bot]@users.noreply.github.com> * Target root can't be 0x00 * Don't use cache for current slot (may not be the right fix) * fixed up tests * Remove some copy for init sync. Not sure if it is safe enough for runtime though... testing... * Align with prev logic for process slots cachedState.Slot() < slot * Fix Initial Sync Flag (#4692) * fixes * fix up some test failures due to lack of nil checks * fix up some test failures due to lack of nil checks * fix up imports * revert some changes * imports Co-authored-by: Raul Jordan <raul@prysmaticlabs.com> * resolving further conflicts * Better skip slot cache (#4694) * Return copy of skip slot cache state, disable skip slot cache on sync * fix * Fix pruning * fix up issues with broken tests Co-authored-by: Nishant Das <nish1993@hotmail.com> Co-authored-by: Preston Van Loon <preston@prysmaticlabs.com> Co-authored-by: shayzluf <thezluf@gmail.com> Co-authored-by: terence tsao <terence@prysmaticlabs.com> Co-authored-by: prylabs-bulldozer[bot] <58059840+prylabs-bulldozer[bot]@users.noreply.github.com>
580 lines
16 KiB
Go
580 lines
16 KiB
Go
package beacon
|
|
|
|
import (
|
|
"context"
|
|
"encoding/binary"
|
|
"fmt"
|
|
"reflect"
|
|
"strconv"
|
|
"strings"
|
|
"testing"
|
|
|
|
"github.com/gogo/protobuf/proto"
|
|
ethpb "github.com/prysmaticlabs/ethereumapis/eth/v1alpha1"
|
|
"github.com/prysmaticlabs/go-ssz"
|
|
mock "github.com/prysmaticlabs/prysm/beacon-chain/blockchain/testing"
|
|
"github.com/prysmaticlabs/prysm/beacon-chain/core/helpers"
|
|
dbTest "github.com/prysmaticlabs/prysm/beacon-chain/db/testing"
|
|
"github.com/prysmaticlabs/prysm/beacon-chain/flags"
|
|
stateTrie "github.com/prysmaticlabs/prysm/beacon-chain/state"
|
|
pbp2p "github.com/prysmaticlabs/prysm/proto/beacon/p2p/v1"
|
|
"github.com/prysmaticlabs/prysm/shared/params"
|
|
)
|
|
|
|
func TestServer_ListAssignments_CannotRequestFutureEpoch(t *testing.T) {
|
|
db := dbTest.SetupDB(t)
|
|
defer dbTest.TeardownDB(t, db)
|
|
|
|
ctx := context.Background()
|
|
st, err := stateTrie.InitializeFromProto(&pbp2p.BeaconState{
|
|
Slot: 0,
|
|
})
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
bs := &Server{
|
|
BeaconDB: db,
|
|
HeadFetcher: &mock.ChainService{
|
|
State: st,
|
|
},
|
|
}
|
|
|
|
wanted := "Cannot retrieve information about an epoch in the future"
|
|
if _, err := bs.ListValidatorAssignments(
|
|
ctx,
|
|
ðpb.ListValidatorAssignmentsRequest{
|
|
QueryFilter: ðpb.ListValidatorAssignmentsRequest_Epoch{
|
|
Epoch: 1,
|
|
},
|
|
},
|
|
); err != nil && !strings.Contains(err.Error(), wanted) {
|
|
t.Errorf("Expected error %v, received %v", wanted, err)
|
|
}
|
|
}
|
|
|
|
func TestServer_ListAssignments_NoResults(t *testing.T) {
|
|
db := dbTest.SetupDB(t)
|
|
defer dbTest.TeardownDB(t, db)
|
|
|
|
ctx := context.Background()
|
|
st, err := stateTrie.InitializeFromProto(&pbp2p.BeaconState{
|
|
Slot: 0,
|
|
RandaoMixes: make([][]byte, params.BeaconConfig().EpochsPerHistoricalVector),
|
|
})
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
bs := &Server{
|
|
BeaconDB: db,
|
|
HeadFetcher: &mock.ChainService{
|
|
State: st,
|
|
},
|
|
}
|
|
wanted := ðpb.ValidatorAssignments{
|
|
Assignments: make([]*ethpb.ValidatorAssignments_CommitteeAssignment, 0),
|
|
TotalSize: int32(0),
|
|
NextPageToken: strconv.Itoa(0),
|
|
}
|
|
res, err := bs.ListValidatorAssignments(
|
|
ctx,
|
|
ðpb.ListValidatorAssignmentsRequest{
|
|
QueryFilter: ðpb.ListValidatorAssignmentsRequest_Genesis{
|
|
Genesis: true,
|
|
},
|
|
},
|
|
)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if !proto.Equal(wanted, res) {
|
|
t.Errorf("Wanted %v, received %v", wanted, res)
|
|
}
|
|
}
|
|
|
|
func TestServer_ListAssignments_Pagination_InputOutOfRange(t *testing.T) {
|
|
db := dbTest.SetupDB(t)
|
|
defer dbTest.TeardownDB(t, db)
|
|
|
|
ctx := context.Background()
|
|
setupValidators(t, db, 1)
|
|
headState, err := db.HeadState(ctx)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
bs := &Server{
|
|
BeaconDB: db,
|
|
HeadFetcher: &mock.ChainService{
|
|
State: headState,
|
|
},
|
|
}
|
|
|
|
wanted := fmt.Sprintf("page start %d >= list %d", 0, 0)
|
|
if _, err := bs.ListValidatorAssignments(
|
|
context.Background(),
|
|
ðpb.ListValidatorAssignmentsRequest{
|
|
QueryFilter: ðpb.ListValidatorAssignmentsRequest_Genesis{Genesis: true},
|
|
},
|
|
); err != nil && !strings.Contains(err.Error(), wanted) {
|
|
t.Errorf("Expected error %v, received %v", wanted, err)
|
|
}
|
|
}
|
|
|
|
func TestServer_ListAssignments_Pagination_ExceedsMaxPageSize(t *testing.T) {
|
|
bs := &Server{}
|
|
exceedsMax := int32(flags.Get().MaxPageSize + 1)
|
|
|
|
wanted := fmt.Sprintf("Requested page size %d can not be greater than max size %d", exceedsMax, flags.Get().MaxPageSize)
|
|
req := ðpb.ListValidatorAssignmentsRequest{
|
|
PageToken: strconv.Itoa(0),
|
|
PageSize: exceedsMax,
|
|
}
|
|
if _, err := bs.ListValidatorAssignments(context.Background(), req); err != nil && !strings.Contains(err.Error(), wanted) {
|
|
t.Errorf("Expected error %v, received %v", wanted, err)
|
|
}
|
|
}
|
|
|
|
func TestServer_ListAssignments_Pagination_DefaultPageSize_NoArchive(t *testing.T) {
|
|
helpers.ClearCache()
|
|
db := dbTest.SetupDB(t)
|
|
defer dbTest.TeardownDB(t, db)
|
|
|
|
ctx := context.Background()
|
|
count := 1000
|
|
validators := make([]*ethpb.Validator, 0, count)
|
|
for i := 0; i < count; i++ {
|
|
pubKey := make([]byte, params.BeaconConfig().BLSPubkeyLength)
|
|
binary.LittleEndian.PutUint64(pubKey, uint64(i))
|
|
if err := db.SaveValidatorIndex(ctx, pubKey, uint64(i)); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
// Mark the validators with index divisible by 3 inactive.
|
|
if i%3 == 0 {
|
|
validators = append(validators, ðpb.Validator{
|
|
PublicKey: pubKey,
|
|
ExitEpoch: 0,
|
|
ActivationEpoch: 0,
|
|
EffectiveBalance: params.BeaconConfig().MaxEffectiveBalance,
|
|
})
|
|
} else {
|
|
validators = append(validators, ðpb.Validator{
|
|
PublicKey: pubKey,
|
|
ExitEpoch: params.BeaconConfig().FarFutureEpoch,
|
|
EffectiveBalance: params.BeaconConfig().MaxEffectiveBalance,
|
|
ActivationEpoch: 0,
|
|
})
|
|
}
|
|
}
|
|
|
|
blk := ðpb.BeaconBlock{
|
|
Slot: 0,
|
|
}
|
|
blockRoot, err := ssz.HashTreeRoot(blk)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
s, err := stateTrie.InitializeFromProto(&pbp2p.BeaconState{
|
|
Validators: validators,
|
|
RandaoMixes: make([][]byte, params.BeaconConfig().EpochsPerHistoricalVector),
|
|
})
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if err := db.SaveState(ctx, s, blockRoot); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if err := db.SaveHeadBlockRoot(ctx, blockRoot); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
bs := &Server{
|
|
BeaconDB: db,
|
|
HeadFetcher: &mock.ChainService{
|
|
State: s,
|
|
},
|
|
FinalizationFetcher: &mock.ChainService{
|
|
FinalizedCheckPoint: ðpb.Checkpoint{
|
|
Epoch: 0,
|
|
},
|
|
},
|
|
}
|
|
|
|
res, err := bs.ListValidatorAssignments(context.Background(), ðpb.ListValidatorAssignmentsRequest{
|
|
QueryFilter: ðpb.ListValidatorAssignmentsRequest_Genesis{Genesis: true},
|
|
})
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Construct the wanted assignments.
|
|
var wanted []*ethpb.ValidatorAssignments_CommitteeAssignment
|
|
|
|
activeIndices, err := helpers.ActiveValidatorIndices(s, 0)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
for _, index := range activeIndices[0:params.BeaconConfig().DefaultPageSize] {
|
|
committee, committeeIndex, attesterSlot, proposerSlot, err := helpers.CommitteeAssignment(s, 0, index)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
val, err := s.ValidatorAtIndex(index)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
wanted = append(wanted, ðpb.ValidatorAssignments_CommitteeAssignment{
|
|
BeaconCommittees: committee,
|
|
CommitteeIndex: committeeIndex,
|
|
AttesterSlot: attesterSlot,
|
|
ProposerSlot: proposerSlot,
|
|
PublicKey: val.PublicKey,
|
|
})
|
|
}
|
|
|
|
if !reflect.DeepEqual(res.Assignments, wanted) {
|
|
t.Error("Did not receive wanted assignments")
|
|
}
|
|
}
|
|
|
|
func TestServer_ListAssignments_Pagination_DefaultPageSize_FromArchive(t *testing.T) {
|
|
helpers.ClearCache()
|
|
db := dbTest.SetupDB(t)
|
|
defer dbTest.TeardownDB(t, db)
|
|
|
|
ctx := context.Background()
|
|
count := 1000
|
|
validators := make([]*ethpb.Validator, 0, count)
|
|
balances := make([]uint64, count)
|
|
for i := 0; i < count; i++ {
|
|
pubKey := make([]byte, params.BeaconConfig().BLSPubkeyLength)
|
|
binary.LittleEndian.PutUint64(pubKey, uint64(i))
|
|
if err := db.SaveValidatorIndex(ctx, pubKey, uint64(i)); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
// Mark the validators with index divisible by 3 inactive.
|
|
if i%3 == 0 {
|
|
validators = append(validators, ðpb.Validator{
|
|
PublicKey: pubKey,
|
|
ExitEpoch: 0,
|
|
EffectiveBalance: params.BeaconConfig().MaxEffectiveBalance,
|
|
})
|
|
} else {
|
|
validators = append(validators, ðpb.Validator{
|
|
PublicKey: pubKey,
|
|
ExitEpoch: params.BeaconConfig().FarFutureEpoch,
|
|
EffectiveBalance: params.BeaconConfig().MaxEffectiveBalance,
|
|
})
|
|
}
|
|
balances[i] = params.BeaconConfig().MaxEffectiveBalance
|
|
}
|
|
|
|
blk := ðpb.BeaconBlock{
|
|
Slot: 0,
|
|
}
|
|
blockRoot, err := ssz.HashTreeRoot(blk)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
s, err := stateTrie.InitializeFromProto(&pbp2p.BeaconState{
|
|
Validators: validators,
|
|
Balances: balances,
|
|
RandaoMixes: make([][]byte, params.BeaconConfig().EpochsPerHistoricalVector),
|
|
})
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if err := db.SaveState(ctx, s, blockRoot); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if err := db.SaveHeadBlockRoot(ctx, blockRoot); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// We tell the beacon chain server that our finalized epoch is 10 so that when
|
|
// we request assignments for epoch 0, it looks within the archived data.
|
|
bs := &Server{
|
|
BeaconDB: db,
|
|
HeadFetcher: &mock.ChainService{
|
|
State: s,
|
|
},
|
|
FinalizationFetcher: &mock.ChainService{
|
|
FinalizedCheckPoint: ðpb.Checkpoint{
|
|
Epoch: 10,
|
|
},
|
|
},
|
|
}
|
|
|
|
// We then store archived data into the DB.
|
|
currentEpoch := helpers.CurrentEpoch(s)
|
|
proposerSeed, err := helpers.Seed(s, currentEpoch, params.BeaconConfig().DomainBeaconProposer)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
attesterSeed, err := helpers.Seed(s, currentEpoch, params.BeaconConfig().DomainBeaconAttester)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if err := db.SaveArchivedCommitteeInfo(context.Background(), 0, &pbp2p.ArchivedCommitteeInfo{
|
|
ProposerSeed: proposerSeed[:],
|
|
AttesterSeed: attesterSeed[:],
|
|
}); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
if err := db.SaveArchivedBalances(context.Background(), 0, balances); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Construct the wanted assignments.
|
|
var wanted []*ethpb.ValidatorAssignments_CommitteeAssignment
|
|
activeIndices, err := helpers.ActiveValidatorIndices(s, 0)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
for _, index := range activeIndices[0:params.BeaconConfig().DefaultPageSize] {
|
|
committee, committeeIndex, attesterSlot, proposerSlot, err := helpers.CommitteeAssignment(s, 0, index)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
val, err := s.ValidatorAtIndex(index)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
assign := ðpb.ValidatorAssignments_CommitteeAssignment{
|
|
BeaconCommittees: committee,
|
|
CommitteeIndex: committeeIndex,
|
|
AttesterSlot: attesterSlot,
|
|
ProposerSlot: proposerSlot,
|
|
PublicKey: val.PublicKey,
|
|
}
|
|
wanted = append(wanted, assign)
|
|
}
|
|
|
|
res, err := bs.ListValidatorAssignments(context.Background(), ðpb.ListValidatorAssignmentsRequest{
|
|
QueryFilter: ðpb.ListValidatorAssignmentsRequest_Genesis{Genesis: true},
|
|
})
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if !reflect.DeepEqual(res.Assignments, wanted) {
|
|
t.Error("Did not receive wanted assignments")
|
|
}
|
|
}
|
|
|
|
func TestServer_ListAssignments_FilterPubkeysIndices_NoPagination(t *testing.T) {
|
|
helpers.ClearCache()
|
|
db := dbTest.SetupDB(t)
|
|
defer dbTest.TeardownDB(t, db)
|
|
|
|
ctx := context.Background()
|
|
count := 100
|
|
validators := make([]*ethpb.Validator, 0, count)
|
|
for i := 0; i < count; i++ {
|
|
pubKey := make([]byte, params.BeaconConfig().BLSPubkeyLength)
|
|
binary.LittleEndian.PutUint64(pubKey, uint64(i))
|
|
if err := db.SaveValidatorIndex(ctx, pubKey, uint64(i)); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
validators = append(validators, ðpb.Validator{PublicKey: pubKey, ExitEpoch: params.BeaconConfig().FarFutureEpoch})
|
|
}
|
|
|
|
blk := ðpb.BeaconBlock{
|
|
Slot: 0,
|
|
}
|
|
blockRoot, err := ssz.HashTreeRoot(blk)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
s, err := stateTrie.InitializeFromProto(&pbp2p.BeaconState{
|
|
Validators: validators,
|
|
RandaoMixes: make([][]byte, params.BeaconConfig().EpochsPerHistoricalVector),
|
|
})
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if err := db.SaveState(ctx, s, blockRoot); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if err := db.SaveHeadBlockRoot(ctx, blockRoot); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
bs := &Server{
|
|
BeaconDB: db,
|
|
HeadFetcher: &mock.ChainService{
|
|
State: s,
|
|
},
|
|
FinalizationFetcher: &mock.ChainService{
|
|
FinalizedCheckPoint: ðpb.Checkpoint{
|
|
Epoch: 0,
|
|
},
|
|
},
|
|
}
|
|
|
|
pubKey1 := make([]byte, params.BeaconConfig().BLSPubkeyLength)
|
|
binary.LittleEndian.PutUint64(pubKey1, 1)
|
|
pubKey2 := make([]byte, params.BeaconConfig().BLSPubkeyLength)
|
|
binary.LittleEndian.PutUint64(pubKey2, 2)
|
|
req := ðpb.ListValidatorAssignmentsRequest{PublicKeys: [][]byte{pubKey1, pubKey2}, Indices: []uint64{2, 3}}
|
|
res, err := bs.ListValidatorAssignments(context.Background(), req)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Construct the wanted assignments.
|
|
var wanted []*ethpb.ValidatorAssignments_CommitteeAssignment
|
|
|
|
activeIndices, err := helpers.ActiveValidatorIndices(s, 0)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
for _, index := range activeIndices[1:4] {
|
|
committee, committeeIndex, attesterSlot, proposerSlot, err := helpers.CommitteeAssignment(s, 0, index)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
val, err := s.ValidatorAtIndex(index)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
wanted = append(wanted, ðpb.ValidatorAssignments_CommitteeAssignment{
|
|
BeaconCommittees: committee,
|
|
CommitteeIndex: committeeIndex,
|
|
AttesterSlot: attesterSlot,
|
|
ProposerSlot: proposerSlot,
|
|
PublicKey: val.PublicKey,
|
|
})
|
|
}
|
|
|
|
if !reflect.DeepEqual(res.Assignments, wanted) {
|
|
t.Error("Did not receive wanted assignments")
|
|
}
|
|
}
|
|
|
|
func TestServer_ListAssignments_CanFilterPubkeysIndices_WithPagination(t *testing.T) {
|
|
db := dbTest.SetupDB(t)
|
|
defer dbTest.TeardownDB(t, db)
|
|
|
|
ctx := context.Background()
|
|
count := 100
|
|
validators := make([]*ethpb.Validator, 0, count)
|
|
for i := 0; i < count; i++ {
|
|
pubKey := make([]byte, params.BeaconConfig().BLSPubkeyLength)
|
|
binary.LittleEndian.PutUint64(pubKey, uint64(i))
|
|
if err := db.SaveValidatorIndex(ctx, pubKey, uint64(i)); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
validators = append(validators, ðpb.Validator{PublicKey: pubKey, ExitEpoch: params.BeaconConfig().FarFutureEpoch})
|
|
}
|
|
|
|
blk := ðpb.BeaconBlock{
|
|
Slot: 0,
|
|
}
|
|
blockRoot, err := ssz.HashTreeRoot(blk)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
s, err := stateTrie.InitializeFromProto(&pbp2p.BeaconState{
|
|
Validators: validators,
|
|
RandaoMixes: make([][]byte, params.BeaconConfig().EpochsPerHistoricalVector),
|
|
})
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if err := db.SaveState(ctx, s, blockRoot); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if err := db.SaveHeadBlockRoot(ctx, blockRoot); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
bs := &Server{
|
|
BeaconDB: db,
|
|
HeadFetcher: &mock.ChainService{
|
|
State: s,
|
|
},
|
|
FinalizationFetcher: &mock.ChainService{
|
|
FinalizedCheckPoint: ðpb.Checkpoint{
|
|
Epoch: 0,
|
|
},
|
|
},
|
|
}
|
|
|
|
req := ðpb.ListValidatorAssignmentsRequest{Indices: []uint64{1, 2, 3, 4, 5, 6}, PageSize: 2, PageToken: "1"}
|
|
res, err := bs.ListValidatorAssignments(context.Background(), req)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Construct the wanted assignments.
|
|
var assignments []*ethpb.ValidatorAssignments_CommitteeAssignment
|
|
|
|
activeIndices, err := helpers.ActiveValidatorIndices(s, 0)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
for _, index := range activeIndices[3:5] {
|
|
committee, committeeIndex, attesterSlot, proposerSlot, err := helpers.CommitteeAssignment(s, 0, index)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
val, err := s.ValidatorAtIndex(index)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
assignments = append(assignments, ðpb.ValidatorAssignments_CommitteeAssignment{
|
|
BeaconCommittees: committee,
|
|
CommitteeIndex: committeeIndex,
|
|
AttesterSlot: attesterSlot,
|
|
ProposerSlot: proposerSlot,
|
|
PublicKey: val.PublicKey,
|
|
})
|
|
}
|
|
|
|
wantedRes := ðpb.ValidatorAssignments{
|
|
Assignments: assignments,
|
|
TotalSize: int32(len(req.Indices)),
|
|
NextPageToken: "2",
|
|
}
|
|
|
|
if !reflect.DeepEqual(res, wantedRes) {
|
|
t.Error("Did not get wanted assignments")
|
|
}
|
|
|
|
// Test the wrap around scenario.
|
|
assignments = nil
|
|
req = ðpb.ListValidatorAssignmentsRequest{Indices: []uint64{1, 2, 3, 4, 5, 6}, PageSize: 5, PageToken: "1"}
|
|
res, err = bs.ListValidatorAssignments(context.Background(), req)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
for _, index := range activeIndices[6:7] {
|
|
committee, committeeIndex, attesterSlot, proposerSlot, err := helpers.CommitteeAssignment(s, 0, index)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
val, err := s.ValidatorAtIndex(index)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
assignments = append(assignments, ðpb.ValidatorAssignments_CommitteeAssignment{
|
|
BeaconCommittees: committee,
|
|
CommitteeIndex: committeeIndex,
|
|
AttesterSlot: attesterSlot,
|
|
ProposerSlot: proposerSlot,
|
|
PublicKey: val.PublicKey,
|
|
})
|
|
}
|
|
|
|
wantedRes = ðpb.ValidatorAssignments{
|
|
Assignments: assignments,
|
|
TotalSize: int32(len(req.Indices)),
|
|
NextPageToken: "",
|
|
}
|
|
|
|
if !reflect.DeepEqual(res, wantedRes) {
|
|
t.Error("Did not receive wanted assignments")
|
|
}
|
|
}
|