mirror of
https://gitlab.com/pulsechaincom/prysm-pulse.git
synced 2025-01-15 22:48:19 +00:00
cc741ed8af
* begin state service * begin on the state trie idea * created beacon state structure * add in the full clone getter * return by value instead * add all setters * new state setters are being completed * arrays roots exposed * close to finishing all these headerssss * functionality complete * added in proto benchmark test * test for compatibility * add test for compat * comments fixed * add clone * add clone * remove underlying copies * make it immutable * integrate it into chainservice * revert * wrap up comments for package * address all comments and godocs * address all comments * clone the pending attestation properly * properly clone remaining items * tests pass fixed bug * begin using it instead of head state * prevent nil pointer exceptions * begin using new struct in db * integrated new type into db package * add proper nil checks * using new state in archiver * refactored much of core * editing all the precompute functions * done with most core refactor * fixed up some bugs in the clone comparisons * append current epoch atts * add missing setters * add new setters * fix other core methods * fix up transition * main service and forkchoice * fix rpc * integrated to powchain * some more changes * fix build * improve processing of deposits * fix error * prevent panic * comment * fix process att * gaz * fix up att process * resolve existing review comments * resolve another batch of gh comments * resolve broken cpt state * revise testutil to use the new state * begin updating the state transition func to pass in more compartmentalized args * finish editing transition function to return errors * block operations pretty much done with refactor * state transition fully refactored * got epoch processing completed * fix build in fork choice * fixing more of the build * fix up broken sync package * it builds nowww it buildssss * revert registry changes * Recompute on Read (#4627) * compute on read * fix up eth1 data votes * looking into slashings bug introduced in core/ * able to advance more slots * add logging * can now sync with testnet yay * remove the leaves algorithm and other merkle imports * expose initialize unsafe funcs * Update beacon-chain/db/kv/state.go * lint Co-authored-by: Raul Jordan <raul@prysmaticlabs.com> * More Optimizations for New State (#4641) * map optimization * more optimizations * use a custom hasher * comment * block operations optimizations * Update beacon-chain/state/types.go Co-Authored-By: Raul Jordan <raul@prysmaticlabs.com> * fixed up various operations to use the validator index map access Co-authored-by: Raul Jordan <raul@prysmaticlabs.com> * archiver tests pass * fixing cache tests * cache tests passing * edited validator tests * powchain tests passing * halfway thru sync tests * more sync test fixes * add in tests for state/ * working through rpc tests * assignments tests passed * almost done with rpc/beacon tests * resolved painful validator test * fixed up even more tests * resolve tests * fix build * reduce a randao mixes copy * fixes under //beacon-chain/blockchain/... * build //beacon-chain/core/... * fixes * Runtime Optimizations (#4648) * parallelize shuffling * clean up * lint * fix build * use callback to read from registry * fix array roots and size map * new improvements * reduce hash allocs * improved shuffling * terence's review * use different method * raul's comment * new array roots * remove clone in pre-compute * Update beacon-chain/state/types.go Co-Authored-By: Raul Jordan <raul@prysmaticlabs.com> * raul's review * lint * fix build issues * fix visibility Co-authored-by: Raul Jordan <raul@prysmaticlabs.com> * fix visibility * build works for all * fix blockchain test * fix a few tests * fix more tests * update validator in slashing * archiver passing * fixed rpc/validator * progress on core tests * resolve broken rpc tests * blockchain tests passed * fix up some tests in core * fix message diff * remove unnecessary save * Save validator after slashing * Update validators one by one * another update * fix everything * fix more precompute tests * fix blocks tests * more elegant fix * more helper fixes * change back ? * fix test * fix skip slot * fix test * reset caches * fix testutil * raceoff fixed * passing * Retrieve cached state in the beginning * lint * Fixed tests part 1 * Fixed rest of the tests * Minor changes to avoid copying, small refactor to reduce deplicated code * Handle att req for slot 0 * New beacon state: Only populate merkle layers as needed, copy merkle layers on copy/clone. (#4689) * Only populate merkle layers as needed, copy merkle layers on copy/clone. * use custom copy * Make maps of correct size * slightly fast, doesn't wait for lock Co-authored-by: prylabs-bulldozer[bot] <58059840+prylabs-bulldozer[bot]@users.noreply.github.com> * Target root can't be 0x00 * Don't use cache for current slot (may not be the right fix) * fixed up tests * Remove some copy for init sync. Not sure if it is safe enough for runtime though... testing... * Align with prev logic for process slots cachedState.Slot() < slot * Fix Initial Sync Flag (#4692) * fixes * fix up some test failures due to lack of nil checks * fix up some test failures due to lack of nil checks * fix up imports * revert some changes * imports Co-authored-by: Raul Jordan <raul@prysmaticlabs.com> * resolving further conflicts * Better skip slot cache (#4694) * Return copy of skip slot cache state, disable skip slot cache on sync * fix * Fix pruning * fix up issues with broken tests Co-authored-by: Nishant Das <nish1993@hotmail.com> Co-authored-by: Preston Van Loon <preston@prysmaticlabs.com> Co-authored-by: shayzluf <thezluf@gmail.com> Co-authored-by: terence tsao <terence@prysmaticlabs.com> Co-authored-by: prylabs-bulldozer[bot] <58059840+prylabs-bulldozer[bot]@users.noreply.github.com>
352 lines
10 KiB
Go
352 lines
10 KiB
Go
package validator
|
|
|
|
import (
|
|
"context"
|
|
"encoding/binary"
|
|
"fmt"
|
|
"strings"
|
|
"testing"
|
|
|
|
ethpb "github.com/prysmaticlabs/ethereumapis/eth/v1alpha1"
|
|
"github.com/prysmaticlabs/go-ssz"
|
|
mockChain "github.com/prysmaticlabs/prysm/beacon-chain/blockchain/testing"
|
|
blk "github.com/prysmaticlabs/prysm/beacon-chain/core/blocks"
|
|
"github.com/prysmaticlabs/prysm/beacon-chain/core/state"
|
|
dbutil "github.com/prysmaticlabs/prysm/beacon-chain/db/testing"
|
|
mockSync "github.com/prysmaticlabs/prysm/beacon-chain/sync/initial-sync/testing"
|
|
"github.com/prysmaticlabs/prysm/shared/bytesutil"
|
|
"github.com/prysmaticlabs/prysm/shared/params"
|
|
"github.com/prysmaticlabs/prysm/shared/testutil"
|
|
)
|
|
|
|
// pubKey is a helper to generate a well-formed public key.
|
|
func pubKey(i uint64) []byte {
|
|
pubKey := make([]byte, params.BeaconConfig().BLSPubkeyLength)
|
|
binary.LittleEndian.PutUint64(pubKey, uint64(i))
|
|
return pubKey
|
|
}
|
|
|
|
func TestGetDuties_NextEpoch_WrongPubkeyLength(t *testing.T) {
|
|
db := dbutil.SetupDB(t)
|
|
defer dbutil.TeardownDB(t, db)
|
|
ctx := context.Background()
|
|
|
|
beaconState, _ := testutil.DeterministicGenesisState(t, 8)
|
|
block := blk.NewGenesisBlock([]byte{})
|
|
if err := db.SaveBlock(ctx, block); err != nil {
|
|
t.Fatalf("Could not save genesis block: %v", err)
|
|
}
|
|
genesisRoot, err := ssz.HashTreeRoot(block.Block)
|
|
if err != nil {
|
|
t.Fatalf("Could not get signing root %v", err)
|
|
}
|
|
|
|
Server := &Server{
|
|
BeaconDB: db,
|
|
HeadFetcher: &mockChain.ChainService{State: beaconState, Root: genesisRoot[:]},
|
|
SyncChecker: &mockSync.Sync{IsSyncing: false},
|
|
}
|
|
req := ðpb.DutiesRequest{
|
|
PublicKeys: [][]byte{{1}},
|
|
Epoch: 0,
|
|
}
|
|
if _, err := Server.GetDuties(context.Background(), req); err != nil && !strings.Contains(err.Error(), "incorrect key length") {
|
|
t.Errorf("Expected \"incorrect key length\", received %v", err)
|
|
}
|
|
}
|
|
|
|
func TestGetDuties_NextEpoch_CantFindValidatorIdx(t *testing.T) {
|
|
db := dbutil.SetupDB(t)
|
|
defer dbutil.TeardownDB(t, db)
|
|
ctx := context.Background()
|
|
beaconState, _ := testutil.DeterministicGenesisState(t, 10)
|
|
|
|
genesis := blk.NewGenesisBlock([]byte{})
|
|
genesisRoot, err := ssz.HashTreeRoot(genesis.Block)
|
|
if err != nil {
|
|
t.Fatalf("Could not get signing root %v", err)
|
|
}
|
|
|
|
vs := &Server{
|
|
BeaconDB: db,
|
|
HeadFetcher: &mockChain.ChainService{State: beaconState, Root: genesisRoot[:]},
|
|
SyncChecker: &mockSync.Sync{IsSyncing: false},
|
|
}
|
|
|
|
pubKey := pubKey(99999)
|
|
req := ðpb.DutiesRequest{
|
|
PublicKeys: [][]byte{pubKey},
|
|
Epoch: 0,
|
|
}
|
|
want := fmt.Sprintf("validator %#x does not exist", req.PublicKeys[0])
|
|
if _, err := vs.GetDuties(ctx, req); err != nil && !strings.Contains(err.Error(), want) {
|
|
t.Errorf("Expected %v, received %v", want, err)
|
|
}
|
|
}
|
|
|
|
func TestGetDuties_OK(t *testing.T) {
|
|
db := dbutil.SetupDB(t)
|
|
defer dbutil.TeardownDB(t, db)
|
|
ctx := context.Background()
|
|
|
|
genesis := blk.NewGenesisBlock([]byte{})
|
|
depChainStart := params.BeaconConfig().MinGenesisActiveValidatorCount
|
|
deposits, _, _ := testutil.DeterministicDepositsAndKeys(depChainStart)
|
|
eth1Data, err := testutil.DeterministicEth1Data(len(deposits))
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
state, err := state.GenesisBeaconState(deposits, 0, eth1Data)
|
|
if err != nil {
|
|
t.Fatalf("Could not setup genesis state: %v", err)
|
|
}
|
|
genesisRoot, err := ssz.HashTreeRoot(genesis.Block)
|
|
if err != nil {
|
|
t.Fatalf("Could not get signing root %v", err)
|
|
}
|
|
|
|
pubKeys := make([][]byte, len(deposits))
|
|
indices := make([]uint64, len(deposits))
|
|
for i := 0; i < len(deposits); i++ {
|
|
pubKeys[i] = deposits[i].Data.PublicKey
|
|
indices[i] = uint64(i)
|
|
}
|
|
|
|
pubkeysAs48ByteType := make([][48]byte, len(pubKeys))
|
|
for i, pk := range pubKeys {
|
|
pubkeysAs48ByteType[i] = bytesutil.ToBytes48(pk)
|
|
}
|
|
if err := db.SaveValidatorIndices(ctx, pubkeysAs48ByteType, indices); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
vs := &Server{
|
|
BeaconDB: db,
|
|
HeadFetcher: &mockChain.ChainService{State: state, Root: genesisRoot[:]},
|
|
SyncChecker: &mockSync.Sync{IsSyncing: false},
|
|
}
|
|
|
|
// Test the first validator in registry.
|
|
req := ðpb.DutiesRequest{
|
|
PublicKeys: [][]byte{deposits[0].Data.PublicKey},
|
|
Epoch: 0,
|
|
}
|
|
res, err := vs.GetDuties(context.Background(), req)
|
|
if err != nil {
|
|
t.Fatalf("Could not call epoch committee assignment %v", err)
|
|
}
|
|
if res.Duties[0].AttesterSlot > state.Slot()+params.BeaconConfig().SlotsPerEpoch {
|
|
t.Errorf("Assigned slot %d can't be higher than %d",
|
|
res.Duties[0].AttesterSlot, state.Slot()+params.BeaconConfig().SlotsPerEpoch)
|
|
}
|
|
|
|
// Test the last validator in registry.
|
|
lastValidatorIndex := depChainStart - 1
|
|
req = ðpb.DutiesRequest{
|
|
PublicKeys: [][]byte{deposits[lastValidatorIndex].Data.PublicKey},
|
|
Epoch: 0,
|
|
}
|
|
res, err = vs.GetDuties(context.Background(), req)
|
|
if err != nil {
|
|
t.Fatalf("Could not call epoch committee assignment %v", err)
|
|
}
|
|
if res.Duties[0].AttesterSlot > state.Slot()+params.BeaconConfig().SlotsPerEpoch {
|
|
t.Errorf("Assigned slot %d can't be higher than %d",
|
|
res.Duties[0].AttesterSlot, state.Slot()+params.BeaconConfig().SlotsPerEpoch)
|
|
}
|
|
|
|
// We request for duties for all validators.
|
|
req = ðpb.DutiesRequest{
|
|
PublicKeys: pubKeys,
|
|
Epoch: 0,
|
|
}
|
|
res, err = vs.GetDuties(context.Background(), req)
|
|
if err != nil {
|
|
t.Fatalf("Could not call epoch committee assignment %v", err)
|
|
}
|
|
for i := 0; i < len(res.Duties); i++ {
|
|
if res.Duties[i].ValidatorIndex != uint64(i) {
|
|
t.Errorf("Wanted %d, received %d", i, res.Duties[i].ValidatorIndex)
|
|
}
|
|
}
|
|
}
|
|
|
|
func TestGetDuties_CurrentEpoch_ShouldNotFail(t *testing.T) {
|
|
db := dbutil.SetupDB(t)
|
|
defer dbutil.TeardownDB(t, db)
|
|
ctx := context.Background()
|
|
|
|
genesis := blk.NewGenesisBlock([]byte{})
|
|
depChainStart := params.BeaconConfig().MinGenesisActiveValidatorCount
|
|
deposits, _, _ := testutil.DeterministicDepositsAndKeys(depChainStart)
|
|
eth1Data, err := testutil.DeterministicEth1Data(len(deposits))
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
bState, err := state.GenesisBeaconState(deposits, 0, eth1Data)
|
|
if err != nil {
|
|
t.Fatalf("Could not setup genesis state: %v", err)
|
|
}
|
|
bState.SetSlot(5) // Set state to non-epoch start slot.
|
|
|
|
genesisRoot, err := ssz.HashTreeRoot(genesis.Block)
|
|
if err != nil {
|
|
t.Fatalf("Could not get signing root %v", err)
|
|
}
|
|
|
|
pubKeys := make([][48]byte, len(deposits))
|
|
indices := make([]uint64, len(deposits))
|
|
for i := 0; i < len(deposits); i++ {
|
|
pubKeys[i] = bytesutil.ToBytes48(deposits[i].Data.PublicKey)
|
|
indices[i] = uint64(i)
|
|
}
|
|
if err := db.SaveValidatorIndices(ctx, pubKeys, indices); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
vs := &Server{
|
|
BeaconDB: db,
|
|
HeadFetcher: &mockChain.ChainService{State: bState, Root: genesisRoot[:]},
|
|
SyncChecker: &mockSync.Sync{IsSyncing: false},
|
|
}
|
|
|
|
// Test the first validator in registry.
|
|
req := ðpb.DutiesRequest{
|
|
PublicKeys: [][]byte{deposits[0].Data.PublicKey},
|
|
Epoch: 0,
|
|
}
|
|
res, err := vs.GetDuties(context.Background(), req)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if len(res.Duties) != 1 {
|
|
t.Error("Expected 1 assignment")
|
|
}
|
|
}
|
|
|
|
func TestGetDuties_MultipleKeys_OK(t *testing.T) {
|
|
db := dbutil.SetupDB(t)
|
|
defer dbutil.TeardownDB(t, db)
|
|
ctx := context.Background()
|
|
|
|
genesis := blk.NewGenesisBlock([]byte{})
|
|
depChainStart := uint64(64)
|
|
deposits, _, _ := testutil.DeterministicDepositsAndKeys(depChainStart)
|
|
eth1Data, err := testutil.DeterministicEth1Data(len(deposits))
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
state, err := state.GenesisBeaconState(deposits, 0, eth1Data)
|
|
if err != nil {
|
|
t.Fatalf("Could not setup genesis state: %v", err)
|
|
}
|
|
genesisRoot, err := ssz.HashTreeRoot(genesis.Block)
|
|
if err != nil {
|
|
t.Fatalf("Could not get signing root %v", err)
|
|
}
|
|
|
|
pubKeys := make([][48]byte, len(deposits))
|
|
indices := make([]uint64, len(deposits))
|
|
for i := 0; i < len(deposits); i++ {
|
|
pubKeys[i] = bytesutil.ToBytes48(deposits[i].Data.PublicKey)
|
|
indices[i] = uint64(i)
|
|
}
|
|
if err := db.SaveValidatorIndices(ctx, pubKeys, indices); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
vs := &Server{
|
|
BeaconDB: db,
|
|
HeadFetcher: &mockChain.ChainService{State: state, Root: genesisRoot[:]},
|
|
SyncChecker: &mockSync.Sync{IsSyncing: false},
|
|
}
|
|
|
|
pubkey0 := deposits[0].Data.PublicKey
|
|
pubkey1 := deposits[1].Data.PublicKey
|
|
|
|
// Test the first validator in registry.
|
|
req := ðpb.DutiesRequest{
|
|
PublicKeys: [][]byte{pubkey0, pubkey1},
|
|
Epoch: 0,
|
|
}
|
|
res, err := vs.GetDuties(context.Background(), req)
|
|
if err != nil {
|
|
t.Fatalf("Could not call epoch committee assignment %v", err)
|
|
}
|
|
|
|
if len(res.Duties) != 2 {
|
|
t.Errorf("expected 2 assignments but got %d", len(res.Duties))
|
|
}
|
|
if res.Duties[0].AttesterSlot != 4 {
|
|
t.Errorf("Expected res.Duties[0].AttesterSlot == 4, got %d", res.Duties[0].AttesterSlot)
|
|
}
|
|
if res.Duties[1].AttesterSlot != 3 {
|
|
t.Errorf("Expected res.Duties[1].AttesterSlot == 3, got %d", res.Duties[0].AttesterSlot)
|
|
}
|
|
}
|
|
|
|
func TestGetDuties_SyncNotReady(t *testing.T) {
|
|
vs := &Server{
|
|
SyncChecker: &mockSync.Sync{IsSyncing: true},
|
|
}
|
|
_, err := vs.GetDuties(context.Background(), ðpb.DutiesRequest{})
|
|
if strings.Contains(err.Error(), "syncing to latest head") {
|
|
t.Error("Did not get wanted error")
|
|
}
|
|
}
|
|
|
|
func BenchmarkCommitteeAssignment(b *testing.B) {
|
|
db := dbutil.SetupDB(b)
|
|
defer dbutil.TeardownDB(b, db)
|
|
ctx := context.Background()
|
|
|
|
genesis := blk.NewGenesisBlock([]byte{})
|
|
depChainStart := uint64(8192 * 2)
|
|
deposits, _, _ := testutil.DeterministicDepositsAndKeys(depChainStart)
|
|
eth1Data, err := testutil.DeterministicEth1Data(len(deposits))
|
|
if err != nil {
|
|
b.Fatal(err)
|
|
}
|
|
state, err := state.GenesisBeaconState(deposits, 0, eth1Data)
|
|
if err != nil {
|
|
b.Fatalf("Could not setup genesis state: %v", err)
|
|
}
|
|
genesisRoot, err := ssz.HashTreeRoot(genesis.Block)
|
|
if err != nil {
|
|
b.Fatalf("Could not get signing root %v", err)
|
|
}
|
|
|
|
pubKeys := make([][48]byte, len(deposits))
|
|
indices := make([]uint64, len(deposits))
|
|
for i := 0; i < len(deposits); i++ {
|
|
pubKeys[i] = bytesutil.ToBytes48(deposits[i].Data.PublicKey)
|
|
indices[i] = uint64(i)
|
|
}
|
|
if err := db.SaveValidatorIndices(ctx, pubKeys, indices); err != nil {
|
|
b.Fatal(err)
|
|
}
|
|
|
|
vs := &Server{
|
|
BeaconDB: db,
|
|
HeadFetcher: &mockChain.ChainService{State: state, Root: genesisRoot[:]},
|
|
SyncChecker: &mockSync.Sync{IsSyncing: false},
|
|
}
|
|
|
|
// Create request for all validators in the system.
|
|
pks := make([][]byte, len(deposits))
|
|
for i, deposit := range deposits {
|
|
pks[i] = deposit.Data.PublicKey
|
|
}
|
|
req := ðpb.DutiesRequest{
|
|
PublicKeys: pks,
|
|
Epoch: 0,
|
|
}
|
|
b.ResetTimer()
|
|
for i := 0; i < b.N; i++ {
|
|
_, err := vs.GetDuties(context.Background(), req)
|
|
if err != nil {
|
|
b.Error(err)
|
|
}
|
|
}
|
|
}
|