mirror of
https://gitlab.com/pulsechaincom/prysm-pulse.git
synced 2025-01-03 16:37:39 +00:00
81c53c26fb
* Update seed domains (#3872) * Remove Transfers (#3870) * Remove active index roots and compact committee roots (#3869) * Update inclusion reward (#3886) * Alter proposer selection logic (#3884) * Fix early committee bias (#3888) * Remove shards and committees (#3896) * Epoch spec tests v0.9 (#3907) * Block spec test v0.9 (#3905) * rm'ed in protobuf * build proto * build proto * build proto * fix core package * Gazelle * Fixed all the tests * Fixed static test * Comment out spec test for now * One more skip * fix-roundRobinSync (#3862) * Starting but need new seed function * Revert initial sync * Updated Proposer Slashing * Fixed all tests * Lint * Update inclusion reward * Fill randao mixes with eth1 data hash * Test * Fixing test part1 * All tests passing * One last test * Updated config * Build proto * Proper skip message * Conflict and fmt * Removed crosslinks and shards. Built * Format and gazelle * Fixed all the block package tests * Fixed all the helper tests * All epoch package tests pass * All core package tests pass * Fixed operation tests * Started fixing rpc test * RPC tests passed! * Fixed all init sync tests * All tests pass * Fixed blockchain tests * Lint * Lint * Preston's feedback * Starting * Remove container * Fixed block spec tests * All passing except for block_processing test * Failing block processing test * Starting * Add AggregateAndProof * All mainnet test passes * Update deposit contract (#3906) * Proto spec tests v0.9 (#3908) * Starting * Add AggregateAndProof * Unskip block util tests (#3910) * rm'ed in protobuf * build proto * build proto * build proto * fix core package * Gazelle * Fixed all the tests * Fixed static test * Comment out spec test for now * One more skip * fix-roundRobinSync (#3862) * Starting but need new seed function * Revert initial sync * Updated Proposer Slashing * Fixed all tests * Lint * Update inclusion reward * Fill randao mixes with eth1 data hash * Test * Fixing test part1 * All tests passing * One last test * Updated config * Build proto * Proper skip message * Conflict and fmt * Removed crosslinks and shards. Built * Format and gazelle * Fixed all the block package tests * Fixed all the helper tests * All epoch package tests pass * All core package tests pass * Fixed operation tests * Started fixing rpc test * RPC tests passed! * Fixed all init sync tests * All tests pass * Fixed blockchain tests * Lint * Lint * Preston's feedback * Starting * Remove container * Fixed block spec tests * All passing except for block_processing test * Failing block processing test * Starting * Add AggregateAndProof * All mainnet test passes * Unskip block util tests * Slot processing spec test V0.9 (#3912) * Starting * Add AggregateAndProof * Unskip slot processing mainnet test * Unskip minimal spec test for finalization (#3920) * Remove outdated interop tests (#3922) * Rm outdated interop tests * Rm test runner * Gazelle * Update validator to use proposer slot (#3919) * Fix committee assignment (#3931) * Replace shard with committee index (#3930) * Conflict * Clean up (#3933) * Remove shard filter in db (#3936) * Remove lightouse compatibility test (#3939) * Update Committee Cache for v0.9 (#3948) * Updated committee cache * Removed shuffled indices cache * Started testing run time * Lint * Fixed test * Safeguard against nil head state * address edge case * add test * Fixed TestRoundRobinSync by doubling the epochs * Unskip TestProtoCompatability (#3958) * Unskip TestProtoCompatability * Update WORKSPACE * Fix minimal config (#3959) * fix minimal configs * fix hardcoded value in test * Simplify verify att time (#3961) * update readme for deposit contract, regen bindings for vyper 0.1.0b12 (#3963) * update readme for deposit contract, regen bindings * medium * Check nil base state (#3964) * Copy Block When Receiving it From Sync (#3966) * copy block * clone for other service methods too * Change logging of Bitfield (#3956) * change logging of bits * preston's review * Unskip Beacon Server Test (#3962) * run test till the end * fix up proto message types * fmt * resolve broken tests * better error handling * fixing new logic to use archived proposer info * fix up logic * clip using the max effective balance * broken build fix with num arg mismatch * amend archive * archival logic changed * rename test * archive both proposer and attester seeds * page size 100 * further experiments * further experimentation, archivedProposerIndex seems wrong * test passes * rem log * fix broken test * fix test * gaz * fix imports * ethapis
344 lines
10 KiB
Go
344 lines
10 KiB
Go
package forkchoice
|
|
|
|
import (
|
|
"context"
|
|
"reflect"
|
|
"strings"
|
|
"testing"
|
|
|
|
"github.com/prysmaticlabs/go-bitfield"
|
|
"github.com/prysmaticlabs/go-ssz"
|
|
"github.com/prysmaticlabs/prysm/beacon-chain/core/blocks"
|
|
"github.com/prysmaticlabs/prysm/beacon-chain/core/helpers"
|
|
"github.com/prysmaticlabs/prysm/beacon-chain/core/state"
|
|
testDB "github.com/prysmaticlabs/prysm/beacon-chain/db/testing"
|
|
pb "github.com/prysmaticlabs/prysm/proto/beacon/p2p/v1"
|
|
ethpb "github.com/prysmaticlabs/prysm/proto/eth/v1alpha1"
|
|
"github.com/prysmaticlabs/prysm/shared/bytesutil"
|
|
"github.com/prysmaticlabs/prysm/shared/featureconfig"
|
|
"github.com/prysmaticlabs/prysm/shared/hashutil"
|
|
"github.com/prysmaticlabs/prysm/shared/params"
|
|
"github.com/prysmaticlabs/prysm/shared/testutil"
|
|
)
|
|
|
|
func init() {
|
|
fc := featureconfig.Get()
|
|
fc.PruneFinalizedStates = true
|
|
featureconfig.Init(fc)
|
|
}
|
|
|
|
func TestStore_OnBlock(t *testing.T) {
|
|
ctx := context.Background()
|
|
db := testDB.SetupDB(t)
|
|
defer testDB.TeardownDB(t, db)
|
|
|
|
store := NewForkChoiceService(ctx, db)
|
|
|
|
roots, err := blockTree1(db)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
randomParentRoot := []byte{'a'}
|
|
if err := store.db.SaveState(ctx, &pb.BeaconState{}, bytesutil.ToBytes32(randomParentRoot)); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
randomParentRoot2 := roots[1]
|
|
if err := store.db.SaveState(ctx, &pb.BeaconState{}, bytesutil.ToBytes32(randomParentRoot2)); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
validGenesisRoot := []byte{'g'}
|
|
if err := store.db.SaveState(ctx, &pb.BeaconState{}, bytesutil.ToBytes32(validGenesisRoot)); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
tests := []struct {
|
|
name string
|
|
blk *ethpb.BeaconBlock
|
|
s *pb.BeaconState
|
|
time uint64
|
|
wantErrString string
|
|
}{
|
|
{
|
|
name: "parent block root does not have a state",
|
|
blk: ðpb.BeaconBlock{},
|
|
s: &pb.BeaconState{},
|
|
wantErrString: "pre state of slot 0 does not exist",
|
|
},
|
|
{
|
|
name: "block is from the feature",
|
|
blk: ðpb.BeaconBlock{ParentRoot: randomParentRoot, Slot: params.BeaconConfig().FarFutureEpoch},
|
|
s: &pb.BeaconState{},
|
|
wantErrString: "could not process slot from the future",
|
|
},
|
|
{
|
|
name: "could not get finalized block",
|
|
blk: ðpb.BeaconBlock{ParentRoot: randomParentRoot},
|
|
s: &pb.BeaconState{},
|
|
wantErrString: "block from slot 0 is not a descendent of the current finalized block",
|
|
},
|
|
{
|
|
name: "same slot as finalized block",
|
|
blk: ðpb.BeaconBlock{Slot: 0, ParentRoot: randomParentRoot2},
|
|
s: &pb.BeaconState{},
|
|
wantErrString: "block is equal or earlier than finalized block, slot 0 < slot 0",
|
|
},
|
|
}
|
|
|
|
for _, tt := range tests {
|
|
t.Run(tt.name, func(t *testing.T) {
|
|
if err := store.GenesisStore(ctx, ðpb.Checkpoint{}, ðpb.Checkpoint{}); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
store.finalizedCheckpt.Root = roots[0]
|
|
|
|
err := store.OnBlock(ctx, tt.blk)
|
|
if !strings.Contains(err.Error(), tt.wantErrString) {
|
|
t.Errorf("Store.OnBlock() error = %v, wantErr = %v", err, tt.wantErrString)
|
|
}
|
|
})
|
|
}
|
|
}
|
|
|
|
func TestStore_SaveNewValidators(t *testing.T) {
|
|
ctx := context.Background()
|
|
db := testDB.SetupDB(t)
|
|
defer testDB.TeardownDB(t, db)
|
|
|
|
store := NewForkChoiceService(ctx, db)
|
|
preCount := 2 // validators 0 and validators 1
|
|
s := &pb.BeaconState{Validators: []*ethpb.Validator{
|
|
{PublicKey: []byte{0}}, {PublicKey: []byte{1}},
|
|
{PublicKey: []byte{2}}, {PublicKey: []byte{3}},
|
|
}}
|
|
if err := store.saveNewValidators(ctx, preCount, s); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
if !db.HasValidatorIndex(ctx, bytesutil.ToBytes48([]byte{2})) {
|
|
t.Error("Wanted validator saved in db")
|
|
}
|
|
if !db.HasValidatorIndex(ctx, bytesutil.ToBytes48([]byte{3})) {
|
|
t.Error("Wanted validator saved in db")
|
|
}
|
|
if db.HasValidatorIndex(ctx, bytesutil.ToBytes48([]byte{1})) {
|
|
t.Error("validator not suppose to be saved in db")
|
|
}
|
|
}
|
|
|
|
func TestStore_UpdateBlockAttestationVote(t *testing.T) {
|
|
ctx := context.Background()
|
|
db := testDB.SetupDB(t)
|
|
defer testDB.TeardownDB(t, db)
|
|
params.UseMinimalConfig()
|
|
|
|
deposits, _, _ := testutil.SetupInitialDeposits(t, 100)
|
|
beaconState, err := state.GenesisBeaconState(deposits, uint64(0), ðpb.Eth1Data{BlockHash: make([]byte, 32)})
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
store := NewForkChoiceService(ctx, db)
|
|
r := [32]byte{'A'}
|
|
att := ðpb.Attestation{
|
|
Data: ðpb.AttestationData{
|
|
Source: ðpb.Checkpoint{Epoch: 0, Root: params.BeaconConfig().ZeroHash[:]},
|
|
Target: ðpb.Checkpoint{Epoch: 0, Root: r[:]},
|
|
},
|
|
AggregationBits: []byte{255},
|
|
CustodyBits: []byte{255},
|
|
}
|
|
if err := store.db.SaveState(ctx, beaconState, r); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
indices, err := blocks.ConvertToIndexed(ctx, beaconState, att)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
var attestedIndices []uint64
|
|
for _, k := range append(indices.CustodyBit_0Indices, indices.CustodyBit_1Indices...) {
|
|
attestedIndices = append(attestedIndices, k)
|
|
}
|
|
|
|
if err := store.updateBlockAttestationVote(ctx, att); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
for _, i := range attestedIndices {
|
|
v, err := store.db.ValidatorLatestVote(ctx, i)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if !reflect.DeepEqual(v.Root, r[:]) {
|
|
t.Error("Attested roots don't match")
|
|
}
|
|
}
|
|
}
|
|
|
|
func TestStore_UpdateBlockAttestationsVote(t *testing.T) {
|
|
ctx := context.Background()
|
|
db := testDB.SetupDB(t)
|
|
defer testDB.TeardownDB(t, db)
|
|
params.UseMinimalConfig()
|
|
|
|
deposits, _, _ := testutil.SetupInitialDeposits(t, 100)
|
|
beaconState, err := state.GenesisBeaconState(deposits, uint64(0), ðpb.Eth1Data{BlockHash: make([]byte, 32)})
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
store := NewForkChoiceService(ctx, db)
|
|
r := [32]byte{'A'}
|
|
atts := make([]*ethpb.Attestation, 5)
|
|
hashes := make([][32]byte, 5)
|
|
for i := 0; i < len(atts); i++ {
|
|
atts[i] = ðpb.Attestation{
|
|
Data: ðpb.AttestationData{
|
|
Source: ðpb.Checkpoint{Epoch: 0, Root: params.BeaconConfig().ZeroHash[:]},
|
|
Target: ðpb.Checkpoint{Epoch: 0, Root: r[:]},
|
|
},
|
|
AggregationBits: []byte{255},
|
|
CustodyBits: []byte{255},
|
|
}
|
|
h, _ := hashutil.HashProto(atts[i])
|
|
hashes[i] = h
|
|
}
|
|
|
|
if err := store.db.SaveState(ctx, beaconState, r); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
if err := store.updateBlockAttestationsVotes(ctx, atts); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
for _, h := range hashes {
|
|
if !store.seenAtts[h] {
|
|
t.Error("Seen attestation did not get recorded")
|
|
}
|
|
}
|
|
}
|
|
|
|
func TestStore_SavesNewBlockAttestations(t *testing.T) {
|
|
ctx := context.Background()
|
|
db := testDB.SetupDB(t)
|
|
defer testDB.TeardownDB(t, db)
|
|
|
|
store := NewForkChoiceService(ctx, db)
|
|
a1 := ðpb.Attestation{Data: ðpb.AttestationData{}, AggregationBits: bitfield.Bitlist{0b101}, CustodyBits: bitfield.NewBitlist(2)}
|
|
a2 := ðpb.Attestation{Data: ðpb.AttestationData{BeaconBlockRoot: []byte{'A'}}, AggregationBits: bitfield.Bitlist{0b110}, CustodyBits: bitfield.NewBitlist(2)}
|
|
r1, _ := ssz.HashTreeRoot(a1.Data)
|
|
r2, _ := ssz.HashTreeRoot(a2.Data)
|
|
|
|
if err := store.saveNewBlockAttestations(ctx, []*ethpb.Attestation{a1, a2}); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
saved, err := store.db.AttestationsByDataRoot(ctx, r1)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if !reflect.DeepEqual([]*ethpb.Attestation{a1}, saved) {
|
|
t.Error("did not retrieve saved attestation")
|
|
}
|
|
|
|
saved, err = store.db.AttestationsByDataRoot(ctx, r2)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if !reflect.DeepEqual([]*ethpb.Attestation{a2}, saved) {
|
|
t.Error("did not retrieve saved attestation")
|
|
}
|
|
|
|
a1 = ðpb.Attestation{Data: ðpb.AttestationData{}, AggregationBits: bitfield.Bitlist{0b111}, CustodyBits: bitfield.NewBitlist(2)}
|
|
a2 = ðpb.Attestation{Data: ðpb.AttestationData{BeaconBlockRoot: []byte{'A'}}, AggregationBits: bitfield.Bitlist{0b111}, CustodyBits: bitfield.NewBitlist(2)}
|
|
|
|
if err := store.saveNewBlockAttestations(ctx, []*ethpb.Attestation{a1, a2}); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
saved, err = store.db.AttestationsByDataRoot(ctx, r1)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if !reflect.DeepEqual([]*ethpb.Attestation{a1}, saved) {
|
|
t.Error("did not retrieve saved attestation")
|
|
}
|
|
|
|
saved, err = store.db.AttestationsByDataRoot(ctx, r2)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if !reflect.DeepEqual([]*ethpb.Attestation{a2}, saved) {
|
|
t.Error("did not retrieve saved attestation")
|
|
}
|
|
}
|
|
|
|
func TestRemoveStateSinceLastFinalized(t *testing.T) {
|
|
ctx := context.Background()
|
|
db := testDB.SetupDB(t)
|
|
defer testDB.TeardownDB(t, db)
|
|
params.UseMinimalConfig()
|
|
defer params.UseMainnetConfig()
|
|
|
|
store := NewForkChoiceService(ctx, db)
|
|
|
|
// Save 100 blocks in DB, each has a state.
|
|
numBlocks := 100
|
|
totalBlocks := make([]*ethpb.BeaconBlock, numBlocks)
|
|
blockRoots := make([][32]byte, 0)
|
|
for i := 0; i < len(totalBlocks); i++ {
|
|
totalBlocks[i] = ðpb.BeaconBlock{
|
|
Slot: uint64(i),
|
|
}
|
|
r, err := ssz.SigningRoot(totalBlocks[i])
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if err := store.db.SaveState(ctx, &pb.BeaconState{Slot: uint64(i)}, r); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if err := store.db.SaveBlock(ctx, totalBlocks[i]); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
blockRoots = append(blockRoots, r)
|
|
}
|
|
|
|
// New finalized epoch: 1
|
|
finalizedEpoch := uint64(1)
|
|
finalizedSlot := finalizedEpoch * params.BeaconConfig().SlotsPerEpoch
|
|
endSlot := helpers.StartSlot(finalizedEpoch+1) - 1 // Inclusive
|
|
if err := store.rmStatesOlderThanLastFinalized(ctx, 0, endSlot); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
for _, r := range blockRoots {
|
|
s, err := store.db.State(ctx, r)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
// Also verifies genesis state didnt get deleted
|
|
if s != nil && s.Slot != finalizedSlot && s.Slot != 0 && s.Slot < endSlot {
|
|
t.Errorf("State with slot %d should not be in DB", s.Slot)
|
|
}
|
|
}
|
|
|
|
// New finalized epoch: 5
|
|
newFinalizedEpoch := uint64(5)
|
|
newFinalizedSlot := newFinalizedEpoch * params.BeaconConfig().SlotsPerEpoch
|
|
endSlot = helpers.StartSlot(newFinalizedEpoch+1) - 1 // Inclusive
|
|
if err := store.rmStatesOlderThanLastFinalized(ctx, helpers.StartSlot(finalizedEpoch+1)-1, endSlot); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
for _, r := range blockRoots {
|
|
s, err := store.db.State(ctx, r)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
// Also verifies genesis state didnt get deleted
|
|
if s != nil && s.Slot != newFinalizedSlot && s.Slot != finalizedSlot && s.Slot != 0 && s.Slot < endSlot {
|
|
t.Errorf("State with slot %d should not be in DB", s.Slot)
|
|
}
|
|
}
|
|
}
|