mirror of
https://gitlab.com/pulsechaincom/prysm-pulse.git
synced 2025-01-06 01:32:18 +00:00
7f7866ff2a
* Starting a quick PoC * Rate limit to one epoch worth of blocks in memory * Proof of concept working * Quick comment out * Save previous finalized checkpoint * Test * Minor fixes * More run time fixes * Remove panic * Feature flag * Removed unused methods * Fixed tests * E2e test * comment * Compatible with current initial sync * Starting * New cache * Cache getters and setters * It should be part of state gen * Need to use cache for DB * Don't have to use finalized state * Rm unused file * some changes to memory mgmt when using mempool * More run time fixes * Can sync to head * Feedback * Revert "some changes to memory mgmt when using mempool" This reverts commit f5b3e7ff4714fef9f0397007f519a45fa259ad24. * Fixed sync tests * Fixed existing tests * Test for state summary getter * Gaz * Fix kafka passthrough * Fixed inputs * Gaz * Fixed build * Fixed visibility * Trying without the ignore * Didn't work.. * Fix kafka Co-authored-by: prylabs-bulldozer[bot] <58059840+prylabs-bulldozer[bot]@users.noreply.github.com> Co-authored-by: Preston Van Loon <preston@prysmaticlabs.com>
219 lines
6.4 KiB
Go
219 lines
6.4 KiB
Go
package stategen
|
|
|
|
import (
|
|
"context"
|
|
"testing"
|
|
|
|
"github.com/gogo/protobuf/proto"
|
|
ethpb "github.com/prysmaticlabs/ethereumapis/eth/v1alpha1"
|
|
"github.com/prysmaticlabs/go-ssz"
|
|
"github.com/prysmaticlabs/prysm/beacon-chain/cache"
|
|
testDB "github.com/prysmaticlabs/prysm/beacon-chain/db/testing"
|
|
pb "github.com/prysmaticlabs/prysm/proto/beacon/p2p/v1"
|
|
|
|
//pb "github.com/prysmaticlabs/prysm/proto/beacon/p2p/v1"
|
|
"github.com/prysmaticlabs/prysm/shared/params"
|
|
"github.com/prysmaticlabs/prysm/shared/testutil"
|
|
logTest "github.com/sirupsen/logrus/hooks/test"
|
|
)
|
|
|
|
func TestSaveHotState_AlreadyHas(t *testing.T) {
|
|
hook := logTest.NewGlobal()
|
|
ctx := context.Background()
|
|
db := testDB.SetupDB(t)
|
|
defer testDB.TeardownDB(t, db)
|
|
service := New(db, cache.NewStateSummaryCache())
|
|
|
|
beaconState, _ := testutil.DeterministicGenesisState(t, 32)
|
|
beaconState.SetSlot(params.BeaconConfig().SlotsPerEpoch)
|
|
r := [32]byte{'A'}
|
|
|
|
// Pre cache the hot state.
|
|
service.hotStateCache.Put(r, beaconState)
|
|
if err := service.saveHotState(ctx, r, beaconState); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Should not save the state and state summary.
|
|
if service.beaconDB.HasState(ctx, r) {
|
|
t.Error("Should not have saved the state")
|
|
}
|
|
if service.beaconDB.HasStateSummary(ctx, r) {
|
|
t.Error("Should have saved the state summary")
|
|
}
|
|
testutil.AssertLogsDoNotContain(t, hook, "Saved full state on epoch boundary")
|
|
}
|
|
|
|
func TestSaveHotState_CanSaveOnEpochBoundary(t *testing.T) {
|
|
hook := logTest.NewGlobal()
|
|
ctx := context.Background()
|
|
db := testDB.SetupDB(t)
|
|
defer testDB.TeardownDB(t, db)
|
|
service := New(db, cache.NewStateSummaryCache())
|
|
|
|
beaconState, _ := testutil.DeterministicGenesisState(t, 32)
|
|
beaconState.SetSlot(params.BeaconConfig().SlotsPerEpoch)
|
|
r := [32]byte{'A'}
|
|
|
|
if err := service.saveHotState(ctx, r, beaconState); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Should save both state and state summary.
|
|
if !service.beaconDB.HasState(ctx, r) {
|
|
t.Error("Should have saved the state")
|
|
}
|
|
if !service.stateSummaryCache.Has(r) {
|
|
t.Error("Should have saved the state summary")
|
|
}
|
|
testutil.AssertLogsContain(t, hook, "Saved full state on epoch boundary")
|
|
}
|
|
|
|
func TestSaveHotState_NoSaveNotEpochBoundary(t *testing.T) {
|
|
hook := logTest.NewGlobal()
|
|
ctx := context.Background()
|
|
db := testDB.SetupDB(t)
|
|
defer testDB.TeardownDB(t, db)
|
|
service := New(db, cache.NewStateSummaryCache())
|
|
|
|
beaconState, _ := testutil.DeterministicGenesisState(t, 32)
|
|
beaconState.SetSlot(params.BeaconConfig().SlotsPerEpoch - 1)
|
|
r := [32]byte{'A'}
|
|
b := ðpb.SignedBeaconBlock{Block: ðpb.BeaconBlock{}}
|
|
if err := db.SaveBlock(ctx, b); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
gRoot, _ := ssz.HashTreeRoot(b.Block)
|
|
if err := db.SaveGenesisBlockRoot(ctx, gRoot); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
if err := service.saveHotState(ctx, r, beaconState); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Should only save state summary.
|
|
if service.beaconDB.HasState(ctx, r) {
|
|
t.Error("Should not have saved the state")
|
|
}
|
|
if !service.stateSummaryCache.Has(r) {
|
|
t.Error("Should have saved the state summary")
|
|
}
|
|
testutil.AssertLogsDoNotContain(t, hook, "Saved full state on epoch boundary")
|
|
}
|
|
|
|
func TestLoadHoteStateByRoot_Cached(t *testing.T) {
|
|
ctx := context.Background()
|
|
db := testDB.SetupDB(t)
|
|
defer testDB.TeardownDB(t, db)
|
|
service := New(db, cache.NewStateSummaryCache())
|
|
|
|
beaconState, _ := testutil.DeterministicGenesisState(t, 32)
|
|
r := [32]byte{'A'}
|
|
service.hotStateCache.Put(r, beaconState)
|
|
|
|
// This tests where hot state was already cached.
|
|
loadedState, err := service.loadHotStateByRoot(ctx, r)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
if !proto.Equal(loadedState.InnerStateUnsafe(), beaconState.InnerStateUnsafe()) {
|
|
t.Error("Did not correctly cache state")
|
|
}
|
|
}
|
|
|
|
func TestLoadHoteStateByRoot_FromDBCanProcess(t *testing.T) {
|
|
ctx := context.Background()
|
|
db := testDB.SetupDB(t)
|
|
defer testDB.TeardownDB(t, db)
|
|
service := New(db, cache.NewStateSummaryCache())
|
|
|
|
beaconState, _ := testutil.DeterministicGenesisState(t, 32)
|
|
blk := ðpb.SignedBeaconBlock{Block: ðpb.BeaconBlock{}}
|
|
blkRoot, _ := ssz.HashTreeRoot(blk.Block)
|
|
service.beaconDB.SaveGenesisBlockRoot(ctx, blkRoot)
|
|
if err := service.beaconDB.SaveState(ctx, beaconState, blkRoot); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
targetSlot := uint64(10)
|
|
targetRoot := [32]byte{'a'}
|
|
if err := service.beaconDB.SaveStateSummary(ctx, &pb.StateSummary{
|
|
Slot: targetSlot,
|
|
Root: targetRoot[:],
|
|
}); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// This tests where hot state was not cached and needs processing.
|
|
loadedState, err := service.loadHotStateByRoot(ctx, targetRoot)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
if loadedState.Slot() != targetSlot {
|
|
t.Error("Did not correctly load state")
|
|
}
|
|
}
|
|
|
|
func TestLoadHoteStateByRoot_FromDBBoundaryCase(t *testing.T) {
|
|
ctx := context.Background()
|
|
db := testDB.SetupDB(t)
|
|
defer testDB.TeardownDB(t, db)
|
|
service := New(db, cache.NewStateSummaryCache())
|
|
|
|
beaconState, _ := testutil.DeterministicGenesisState(t, 32)
|
|
blk := ðpb.SignedBeaconBlock{Block: ðpb.BeaconBlock{}}
|
|
blkRoot, _ := ssz.HashTreeRoot(blk.Block)
|
|
service.beaconDB.SaveGenesisBlockRoot(ctx, blkRoot)
|
|
if err := service.beaconDB.SaveState(ctx, beaconState, blkRoot); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
targetSlot := uint64(0)
|
|
if err := service.beaconDB.SaveStateSummary(ctx, &pb.StateSummary{
|
|
Slot: targetSlot,
|
|
Root: blkRoot[:],
|
|
}); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// This tests where hot state was not cached but doesn't need processing
|
|
// because it on the epoch boundary slot.
|
|
loadedState, err := service.loadHotStateByRoot(ctx, blkRoot)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
if loadedState.Slot() != targetSlot {
|
|
t.Error("Did not correctly load state")
|
|
}
|
|
}
|
|
|
|
func TestLoadHoteStateBySlot_CanAdvanceSlotUsingDB(t *testing.T) {
|
|
ctx := context.Background()
|
|
db := testDB.SetupDB(t)
|
|
defer testDB.TeardownDB(t, db)
|
|
service := New(db, cache.NewStateSummaryCache())
|
|
beaconState, _ := testutil.DeterministicGenesisState(t, 32)
|
|
b := ðpb.SignedBeaconBlock{Block: ðpb.BeaconBlock{}}
|
|
if err := service.beaconDB.SaveBlock(ctx, b); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
gRoot, _ := ssz.HashTreeRoot(b.Block)
|
|
if err := service.beaconDB.SaveGenesisBlockRoot(ctx, gRoot); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if err := service.beaconDB.SaveState(ctx, beaconState, gRoot); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
slot := uint64(10)
|
|
loadedState, err := service.loadHotStateBySlot(ctx, slot)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if loadedState.Slot() != slot {
|
|
t.Error("Did not correctly load state")
|
|
}
|
|
}
|