package sync import ( "context" "sync" "testing" "time" "github.com/ethereum/go-ethereum/p2p/enr" "github.com/kevinms/leakybucket-go" "github.com/libp2p/go-libp2p-core/network" "github.com/libp2p/go-libp2p-core/protocol" types "github.com/prysmaticlabs/eth2-types" mock "github.com/prysmaticlabs/prysm/beacon-chain/blockchain/testing" "github.com/prysmaticlabs/prysm/beacon-chain/core/transition" "github.com/prysmaticlabs/prysm/beacon-chain/db/kv" testingDB "github.com/prysmaticlabs/prysm/beacon-chain/db/testing" "github.com/prysmaticlabs/prysm/beacon-chain/p2p" "github.com/prysmaticlabs/prysm/beacon-chain/p2p/peers" p2ptest "github.com/prysmaticlabs/prysm/beacon-chain/p2p/testing" p2ptypes "github.com/prysmaticlabs/prysm/beacon-chain/p2p/types" v1 "github.com/prysmaticlabs/prysm/beacon-chain/state/v1" mockSync "github.com/prysmaticlabs/prysm/beacon-chain/sync/initial-sync/testing" ethpb "github.com/prysmaticlabs/prysm/proto/prysm/v1alpha1" pb "github.com/prysmaticlabs/prysm/proto/prysm/v1alpha1" "github.com/prysmaticlabs/prysm/proto/prysm/v1alpha1/block" "github.com/prysmaticlabs/prysm/proto/prysm/v1alpha1/wrapper" p2pWrapper "github.com/prysmaticlabs/prysm/proto/prysm/v1alpha1/wrapper" "github.com/prysmaticlabs/prysm/shared/bytesutil" "github.com/prysmaticlabs/prysm/shared/params" "github.com/prysmaticlabs/prysm/shared/testutil" "github.com/prysmaticlabs/prysm/shared/testutil/assert" "github.com/prysmaticlabs/prysm/shared/testutil/require" "github.com/prysmaticlabs/prysm/shared/timeutils" "google.golang.org/protobuf/proto" ) func TestStatusRPCHandler_Disconnects_OnForkVersionMismatch(t *testing.T) { p1 := p2ptest.NewTestP2P(t) p2 := p2ptest.NewTestP2P(t) p1.Connect(p2) assert.Equal(t, 1, len(p1.BHost.Network().Peers()), "Expected peers to be connected") root := [32]byte{'C'} r := &Service{ cfg: &Config{ P2P: p1, Chain: &mock.ChainService{ Fork: ðpb.Fork{ PreviousVersion: params.BeaconConfig().GenesisForkVersion, CurrentVersion: params.BeaconConfig().GenesisForkVersion, }, FinalizedCheckPoint: ðpb.Checkpoint{ Epoch: 0, Root: root[:], }, Genesis: time.Now(), ValidatorsRoot: [32]byte{'A'}, Root: make([]byte, 32), }, }, rateLimiter: newRateLimiter(p1), } pcl := protocol.ID(p2p.RPCStatusTopicV1) topic := string(pcl) r.rateLimiter.limiterMap[topic] = leakybucket.NewCollector(1, 1, false) var wg sync.WaitGroup wg.Add(1) p2.BHost.SetStreamHandler(pcl, func(stream network.Stream) { defer wg.Done() expectSuccess(t, stream) out := &pb.Status{} assert.NoError(t, r.cfg.P2P.Encoding().DecodeWithMaxLength(stream, out)) assert.DeepEqual(t, root[:], out.FinalizedRoot) assert.NoError(t, stream.Close()) }) pcl2 := protocol.ID("/eth2/beacon_chain/req/goodbye/1/ssz_snappy") topic = string(pcl2) r.rateLimiter.limiterMap[topic] = leakybucket.NewCollector(1, 1, false) var wg2 sync.WaitGroup wg2.Add(1) p2.BHost.SetStreamHandler(pcl2, func(stream network.Stream) { defer wg2.Done() msg := new(types.SSZUint64) assert.NoError(t, r.cfg.P2P.Encoding().DecodeWithMaxLength(stream, msg)) assert.Equal(t, p2ptypes.GoodbyeCodeWrongNetwork, *msg) assert.NoError(t, stream.Close()) }) stream1, err := p1.BHost.NewStream(context.Background(), p2.BHost.ID(), pcl) require.NoError(t, err) assert.NoError(t, r.statusRPCHandler(context.Background(), &pb.Status{ForkDigest: bytesutil.PadTo([]byte("f"), 4), HeadRoot: make([]byte, 32), FinalizedRoot: make([]byte, 32)}, stream1)) if testutil.WaitTimeout(&wg, 1*time.Second) { t.Fatal("Did not receive stream within 1 sec") } if testutil.WaitTimeout(&wg2, 1*time.Second) { t.Fatal("Did not receive stream within 1 sec") } assert.Equal(t, 0, len(p1.BHost.Network().Peers()), "handler did not disconnect peer") } func TestStatusRPCHandler_ConnectsOnGenesis(t *testing.T) { p1 := p2ptest.NewTestP2P(t) p2 := p2ptest.NewTestP2P(t) p1.Connect(p2) assert.Equal(t, 1, len(p1.BHost.Network().Peers()), "Expected peers to be connected") root := [32]byte{} r := &Service{ cfg: &Config{ P2P: p1, Chain: &mock.ChainService{ Fork: ðpb.Fork{ PreviousVersion: params.BeaconConfig().GenesisForkVersion, CurrentVersion: params.BeaconConfig().GenesisForkVersion, }, FinalizedCheckPoint: ðpb.Checkpoint{ Epoch: 0, Root: params.BeaconConfig().ZeroHash[:], }, Genesis: time.Now(), ValidatorsRoot: [32]byte{'A'}, Root: make([]byte, 32), }, }, rateLimiter: newRateLimiter(p1), } pcl := protocol.ID(p2p.RPCStatusTopicV1) topic := string(pcl) r.rateLimiter.limiterMap[topic] = leakybucket.NewCollector(1, 1, false) var wg sync.WaitGroup wg.Add(1) p2.BHost.SetStreamHandler(pcl, func(stream network.Stream) { defer wg.Done() expectSuccess(t, stream) out := &pb.Status{} assert.NoError(t, r.cfg.P2P.Encoding().DecodeWithMaxLength(stream, out)) assert.DeepEqual(t, root[:], out.FinalizedRoot) }) stream1, err := p1.BHost.NewStream(context.Background(), p2.BHost.ID(), pcl) require.NoError(t, err) digest, err := r.currentForkDigest() require.NoError(t, err) err = r.statusRPCHandler(context.Background(), &pb.Status{ForkDigest: digest[:], FinalizedRoot: params.BeaconConfig().ZeroHash[:]}, stream1) assert.NoError(t, err) if testutil.WaitTimeout(&wg, 1*time.Second) { t.Fatal("Did not receive stream within 1 sec") } assert.Equal(t, 1, len(p1.BHost.Network().Peers()), "Handler disconnected with peer") } func TestStatusRPCHandler_ReturnsHelloMessage(t *testing.T) { p1 := p2ptest.NewTestP2P(t) p2 := p2ptest.NewTestP2P(t) p1.Connect(p2) assert.Equal(t, 1, len(p1.BHost.Network().Peers()), "Expected peers to be connected") db := testingDB.SetupDB(t) // Set up a head state with data we expect. head := testutil.NewBeaconBlock() head.Block.Slot = 111 headRoot, err := head.Block.HashTreeRoot() require.NoError(t, err) blkSlot := 3 * params.BeaconConfig().SlotsPerEpoch finalized := testutil.NewBeaconBlock() finalized.Block.Slot = blkSlot finalizedRoot, err := finalized.Block.HashTreeRoot() require.NoError(t, err) genesisState, err := transition.GenesisBeaconState(context.Background(), nil, 0, ðpb.Eth1Data{}) require.NoError(t, err) require.NoError(t, genesisState.SetSlot(111)) require.NoError(t, genesisState.UpdateBlockRootAtIndex(111%uint64(params.BeaconConfig().SlotsPerHistoricalRoot), headRoot)) require.NoError(t, db.SaveBlock(context.Background(), wrapper.WrappedPhase0SignedBeaconBlock(finalized))) require.NoError(t, db.SaveGenesisBlockRoot(context.Background(), finalizedRoot)) finalizedCheckpt := ðpb.Checkpoint{ Epoch: 3, Root: finalizedRoot[:], } totalSec := int64(params.BeaconConfig().SlotsPerEpoch.Mul(5 * params.BeaconConfig().SecondsPerSlot)) genTime := time.Now().Unix() - totalSec r := &Service{ cfg: &Config{ P2P: p1, Chain: &mock.ChainService{ State: genesisState, FinalizedCheckPoint: finalizedCheckpt, Root: headRoot[:], Fork: ðpb.Fork{ PreviousVersion: params.BeaconConfig().GenesisForkVersion, CurrentVersion: params.BeaconConfig().GenesisForkVersion, }, ValidatorsRoot: [32]byte{'A'}, Genesis: time.Unix(genTime, 0), }, DB: db, }, rateLimiter: newRateLimiter(p1), } digest, err := r.currentForkDigest() require.NoError(t, err) // Setup streams pcl := protocol.ID(p2p.RPCStatusTopicV1) topic := string(pcl) r.rateLimiter.limiterMap[topic] = leakybucket.NewCollector(1, 1, false) var wg sync.WaitGroup wg.Add(1) p2.BHost.SetStreamHandler(pcl, func(stream network.Stream) { defer wg.Done() expectSuccess(t, stream) out := &pb.Status{} assert.NoError(t, r.cfg.P2P.Encoding().DecodeWithMaxLength(stream, out)) expected := &pb.Status{ ForkDigest: digest[:], HeadSlot: genesisState.Slot(), HeadRoot: headRoot[:], FinalizedEpoch: 3, FinalizedRoot: finalizedRoot[:], } if !proto.Equal(out, expected) { t.Errorf("Did not receive expected message. Got %+v wanted %+v", out, expected) } }) stream1, err := p1.BHost.NewStream(context.Background(), p2.BHost.ID(), pcl) require.NoError(t, err) err = r.statusRPCHandler(context.Background(), &pb.Status{ ForkDigest: digest[:], FinalizedRoot: finalizedRoot[:], FinalizedEpoch: 3, }, stream1) assert.NoError(t, err) if testutil.WaitTimeout(&wg, 1*time.Second) { t.Fatal("Did not receive stream within 1 sec") } } func TestHandshakeHandlers_Roundtrip(t *testing.T) { // Scenario is that p1 and p2 connect, exchange handshakes. // p2 disconnects and p1 should forget the handshake status. p1 := p2ptest.NewTestP2P(t) p2 := p2ptest.NewTestP2P(t) db := testingDB.SetupDB(t) p1.LocalMetadata = p2pWrapper.WrappedMetadataV0(&pb.MetaDataV0{ SeqNumber: 2, Attnets: bytesutil.PadTo([]byte{'A', 'B'}, 8), }) p2.LocalMetadata = p2pWrapper.WrappedMetadataV0(&pb.MetaDataV0{ SeqNumber: 2, Attnets: bytesutil.PadTo([]byte{'C', 'D'}, 8), }) st, err := v1.InitializeFromProto(ðpb.BeaconState{ Slot: 5, }) require.NoError(t, err) blk := testutil.NewBeaconBlock() blk.Block.Slot = 0 require.NoError(t, db.SaveBlock(context.Background(), wrapper.WrappedPhase0SignedBeaconBlock(blk))) finalizedRoot, err := blk.Block.HashTreeRoot() require.NoError(t, err) require.NoError(t, db.SaveGenesisBlockRoot(context.Background(), finalizedRoot)) r := &Service{ cfg: &Config{ P2P: p1, Chain: &mock.ChainService{ State: st, FinalizedCheckPoint: ðpb.Checkpoint{Epoch: 0, Root: finalizedRoot[:]}, Fork: ðpb.Fork{ PreviousVersion: params.BeaconConfig().GenesisForkVersion, CurrentVersion: params.BeaconConfig().GenesisForkVersion, }, Genesis: time.Now(), ValidatorsRoot: [32]byte{'A'}, Root: make([]byte, 32), }, DB: db, }, ctx: context.Background(), rateLimiter: newRateLimiter(p1), } p1.Digest, err = r.currentForkDigest() require.NoError(t, err) r2 := &Service{ cfg: &Config{ Chain: &mock.ChainService{ FinalizedCheckPoint: ðpb.Checkpoint{Epoch: 0, Root: finalizedRoot[:]}, }, P2P: p2, }, rateLimiter: newRateLimiter(p2), } p2.Digest, err = r.currentForkDigest() require.NoError(t, err) r.Start() // Setup streams pcl := protocol.ID("/eth2/beacon_chain/req/status/1/ssz_snappy") topic := string(pcl) r.rateLimiter.limiterMap[topic] = leakybucket.NewCollector(1, 1, false) var wg sync.WaitGroup wg.Add(1) p2.BHost.SetStreamHandler(pcl, func(stream network.Stream) { defer wg.Done() out := &pb.Status{} assert.NoError(t, r.cfg.P2P.Encoding().DecodeWithMaxLength(stream, out)) log.WithField("status", out).Warn("received status") resp := &pb.Status{HeadSlot: 100, HeadRoot: make([]byte, 32), ForkDigest: p2.Digest[:], FinalizedRoot: finalizedRoot[:], FinalizedEpoch: 0} _, err := stream.Write([]byte{responseCodeSuccess}) assert.NoError(t, err) _, err = r.cfg.P2P.Encoding().EncodeWithMaxLength(stream, resp) assert.NoError(t, err) log.WithField("status", out).Warn("sending status") if err := stream.Close(); err != nil { t.Log(err) } }) pcl = "/eth2/beacon_chain/req/ping/1/ssz_snappy" topic = string(pcl) r2.rateLimiter.limiterMap[topic] = leakybucket.NewCollector(1, 1, false) var wg2 sync.WaitGroup wg2.Add(1) p2.BHost.SetStreamHandler(pcl, func(stream network.Stream) { defer wg2.Done() out := new(types.SSZUint64) assert.NoError(t, r.cfg.P2P.Encoding().DecodeWithMaxLength(stream, out)) assert.Equal(t, uint64(2), uint64(*out)) assert.NoError(t, r2.pingHandler(context.Background(), out, stream)) assert.NoError(t, stream.Close()) }) numInactive1 := len(p1.Peers().Inactive()) numActive1 := len(p1.Peers().Active()) p1.Connect(p2) p1.Peers().Add(new(enr.Record), p2.BHost.ID(), p2.BHost.Addrs()[0], network.DirUnknown) p1.Peers().SetMetadata(p2.BHost.ID(), p2.LocalMetadata) p2.Peers().Add(new(enr.Record), p1.BHost.ID(), p1.BHost.Addrs()[0], network.DirUnknown) p2.Peers().SetMetadata(p1.BHost.ID(), p1.LocalMetadata) if testutil.WaitTimeout(&wg, 1*time.Second) { t.Fatal("Did not receive stream within 1 sec") } if testutil.WaitTimeout(&wg2, 1*time.Second) { t.Fatal("Did not receive stream within 1 sec") } // Wait for stream buffer to be read. time.Sleep(200 * time.Millisecond) numInactive2 := len(p1.Peers().Inactive()) numActive2 := len(p1.Peers().Active()) assert.Equal(t, numInactive1, numInactive1, "Number of inactive peers changed unexpectedly") assert.Equal(t, numActive1+1, numActive2, "Number of active peers unexpected") require.NoError(t, p2.Disconnect(p1.PeerID())) p1.Peers().SetConnectionState(p2.PeerID(), peers.PeerDisconnected) // Wait for disconnect event to trigger. time.Sleep(200 * time.Millisecond) numInactive3 := len(p1.Peers().Inactive()) numActive3 := len(p1.Peers().Active()) assert.Equal(t, numInactive2+1, numInactive3, "Number of inactive peers unexpected") assert.Equal(t, numActive2-1, numActive3, "Number of active peers unexpected") } func TestStatusRPCRequest_RequestSent(t *testing.T) { p1 := p2ptest.NewTestP2P(t) p2 := p2ptest.NewTestP2P(t) // Set up a head state with data we expect. head := testutil.NewBeaconBlock() head.Block.Slot = 111 headRoot, err := head.Block.HashTreeRoot() require.NoError(t, err) finalized := testutil.NewBeaconBlock() finalized.Block.Slot = 40 finalizedRoot, err := finalized.Block.HashTreeRoot() require.NoError(t, err) genesisState, err := transition.GenesisBeaconState(context.Background(), nil, 0, ðpb.Eth1Data{}) require.NoError(t, err) require.NoError(t, genesisState.SetSlot(111)) require.NoError(t, genesisState.UpdateBlockRootAtIndex(111%uint64(params.BeaconConfig().SlotsPerHistoricalRoot), headRoot)) finalizedCheckpt := ðpb.Checkpoint{ Epoch: 5, Root: finalizedRoot[:], } r := &Service{ cfg: &Config{ P2P: p1, Chain: &mock.ChainService{ State: genesisState, FinalizedCheckPoint: finalizedCheckpt, Root: headRoot[:], Fork: ðpb.Fork{ PreviousVersion: params.BeaconConfig().GenesisForkVersion, CurrentVersion: params.BeaconConfig().GenesisForkVersion, }, Genesis: time.Now(), ValidatorsRoot: [32]byte{'A'}, }, }, ctx: context.Background(), rateLimiter: newRateLimiter(p1), } // Setup streams pcl := protocol.ID("/eth2/beacon_chain/req/status/1/ssz_snappy") topic := string(pcl) r.rateLimiter.limiterMap[topic] = leakybucket.NewCollector(1, 1, false) var wg sync.WaitGroup wg.Add(1) p2.BHost.SetStreamHandler(pcl, func(stream network.Stream) { defer wg.Done() out := &pb.Status{} assert.NoError(t, r.cfg.P2P.Encoding().DecodeWithMaxLength(stream, out)) digest, err := r.currentForkDigest() assert.NoError(t, err) expected := &pb.Status{ ForkDigest: digest[:], HeadSlot: genesisState.Slot(), HeadRoot: headRoot[:], FinalizedEpoch: 5, FinalizedRoot: finalizedRoot[:], } if !proto.Equal(out, expected) { t.Errorf("Did not receive expected message. Got %+v wanted %+v", out, expected) } }) p1.AddConnectionHandler(r.sendRPCStatusRequest, nil) p1.Connect(p2) if testutil.WaitTimeout(&wg, 1*time.Second) { t.Fatal("Did not receive stream within 1 sec") } assert.Equal(t, 1, len(p1.BHost.Network().Peers()), "Expected peers to continue being connected") } func TestStatusRPCRequest_FinalizedBlockExists(t *testing.T) { p1 := p2ptest.NewTestP2P(t) p2 := p2ptest.NewTestP2P(t) db := testingDB.SetupDB(t) // Set up a head state with data we expect. head := testutil.NewBeaconBlock() head.Block.Slot = 111 headRoot, err := head.Block.HashTreeRoot() require.NoError(t, err) blkSlot := 3 * params.BeaconConfig().SlotsPerEpoch finalized := testutil.NewBeaconBlock() finalized.Block.Slot = blkSlot finalizedRoot, err := finalized.Block.HashTreeRoot() require.NoError(t, err) genesisState, err := transition.GenesisBeaconState(context.Background(), nil, 0, ðpb.Eth1Data{DepositRoot: make([]byte, 32), BlockHash: make([]byte, 32)}) require.NoError(t, err) require.NoError(t, genesisState.SetSlot(111)) require.NoError(t, genesisState.UpdateBlockRootAtIndex(111%uint64(params.BeaconConfig().SlotsPerHistoricalRoot), headRoot)) blk := testutil.NewBeaconBlock() blk.Block.Slot = blkSlot require.NoError(t, db.SaveBlock(context.Background(), wrapper.WrappedPhase0SignedBeaconBlock(blk))) require.NoError(t, db.SaveGenesisBlockRoot(context.Background(), finalizedRoot)) finalizedCheckpt := ðpb.Checkpoint{ Epoch: 3, Root: finalizedRoot[:], } totalSec := int64(params.BeaconConfig().SlotsPerEpoch.Mul(5 * params.BeaconConfig().SecondsPerSlot)) genTime := time.Now().Unix() - totalSec r := &Service{ cfg: &Config{ P2P: p1, Chain: &mock.ChainService{ State: genesisState, FinalizedCheckPoint: finalizedCheckpt, Root: headRoot[:], Fork: ðpb.Fork{ PreviousVersion: params.BeaconConfig().GenesisForkVersion, CurrentVersion: params.BeaconConfig().GenesisForkVersion, }, Genesis: time.Unix(genTime, 0), ValidatorsRoot: [32]byte{'A'}, }, }, ctx: context.Background(), rateLimiter: newRateLimiter(p1), } r2 := &Service{ cfg: &Config{ P2P: p1, Chain: &mock.ChainService{ State: genesisState, FinalizedCheckPoint: finalizedCheckpt, Root: headRoot[:], Fork: ðpb.Fork{ PreviousVersion: params.BeaconConfig().GenesisForkVersion, CurrentVersion: params.BeaconConfig().GenesisForkVersion, }, Genesis: time.Unix(genTime, 0), ValidatorsRoot: [32]byte{'A'}, }, DB: db, }, ctx: context.Background(), rateLimiter: newRateLimiter(p1), } // Setup streams pcl := protocol.ID("/eth2/beacon_chain/req/status/1/ssz_snappy") topic := string(pcl) r.rateLimiter.limiterMap[topic] = leakybucket.NewCollector(1, 1, false) var wg sync.WaitGroup wg.Add(1) p2.BHost.SetStreamHandler(pcl, func(stream network.Stream) { defer wg.Done() out := &pb.Status{} assert.NoError(t, r.cfg.P2P.Encoding().DecodeWithMaxLength(stream, out)) assert.NoError(t, r2.validateStatusMessage(context.Background(), out)) }) p1.AddConnectionHandler(r.sendRPCStatusRequest, nil) p1.Connect(p2) if testutil.WaitTimeout(&wg, 1*time.Second) { t.Fatal("Did not receive stream within 1 sec") } assert.Equal(t, 1, len(p1.BHost.Network().Peers()), "Expected peers to continue being connected") } func TestStatusRPCRequest_FinalizedBlockSkippedSlots(t *testing.T) { db, err := kv.NewKVStore(context.Background(), t.TempDir(), &kv.Config{}) require.NoError(t, err) bState, err := transition.GenesisBeaconState(context.Background(), nil, 0, ðpb.Eth1Data{DepositRoot: make([]byte, 32), BlockHash: make([]byte, 32)}) require.NoError(t, err) blk := testutil.NewBeaconBlock() blk.Block.Slot = 0 genRoot, err := blk.Block.HashTreeRoot() require.NoError(t, err) require.NoError(t, db.SaveBlock(context.Background(), wrapper.WrappedPhase0SignedBeaconBlock(blk))) require.NoError(t, db.SaveGenesisBlockRoot(context.Background(), genRoot)) blocksTillHead := makeBlocks(t, 1, 1000, genRoot) require.NoError(t, db.SaveBlocks(context.Background(), blocksTillHead)) stateSummaries := make([]*ethpb.StateSummary, len(blocksTillHead)) for i, b := range blocksTillHead { bRoot, err := b.Block().HashTreeRoot() require.NoError(t, err) stateSummaries[i] = ðpb.StateSummary{ Slot: b.Block().Slot(), Root: bRoot[:], } } require.NoError(t, db.SaveStateSummaries(context.Background(), stateSummaries)) rootFetcher := func(slot types.Slot) [32]byte { rt, err := blocksTillHead[slot-1].Block().HashTreeRoot() require.NoError(t, err) return rt } tests := []struct { name string expectedFinalizedEpoch types.Epoch expectedFinalizedRoot [32]byte headSlot types.Slot remoteFinalizedEpoch types.Epoch remoteFinalizedRoot [32]byte remoteHeadSlot types.Slot expectError bool }{ { name: "valid finalized epoch", expectedFinalizedEpoch: 3, expectedFinalizedRoot: rootFetcher(3 * params.BeaconConfig().SlotsPerEpoch), headSlot: 111, remoteFinalizedEpoch: 3, remoteFinalizedRoot: rootFetcher(3 * params.BeaconConfig().SlotsPerEpoch), remoteHeadSlot: 100, expectError: false, }, { name: "invalid finalized epoch", expectedFinalizedEpoch: 3, expectedFinalizedRoot: rootFetcher(3 * params.BeaconConfig().SlotsPerEpoch), headSlot: 111, remoteFinalizedEpoch: 3, // give an incorrect root relative to the finalized epoch. remoteFinalizedRoot: rootFetcher(2 * params.BeaconConfig().SlotsPerEpoch), remoteHeadSlot: 120, expectError: true, }, { name: "invalid finalized root", expectedFinalizedEpoch: 3, expectedFinalizedRoot: rootFetcher(3 * params.BeaconConfig().SlotsPerEpoch), headSlot: 111, remoteFinalizedEpoch: 3, // give a bad finalized root, and the beacon node verifies that // it is indeed incorrect. remoteFinalizedRoot: [32]byte{'a', 'b', 'c'}, remoteHeadSlot: 120, expectError: true, }, } for _, tt := range tests { p1 := p2ptest.NewTestP2P(t) p2 := p2ptest.NewTestP2P(t) expectedFinalizedEpoch := tt.expectedFinalizedEpoch headSlot := tt.headSlot nState := bState.Copy() // Set up a head state with data we expect. head := blocksTillHead[len(blocksTillHead)-1] headRoot, err := head.Block().HashTreeRoot() require.NoError(t, err) rHead := blocksTillHead[tt.remoteHeadSlot-1] rHeadRoot, err := rHead.Block().HashTreeRoot() require.NoError(t, err) require.NoError(t, nState.SetSlot(headSlot)) require.NoError(t, nState.UpdateBlockRootAtIndex(uint64(headSlot.ModSlot(params.BeaconConfig().SlotsPerHistoricalRoot)), headRoot)) finalizedCheckpt := ðpb.Checkpoint{ Epoch: expectedFinalizedEpoch, Root: tt.expectedFinalizedRoot[:], } remoteFinalizedChkpt := ðpb.Checkpoint{ Epoch: tt.remoteFinalizedEpoch, Root: tt.remoteFinalizedRoot[:], } require.NoError(t, db.SaveFinalizedCheckpoint(context.Background(), finalizedCheckpt)) epoch := expectedFinalizedEpoch.Add(2) totalSec := uint64(params.BeaconConfig().SlotsPerEpoch.Mul(uint64(epoch) * params.BeaconConfig().SecondsPerSlot)) genTime := time.Now().Unix() - int64(totalSec) r := &Service{ cfg: &Config{ P2P: p1, Chain: &mock.ChainService{ State: nState, FinalizedCheckPoint: remoteFinalizedChkpt, Root: rHeadRoot[:], Fork: ðpb.Fork{ PreviousVersion: params.BeaconConfig().GenesisForkVersion, CurrentVersion: params.BeaconConfig().GenesisForkVersion, }, Genesis: time.Unix(genTime, 0), ValidatorsRoot: [32]byte{'A'}, }, }, ctx: context.Background(), rateLimiter: newRateLimiter(p1), } r2 := &Service{ cfg: &Config{ P2P: p2, Chain: &mock.ChainService{ State: nState, FinalizedCheckPoint: finalizedCheckpt, Root: headRoot[:], Fork: ðpb.Fork{ PreviousVersion: params.BeaconConfig().GenesisForkVersion, CurrentVersion: params.BeaconConfig().GenesisForkVersion, }, Genesis: time.Unix(genTime, 0), ValidatorsRoot: [32]byte{'A'}, }, DB: db, }, ctx: context.Background(), rateLimiter: newRateLimiter(p1), } // Setup streams pcl := protocol.ID("/eth2/beacon_chain/req/status/1/ssz_snappy") topic := string(pcl) r.rateLimiter.limiterMap[topic] = leakybucket.NewCollector(1, 1, false) var wg sync.WaitGroup wg.Add(1) p2.BHost.SetStreamHandler(pcl, func(stream network.Stream) { defer wg.Done() out := &pb.Status{} assert.NoError(t, r.cfg.P2P.Encoding().DecodeWithMaxLength(stream, out)) assert.Equal(t, tt.expectError, r2.validateStatusMessage(context.Background(), out) != nil) }) p1.AddConnectionHandler(r.sendRPCStatusRequest, nil) p1.Connect(p2) if testutil.WaitTimeout(&wg, 1*time.Second) { t.Fatal("Did not receive stream within 1 sec") } assert.Equal(t, 1, len(p1.BHost.Network().Peers()), "Expected peers to continue being connected") assert.NoError(t, p1.Disconnect(p2.PeerID())) } assert.NoError(t, db.Close()) } func TestStatusRPCRequest_BadPeerHandshake(t *testing.T) { p1 := p2ptest.NewTestP2P(t) p2 := p2ptest.NewTestP2P(t) // Set up a head state with data we expect. head := testutil.NewBeaconBlock() head.Block.Slot = 111 headRoot, err := head.Block.HashTreeRoot() require.NoError(t, err) finalized := testutil.NewBeaconBlock() finalizedRoot, err := finalized.Block.HashTreeRoot() require.NoError(t, err) genesisState, err := transition.GenesisBeaconState(context.Background(), nil, 0, ðpb.Eth1Data{}) require.NoError(t, err) require.NoError(t, genesisState.SetSlot(111)) require.NoError(t, genesisState.UpdateBlockRootAtIndex(111%uint64(params.BeaconConfig().SlotsPerHistoricalRoot), headRoot)) finalizedCheckpt := ðpb.Checkpoint{ Epoch: 5, Root: finalizedRoot[:], } r := &Service{ cfg: &Config{ P2P: p1, Chain: &mock.ChainService{ State: genesisState, FinalizedCheckPoint: finalizedCheckpt, Root: headRoot[:], Fork: ðpb.Fork{ PreviousVersion: params.BeaconConfig().GenesisForkVersion, CurrentVersion: params.BeaconConfig().GenesisForkVersion, }, Genesis: time.Now(), ValidatorsRoot: [32]byte{'A'}, }, }, ctx: context.Background(), rateLimiter: newRateLimiter(p1), } r.Start() // Setup streams pcl := protocol.ID("/eth2/beacon_chain/req/status/1/ssz_snappy") topic := string(pcl) r.rateLimiter.limiterMap[topic] = leakybucket.NewCollector(1, 1, false) var wg sync.WaitGroup wg.Add(1) p2.BHost.SetStreamHandler(pcl, func(stream network.Stream) { defer wg.Done() out := &pb.Status{} assert.NoError(t, r.cfg.P2P.Encoding().DecodeWithMaxLength(stream, out)) expected := &pb.Status{ ForkDigest: []byte{1, 1, 1, 1}, HeadSlot: genesisState.Slot(), HeadRoot: headRoot[:], FinalizedEpoch: 5, FinalizedRoot: finalizedRoot[:], } if _, err := stream.Write([]byte{responseCodeSuccess}); err != nil { log.WithError(err).Debug("Could not write to stream") } _, err := r.cfg.P2P.Encoding().EncodeWithMaxLength(stream, expected) assert.NoError(t, err) }) assert.Equal(t, false, p1.Peers().Scorers().IsBadPeer(p2.PeerID()), "Peer is marked as bad") p1.Connect(p2) if testutil.WaitTimeout(&wg, time.Second) { t.Fatal("Did not receive stream within 1 sec") } time.Sleep(100 * time.Millisecond) connectionState, err := p1.Peers().ConnectionState(p2.PeerID()) require.NoError(t, err, "Could not obtain peer connection state") assert.Equal(t, peers.PeerDisconnected, connectionState, "Expected peer to be disconnected") assert.Equal(t, true, p1.Peers().Scorers().IsBadPeer(p2.PeerID()), "Peer is not marked as bad") } func TestStatusRPC_ValidGenesisMessage(t *testing.T) { // Set up a head state with data we expect. head := testutil.NewBeaconBlock() head.Block.Slot = 111 headRoot, err := head.Block.HashTreeRoot() require.NoError(t, err) blkSlot := 3 * params.BeaconConfig().SlotsPerEpoch finalized := testutil.NewBeaconBlock() finalized.Block.Slot = blkSlot finalizedRoot, err := finalized.Block.HashTreeRoot() require.NoError(t, err) genesisState, err := transition.GenesisBeaconState(context.Background(), nil, 0, ðpb.Eth1Data{}) require.NoError(t, err) require.NoError(t, genesisState.SetSlot(111)) require.NoError(t, genesisState.UpdateBlockRootAtIndex(111%uint64(params.BeaconConfig().SlotsPerHistoricalRoot), headRoot)) finalizedCheckpt := ðpb.Checkpoint{ Epoch: 5, Root: finalizedRoot[:], } r := &Service{ cfg: &Config{ Chain: &mock.ChainService{ State: genesisState, FinalizedCheckPoint: finalizedCheckpt, Root: headRoot[:], Fork: ðpb.Fork{ PreviousVersion: params.BeaconConfig().GenesisForkVersion, CurrentVersion: params.BeaconConfig().GenesisForkVersion, }, Genesis: time.Now(), ValidatorsRoot: [32]byte{'A'}, }, }, ctx: context.Background(), } digest, err := r.currentForkDigest() require.NoError(t, err) // There should be no error for a status message // with a genesis checkpoint. err = r.validateStatusMessage(r.ctx, &pb.Status{ ForkDigest: digest[:], FinalizedRoot: params.BeaconConfig().ZeroHash[:], FinalizedEpoch: 0, HeadRoot: headRoot[:], HeadSlot: 111, }) require.NoError(t, err) } func TestShouldResync(t *testing.T) { type args struct { genesis time.Time syncing bool headSlot types.Slot } tests := []struct { name string args args want bool }{ { name: "genesis epoch should not resync when syncing is true", args: args{ headSlot: 31, genesis: timeutils.Now(), syncing: true, }, want: false, }, { name: "genesis epoch should not resync when syncing is false", args: args{ headSlot: 31, genesis: timeutils.Now(), syncing: false, }, want: false, }, { name: "two epochs behind, resync ok", args: args{ headSlot: 31, genesis: timeutils.Now().Add(-1 * 96 * time.Duration(params.BeaconConfig().SecondsPerSlot) * time.Second), syncing: false, }, want: true, }, { name: "two epochs behind, already syncing", args: args{ headSlot: 31, genesis: timeutils.Now().Add(-1 * 96 * time.Duration(params.BeaconConfig().SecondsPerSlot) * time.Second), syncing: true, }, want: false, }, } for _, tt := range tests { headState, err := transition.GenesisBeaconState(context.Background(), nil, 0, ðpb.Eth1Data{}) require.NoError(t, err) require.NoError(t, headState.SetSlot(tt.args.headSlot)) r := &Service{ cfg: &Config{ Chain: &mock.ChainService{ State: headState, Genesis: tt.args.genesis, }, InitialSync: &mockSync.Sync{IsSyncing: tt.args.syncing}, }, ctx: context.Background(), } t.Run(tt.name, func(t *testing.T) { if got := r.shouldReSync(); got != tt.want { t.Errorf("shouldReSync() = %v, want %v", got, tt.want) } }) } } func makeBlocks(t *testing.T, i, n uint64, previousRoot [32]byte) []block.SignedBeaconBlock { blocks := make([]*ethpb.SignedBeaconBlock, n) ifaceBlocks := make([]block.SignedBeaconBlock, n) for j := i; j < n+i; j++ { parentRoot := make([]byte, 32) copy(parentRoot, previousRoot[:]) blocks[j-i] = testutil.NewBeaconBlock() blocks[j-i].Block.Slot = types.Slot(j + 1) blocks[j-i].Block.ParentRoot = parentRoot var err error previousRoot, err = blocks[j-i].Block.HashTreeRoot() require.NoError(t, err) ifaceBlocks[j-i] = wrapper.WrappedPhase0SignedBeaconBlock(blocks[j-i]) } return ifaceBlocks }