mirror of
https://gitlab.com/pulsechaincom/prysm-pulse.git
synced 2024-12-25 04:47:18 +00:00
948 lines
30 KiB
Go
948 lines
30 KiB
Go
package sync
|
|
|
|
import (
|
|
"context"
|
|
"sync"
|
|
"testing"
|
|
"time"
|
|
|
|
"github.com/ethereum/go-ethereum/p2p/enr"
|
|
"github.com/kevinms/leakybucket-go"
|
|
"github.com/libp2p/go-libp2p-core/network"
|
|
"github.com/libp2p/go-libp2p-core/protocol"
|
|
types "github.com/prysmaticlabs/eth2-types"
|
|
mock "github.com/prysmaticlabs/prysm/beacon-chain/blockchain/testing"
|
|
"github.com/prysmaticlabs/prysm/beacon-chain/core/transition"
|
|
"github.com/prysmaticlabs/prysm/beacon-chain/db/kv"
|
|
testingDB "github.com/prysmaticlabs/prysm/beacon-chain/db/testing"
|
|
"github.com/prysmaticlabs/prysm/beacon-chain/p2p"
|
|
"github.com/prysmaticlabs/prysm/beacon-chain/p2p/peers"
|
|
p2ptest "github.com/prysmaticlabs/prysm/beacon-chain/p2p/testing"
|
|
p2ptypes "github.com/prysmaticlabs/prysm/beacon-chain/p2p/types"
|
|
v1 "github.com/prysmaticlabs/prysm/beacon-chain/state/v1"
|
|
mockSync "github.com/prysmaticlabs/prysm/beacon-chain/sync/initial-sync/testing"
|
|
ethpb "github.com/prysmaticlabs/prysm/proto/prysm/v1alpha1"
|
|
pb "github.com/prysmaticlabs/prysm/proto/prysm/v1alpha1"
|
|
"github.com/prysmaticlabs/prysm/proto/prysm/v1alpha1/block"
|
|
"github.com/prysmaticlabs/prysm/proto/prysm/v1alpha1/wrapper"
|
|
p2pWrapper "github.com/prysmaticlabs/prysm/proto/prysm/v1alpha1/wrapper"
|
|
"github.com/prysmaticlabs/prysm/shared/bytesutil"
|
|
"github.com/prysmaticlabs/prysm/shared/params"
|
|
"github.com/prysmaticlabs/prysm/shared/testutil"
|
|
"github.com/prysmaticlabs/prysm/shared/testutil/assert"
|
|
"github.com/prysmaticlabs/prysm/shared/testutil/require"
|
|
"github.com/prysmaticlabs/prysm/shared/timeutils"
|
|
"google.golang.org/protobuf/proto"
|
|
)
|
|
|
|
func TestStatusRPCHandler_Disconnects_OnForkVersionMismatch(t *testing.T) {
|
|
p1 := p2ptest.NewTestP2P(t)
|
|
p2 := p2ptest.NewTestP2P(t)
|
|
p1.Connect(p2)
|
|
assert.Equal(t, 1, len(p1.BHost.Network().Peers()), "Expected peers to be connected")
|
|
root := [32]byte{'C'}
|
|
|
|
r := &Service{
|
|
cfg: &Config{
|
|
P2P: p1,
|
|
Chain: &mock.ChainService{
|
|
Fork: ðpb.Fork{
|
|
PreviousVersion: params.BeaconConfig().GenesisForkVersion,
|
|
CurrentVersion: params.BeaconConfig().GenesisForkVersion,
|
|
},
|
|
FinalizedCheckPoint: ðpb.Checkpoint{
|
|
Epoch: 0,
|
|
Root: root[:],
|
|
},
|
|
Genesis: time.Now(),
|
|
ValidatorsRoot: [32]byte{'A'},
|
|
Root: make([]byte, 32),
|
|
},
|
|
},
|
|
rateLimiter: newRateLimiter(p1),
|
|
}
|
|
pcl := protocol.ID(p2p.RPCStatusTopicV1)
|
|
topic := string(pcl)
|
|
r.rateLimiter.limiterMap[topic] = leakybucket.NewCollector(1, 1, false)
|
|
|
|
var wg sync.WaitGroup
|
|
wg.Add(1)
|
|
p2.BHost.SetStreamHandler(pcl, func(stream network.Stream) {
|
|
defer wg.Done()
|
|
expectSuccess(t, stream)
|
|
out := &pb.Status{}
|
|
assert.NoError(t, r.cfg.P2P.Encoding().DecodeWithMaxLength(stream, out))
|
|
assert.DeepEqual(t, root[:], out.FinalizedRoot)
|
|
assert.NoError(t, stream.Close())
|
|
})
|
|
|
|
pcl2 := protocol.ID("/eth2/beacon_chain/req/goodbye/1/ssz_snappy")
|
|
topic = string(pcl2)
|
|
r.rateLimiter.limiterMap[topic] = leakybucket.NewCollector(1, 1, false)
|
|
var wg2 sync.WaitGroup
|
|
wg2.Add(1)
|
|
p2.BHost.SetStreamHandler(pcl2, func(stream network.Stream) {
|
|
defer wg2.Done()
|
|
msg := new(types.SSZUint64)
|
|
assert.NoError(t, r.cfg.P2P.Encoding().DecodeWithMaxLength(stream, msg))
|
|
assert.Equal(t, p2ptypes.GoodbyeCodeWrongNetwork, *msg)
|
|
assert.NoError(t, stream.Close())
|
|
})
|
|
|
|
stream1, err := p1.BHost.NewStream(context.Background(), p2.BHost.ID(), pcl)
|
|
require.NoError(t, err)
|
|
assert.NoError(t, r.statusRPCHandler(context.Background(), &pb.Status{ForkDigest: bytesutil.PadTo([]byte("f"), 4), HeadRoot: make([]byte, 32), FinalizedRoot: make([]byte, 32)}, stream1))
|
|
|
|
if testutil.WaitTimeout(&wg, 1*time.Second) {
|
|
t.Fatal("Did not receive stream within 1 sec")
|
|
}
|
|
if testutil.WaitTimeout(&wg2, 1*time.Second) {
|
|
t.Fatal("Did not receive stream within 1 sec")
|
|
}
|
|
|
|
assert.Equal(t, 0, len(p1.BHost.Network().Peers()), "handler did not disconnect peer")
|
|
}
|
|
|
|
func TestStatusRPCHandler_ConnectsOnGenesis(t *testing.T) {
|
|
p1 := p2ptest.NewTestP2P(t)
|
|
p2 := p2ptest.NewTestP2P(t)
|
|
p1.Connect(p2)
|
|
assert.Equal(t, 1, len(p1.BHost.Network().Peers()), "Expected peers to be connected")
|
|
root := [32]byte{}
|
|
|
|
r := &Service{
|
|
cfg: &Config{
|
|
P2P: p1,
|
|
Chain: &mock.ChainService{
|
|
Fork: ðpb.Fork{
|
|
PreviousVersion: params.BeaconConfig().GenesisForkVersion,
|
|
CurrentVersion: params.BeaconConfig().GenesisForkVersion,
|
|
},
|
|
FinalizedCheckPoint: ðpb.Checkpoint{
|
|
Epoch: 0,
|
|
Root: params.BeaconConfig().ZeroHash[:],
|
|
},
|
|
Genesis: time.Now(),
|
|
ValidatorsRoot: [32]byte{'A'},
|
|
Root: make([]byte, 32),
|
|
},
|
|
},
|
|
rateLimiter: newRateLimiter(p1),
|
|
}
|
|
pcl := protocol.ID(p2p.RPCStatusTopicV1)
|
|
topic := string(pcl)
|
|
r.rateLimiter.limiterMap[topic] = leakybucket.NewCollector(1, 1, false)
|
|
|
|
var wg sync.WaitGroup
|
|
wg.Add(1)
|
|
p2.BHost.SetStreamHandler(pcl, func(stream network.Stream) {
|
|
defer wg.Done()
|
|
expectSuccess(t, stream)
|
|
out := &pb.Status{}
|
|
assert.NoError(t, r.cfg.P2P.Encoding().DecodeWithMaxLength(stream, out))
|
|
assert.DeepEqual(t, root[:], out.FinalizedRoot)
|
|
})
|
|
|
|
stream1, err := p1.BHost.NewStream(context.Background(), p2.BHost.ID(), pcl)
|
|
require.NoError(t, err)
|
|
digest, err := r.currentForkDigest()
|
|
require.NoError(t, err)
|
|
|
|
err = r.statusRPCHandler(context.Background(), &pb.Status{ForkDigest: digest[:], FinalizedRoot: params.BeaconConfig().ZeroHash[:]}, stream1)
|
|
assert.NoError(t, err)
|
|
|
|
if testutil.WaitTimeout(&wg, 1*time.Second) {
|
|
t.Fatal("Did not receive stream within 1 sec")
|
|
}
|
|
|
|
assert.Equal(t, 1, len(p1.BHost.Network().Peers()), "Handler disconnected with peer")
|
|
}
|
|
|
|
func TestStatusRPCHandler_ReturnsHelloMessage(t *testing.T) {
|
|
p1 := p2ptest.NewTestP2P(t)
|
|
p2 := p2ptest.NewTestP2P(t)
|
|
p1.Connect(p2)
|
|
assert.Equal(t, 1, len(p1.BHost.Network().Peers()), "Expected peers to be connected")
|
|
db := testingDB.SetupDB(t)
|
|
|
|
// Set up a head state with data we expect.
|
|
head := testutil.NewBeaconBlock()
|
|
head.Block.Slot = 111
|
|
headRoot, err := head.Block.HashTreeRoot()
|
|
require.NoError(t, err)
|
|
blkSlot := 3 * params.BeaconConfig().SlotsPerEpoch
|
|
finalized := testutil.NewBeaconBlock()
|
|
finalized.Block.Slot = blkSlot
|
|
finalizedRoot, err := finalized.Block.HashTreeRoot()
|
|
require.NoError(t, err)
|
|
genesisState, err := transition.GenesisBeaconState(context.Background(), nil, 0, ðpb.Eth1Data{})
|
|
require.NoError(t, err)
|
|
require.NoError(t, genesisState.SetSlot(111))
|
|
require.NoError(t, genesisState.UpdateBlockRootAtIndex(111%uint64(params.BeaconConfig().SlotsPerHistoricalRoot), headRoot))
|
|
require.NoError(t, db.SaveBlock(context.Background(), wrapper.WrappedPhase0SignedBeaconBlock(finalized)))
|
|
require.NoError(t, db.SaveGenesisBlockRoot(context.Background(), finalizedRoot))
|
|
finalizedCheckpt := ðpb.Checkpoint{
|
|
Epoch: 3,
|
|
Root: finalizedRoot[:],
|
|
}
|
|
totalSec := int64(params.BeaconConfig().SlotsPerEpoch.Mul(5 * params.BeaconConfig().SecondsPerSlot))
|
|
genTime := time.Now().Unix() - totalSec
|
|
|
|
r := &Service{
|
|
cfg: &Config{
|
|
P2P: p1,
|
|
Chain: &mock.ChainService{
|
|
State: genesisState,
|
|
FinalizedCheckPoint: finalizedCheckpt,
|
|
Root: headRoot[:],
|
|
Fork: ðpb.Fork{
|
|
PreviousVersion: params.BeaconConfig().GenesisForkVersion,
|
|
CurrentVersion: params.BeaconConfig().GenesisForkVersion,
|
|
},
|
|
ValidatorsRoot: [32]byte{'A'},
|
|
Genesis: time.Unix(genTime, 0),
|
|
},
|
|
DB: db,
|
|
},
|
|
rateLimiter: newRateLimiter(p1),
|
|
}
|
|
digest, err := r.currentForkDigest()
|
|
require.NoError(t, err)
|
|
|
|
// Setup streams
|
|
pcl := protocol.ID(p2p.RPCStatusTopicV1)
|
|
topic := string(pcl)
|
|
r.rateLimiter.limiterMap[topic] = leakybucket.NewCollector(1, 1, false)
|
|
var wg sync.WaitGroup
|
|
wg.Add(1)
|
|
p2.BHost.SetStreamHandler(pcl, func(stream network.Stream) {
|
|
defer wg.Done()
|
|
expectSuccess(t, stream)
|
|
out := &pb.Status{}
|
|
assert.NoError(t, r.cfg.P2P.Encoding().DecodeWithMaxLength(stream, out))
|
|
expected := &pb.Status{
|
|
ForkDigest: digest[:],
|
|
HeadSlot: genesisState.Slot(),
|
|
HeadRoot: headRoot[:],
|
|
FinalizedEpoch: 3,
|
|
FinalizedRoot: finalizedRoot[:],
|
|
}
|
|
if !proto.Equal(out, expected) {
|
|
t.Errorf("Did not receive expected message. Got %+v wanted %+v", out, expected)
|
|
}
|
|
})
|
|
stream1, err := p1.BHost.NewStream(context.Background(), p2.BHost.ID(), pcl)
|
|
require.NoError(t, err)
|
|
|
|
err = r.statusRPCHandler(context.Background(), &pb.Status{
|
|
ForkDigest: digest[:],
|
|
FinalizedRoot: finalizedRoot[:],
|
|
FinalizedEpoch: 3,
|
|
}, stream1)
|
|
assert.NoError(t, err)
|
|
|
|
if testutil.WaitTimeout(&wg, 1*time.Second) {
|
|
t.Fatal("Did not receive stream within 1 sec")
|
|
}
|
|
}
|
|
|
|
func TestHandshakeHandlers_Roundtrip(t *testing.T) {
|
|
// Scenario is that p1 and p2 connect, exchange handshakes.
|
|
// p2 disconnects and p1 should forget the handshake status.
|
|
p1 := p2ptest.NewTestP2P(t)
|
|
p2 := p2ptest.NewTestP2P(t)
|
|
db := testingDB.SetupDB(t)
|
|
|
|
p1.LocalMetadata = p2pWrapper.WrappedMetadataV0(&pb.MetaDataV0{
|
|
SeqNumber: 2,
|
|
Attnets: bytesutil.PadTo([]byte{'A', 'B'}, 8),
|
|
})
|
|
|
|
p2.LocalMetadata = p2pWrapper.WrappedMetadataV0(&pb.MetaDataV0{
|
|
SeqNumber: 2,
|
|
Attnets: bytesutil.PadTo([]byte{'C', 'D'}, 8),
|
|
})
|
|
|
|
st, err := v1.InitializeFromProto(ðpb.BeaconState{
|
|
Slot: 5,
|
|
})
|
|
require.NoError(t, err)
|
|
blk := testutil.NewBeaconBlock()
|
|
blk.Block.Slot = 0
|
|
require.NoError(t, db.SaveBlock(context.Background(), wrapper.WrappedPhase0SignedBeaconBlock(blk)))
|
|
finalizedRoot, err := blk.Block.HashTreeRoot()
|
|
require.NoError(t, err)
|
|
require.NoError(t, db.SaveGenesisBlockRoot(context.Background(), finalizedRoot))
|
|
r := &Service{
|
|
cfg: &Config{
|
|
P2P: p1,
|
|
Chain: &mock.ChainService{
|
|
State: st,
|
|
FinalizedCheckPoint: ðpb.Checkpoint{Epoch: 0, Root: finalizedRoot[:]},
|
|
Fork: ðpb.Fork{
|
|
PreviousVersion: params.BeaconConfig().GenesisForkVersion,
|
|
CurrentVersion: params.BeaconConfig().GenesisForkVersion,
|
|
},
|
|
Genesis: time.Now(),
|
|
ValidatorsRoot: [32]byte{'A'},
|
|
Root: make([]byte, 32),
|
|
},
|
|
DB: db,
|
|
},
|
|
ctx: context.Background(),
|
|
rateLimiter: newRateLimiter(p1),
|
|
}
|
|
p1.Digest, err = r.currentForkDigest()
|
|
require.NoError(t, err)
|
|
|
|
r2 := &Service{
|
|
cfg: &Config{
|
|
Chain: &mock.ChainService{
|
|
FinalizedCheckPoint: ðpb.Checkpoint{Epoch: 0, Root: finalizedRoot[:]},
|
|
},
|
|
P2P: p2,
|
|
},
|
|
rateLimiter: newRateLimiter(p2),
|
|
}
|
|
p2.Digest, err = r.currentForkDigest()
|
|
require.NoError(t, err)
|
|
|
|
r.Start()
|
|
|
|
// Setup streams
|
|
pcl := protocol.ID("/eth2/beacon_chain/req/status/1/ssz_snappy")
|
|
topic := string(pcl)
|
|
r.rateLimiter.limiterMap[topic] = leakybucket.NewCollector(1, 1, false)
|
|
var wg sync.WaitGroup
|
|
wg.Add(1)
|
|
p2.BHost.SetStreamHandler(pcl, func(stream network.Stream) {
|
|
defer wg.Done()
|
|
out := &pb.Status{}
|
|
assert.NoError(t, r.cfg.P2P.Encoding().DecodeWithMaxLength(stream, out))
|
|
log.WithField("status", out).Warn("received status")
|
|
resp := &pb.Status{HeadSlot: 100, HeadRoot: make([]byte, 32), ForkDigest: p2.Digest[:],
|
|
FinalizedRoot: finalizedRoot[:], FinalizedEpoch: 0}
|
|
_, err := stream.Write([]byte{responseCodeSuccess})
|
|
assert.NoError(t, err)
|
|
_, err = r.cfg.P2P.Encoding().EncodeWithMaxLength(stream, resp)
|
|
assert.NoError(t, err)
|
|
log.WithField("status", out).Warn("sending status")
|
|
if err := stream.Close(); err != nil {
|
|
t.Log(err)
|
|
}
|
|
})
|
|
|
|
pcl = "/eth2/beacon_chain/req/ping/1/ssz_snappy"
|
|
topic = string(pcl)
|
|
r2.rateLimiter.limiterMap[topic] = leakybucket.NewCollector(1, 1, false)
|
|
var wg2 sync.WaitGroup
|
|
wg2.Add(1)
|
|
p2.BHost.SetStreamHandler(pcl, func(stream network.Stream) {
|
|
defer wg2.Done()
|
|
out := new(types.SSZUint64)
|
|
assert.NoError(t, r.cfg.P2P.Encoding().DecodeWithMaxLength(stream, out))
|
|
assert.Equal(t, uint64(2), uint64(*out))
|
|
assert.NoError(t, r2.pingHandler(context.Background(), out, stream))
|
|
assert.NoError(t, stream.Close())
|
|
})
|
|
|
|
numInactive1 := len(p1.Peers().Inactive())
|
|
numActive1 := len(p1.Peers().Active())
|
|
|
|
p1.Connect(p2)
|
|
|
|
p1.Peers().Add(new(enr.Record), p2.BHost.ID(), p2.BHost.Addrs()[0], network.DirUnknown)
|
|
p1.Peers().SetMetadata(p2.BHost.ID(), p2.LocalMetadata)
|
|
|
|
p2.Peers().Add(new(enr.Record), p1.BHost.ID(), p1.BHost.Addrs()[0], network.DirUnknown)
|
|
p2.Peers().SetMetadata(p1.BHost.ID(), p1.LocalMetadata)
|
|
|
|
if testutil.WaitTimeout(&wg, 1*time.Second) {
|
|
t.Fatal("Did not receive stream within 1 sec")
|
|
}
|
|
if testutil.WaitTimeout(&wg2, 1*time.Second) {
|
|
t.Fatal("Did not receive stream within 1 sec")
|
|
}
|
|
|
|
// Wait for stream buffer to be read.
|
|
time.Sleep(200 * time.Millisecond)
|
|
|
|
numInactive2 := len(p1.Peers().Inactive())
|
|
numActive2 := len(p1.Peers().Active())
|
|
|
|
assert.Equal(t, numInactive1, numInactive1, "Number of inactive peers changed unexpectedly")
|
|
assert.Equal(t, numActive1+1, numActive2, "Number of active peers unexpected")
|
|
|
|
require.NoError(t, p2.Disconnect(p1.PeerID()))
|
|
p1.Peers().SetConnectionState(p2.PeerID(), peers.PeerDisconnected)
|
|
|
|
// Wait for disconnect event to trigger.
|
|
time.Sleep(200 * time.Millisecond)
|
|
|
|
numInactive3 := len(p1.Peers().Inactive())
|
|
numActive3 := len(p1.Peers().Active())
|
|
assert.Equal(t, numInactive2+1, numInactive3, "Number of inactive peers unexpected")
|
|
assert.Equal(t, numActive2-1, numActive3, "Number of active peers unexpected")
|
|
}
|
|
|
|
func TestStatusRPCRequest_RequestSent(t *testing.T) {
|
|
p1 := p2ptest.NewTestP2P(t)
|
|
p2 := p2ptest.NewTestP2P(t)
|
|
|
|
// Set up a head state with data we expect.
|
|
head := testutil.NewBeaconBlock()
|
|
head.Block.Slot = 111
|
|
headRoot, err := head.Block.HashTreeRoot()
|
|
require.NoError(t, err)
|
|
finalized := testutil.NewBeaconBlock()
|
|
finalized.Block.Slot = 40
|
|
finalizedRoot, err := finalized.Block.HashTreeRoot()
|
|
require.NoError(t, err)
|
|
genesisState, err := transition.GenesisBeaconState(context.Background(), nil, 0, ðpb.Eth1Data{})
|
|
require.NoError(t, err)
|
|
require.NoError(t, genesisState.SetSlot(111))
|
|
require.NoError(t, genesisState.UpdateBlockRootAtIndex(111%uint64(params.BeaconConfig().SlotsPerHistoricalRoot), headRoot))
|
|
finalizedCheckpt := ðpb.Checkpoint{
|
|
Epoch: 5,
|
|
Root: finalizedRoot[:],
|
|
}
|
|
|
|
r := &Service{
|
|
cfg: &Config{
|
|
P2P: p1,
|
|
Chain: &mock.ChainService{
|
|
State: genesisState,
|
|
FinalizedCheckPoint: finalizedCheckpt,
|
|
Root: headRoot[:],
|
|
Fork: ðpb.Fork{
|
|
PreviousVersion: params.BeaconConfig().GenesisForkVersion,
|
|
CurrentVersion: params.BeaconConfig().GenesisForkVersion,
|
|
},
|
|
Genesis: time.Now(),
|
|
ValidatorsRoot: [32]byte{'A'},
|
|
},
|
|
},
|
|
ctx: context.Background(),
|
|
rateLimiter: newRateLimiter(p1),
|
|
}
|
|
|
|
// Setup streams
|
|
pcl := protocol.ID("/eth2/beacon_chain/req/status/1/ssz_snappy")
|
|
topic := string(pcl)
|
|
r.rateLimiter.limiterMap[topic] = leakybucket.NewCollector(1, 1, false)
|
|
var wg sync.WaitGroup
|
|
wg.Add(1)
|
|
p2.BHost.SetStreamHandler(pcl, func(stream network.Stream) {
|
|
defer wg.Done()
|
|
out := &pb.Status{}
|
|
assert.NoError(t, r.cfg.P2P.Encoding().DecodeWithMaxLength(stream, out))
|
|
digest, err := r.currentForkDigest()
|
|
assert.NoError(t, err)
|
|
expected := &pb.Status{
|
|
ForkDigest: digest[:],
|
|
HeadSlot: genesisState.Slot(),
|
|
HeadRoot: headRoot[:],
|
|
FinalizedEpoch: 5,
|
|
FinalizedRoot: finalizedRoot[:],
|
|
}
|
|
if !proto.Equal(out, expected) {
|
|
t.Errorf("Did not receive expected message. Got %+v wanted %+v", out, expected)
|
|
}
|
|
})
|
|
|
|
p1.AddConnectionHandler(r.sendRPCStatusRequest, nil)
|
|
p1.Connect(p2)
|
|
|
|
if testutil.WaitTimeout(&wg, 1*time.Second) {
|
|
t.Fatal("Did not receive stream within 1 sec")
|
|
}
|
|
|
|
assert.Equal(t, 1, len(p1.BHost.Network().Peers()), "Expected peers to continue being connected")
|
|
}
|
|
|
|
func TestStatusRPCRequest_FinalizedBlockExists(t *testing.T) {
|
|
p1 := p2ptest.NewTestP2P(t)
|
|
p2 := p2ptest.NewTestP2P(t)
|
|
db := testingDB.SetupDB(t)
|
|
|
|
// Set up a head state with data we expect.
|
|
head := testutil.NewBeaconBlock()
|
|
head.Block.Slot = 111
|
|
headRoot, err := head.Block.HashTreeRoot()
|
|
require.NoError(t, err)
|
|
blkSlot := 3 * params.BeaconConfig().SlotsPerEpoch
|
|
finalized := testutil.NewBeaconBlock()
|
|
finalized.Block.Slot = blkSlot
|
|
finalizedRoot, err := finalized.Block.HashTreeRoot()
|
|
require.NoError(t, err)
|
|
genesisState, err := transition.GenesisBeaconState(context.Background(), nil, 0, ðpb.Eth1Data{DepositRoot: make([]byte, 32), BlockHash: make([]byte, 32)})
|
|
require.NoError(t, err)
|
|
require.NoError(t, genesisState.SetSlot(111))
|
|
require.NoError(t, genesisState.UpdateBlockRootAtIndex(111%uint64(params.BeaconConfig().SlotsPerHistoricalRoot), headRoot))
|
|
blk := testutil.NewBeaconBlock()
|
|
blk.Block.Slot = blkSlot
|
|
require.NoError(t, db.SaveBlock(context.Background(), wrapper.WrappedPhase0SignedBeaconBlock(blk)))
|
|
require.NoError(t, db.SaveGenesisBlockRoot(context.Background(), finalizedRoot))
|
|
finalizedCheckpt := ðpb.Checkpoint{
|
|
Epoch: 3,
|
|
Root: finalizedRoot[:],
|
|
}
|
|
totalSec := int64(params.BeaconConfig().SlotsPerEpoch.Mul(5 * params.BeaconConfig().SecondsPerSlot))
|
|
genTime := time.Now().Unix() - totalSec
|
|
r := &Service{
|
|
cfg: &Config{
|
|
P2P: p1,
|
|
Chain: &mock.ChainService{
|
|
State: genesisState,
|
|
FinalizedCheckPoint: finalizedCheckpt,
|
|
Root: headRoot[:],
|
|
Fork: ðpb.Fork{
|
|
PreviousVersion: params.BeaconConfig().GenesisForkVersion,
|
|
CurrentVersion: params.BeaconConfig().GenesisForkVersion,
|
|
},
|
|
Genesis: time.Unix(genTime, 0),
|
|
ValidatorsRoot: [32]byte{'A'},
|
|
},
|
|
},
|
|
ctx: context.Background(),
|
|
rateLimiter: newRateLimiter(p1),
|
|
}
|
|
|
|
r2 := &Service{
|
|
cfg: &Config{
|
|
P2P: p1,
|
|
Chain: &mock.ChainService{
|
|
State: genesisState,
|
|
FinalizedCheckPoint: finalizedCheckpt,
|
|
Root: headRoot[:],
|
|
Fork: ðpb.Fork{
|
|
PreviousVersion: params.BeaconConfig().GenesisForkVersion,
|
|
CurrentVersion: params.BeaconConfig().GenesisForkVersion,
|
|
},
|
|
Genesis: time.Unix(genTime, 0),
|
|
ValidatorsRoot: [32]byte{'A'},
|
|
},
|
|
DB: db,
|
|
},
|
|
ctx: context.Background(),
|
|
rateLimiter: newRateLimiter(p1),
|
|
}
|
|
|
|
// Setup streams
|
|
pcl := protocol.ID("/eth2/beacon_chain/req/status/1/ssz_snappy")
|
|
topic := string(pcl)
|
|
r.rateLimiter.limiterMap[topic] = leakybucket.NewCollector(1, 1, false)
|
|
var wg sync.WaitGroup
|
|
wg.Add(1)
|
|
p2.BHost.SetStreamHandler(pcl, func(stream network.Stream) {
|
|
defer wg.Done()
|
|
out := &pb.Status{}
|
|
assert.NoError(t, r.cfg.P2P.Encoding().DecodeWithMaxLength(stream, out))
|
|
assert.NoError(t, r2.validateStatusMessage(context.Background(), out))
|
|
})
|
|
|
|
p1.AddConnectionHandler(r.sendRPCStatusRequest, nil)
|
|
p1.Connect(p2)
|
|
|
|
if testutil.WaitTimeout(&wg, 1*time.Second) {
|
|
t.Fatal("Did not receive stream within 1 sec")
|
|
}
|
|
|
|
assert.Equal(t, 1, len(p1.BHost.Network().Peers()), "Expected peers to continue being connected")
|
|
}
|
|
|
|
func TestStatusRPCRequest_FinalizedBlockSkippedSlots(t *testing.T) {
|
|
db, err := kv.NewKVStore(context.Background(), t.TempDir(), &kv.Config{})
|
|
require.NoError(t, err)
|
|
bState, err := transition.GenesisBeaconState(context.Background(), nil, 0, ðpb.Eth1Data{DepositRoot: make([]byte, 32), BlockHash: make([]byte, 32)})
|
|
require.NoError(t, err)
|
|
|
|
blk := testutil.NewBeaconBlock()
|
|
blk.Block.Slot = 0
|
|
genRoot, err := blk.Block.HashTreeRoot()
|
|
require.NoError(t, err)
|
|
|
|
require.NoError(t, db.SaveBlock(context.Background(), wrapper.WrappedPhase0SignedBeaconBlock(blk)))
|
|
require.NoError(t, db.SaveGenesisBlockRoot(context.Background(), genRoot))
|
|
blocksTillHead := makeBlocks(t, 1, 1000, genRoot)
|
|
require.NoError(t, db.SaveBlocks(context.Background(), blocksTillHead))
|
|
|
|
stateSummaries := make([]*ethpb.StateSummary, len(blocksTillHead))
|
|
for i, b := range blocksTillHead {
|
|
bRoot, err := b.Block().HashTreeRoot()
|
|
require.NoError(t, err)
|
|
stateSummaries[i] = ðpb.StateSummary{
|
|
Slot: b.Block().Slot(),
|
|
Root: bRoot[:],
|
|
}
|
|
}
|
|
require.NoError(t, db.SaveStateSummaries(context.Background(), stateSummaries))
|
|
|
|
rootFetcher := func(slot types.Slot) [32]byte {
|
|
rt, err := blocksTillHead[slot-1].Block().HashTreeRoot()
|
|
require.NoError(t, err)
|
|
return rt
|
|
}
|
|
tests := []struct {
|
|
name string
|
|
expectedFinalizedEpoch types.Epoch
|
|
expectedFinalizedRoot [32]byte
|
|
headSlot types.Slot
|
|
remoteFinalizedEpoch types.Epoch
|
|
remoteFinalizedRoot [32]byte
|
|
remoteHeadSlot types.Slot
|
|
expectError bool
|
|
}{
|
|
{
|
|
name: "valid finalized epoch",
|
|
expectedFinalizedEpoch: 3,
|
|
expectedFinalizedRoot: rootFetcher(3 * params.BeaconConfig().SlotsPerEpoch),
|
|
headSlot: 111,
|
|
remoteFinalizedEpoch: 3,
|
|
remoteFinalizedRoot: rootFetcher(3 * params.BeaconConfig().SlotsPerEpoch),
|
|
remoteHeadSlot: 100,
|
|
expectError: false,
|
|
},
|
|
{
|
|
name: "invalid finalized epoch",
|
|
expectedFinalizedEpoch: 3,
|
|
expectedFinalizedRoot: rootFetcher(3 * params.BeaconConfig().SlotsPerEpoch),
|
|
headSlot: 111,
|
|
remoteFinalizedEpoch: 3,
|
|
// give an incorrect root relative to the finalized epoch.
|
|
remoteFinalizedRoot: rootFetcher(2 * params.BeaconConfig().SlotsPerEpoch),
|
|
remoteHeadSlot: 120,
|
|
expectError: true,
|
|
},
|
|
{
|
|
name: "invalid finalized root",
|
|
expectedFinalizedEpoch: 3,
|
|
expectedFinalizedRoot: rootFetcher(3 * params.BeaconConfig().SlotsPerEpoch),
|
|
headSlot: 111,
|
|
remoteFinalizedEpoch: 3,
|
|
// give a bad finalized root, and the beacon node verifies that
|
|
// it is indeed incorrect.
|
|
remoteFinalizedRoot: [32]byte{'a', 'b', 'c'},
|
|
remoteHeadSlot: 120,
|
|
expectError: true,
|
|
},
|
|
}
|
|
|
|
for _, tt := range tests {
|
|
p1 := p2ptest.NewTestP2P(t)
|
|
p2 := p2ptest.NewTestP2P(t)
|
|
|
|
expectedFinalizedEpoch := tt.expectedFinalizedEpoch
|
|
headSlot := tt.headSlot
|
|
|
|
nState := bState.Copy()
|
|
// Set up a head state with data we expect.
|
|
head := blocksTillHead[len(blocksTillHead)-1]
|
|
headRoot, err := head.Block().HashTreeRoot()
|
|
require.NoError(t, err)
|
|
|
|
rHead := blocksTillHead[tt.remoteHeadSlot-1]
|
|
rHeadRoot, err := rHead.Block().HashTreeRoot()
|
|
require.NoError(t, err)
|
|
|
|
require.NoError(t, nState.SetSlot(headSlot))
|
|
require.NoError(t, nState.UpdateBlockRootAtIndex(uint64(headSlot.ModSlot(params.BeaconConfig().SlotsPerHistoricalRoot)), headRoot))
|
|
|
|
finalizedCheckpt := ðpb.Checkpoint{
|
|
Epoch: expectedFinalizedEpoch,
|
|
Root: tt.expectedFinalizedRoot[:],
|
|
}
|
|
|
|
remoteFinalizedChkpt := ðpb.Checkpoint{
|
|
Epoch: tt.remoteFinalizedEpoch,
|
|
Root: tt.remoteFinalizedRoot[:],
|
|
}
|
|
require.NoError(t, db.SaveFinalizedCheckpoint(context.Background(), finalizedCheckpt))
|
|
|
|
epoch := expectedFinalizedEpoch.Add(2)
|
|
totalSec := uint64(params.BeaconConfig().SlotsPerEpoch.Mul(uint64(epoch) * params.BeaconConfig().SecondsPerSlot))
|
|
genTime := time.Now().Unix() - int64(totalSec)
|
|
r := &Service{
|
|
cfg: &Config{
|
|
P2P: p1,
|
|
Chain: &mock.ChainService{
|
|
State: nState,
|
|
FinalizedCheckPoint: remoteFinalizedChkpt,
|
|
Root: rHeadRoot[:],
|
|
Fork: ðpb.Fork{
|
|
PreviousVersion: params.BeaconConfig().GenesisForkVersion,
|
|
CurrentVersion: params.BeaconConfig().GenesisForkVersion,
|
|
},
|
|
Genesis: time.Unix(genTime, 0),
|
|
ValidatorsRoot: [32]byte{'A'},
|
|
},
|
|
},
|
|
ctx: context.Background(),
|
|
rateLimiter: newRateLimiter(p1),
|
|
}
|
|
|
|
r2 := &Service{
|
|
cfg: &Config{
|
|
P2P: p2,
|
|
Chain: &mock.ChainService{
|
|
State: nState,
|
|
FinalizedCheckPoint: finalizedCheckpt,
|
|
Root: headRoot[:],
|
|
Fork: ðpb.Fork{
|
|
PreviousVersion: params.BeaconConfig().GenesisForkVersion,
|
|
CurrentVersion: params.BeaconConfig().GenesisForkVersion,
|
|
},
|
|
Genesis: time.Unix(genTime, 0),
|
|
ValidatorsRoot: [32]byte{'A'},
|
|
},
|
|
DB: db,
|
|
},
|
|
|
|
ctx: context.Background(),
|
|
rateLimiter: newRateLimiter(p1),
|
|
}
|
|
|
|
// Setup streams
|
|
pcl := protocol.ID("/eth2/beacon_chain/req/status/1/ssz_snappy")
|
|
topic := string(pcl)
|
|
r.rateLimiter.limiterMap[topic] = leakybucket.NewCollector(1, 1, false)
|
|
var wg sync.WaitGroup
|
|
wg.Add(1)
|
|
p2.BHost.SetStreamHandler(pcl, func(stream network.Stream) {
|
|
defer wg.Done()
|
|
out := &pb.Status{}
|
|
assert.NoError(t, r.cfg.P2P.Encoding().DecodeWithMaxLength(stream, out))
|
|
assert.Equal(t, tt.expectError, r2.validateStatusMessage(context.Background(), out) != nil)
|
|
})
|
|
|
|
p1.AddConnectionHandler(r.sendRPCStatusRequest, nil)
|
|
p1.Connect(p2)
|
|
|
|
if testutil.WaitTimeout(&wg, 1*time.Second) {
|
|
t.Fatal("Did not receive stream within 1 sec")
|
|
}
|
|
|
|
assert.Equal(t, 1, len(p1.BHost.Network().Peers()), "Expected peers to continue being connected")
|
|
assert.NoError(t, p1.Disconnect(p2.PeerID()))
|
|
}
|
|
assert.NoError(t, db.Close())
|
|
}
|
|
|
|
func TestStatusRPCRequest_BadPeerHandshake(t *testing.T) {
|
|
p1 := p2ptest.NewTestP2P(t)
|
|
p2 := p2ptest.NewTestP2P(t)
|
|
|
|
// Set up a head state with data we expect.
|
|
head := testutil.NewBeaconBlock()
|
|
head.Block.Slot = 111
|
|
headRoot, err := head.Block.HashTreeRoot()
|
|
require.NoError(t, err)
|
|
finalized := testutil.NewBeaconBlock()
|
|
finalizedRoot, err := finalized.Block.HashTreeRoot()
|
|
require.NoError(t, err)
|
|
genesisState, err := transition.GenesisBeaconState(context.Background(), nil, 0, ðpb.Eth1Data{})
|
|
require.NoError(t, err)
|
|
require.NoError(t, genesisState.SetSlot(111))
|
|
require.NoError(t, genesisState.UpdateBlockRootAtIndex(111%uint64(params.BeaconConfig().SlotsPerHistoricalRoot), headRoot))
|
|
finalizedCheckpt := ðpb.Checkpoint{
|
|
Epoch: 5,
|
|
Root: finalizedRoot[:],
|
|
}
|
|
|
|
r := &Service{
|
|
cfg: &Config{
|
|
P2P: p1,
|
|
Chain: &mock.ChainService{
|
|
State: genesisState,
|
|
FinalizedCheckPoint: finalizedCheckpt,
|
|
Root: headRoot[:],
|
|
Fork: ðpb.Fork{
|
|
PreviousVersion: params.BeaconConfig().GenesisForkVersion,
|
|
CurrentVersion: params.BeaconConfig().GenesisForkVersion,
|
|
},
|
|
Genesis: time.Now(),
|
|
ValidatorsRoot: [32]byte{'A'},
|
|
},
|
|
},
|
|
|
|
ctx: context.Background(),
|
|
rateLimiter: newRateLimiter(p1),
|
|
}
|
|
|
|
r.Start()
|
|
|
|
// Setup streams
|
|
pcl := protocol.ID("/eth2/beacon_chain/req/status/1/ssz_snappy")
|
|
topic := string(pcl)
|
|
r.rateLimiter.limiterMap[topic] = leakybucket.NewCollector(1, 1, false)
|
|
var wg sync.WaitGroup
|
|
wg.Add(1)
|
|
p2.BHost.SetStreamHandler(pcl, func(stream network.Stream) {
|
|
defer wg.Done()
|
|
out := &pb.Status{}
|
|
assert.NoError(t, r.cfg.P2P.Encoding().DecodeWithMaxLength(stream, out))
|
|
expected := &pb.Status{
|
|
ForkDigest: []byte{1, 1, 1, 1},
|
|
HeadSlot: genesisState.Slot(),
|
|
HeadRoot: headRoot[:],
|
|
FinalizedEpoch: 5,
|
|
FinalizedRoot: finalizedRoot[:],
|
|
}
|
|
if _, err := stream.Write([]byte{responseCodeSuccess}); err != nil {
|
|
log.WithError(err).Debug("Could not write to stream")
|
|
}
|
|
_, err := r.cfg.P2P.Encoding().EncodeWithMaxLength(stream, expected)
|
|
assert.NoError(t, err)
|
|
})
|
|
|
|
assert.Equal(t, false, p1.Peers().Scorers().IsBadPeer(p2.PeerID()), "Peer is marked as bad")
|
|
p1.Connect(p2)
|
|
|
|
if testutil.WaitTimeout(&wg, time.Second) {
|
|
t.Fatal("Did not receive stream within 1 sec")
|
|
}
|
|
time.Sleep(100 * time.Millisecond)
|
|
|
|
connectionState, err := p1.Peers().ConnectionState(p2.PeerID())
|
|
require.NoError(t, err, "Could not obtain peer connection state")
|
|
assert.Equal(t, peers.PeerDisconnected, connectionState, "Expected peer to be disconnected")
|
|
|
|
assert.Equal(t, true, p1.Peers().Scorers().IsBadPeer(p2.PeerID()), "Peer is not marked as bad")
|
|
}
|
|
|
|
func TestStatusRPC_ValidGenesisMessage(t *testing.T) {
|
|
// Set up a head state with data we expect.
|
|
head := testutil.NewBeaconBlock()
|
|
head.Block.Slot = 111
|
|
headRoot, err := head.Block.HashTreeRoot()
|
|
require.NoError(t, err)
|
|
blkSlot := 3 * params.BeaconConfig().SlotsPerEpoch
|
|
finalized := testutil.NewBeaconBlock()
|
|
finalized.Block.Slot = blkSlot
|
|
finalizedRoot, err := finalized.Block.HashTreeRoot()
|
|
require.NoError(t, err)
|
|
genesisState, err := transition.GenesisBeaconState(context.Background(), nil, 0, ðpb.Eth1Data{})
|
|
require.NoError(t, err)
|
|
require.NoError(t, genesisState.SetSlot(111))
|
|
require.NoError(t, genesisState.UpdateBlockRootAtIndex(111%uint64(params.BeaconConfig().SlotsPerHistoricalRoot), headRoot))
|
|
finalizedCheckpt := ðpb.Checkpoint{
|
|
Epoch: 5,
|
|
Root: finalizedRoot[:],
|
|
}
|
|
r := &Service{
|
|
cfg: &Config{
|
|
Chain: &mock.ChainService{
|
|
State: genesisState,
|
|
FinalizedCheckPoint: finalizedCheckpt,
|
|
Root: headRoot[:],
|
|
Fork: ðpb.Fork{
|
|
PreviousVersion: params.BeaconConfig().GenesisForkVersion,
|
|
CurrentVersion: params.BeaconConfig().GenesisForkVersion,
|
|
},
|
|
Genesis: time.Now(),
|
|
ValidatorsRoot: [32]byte{'A'},
|
|
},
|
|
},
|
|
ctx: context.Background(),
|
|
}
|
|
digest, err := r.currentForkDigest()
|
|
require.NoError(t, err)
|
|
// There should be no error for a status message
|
|
// with a genesis checkpoint.
|
|
err = r.validateStatusMessage(r.ctx, &pb.Status{
|
|
ForkDigest: digest[:],
|
|
FinalizedRoot: params.BeaconConfig().ZeroHash[:],
|
|
FinalizedEpoch: 0,
|
|
HeadRoot: headRoot[:],
|
|
HeadSlot: 111,
|
|
})
|
|
require.NoError(t, err)
|
|
}
|
|
|
|
func TestShouldResync(t *testing.T) {
|
|
type args struct {
|
|
genesis time.Time
|
|
syncing bool
|
|
headSlot types.Slot
|
|
}
|
|
tests := []struct {
|
|
name string
|
|
args args
|
|
want bool
|
|
}{
|
|
{
|
|
name: "genesis epoch should not resync when syncing is true",
|
|
args: args{
|
|
headSlot: 31,
|
|
genesis: timeutils.Now(),
|
|
syncing: true,
|
|
},
|
|
want: false,
|
|
},
|
|
{
|
|
name: "genesis epoch should not resync when syncing is false",
|
|
args: args{
|
|
headSlot: 31,
|
|
genesis: timeutils.Now(),
|
|
syncing: false,
|
|
},
|
|
want: false,
|
|
},
|
|
{
|
|
name: "two epochs behind, resync ok",
|
|
args: args{
|
|
headSlot: 31,
|
|
genesis: timeutils.Now().Add(-1 * 96 * time.Duration(params.BeaconConfig().SecondsPerSlot) * time.Second),
|
|
syncing: false,
|
|
},
|
|
want: true,
|
|
},
|
|
{
|
|
name: "two epochs behind, already syncing",
|
|
args: args{
|
|
headSlot: 31,
|
|
genesis: timeutils.Now().Add(-1 * 96 * time.Duration(params.BeaconConfig().SecondsPerSlot) * time.Second),
|
|
syncing: true,
|
|
},
|
|
want: false,
|
|
},
|
|
}
|
|
for _, tt := range tests {
|
|
headState, err := transition.GenesisBeaconState(context.Background(), nil, 0, ðpb.Eth1Data{})
|
|
require.NoError(t, err)
|
|
require.NoError(t, headState.SetSlot(tt.args.headSlot))
|
|
r := &Service{
|
|
cfg: &Config{
|
|
Chain: &mock.ChainService{
|
|
State: headState,
|
|
Genesis: tt.args.genesis,
|
|
},
|
|
InitialSync: &mockSync.Sync{IsSyncing: tt.args.syncing},
|
|
},
|
|
ctx: context.Background(),
|
|
}
|
|
t.Run(tt.name, func(t *testing.T) {
|
|
if got := r.shouldReSync(); got != tt.want {
|
|
t.Errorf("shouldReSync() = %v, want %v", got, tt.want)
|
|
}
|
|
})
|
|
}
|
|
}
|
|
|
|
func makeBlocks(t *testing.T, i, n uint64, previousRoot [32]byte) []block.SignedBeaconBlock {
|
|
blocks := make([]*ethpb.SignedBeaconBlock, n)
|
|
ifaceBlocks := make([]block.SignedBeaconBlock, n)
|
|
for j := i; j < n+i; j++ {
|
|
parentRoot := make([]byte, 32)
|
|
copy(parentRoot, previousRoot[:])
|
|
blocks[j-i] = testutil.NewBeaconBlock()
|
|
blocks[j-i].Block.Slot = types.Slot(j + 1)
|
|
blocks[j-i].Block.ParentRoot = parentRoot
|
|
var err error
|
|
previousRoot, err = blocks[j-i].Block.HashTreeRoot()
|
|
require.NoError(t, err)
|
|
ifaceBlocks[j-i] = wrapper.WrappedPhase0SignedBeaconBlock(blocks[j-i])
|
|
}
|
|
return ifaceBlocks
|
|
}
|