package sync import ( "context" "fmt" "reflect" "sync" "testing" "time" pubsub "github.com/libp2p/go-libp2p-pubsub" pubsubpb "github.com/libp2p/go-libp2p-pubsub/pb" "github.com/libp2p/go-libp2p/core/peer" "github.com/prysmaticlabs/prysm/v4/async/abool" mockChain "github.com/prysmaticlabs/prysm/v4/beacon-chain/blockchain/testing" "github.com/prysmaticlabs/prysm/v4/beacon-chain/cache" "github.com/prysmaticlabs/prysm/v4/beacon-chain/core/signing" db "github.com/prysmaticlabs/prysm/v4/beacon-chain/db/testing" "github.com/prysmaticlabs/prysm/v4/beacon-chain/operations/slashings" "github.com/prysmaticlabs/prysm/v4/beacon-chain/p2p" "github.com/prysmaticlabs/prysm/v4/beacon-chain/p2p/encoder" p2ptest "github.com/prysmaticlabs/prysm/v4/beacon-chain/p2p/testing" "github.com/prysmaticlabs/prysm/v4/beacon-chain/startup" mockSync "github.com/prysmaticlabs/prysm/v4/beacon-chain/sync/initial-sync/testing" lruwrpr "github.com/prysmaticlabs/prysm/v4/cache/lru" "github.com/prysmaticlabs/prysm/v4/cmd/beacon-chain/flags" fieldparams "github.com/prysmaticlabs/prysm/v4/config/fieldparams" "github.com/prysmaticlabs/prysm/v4/config/params" "github.com/prysmaticlabs/prysm/v4/consensus-types/primitives" "github.com/prysmaticlabs/prysm/v4/encoding/bytesutil" "github.com/prysmaticlabs/prysm/v4/network/forks" pb "github.com/prysmaticlabs/prysm/v4/proto/prysm/v1alpha1" "github.com/prysmaticlabs/prysm/v4/testing/assert" "github.com/prysmaticlabs/prysm/v4/testing/require" "github.com/prysmaticlabs/prysm/v4/testing/util" "github.com/prysmaticlabs/prysm/v4/time/slots" logTest "github.com/sirupsen/logrus/hooks/test" "google.golang.org/protobuf/proto" ) func TestSubscribe_ReceivesValidMessage(t *testing.T) { p2pService := p2ptest.NewTestP2P(t) gt := time.Now() vr := [32]byte{'A'} r := Service{ ctx: context.Background(), cfg: &config{ p2p: p2pService, initialSync: &mockSync.Sync{IsSyncing: false}, chain: &mockChain.ChainService{ ValidatorsRoot: vr, Genesis: gt, }, clock: startup.NewClock(gt, vr), }, subHandler: newSubTopicHandler(), chainStarted: abool.New(), } var err error p2pService.Digest, err = r.currentForkDigest() require.NoError(t, err) topic := "/eth2/%x/voluntary_exit" var wg sync.WaitGroup wg.Add(1) r.subscribe(topic, r.noopValidator, func(_ context.Context, msg proto.Message) error { m, ok := msg.(*pb.SignedVoluntaryExit) assert.Equal(t, true, ok, "Object is not of type *pb.SignedVoluntaryExit") if m.Exit == nil || m.Exit.Epoch != 55 { t.Errorf("Unexpected incoming message: %+v", m) } wg.Done() return nil }, p2pService.Digest) r.markForChainStart() p2pService.ReceivePubSub(topic, &pb.SignedVoluntaryExit{Exit: &pb.VoluntaryExit{Epoch: 55}, Signature: make([]byte, fieldparams.BLSSignatureLength)}) if util.WaitTimeout(&wg, time.Second) { t.Fatal("Did not receive PubSub in 1 second") } } func TestSubscribe_UnsubscribeTopic(t *testing.T) { p2pService := p2ptest.NewTestP2P(t) gt := time.Now() vr := [32]byte{'A'} r := Service{ ctx: context.Background(), cfg: &config{ p2p: p2pService, initialSync: &mockSync.Sync{IsSyncing: false}, chain: &mockChain.ChainService{ ValidatorsRoot: vr, Genesis: gt, }, clock: startup.NewClock(gt, vr), }, chainStarted: abool.New(), subHandler: newSubTopicHandler(), } var err error p2pService.Digest, err = r.currentForkDigest() require.NoError(t, err) topic := "/eth2/%x/voluntary_exit" r.subscribe(topic, r.noopValidator, func(_ context.Context, msg proto.Message) error { return nil }, p2pService.Digest) r.markForChainStart() fullTopic := fmt.Sprintf(topic, p2pService.Digest) + p2pService.Encoding().ProtocolSuffix() assert.Equal(t, true, r.subHandler.topicExists(fullTopic)) topics := p2pService.PubSub().GetTopics() assert.Equal(t, fullTopic, topics[0]) r.unSubscribeFromTopic(fullTopic) assert.Equal(t, false, r.subHandler.topicExists(fullTopic)) assert.Equal(t, 0, len(p2pService.PubSub().GetTopics())) } func TestSubscribe_ReceivesAttesterSlashing(t *testing.T) { params.SetupTestConfigCleanup(t) cfg := params.MainnetConfig().Copy() cfg.SecondsPerSlot = 1 params.OverrideBeaconConfig(cfg) p2pService := p2ptest.NewTestP2P(t) ctx := context.Background() d := db.SetupDB(t) gt := time.Now() vr := [32]byte{'A'} chainService := &mockChain.ChainService{ Genesis: gt, ValidatorsRoot: vr, } r := Service{ ctx: ctx, cfg: &config{ p2p: p2pService, initialSync: &mockSync.Sync{IsSyncing: false}, slashingPool: slashings.NewPool(), chain: chainService, clock: startup.NewClock(gt, vr), beaconDB: d, }, seenAttesterSlashingCache: make(map[uint64]bool), chainStarted: abool.New(), subHandler: newSubTopicHandler(), } topic := "/eth2/%x/attester_slashing" var wg sync.WaitGroup wg.Add(1) var err error p2pService.Digest, err = r.currentForkDigest() require.NoError(t, err) r.subscribe(topic, r.noopValidator, func(ctx context.Context, msg proto.Message) error { require.NoError(t, r.attesterSlashingSubscriber(ctx, msg)) wg.Done() return nil }, p2pService.Digest) beaconState, privKeys := util.DeterministicGenesisState(t, 64) chainService.State = beaconState r.markForChainStart() attesterSlashing, err := util.GenerateAttesterSlashingForValidator( beaconState, privKeys[1], 1, /* validator index */ ) require.NoError(t, err, "Error generating attester slashing") err = r.cfg.beaconDB.SaveState(ctx, beaconState, bytesutil.ToBytes32(attesterSlashing.Attestation_1.Data.BeaconBlockRoot)) require.NoError(t, err) p2pService.ReceivePubSub(topic, attesterSlashing) if util.WaitTimeout(&wg, time.Second) { t.Fatal("Did not receive PubSub in 1 second") } as := r.cfg.slashingPool.PendingAttesterSlashings(ctx, beaconState, false /*noLimit*/) assert.Equal(t, 1, len(as), "Expected attester slashing") } func TestSubscribe_ReceivesProposerSlashing(t *testing.T) { p2pService := p2ptest.NewTestP2P(t) ctx := context.Background() chainService := &mockChain.ChainService{ ValidatorsRoot: [32]byte{'A'}, Genesis: time.Now(), } d := db.SetupDB(t) r := Service{ ctx: ctx, cfg: &config{ p2p: p2pService, initialSync: &mockSync.Sync{IsSyncing: false}, slashingPool: slashings.NewPool(), chain: chainService, beaconDB: d, clock: startup.NewClock(chainService.Genesis, chainService.ValidatorsRoot), }, seenProposerSlashingCache: lruwrpr.New(10), chainStarted: abool.New(), subHandler: newSubTopicHandler(), } topic := "/eth2/%x/proposer_slashing" var wg sync.WaitGroup wg.Add(1) params.SetupTestConfigCleanup(t) params.OverrideBeaconConfig(params.MainnetConfig()) var err error p2pService.Digest, err = r.currentForkDigest() require.NoError(t, err) r.subscribe(topic, r.noopValidator, func(ctx context.Context, msg proto.Message) error { require.NoError(t, r.proposerSlashingSubscriber(ctx, msg)) wg.Done() return nil }, p2pService.Digest) beaconState, privKeys := util.DeterministicGenesisState(t, 64) chainService.State = beaconState r.markForChainStart() proposerSlashing, err := util.GenerateProposerSlashingForValidator( beaconState, privKeys[1], 1, /* validator index */ ) require.NoError(t, err, "Error generating proposer slashing") p2pService.ReceivePubSub(topic, proposerSlashing) if util.WaitTimeout(&wg, time.Second) { t.Fatal("Did not receive PubSub in 1 second") } ps := r.cfg.slashingPool.PendingProposerSlashings(ctx, beaconState, false /*noLimit*/) assert.Equal(t, 1, len(ps), "Expected proposer slashing") } func TestSubscribe_HandlesPanic(t *testing.T) { p := p2ptest.NewTestP2P(t) chain := &mockChain.ChainService{ Genesis: time.Now(), ValidatorsRoot: [32]byte{'A'}, } r := Service{ ctx: context.Background(), cfg: &config{ chain: chain, clock: startup.NewClock(chain.Genesis, chain.ValidatorsRoot), p2p: p, }, subHandler: newSubTopicHandler(), chainStarted: abool.New(), } var err error p.Digest, err = r.currentForkDigest() require.NoError(t, err) topic := p2p.GossipTypeMapping[reflect.TypeOf(&pb.SignedVoluntaryExit{})] var wg sync.WaitGroup wg.Add(1) r.subscribe(topic, r.noopValidator, func(_ context.Context, msg proto.Message) error { defer wg.Done() panic("bad") }, p.Digest) r.markForChainStart() p.ReceivePubSub(topic, &pb.SignedVoluntaryExit{Exit: &pb.VoluntaryExit{Epoch: 55}, Signature: make([]byte, fieldparams.BLSSignatureLength)}) if util.WaitTimeout(&wg, time.Second) { t.Fatal("Did not receive PubSub in 1 second") } } func TestRevalidateSubscription_CorrectlyFormatsTopic(t *testing.T) { p := p2ptest.NewTestP2P(t) hook := logTest.NewGlobal() chain := &mockChain.ChainService{ Genesis: time.Now(), ValidatorsRoot: [32]byte{'A'}, } r := Service{ ctx: context.Background(), cfg: &config{ chain: chain, clock: startup.NewClock(chain.Genesis, chain.ValidatorsRoot), p2p: p, }, chainStarted: abool.New(), subHandler: newSubTopicHandler(), } digest, err := r.currentForkDigest() require.NoError(t, err) subscriptions := make(map[uint64]*pubsub.Subscription, params.BeaconConfig().MaxCommitteesPerSlot) defaultTopic := "/eth2/testing/%#x/committee%d" // committee index 1 fullTopic := fmt.Sprintf(defaultTopic, digest, 1) + r.cfg.p2p.Encoding().ProtocolSuffix() _, topVal := r.wrapAndReportValidation(fullTopic, r.noopValidator) require.NoError(t, r.cfg.p2p.PubSub().RegisterTopicValidator(fullTopic, topVal)) subscriptions[1], err = r.cfg.p2p.SubscribeToTopic(fullTopic) require.NoError(t, err) // committee index 2 fullTopic = fmt.Sprintf(defaultTopic, digest, 2) + r.cfg.p2p.Encoding().ProtocolSuffix() _, topVal = r.wrapAndReportValidation(fullTopic, r.noopValidator) err = r.cfg.p2p.PubSub().RegisterTopicValidator(fullTopic, topVal) require.NoError(t, err) subscriptions[2], err = r.cfg.p2p.SubscribeToTopic(fullTopic) require.NoError(t, err) r.reValidateSubscriptions(subscriptions, []uint64{2}, defaultTopic, digest) require.LogsDoNotContain(t, hook, "Could not unregister topic validator") } func TestStaticSubnets(t *testing.T) { p := p2ptest.NewTestP2P(t) ctx, cancel := context.WithCancel(context.Background()) chain := &mockChain.ChainService{ Genesis: time.Now(), ValidatorsRoot: [32]byte{'A'}, } r := Service{ ctx: ctx, cfg: &config{ chain: chain, clock: startup.NewClock(chain.Genesis, chain.ValidatorsRoot), p2p: p, }, chainStarted: abool.New(), subHandler: newSubTopicHandler(), } defaultTopic := "/eth2/%x/beacon_attestation_%d" d, err := r.currentForkDigest() assert.NoError(t, err) r.subscribeStaticWithSubnets(defaultTopic, r.noopValidator, func(_ context.Context, msg proto.Message) error { // no-op return nil }, d, params.BeaconConfig().AttestationSubnetCount) topics := r.cfg.p2p.PubSub().GetTopics() if uint64(len(topics)) != params.BeaconConfig().AttestationSubnetCount { t.Errorf("Wanted the number of subnet topics registered to be %d but got %d", params.BeaconConfig().AttestationSubnetCount, len(topics)) } cancel() } func Test_wrapAndReportValidation(t *testing.T) { mChain := &mockChain.ChainService{ Genesis: time.Now(), ValidatorsRoot: [32]byte{0x01}, } fd, err := forks.CreateForkDigest(mChain.GenesisTime(), mChain.ValidatorsRoot[:]) assert.NoError(t, err) mockTopic := fmt.Sprintf(p2p.BlockSubnetTopicFormat, fd) + encoder.SszNetworkEncoder{}.ProtocolSuffix() type args struct { topic string v wrappedVal chainstarted bool pid peer.ID msg *pubsub.Message } tests := []struct { name string args args want pubsub.ValidationResult }{ { name: "validator Before chainstart", args: args{ topic: "foo", v: func(ctx context.Context, id peer.ID, message *pubsub.Message) (pubsub.ValidationResult, error) { return pubsub.ValidationAccept, nil }, msg: &pubsub.Message{ Message: &pubsubpb.Message{ Topic: func() *string { s := "foo" return &s }(), }, }, chainstarted: false, }, want: pubsub.ValidationIgnore, }, { name: "validator panicked", args: args{ topic: "foo", v: func(ctx context.Context, id peer.ID, message *pubsub.Message) (pubsub.ValidationResult, error) { panic("oh no!") }, chainstarted: true, msg: &pubsub.Message{ Message: &pubsubpb.Message{ Topic: func() *string { s := "foo" return &s }(), }, }, }, want: pubsub.ValidationIgnore, }, { name: "validator OK", args: args{ topic: mockTopic, v: func(ctx context.Context, id peer.ID, message *pubsub.Message) (pubsub.ValidationResult, error) { return pubsub.ValidationAccept, nil }, chainstarted: true, msg: &pubsub.Message{ Message: &pubsubpb.Message{ Topic: func() *string { s := mockTopic return &s }(), }, }, }, want: pubsub.ValidationAccept, }, { name: "nil topic", args: args{ topic: "foo", v: func(ctx context.Context, id peer.ID, message *pubsub.Message) (pubsub.ValidationResult, error) { return pubsub.ValidationAccept, nil }, msg: &pubsub.Message{ Message: &pubsubpb.Message{ Topic: nil, }, }, }, want: pubsub.ValidationReject, }, } for _, tt := range tests { t.Run(tt.name, func(t *testing.T) { chainStarted := abool.New() chainStarted.SetTo(tt.args.chainstarted) s := &Service{ chainStarted: chainStarted, cfg: &config{ chain: mChain, clock: startup.NewClock(mChain.Genesis, mChain.ValidatorsRoot), }, subHandler: newSubTopicHandler(), } _, v := s.wrapAndReportValidation(tt.args.topic, tt.args.v) got := v(context.Background(), tt.args.pid, tt.args.msg) if got != tt.want { t.Errorf("wrapAndReportValidation() got = %v, want %v", got, tt.want) } }) } } func TestFilterSubnetPeers(t *testing.T) { params.SetupTestConfigCleanup(t) cfg := params.MainnetConfig().Copy() cfg.SecondsPerSlot = 1 params.OverrideBeaconConfig(cfg) gFlags := new(flags.GlobalFlags) gFlags.MinimumPeersPerSubnet = 4 flags.Init(gFlags) // Reset config. defer flags.Init(new(flags.GlobalFlags)) p := p2ptest.NewTestP2P(t) ctx, cancel := context.WithCancel(context.Background()) defer cancel() currSlot := primitives.Slot(100) gt := time.Now() genPlus100 := func() time.Time { return gt.Add(time.Second * time.Duration(uint64(currSlot)*params.BeaconConfig().SecondsPerSlot)) } chain := &mockChain.ChainService{ Genesis: gt, ValidatorsRoot: [32]byte{'A'}, FinalizedRoots: map[[32]byte]bool{ {}: true, }, } clock := startup.NewClock(chain.Genesis, chain.ValidatorsRoot, startup.WithNower(genPlus100)) require.Equal(t, currSlot, clock.CurrentSlot()) r := Service{ ctx: ctx, cfg: &config{ chain: chain, clock: clock, p2p: p, }, chainStarted: abool.New(), subHandler: newSubTopicHandler(), } // Empty cache at the end of the test. defer cache.SubnetIDs.EmptyAllCaches() digest, err := r.currentForkDigest() assert.NoError(t, err) defaultTopic := "/eth2/%x/beacon_attestation_%d" + r.cfg.p2p.Encoding().ProtocolSuffix() subnet10 := r.addDigestAndIndexToTopic(defaultTopic, digest, 10) cache.SubnetIDs.AddAggregatorSubnetID(currSlot, 10) subnet20 := r.addDigestAndIndexToTopic(defaultTopic, digest, 20) cache.SubnetIDs.AddAttesterSubnetID(currSlot, 20) p1 := createPeer(t, subnet10) p2 := createPeer(t, subnet10, subnet20) p3 := createPeer(t) // Connect to all // peers. p.Connect(p1) p.Connect(p2) p.Connect(p3) // Sleep a while to allow peers to connect. time.Sleep(100 * time.Millisecond) wantedPeers := []peer.ID{p1.PeerID(), p2.PeerID(), p3.PeerID()} // Expect Peer 3 to be marked as suitable. recPeers := r.filterNeededPeers(wantedPeers) assert.DeepEqual(t, []peer.ID{p3.PeerID()}, recPeers) // Try with only peers from subnet 20. wantedPeers = []peer.ID{p2.BHost.ID()} // Connect an excess amount of peers in the particular subnet. for i := 1; i <= flags.Get().MinimumPeersPerSubnet; i++ { nPeer := createPeer(t, subnet20) p.Connect(nPeer) wantedPeers = append(wantedPeers, nPeer.BHost.ID()) time.Sleep(100 * time.Millisecond) } recPeers = r.filterNeededPeers(wantedPeers) assert.Equal(t, 1, len(recPeers), "expected at least 1 suitable peer to prune") } func TestSubscribeWithSyncSubnets_StaticOK(t *testing.T) { params.SetupTestConfigCleanup(t) cfg := params.MainnetTestConfig().Copy() cfg.SecondsPerSlot = 1 params.OverrideBeaconConfig(cfg) p := p2ptest.NewTestP2P(t) ctx, cancel := context.WithCancel(context.Background()) chain := &mockChain.ChainService{ Genesis: time.Now(), ValidatorsRoot: [32]byte{'A'}, } r := Service{ ctx: ctx, cfg: &config{ chain: chain, clock: startup.NewClock(chain.Genesis, chain.ValidatorsRoot), p2p: p, }, chainStarted: abool.New(), subHandler: newSubTopicHandler(), } // Empty cache at the end of the test. defer cache.SyncSubnetIDs.EmptyAllCaches() digest, err := r.currentForkDigest() assert.NoError(t, err) r.subscribeStaticWithSyncSubnets(p2p.SyncCommitteeSubnetTopicFormat, nil, nil, digest) assert.Equal(t, int(params.BeaconConfig().SyncCommitteeSubnetCount), len(r.cfg.p2p.PubSub().GetTopics())) cancel() } func TestSubscribeWithSyncSubnets_DynamicOK(t *testing.T) { params.SetupTestConfigCleanup(t) cfg := params.MainnetConfig().Copy() cfg.SecondsPerSlot = 1 params.OverrideBeaconConfig(cfg) p := p2ptest.NewTestP2P(t) ctx, cancel := context.WithCancel(context.Background()) gt := time.Now() vr := [32]byte{'A'} r := Service{ ctx: ctx, cfg: &config{ chain: &mockChain.ChainService{ Genesis: gt, ValidatorsRoot: vr, }, p2p: p, clock: startup.NewClock(gt, vr), }, chainStarted: abool.New(), subHandler: newSubTopicHandler(), } // Empty cache at the end of the test. defer cache.SyncSubnetIDs.EmptyAllCaches() slot := r.cfg.clock.CurrentSlot() currEpoch := slots.ToEpoch(slot) cache.SyncSubnetIDs.AddSyncCommitteeSubnets([]byte("pubkey"), currEpoch, []uint64{0, 1}, 10*time.Second) digest, err := r.currentForkDigest() assert.NoError(t, err) r.subscribeDynamicWithSyncSubnets(p2p.SyncCommitteeSubnetTopicFormat, nil, nil, digest) time.Sleep(2 * time.Second) assert.Equal(t, 2, len(r.cfg.p2p.PubSub().GetTopics())) topicMap := map[string]bool{} for _, t := range r.cfg.p2p.PubSub().GetTopics() { topicMap[t] = true } firstSub := fmt.Sprintf(p2p.SyncCommitteeSubnetTopicFormat, digest, 0) + r.cfg.p2p.Encoding().ProtocolSuffix() assert.Equal(t, true, topicMap[firstSub]) secondSub := fmt.Sprintf(p2p.SyncCommitteeSubnetTopicFormat, digest, 1) + r.cfg.p2p.Encoding().ProtocolSuffix() assert.Equal(t, true, topicMap[secondSub]) cancel() } func TestSubscribeWithSyncSubnets_StaticSwitchFork(t *testing.T) { p := p2ptest.NewTestP2P(t) params.SetupTestConfigCleanup(t) cfg := params.BeaconConfig() cfg.AltairForkEpoch = 1 cfg.SecondsPerSlot = 1 params.OverrideBeaconConfig(cfg) params.BeaconConfig().InitializeForkSchedule() ctx, cancel := context.WithCancel(context.Background()) currSlot := primitives.Slot(100) chain := &mockChain.ChainService{ Genesis: time.Now().Add(-time.Duration(uint64(params.BeaconConfig().SlotsPerEpoch)*params.BeaconConfig().SecondsPerSlot) * time.Second), ValidatorsRoot: [32]byte{'A'}, Slot: &currSlot, } r := Service{ ctx: ctx, cfg: &config{ chain: chain, clock: startup.NewClock(chain.Genesis, chain.ValidatorsRoot), p2p: p, }, chainStarted: abool.New(), subHandler: newSubTopicHandler(), } // Empty cache at the end of the test. defer cache.SyncSubnetIDs.EmptyAllCaches() genRoot := r.cfg.clock.GenesisValidatorsRoot() digest, err := signing.ComputeForkDigest(params.BeaconConfig().GenesisForkVersion, genRoot[:]) assert.NoError(t, err) r.subscribeStaticWithSyncSubnets(p2p.SyncCommitteeSubnetTopicFormat, nil, nil, digest) assert.Equal(t, int(params.BeaconConfig().SyncCommitteeSubnetCount), len(r.cfg.p2p.PubSub().GetTopics())) // Expect that all old topics will be unsubscribed. time.Sleep(2 * time.Second) assert.Equal(t, 0, len(r.cfg.p2p.PubSub().GetTopics())) cancel() } func TestSubscribeWithSyncSubnets_DynamicSwitchFork(t *testing.T) { params.SetupTestConfigCleanup(t) p := p2ptest.NewTestP2P(t) cfg := params.BeaconConfig().Copy() cfg.AltairForkEpoch = 1 cfg.SecondsPerSlot = 1 cfg.SlotsPerEpoch = 4 params.OverrideBeaconConfig(cfg) params.BeaconConfig().InitializeForkSchedule() ctx, cancel := context.WithCancel(context.Background()) currSlot := primitives.Slot(100) gt := time.Now().Add(-time.Duration(params.BeaconConfig().SecondsPerSlot) * time.Second) vr := [32]byte{'A'} r := Service{ ctx: ctx, cfg: &config{ chain: &mockChain.ChainService{ Genesis: gt, ValidatorsRoot: vr, Slot: &currSlot, }, clock: startup.NewClock(gt, vr), p2p: p, }, chainStarted: abool.New(), subHandler: newSubTopicHandler(), } // Empty cache at the end of the test. defer cache.SyncSubnetIDs.EmptyAllCaches() cache.SyncSubnetIDs.AddSyncCommitteeSubnets([]byte("pubkey"), 0, []uint64{0, 1}, 10*time.Second) genRoot := r.cfg.clock.GenesisValidatorsRoot() digest, err := signing.ComputeForkDigest(params.BeaconConfig().GenesisForkVersion, genRoot[:]) assert.NoError(t, err) r.subscribeDynamicWithSyncSubnets(p2p.SyncCommitteeSubnetTopicFormat, nil, nil, digest) time.Sleep(2 * time.Second) assert.Equal(t, 2, len(r.cfg.p2p.PubSub().GetTopics())) topicMap := map[string]bool{} for _, t := range r.cfg.p2p.PubSub().GetTopics() { topicMap[t] = true } firstSub := fmt.Sprintf(p2p.SyncCommitteeSubnetTopicFormat, digest, 0) + r.cfg.p2p.Encoding().ProtocolSuffix() assert.Equal(t, true, topicMap[firstSub]) secondSub := fmt.Sprintf(p2p.SyncCommitteeSubnetTopicFormat, digest, 1) + r.cfg.p2p.Encoding().ProtocolSuffix() assert.Equal(t, true, topicMap[secondSub]) // Expect that all old topics will be unsubscribed. time.Sleep(2 * time.Second) assert.Equal(t, 0, len(r.cfg.p2p.PubSub().GetTopics())) cancel() } func TestIsDigestValid(t *testing.T) { genRoot := [32]byte{'A'} digest, err := signing.ComputeForkDigest(params.BeaconConfig().GenesisForkVersion, genRoot[:]) assert.NoError(t, err) valid, err := isDigestValid(digest, time.Now().Add(-100*time.Second), genRoot) assert.NoError(t, err) assert.Equal(t, true, valid) // Compute future fork digest that will be invalid currently. digest, err = signing.ComputeForkDigest(params.BeaconConfig().AltairForkVersion, genRoot[:]) assert.NoError(t, err) valid, err = isDigestValid(digest, time.Now().Add(-100*time.Second), genRoot) assert.NoError(t, err) assert.Equal(t, false, valid) } // Create peer and register them to provided topics. func createPeer(t *testing.T, topics ...string) *p2ptest.TestP2P { p := p2ptest.NewTestP2P(t) for _, tp := range topics { jTop, err := p.PubSub().Join(tp) if err != nil { t.Fatal(err) } _, err = jTop.Subscribe() if err != nil { t.Fatal(err) } } return p }