// Package endtoend performs full a end-to-end test for Prysm, // including spinning up an ETH1 dev chain, sending deposits to the deposit // contract, and making sure the beacon node and validators are running and // performing properly for a few epochs. package endtoend import ( "context" "fmt" "os" "path" "testing" "time" ptypes "github.com/gogo/protobuf/types" eth "github.com/prysmaticlabs/ethereumapis/eth/v1alpha1" "github.com/prysmaticlabs/prysm/beacon-chain/core/state" "github.com/prysmaticlabs/prysm/endtoend/components" ev "github.com/prysmaticlabs/prysm/endtoend/evaluators" "github.com/prysmaticlabs/prysm/endtoend/helpers" e2e "github.com/prysmaticlabs/prysm/endtoend/params" "github.com/prysmaticlabs/prysm/endtoend/types" "github.com/prysmaticlabs/prysm/shared/params" "github.com/prysmaticlabs/prysm/shared/slotutil" "google.golang.org/grpc" ) func init() { state.SkipSlotCache.Disable() } func runEndToEndTest(t *testing.T, config *types.E2EConfig) { t.Logf("Shard index: %d\n", e2e.TestParams.TestShardIndex) t.Logf("Starting time: %s\n", time.Now().String()) t.Logf("Log Path: %s\n\n", e2e.TestParams.LogPath) keystorePath, eth1PID := components.StartEth1Node(t) bootnodeENR, bootnodePID := components.StartBootnode(t) bProcessIDs := components.StartBeaconNodes(t, config, bootnodeENR) valProcessIDs := components.StartValidatorClients(t, config, keystorePath) processIDs := append(valProcessIDs, bProcessIDs...) processIDs = append(processIDs, []int{eth1PID, bootnodePID}...) defer helpers.LogOutput(t, config) defer helpers.KillProcesses(t, processIDs) // Sleep depending on the count of validators, as generating the genesis state could take some time. time.Sleep(time.Duration(params.BeaconConfig().MinGenesisDelay) * time.Second) beaconLogFile, err := os.Open(path.Join(e2e.TestParams.LogPath, fmt.Sprintf(e2e.BeaconNodeLogFileName, 0))) if err != nil { t.Fatal(err) } t.Run("chain started", func(t *testing.T) { if err := helpers.WaitForTextInFile(beaconLogFile, "Chain started within the last epoch"); err != nil { t.Fatalf("failed to find chain start in logs, this means the chain did not start: %v", err) } }) // Failing early in case chain doesn't start. if t.Failed() { return } if config.TestSlasher { slasherPIDs := components.StartSlashers(t) defer helpers.KillProcesses(t, slasherPIDs) } if config.TestDeposits { valCount := int(params.BeaconConfig().MinGenesisActiveValidatorCount) / e2e.TestParams.BeaconNodeCount valPid := components.StartNewValidatorClient(t, config, valCount, e2e.TestParams.BeaconNodeCount) defer helpers.KillProcesses(t, []int{valPid}) components.SendAndMineDeposits(t, keystorePath, valCount, int(params.BeaconConfig().MinGenesisActiveValidatorCount)) } conns := make([]*grpc.ClientConn, e2e.TestParams.BeaconNodeCount) for i := 0; i < len(conns); i++ { conn, err := grpc.Dial(fmt.Sprintf("127.0.0.1:%d", e2e.TestParams.BeaconNodeRPCPort+i), grpc.WithInsecure()) if err != nil { t.Fatalf("Failed to dial: %v", err) } conns[i] = conn defer func() { if err := conn.Close(); err != nil { t.Log(err) } }() } nodeClient := eth.NewNodeClient(conns[0]) genesis, err := nodeClient.GetGenesis(context.Background(), &ptypes.Empty{}) if err != nil { t.Fatal(err) } epochSeconds := params.BeaconConfig().SecondsPerSlot * params.BeaconConfig().SlotsPerEpoch epochSecondsHalf := time.Duration(int64(epochSeconds*1000)/2) * time.Millisecond // Adding a half slot here to ensure the requests are in the middle of an epoch. middleOfEpoch := epochSecondsHalf + slotutil.DivideSlotBy(2 /* half a slot */) genesisTime := time.Unix(genesis.GenesisTime.Seconds, 0) // Offsetting the ticker from genesis so it ticks in the middle of an epoch, in order to keep results consistent. tickingStartTime := genesisTime.Add(middleOfEpoch) ticker := helpers.GetEpochTicker(tickingStartTime, epochSeconds) for currentEpoch := range ticker.C() { for _, evaluator := range config.Evaluators { // Only run if the policy says so. if !evaluator.Policy(currentEpoch) { continue } t.Run(fmt.Sprintf(evaluator.Name, currentEpoch), func(t *testing.T) { if err := evaluator.Evaluation(conns...); err != nil { t.Errorf("evaluation failed for epoch %d: %v", currentEpoch, err) } }) } if t.Failed() || currentEpoch >= config.EpochsToRun-1 { ticker.Done() if t.Failed() { return } break } } if !config.TestSync { return } index := e2e.TestParams.BeaconNodeCount processID := components.StartNewBeaconNode(t, config, index, bootnodeENR) syncConn, err := grpc.Dial(fmt.Sprintf("127.0.0.1:%d", e2e.TestParams.BeaconNodeRPCPort+index), grpc.WithInsecure()) if err != nil { t.Fatalf("Failed to dial: %v", err) } conns = append(conns, syncConn) // Sleep a second for every 8 blocks that need to be synced for the newly started node. extraSecondsToSync := (config.EpochsToRun)*epochSeconds + (params.BeaconConfig().SlotsPerEpoch / 8 * config.EpochsToRun) waitForSync := tickingStartTime.Add(time.Duration(extraSecondsToSync) * time.Second) time.Sleep(time.Until(waitForSync)) syncLogFile, err := os.Open(path.Join(e2e.TestParams.LogPath, fmt.Sprintf(e2e.BeaconNodeLogFileName, index))) if err != nil { t.Fatal(err) } defer helpers.LogErrorOutput(t, syncLogFile, "beacon chain node", index) defer helpers.KillProcesses(t, []int{processID}) t.Run("sync completed", func(t *testing.T) { if err := helpers.WaitForTextInFile(syncLogFile, "Synced up to"); err != nil { t.Errorf("Failed to sync: %v", err) } }) if t.Failed() { return } syncEvaluators := []types.Evaluator{ev.FinishedSyncing, ev.AllNodesHaveSameHead} for _, evaluator := range syncEvaluators { t.Run(evaluator.Name, func(t *testing.T) { if err := evaluator.Evaluation(conns...); err != nil { t.Errorf("evaluation failed for sync node: %v", err) } }) } }