mirror of
https://gitlab.com/pulsechaincom/erigon-pulse.git
synced 2024-12-24 12:37:16 +00:00
2cdf6ee7e0
This PR enables the indexers to work in light client mode by downloading a part of these tries (the Merkle proofs of the last values of the last known section) in order to be able to add new values and recalculate subsequent hashes. It also adds CHT data to NodeInfo.
240 lines
7.2 KiB
Go
240 lines
7.2 KiB
Go
// Copyright 2017 The go-ethereum Authors
|
|
// This file is part of the go-ethereum library.
|
|
//
|
|
// The go-ethereum library is free software: you can redistribute it and/or modify
|
|
// it under the terms of the GNU Lesser General Public License as published by
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
// (at your option) any later version.
|
|
//
|
|
// The go-ethereum library is distributed in the hope that it will be useful,
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
// GNU Lesser General Public License for more details.
|
|
//
|
|
// You should have received a copy of the GNU Lesser General Public License
|
|
// along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
package core
|
|
|
|
import (
|
|
"context"
|
|
"fmt"
|
|
"math/big"
|
|
"math/rand"
|
|
"testing"
|
|
"time"
|
|
|
|
"github.com/ethereum/go-ethereum/common"
|
|
"github.com/ethereum/go-ethereum/core/rawdb"
|
|
"github.com/ethereum/go-ethereum/core/types"
|
|
"github.com/ethereum/go-ethereum/ethdb"
|
|
)
|
|
|
|
// Runs multiple tests with randomized parameters.
|
|
func TestChainIndexerSingle(t *testing.T) {
|
|
for i := 0; i < 10; i++ {
|
|
testChainIndexer(t, 1)
|
|
}
|
|
}
|
|
|
|
// Runs multiple tests with randomized parameters and different number of
|
|
// chain backends.
|
|
func TestChainIndexerWithChildren(t *testing.T) {
|
|
for i := 2; i < 8; i++ {
|
|
testChainIndexer(t, i)
|
|
}
|
|
}
|
|
|
|
// testChainIndexer runs a test with either a single chain indexer or a chain of
|
|
// multiple backends. The section size and required confirmation count parameters
|
|
// are randomized.
|
|
func testChainIndexer(t *testing.T, count int) {
|
|
db := ethdb.NewMemDatabase()
|
|
defer db.Close()
|
|
|
|
// Create a chain of indexers and ensure they all report empty
|
|
backends := make([]*testChainIndexBackend, count)
|
|
for i := 0; i < count; i++ {
|
|
var (
|
|
sectionSize = uint64(rand.Intn(100) + 1)
|
|
confirmsReq = uint64(rand.Intn(10))
|
|
)
|
|
backends[i] = &testChainIndexBackend{t: t, processCh: make(chan uint64)}
|
|
backends[i].indexer = NewChainIndexer(db, ethdb.NewTable(db, string([]byte{byte(i)})), backends[i], sectionSize, confirmsReq, 0, fmt.Sprintf("indexer-%d", i))
|
|
|
|
if sections, _, _ := backends[i].indexer.Sections(); sections != 0 {
|
|
t.Fatalf("Canonical section count mismatch: have %v, want %v", sections, 0)
|
|
}
|
|
if i > 0 {
|
|
backends[i-1].indexer.AddChildIndexer(backends[i].indexer)
|
|
}
|
|
}
|
|
defer backends[0].indexer.Close() // parent indexer shuts down children
|
|
// notify pings the root indexer about a new head or reorg, then expect
|
|
// processed blocks if a section is processable
|
|
notify := func(headNum, failNum uint64, reorg bool) {
|
|
backends[0].indexer.newHead(headNum, reorg)
|
|
if reorg {
|
|
for _, backend := range backends {
|
|
headNum = backend.reorg(headNum)
|
|
backend.assertSections()
|
|
}
|
|
return
|
|
}
|
|
var cascade bool
|
|
for _, backend := range backends {
|
|
headNum, cascade = backend.assertBlocks(headNum, failNum)
|
|
if !cascade {
|
|
break
|
|
}
|
|
backend.assertSections()
|
|
}
|
|
}
|
|
// inject inserts a new random canonical header into the database directly
|
|
inject := func(number uint64) {
|
|
header := &types.Header{Number: big.NewInt(int64(number)), Extra: big.NewInt(rand.Int63()).Bytes()}
|
|
if number > 0 {
|
|
header.ParentHash = rawdb.ReadCanonicalHash(db, number-1)
|
|
}
|
|
rawdb.WriteHeader(db, header)
|
|
rawdb.WriteCanonicalHash(db, header.Hash(), number)
|
|
}
|
|
// Start indexer with an already existing chain
|
|
for i := uint64(0); i <= 100; i++ {
|
|
inject(i)
|
|
}
|
|
notify(100, 100, false)
|
|
|
|
// Add new blocks one by one
|
|
for i := uint64(101); i <= 1000; i++ {
|
|
inject(i)
|
|
notify(i, i, false)
|
|
}
|
|
// Do a reorg
|
|
notify(500, 500, true)
|
|
|
|
// Create new fork
|
|
for i := uint64(501); i <= 1000; i++ {
|
|
inject(i)
|
|
notify(i, i, false)
|
|
}
|
|
for i := uint64(1001); i <= 1500; i++ {
|
|
inject(i)
|
|
}
|
|
// Failed processing scenario where less blocks are available than notified
|
|
notify(2000, 1500, false)
|
|
|
|
// Notify about a reorg (which could have caused the missing blocks if happened during processing)
|
|
notify(1500, 1500, true)
|
|
|
|
// Create new fork
|
|
for i := uint64(1501); i <= 2000; i++ {
|
|
inject(i)
|
|
notify(i, i, false)
|
|
}
|
|
}
|
|
|
|
// testChainIndexBackend implements ChainIndexerBackend
|
|
type testChainIndexBackend struct {
|
|
t *testing.T
|
|
indexer *ChainIndexer
|
|
section, headerCnt, stored uint64
|
|
processCh chan uint64
|
|
}
|
|
|
|
// assertSections verifies if a chain indexer has the correct number of section.
|
|
func (b *testChainIndexBackend) assertSections() {
|
|
// Keep trying for 3 seconds if it does not match
|
|
var sections uint64
|
|
for i := 0; i < 300; i++ {
|
|
sections, _, _ = b.indexer.Sections()
|
|
if sections == b.stored {
|
|
return
|
|
}
|
|
time.Sleep(10 * time.Millisecond)
|
|
}
|
|
b.t.Fatalf("Canonical section count mismatch: have %v, want %v", sections, b.stored)
|
|
}
|
|
|
|
// assertBlocks expects processing calls after new blocks have arrived. If the
|
|
// failNum < headNum then we are simulating a scenario where a reorg has happened
|
|
// after the processing has started and the processing of a section fails.
|
|
func (b *testChainIndexBackend) assertBlocks(headNum, failNum uint64) (uint64, bool) {
|
|
var sections uint64
|
|
if headNum >= b.indexer.confirmsReq {
|
|
sections = (headNum + 1 - b.indexer.confirmsReq) / b.indexer.sectionSize
|
|
if sections > b.stored {
|
|
// expect processed blocks
|
|
for expectd := b.stored * b.indexer.sectionSize; expectd < sections*b.indexer.sectionSize; expectd++ {
|
|
if expectd > failNum {
|
|
// rolled back after processing started, no more process calls expected
|
|
// wait until updating is done to make sure that processing actually fails
|
|
var updating bool
|
|
for i := 0; i < 300; i++ {
|
|
b.indexer.lock.Lock()
|
|
updating = b.indexer.knownSections > b.indexer.storedSections
|
|
b.indexer.lock.Unlock()
|
|
if !updating {
|
|
break
|
|
}
|
|
time.Sleep(10 * time.Millisecond)
|
|
}
|
|
if updating {
|
|
b.t.Fatalf("update did not finish")
|
|
}
|
|
sections = expectd / b.indexer.sectionSize
|
|
break
|
|
}
|
|
select {
|
|
case <-time.After(10 * time.Second):
|
|
b.t.Fatalf("Expected processed block #%d, got nothing", expectd)
|
|
case processed := <-b.processCh:
|
|
if processed != expectd {
|
|
b.t.Errorf("Expected processed block #%d, got #%d", expectd, processed)
|
|
}
|
|
}
|
|
}
|
|
b.stored = sections
|
|
}
|
|
}
|
|
if b.stored == 0 {
|
|
return 0, false
|
|
}
|
|
return b.stored*b.indexer.sectionSize - 1, true
|
|
}
|
|
|
|
func (b *testChainIndexBackend) reorg(headNum uint64) uint64 {
|
|
firstChanged := headNum / b.indexer.sectionSize
|
|
if firstChanged < b.stored {
|
|
b.stored = firstChanged
|
|
}
|
|
return b.stored * b.indexer.sectionSize
|
|
}
|
|
|
|
func (b *testChainIndexBackend) Reset(ctx context.Context, section uint64, prevHead common.Hash) error {
|
|
b.section = section
|
|
b.headerCnt = 0
|
|
return nil
|
|
}
|
|
|
|
func (b *testChainIndexBackend) Process(ctx context.Context, header *types.Header) error {
|
|
b.headerCnt++
|
|
if b.headerCnt > b.indexer.sectionSize {
|
|
b.t.Error("Processing too many headers")
|
|
}
|
|
//t.processCh <- header.Number.Uint64()
|
|
select {
|
|
case <-time.After(10 * time.Second):
|
|
b.t.Fatal("Unexpected call to Process")
|
|
case b.processCh <- header.Number.Uint64():
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (b *testChainIndexBackend) Commit() error {
|
|
if b.headerCnt != b.indexer.sectionSize {
|
|
b.t.Error("Not enough headers processed")
|
|
}
|
|
return nil
|
|
}
|