mirror of
https://gitlab.com/pulsechaincom/go-pulse.git
synced 2024-12-22 11:31:02 +00:00
6eef141aef
This change introduces garbage collection for the light client. Historical chain data is deleted periodically. If you want to disable the GC, use the --light.nopruning flag.
161 lines
6.9 KiB
Go
161 lines
6.9 KiB
Go
// Copyright 2016 The go-ethereum Authors
|
|
// This file is part of the go-ethereum library.
|
|
//
|
|
// The go-ethereum library is free software: you can redistribute it and/or modify
|
|
// it under the terms of the GNU Lesser General Public License as published by
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
// (at your option) any later version.
|
|
//
|
|
// The go-ethereum library is distributed in the hope that it will be useful,
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
// GNU Lesser General Public License for more details.
|
|
//
|
|
// You should have received a copy of the GNU Lesser General Public License
|
|
// along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
package core
|
|
|
|
import (
|
|
"math/big"
|
|
"testing"
|
|
|
|
"github.com/ethereum/go-ethereum/consensus/ethash"
|
|
"github.com/ethereum/go-ethereum/core/rawdb"
|
|
"github.com/ethereum/go-ethereum/core/vm"
|
|
"github.com/ethereum/go-ethereum/params"
|
|
)
|
|
|
|
// Tests that DAO-fork enabled clients can properly filter out fork-commencing
|
|
// blocks based on their extradata fields.
|
|
func TestDAOForkRangeExtradata(t *testing.T) {
|
|
forkBlock := big.NewInt(32)
|
|
|
|
// Generate a common prefix for both pro-forkers and non-forkers
|
|
db := rawdb.NewMemoryDatabase()
|
|
gspec := new(Genesis)
|
|
genesis := gspec.MustCommit(db)
|
|
prefix, _ := GenerateChain(params.TestChainConfig, genesis, ethash.NewFaker(), db, int(forkBlock.Int64()-1), func(i int, gen *BlockGen) {})
|
|
|
|
// Create the concurrent, conflicting two nodes
|
|
proDb := rawdb.NewMemoryDatabase()
|
|
gspec.MustCommit(proDb)
|
|
|
|
proConf := *params.TestChainConfig
|
|
proConf.DAOForkBlock = forkBlock
|
|
proConf.DAOForkSupport = true
|
|
|
|
proBc, _ := NewBlockChain(proDb, nil, &proConf, ethash.NewFaker(), vm.Config{}, nil, nil)
|
|
defer proBc.Stop()
|
|
|
|
conDb := rawdb.NewMemoryDatabase()
|
|
gspec.MustCommit(conDb)
|
|
|
|
conConf := *params.TestChainConfig
|
|
conConf.DAOForkBlock = forkBlock
|
|
conConf.DAOForkSupport = false
|
|
|
|
conBc, _ := NewBlockChain(conDb, nil, &conConf, ethash.NewFaker(), vm.Config{}, nil, nil)
|
|
defer conBc.Stop()
|
|
|
|
if _, err := proBc.InsertChain(prefix); err != nil {
|
|
t.Fatalf("pro-fork: failed to import chain prefix: %v", err)
|
|
}
|
|
if _, err := conBc.InsertChain(prefix); err != nil {
|
|
t.Fatalf("con-fork: failed to import chain prefix: %v", err)
|
|
}
|
|
// Try to expand both pro-fork and non-fork chains iteratively with other camp's blocks
|
|
for i := int64(0); i < params.DAOForkExtraRange.Int64(); i++ {
|
|
// Create a pro-fork block, and try to feed into the no-fork chain
|
|
db = rawdb.NewMemoryDatabase()
|
|
gspec.MustCommit(db)
|
|
bc, _ := NewBlockChain(db, nil, &conConf, ethash.NewFaker(), vm.Config{}, nil, nil)
|
|
defer bc.Stop()
|
|
|
|
blocks := conBc.GetBlocksFromHash(conBc.CurrentBlock().Hash(), int(conBc.CurrentBlock().NumberU64()))
|
|
for j := 0; j < len(blocks)/2; j++ {
|
|
blocks[j], blocks[len(blocks)-1-j] = blocks[len(blocks)-1-j], blocks[j]
|
|
}
|
|
if _, err := bc.InsertChain(blocks); err != nil {
|
|
t.Fatalf("failed to import contra-fork chain for expansion: %v", err)
|
|
}
|
|
if err := bc.stateCache.TrieDB().Commit(bc.CurrentHeader().Root, true, nil); err != nil {
|
|
t.Fatalf("failed to commit contra-fork head for expansion: %v", err)
|
|
}
|
|
blocks, _ = GenerateChain(&proConf, conBc.CurrentBlock(), ethash.NewFaker(), db, 1, func(i int, gen *BlockGen) {})
|
|
if _, err := conBc.InsertChain(blocks); err == nil {
|
|
t.Fatalf("contra-fork chain accepted pro-fork block: %v", blocks[0])
|
|
}
|
|
// Create a proper no-fork block for the contra-forker
|
|
blocks, _ = GenerateChain(&conConf, conBc.CurrentBlock(), ethash.NewFaker(), db, 1, func(i int, gen *BlockGen) {})
|
|
if _, err := conBc.InsertChain(blocks); err != nil {
|
|
t.Fatalf("contra-fork chain didn't accepted no-fork block: %v", err)
|
|
}
|
|
// Create a no-fork block, and try to feed into the pro-fork chain
|
|
db = rawdb.NewMemoryDatabase()
|
|
gspec.MustCommit(db)
|
|
bc, _ = NewBlockChain(db, nil, &proConf, ethash.NewFaker(), vm.Config{}, nil, nil)
|
|
defer bc.Stop()
|
|
|
|
blocks = proBc.GetBlocksFromHash(proBc.CurrentBlock().Hash(), int(proBc.CurrentBlock().NumberU64()))
|
|
for j := 0; j < len(blocks)/2; j++ {
|
|
blocks[j], blocks[len(blocks)-1-j] = blocks[len(blocks)-1-j], blocks[j]
|
|
}
|
|
if _, err := bc.InsertChain(blocks); err != nil {
|
|
t.Fatalf("failed to import pro-fork chain for expansion: %v", err)
|
|
}
|
|
if err := bc.stateCache.TrieDB().Commit(bc.CurrentHeader().Root, true, nil); err != nil {
|
|
t.Fatalf("failed to commit pro-fork head for expansion: %v", err)
|
|
}
|
|
blocks, _ = GenerateChain(&conConf, proBc.CurrentBlock(), ethash.NewFaker(), db, 1, func(i int, gen *BlockGen) {})
|
|
if _, err := proBc.InsertChain(blocks); err == nil {
|
|
t.Fatalf("pro-fork chain accepted contra-fork block: %v", blocks[0])
|
|
}
|
|
// Create a proper pro-fork block for the pro-forker
|
|
blocks, _ = GenerateChain(&proConf, proBc.CurrentBlock(), ethash.NewFaker(), db, 1, func(i int, gen *BlockGen) {})
|
|
if _, err := proBc.InsertChain(blocks); err != nil {
|
|
t.Fatalf("pro-fork chain didn't accepted pro-fork block: %v", err)
|
|
}
|
|
}
|
|
// Verify that contra-forkers accept pro-fork extra-datas after forking finishes
|
|
db = rawdb.NewMemoryDatabase()
|
|
gspec.MustCommit(db)
|
|
bc, _ := NewBlockChain(db, nil, &conConf, ethash.NewFaker(), vm.Config{}, nil, nil)
|
|
defer bc.Stop()
|
|
|
|
blocks := conBc.GetBlocksFromHash(conBc.CurrentBlock().Hash(), int(conBc.CurrentBlock().NumberU64()))
|
|
for j := 0; j < len(blocks)/2; j++ {
|
|
blocks[j], blocks[len(blocks)-1-j] = blocks[len(blocks)-1-j], blocks[j]
|
|
}
|
|
if _, err := bc.InsertChain(blocks); err != nil {
|
|
t.Fatalf("failed to import contra-fork chain for expansion: %v", err)
|
|
}
|
|
if err := bc.stateCache.TrieDB().Commit(bc.CurrentHeader().Root, true, nil); err != nil {
|
|
t.Fatalf("failed to commit contra-fork head for expansion: %v", err)
|
|
}
|
|
blocks, _ = GenerateChain(&proConf, conBc.CurrentBlock(), ethash.NewFaker(), db, 1, func(i int, gen *BlockGen) {})
|
|
if _, err := conBc.InsertChain(blocks); err != nil {
|
|
t.Fatalf("contra-fork chain didn't accept pro-fork block post-fork: %v", err)
|
|
}
|
|
// Verify that pro-forkers accept contra-fork extra-datas after forking finishes
|
|
db = rawdb.NewMemoryDatabase()
|
|
gspec.MustCommit(db)
|
|
bc, _ = NewBlockChain(db, nil, &proConf, ethash.NewFaker(), vm.Config{}, nil, nil)
|
|
defer bc.Stop()
|
|
|
|
blocks = proBc.GetBlocksFromHash(proBc.CurrentBlock().Hash(), int(proBc.CurrentBlock().NumberU64()))
|
|
for j := 0; j < len(blocks)/2; j++ {
|
|
blocks[j], blocks[len(blocks)-1-j] = blocks[len(blocks)-1-j], blocks[j]
|
|
}
|
|
if _, err := bc.InsertChain(blocks); err != nil {
|
|
t.Fatalf("failed to import pro-fork chain for expansion: %v", err)
|
|
}
|
|
if err := bc.stateCache.TrieDB().Commit(bc.CurrentHeader().Root, true, nil); err != nil {
|
|
t.Fatalf("failed to commit pro-fork head for expansion: %v", err)
|
|
}
|
|
blocks, _ = GenerateChain(&conConf, proBc.CurrentBlock(), ethash.NewFaker(), db, 1, func(i int, gen *BlockGen) {})
|
|
if _, err := proBc.InsertChain(blocks); err != nil {
|
|
t.Fatalf("pro-fork chain didn't accept contra-fork block post-fork: %v", err)
|
|
}
|
|
}
|