mirror of
https://gitlab.com/pulsechaincom/erigon-pulse.git
synced 2024-12-29 07:07:16 +00:00
fd77eaf86a
* introduce PlainStateReader with fallbacks * no 10.000 changes in tests * even less iterations * remove even more iterations * add `go run ./cmd/geth --syncmode staged --plainstate` flag * fix serialization calls * make a more sensible file default doesn’t affect anything, because this flag is always overriden when parsing CLI. but still.
227 lines
6.5 KiB
Go
227 lines
6.5 KiB
Go
package state
|
|
|
|
import (
|
|
"context"
|
|
"fmt"
|
|
|
|
"github.com/ledgerwatch/turbo-geth/common"
|
|
"github.com/ledgerwatch/turbo-geth/common/changeset"
|
|
"github.com/ledgerwatch/turbo-geth/common/dbutils"
|
|
"github.com/ledgerwatch/turbo-geth/core/rawdb"
|
|
"github.com/ledgerwatch/turbo-geth/core/types/accounts"
|
|
"github.com/ledgerwatch/turbo-geth/ethdb"
|
|
"github.com/ledgerwatch/turbo-geth/trie"
|
|
)
|
|
|
|
var _ WriterWithChangeSets = (*DbStateWriter)(nil)
|
|
|
|
func NewDbStateWriter(db ethdb.Database, blockNr uint64, incarnationMap map[common.Address]uint64) *DbStateWriter {
|
|
return &DbStateWriter{
|
|
db: db,
|
|
blockNr: blockNr,
|
|
pw: &PreimageWriter{db: db, savePreimages: false},
|
|
csw: NewChangeSetWriter(),
|
|
incarnationMap: incarnationMap,
|
|
}
|
|
}
|
|
|
|
type DbStateWriter struct {
|
|
db ethdb.Database
|
|
pw *PreimageWriter
|
|
blockNr uint64
|
|
csw *ChangeSetWriter
|
|
incarnationMap map[common.Address]uint64
|
|
}
|
|
|
|
func originalAccountData(original *accounts.Account, omitHashes bool) []byte {
|
|
var originalData []byte
|
|
if !original.Initialised {
|
|
originalData = []byte{}
|
|
} else if omitHashes {
|
|
testAcc := original.SelfCopy()
|
|
copy(testAcc.CodeHash[:], emptyCodeHash)
|
|
testAcc.Root = trie.EmptyRoot
|
|
originalDataLen := testAcc.EncodingLengthForStorage()
|
|
originalData = make([]byte, originalDataLen)
|
|
testAcc.EncodeForStorage(originalData)
|
|
} else {
|
|
originalDataLen := original.EncodingLengthForStorage()
|
|
originalData = make([]byte, originalDataLen)
|
|
original.EncodeForStorage(originalData)
|
|
}
|
|
return originalData
|
|
}
|
|
|
|
func (dsw *DbStateWriter) UpdateAccountData(ctx context.Context, address common.Address, original, account *accounts.Account) error {
|
|
if err := dsw.csw.UpdateAccountData(ctx, address, original, account); err != nil {
|
|
return err
|
|
}
|
|
addrHash, err := dsw.pw.HashAddress(address, true /*save*/)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
if err := rawdb.WriteAccount(dsw.db, addrHash, *account); err != nil {
|
|
return err
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (dsw *DbStateWriter) DeleteAccount(ctx context.Context, address common.Address, original *accounts.Account) error {
|
|
if err := dsw.csw.DeleteAccount(ctx, address, original); err != nil {
|
|
return err
|
|
}
|
|
addrHash, err := dsw.pw.HashAddress(address, true /*save*/)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
if err := rawdb.DeleteAccount(dsw.db, addrHash); err != nil {
|
|
return err
|
|
}
|
|
if original.Incarnation > 0 {
|
|
dsw.incarnationMap[address] = original.Incarnation
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (dsw *DbStateWriter) UpdateAccountCode(address common.Address, incarnation uint64, codeHash common.Hash, code []byte) error {
|
|
if err := dsw.csw.UpdateAccountCode(address, incarnation, codeHash, code); err != nil {
|
|
return err
|
|
}
|
|
//save contract code mapping
|
|
if err := dsw.db.Put(dbutils.CodeBucket, codeHash[:], code); err != nil {
|
|
return err
|
|
}
|
|
addrHash, err := common.HashData(address.Bytes())
|
|
if err != nil {
|
|
return err
|
|
}
|
|
//save contract to codeHash mapping
|
|
return dsw.db.Put(dbutils.ContractCodeBucket, dbutils.GenerateStoragePrefix(addrHash[:], incarnation), codeHash[:])
|
|
}
|
|
|
|
func (dsw *DbStateWriter) WriteAccountStorage(ctx context.Context, address common.Address, incarnation uint64, key, original, value *common.Hash) error {
|
|
// We delegate here first to let the changeSetWrite make its own decision on whether to proceed in case *original == *value
|
|
if err := dsw.csw.WriteAccountStorage(ctx, address, incarnation, key, original, value); err != nil {
|
|
return err
|
|
}
|
|
if *original == *value {
|
|
return nil
|
|
}
|
|
seckey, err := dsw.pw.HashKey(key, true /*save*/)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
addrHash, err := dsw.pw.HashAddress(address, false /*save*/)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
compositeKey := dbutils.GenerateCompositeStorageKey(addrHash, incarnation, seckey)
|
|
|
|
v := cleanUpTrailingZeroes(value[:])
|
|
if len(v) == 0 {
|
|
return dsw.db.Delete(dbutils.CurrentStateBucket, compositeKey)
|
|
} else {
|
|
vv := make([]byte, len(v))
|
|
copy(vv, v)
|
|
return dsw.db.Put(dbutils.CurrentStateBucket, compositeKey, vv)
|
|
}
|
|
}
|
|
|
|
func (dsw *DbStateWriter) CreateContract(address common.Address) error {
|
|
return dsw.csw.CreateContract(address)
|
|
}
|
|
|
|
// WriteChangeSets causes accumulated change sets to be written into
|
|
// the database (or batch) associated with the `dsw`
|
|
func (dsw *DbStateWriter) WriteChangeSets() error {
|
|
accountChanges, err := dsw.csw.GetAccountChanges()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
var accountSerialised []byte
|
|
accountSerialised, err = changeset.EncodeAccounts(accountChanges)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
key := dbutils.EncodeTimestamp(dsw.blockNr)
|
|
if err = dsw.db.Put(dbutils.AccountChangeSetBucket, key, accountSerialised); err != nil {
|
|
return err
|
|
}
|
|
storageChanges, err := dsw.csw.GetStorageChanges()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
var storageSerialized []byte
|
|
if storageChanges.Len() > 0 {
|
|
storageSerialized, err = changeset.EncodeStorage(storageChanges)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
if err = dsw.db.Put(dbutils.StorageChangeSetBucket, key, storageSerialized); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (dsw *DbStateWriter) WriteHistory() error {
|
|
accountChanges, err := dsw.csw.GetAccountChanges()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
err = dsw.writeIndex(accountChanges, dbutils.AccountsHistoryBucket)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
storageChanges, err := dsw.csw.GetStorageChanges()
|
|
if err != nil {
|
|
return err
|
|
}
|
|
err = dsw.writeIndex(storageChanges, dbutils.StorageHistoryBucket)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
func (dsw *DbStateWriter) writeIndex(changes *changeset.ChangeSet, bucket []byte) error {
|
|
for _, change := range changes.Changes {
|
|
currentChunkKey := dbutils.IndexChunkKey(change.Key, ^uint64(0))
|
|
indexBytes, err := dsw.db.Get(bucket, currentChunkKey)
|
|
if err != nil && err != ethdb.ErrKeyNotFound {
|
|
return fmt.Errorf("find chunk failed: %w", err)
|
|
}
|
|
v := dsw.blockNr
|
|
|
|
var index dbutils.HistoryIndexBytes
|
|
if len(indexBytes) == 0 {
|
|
index = dbutils.NewHistoryIndex()
|
|
} else if dbutils.CheckNewIndexChunk(indexBytes, v) {
|
|
// Chunk overflow, need to write the "old" current chunk under its key derived from the last element
|
|
index = dbutils.WrapHistoryIndex(indexBytes)
|
|
indexKey, err := index.Key(change.Key)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
// Flush the old chunk
|
|
if err := dsw.db.Put(bucket, indexKey, index); err != nil {
|
|
return err
|
|
}
|
|
// Start a new chunk
|
|
index = dbutils.NewHistoryIndex()
|
|
} else {
|
|
index = dbutils.WrapHistoryIndex(indexBytes)
|
|
}
|
|
index = index.Append(v, len(change.Value) == 0)
|
|
|
|
if err := dsw.db.Put(bucket, currentChunkKey, index); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
|
|
return nil
|
|
}
|