mirror of
https://gitlab.com/pulsechaincom/erigon-pulse.git
synced 2024-12-28 06:37:17 +00:00
9712e858e2
* aloow_delete_dupsort_values * check canonical receipt * exclusive lock for mdbx * readme_load_dump_tables * readme_load_dump_tables * readme_load_dump_tables * Remove IdealBatchSize change Co-authored-by: Alexey Akhunov <akhounov@gmail.com>
745 lines
25 KiB
Go
745 lines
25 KiB
Go
// Copyright 2018 The go-ethereum Authors
|
|
// This file is part of the go-ethereum library.
|
|
//
|
|
// The go-ethereum library is free software: you can redistribute it and/or modify
|
|
// it under the terms of the GNU Lesser General Public License as published by
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
// (at your option) any later version.
|
|
//
|
|
// The go-ethereum library is distributed in the hope that it will be useful,
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
// GNU Lesser General Public License for more details.
|
|
//
|
|
// You should have received a copy of the GNU Lesser General Public License
|
|
// along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
package rawdb
|
|
|
|
import (
|
|
"bytes"
|
|
"context"
|
|
"encoding/binary"
|
|
"errors"
|
|
"fmt"
|
|
"math/big"
|
|
|
|
"github.com/ledgerwatch/turbo-geth/ethdb"
|
|
"github.com/ledgerwatch/turbo-geth/ethdb/cbor"
|
|
|
|
"github.com/ledgerwatch/turbo-geth/common"
|
|
"github.com/ledgerwatch/turbo-geth/common/dbutils"
|
|
"github.com/ledgerwatch/turbo-geth/common/debug"
|
|
"github.com/ledgerwatch/turbo-geth/core/types"
|
|
"github.com/ledgerwatch/turbo-geth/log"
|
|
"github.com/ledgerwatch/turbo-geth/rlp"
|
|
|
|
"github.com/golang/snappy"
|
|
)
|
|
|
|
// ReadCanonicalHash retrieves the hash assigned to a canonical block number.
|
|
func ReadCanonicalHash(db DatabaseReader, number uint64) (common.Hash, error) {
|
|
data, err := db.Get(dbutils.HeaderPrefix, dbutils.HeaderHashKey(number))
|
|
if err != nil && !errors.Is(err, ethdb.ErrKeyNotFound) {
|
|
return common.Hash{}, fmt.Errorf("failed ReadCanonicalHash: %w, number=%d", err, number)
|
|
}
|
|
if len(data) == 0 {
|
|
return common.Hash{}, nil
|
|
}
|
|
return common.BytesToHash(data), nil
|
|
}
|
|
|
|
// WriteCanonicalHash stores the hash assigned to a canonical block number.
|
|
func WriteCanonicalHash(db DatabaseWriter, hash common.Hash, number uint64) error {
|
|
if err := db.Put(dbutils.HeaderPrefix, dbutils.HeaderHashKey(number), hash.Bytes()); err != nil {
|
|
return fmt.Errorf("failed to store number to hash mapping: %w", err)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// DeleteCanonicalHash removes the number to hash canonical mapping.
|
|
func DeleteCanonicalHash(db DatabaseDeleter, number uint64) error {
|
|
if err := db.Delete(dbutils.HeaderPrefix, dbutils.HeaderHashKey(number), nil); err != nil {
|
|
return fmt.Errorf("failed to delete number to hash mapping: %w", err)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// ReadAllHashes retrieves all the hashes assigned to blocks at a certain heights,
|
|
// both canonical and reorged forks included.
|
|
func ReadAllHashes(db DatabaseReader, number uint64) []common.Hash {
|
|
//prefix := headerKeyPrefix(number)
|
|
|
|
hashes := make([]common.Hash, 0, 1)
|
|
/*
|
|
it := db.NewIteratorWithPrefix(prefix)
|
|
defer it.Release()
|
|
|
|
for it.Next() {
|
|
if key := it.Key(); len(key) == len(prefix)+32 {
|
|
hashes = append(hashes, common.BytesToHash(key[len(key)-32:]))
|
|
}
|
|
}
|
|
*/
|
|
return hashes
|
|
}
|
|
|
|
// ReadHeaderNumber returns the header number assigned to a hash.
|
|
func ReadHeaderNumber(db DatabaseReader, hash common.Hash) *uint64 {
|
|
data, err := db.Get(dbutils.HeaderNumberPrefix, hash.Bytes())
|
|
if err != nil && !errors.Is(err, ethdb.ErrKeyNotFound) {
|
|
log.Error("ReadHeaderNumber failed", "err", err)
|
|
}
|
|
if len(data) == 0 {
|
|
return nil
|
|
}
|
|
if len(data) != 8 {
|
|
log.Error("ReadHeaderNumber got wrong data len", "len", len(data))
|
|
return nil
|
|
}
|
|
number := binary.BigEndian.Uint64(data)
|
|
return &number
|
|
}
|
|
|
|
// WriteHeaderNumber stores the hash->number mapping.
|
|
func WriteHeaderNumber(db DatabaseWriter, hash common.Hash, number uint64) {
|
|
enc := dbutils.EncodeBlockNumber(number)
|
|
if err := db.Put(dbutils.HeaderNumberPrefix, hash[:], enc); err != nil {
|
|
log.Crit("Failed to store hash to number mapping", "err", err)
|
|
}
|
|
}
|
|
|
|
// DeleteHeaderNumber removes hash->number mapping.
|
|
func DeleteHeaderNumber(db DatabaseDeleter, hash common.Hash) {
|
|
if err := db.Delete(dbutils.HeaderNumberPrefix, hash[:], nil); err != nil {
|
|
log.Crit("Failed to delete hash to number mapping", "err", err)
|
|
}
|
|
}
|
|
|
|
// ReadHeadHeaderHash retrieves the hash of the current canonical head header.
|
|
func ReadHeadHeaderHash(db DatabaseReader) common.Hash {
|
|
data, err := db.Get(dbutils.HeadHeaderKey, []byte(dbutils.HeadHeaderKey))
|
|
if err != nil && !errors.Is(err, ethdb.ErrKeyNotFound) {
|
|
log.Error("ReadHeadHeaderHash failed", "err", err)
|
|
}
|
|
if len(data) == 0 {
|
|
return common.Hash{}
|
|
}
|
|
return common.BytesToHash(data)
|
|
}
|
|
|
|
// WriteHeadHeaderHash stores the hash of the current canonical head header.
|
|
func WriteHeadHeaderHash(db DatabaseWriter, hash common.Hash) {
|
|
if err := db.Put(dbutils.HeadHeaderKey, []byte(dbutils.HeadHeaderKey), hash.Bytes()); err != nil {
|
|
log.Crit("Failed to store last header's hash", "err", err)
|
|
}
|
|
}
|
|
|
|
// ReadHeadBlockHash retrieves the hash of the current canonical head block.
|
|
func ReadHeadBlockHash(db DatabaseReader) common.Hash {
|
|
data, err := db.Get(dbutils.HeadBlockKey, []byte(dbutils.HeadBlockKey))
|
|
if err != nil && !errors.Is(err, ethdb.ErrKeyNotFound) {
|
|
log.Error("ReadHeadBlockHash failed", "err", err)
|
|
}
|
|
if len(data) == 0 {
|
|
return common.Hash{}
|
|
}
|
|
return common.BytesToHash(data)
|
|
}
|
|
|
|
// WriteHeadBlockHash stores the head block's hash.
|
|
func WriteHeadBlockHash(db DatabaseWriter, hash common.Hash) {
|
|
if err := db.Put(dbutils.HeadBlockKey, []byte(dbutils.HeadBlockKey), hash.Bytes()); err != nil {
|
|
log.Crit("Failed to store last block's hash", "err", err)
|
|
}
|
|
}
|
|
|
|
// ReadHeadFastBlockHash retrieves the hash of the current fast-sync head block.
|
|
func ReadHeadFastBlockHash(db DatabaseReader) common.Hash {
|
|
data, err := db.Get(dbutils.HeadFastBlockKey, []byte(dbutils.HeadFastBlockKey))
|
|
if err != nil && !errors.Is(err, ethdb.ErrKeyNotFound) {
|
|
log.Error("ReadHeadFastBlockHash failed", "err", err)
|
|
}
|
|
if len(data) == 0 {
|
|
return common.Hash{}
|
|
}
|
|
return common.BytesToHash(data)
|
|
}
|
|
|
|
// WriteHeadFastBlockHash stores the hash of the current fast-sync head block.
|
|
func WriteHeadFastBlockHash(db DatabaseWriter, hash common.Hash) {
|
|
if err := db.Put(dbutils.HeadFastBlockKey, []byte(dbutils.HeadFastBlockKey), hash.Bytes()); err != nil {
|
|
log.Crit("Failed to store last fast block's hash", "err", err)
|
|
}
|
|
}
|
|
|
|
// ReadFastTrieProgress retrieves the number of tries nodes fast synced to allow
|
|
// reporting correct numbers across restarts.
|
|
func ReadFastTrieProgress(db DatabaseReader) uint64 {
|
|
data, err := db.Get(dbutils.FastTrieProgressKey, []byte(dbutils.FastTrieProgressKey))
|
|
if err != nil && !errors.Is(err, ethdb.ErrKeyNotFound) {
|
|
log.Error("ReadFastTrieProgress failed", "err", err)
|
|
}
|
|
if len(data) == 0 {
|
|
return 0
|
|
}
|
|
return new(big.Int).SetBytes(data).Uint64()
|
|
}
|
|
|
|
// WriteFastTrieProgress stores the fast sync trie process counter to support
|
|
// retrieving it across restarts.
|
|
func WriteFastTrieProgress(db DatabaseWriter, count uint64) {
|
|
if err := db.Put(dbutils.FastTrieProgressKey, []byte(dbutils.FastTrieProgressKey), new(big.Int).SetUint64(count).Bytes()); err != nil {
|
|
log.Crit("Failed to store fast sync trie progress", "err", err)
|
|
}
|
|
}
|
|
|
|
// ReadHeaderRLP retrieves a block header in its raw RLP database encoding.
|
|
func ReadHeaderRLP(db DatabaseReader, hash common.Hash, number uint64) rlp.RawValue {
|
|
data, err := db.Get(dbutils.HeaderPrefix, dbutils.HeaderKey(number, hash))
|
|
if err != nil && !errors.Is(err, ethdb.ErrKeyNotFound) {
|
|
log.Error("ReadHeaderRLP failed", "err", err)
|
|
}
|
|
return data
|
|
}
|
|
|
|
// HasHeader verifies the existence of a block header corresponding to the hash.
|
|
func HasHeader(db DatabaseReader, hash common.Hash, number uint64) bool {
|
|
if has, err := db.Has(dbutils.HeaderPrefix, dbutils.HeaderKey(number, hash)); !has || err != nil {
|
|
return false
|
|
}
|
|
return true
|
|
}
|
|
|
|
// ReadHeader retrieves the block header corresponding to the hash.
|
|
func ReadHeader(db DatabaseReader, hash common.Hash, number uint64) *types.Header {
|
|
data := ReadHeaderRLP(db, hash, number)
|
|
if len(data) == 0 {
|
|
return nil
|
|
}
|
|
header := new(types.Header)
|
|
if err := rlp.Decode(bytes.NewReader(data), header); err != nil {
|
|
log.Error("Invalid block header RLP", "hash", hash, "err", err)
|
|
return nil
|
|
}
|
|
return header
|
|
}
|
|
|
|
// WriteHeader stores a block header into the database and also stores the hash-
|
|
// to-number mapping.
|
|
func WriteHeader(ctx context.Context, db DatabaseWriter, header *types.Header) {
|
|
var (
|
|
hash = header.Hash()
|
|
number = header.Number.Uint64()
|
|
encoded = dbutils.EncodeBlockNumber(number)
|
|
)
|
|
if common.IsCanceled(ctx) {
|
|
return
|
|
}
|
|
if err := db.Put(dbutils.HeaderNumberPrefix, hash[:], encoded); err != nil {
|
|
log.Crit("Failed to store hash to number mapping", "err", err)
|
|
}
|
|
// Write the encoded header
|
|
data, err := rlp.EncodeToBytes(header)
|
|
if err != nil {
|
|
log.Crit("Failed to RLP encode header", "err", err)
|
|
}
|
|
if err := db.Put(dbutils.HeaderPrefix, dbutils.HeaderKey(number, hash), data); err != nil {
|
|
log.Crit("Failed to store header", "err", err)
|
|
}
|
|
}
|
|
|
|
// DeleteHeader removes all block header data associated with a hash.
|
|
func DeleteHeader(db DatabaseDeleter, hash common.Hash, number uint64) {
|
|
if err := db.Delete(dbutils.HeaderPrefix, dbutils.HeaderKey(number, hash), nil); err != nil {
|
|
log.Crit("Failed to delete header", "err", err)
|
|
}
|
|
if err := db.Delete(dbutils.HeaderNumberPrefix, hash.Bytes(), nil); err != nil {
|
|
log.Crit("Failed to delete hash to number mapping", "err", err)
|
|
}
|
|
}
|
|
|
|
// deleteHeaderWithoutNumber removes only the block header but does not remove
|
|
// the hash to number mapping.
|
|
func deleteHeaderWithoutNumber(db DatabaseDeleter, hash common.Hash, number uint64) {
|
|
if err := db.Delete(dbutils.HeaderPrefix, dbutils.HeaderKey(number, hash), nil); err != nil {
|
|
log.Crit("Failed to delete header", "err", err)
|
|
}
|
|
}
|
|
|
|
// ReadBodyRLP retrieves the block body (transactions and uncles) in RLP encoding.
|
|
func ReadBodyRLP(db DatabaseReader, hash common.Hash, number uint64) rlp.RawValue {
|
|
data, err1 := db.Get(dbutils.BlockBodyPrefix, dbutils.BlockBodyKey(number, hash))
|
|
if err1 != nil && !errors.Is(err1, ethdb.ErrKeyNotFound) {
|
|
log.Error("ReadBodyRLP failed", "err", err1)
|
|
}
|
|
bodyRlp, err := DecompressBlockBody(data)
|
|
if err != nil {
|
|
log.Warn("err on decode block", "err", err)
|
|
}
|
|
return bodyRlp
|
|
}
|
|
|
|
// WriteBodyRLP stores an RLP encoded block body into the database.
|
|
func WriteBodyRLP(ctx context.Context, db DatabaseWriter, hash common.Hash, number uint64, rlp rlp.RawValue) {
|
|
if common.IsCanceled(ctx) {
|
|
return
|
|
}
|
|
if debug.IsBlockCompressionEnabled() {
|
|
rlp = snappy.Encode(nil, rlp)
|
|
}
|
|
if err := db.Put(dbutils.BlockBodyPrefix, dbutils.BlockBodyKey(number, hash), rlp); err != nil {
|
|
log.Crit("Failed to store block body", "err", err)
|
|
}
|
|
}
|
|
|
|
// HasBody verifies the existence of a block body corresponding to the hash.
|
|
func HasBody(db DatabaseReader, hash common.Hash, number uint64) bool {
|
|
if has, err := db.Has(dbutils.BlockBodyPrefix, dbutils.BlockBodyKey(number, hash)); !has || err != nil {
|
|
return false
|
|
}
|
|
return true
|
|
}
|
|
|
|
// ReadBody retrieves the block body corresponding to the hash.
|
|
func ReadBody(db DatabaseReader, hash common.Hash, number uint64) *types.Body {
|
|
data := ReadBodyRLP(db, hash, number)
|
|
if len(data) == 0 {
|
|
return nil
|
|
}
|
|
body := new(types.Body)
|
|
if err := rlp.Decode(bytes.NewReader(data), body); err != nil {
|
|
log.Error("Invalid block body RLP", "hash", hash, "err", err)
|
|
return nil
|
|
}
|
|
return body
|
|
}
|
|
|
|
func ReadSenders(db DatabaseReader, hash common.Hash, number uint64) []common.Address {
|
|
data, err := db.Get(dbutils.Senders, dbutils.BlockBodyKey(number, hash))
|
|
if err != nil && !errors.Is(err, ethdb.ErrKeyNotFound) {
|
|
log.Error("ReadSenders failed", "err", err)
|
|
}
|
|
senders := make([]common.Address, len(data)/common.AddressLength)
|
|
for i := 0; i < len(senders); i++ {
|
|
copy(senders[i][:], data[i*common.AddressLength:])
|
|
}
|
|
return senders
|
|
}
|
|
|
|
// WriteBody storea a block body into the database.
|
|
func WriteBody(ctx context.Context, db DatabaseWriter, hash common.Hash, number uint64, body *types.Body) {
|
|
if common.IsCanceled(ctx) {
|
|
return
|
|
}
|
|
// Pre-processing
|
|
body.SendersFromTxs()
|
|
data, err := rlp.EncodeToBytes(body)
|
|
if err != nil {
|
|
log.Crit("Failed to RLP encode body", "err", err)
|
|
}
|
|
WriteBodyRLP(ctx, db, hash, number, data)
|
|
}
|
|
|
|
func WriteSenders(ctx context.Context, db DatabaseWriter, hash common.Hash, number uint64, senders []common.Address) {
|
|
if common.IsCanceled(ctx) {
|
|
return
|
|
}
|
|
data := make([]byte, common.AddressLength*len(senders))
|
|
for i, sender := range senders {
|
|
copy(data[i*common.AddressLength:], sender[:])
|
|
}
|
|
if err := db.Put(dbutils.Senders, dbutils.BlockBodyKey(number, hash), data); err != nil {
|
|
log.Crit("Failed to store block senders", "err", err)
|
|
}
|
|
}
|
|
|
|
// DeleteBody removes all block body data associated with a hash.
|
|
func DeleteBody(db DatabaseDeleter, hash common.Hash, number uint64) {
|
|
if err := db.Delete(dbutils.BlockBodyPrefix, dbutils.BlockBodyKey(number, hash), nil); err != nil {
|
|
log.Crit("Failed to delete block body", "err", err)
|
|
}
|
|
}
|
|
|
|
// ReadTdRLP retrieves a block's total difficulty corresponding to the hash in RLP encoding.
|
|
func ReadTdRLP(db DatabaseReader, hash common.Hash, number uint64) rlp.RawValue {
|
|
//data, _ := db.Ancient(freezerDifficultyTable, number)
|
|
data := []byte{}
|
|
if len(data) == 0 {
|
|
data, _ = db.Get(dbutils.HeaderPrefix, dbutils.HeaderTDKey(number, hash))
|
|
// In the background freezer is moving data from leveldb to flatten files.
|
|
// So during the first check for ancient db, the data is not yet in there,
|
|
// but when we reach into leveldb, the data was already moved. That would
|
|
// result in a not found error.
|
|
if len(data) == 0 {
|
|
//data, _ = db.Ancient(freezerDifficultyTable, number)
|
|
}
|
|
}
|
|
return nil // Can't find the data anywhere.
|
|
}
|
|
|
|
// ReadTd retrieves a block's total difficulty corresponding to the hash.
|
|
func ReadTd(db DatabaseReader, hash common.Hash, number uint64) (*big.Int, error) {
|
|
data, err := db.Get(dbutils.HeaderPrefix, dbutils.HeaderTDKey(number, hash))
|
|
if err != nil && !errors.Is(err, ethdb.ErrKeyNotFound) {
|
|
return nil, fmt.Errorf("failed ReadTd: %w", err)
|
|
}
|
|
if len(data) == 0 {
|
|
return nil, nil
|
|
}
|
|
td := new(big.Int)
|
|
if err := rlp.Decode(bytes.NewReader(data), td); err != nil {
|
|
return nil, fmt.Errorf("invalid block total difficulty RLP: %x, %w", hash, err)
|
|
}
|
|
return td, nil
|
|
}
|
|
|
|
// WriteTd stores the total difficulty of a block into the database.
|
|
func WriteTd(db DatabaseWriter, hash common.Hash, number uint64, td *big.Int) error {
|
|
data, err := rlp.EncodeToBytes(td)
|
|
if err != nil {
|
|
return fmt.Errorf("failed to RLP encode block total difficulty: %w", err)
|
|
}
|
|
if err := db.Put(dbutils.HeaderPrefix, dbutils.HeaderTDKey(number, hash), data); err != nil {
|
|
return fmt.Errorf("failed to store block total difficulty: %w", err)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// DeleteTd removes all block total difficulty data associated with a hash.
|
|
func DeleteTd(db DatabaseDeleter, hash common.Hash, number uint64) error {
|
|
if err := db.Delete(dbutils.HeaderPrefix, dbutils.HeaderTDKey(number, hash), nil); err != nil {
|
|
return fmt.Errorf("failed to delete block total difficulty: %w", err)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// HasReceipts verifies the existence of all the transaction receipts belonging
|
|
// to a block.
|
|
func HasReceipts(db DatabaseReader, hash common.Hash, number uint64) bool {
|
|
if has, err := db.Has(dbutils.BlockReceiptsPrefix, dbutils.ReceiptsKey(number)); !has || err != nil {
|
|
return false
|
|
}
|
|
return true
|
|
}
|
|
|
|
// ReadRawReceipts retrieves all the transaction receipts belonging to a block.
|
|
// The receipt metadata fields are not guaranteed to be populated, so they
|
|
// should not be used. Use ReadReceipts instead if the metadata is needed.
|
|
func ReadRawReceipts(db ethdb.Database, hash common.Hash, number uint64) types.Receipts {
|
|
// Retrieve the flattened receipt slice
|
|
data, err := db.Get(dbutils.BlockReceiptsPrefix, dbutils.ReceiptsKey(number))
|
|
if err != nil && !errors.Is(err, ethdb.ErrKeyNotFound) {
|
|
log.Error("ReadRawReceipts failed", "err", err)
|
|
}
|
|
if len(data) == 0 {
|
|
return nil
|
|
}
|
|
var receipts types.Receipts
|
|
if err := cbor.Unmarshal(&receipts, bytes.NewReader(data)); err != nil {
|
|
log.Error("receipt unmarshal failed", "hash", hash, "err", err)
|
|
return nil
|
|
}
|
|
|
|
if err := db.Walk(dbutils.Log, dbutils.LogKey(number, 0), 8*8, func(k, v []byte) (bool, error) {
|
|
var logs types.Logs
|
|
if err := cbor.Unmarshal(&logs, bytes.NewReader(v)); err != nil {
|
|
return false, fmt.Errorf("receipt unmarshal failed: %x, %w", hash, err)
|
|
}
|
|
|
|
receipts[binary.BigEndian.Uint32(k[8:])].Logs = logs
|
|
return true, nil
|
|
}); err != nil {
|
|
log.Error("logs fetching failed", "hash", hash, "err", err)
|
|
return nil
|
|
}
|
|
|
|
return receipts
|
|
}
|
|
|
|
// ReadReceipts retrieves all the transaction receipts belonging to a block, including
|
|
// its correspoinding metadata fields. If it is unable to populate these metadata
|
|
// fields then nil is returned.
|
|
//
|
|
// The current implementation populates these metadata fields by reading the receipts'
|
|
// corresponding block body, so if the block body is not found it will return nil even
|
|
// if the receipt itself is stored.
|
|
func ReadReceipts(db ethdb.Database, hash common.Hash, number uint64) types.Receipts {
|
|
// We're deriving many fields from the block body, retrieve beside the receipt
|
|
receipts := ReadRawReceipts(db, hash, number)
|
|
if receipts == nil {
|
|
return nil
|
|
}
|
|
body := ReadBody(db, hash, number)
|
|
if body == nil {
|
|
log.Error("Missing body but have receipt", "hash", hash, "number", number)
|
|
return nil
|
|
}
|
|
senders := ReadSenders(db, hash, number)
|
|
if err := receipts.DeriveFields(hash, number, body.Transactions, senders); err != nil {
|
|
log.Error("Failed to derive block receipts fields", "hash", hash, "number", number, "err", err)
|
|
return nil
|
|
}
|
|
return receipts
|
|
}
|
|
|
|
// WriteReceipts stores all the transaction receipts belonging to a block.
|
|
func WriteReceipts(tx DatabaseWriter, number uint64, receipts types.Receipts) error {
|
|
buf := bytes.NewBuffer(make([]byte, 0, 1024))
|
|
for txId, r := range receipts {
|
|
if len(r.Logs) == 0 {
|
|
continue
|
|
}
|
|
|
|
buf.Reset()
|
|
err := cbor.Marshal(buf, r.Logs)
|
|
if err != nil {
|
|
return fmt.Errorf("encode block logs for block %d: %v", number, err)
|
|
}
|
|
|
|
if err = tx.Put(dbutils.Log, dbutils.LogKey(number, uint32(txId)), buf.Bytes()); err != nil {
|
|
return fmt.Errorf("writing logs for block %d: %v", number, err)
|
|
}
|
|
}
|
|
|
|
buf.Reset()
|
|
err := cbor.Marshal(buf, receipts)
|
|
if err != nil {
|
|
return fmt.Errorf("encode block receipts for block %d: %v", number, err)
|
|
}
|
|
|
|
if err = tx.Put(dbutils.BlockReceiptsPrefix, dbutils.ReceiptsKey(number), buf.Bytes()); err != nil {
|
|
return fmt.Errorf("writing receipts for block %d: %v", number, err)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// WriteReceipts stores all the transaction receipts belonging to a block.
|
|
func AppendReceipts(tx ethdb.DbWithPendingMutations, blockNumber uint64, receipts types.Receipts) error {
|
|
buf := bytes.NewBuffer(make([]byte, 0, 1024))
|
|
for txId, r := range receipts {
|
|
if len(r.Logs) == 0 {
|
|
continue
|
|
}
|
|
|
|
buf.Reset()
|
|
err := cbor.Marshal(buf, r.Logs)
|
|
if err != nil {
|
|
return fmt.Errorf("encode block receipts for block %d: %v", blockNumber, err)
|
|
}
|
|
|
|
if err = tx.Append(dbutils.Log, dbutils.LogKey(blockNumber, uint32(txId)), buf.Bytes()); err != nil {
|
|
return fmt.Errorf("writing receipts for block %d: %v", blockNumber, err)
|
|
}
|
|
}
|
|
|
|
buf.Reset()
|
|
err := cbor.Marshal(buf, receipts)
|
|
if err != nil {
|
|
return fmt.Errorf("encode block receipts for block %d: %v", blockNumber, err)
|
|
}
|
|
|
|
if err = tx.Append(dbutils.BlockReceiptsPrefix, dbutils.ReceiptsKey(blockNumber), buf.Bytes()); err != nil {
|
|
return fmt.Errorf("writing receipts for block %d: %v", blockNumber, err)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// DeleteReceipts removes all receipt data associated with a block hash.
|
|
func DeleteReceipts(db ethdb.Database, number uint64) error {
|
|
if err := db.Delete(dbutils.BlockReceiptsPrefix, dbutils.ReceiptsKey(number), nil); err != nil {
|
|
return fmt.Errorf("receipts delete failed: %d, %w", number, err)
|
|
}
|
|
|
|
if err := db.Walk(dbutils.Log, dbutils.LogKey(number, 0), 8*8, func(k, v []byte) (bool, error) {
|
|
if err := db.Delete(dbutils.Log, k, nil); err != nil {
|
|
return false, err
|
|
}
|
|
return true, nil
|
|
}); err != nil {
|
|
return fmt.Errorf("logs delete failed: %d, %w", number, err)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// DeleteNewerReceipts removes all receipt for given block number or newer
|
|
func DeleteNewerReceipts(db ethdb.Database, number uint64) error {
|
|
if err := db.Walk(dbutils.BlockReceiptsPrefix, dbutils.ReceiptsKey(number), 0, func(k, v []byte) (bool, error) {
|
|
if err := db.Delete(dbutils.BlockReceiptsPrefix, k, nil); err != nil {
|
|
return false, err
|
|
}
|
|
return true, nil
|
|
}); err != nil {
|
|
return fmt.Errorf("delete newer receipts failed: %d, %w", number, err)
|
|
}
|
|
|
|
if err := db.Walk(dbutils.Log, dbutils.LogKey(number, 0), 0, func(k, v []byte) (bool, error) {
|
|
if err := db.Delete(dbutils.Log, k, nil); err != nil {
|
|
return false, err
|
|
}
|
|
return true, nil
|
|
}); err != nil {
|
|
return fmt.Errorf("delete newer logs failed: %d, %w", number, err)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// ReadBlock retrieves an entire block corresponding to the hash, assembling it
|
|
// back from the stored header and body. If either the header or body could not
|
|
// be retrieved nil is returned.
|
|
//
|
|
// Note, due to concurrent download of header and block body the header and thus
|
|
// canonical hash can be stored in the database but the body data not (yet).
|
|
func ReadBlock(db DatabaseReader, hash common.Hash, number uint64) *types.Block {
|
|
header := ReadHeader(db, hash, number)
|
|
if header == nil {
|
|
return nil
|
|
}
|
|
body := ReadBody(db, hash, number)
|
|
if body == nil {
|
|
return nil
|
|
}
|
|
return types.NewBlockWithHeader(header).WithBody(body.Transactions, body.Uncles)
|
|
}
|
|
|
|
// WriteBlock serializes a block into the database, header and body separately.
|
|
func WriteBlock(ctx context.Context, db DatabaseWriter, block *types.Block) error {
|
|
WriteBody(ctx, db, block.Hash(), block.NumberU64(), block.Body())
|
|
WriteHeader(ctx, db, block.Header())
|
|
return nil
|
|
}
|
|
|
|
// WriteAncientBlock writes entire block data into ancient store and returns the total written size.
|
|
/*
|
|
func WriteAncientBlock(db ethdb.AncientWriter, block *types.Block, receipts types.Receipts, td *big.Int) int {
|
|
// Encode all block components to RLP format.
|
|
headerBlob, err := rlp.EncodeToBytes(block.Header())
|
|
if err != nil {
|
|
log.Crit("Failed to RLP encode block header", "err", err)
|
|
}
|
|
bodyBlob, err := rlp.EncodeToBytes(block.Body())
|
|
if err != nil {
|
|
log.Crit("Failed to RLP encode body", "err", err)
|
|
}
|
|
storageReceipts := make([]*types.ReceiptForStorage, len(receipts))
|
|
for i, receipt := range receipts {
|
|
storageReceipts[i] = (*types.ReceiptForStorage)(receipt)
|
|
}
|
|
receiptBlob, err := rlp.EncodeToBytes(storageReceipts)
|
|
if err != nil {
|
|
log.Crit("Failed to RLP encode block receipts", "err", err)
|
|
}
|
|
tdBlob, err := rlp.EncodeToBytes(td)
|
|
if err != nil {
|
|
log.Crit("Failed to RLP encode block total difficulty", "err", err)
|
|
}
|
|
// Write all blob to flatten files.
|
|
err = db.AppendAncient(block.NumberU64(), block.Hash().Bytes(), headerBlob, bodyBlob, receiptBlob, tdBlob)
|
|
if err != nil {
|
|
log.Crit("Failed to write block data to ancient store", "err", err)
|
|
}
|
|
return len(headerBlob) + len(bodyBlob) + len(receiptBlob) + len(tdBlob) + common.HashLength
|
|
}
|
|
*/
|
|
|
|
// DeleteBlock removes all block data associated with a hash.
|
|
func DeleteBlock(db ethdb.Database, hash common.Hash, number uint64) error {
|
|
if err := DeleteReceipts(db, number); err != nil {
|
|
return err
|
|
}
|
|
DeleteHeader(db, hash, number)
|
|
DeleteBody(db, hash, number)
|
|
if err := DeleteTd(db, hash, number); err != nil {
|
|
return err
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// DeleteBlockWithoutNumber removes all block data associated with a hash, except
|
|
// the hash to number mapping.
|
|
func DeleteBlockWithoutNumber(db ethdb.Database, hash common.Hash, number uint64) error {
|
|
if err := DeleteReceipts(db, number); err != nil {
|
|
return err
|
|
}
|
|
deleteHeaderWithoutNumber(db, hash, number)
|
|
DeleteBody(db, hash, number)
|
|
if err := DeleteTd(db, hash, number); err != nil {
|
|
return err
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// FindCommonAncestor returns the last common ancestor of two block headers
|
|
func FindCommonAncestor(db DatabaseReader, a, b *types.Header) *types.Header {
|
|
for bn := b.Number.Uint64(); a.Number.Uint64() > bn; {
|
|
a = ReadHeader(db, a.ParentHash, a.Number.Uint64()-1)
|
|
if a == nil {
|
|
return nil
|
|
}
|
|
}
|
|
for an := a.Number.Uint64(); an < b.Number.Uint64(); {
|
|
b = ReadHeader(db, b.ParentHash, b.Number.Uint64()-1)
|
|
if b == nil {
|
|
return nil
|
|
}
|
|
}
|
|
for a.Hash() != b.Hash() {
|
|
a = ReadHeader(db, a.ParentHash, a.Number.Uint64()-1)
|
|
if a == nil {
|
|
return nil
|
|
}
|
|
b = ReadHeader(db, b.ParentHash, b.Number.Uint64()-1)
|
|
if b == nil {
|
|
return nil
|
|
}
|
|
}
|
|
return a
|
|
}
|
|
|
|
func ReadBlockByNumber(db DatabaseReader, number uint64) (*types.Block, error) {
|
|
hash, err := ReadCanonicalHash(db, number)
|
|
if err != nil {
|
|
return nil, fmt.Errorf("failed ReadCanonicalHash: %w", err)
|
|
}
|
|
if hash == (common.Hash{}) {
|
|
return nil, nil
|
|
}
|
|
|
|
return ReadBlock(db, hash, number), nil
|
|
}
|
|
|
|
func ReadBlockByHash(db DatabaseReader, hash common.Hash) (*types.Block, error) {
|
|
number := ReadHeaderNumber(db, hash)
|
|
if number == nil {
|
|
return nil, nil
|
|
}
|
|
return ReadBlock(db, hash, *number), nil
|
|
}
|
|
|
|
func ReadHeaderByNumber(db DatabaseReader, number uint64) *types.Header {
|
|
hash, err := ReadCanonicalHash(db, number)
|
|
if err != nil {
|
|
log.Error("ReadCanonicalHash failed", "err", err)
|
|
return nil
|
|
}
|
|
if hash == (common.Hash{}) {
|
|
return nil
|
|
}
|
|
|
|
return ReadHeader(db, hash, number)
|
|
}
|
|
|
|
func ReadHeaderByHash(db DatabaseReader, hash common.Hash) (*types.Header, error) {
|
|
number := ReadHeaderNumber(db, hash)
|
|
if number == nil {
|
|
return nil, nil
|
|
}
|
|
return ReadHeader(db, hash, *number), nil
|
|
}
|
|
|
|
// FIXME: implement in Turbo-Geth
|
|
// WriteAncientBlock writes entire block data into ancient store and returns the total written size.
|
|
func WriteAncientBlock(db DatabaseWriter, block *types.Block, receipts types.Receipts, td *big.Int) int {
|
|
panic("not implemented")
|
|
}
|