Snapshots: retire testing tool (#3684)

This commit is contained in:
Alex Sharov 2022-03-12 17:26:11 +07:00 committed by GitHub
parent aed679c8a0
commit f03d08c5ce
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
11 changed files with 34 additions and 2792 deletions

View File

@ -1,375 +0,0 @@
package debug
import (
"context"
"encoding/binary"
"fmt"
"os"
"testing"
"time"
"github.com/holiman/uint256"
"github.com/ledgerwatch/erigon-lib/kv"
"github.com/ledgerwatch/erigon-lib/kv/mdbx"
"github.com/ledgerwatch/erigon/common"
"github.com/ledgerwatch/erigon/common/dbutils"
"github.com/ledgerwatch/erigon/consensus/ethash"
"github.com/ledgerwatch/erigon/core"
"github.com/ledgerwatch/erigon/core/rawdb"
"github.com/ledgerwatch/erigon/core/state"
"github.com/ledgerwatch/erigon/core/types"
"github.com/ledgerwatch/erigon/core/types/accounts"
"github.com/ledgerwatch/erigon/core/vm"
"github.com/ledgerwatch/erigon/ethdb"
"github.com/ledgerwatch/erigon/ethdb/snapshotdb"
"github.com/ledgerwatch/erigon/rlp"
"github.com/ledgerwatch/log/v3"
)
const (
AccountDiff = "accdiff"
StorageDiff = "stdiff"
ContractDiff = "contractdiff"
Deleted = "it is deleted"
)
func WithBlock(block uint64, key []byte) []byte {
b := make([]byte, 8)
binary.BigEndian.PutUint64(b, block)
return append(b, key...)
}
func TestMatreshkaStream(t *testing.T) {
t.Skip()
chaindatadir := "/media/b00ris/nvme/fresh_sync/tg/chaindata"
tmpDbDir := "/home/b00ris/event_stream"
chaindata, err := mdbx.Open(chaindatadir, log.New(), true)
if err != nil {
t.Fatal(err)
}
//tmpDb:=ethdb.NewMemDatabase()
os.RemoveAll(tmpDbDir)
db, err := mdbx.NewMDBX(log.New()).Path(tmpDbDir).WithTablessCfg(func(defaultBuckets kv.TableCfg) kv.TableCfg {
defaultBuckets[AccountDiff] = kv.TableCfgItem{}
defaultBuckets[StorageDiff] = kv.TableCfgItem{}
defaultBuckets[ContractDiff] = kv.TableCfgItem{}
return defaultBuckets
}).Open()
if err != nil {
t.Fatal(err)
}
chainConfig, _, genesisErr := core.CommitGenesisBlock(db, core.DefaultGenesisBlock())
if genesisErr != nil {
t.Fatal(err)
}
if err := db.Update(context.Background(), func(tx kv.RwTx) error {
return tx.ClearBucket(kv.HeadHeaderKey)
}); err != nil {
t.Fatal(err)
}
snkv := snapshotdb.NewSnapshotKV().DB(db).
//broken
//SnapshotDB([]string{dbutils.Headers, dbutils.HeaderCanonical, dbutils.HeaderTD, dbutils.HeaderNumber, dbutils.BlockBody, dbutils.HeadHeaderKey, dbutils.Senders}, chaindata.RwDB()).
Open()
_ = chaindata
defer snkv.Close()
tx, err := snkv.BeginRw(context.Background())
if err != nil {
t.Fatal(err)
}
defer tx.Rollback()
//
//tx, err := db.Begin(context.Background(), ethdb.RW)
//if err != nil {
// t.Fatal(err)
//}
psCursor, err := tx.Cursor(kv.PlainState)
if err != nil {
t.Fatal(err)
}
i := 5
err = ethdb.Walk(psCursor, []byte{}, 0, func(k, v []byte) (bool, error) {
fmt.Println(common.Bytes2Hex(k))
i--
if i == 0 {
return false, nil
}
return true, nil
})
if err != nil {
t.Fatal(err)
}
currentBlock := rawdb.ReadCurrentHeader(tx)
fmt.Println("currentBlock", currentBlock.Number.Uint64())
blockNum := uint64(1)
limit := currentBlock.Number.Uint64()
getHeader := func(hash common.Hash, number uint64) *types.Header { return rawdb.ReadHeader(tx, hash, number) }
stateReaderWriter := NewDebugReaderWriter(state.NewPlainStateReader(tx), state.NewPlainStateWriter(tx, tx, blockNum))
tt := time.Now()
ttt := time.Now()
for currentBlock := blockNum; currentBlock < blockNum+limit; currentBlock++ {
stateReaderWriter.UpdateWriter(state.NewPlainStateWriter(tx, tx, currentBlock))
block, err := rawdb.ReadBlockByNumber(tx, currentBlock)
if err != nil {
t.Fatal(err, currentBlock)
}
contractHasTEVM := ethdb.GetHasTEVM(tx)
_, _, err = core.ExecuteBlockEphemerally(chainConfig, &vm.Config{NoReceipts: true}, getHeader, ethash.NewFaker(), block, stateReaderWriter, stateReaderWriter, nil, nil, contractHasTEVM)
if err != nil {
t.Fatal(err, currentBlock)
}
cs := stateReaderWriter.UpdatedAccouts()
accDiffLen := len(cs)
for i := range cs {
if len(cs[i].Value) == 0 {
cs[i].Value = []byte(Deleted)
}
err = tx.Put(AccountDiff, WithBlock(currentBlock, cs[i].Key), cs[i].Value)
if err != nil {
t.Fatal(err, cs[i].Key, currentBlock)
}
}
cs = stateReaderWriter.UpdatedStorage()
stDiffLen := len(cs)
for i := range cs {
if len(cs[i].Value) == 0 {
cs[i].Value = []byte(Deleted)
}
err = tx.Put(StorageDiff, WithBlock(currentBlock, cs[i].Key), cs[i].Value)
if err != nil {
t.Fatal(err, cs[i].Key, currentBlock)
}
}
cs = stateReaderWriter.UpdatedCodes()
codesDiffLen := len(cs)
for i := range cs {
if len(cs[i].Value) == 0 {
cs[i].Value = []byte(Deleted)
}
err = tx.Put(ContractDiff, WithBlock(currentBlock, cs[i].Key), cs[i].Value)
if err != nil {
t.Fatal(err, cs[i].Key, currentBlock)
}
}
stateReaderWriter.Reset()
if currentBlock%10000 == 0 {
err = tx.Commit()
if err != nil {
t.Fatal(err, currentBlock)
}
tx, err = snkv.BeginRw(context.Background())
if err != nil {
t.Fatal(err, currentBlock)
}
defer tx.Rollback()
dr := time.Since(ttt)
fmt.Println(currentBlock, "finished", "acc-", accDiffLen, "st-", stDiffLen, "codes - ", codesDiffLen, "all -", time.Since(tt), "chunk - ", dr, "blocks/s", 10000/dr.Seconds())
ttt = time.Now()
}
}
err = tx.Commit()
if err != nil {
t.Fatal(err)
}
fmt.Println("End")
//spew.Dump("readAcc",len(stateReaderWriter.readAcc))
//spew.Dump("readStr",len(stateReaderWriter.readStorage))
//spew.Dump("createdContracts", len(stateReaderWriter.createdContracts))
//spew.Dump("deleted",len(stateReaderWriter.deletedAcc))
}
var _ state.StateReader = &DebugReaderWriter{}
var _ state.WriterWithChangeSets = &DebugReaderWriter{}
func NewDebugReaderWriter(r state.StateReader, w state.WriterWithChangeSets) *DebugReaderWriter {
return &DebugReaderWriter{
r: r,
w: w,
//readAcc: make(map[common.Address]struct{}),
//readStorage: make(map[string]struct{}),
//readCodes: make(map[common.Hash]struct{}),
//readIncarnations: make(map[common.Address]struct{}),
updatedAcc: make(map[common.Address][]byte),
updatedStorage: make(map[string][]byte),
updatedCodes: make(map[common.Hash][]byte),
//deletedAcc: make(map[common.Address]struct{}),
//createdContracts: make(map[common.Address]struct{}),
}
}
type DebugReaderWriter struct {
r state.StateReader
w state.WriterWithChangeSets
//readAcc map[common.Address]struct{}
//readStorage map[string]struct{}
//readCodes map[common.Hash] struct{}
//readIncarnations map[common.Address] struct{}
updatedAcc map[common.Address][]byte
updatedStorage map[string][]byte
updatedCodes map[common.Hash][]byte
//deletedAcc map[common.Address]struct{}
//createdContracts map[common.Address]struct{}
}
func (d *DebugReaderWriter) Reset() {
d.updatedAcc = map[common.Address][]byte{}
d.updatedStorage = map[string][]byte{}
d.updatedCodes = map[common.Hash][]byte{}
}
func (d *DebugReaderWriter) UpdateWriter(w state.WriterWithChangeSets) {
d.w = w
}
func (d *DebugReaderWriter) ReadAccountData(address common.Address) (*accounts.Account, error) {
//d.readAcc[address] = struct{}{}
return d.r.ReadAccountData(address)
}
func (d *DebugReaderWriter) ReadAccountStorage(address common.Address, incarnation uint64, key *common.Hash) ([]byte, error) {
//d.readStorage[string(dbutils.PlainGenerateCompositeStorageKey(address.Bytes(),incarnation, key.Bytes()))] = struct{}{}
return d.r.ReadAccountStorage(address, incarnation, key)
}
func (d *DebugReaderWriter) ReadAccountCode(address common.Address, incarnation uint64, codeHash common.Hash) ([]byte, error) {
//d.readCodes[codeHash] = struct{}{}
return d.r.ReadAccountCode(address, incarnation, codeHash)
}
func (d *DebugReaderWriter) ReadAccountCodeSize(address common.Address, incarnation uint64, codeHash common.Hash) (int, error) {
return d.r.ReadAccountCodeSize(address, incarnation, codeHash)
}
func (d *DebugReaderWriter) ReadAccountIncarnation(address common.Address) (uint64, error) {
//d.readIncarnations[address] = struct{}{}
return d.r.ReadAccountIncarnation(address)
}
func (d *DebugReaderWriter) WriteChangeSets() error {
return d.w.WriteChangeSets()
}
func (d *DebugReaderWriter) WriteHistory() error {
return d.w.WriteHistory()
}
func (d *DebugReaderWriter) UpdateAccountData(address common.Address, original, account *accounts.Account) error {
b, err := rlp.EncodeToBytes(account)
if err != nil {
return err
}
d.updatedAcc[address] = b
return d.w.UpdateAccountData(address, original, account)
}
func (d *DebugReaderWriter) UpdateAccountCode(address common.Address, incarnation uint64, codeHash common.Hash, code []byte) error {
d.updatedCodes[codeHash] = code
return d.w.UpdateAccountCode(address, incarnation, codeHash, code)
}
func (d *DebugReaderWriter) DeleteAccount(address common.Address, original *accounts.Account) error {
d.updatedAcc[address] = nil
//d.deletedAcc[address]= struct{}{}
return d.w.DeleteAccount(address, original)
}
func (d *DebugReaderWriter) WriteAccountStorage(address common.Address, incarnation uint64, key *common.Hash, original, value *uint256.Int) error {
d.updatedStorage[string(dbutils.PlainGenerateCompositeStorageKey(address.Bytes(), incarnation, key.Bytes()))] = value.Bytes()
return d.w.WriteAccountStorage(address, incarnation, key, original, value)
}
func (d *DebugReaderWriter) CreateContract(address common.Address) error {
//d.createdContracts[address] = struct{}{}
return d.w.CreateContract(address)
}
type Change struct {
Key []byte
Value []byte
}
func (d *DebugReaderWriter) UpdatedAccouts() []Change {
ch := make([]Change, 0, len(d.updatedAcc))
for k, v := range d.updatedAcc {
ch = append(ch, Change{
Key: common.CopyBytes(k.Bytes()),
Value: common.CopyBytes(v),
})
}
return ch
}
func (d *DebugReaderWriter) UpdatedStorage() []Change {
ch := make([]Change, 0, len(d.updatedStorage))
for k, v := range d.updatedStorage {
ch = append(ch, Change{
Key: common.CopyBytes([]byte(k)),
Value: common.CopyBytes(v),
})
}
return ch
}
func (d *DebugReaderWriter) UpdatedCodes() []Change {
ch := make([]Change, 0, len(d.updatedCodes))
for k, v := range d.updatedCodes {
ch = append(ch, Change{
Key: common.CopyBytes(k.Bytes()),
Value: common.CopyBytes(v),
})
}
return ch
}
//func (d *DebugReaderWriter) AllAccounts() map[common.Address]struct{} {
// accs:=make(map[common.Address]struct{})
// for i:=range d.readAcc {
// accs[i]=struct{}{}
// }
// for i:=range d.updatedAcc {
// accs[i]=struct{}{}
// }
// for i:=range d.readIncarnations {
// accs[i]=struct{}{}
// }
// for i:=range d.deletedAcc {
// accs[i]=struct{}{}
// }
// for i:=range d.createdContracts {
// accs[i]=struct{}{}
// }
// return accs
//}
//func (d *DebugReaderWriter) AllStorage() map[string]struct{} {
// st:=make(map[string]struct{})
// for i:=range d.readStorage {
// st[i]=struct{}{}
// }
// for i:=range d.updatedStorage {
// st[i]=struct{}{}
// }
// return st
//}
//func (d *DebugReaderWriter) AllCodes() map[common.Hash]struct{} {
// c:=make(map[common.Hash]struct{})
// for i:=range d.readCodes {
// c[i]=struct{}{}
// }
// for i:=range d.updatedCodes {
// c[i]=struct{}{}
// }
// return c
//}

View File

@ -243,11 +243,11 @@ func AddTorrentFiles(ctx context.Context, snapshotsDir *dir.Rw, torrentClient *t
// ResolveAbsentTorrents - add hard-coded hashes (if client doesn't have) as magnet links and download everything
func ResolveAbsentTorrents(ctx context.Context, torrentClient *torrent.Client, preverifiedHashes []metainfo.Hash, snapshotDir *dir.Rw, silent bool) error {
mi := &metainfo.MetaInfo{AnnounceList: Trackers}
for _, infoHash := range preverifiedHashes {
if _, ok := torrentClient.Torrent(infoHash); ok {
for i := range preverifiedHashes {
if _, ok := torrentClient.Torrent(preverifiedHashes[i]); ok {
continue
}
magnet := mi.Magnet(&infoHash, nil)
magnet := mi.Magnet(&preverifiedHashes[i], nil)
t, err := torrentClient.AddMagnet(magnet.String())
if err != nil {
return err

View File

@ -3,7 +3,7 @@ package downloader
import (
"bytes"
"context"
"crypto/sha1"
"crypto/sha1" //nolint:gosec
"errors"
"fmt"
"io"
@ -206,7 +206,7 @@ func verifyTorrent(info *metainfo.Info, root string, consumer func(i int, good b
span.InitIndex()
for i, numPieces := 0, info.NumPieces(); i < numPieces; i += 1 {
p := info.Piece(i)
hash := sha1.New()
hash := sha1.New() //nolint:gosec
_, err := io.Copy(hash, io.NewSectionReader(span, p.Offset(), p.Length()))
if err != nil {
return err

View File

@ -403,7 +403,6 @@ func retireBlocks(s *PruneState, tx kv.RwTx, cfg SendersCfg, ctx context.Context
}
//TODO: avoid too large deletes
log.Info("[snapshots] Retire blocks", "from", blockFrom, "to", blockTo)
chainID, _ := uint256.FromBig(cfg.chainConfig.ChainID)
wg := sync.WaitGroup{}
wg.Add(1)
@ -432,6 +431,5 @@ func retireBlocks(s *PruneState, tx kv.RwTx, cfg SendersCfg, ctx context.Context
defer wg.Done()
}()
wg.Wait()
fmt.Printf("sn runtime dump: %d-%d\n", blockFrom, blockTo)
return nil
}

File diff suppressed because it is too large Load Diff

File diff suppressed because it is too large Load Diff

2
go.mod
View File

@ -40,7 +40,7 @@ require (
github.com/json-iterator/go v1.1.12
github.com/julienschmidt/httprouter v1.3.0
github.com/kevinburke/go-bindata v3.21.0+incompatible
github.com/ledgerwatch/erigon-lib v0.0.0-20220310121515-3123b6d895c5
github.com/ledgerwatch/erigon-lib v0.0.0-20220312093458-c1f1365f9224
github.com/ledgerwatch/log/v3 v3.4.1
github.com/ledgerwatch/secp256k1 v1.0.0
github.com/niemeyer/pretty v0.0.0-20200227124842-a10e7caefd8e // indirect

4
go.sum
View File

@ -641,8 +641,8 @@ github.com/kylelemons/godebug v0.0.0-20170224010052-a616ab194758 h1:0D5M2HQSGD3P
github.com/kylelemons/godebug v0.0.0-20170224010052-a616ab194758/go.mod h1:B69LEHPfb2qLo0BaaOLcbitczOKLWTsrBG9LczfCD4k=
github.com/leanovate/gopter v0.2.9 h1:fQjYxZaynp97ozCzfOyOuAGOU4aU/z37zf/tOujFk7c=
github.com/leanovate/gopter v0.2.9/go.mod h1:U2L/78B+KVFIx2VmW6onHJQzXtFb+p5y3y2Sh+Jxxv8=
github.com/ledgerwatch/erigon-lib v0.0.0-20220310121515-3123b6d895c5 h1:f81JYvbRxP0T//AG+wmfZNxiwz2mDPK1cVJuhwcAiYc=
github.com/ledgerwatch/erigon-lib v0.0.0-20220310121515-3123b6d895c5/go.mod h1:mag5WaGOUTVOLvFkT4wpjR5YHMmm4hynWJ3YfQ44Elg=
github.com/ledgerwatch/erigon-lib v0.0.0-20220312093458-c1f1365f9224 h1:GqWMSF+pwVuW7cAosXMOzH/h3sIrEQ3TM4Ivwdl+D4M=
github.com/ledgerwatch/erigon-lib v0.0.0-20220312093458-c1f1365f9224/go.mod h1:mag5WaGOUTVOLvFkT4wpjR5YHMmm4hynWJ3YfQ44Elg=
github.com/ledgerwatch/log/v3 v3.4.1 h1:/xGwlVulXnsO9Uq+tzaExc8OWmXXHU0dnLalpbnY5Bc=
github.com/ledgerwatch/log/v3 v3.4.1/go.mod h1:VXcz6Ssn6XEeU92dCMc39/g1F0OYAjw1Mt+dGP5DjXY=
github.com/ledgerwatch/secp256k1 v1.0.0 h1:Usvz87YoTG0uePIV8woOof5cQnLXGYa162rFf3YnwaQ=

View File

@ -72,6 +72,7 @@ var snapshotCommand = cli.Command{
utils.DataDirFlag,
SnapshotFromFlag,
SnapshotToFlag,
SnapshotEveryFlag,
}, debug.Flags...),
},
},
@ -88,6 +89,11 @@ var (
Usage: "To block number. Zero - means unlimited.",
Value: 0,
}
SnapshotEveryFlag = cli.Uint64Flag{
Name: "every",
Usage: "Do operation every N blocks",
Value: 1_000,
}
SnapshotSegmentSizeFlag = cli.Uint64Flag{
Name: "segment.size",
Usage: "Amount of blocks in each segment",
@ -135,8 +141,9 @@ func doRetireCommand(cliCtx *cli.Context) error {
tmpDir := filepath.Join(datadir, etl.TmpDirName)
from := cliCtx.Uint64(SnapshotFromFlag.Name)
to := cliCtx.Uint64(SnapshotToFlag.Name)
every := cliCtx.Uint64(SnapshotEveryFlag.Name)
chainDB := mdbx.NewMDBX(log.New()).Path(path.Join(datadir, "chaindata")).Readonly().MustOpen()
chainDB := mdbx.NewMDBX(log.New()).Label(kv.ChainDB).Path(path.Join(datadir, "chaindata")).Readonly().MustOpen()
defer chainDB.Close()
cfg := ethconfig.NewSnapshotCfg(true, true)
@ -148,11 +155,14 @@ func doRetireCommand(cliCtx *cli.Context) error {
chainConfig := tool.ChainConfigFromDB(chainDB)
chainID, _ := uint256.FromBig(chainConfig.ChainID)
snapshots := snapshotsync.NewRoSnapshots(cfg, snapshotDir)
snapshots.ReopenSegments()
if err := snapshotsync.RetireBlocks(ctx, from, to, *chainID, tmpDir, snapshots, chainDB, 1, log.LvlInfo); err != nil {
for i := from; i < to; i += every {
if err := snapshotsync.RetireBlocks(ctx, i, i+every, *chainID, tmpDir, snapshots, chainDB, runtime.NumCPU()/2, log.LvlInfo); err != nil {
panic(err)
//return err
}
}
return nil
}
@ -171,7 +181,7 @@ func doSnapshotCommand(cliCtx *cli.Context) error {
tmpDir := filepath.Join(datadir, etl.TmpDirName)
dir.MustExist(tmpDir)
chainDB := mdbx.NewMDBX(log.New()).Path(filepath.Join(datadir, "chaindata")).Readonly().MustOpen()
chainDB := mdbx.NewMDBX(log.New()).Label(kv.ChainDB).Path(filepath.Join(datadir, "chaindata")).Readonly().MustOpen()
defer chainDB.Close()
if err := snapshotBlocks(ctx, chainDB, fromBlock, toBlock, segmentSize, snapshotDir, tmpDir); err != nil {

View File

@ -666,7 +666,7 @@ func min(a, b uint64) uint64 {
}
func RetireBlocks(ctx context.Context, blockFrom, blockTo uint64, chainID uint256.Int, tmpDir string, snapshots *RoSnapshots, db kv.RoDB, workers int, lvl log.Lvl) error {
log.Log(lvl, "[snapshots] Retire Blocks", "from", blockFrom, "to", blockTo)
log.Log(lvl, "[snapshots] Retire Blocks", "range", fmt.Sprintf("%dk-%dk", blockFrom/1000, blockTo/1000))
// in future we will do it in background
if err := DumpBlocks(ctx, blockFrom, blockTo, DEFAULT_SEGMENT_SIZE, tmpDir, snapshots.Dir(), db, workers, lvl); err != nil {
return fmt.Errorf("DumpBlocks: %w", err)
@ -679,7 +679,6 @@ func RetireBlocks(ctx context.Context, blockFrom, blockTo uint64, chainID uint25
if len(ranges) == 0 {
return nil
}
if err := merger.Merge(ctx, snapshots, ranges, &dir.Rw{Path: snapshots.Dir()}); err != nil {
return err
}
@ -878,7 +877,8 @@ func DumpTxs(ctx context.Context, db kv.RoDB, segmentFile, tmpDir string, blockF
}
_, fileName := filepath.Split(segmentFile)
log.Log(lvl, "[snapshots] Compression", "ratio", f.Ratio.String(), "file", fileName)
ext := filepath.Ext(fileName)
log.Log(lvl, "[snapshots] Compression", "ratio", f.Ratio.String(), "file", fileName[:len(fileName)-len(ext)])
return firstTxID, nil
}
@ -1427,6 +1427,8 @@ type mergeRange struct {
from, to uint64
}
func (r mergeRange) String() string { return fmt.Sprintf("%dk-%dk", r.from/1000, r.to/1000) }
func (*Merger) FindMergeRanges(snapshots *RoSnapshots) (res []mergeRange) {
for i := len(snapshots.blocks) - 1; i > 0; i-- {
sn := snapshots.blocks[i]
@ -1469,6 +1471,7 @@ func (m *Merger) filesByRange(snapshots *RoSnapshots, from, to uint64) (toMergeH
}
func (m *Merger) Merge(ctx context.Context, snapshots *RoSnapshots, mergeRanges []mergeRange, snapshotDir *dir.Rw) error {
log.Log(m.lvl, "[snapshots] Merge segments", "ranges", fmt.Sprintf("%v", mergeRanges))
for _, r := range mergeRanges {
toMergeHeaders, toMergeBodies, toMergeTxs := m.filesByRange(snapshots, r.from, r.to)
if err := m.merge(ctx, toMergeBodies, filepath.Join(snapshotDir.Path, SegmentFileName(r.from, r.to, Bodies))); err != nil {
@ -1497,8 +1500,6 @@ func (m *Merger) merge(ctx context.Context, toMerge []string, targetFile string)
_, fName := filepath.Split(f)
fileNames[i] = fName
}
_, fName := filepath.Split(targetFile)
log.Log(m.lvl, "[snapshots] Merging", "files", fileNames, "to", fName)
f, err := compress.NewCompressor(ctx, "merge", targetFile, m.tmpDir, compress.MinPatternScore, m.workers)
if err != nil {
return err
@ -1518,6 +1519,11 @@ func (m *Merger) merge(ctx context.Context, toMerge []string, targetFile string)
if err := f.AddWord(word); err != nil {
return err
}
select {
case <-ctx.Done():
return ctx.Err()
default:
}
}
return nil
}); err != nil {

View File

@ -1,60 +0,0 @@
package snapshotsync
/*
import (
"github.com/ledgerwatch/erigon-lib/gointerfaces/snapshotsync"
"github.com/ledgerwatch/erigon-lib/kv"
kv2 "github.com/ledgerwatch/erigon-lib/kv/mdbx"
"github.com/ledgerwatch/erigon/ethdb/snapshotdb"
"github.com/ledgerwatch/log/v3"
)
var (
BucketConfigs = map[snapshotsync.Type]kv.TableCfg{
snapshotsync.SnapshotType_bodies: {
kv.BlockBody: kv.TableCfgItem{},
kv.EthTx: kv.TableCfgItem{},
},
snapshotsync.SnapshotType_headers: {
kv.Headers: kv.TableCfgItem{},
},
snapshotsync.SnapshotType_state: {
kv.PlainState: kv.TableCfgItem{
Flags: kv.DupSort,
AutoDupSortKeysConversion: true,
DupFromLen: 60,
DupToLen: 28,
},
kv.PlainContractCode: kv.TableCfgItem{},
kv.Code: kv.TableCfgItem{},
},
}
)
func WrapBySnapshotsFromDownloader(db kv.RwDB, snapshots map[snapshotsync.Type]*snapshotsync.SnapshotsInfo) (kv.RwDB, error) {
snKV := snapshotdb.NewSnapshotKV().DB(db)
for k, v := range snapshots {
log.Info("Wrap db by", "snapshot", k.String(), "dir", v.Dbpath)
chainSnapshotCfg := BucketConfigs[k]
snapshotKV, err := kv2.NewMDBX(log.New()).Readonly().Path(v.Dbpath).WithTablessCfg(func(defaultBuckets kv.TableCfg) kv.TableCfg {
return chainSnapshotCfg
}).Open()
if err != nil {
log.Error("Can't open snapshot", "err", err)
return nil, err
} else { //nolint
switch k {
case snapshotsync.SnapshotType_headers:
snKV = snKV.HeadersSnapshot(snapshotKV)
case snapshotsync.SnapshotType_bodies:
snKV = snKV.BodiesSnapshot(snapshotKV)
case snapshotsync.SnapshotType_state:
snKV = snKV.StateSnapshot(snapshotKV)
}
}
}
return snKV.Open(), nil
}
*/