erigon-pulse/trie/trie_root.go

871 lines
24 KiB
Go

package trie
import (
"bytes"
"context"
"encoding/binary"
"fmt"
"time"
"github.com/ledgerwatch/turbo-geth/common"
"github.com/ledgerwatch/turbo-geth/common/dbutils"
"github.com/ledgerwatch/turbo-geth/core/types/accounts"
"github.com/ledgerwatch/turbo-geth/ethdb"
"github.com/ledgerwatch/turbo-geth/log"
"github.com/ledgerwatch/turbo-geth/trie/rlphacks"
)
/*
**Theoretically:** "Merkle trie root calculation" starts from state, build from state keys - trie,
on each level of trie calculates intermediate hash of underlying data.
**Practically:** It can be implemented as "Preorder trie traversal" (Preorder - visit Root, visit Left, visit Right).
But, let's make couple observations to make traversal over huge state efficient.
**Observation 1:** `CurrentStateBucket` already stores state keys in sorted way.
Iteration over this bucket will retrieve keys in same order as "Preorder trie traversal".
**Observation 2:** each Eth block - changes not big part of state - it means most of Merkle trie intermediate hashes will not change.
It means we effectively can cache them. `IntermediateTrieHashBucket` stores "Intermediate hashes of all Merkle trie levels".
It also sorted and Iteration over `IntermediateTrieHashBucket` will retrieve keys in same order as "Preorder trie traversal".
**Implementation:** by opening 1 Cursor on state and 1 more Cursor on intermediate hashes bucket - we will receive data in
order of "Preorder trie traversal". Cursors will only do "sequential reads" and "jumps forward" - been hardware-friendly.
1 stack keeps all accumulated hashes, when sub-trie traverse ends - all hashes pulled from stack -> hashed -> new hash puts on stack - it's hash of visited sub-trie (it emulates recursive nature of "Preorder trie traversal" algo).
Imagine that account with key 0000....00 (64 zeroes, 32 bytes of zeroes) changed.
Here is an example sequence which can be seen by running 2 Cursors:
```
00 // key which came from cache, can't use it - because account with this prefix changed
0000 // key which came from cache, can't use it - because account with this prefix changed
...
{30 zero bytes}00 // key which came from cache, can't use it - because account with this prefix changed
{30 zero bytes}0000 // Account which came from state, use it - calculate hash, jump to "next sub-trie"
{30 zero bytes}01 // key which came from cache, it is "next sub-trie", use it, jump to "next sub-trie"
{30 zero bytes}02 // key which came from cache, it is "next sub-trie", use it, jump to "next sub-trie"
...
{30 zero bytes}ff // key which came from cache, it is "next sub-trie", use it, jump to "next sub-trie"
{29 zero bytes}01 // key which came from cache, it is "next sub-trie" (1 byte shorter key), use it, jump to "next sub-trie"
{29 zero bytes}02 // key which came from cache, it is "next sub-trie" (1 byte shorter key), use it, jump to "next sub-trie"
...
ff // key which came from cache, it is "next sub-trie" (1 byte shorter key), use it, jump to "next sub-trie"
nil // db returned nil - means no more keys there, done
```
On practice Trie is no full - it means after account key `{30 zero bytes}0000` may come `{5 zero bytes}01` and amount of iterations will not be big.
### Attack - by delete account with huge state
It's possible to create Account with very big storage (increase storage size during many blocks).
Then delete this account (SELFDESTRUCT).
Naive storage deletion may take several minutes - depends on Disk speed - means every Eth client
will not process any incoming block that time. To protect against this attack:
PlainState, HashedState and IntermediateTrieHash buckets have "incarnations". Account entity has field "Incarnation" -
just a digit which increasing each SELFDESTRUCT or CREATE2 opcodes. Storage key formed by:
`{account_key}{incarnation}{storage_hash}`. And [trie/trie_root.go](../../trie/trie_root.go) has logic - every time
when Account visited - we save it to `accAddrHashWithInc` variable and skip any Storage or IntermediateTrieHashes with another incarnation.
*/
// FlatDBTrieLoader reads state and intermediate trie hashes in order equal to "Preorder trie traversal"
// (Preorder - visit Root, visit Left, visit Right)
//
// It produces stream of values and send this stream to `defaultReceiver`
// It skips storage with incorrect incarnations
//
// Each intermediate hash key firstly pass to RetainDecider, only if it returns "false" - such IH can be used.
type FlatDBTrieLoader struct {
trace bool
itemPresent bool
itemType StreamItem
stateBucket string
intermediateHashesBucket string
rd RetainDecider
accAddrHashWithInc [40]byte // Concatenation of addrHash of the currently build account with its incarnation encoding
nextAccountKey [32]byte
k, v []byte
ihK, ihV []byte
// Storage item buffer
storageKey []byte
storageValue []byte
// Account item buffer
accountKey []byte
accountValue accounts.Account
hashValue []byte
receiver StreamReceiver
defaultReceiver *RootHashAggregator
hc HashCollector
}
// RootHashAggregator - calculates Merkle trie root hash from incoming data stream
type RootHashAggregator struct {
trace bool
wasIH bool
wasIHStorage bool
root common.Hash
hc HashCollector
currStorage bytes.Buffer // Current key for the structure generation algorithm, as well as the input tape for the hash builder
succStorage bytes.Buffer
valueStorage []byte // Current value to be used as the value tape for the hash builder
curr bytes.Buffer // Current key for the structure generation algorithm, as well as the input tape for the hash builder
succ bytes.Buffer
value []byte // Current value to be used as the value tape for the hash builder
groups []uint16 // `groups` parameter is the map of the stack. each element of the `groups` slice is a bitmask, one bit per element currently on the stack. See `GenStructStep` docs
hb *HashBuilder
hashData GenStructStepHashData
a accounts.Account
leafData GenStructStepLeafData
accData GenStructStepAccountData
}
func NewRootHashAggregator() *RootHashAggregator {
return &RootHashAggregator{
hb: NewHashBuilder(false),
}
}
func NewFlatDBTrieLoader(stateBucket, intermediateHashesBucket string) *FlatDBTrieLoader {
return &FlatDBTrieLoader{
defaultReceiver: NewRootHashAggregator(),
stateBucket: stateBucket,
intermediateHashesBucket: intermediateHashesBucket,
}
}
// Reset prepares the loader for reuse
func (l *FlatDBTrieLoader) Reset(rd RetainDecider, hc HashCollector, trace bool) error {
l.defaultReceiver.Reset(hc, trace)
l.hc = hc
l.receiver = l.defaultReceiver
l.trace = trace
l.rd = rd
l.itemPresent = false
if l.trace {
fmt.Printf("----------\n")
fmt.Printf("CalcTrieRoot\n")
}
return nil
}
func (l *FlatDBTrieLoader) SetStreamReceiver(receiver StreamReceiver) {
l.receiver = receiver
}
// iteration moves through the database buckets and creates at most
// one stream item, which is indicated by setting the field fstl.itemPresent to true
func (l *FlatDBTrieLoader) iteration(c ethdb.Cursor, ih *IHCursor, first bool) error {
var isIH, isIHSequence bool
var err error
if first {
if l.ihK, l.ihV, isIHSequence, err = ih.Seek([]byte{}); err != nil {
return err
}
if isIHSequence {
l.k = l.ihK
return nil
}
if l.k, l.v, err = c.Seek([]byte{}); err != nil {
return err
}
// Skip wrong incarnation
if len(l.k) > common.HashLength {
if nextAccount(l.k, l.nextAccountKey[:]) {
if l.k, l.v, err = c.Seek(l.nextAccountKey[:]); err != nil {
return err
}
} else {
l.k = nil
}
}
return nil
}
if l.ihK == nil && l.k == nil { // loop termination
l.itemPresent = true
l.itemType = CutoffStreamItem
l.accountKey = nil
l.storageKey = nil
l.storageValue = nil
l.hashValue = nil
return nil
}
isIH, _ = keyIsBeforeOrEqual(l.ihK, l.k)
if !isIH {
// skip wrong incarnation
if len(l.k) > common.HashLength && !bytes.HasPrefix(l.k, l.accAddrHashWithInc[:]) {
if bytes.Compare(l.k, l.accAddrHashWithInc[:]) < 0 {
// Skip all the irrelevant storage in the middle
if l.k, l.v, err = c.Seek(l.accAddrHashWithInc[:]); err != nil {
return err
}
} else {
if nextAccount(l.k, l.nextAccountKey[:]) {
if l.k, l.v, err = c.Seek(l.nextAccountKey[:]); err != nil {
return err
}
} else {
l.k = nil
}
}
return nil
}
l.itemPresent = true
if len(l.k) > common.HashLength {
l.itemType = StorageStreamItem
l.accountKey = nil
l.storageKey = l.k // no reason to copy, because this "pointer and data" will valid until end of transaction
l.hashValue = nil
l.storageValue = l.v
if l.k, l.v, err = c.Next(); err != nil {
return err
}
if l.trace {
fmt.Printf("k after storageWalker and Next: %x\n", l.k)
}
} else if len(l.k) > 0 {
l.itemType = AccountStreamItem
l.accountKey = l.k
l.storageKey = nil
l.storageValue = nil
l.hashValue = nil
if err = l.accountValue.DecodeForStorage(l.v); err != nil {
return fmt.Errorf("fail DecodeForStorage: %w", err)
}
copy(l.accAddrHashWithInc[:], l.k)
binary.BigEndian.PutUint64(l.accAddrHashWithInc[32:], l.accountValue.Incarnation)
// Now we know the correct incarnation of the account, and we can skip all irrelevant storage records
// Since 0 incarnation if 0xfff...fff, and we do not expect any records like that, this automatically
// skips over all storage items
if l.k, l.v, err = c.Seek(l.accAddrHashWithInc[:]); err != nil {
return err
}
if l.trace {
fmt.Printf("k after accountWalker and Seek: %x\n", l.k)
}
if keyIsBefore(l.ihK, l.accAddrHashWithInc[:]) {
if l.ihK, l.ihV, _, err = ih.Seek(l.accAddrHashWithInc[:]); err != nil {
return err
}
}
}
return nil
}
// ih part
if l.trace {
fmt.Printf("l.ihK %x, l.accAddrHashWithInc %x\n", l.ihK, l.accAddrHashWithInc[:])
}
// Skip IH with wrong incarnation
if len(l.ihK) > common.HashLength && !bytes.HasPrefix(l.ihK, l.accAddrHashWithInc[:]) {
if bytes.Compare(l.ihK, l.accAddrHashWithInc[:]) < 0 {
// Skip all the irrelevant storage in the middle
if l.ihK, l.ihV, _, err = ih.Seek(l.accAddrHashWithInc[:]); err != nil {
return err
}
} else {
if nextAccount(l.ihK, l.nextAccountKey[:]) {
if l.ihK, l.ihV, _, err = ih.Seek(l.nextAccountKey[:]); err != nil {
return err
}
} else {
l.ihK = nil
}
}
return nil
}
l.itemPresent = true
if len(l.ihK) > common.HashLength {
l.itemType = SHashStreamItem
l.accountKey = nil
l.storageKey = l.ihK
l.hashValue = l.ihV
l.storageValue = nil
} else {
l.itemType = AHashStreamItem
l.accountKey = l.ihK
l.storageKey = nil
l.storageValue = nil
l.hashValue = l.ihV
}
// skip subtree
next, ok := dbutils.NextSubtree(l.ihK)
if !ok { // no siblings left
l.k, l.ihK, l.ihV = nil, nil, nil
return nil
}
if l.trace {
fmt.Printf("next: %x\n", next)
}
if l.ihK, l.ihV, isIHSequence, err = ih.Seek(next); err != nil {
return err
}
if isIHSequence {
l.k = l.ihK
return nil
}
if l.k, l.v, err = c.Seek(next); err != nil {
return err
}
// Skip wrong incarnation
if len(next) <= common.HashLength && len(l.k) > common.HashLength {
// Advance past the storage to the first account
if nextAccount(l.k, l.nextAccountKey[:]) {
if l.k, l.v, err = c.Seek(l.nextAccountKey[:]); err != nil {
return err
}
} else {
l.k = nil
}
}
if l.trace {
fmt.Printf("k after next: %x\n", l.k)
}
return nil
}
// CalcTrieRoot - spawn 2 cursors (IntermediateHashes and HashedState)
// Wrap IntermediateHashes cursor to IH class - this class will return only keys which passed RetainDecider check
// If RetainDecider check not passed, then such key must be deleted - HashCollector receiving nil for such key.
func (l *FlatDBTrieLoader) CalcTrieRoot(db ethdb.Database, quit <-chan struct{}) (common.Hash, error) {
var (
tx ethdb.Tx
kv ethdb.KV
)
// If method executed within transaction - use it, or open new read transaction
if hasTx, ok := db.(ethdb.HasTx); ok && hasTx.Tx() != nil {
tx = hasTx.Tx()
} else {
if hasKV, ok := db.(ethdb.HasKV); ok {
kv = hasKV.KV()
var err error
tx, err = kv.Begin(context.Background(), nil, false)
if err != nil {
return EmptyRoot, err
}
defer tx.Rollback()
} else {
return EmptyRoot, fmt.Errorf("database doest not implement KV: %T", db)
}
}
c := tx.Cursor(l.stateBucket)
var filter = func(k []byte) (bool, error) {
if l.rd.Retain(k) {
if l.hc != nil {
if err := l.hc(k, nil); err != nil {
return false, err
}
}
return false, nil
}
return true, nil
}
ih := IH(Filter(filter, tx.Cursor(l.intermediateHashesBucket)))
if err := l.iteration(c, ih, true /* first */); err != nil {
return EmptyRoot, err
}
logEvery := time.NewTicker(30 * time.Second)
defer logEvery.Stop()
for l.itemType != CutoffStreamItem {
if err := common.Stopped(quit); err != nil {
return EmptyRoot, err
}
for !l.itemPresent {
if err := l.iteration(c, ih, false /* first */); err != nil {
return EmptyRoot, err
}
}
if err := l.receiver.Receive(l.itemType, l.accountKey, l.storageKey, &l.accountValue, l.storageValue, l.hashValue, 0); err != nil {
return EmptyRoot, err
}
l.itemPresent = false
select {
default:
case <-logEvery.C:
l.logProgress()
}
}
return l.receiver.Root(), nil
}
func (l *FlatDBTrieLoader) logProgress() {
var k string
if l.accountKey != nil {
k = makeCurrentKeyStr(l.accountKey)
} else {
k = makeCurrentKeyStr(l.ihK)
}
log.Info("Calculating Merkle root", "current key", k)
}
func (r *RootHashAggregator) RetainNothing(prefix []byte) bool {
return false
}
func (r *RootHashAggregator) Reset(hc HashCollector, trace bool) {
r.hc = hc
r.curr.Reset()
r.succ.Reset()
r.value = nil
r.groups = r.groups[:0]
r.a.Reset()
r.hb.Reset()
r.wasIH = false
r.currStorage.Reset()
r.succStorage.Reset()
r.valueStorage = nil
r.wasIHStorage = false
r.root = common.Hash{}
r.trace = trace
r.hb.trace = trace
}
func (r *RootHashAggregator) Receive(itemType StreamItem,
accountKey []byte,
storageKey []byte,
accountValue *accounts.Account,
storageValue []byte,
hash []byte,
cutoff int,
) error {
switch itemType {
case StorageStreamItem:
r.advanceKeysStorage(storageKey, true /* terminator */)
if r.currStorage.Len() > 0 {
if err := r.genStructStorage(); err != nil {
return err
}
}
r.saveValueStorage(false, storageValue, hash)
case SHashStreamItem:
r.advanceKeysStorage(storageKey, false /* terminator */)
if r.currStorage.Len() > 0 {
if err := r.genStructStorage(); err != nil {
return err
}
}
r.saveValueStorage(true, storageValue, hash)
case AccountStreamItem:
r.advanceKeysAccount(accountKey, true /* terminator */)
if r.curr.Len() > 0 && !r.wasIH {
r.cutoffKeysStorage(2 * (common.HashLength + common.IncarnationLength))
if r.currStorage.Len() > 0 {
if err := r.genStructStorage(); err != nil {
return err
}
}
if r.currStorage.Len() > 0 {
if len(r.groups) >= 2*common.HashLength {
r.groups = r.groups[:2*common.HashLength-1]
}
for len(r.groups) > 0 && r.groups[len(r.groups)-1] == 0 {
r.groups = r.groups[:len(r.groups)-1]
}
r.currStorage.Reset()
r.succStorage.Reset()
r.wasIHStorage = false
// There are some storage items
r.accData.FieldSet |= AccountFieldStorageOnly
}
}
if r.curr.Len() > 0 {
if err := r.genStructAccount(); err != nil {
return err
}
}
if err := r.saveValueAccount(false, accountValue, hash); err != nil {
return err
}
case AHashStreamItem:
r.advanceKeysAccount(accountKey, false /* terminator */)
if r.curr.Len() > 0 && !r.wasIH {
r.cutoffKeysStorage(2 * (common.HashLength + common.IncarnationLength))
if r.currStorage.Len() > 0 {
if err := r.genStructStorage(); err != nil {
return err
}
}
if r.currStorage.Len() > 0 {
if len(r.groups) >= 2*common.HashLength {
r.groups = r.groups[:2*common.HashLength-1]
}
for len(r.groups) > 0 && r.groups[len(r.groups)-1] == 0 {
r.groups = r.groups[:len(r.groups)-1]
}
r.currStorage.Reset()
r.succStorage.Reset()
r.wasIHStorage = false
// There are some storage items
r.accData.FieldSet |= AccountFieldStorageOnly
}
}
if r.curr.Len() > 0 {
if err := r.genStructAccount(); err != nil {
return err
}
}
if err := r.saveValueAccount(true, accountValue, hash); err != nil {
return err
}
case CutoffStreamItem:
if r.trace {
fmt.Printf("storage cuttoff %d\n", cutoff)
}
r.cutoffKeysAccount(cutoff)
if r.curr.Len() > 0 && !r.wasIH {
r.cutoffKeysStorage(2 * (common.HashLength + common.IncarnationLength))
if r.currStorage.Len() > 0 {
if err := r.genStructStorage(); err != nil {
return err
}
}
if r.currStorage.Len() > 0 {
if len(r.groups) >= 2*common.HashLength {
r.groups = r.groups[:2*common.HashLength-1]
}
for len(r.groups) > 0 && r.groups[len(r.groups)-1] == 0 {
r.groups = r.groups[:len(r.groups)-1]
}
r.currStorage.Reset()
r.succStorage.Reset()
r.wasIHStorage = false
// There are some storage items
r.accData.FieldSet |= AccountFieldStorageOnly
}
}
if r.curr.Len() > 0 {
if err := r.genStructAccount(); err != nil {
return err
}
}
if r.curr.Len() > 0 {
if len(r.groups) > cutoff {
r.groups = r.groups[:cutoff]
}
for len(r.groups) > 0 && r.groups[len(r.groups)-1] == 0 {
r.groups = r.groups[:len(r.groups)-1]
}
}
if r.hb.hasRoot() {
r.root = r.hb.rootHash()
} else {
r.root = EmptyRoot
}
r.groups = r.groups[:0]
r.hb.Reset()
r.wasIH = false
r.wasIHStorage = false
r.curr.Reset()
r.succ.Reset()
r.currStorage.Reset()
r.succStorage.Reset()
}
return nil
}
func (r *RootHashAggregator) Result() SubTries {
panic("don't call me")
}
func (r *RootHashAggregator) Root() common.Hash {
return r.root
}
func (r *RootHashAggregator) advanceKeysStorage(k []byte, terminator bool) {
r.currStorage.Reset()
r.currStorage.Write(r.succStorage.Bytes())
r.succStorage.Reset()
// Transform k to nibbles, but skip the incarnation part in the middle
keyToNibbles(k, &r.succStorage)
if terminator {
r.succStorage.WriteByte(16)
}
}
func (r *RootHashAggregator) cutoffKeysStorage(cutoff int) {
r.currStorage.Reset()
r.currStorage.Write(r.succStorage.Bytes())
r.succStorage.Reset()
if r.currStorage.Len() > 0 {
r.succStorage.Write(r.currStorage.Bytes()[:cutoff-1])
r.succStorage.WriteByte(r.currStorage.Bytes()[cutoff-1] + 1) // Modify last nibble in the incarnation part of the `currStorage`
}
}
func (r *RootHashAggregator) genStructStorage() error {
var err error
var data GenStructStepData
if r.wasIHStorage {
r.hashData.Hash = common.BytesToHash(r.valueStorage)
data = &r.hashData
} else {
r.leafData.Value = rlphacks.RlpSerializableBytes(r.valueStorage)
data = &r.leafData
}
r.groups, err = GenStructStep(r.RetainNothing, r.currStorage.Bytes(), r.succStorage.Bytes(), r.hb, r.hc, data, r.groups, r.trace)
if err != nil {
return err
}
return nil
}
func (r *RootHashAggregator) saveValueStorage(isIH bool, v, h []byte) {
// Remember the current value
r.wasIHStorage = isIH
r.valueStorage = nil
if isIH {
r.valueStorage = h
} else {
r.valueStorage = v
}
}
func (r *RootHashAggregator) advanceKeysAccount(k []byte, terminator bool) {
r.curr.Reset()
r.curr.Write(r.succ.Bytes())
r.succ.Reset()
for _, b := range k {
r.succ.WriteByte(b / 16)
r.succ.WriteByte(b % 16)
}
if terminator {
r.succ.WriteByte(16)
}
}
func (r *RootHashAggregator) cutoffKeysAccount(cutoff int) {
r.curr.Reset()
r.curr.Write(r.succ.Bytes())
r.succ.Reset()
if r.curr.Len() > 0 && cutoff > 0 {
r.succ.Write(r.curr.Bytes()[:cutoff-1])
r.succ.WriteByte(r.curr.Bytes()[cutoff-1] + 1) // Modify last nibble before the cutoff point
}
}
func (r *RootHashAggregator) genStructAccount() error {
var data GenStructStepData
if r.wasIH {
copy(r.hashData.Hash[:], r.value)
data = &r.hashData
} else {
r.accData.Balance.Set(&r.a.Balance)
if r.a.Balance.Sign() != 0 {
r.accData.FieldSet |= AccountFieldBalanceOnly
}
r.accData.Nonce = r.a.Nonce
if r.a.Nonce != 0 {
r.accData.FieldSet |= AccountFieldNonceOnly
}
r.accData.Incarnation = r.a.Incarnation
data = &r.accData
}
r.wasIHStorage = false
r.currStorage.Reset()
r.succStorage.Reset()
var err error
if r.groups, err = GenStructStep(r.RetainNothing, r.curr.Bytes(), r.succ.Bytes(), r.hb, r.hc, data, r.groups, r.trace); err != nil {
return err
}
r.accData.FieldSet = 0
return nil
}
func (r *RootHashAggregator) saveValueAccount(isIH bool, v *accounts.Account, h []byte) error {
r.wasIH = isIH
if isIH {
r.value = h
return nil
}
r.a.Copy(v)
// Place code on the stack first, the storage will follow
if !r.a.IsEmptyCodeHash() {
// the first item ends up deepest on the stack, the second item - on the top
r.accData.FieldSet |= AccountFieldCodeOnly
if err := r.hb.hash(r.a.CodeHash[:]); err != nil {
return err
}
}
return nil
}
// FilterCursor - call .filter() and if it returns false - skip element
type FilterCursor struct {
c ethdb.Cursor
k, kHex, v []byte
filter func(k []byte) (bool, error)
}
func Filter(filter func(k []byte) (bool, error), c ethdb.Cursor) *FilterCursor {
return &FilterCursor{c: c, filter: filter}
}
func (c *FilterCursor) _seek(seek []byte) (err error) {
c.k, c.v, err = c.c.Seek(seek)
if err != nil {
return err
}
if c.k == nil {
return nil
}
DecompressNibbles(c.k, &c.kHex)
if ok, err := c.filter(c.kHex); err != nil {
return err
} else if ok {
return nil
}
return c._next()
}
func (c *FilterCursor) _next() (err error) {
c.k, c.v, err = c.c.Next()
if err != nil {
return err
}
for {
if c.k == nil {
return nil
}
DecompressNibbles(c.k, &c.kHex)
var ok bool
ok, err = c.filter(c.kHex)
if err != nil {
return err
} else if ok {
return nil
}
c.k, c.v, err = c.c.Next()
if err != nil {
return err
}
}
}
func (c *FilterCursor) Seek(seek []byte) ([]byte, []byte, error) {
if err := c._seek(seek); err != nil {
return []byte{}, nil, err
}
return c.k, c.v, nil
}
// IHCursor - holds logic related to iteration over IH bucket
type IHCursor struct {
c *FilterCursor
}
func IH(c *FilterCursor) *IHCursor {
return &IHCursor{c: c}
}
func (c *IHCursor) Seek(seek []byte) ([]byte, []byte, bool, error) {
k, v, err := c.c.Seek(seek)
if err != nil {
return []byte{}, nil, false, err
}
if k == nil {
return k, v, false, nil
}
return k, v, isSequence(seek, k), nil
}
/*
Sequence - if between 2 IH records not possible insert any state record - then they form "sequence"
Example1:
1234
1235
Example2:
12ff
13
Example3:
12ff
13000000
If 2 IH records form "sequence" then it can be consumed without moving StateCursor
*/
func isSequence(prev []byte, next []byte) bool {
isSequence := false
if bytes.HasPrefix(next, prev) {
tail := next[len(prev):] // if tail has only zeroes, then no state records can be between fstl.nextHex and fstl.ihK
isSequence = true
for _, n := range tail {
if n != 0 {
isSequence = false
break
}
}
}
return isSequence
}
func nextAccount(in, out []byte) bool {
copy(out, in)
for i := len(out) - 1; i >= 0; i-- {
if out[i] != 255 {
out[i]++
return true
}
out[i] = 0
}
return false
}
// keyIsBefore - kind of bytes.Compare, but nil is the last key. And return
func keyIsBeforeOrEqual(k1, k2 []byte) (bool, []byte) {
if k1 == nil {
return false, k2
}
if k2 == nil {
return true, k1
}
switch bytes.Compare(k1, k2) {
case -1, 0:
return true, k1
default:
return false, k2
}
}
// keyIsBefore - kind of bytes.Compare, but nil is the last key. And return
func keyIsBefore(k1, k2 []byte) bool {
if k1 == nil {
return false
}
if k2 == nil {
return true
}
switch bytes.Compare(k1, k2) {
case -1:
return true
default:
return false
}
}