mirror of
https://gitlab.com/pulsechaincom/erigon-pulse.git
synced 2024-12-27 22:28:21 +00:00
fd52a788b7
* Print snapshot prune * More print * Print * Print * Print * Move snapshots stage forward * Cleanup * Fix tests * Print * Too much logging * Remove print * Log, check * Revert * No panic, print * Fix tx numbering * Harder condition to start retiring blocks * Disable Pow verification after TTD is reached * Fix POW verifying * Print * Prints * Fix? * cleanup * Add migrations and hack * More diagnostics * More print * Reset sequence only once * Fix migration * Remove print reset * Fix lint Co-authored-by: Alexey Sharp <alexeysharp@Alexeys-iMac.local> Co-authored-by: Alex Sharp <alexsharp@Alexs-MacBook-Pro.local>
287 lines
8.3 KiB
Go
287 lines
8.3 KiB
Go
package migrations
|
|
|
|
import (
|
|
"bytes"
|
|
"context"
|
|
"encoding/binary"
|
|
"fmt"
|
|
"path/filepath"
|
|
|
|
"github.com/ledgerwatch/erigon-lib/kv"
|
|
"github.com/ledgerwatch/erigon/common"
|
|
"github.com/ledgerwatch/erigon/eth/stagedsync/stages"
|
|
"github.com/ledgerwatch/erigon/node/nodecfg/datadir"
|
|
"github.com/ledgerwatch/log/v3"
|
|
"github.com/ugorji/go/codec"
|
|
)
|
|
|
|
// migrations apply sequentially in order of this array, skips applied migrations
|
|
// it allows - don't worry about merge conflicts and use switch branches
|
|
// see also dbutils.Migrations - it stores context in which each transaction was exectured - useful for bug-reports
|
|
//
|
|
// Idempotency is expected
|
|
// Best practices to achieve Idempotency:
|
|
// - in dbutils/bucket.go add suffix for existing bucket variable, create new bucket with same variable name.
|
|
// Example:
|
|
// - SyncStageProgress = []byte("SSP1")
|
|
// - SyncStageProgressOld1 = []byte("SSP1")
|
|
// - SyncStageProgress = []byte("SSP2")
|
|
// - in the beginning of migration: check that old bucket exists, clear new bucket
|
|
// - in the end:drop old bucket (not in defer!).
|
|
// - if you need migrate multiple buckets - create separate migration for each bucket
|
|
// - write test - and check that it's safe to apply same migration twice
|
|
var migrations = map[kv.Label][]Migration{
|
|
kv.ChainDB: {
|
|
dbSchemaVersion5,
|
|
txsBeginEnd,
|
|
resetBlocks4,
|
|
},
|
|
kv.TxPoolDB: {},
|
|
kv.SentryDB: {},
|
|
}
|
|
|
|
type Callback func(tx kv.RwTx, progress []byte, isDone bool) error
|
|
type Migration struct {
|
|
Name string
|
|
Up func(db kv.RwDB, dirs datadir.Dirs, progress []byte, BeforeCommit Callback) error
|
|
}
|
|
|
|
var (
|
|
ErrMigrationNonUniqueName = fmt.Errorf("please provide unique migration name")
|
|
ErrMigrationCommitNotCalled = fmt.Errorf("migration before-commit function was not called")
|
|
ErrMigrationETLFilesDeleted = fmt.Errorf("db migration progress was interrupted after extraction step and ETL files was deleted, please contact development team for help or re-sync from scratch")
|
|
)
|
|
|
|
func NewMigrator(label kv.Label) *Migrator {
|
|
return &Migrator{
|
|
Migrations: migrations[label],
|
|
}
|
|
}
|
|
|
|
type Migrator struct {
|
|
Migrations []Migration
|
|
}
|
|
|
|
func AppliedMigrations(tx kv.Tx, withPayload bool) (map[string][]byte, error) {
|
|
applied := map[string][]byte{}
|
|
err := tx.ForEach(kv.Migrations, nil, func(k []byte, v []byte) error {
|
|
if bytes.HasPrefix(k, []byte("_progress_")) {
|
|
return nil
|
|
}
|
|
if withPayload {
|
|
applied[string(common.CopyBytes(k))] = common.CopyBytes(v)
|
|
} else {
|
|
applied[string(common.CopyBytes(k))] = []byte{}
|
|
}
|
|
return nil
|
|
})
|
|
return applied, err
|
|
}
|
|
|
|
func (m *Migrator) HasPendingMigrations(db kv.RwDB) (bool, error) {
|
|
var has bool
|
|
if err := db.View(context.Background(), func(tx kv.Tx) error {
|
|
pending, err := m.PendingMigrations(tx)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
has = len(pending) > 0
|
|
return nil
|
|
}); err != nil {
|
|
return false, err
|
|
}
|
|
return has, nil
|
|
}
|
|
|
|
func (m *Migrator) PendingMigrations(tx kv.Tx) ([]Migration, error) {
|
|
applied, err := AppliedMigrations(tx, false)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
counter := 0
|
|
for i := range m.Migrations {
|
|
v := m.Migrations[i]
|
|
if _, ok := applied[v.Name]; ok {
|
|
continue
|
|
}
|
|
counter++
|
|
}
|
|
|
|
pending := make([]Migration, 0, counter)
|
|
for i := range m.Migrations {
|
|
v := m.Migrations[i]
|
|
if _, ok := applied[v.Name]; ok {
|
|
continue
|
|
}
|
|
pending = append(pending, v)
|
|
}
|
|
return pending, nil
|
|
}
|
|
|
|
func (m *Migrator) VerifyVersion(db kv.RwDB) error {
|
|
if err := db.View(context.Background(), func(tx kv.Tx) error {
|
|
var err error
|
|
existingVersion, err := tx.GetOne(kv.DatabaseInfo, kv.DBSchemaVersionKey)
|
|
if err != nil {
|
|
return fmt.Errorf("reading DB schema version: %w", err)
|
|
}
|
|
if len(existingVersion) != 0 && len(existingVersion) != 12 {
|
|
return fmt.Errorf("incorrect length of DB schema version: %d", len(existingVersion))
|
|
}
|
|
if len(existingVersion) == 12 {
|
|
major := binary.BigEndian.Uint32(existingVersion)
|
|
minor := binary.BigEndian.Uint32(existingVersion[4:])
|
|
if major > kv.DBSchemaVersion.Major {
|
|
return fmt.Errorf("cannot downgrade major DB version from %d to %d", major, kv.DBSchemaVersion.Major)
|
|
} else if major == kv.DBSchemaVersion.Major {
|
|
if minor > kv.DBSchemaVersion.Minor {
|
|
return fmt.Errorf("cannot downgrade minor DB version from %d.%d to %d.%d", major, minor, kv.DBSchemaVersion.Major, kv.DBSchemaVersion.Major)
|
|
}
|
|
} else {
|
|
// major < kv.DBSchemaVersion.Major
|
|
if kv.DBSchemaVersion.Major-major > 1 {
|
|
return fmt.Errorf("cannot upgrade major DB version for more than 1 version from %d to %d, use integration tool if you know what you are doing", major, kv.DBSchemaVersion.Major)
|
|
}
|
|
}
|
|
}
|
|
return nil
|
|
}); err != nil {
|
|
return fmt.Errorf("migrator.VerifyVersion: %w", err)
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
func (m *Migrator) Apply(db kv.RwDB, dataDir string) error {
|
|
if len(m.Migrations) == 0 {
|
|
return nil
|
|
}
|
|
dirs := datadir.New(dataDir)
|
|
|
|
var applied map[string][]byte
|
|
if err := db.View(context.Background(), func(tx kv.Tx) error {
|
|
var err error
|
|
applied, err = AppliedMigrations(tx, false)
|
|
if err != nil {
|
|
return fmt.Errorf("reading applied migrations: %w", err)
|
|
}
|
|
return nil
|
|
}); err != nil {
|
|
return err
|
|
}
|
|
if err := m.VerifyVersion(db); err != nil {
|
|
return fmt.Errorf("migrator.Apply: %w", err)
|
|
}
|
|
|
|
// migration names must be unique, protection against people's mistake
|
|
uniqueNameCheck := map[string]bool{}
|
|
for i := range m.Migrations {
|
|
_, ok := uniqueNameCheck[m.Migrations[i].Name]
|
|
if ok {
|
|
return fmt.Errorf("%w, duplicate: %s", ErrMigrationNonUniqueName, m.Migrations[i].Name)
|
|
}
|
|
uniqueNameCheck[m.Migrations[i].Name] = true
|
|
}
|
|
|
|
for i := range m.Migrations {
|
|
v := m.Migrations[i]
|
|
if _, ok := applied[v.Name]; ok {
|
|
continue
|
|
}
|
|
|
|
callbackCalled := false // commit function must be called if no error, protection against people's mistake
|
|
|
|
log.Info("Apply migration", "name", v.Name)
|
|
var progress []byte
|
|
if err := db.View(context.Background(), func(tx kv.Tx) (err error) {
|
|
progress, err = tx.GetOne(kv.Migrations, []byte("_progress_"+v.Name))
|
|
return err
|
|
}); err != nil {
|
|
return fmt.Errorf("migrator.Apply: %w", err)
|
|
}
|
|
|
|
dirs.Tmp = filepath.Join(dirs.DataDir, "migrations", v.Name)
|
|
if err := v.Up(db, dirs, progress, func(tx kv.RwTx, key []byte, isDone bool) error {
|
|
if !isDone {
|
|
if key != nil {
|
|
if err := tx.Put(kv.Migrations, []byte("_progress_"+v.Name), key); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
return nil
|
|
}
|
|
callbackCalled = true
|
|
|
|
stagesProgress, err := MarshalMigrationPayload(tx)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
err = tx.Put(kv.Migrations, []byte(v.Name), stagesProgress)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
err = tx.Delete(kv.Migrations, []byte("_progress_"+v.Name))
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
return nil
|
|
}); err != nil {
|
|
return fmt.Errorf("migrator.Apply.Up: %s, %w", v.Name, err)
|
|
}
|
|
|
|
if !callbackCalled {
|
|
return fmt.Errorf("%w: %s", ErrMigrationCommitNotCalled, v.Name)
|
|
}
|
|
log.Info("Applied migration", "name", v.Name)
|
|
}
|
|
// Write DB schema version
|
|
var version [12]byte
|
|
binary.BigEndian.PutUint32(version[:], kv.DBSchemaVersion.Major)
|
|
binary.BigEndian.PutUint32(version[4:], kv.DBSchemaVersion.Minor)
|
|
binary.BigEndian.PutUint32(version[8:], kv.DBSchemaVersion.Patch)
|
|
if err := db.Update(context.Background(), func(tx kv.RwTx) error {
|
|
if err := tx.Put(kv.DatabaseInfo, kv.DBSchemaVersionKey, version[:]); err != nil {
|
|
return fmt.Errorf("writing DB schema version: %w", err)
|
|
}
|
|
return nil
|
|
}); err != nil {
|
|
return fmt.Errorf("migrator.Apply: %w", err)
|
|
}
|
|
log.Info("Updated DB schema to", "version", fmt.Sprintf("%d.%d.%d", kv.DBSchemaVersion.Major, kv.DBSchemaVersion.Minor, kv.DBSchemaVersion.Patch))
|
|
return nil
|
|
}
|
|
|
|
func MarshalMigrationPayload(db kv.Getter) ([]byte, error) {
|
|
s := map[string][]byte{}
|
|
|
|
buf := bytes.NewBuffer(nil)
|
|
encoder := codec.NewEncoder(buf, &codec.CborHandle{})
|
|
|
|
for _, stage := range stages.AllStages {
|
|
v, err := db.GetOne(kv.SyncStageProgress, []byte(stage))
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
if len(v) > 0 {
|
|
s[string(stage)] = common.CopyBytes(v)
|
|
}
|
|
}
|
|
|
|
if err := encoder.Encode(s); err != nil {
|
|
return nil, err
|
|
}
|
|
return buf.Bytes(), nil
|
|
}
|
|
|
|
func UnmarshalMigrationPayload(data []byte) (map[string][]byte, error) {
|
|
s := map[string][]byte{}
|
|
|
|
if err := codec.NewDecoder(bytes.NewReader(data), &codec.CborHandle{}).Decode(&s); err != nil {
|
|
return nil, err
|
|
}
|
|
return s, nil
|
|
}
|