mirror of
https://gitlab.com/pulsechaincom/go-pulse.git
synced 2024-12-25 21:07:17 +00:00
289b30715d
This commit converts the dependency management from Godeps to the vendor folder, also switching the tool from godep to trash. Since the upstream tool lacks a few features proposed via a few PRs, until those PRs are merged in (if), use github.com/karalabe/trash. You can update dependencies via trash --update. All dependencies have been updated to their latest version. Parts of the build system are reworked to drop old notions of Godeps and invocation of the go vet command so that it doesn't run against the vendor folder, as that will just blow up during vetting. The conversion drops OpenCL (and hence GPU mining support) from ethash and our codebase. The short reasoning is that there's noone to maintain and having opencl libs in our deps messes up builds as go install ./... tries to build them, failing with unsatisfied link errors for the C OpenCL deps. golang.org/x/net/context is not vendored in. We expect it to be fetched by the user (i.e. using go get). To keep ci.go builds reproducible the package is "vendored" in build/_vendor.
350 lines
8.3 KiB
Go
350 lines
8.3 KiB
Go
// Copyright (c) 2012, Suryandaru Triandana <syndtr@gmail.com>
|
|
// All rights reserved.
|
|
//
|
|
// Use of this source code is governed by a BSD-style license that can be
|
|
// found in the LICENSE file.
|
|
|
|
package leveldb
|
|
|
|
import (
|
|
"encoding/binary"
|
|
"fmt"
|
|
"io"
|
|
|
|
"github.com/syndtr/goleveldb/leveldb/errors"
|
|
"github.com/syndtr/goleveldb/leveldb/memdb"
|
|
"github.com/syndtr/goleveldb/leveldb/storage"
|
|
)
|
|
|
|
// ErrBatchCorrupted records reason of batch corruption. This error will be
|
|
// wrapped with errors.ErrCorrupted.
|
|
type ErrBatchCorrupted struct {
|
|
Reason string
|
|
}
|
|
|
|
func (e *ErrBatchCorrupted) Error() string {
|
|
return fmt.Sprintf("leveldb: batch corrupted: %s", e.Reason)
|
|
}
|
|
|
|
func newErrBatchCorrupted(reason string) error {
|
|
return errors.NewErrCorrupted(storage.FileDesc{}, &ErrBatchCorrupted{reason})
|
|
}
|
|
|
|
const (
|
|
batchHeaderLen = 8 + 4
|
|
batchGrowRec = 3000
|
|
batchBufioSize = 16
|
|
)
|
|
|
|
// BatchReplay wraps basic batch operations.
|
|
type BatchReplay interface {
|
|
Put(key, value []byte)
|
|
Delete(key []byte)
|
|
}
|
|
|
|
type batchIndex struct {
|
|
keyType keyType
|
|
keyPos, keyLen int
|
|
valuePos, valueLen int
|
|
}
|
|
|
|
func (index batchIndex) k(data []byte) []byte {
|
|
return data[index.keyPos : index.keyPos+index.keyLen]
|
|
}
|
|
|
|
func (index batchIndex) v(data []byte) []byte {
|
|
if index.valueLen != 0 {
|
|
return data[index.valuePos : index.valuePos+index.valueLen]
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (index batchIndex) kv(data []byte) (key, value []byte) {
|
|
return index.k(data), index.v(data)
|
|
}
|
|
|
|
// Batch is a write batch.
|
|
type Batch struct {
|
|
data []byte
|
|
index []batchIndex
|
|
|
|
// internalLen is sums of key/value pair length plus 8-bytes internal key.
|
|
internalLen int
|
|
}
|
|
|
|
func (b *Batch) grow(n int) {
|
|
o := len(b.data)
|
|
if cap(b.data)-o < n {
|
|
div := 1
|
|
if len(b.index) > batchGrowRec {
|
|
div = len(b.index) / batchGrowRec
|
|
}
|
|
ndata := make([]byte, o, o+n+o/div)
|
|
copy(ndata, b.data)
|
|
b.data = ndata
|
|
}
|
|
}
|
|
|
|
func (b *Batch) appendRec(kt keyType, key, value []byte) {
|
|
n := 1 + binary.MaxVarintLen32 + len(key)
|
|
if kt == keyTypeVal {
|
|
n += binary.MaxVarintLen32 + len(value)
|
|
}
|
|
b.grow(n)
|
|
index := batchIndex{keyType: kt}
|
|
o := len(b.data)
|
|
data := b.data[:o+n]
|
|
data[o] = byte(kt)
|
|
o++
|
|
o += binary.PutUvarint(data[o:], uint64(len(key)))
|
|
index.keyPos = o
|
|
index.keyLen = len(key)
|
|
o += copy(data[o:], key)
|
|
if kt == keyTypeVal {
|
|
o += binary.PutUvarint(data[o:], uint64(len(value)))
|
|
index.valuePos = o
|
|
index.valueLen = len(value)
|
|
o += copy(data[o:], value)
|
|
}
|
|
b.data = data[:o]
|
|
b.index = append(b.index, index)
|
|
b.internalLen += index.keyLen + index.valueLen + 8
|
|
}
|
|
|
|
// Put appends 'put operation' of the given key/value pair to the batch.
|
|
// It is safe to modify the contents of the argument after Put returns but not
|
|
// before.
|
|
func (b *Batch) Put(key, value []byte) {
|
|
b.appendRec(keyTypeVal, key, value)
|
|
}
|
|
|
|
// Delete appends 'delete operation' of the given key to the batch.
|
|
// It is safe to modify the contents of the argument after Delete returns but
|
|
// not before.
|
|
func (b *Batch) Delete(key []byte) {
|
|
b.appendRec(keyTypeDel, key, nil)
|
|
}
|
|
|
|
// Dump dumps batch contents. The returned slice can be loaded into the
|
|
// batch using Load method.
|
|
// The returned slice is not its own copy, so the contents should not be
|
|
// modified.
|
|
func (b *Batch) Dump() []byte {
|
|
return b.data
|
|
}
|
|
|
|
// Load loads given slice into the batch. Previous contents of the batch
|
|
// will be discarded.
|
|
// The given slice will not be copied and will be used as batch buffer, so
|
|
// it is not safe to modify the contents of the slice.
|
|
func (b *Batch) Load(data []byte) error {
|
|
return b.decode(data, -1)
|
|
}
|
|
|
|
// Replay replays batch contents.
|
|
func (b *Batch) Replay(r BatchReplay) error {
|
|
for _, index := range b.index {
|
|
switch index.keyType {
|
|
case keyTypeVal:
|
|
r.Put(index.k(b.data), index.v(b.data))
|
|
case keyTypeDel:
|
|
r.Delete(index.k(b.data))
|
|
}
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// Len returns number of records in the batch.
|
|
func (b *Batch) Len() int {
|
|
return len(b.index)
|
|
}
|
|
|
|
// Reset resets the batch.
|
|
func (b *Batch) Reset() {
|
|
b.data = b.data[:0]
|
|
b.index = b.index[:0]
|
|
b.internalLen = 0
|
|
}
|
|
|
|
func (b *Batch) replayInternal(fn func(i int, kt keyType, k, v []byte) error) error {
|
|
for i, index := range b.index {
|
|
if err := fn(i, index.keyType, index.k(b.data), index.v(b.data)); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (b *Batch) append(p *Batch) {
|
|
ob := len(b.data)
|
|
oi := len(b.index)
|
|
b.data = append(b.data, p.data...)
|
|
b.index = append(b.index, p.index...)
|
|
b.internalLen += p.internalLen
|
|
|
|
// Updating index offset.
|
|
if ob != 0 {
|
|
for ; oi < len(b.index); oi++ {
|
|
index := &b.index[oi]
|
|
index.keyPos += ob
|
|
if index.valueLen != 0 {
|
|
index.valuePos += ob
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
func (b *Batch) decode(data []byte, expectedLen int) error {
|
|
b.data = data
|
|
b.index = b.index[:0]
|
|
b.internalLen = 0
|
|
err := decodeBatch(data, func(i int, index batchIndex) error {
|
|
b.index = append(b.index, index)
|
|
b.internalLen += index.keyLen + index.valueLen + 8
|
|
return nil
|
|
})
|
|
if err != nil {
|
|
return err
|
|
}
|
|
if expectedLen >= 0 && len(b.index) != expectedLen {
|
|
return newErrBatchCorrupted(fmt.Sprintf("invalid records length: %d vs %d", expectedLen, len(b.index)))
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (b *Batch) putMem(seq uint64, mdb *memdb.DB) error {
|
|
var ik []byte
|
|
for i, index := range b.index {
|
|
ik = makeInternalKey(ik, index.k(b.data), seq+uint64(i), index.keyType)
|
|
if err := mdb.Put(ik, index.v(b.data)); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (b *Batch) revertMem(seq uint64, mdb *memdb.DB) error {
|
|
var ik []byte
|
|
for i, index := range b.index {
|
|
ik = makeInternalKey(ik, index.k(b.data), seq+uint64(i), index.keyType)
|
|
if err := mdb.Delete(ik); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func newBatch() interface{} {
|
|
return &Batch{}
|
|
}
|
|
|
|
func decodeBatch(data []byte, fn func(i int, index batchIndex) error) error {
|
|
var index batchIndex
|
|
for i, o := 0, 0; o < len(data); i++ {
|
|
// Key type.
|
|
index.keyType = keyType(data[o])
|
|
if index.keyType > keyTypeVal {
|
|
return newErrBatchCorrupted(fmt.Sprintf("bad record: invalid type %#x", uint(index.keyType)))
|
|
}
|
|
o++
|
|
|
|
// Key.
|
|
x, n := binary.Uvarint(data[o:])
|
|
o += n
|
|
if n <= 0 || o+int(x) > len(data) {
|
|
return newErrBatchCorrupted("bad record: invalid key length")
|
|
}
|
|
index.keyPos = o
|
|
index.keyLen = int(x)
|
|
o += index.keyLen
|
|
|
|
// Value.
|
|
if index.keyType == keyTypeVal {
|
|
x, n = binary.Uvarint(data[o:])
|
|
o += n
|
|
if n <= 0 || o+int(x) > len(data) {
|
|
return newErrBatchCorrupted("bad record: invalid value length")
|
|
}
|
|
index.valuePos = o
|
|
index.valueLen = int(x)
|
|
o += index.valueLen
|
|
} else {
|
|
index.valuePos = 0
|
|
index.valueLen = 0
|
|
}
|
|
|
|
if err := fn(i, index); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func decodeBatchToMem(data []byte, expectSeq uint64, mdb *memdb.DB) (seq uint64, batchLen int, err error) {
|
|
seq, batchLen, err = decodeBatchHeader(data)
|
|
if err != nil {
|
|
return 0, 0, err
|
|
}
|
|
if seq < expectSeq {
|
|
return 0, 0, newErrBatchCorrupted("invalid sequence number")
|
|
}
|
|
data = data[batchHeaderLen:]
|
|
var ik []byte
|
|
var decodedLen int
|
|
err = decodeBatch(data, func(i int, index batchIndex) error {
|
|
if i >= batchLen {
|
|
return newErrBatchCorrupted("invalid records length")
|
|
}
|
|
ik = makeInternalKey(ik, index.k(data), seq+uint64(i), index.keyType)
|
|
if err := mdb.Put(ik, index.v(data)); err != nil {
|
|
return err
|
|
}
|
|
decodedLen++
|
|
return nil
|
|
})
|
|
if err == nil && decodedLen != batchLen {
|
|
err = newErrBatchCorrupted(fmt.Sprintf("invalid records length: %d vs %d", batchLen, decodedLen))
|
|
}
|
|
return
|
|
}
|
|
|
|
func encodeBatchHeader(dst []byte, seq uint64, batchLen int) []byte {
|
|
dst = ensureBuffer(dst, batchHeaderLen)
|
|
binary.LittleEndian.PutUint64(dst, seq)
|
|
binary.LittleEndian.PutUint32(dst[8:], uint32(batchLen))
|
|
return dst
|
|
}
|
|
|
|
func decodeBatchHeader(data []byte) (seq uint64, batchLen int, err error) {
|
|
if len(data) < batchHeaderLen {
|
|
return 0, 0, newErrBatchCorrupted("too short")
|
|
}
|
|
|
|
seq = binary.LittleEndian.Uint64(data)
|
|
batchLen = int(binary.LittleEndian.Uint32(data[8:]))
|
|
if batchLen < 0 {
|
|
return 0, 0, newErrBatchCorrupted("invalid records length")
|
|
}
|
|
return
|
|
}
|
|
|
|
func batchesLen(batches []*Batch) int {
|
|
batchLen := 0
|
|
for _, batch := range batches {
|
|
batchLen += batch.Len()
|
|
}
|
|
return batchLen
|
|
}
|
|
|
|
func writeBatchesWithHeader(wr io.Writer, batches []*Batch, seq uint64) error {
|
|
if _, err := wr.Write(encodeBatchHeader(nil, seq, batchesLen(batches))); err != nil {
|
|
return err
|
|
}
|
|
for _, batch := range batches {
|
|
if _, err := wr.Write(batch.data); err != nil {
|
|
return err
|
|
}
|
|
}
|
|
return nil
|
|
}
|