all: replace fmt.Errorf() with errors.New() if no param required (#29126)
replace-fmt-errorf Co-authored-by: yzb@example.cn <yzb@example.cn>
This commit is contained in:
parent
28d55218f7
commit
db4cf69166
@ -18,6 +18,7 @@ package main
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
"math/big"
|
"math/big"
|
||||||
"os"
|
"os"
|
||||||
@ -182,7 +183,7 @@ func open(ctx *cli.Context, epoch uint64) (*era.Era, error) {
|
|||||||
// that the accumulator matches the expected value.
|
// that the accumulator matches the expected value.
|
||||||
func verify(ctx *cli.Context) error {
|
func verify(ctx *cli.Context) error {
|
||||||
if ctx.Args().Len() != 1 {
|
if ctx.Args().Len() != 1 {
|
||||||
return fmt.Errorf("missing accumulators file")
|
return errors.New("missing accumulators file")
|
||||||
}
|
}
|
||||||
|
|
||||||
roots, err := readHashes(ctx.Args().First())
|
roots, err := readHashes(ctx.Args().First())
|
||||||
@ -203,7 +204,7 @@ func verify(ctx *cli.Context) error {
|
|||||||
}
|
}
|
||||||
|
|
||||||
if len(entries) != len(roots) {
|
if len(entries) != len(roots) {
|
||||||
return fmt.Errorf("number of era1 files should match the number of accumulator hashes")
|
return errors.New("number of era1 files should match the number of accumulator hashes")
|
||||||
}
|
}
|
||||||
|
|
||||||
// Verify each epoch matches the expected root.
|
// Verify each epoch matches the expected root.
|
||||||
@ -308,7 +309,7 @@ func checkAccumulator(e *era.Era) error {
|
|||||||
func readHashes(f string) ([]common.Hash, error) {
|
func readHashes(f string) ([]common.Hash, error) {
|
||||||
b, err := os.ReadFile(f)
|
b, err := os.ReadFile(f)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, fmt.Errorf("unable to open accumulators file")
|
return nil, errors.New("unable to open accumulators file")
|
||||||
}
|
}
|
||||||
s := strings.Split(string(b), "\n")
|
s := strings.Split(string(b), "\n")
|
||||||
// Remove empty last element, if present.
|
// Remove empty last element, if present.
|
||||||
|
@ -444,7 +444,7 @@ func importHistory(ctx *cli.Context) error {
|
|||||||
return fmt.Errorf("no era1 files found in %s", dir)
|
return fmt.Errorf("no era1 files found in %s", dir)
|
||||||
}
|
}
|
||||||
if len(networks) > 1 {
|
if len(networks) > 1 {
|
||||||
return fmt.Errorf("multiple networks found, use a network flag to specify desired network")
|
return errors.New("multiple networks found, use a network flag to specify desired network")
|
||||||
}
|
}
|
||||||
network = networks[0]
|
network = networks[0]
|
||||||
}
|
}
|
||||||
|
@ -245,7 +245,7 @@ func readList(filename string) ([]string, error) {
|
|||||||
// starting from genesis.
|
// starting from genesis.
|
||||||
func ImportHistory(chain *core.BlockChain, db ethdb.Database, dir string, network string) error {
|
func ImportHistory(chain *core.BlockChain, db ethdb.Database, dir string, network string) error {
|
||||||
if chain.CurrentSnapBlock().Number.BitLen() != 0 {
|
if chain.CurrentSnapBlock().Number.BitLen() != 0 {
|
||||||
return fmt.Errorf("history import only supported when starting from genesis")
|
return errors.New("history import only supported when starting from genesis")
|
||||||
}
|
}
|
||||||
entries, err := era.ReadDir(dir, network)
|
entries, err := era.ReadDir(dir, network)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
@ -18,6 +18,7 @@ package txpool
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"crypto/sha256"
|
"crypto/sha256"
|
||||||
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
"math/big"
|
"math/big"
|
||||||
|
|
||||||
@ -120,13 +121,13 @@ func ValidateTransaction(tx *types.Transaction, head *types.Header, signer types
|
|||||||
}
|
}
|
||||||
sidecar := tx.BlobTxSidecar()
|
sidecar := tx.BlobTxSidecar()
|
||||||
if sidecar == nil {
|
if sidecar == nil {
|
||||||
return fmt.Errorf("missing sidecar in blob transaction")
|
return errors.New("missing sidecar in blob transaction")
|
||||||
}
|
}
|
||||||
// Ensure the number of items in the blob transaction and various side
|
// Ensure the number of items in the blob transaction and various side
|
||||||
// data match up before doing any expensive validations
|
// data match up before doing any expensive validations
|
||||||
hashes := tx.BlobHashes()
|
hashes := tx.BlobHashes()
|
||||||
if len(hashes) == 0 {
|
if len(hashes) == 0 {
|
||||||
return fmt.Errorf("blobless blob transaction")
|
return errors.New("blobless blob transaction")
|
||||||
}
|
}
|
||||||
if len(hashes) > params.MaxBlobGasPerBlock/params.BlobTxBlobGasPerBlob {
|
if len(hashes) > params.MaxBlobGasPerBlock/params.BlobTxBlobGasPerBlob {
|
||||||
return fmt.Errorf("too many blobs in transaction: have %d, permitted %d", len(hashes), params.MaxBlobGasPerBlock/params.BlobTxBlobGasPerBlob)
|
return fmt.Errorf("too many blobs in transaction: have %d, permitted %d", len(hashes), params.MaxBlobGasPerBlock/params.BlobTxBlobGasPerBlob)
|
||||||
|
@ -17,6 +17,7 @@
|
|||||||
package era
|
package era
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
"math/big"
|
"math/big"
|
||||||
|
|
||||||
@ -28,7 +29,7 @@ import (
|
|||||||
// accumulator of header records.
|
// accumulator of header records.
|
||||||
func ComputeAccumulator(hashes []common.Hash, tds []*big.Int) (common.Hash, error) {
|
func ComputeAccumulator(hashes []common.Hash, tds []*big.Int) (common.Hash, error) {
|
||||||
if len(hashes) != len(tds) {
|
if len(hashes) != len(tds) {
|
||||||
return common.Hash{}, fmt.Errorf("must have equal number hashes as td values")
|
return common.Hash{}, errors.New("must have equal number hashes as td values")
|
||||||
}
|
}
|
||||||
if len(hashes) > MaxEra1Size {
|
if len(hashes) > MaxEra1Size {
|
||||||
return common.Hash{}, fmt.Errorf("too many records: have %d, max %d", len(hashes), MaxEra1Size)
|
return common.Hash{}, fmt.Errorf("too many records: have %d, max %d", len(hashes), MaxEra1Size)
|
||||||
|
@ -18,6 +18,7 @@ package era
|
|||||||
import (
|
import (
|
||||||
"bytes"
|
"bytes"
|
||||||
"encoding/binary"
|
"encoding/binary"
|
||||||
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
"io"
|
"io"
|
||||||
"math/big"
|
"math/big"
|
||||||
@ -158,7 +159,7 @@ func (b *Builder) AddRLP(header, body, receipts []byte, number uint64, hash comm
|
|||||||
// corresponding e2store entries.
|
// corresponding e2store entries.
|
||||||
func (b *Builder) Finalize() (common.Hash, error) {
|
func (b *Builder) Finalize() (common.Hash, error) {
|
||||||
if b.startNum == nil {
|
if b.startNum == nil {
|
||||||
return common.Hash{}, fmt.Errorf("finalize called on empty builder")
|
return common.Hash{}, errors.New("finalize called on empty builder")
|
||||||
}
|
}
|
||||||
// Compute accumulator root and write entry.
|
// Compute accumulator root and write entry.
|
||||||
root, err := ComputeAccumulator(b.hashes, b.tds)
|
root, err := ComputeAccumulator(b.hashes, b.tds)
|
||||||
|
@ -18,6 +18,7 @@ package e2store
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"encoding/binary"
|
"encoding/binary"
|
||||||
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
"io"
|
"io"
|
||||||
)
|
)
|
||||||
@ -160,7 +161,7 @@ func (r *Reader) ReadMetadataAt(off int64) (typ uint16, length uint32, err error
|
|||||||
|
|
||||||
// Check reserved bytes of header.
|
// Check reserved bytes of header.
|
||||||
if b[6] != 0 || b[7] != 0 {
|
if b[6] != 0 || b[7] != 0 {
|
||||||
return 0, 0, fmt.Errorf("reserved bytes are non-zero")
|
return 0, 0, errors.New("reserved bytes are non-zero")
|
||||||
}
|
}
|
||||||
|
|
||||||
return typ, length, nil
|
return typ, length, nil
|
||||||
|
@ -18,7 +18,7 @@ package e2store
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"bytes"
|
"bytes"
|
||||||
"fmt"
|
"errors"
|
||||||
"io"
|
"io"
|
||||||
"testing"
|
"testing"
|
||||||
|
|
||||||
@ -92,7 +92,7 @@ func TestDecode(t *testing.T) {
|
|||||||
},
|
},
|
||||||
{ // basic invalid decoding
|
{ // basic invalid decoding
|
||||||
have: "ffff000000000001",
|
have: "ffff000000000001",
|
||||||
err: fmt.Errorf("reserved bytes are non-zero"),
|
err: errors.New("reserved bytes are non-zero"),
|
||||||
},
|
},
|
||||||
{ // no more entries to read, returns EOF
|
{ // no more entries to read, returns EOF
|
||||||
have: "",
|
have: "",
|
||||||
|
@ -18,6 +18,7 @@ package era
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"encoding/binary"
|
"encoding/binary"
|
||||||
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
"io"
|
"io"
|
||||||
"math/big"
|
"math/big"
|
||||||
@ -127,7 +128,7 @@ func (e *Era) Close() error {
|
|||||||
|
|
||||||
func (e *Era) GetBlockByNumber(num uint64) (*types.Block, error) {
|
func (e *Era) GetBlockByNumber(num uint64) (*types.Block, error) {
|
||||||
if e.m.start > num || e.m.start+e.m.count <= num {
|
if e.m.start > num || e.m.start+e.m.count <= num {
|
||||||
return nil, fmt.Errorf("out-of-bounds")
|
return nil, errors.New("out-of-bounds")
|
||||||
}
|
}
|
||||||
off, err := e.readOffset(num)
|
off, err := e.readOffset(num)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
@ -17,6 +17,7 @@
|
|||||||
package era
|
package era
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
"io"
|
"io"
|
||||||
"math/big"
|
"math/big"
|
||||||
@ -61,7 +62,7 @@ func (it *Iterator) Error() error {
|
|||||||
// Block returns the block for the iterator's current position.
|
// Block returns the block for the iterator's current position.
|
||||||
func (it *Iterator) Block() (*types.Block, error) {
|
func (it *Iterator) Block() (*types.Block, error) {
|
||||||
if it.inner.Header == nil || it.inner.Body == nil {
|
if it.inner.Header == nil || it.inner.Body == nil {
|
||||||
return nil, fmt.Errorf("header and body must be non-nil")
|
return nil, errors.New("header and body must be non-nil")
|
||||||
}
|
}
|
||||||
var (
|
var (
|
||||||
header types.Header
|
header types.Header
|
||||||
|
@ -947,7 +947,7 @@ func (w *worker) prepareWork(genParams *generateParams) (*environment, error) {
|
|||||||
if genParams.parentHash != (common.Hash{}) {
|
if genParams.parentHash != (common.Hash{}) {
|
||||||
block := w.chain.GetBlockByHash(genParams.parentHash)
|
block := w.chain.GetBlockByHash(genParams.parentHash)
|
||||||
if block == nil {
|
if block == nil {
|
||||||
return nil, fmt.Errorf("missing parent")
|
return nil, errors.New("missing parent")
|
||||||
}
|
}
|
||||||
parent = block.Header()
|
parent = block.Header()
|
||||||
}
|
}
|
||||||
|
@ -19,6 +19,7 @@ package node
|
|||||||
import (
|
import (
|
||||||
"compress/gzip"
|
"compress/gzip"
|
||||||
"context"
|
"context"
|
||||||
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
"io"
|
"io"
|
||||||
"net"
|
"net"
|
||||||
@ -299,7 +300,7 @@ func (h *httpServer) enableRPC(apis []rpc.API, config httpConfig) error {
|
|||||||
defer h.mu.Unlock()
|
defer h.mu.Unlock()
|
||||||
|
|
||||||
if h.rpcAllowed() {
|
if h.rpcAllowed() {
|
||||||
return fmt.Errorf("JSON-RPC over HTTP is already enabled")
|
return errors.New("JSON-RPC over HTTP is already enabled")
|
||||||
}
|
}
|
||||||
|
|
||||||
// Create RPC server and handler.
|
// Create RPC server and handler.
|
||||||
@ -335,7 +336,7 @@ func (h *httpServer) enableWS(apis []rpc.API, config wsConfig) error {
|
|||||||
defer h.mu.Unlock()
|
defer h.mu.Unlock()
|
||||||
|
|
||||||
if h.wsAllowed() {
|
if h.wsAllowed() {
|
||||||
return fmt.Errorf("JSON-RPC over WebSocket is already enabled")
|
return errors.New("JSON-RPC over WebSocket is already enabled")
|
||||||
}
|
}
|
||||||
// Create RPC server and handler.
|
// Create RPC server and handler.
|
||||||
srv := rpc.NewServer()
|
srv := rpc.NewServer()
|
||||||
|
@ -364,7 +364,7 @@ func (t *UDPv4) RequestENR(n *enode.Node) (*enode.Node, error) {
|
|||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
if respN.ID() != n.ID() {
|
if respN.ID() != n.ID() {
|
||||||
return nil, fmt.Errorf("invalid ID in response record")
|
return nil, errors.New("invalid ID in response record")
|
||||||
}
|
}
|
||||||
if respN.Seq() < n.Seq() {
|
if respN.Seq() < n.Seq() {
|
||||||
return n, nil // response record is older
|
return n, nil // response record is older
|
||||||
|
@ -442,7 +442,7 @@ func (t *UDPv5) verifyResponseNode(c *callV5, r *enr.Record, distances []uint, s
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
if _, ok := seen[node.ID()]; ok {
|
if _, ok := seen[node.ID()]; ok {
|
||||||
return nil, fmt.Errorf("duplicate record")
|
return nil, errors.New("duplicate record")
|
||||||
}
|
}
|
||||||
seen[node.ID()] = struct{}{}
|
seen[node.ID()] = struct{}{}
|
||||||
return node, nil
|
return node, nil
|
||||||
|
@ -367,11 +367,11 @@ func (c *Codec) makeHandshakeAuth(toID enode.ID, addr string, challenge *Whoarey
|
|||||||
// key is part of the ID nonce signature.
|
// key is part of the ID nonce signature.
|
||||||
var remotePubkey = new(ecdsa.PublicKey)
|
var remotePubkey = new(ecdsa.PublicKey)
|
||||||
if err := challenge.Node.Load((*enode.Secp256k1)(remotePubkey)); err != nil {
|
if err := challenge.Node.Load((*enode.Secp256k1)(remotePubkey)); err != nil {
|
||||||
return nil, nil, fmt.Errorf("can't find secp256k1 key for recipient")
|
return nil, nil, errors.New("can't find secp256k1 key for recipient")
|
||||||
}
|
}
|
||||||
ephkey, err := c.sc.ephemeralKeyGen()
|
ephkey, err := c.sc.ephemeralKeyGen()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, nil, fmt.Errorf("can't generate ephemeral key")
|
return nil, nil, errors.New("can't generate ephemeral key")
|
||||||
}
|
}
|
||||||
ephpubkey := EncodePubkey(&ephkey.PublicKey)
|
ephpubkey := EncodePubkey(&ephkey.PublicKey)
|
||||||
auth.pubkey = ephpubkey[:]
|
auth.pubkey = ephpubkey[:]
|
||||||
@ -395,7 +395,7 @@ func (c *Codec) makeHandshakeAuth(toID enode.ID, addr string, challenge *Whoarey
|
|||||||
// Create session keys.
|
// Create session keys.
|
||||||
sec := deriveKeys(sha256.New, ephkey, remotePubkey, c.localnode.ID(), challenge.Node.ID(), cdata)
|
sec := deriveKeys(sha256.New, ephkey, remotePubkey, c.localnode.ID(), challenge.Node.ID(), cdata)
|
||||||
if sec == nil {
|
if sec == nil {
|
||||||
return nil, nil, fmt.Errorf("key derivation failed")
|
return nil, nil, errors.New("key derivation failed")
|
||||||
}
|
}
|
||||||
return auth, sec, err
|
return auth, sec, err
|
||||||
}
|
}
|
||||||
|
@ -191,7 +191,7 @@ func (c *Client) resolveEntry(ctx context.Context, domain, hash string) (entry,
|
|||||||
func (c *Client) doResolveEntry(ctx context.Context, domain, hash string) (entry, error) {
|
func (c *Client) doResolveEntry(ctx context.Context, domain, hash string) (entry, error) {
|
||||||
wantHash, err := b32format.DecodeString(hash)
|
wantHash, err := b32format.DecodeString(hash)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, fmt.Errorf("invalid base32 hash")
|
return nil, errors.New("invalid base32 hash")
|
||||||
}
|
}
|
||||||
name := hash + "." + domain
|
name := hash + "." + domain
|
||||||
txts, err := c.cfg.Resolver.LookupTXT(ctx, hash+"."+domain)
|
txts, err := c.cfg.Resolver.LookupTXT(ctx, hash+"."+domain)
|
||||||
|
@ -21,6 +21,7 @@ import (
|
|||||||
"crypto/ecdsa"
|
"crypto/ecdsa"
|
||||||
"encoding/base32"
|
"encoding/base32"
|
||||||
"encoding/base64"
|
"encoding/base64"
|
||||||
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
"io"
|
"io"
|
||||||
"strings"
|
"strings"
|
||||||
@ -341,7 +342,7 @@ func parseLinkEntry(e string) (entry, error) {
|
|||||||
|
|
||||||
func parseLink(e string) (*linkEntry, error) {
|
func parseLink(e string) (*linkEntry, error) {
|
||||||
if !strings.HasPrefix(e, linkPrefix) {
|
if !strings.HasPrefix(e, linkPrefix) {
|
||||||
return nil, fmt.Errorf("wrong/missing scheme 'enrtree' in URL")
|
return nil, errors.New("wrong/missing scheme 'enrtree' in URL")
|
||||||
}
|
}
|
||||||
e = e[len(linkPrefix):]
|
e = e[len(linkPrefix):]
|
||||||
|
|
||||||
|
@ -18,7 +18,7 @@ package enode
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"crypto/ecdsa"
|
"crypto/ecdsa"
|
||||||
"fmt"
|
"errors"
|
||||||
"io"
|
"io"
|
||||||
|
|
||||||
"github.com/ethereum/go-ethereum/common/math"
|
"github.com/ethereum/go-ethereum/common/math"
|
||||||
@ -67,7 +67,7 @@ func (V4ID) Verify(r *enr.Record, sig []byte) error {
|
|||||||
if err := r.Load(&entry); err != nil {
|
if err := r.Load(&entry); err != nil {
|
||||||
return err
|
return err
|
||||||
} else if len(entry) != 33 {
|
} else if len(entry) != 33 {
|
||||||
return fmt.Errorf("invalid public key")
|
return errors.New("invalid public key")
|
||||||
}
|
}
|
||||||
|
|
||||||
h := sha3.NewLegacyKeccak256()
|
h := sha3.NewLegacyKeccak256()
|
||||||
|
@ -17,6 +17,7 @@
|
|||||||
package nat
|
package nat
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
"net"
|
"net"
|
||||||
"strings"
|
"strings"
|
||||||
@ -46,7 +47,7 @@ func (n *pmp) ExternalIP() (net.IP, error) {
|
|||||||
|
|
||||||
func (n *pmp) AddMapping(protocol string, extport, intport int, name string, lifetime time.Duration) (uint16, error) {
|
func (n *pmp) AddMapping(protocol string, extport, intport int, name string, lifetime time.Duration) (uint16, error) {
|
||||||
if lifetime <= 0 {
|
if lifetime <= 0 {
|
||||||
return 0, fmt.Errorf("lifetime must not be <= 0")
|
return 0, errors.New("lifetime must not be <= 0")
|
||||||
}
|
}
|
||||||
// Note order of port arguments is switched between our
|
// Note order of port arguments is switched between our
|
||||||
// AddMapping and the client's AddPortMapping.
|
// AddMapping and the client's AddPortMapping.
|
||||||
|
@ -460,7 +460,7 @@ func startExecNodeStack() (*node.Node, error) {
|
|||||||
// decode the config
|
// decode the config
|
||||||
confEnv := os.Getenv(envNodeConfig)
|
confEnv := os.Getenv(envNodeConfig)
|
||||||
if confEnv == "" {
|
if confEnv == "" {
|
||||||
return nil, fmt.Errorf("missing " + envNodeConfig)
|
return nil, errors.New("missing " + envNodeConfig)
|
||||||
}
|
}
|
||||||
var conf execNodeConfig
|
var conf execNodeConfig
|
||||||
if err := json.Unmarshal([]byte(confEnv), &conf); err != nil {
|
if err := json.Unmarshal([]byte(confEnv), &conf); err != nil {
|
||||||
|
@ -708,7 +708,7 @@ func formatPrimitiveValue(encType string, encValue interface{}) (string, error)
|
|||||||
func (t Types) validate() error {
|
func (t Types) validate() error {
|
||||||
for typeKey, typeArr := range t {
|
for typeKey, typeArr := range t {
|
||||||
if len(typeKey) == 0 {
|
if len(typeKey) == 0 {
|
||||||
return fmt.Errorf("empty type key")
|
return errors.New("empty type key")
|
||||||
}
|
}
|
||||||
for i, typeObj := range typeArr {
|
for i, typeObj := range typeArr {
|
||||||
if len(typeObj.Type) == 0 {
|
if len(typeObj.Type) == 0 {
|
||||||
|
@ -556,7 +556,7 @@ func runRandTest(rt randTest) error {
|
|||||||
checktr.MustUpdate(it.Key, it.Value)
|
checktr.MustUpdate(it.Key, it.Value)
|
||||||
}
|
}
|
||||||
if tr.Hash() != checktr.Hash() {
|
if tr.Hash() != checktr.Hash() {
|
||||||
rt[i].err = fmt.Errorf("hash mismatch in opItercheckhash")
|
rt[i].err = errors.New("hash mismatch in opItercheckhash")
|
||||||
}
|
}
|
||||||
case opNodeDiff:
|
case opNodeDiff:
|
||||||
var (
|
var (
|
||||||
@ -594,19 +594,19 @@ func runRandTest(rt randTest) error {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
if len(insertExp) != len(tr.tracer.inserts) {
|
if len(insertExp) != len(tr.tracer.inserts) {
|
||||||
rt[i].err = fmt.Errorf("insert set mismatch")
|
rt[i].err = errors.New("insert set mismatch")
|
||||||
}
|
}
|
||||||
if len(deleteExp) != len(tr.tracer.deletes) {
|
if len(deleteExp) != len(tr.tracer.deletes) {
|
||||||
rt[i].err = fmt.Errorf("delete set mismatch")
|
rt[i].err = errors.New("delete set mismatch")
|
||||||
}
|
}
|
||||||
for insert := range tr.tracer.inserts {
|
for insert := range tr.tracer.inserts {
|
||||||
if _, present := insertExp[insert]; !present {
|
if _, present := insertExp[insert]; !present {
|
||||||
rt[i].err = fmt.Errorf("missing inserted node")
|
rt[i].err = errors.New("missing inserted node")
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
for del := range tr.tracer.deletes {
|
for del := range tr.tracer.deletes {
|
||||||
if _, present := deleteExp[del]; !present {
|
if _, present := deleteExp[del]; !present {
|
||||||
rt[i].err = fmt.Errorf("missing deleted node")
|
rt[i].err = errors.New("missing deleted node")
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -215,7 +215,7 @@ func (m *meta) encode() []byte {
|
|||||||
// decode unpacks the meta object from byte stream.
|
// decode unpacks the meta object from byte stream.
|
||||||
func (m *meta) decode(blob []byte) error {
|
func (m *meta) decode(blob []byte) error {
|
||||||
if len(blob) < 1 {
|
if len(blob) < 1 {
|
||||||
return fmt.Errorf("no version tag")
|
return errors.New("no version tag")
|
||||||
}
|
}
|
||||||
switch blob[0] {
|
switch blob[0] {
|
||||||
case stateHistoryVersion:
|
case stateHistoryVersion:
|
||||||
|
Loading…
Reference in New Issue
Block a user