2016-09-25 20:49:02 +02:00
|
|
|
// Copyright 2016 The go-ethereum Authors
|
|
|
|
// This file is part of the go-ethereum library.
|
|
|
|
//
|
|
|
|
// The go-ethereum library is free software: you can redistribute it and/or modify
|
|
|
|
// it under the terms of the GNU Lesser General Public License as published by
|
|
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
|
|
// (at your option) any later version.
|
|
|
|
//
|
|
|
|
// The go-ethereum library is distributed in the hope that it will be useful,
|
|
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
|
|
// GNU Lesser General Public License for more details.
|
|
|
|
//
|
|
|
|
// You should have received a copy of the GNU Lesser General Public License
|
|
|
|
// along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
|
|
|
|
package trie
|
|
|
|
|
|
|
|
import (
|
|
|
|
"hash"
|
|
|
|
"sync"
|
|
|
|
|
|
|
|
"github.com/ethereum/go-ethereum/common"
|
|
|
|
"github.com/ethereum/go-ethereum/crypto/sha3"
|
|
|
|
"github.com/ethereum/go-ethereum/rlp"
|
|
|
|
)
|
|
|
|
|
2018-01-15 15:32:14 +02:00
|
|
|
type hasher struct {
|
2018-06-05 14:06:29 +02:00
|
|
|
tmp sliceBuffer
|
|
|
|
sha keccakState
|
2018-02-05 18:40:32 +02:00
|
|
|
cachegen uint16
|
|
|
|
cachelimit uint16
|
|
|
|
onleaf LeafCallback
|
2016-09-25 20:49:02 +02:00
|
|
|
}
|
|
|
|
|
2018-06-05 14:06:29 +02:00
|
|
|
// keccakState wraps sha3.state. In addition to the usual hash methods, it also supports
|
|
|
|
// Read to get a variable amount of data from the hash state. Read is faster than Sum
|
|
|
|
// because it doesn't copy the internal state, but also modifies the internal state.
|
|
|
|
type keccakState interface {
|
|
|
|
hash.Hash
|
|
|
|
Read([]byte) (int, error)
|
|
|
|
}
|
|
|
|
|
|
|
|
type sliceBuffer []byte
|
|
|
|
|
|
|
|
func (b *sliceBuffer) Write(data []byte) (n int, err error) {
|
|
|
|
*b = append(*b, data...)
|
|
|
|
return len(data), nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func (b *sliceBuffer) Reset() {
|
|
|
|
*b = (*b)[:0]
|
|
|
|
}
|
|
|
|
|
2018-02-05 18:40:32 +02:00
|
|
|
// hashers live in a global db.
|
2018-01-15 15:32:14 +02:00
|
|
|
var hasherPool = sync.Pool{
|
2016-09-25 20:49:02 +02:00
|
|
|
New: func() interface{} {
|
2018-06-05 14:06:29 +02:00
|
|
|
return &hasher{
|
|
|
|
tmp: make(sliceBuffer, 0, 550), // cap is as large as a full fullNode.
|
|
|
|
sha: sha3.NewKeccak256().(keccakState),
|
|
|
|
}
|
2016-09-25 20:49:02 +02:00
|
|
|
},
|
|
|
|
}
|
|
|
|
|
2018-02-05 18:40:32 +02:00
|
|
|
func newHasher(cachegen, cachelimit uint16, onleaf LeafCallback) *hasher {
|
2018-01-15 15:32:14 +02:00
|
|
|
h := hasherPool.Get().(*hasher)
|
2018-02-05 18:40:32 +02:00
|
|
|
h.cachegen, h.cachelimit, h.onleaf = cachegen, cachelimit, onleaf
|
2016-10-14 18:04:33 +02:00
|
|
|
return h
|
2016-09-25 20:49:02 +02:00
|
|
|
}
|
|
|
|
|
2018-01-15 15:32:14 +02:00
|
|
|
func returnHasherToPool(h *hasher) {
|
|
|
|
hasherPool.Put(h)
|
2016-09-25 20:49:02 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
// hash collapses a node down into a hash node, also returning a copy of the
|
2017-06-12 14:45:17 +02:00
|
|
|
// original node initialized with the computed hash to replace the original one.
|
2018-02-05 18:40:32 +02:00
|
|
|
func (h *hasher) hash(n node, db *Database, force bool) (node, node, error) {
|
2017-01-06 19:44:35 +02:00
|
|
|
// If we're not storing the node, just hashing, use available cached data
|
2016-10-14 18:04:33 +02:00
|
|
|
if hash, dirty := n.cache(); hash != nil {
|
|
|
|
if db == nil {
|
|
|
|
return hash, n, nil
|
|
|
|
}
|
|
|
|
if n.canUnload(h.cachegen, h.cachelimit) {
|
2016-10-17 16:13:50 +02:00
|
|
|
// Unload the node from cache. All of its subnodes will have a lower or equal
|
2016-10-14 18:04:33 +02:00
|
|
|
// cache generation number.
|
2016-10-19 16:07:19 +02:00
|
|
|
cacheUnloadCounter.Inc(1)
|
2016-10-14 18:04:33 +02:00
|
|
|
return hash, hash, nil
|
|
|
|
}
|
|
|
|
if !dirty {
|
|
|
|
return hash, n, nil
|
|
|
|
}
|
2016-09-25 20:49:02 +02:00
|
|
|
}
|
|
|
|
// Trie not processed yet or needs storage, walk the children
|
|
|
|
collapsed, cached, err := h.hashChildren(n, db)
|
|
|
|
if err != nil {
|
|
|
|
return hashNode{}, n, err
|
|
|
|
}
|
|
|
|
hashed, err := h.store(collapsed, db, force)
|
|
|
|
if err != nil {
|
|
|
|
return hashNode{}, n, err
|
|
|
|
}
|
2017-09-19 00:07:19 +03:00
|
|
|
// Cache the hash of the node for later reuse and remove
|
2016-10-17 23:01:29 +02:00
|
|
|
// the dirty flag in commit mode. It's fine to assign these values directly
|
|
|
|
// without copying the node first because hashChildren copies it.
|
|
|
|
cachedHash, _ := hashed.(hashNode)
|
|
|
|
switch cn := cached.(type) {
|
|
|
|
case *shortNode:
|
|
|
|
cn.flags.hash = cachedHash
|
|
|
|
if db != nil {
|
|
|
|
cn.flags.dirty = false
|
|
|
|
}
|
|
|
|
case *fullNode:
|
|
|
|
cn.flags.hash = cachedHash
|
|
|
|
if db != nil {
|
|
|
|
cn.flags.dirty = false
|
2016-09-25 20:49:02 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
return hashed, cached, nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// hashChildren replaces the children of a node with their hashes if the encoded
|
|
|
|
// size of the child is larger than a hash, returning the collapsed node as well
|
|
|
|
// as a replacement for the original node with the child hashes cached in.
|
2018-02-05 18:40:32 +02:00
|
|
|
func (h *hasher) hashChildren(original node, db *Database) (node, node, error) {
|
2016-09-25 20:49:02 +02:00
|
|
|
var err error
|
|
|
|
|
|
|
|
switch n := original.(type) {
|
2016-10-14 18:04:33 +02:00
|
|
|
case *shortNode:
|
2016-09-25 20:49:02 +02:00
|
|
|
// Hash the short node's child, caching the newly hashed subtree
|
2016-10-14 18:04:33 +02:00
|
|
|
collapsed, cached := n.copy(), n.copy()
|
2017-04-18 13:25:07 +02:00
|
|
|
collapsed.Key = hexToCompact(n.Key)
|
2016-10-14 18:04:33 +02:00
|
|
|
cached.Key = common.CopyBytes(n.Key)
|
2016-09-25 20:49:02 +02:00
|
|
|
|
|
|
|
if _, ok := n.Val.(valueNode); !ok {
|
2016-10-14 18:04:33 +02:00
|
|
|
collapsed.Val, cached.Val, err = h.hash(n.Val, db, false)
|
|
|
|
if err != nil {
|
|
|
|
return original, original, err
|
2016-09-25 20:49:02 +02:00
|
|
|
}
|
|
|
|
}
|
2016-10-14 18:04:33 +02:00
|
|
|
return collapsed, cached, nil
|
2016-09-25 20:49:02 +02:00
|
|
|
|
2016-10-14 18:04:33 +02:00
|
|
|
case *fullNode:
|
2016-09-25 20:49:02 +02:00
|
|
|
// Hash the full node's children, caching the newly hashed subtrees
|
2016-10-14 18:04:33 +02:00
|
|
|
collapsed, cached := n.copy(), n.copy()
|
2016-09-25 20:49:02 +02:00
|
|
|
|
2018-01-15 15:32:14 +02:00
|
|
|
for i := 0; i < 16; i++ {
|
|
|
|
if n.Children[i] != nil {
|
|
|
|
collapsed.Children[i], cached.Children[i], err = h.hash(n.Children[i], db, false)
|
|
|
|
if err != nil {
|
|
|
|
return original, original, err
|
|
|
|
}
|
2016-09-25 20:49:02 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
cached.Children[16] = n.Children[16]
|
2016-10-14 18:04:33 +02:00
|
|
|
return collapsed, cached, nil
|
2016-09-25 20:49:02 +02:00
|
|
|
|
|
|
|
default:
|
|
|
|
// Value and hash nodes don't have children so they're left as were
|
|
|
|
return n, original, nil
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2018-02-05 18:40:32 +02:00
|
|
|
// store hashes the node n and if we have a storage layer specified, it writes
|
|
|
|
// the key/value pair to it and tracks any node->child references as well as any
|
|
|
|
// node->external trie references.
|
|
|
|
func (h *hasher) store(n node, db *Database, force bool) (node, error) {
|
2016-09-25 20:49:02 +02:00
|
|
|
// Don't store hashes or empty nodes.
|
|
|
|
if _, isHash := n.(hashNode); n == nil || isHash {
|
|
|
|
return n, nil
|
|
|
|
}
|
|
|
|
// Generate the RLP encoding of the node
|
2018-01-15 15:32:14 +02:00
|
|
|
h.tmp.Reset()
|
2018-06-05 14:06:29 +02:00
|
|
|
if err := rlp.Encode(&h.tmp, n); err != nil {
|
2016-09-25 20:49:02 +02:00
|
|
|
panic("encode error: " + err.Error())
|
|
|
|
}
|
2018-06-05 14:06:29 +02:00
|
|
|
if len(h.tmp) < 32 && !force {
|
2016-09-25 20:49:02 +02:00
|
|
|
return n, nil // Nodes smaller than 32 bytes are stored inside their parent
|
|
|
|
}
|
|
|
|
// Larger nodes are replaced by their hash and stored in the database.
|
|
|
|
hash, _ := n.cache()
|
|
|
|
if hash == nil {
|
2018-06-05 14:06:29 +02:00
|
|
|
hash = h.makeHashNode(h.tmp)
|
2016-09-25 20:49:02 +02:00
|
|
|
}
|
2018-06-05 14:06:29 +02:00
|
|
|
|
2016-09-25 20:49:02 +02:00
|
|
|
if db != nil {
|
2018-02-05 18:40:32 +02:00
|
|
|
// We are pooling the trie nodes into an intermediate memory cache
|
|
|
|
hash := common.BytesToHash(hash)
|
2018-06-21 12:28:05 +03:00
|
|
|
|
|
|
|
db.lock.Lock()
|
|
|
|
db.insert(hash, h.tmp, n)
|
2018-02-05 18:40:32 +02:00
|
|
|
db.lock.Unlock()
|
|
|
|
|
|
|
|
// Track external references from account->storage trie
|
|
|
|
if h.onleaf != nil {
|
|
|
|
switch n := n.(type) {
|
|
|
|
case *shortNode:
|
2018-06-21 12:28:05 +03:00
|
|
|
if child, ok := n.Val.(valueNode); ok {
|
2018-02-05 18:40:32 +02:00
|
|
|
h.onleaf(child, hash)
|
|
|
|
}
|
|
|
|
case *fullNode:
|
|
|
|
for i := 0; i < 16; i++ {
|
2018-06-21 12:28:05 +03:00
|
|
|
if child, ok := n.Children[i].(valueNode); ok {
|
2018-02-05 18:40:32 +02:00
|
|
|
h.onleaf(child, hash)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2016-09-25 20:49:02 +02:00
|
|
|
}
|
|
|
|
return hash, nil
|
|
|
|
}
|
2018-06-05 14:06:29 +02:00
|
|
|
|
|
|
|
func (h *hasher) makeHashNode(data []byte) hashNode {
|
|
|
|
n := make(hashNode, h.sha.Size())
|
|
|
|
h.sha.Reset()
|
|
|
|
h.sha.Write(data)
|
|
|
|
h.sha.Read(n)
|
|
|
|
return n
|
|
|
|
}
|