2020-12-14 12:27:15 +03:00
|
|
|
// Copyright 2020 The go-ethereum Authors
|
|
|
|
// This file is part of the go-ethereum library.
|
|
|
|
//
|
|
|
|
// The go-ethereum library is free software: you can redistribute it and/or modify
|
|
|
|
// it under the terms of the GNU Lesser General Public License as published by
|
|
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
|
|
// (at your option) any later version.
|
|
|
|
//
|
|
|
|
// The go-ethereum library is distributed in the hope that it will be useful,
|
|
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
|
|
// GNU Lesser General Public License for more details.
|
|
|
|
//
|
|
|
|
// You should have received a copy of the GNU Lesser General Public License
|
|
|
|
// along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
|
|
|
|
package snap
|
|
|
|
|
|
|
|
import (
|
|
|
|
"bytes"
|
|
|
|
"encoding/json"
|
|
|
|
"errors"
|
|
|
|
"fmt"
|
2022-09-09 11:42:57 +03:00
|
|
|
gomath "math"
|
2020-12-14 12:27:15 +03:00
|
|
|
"math/big"
|
|
|
|
"math/rand"
|
2021-04-27 17:19:59 +03:00
|
|
|
"sort"
|
2020-12-14 12:27:15 +03:00
|
|
|
"sync"
|
2022-09-09 11:42:57 +03:00
|
|
|
"sync/atomic"
|
2020-12-14 12:27:15 +03:00
|
|
|
"time"
|
|
|
|
|
|
|
|
"github.com/ethereum/go-ethereum/common"
|
2021-04-27 17:19:59 +03:00
|
|
|
"github.com/ethereum/go-ethereum/common/math"
|
2020-12-14 12:27:15 +03:00
|
|
|
"github.com/ethereum/go-ethereum/core/rawdb"
|
|
|
|
"github.com/ethereum/go-ethereum/core/state"
|
2021-09-28 11:48:07 +03:00
|
|
|
"github.com/ethereum/go-ethereum/core/types"
|
2020-12-14 12:27:15 +03:00
|
|
|
"github.com/ethereum/go-ethereum/crypto"
|
|
|
|
"github.com/ethereum/go-ethereum/ethdb"
|
|
|
|
"github.com/ethereum/go-ethereum/event"
|
|
|
|
"github.com/ethereum/go-ethereum/log"
|
2021-05-19 15:09:03 +03:00
|
|
|
"github.com/ethereum/go-ethereum/p2p/msgrate"
|
2020-12-14 12:27:15 +03:00
|
|
|
"github.com/ethereum/go-ethereum/rlp"
|
|
|
|
"github.com/ethereum/go-ethereum/trie"
|
2023-10-10 11:30:47 +03:00
|
|
|
"github.com/ethereum/go-ethereum/trie/trienode"
|
2020-12-14 12:27:15 +03:00
|
|
|
"golang.org/x/crypto/sha3"
|
|
|
|
)
|
|
|
|
|
|
|
|
const (
|
2021-05-19 15:09:03 +03:00
|
|
|
// minRequestSize is the minimum number of bytes to request from a remote peer.
|
|
|
|
// This number is used as the low cap for account and storage range requests.
|
|
|
|
// Bytecode and trienode are limited inherently by item count (1).
|
|
|
|
minRequestSize = 64 * 1024
|
2020-12-14 12:27:15 +03:00
|
|
|
|
2021-05-19 15:09:03 +03:00
|
|
|
// maxRequestSize is the maximum number of bytes to request from a remote peer.
|
|
|
|
// This number is used as the high cap for account and storage range requests.
|
|
|
|
// Bytecode and trienode are limited more explicitly by the caps below.
|
|
|
|
maxRequestSize = 512 * 1024
|
2020-12-14 12:27:15 +03:00
|
|
|
|
|
|
|
// maxCodeRequestCount is the maximum number of bytecode blobs to request in a
|
|
|
|
// single query. If this number is too low, we're not filling responses fully
|
|
|
|
// and waste round trip times. If it's too high, we're capping responses and
|
|
|
|
// waste bandwidth.
|
|
|
|
//
|
2022-10-11 10:37:00 +03:00
|
|
|
// Deployed bytecodes are currently capped at 24KB, so the minimum request
|
2020-12-14 12:27:15 +03:00
|
|
|
// size should be maxRequestSize / 24K. Assuming that most contracts do not
|
|
|
|
// come close to that, requesting 4x should be a good approximation.
|
|
|
|
maxCodeRequestCount = maxRequestSize / (24 * 1024) * 4
|
|
|
|
|
|
|
|
// maxTrieRequestCount is the maximum number of trie node blobs to request in
|
|
|
|
// a single query. If this number is too low, we're not filling responses fully
|
|
|
|
// and waste round trip times. If it's too high, we're capping responses and
|
|
|
|
// waste bandwidth.
|
2021-05-19 15:09:03 +03:00
|
|
|
maxTrieRequestCount = maxRequestSize / 512
|
2022-09-09 11:42:57 +03:00
|
|
|
|
|
|
|
// trienodeHealRateMeasurementImpact is the impact a single measurement has on
|
|
|
|
// the local node's trienode processing capacity. A value closer to 0 reacts
|
|
|
|
// slower to sudden changes, but it is also more stable against temporary hiccups.
|
|
|
|
trienodeHealRateMeasurementImpact = 0.005
|
|
|
|
|
|
|
|
// minTrienodeHealThrottle is the minimum divisor for throttling trie node
|
2022-10-11 10:37:00 +03:00
|
|
|
// heal requests to avoid overloading the local node and excessively expanding
|
|
|
|
// the state trie breadth wise.
|
2022-09-09 11:42:57 +03:00
|
|
|
minTrienodeHealThrottle = 1
|
|
|
|
|
|
|
|
// maxTrienodeHealThrottle is the maximum divisor for throttling trie node
|
|
|
|
// heal requests to avoid overloading the local node and exessively expanding
|
|
|
|
// the state trie bedth wise.
|
|
|
|
maxTrienodeHealThrottle = maxTrieRequestCount
|
|
|
|
|
|
|
|
// trienodeHealThrottleIncrease is the multiplier for the throttle when the
|
|
|
|
// rate of arriving data is higher than the rate of processing it.
|
|
|
|
trienodeHealThrottleIncrease = 1.33
|
|
|
|
|
|
|
|
// trienodeHealThrottleDecrease is the divisor for the throttle when the
|
|
|
|
// rate of arriving data is lower than the rate of processing it.
|
|
|
|
trienodeHealThrottleDecrease = 1.25
|
2021-04-27 17:19:59 +03:00
|
|
|
)
|
2020-12-14 12:27:15 +03:00
|
|
|
|
2021-04-27 17:19:59 +03:00
|
|
|
var (
|
2020-12-14 12:27:15 +03:00
|
|
|
// accountConcurrency is the number of chunks to split the account trie into
|
|
|
|
// to allow concurrent retrievals.
|
|
|
|
accountConcurrency = 16
|
|
|
|
|
|
|
|
// storageConcurrency is the number of chunks to split the a large contract
|
|
|
|
// storage trie into to allow concurrent retrievals.
|
|
|
|
storageConcurrency = 16
|
2021-01-25 09:17:05 +03:00
|
|
|
)
|
|
|
|
|
2021-02-16 17:11:33 +03:00
|
|
|
// ErrCancelled is returned from snap syncing if the operation was prematurely
|
|
|
|
// terminated.
|
|
|
|
var ErrCancelled = errors.New("sync cancelled")
|
|
|
|
|
2020-12-14 12:27:15 +03:00
|
|
|
// accountRequest tracks a pending account range request to ensure responses are
|
|
|
|
// to actual requests and to validate any security constraints.
|
|
|
|
//
|
|
|
|
// Concurrency note: account requests and responses are handled concurrently from
|
|
|
|
// the main runloop to allow Merkle proof verifications on the peer's thread and
|
|
|
|
// to drop on invalid response. The request struct must contain all the data to
|
|
|
|
// construct the response without accessing runloop internals (i.e. task). That
|
|
|
|
// is only included to allow the runloop to match a response to the task being
|
|
|
|
// synced without having yet another set of maps.
|
|
|
|
type accountRequest struct {
|
2021-05-19 15:09:03 +03:00
|
|
|
peer string // Peer to which this request is assigned
|
|
|
|
id uint64 // Request ID of this request
|
|
|
|
time time.Time // Timestamp when the request was sent
|
2020-12-14 12:27:15 +03:00
|
|
|
|
2021-04-15 21:01:16 +03:00
|
|
|
deliver chan *accountResponse // Channel to deliver successful response on
|
|
|
|
revert chan *accountRequest // Channel to deliver request failure on
|
|
|
|
cancel chan struct{} // Channel to track sync cancellation
|
|
|
|
timeout *time.Timer // Timer to track delivery timeout
|
|
|
|
stale chan struct{} // Channel to signal the request was dropped
|
2020-12-14 12:27:15 +03:00
|
|
|
|
|
|
|
origin common.Hash // First account requested to allow continuation checks
|
|
|
|
limit common.Hash // Last account requested to allow non-overlapping chunking
|
|
|
|
|
|
|
|
task *accountTask // Task which this request is filling (only access fields through the runloop!!)
|
|
|
|
}
|
|
|
|
|
|
|
|
// accountResponse is an already Merkle-verified remote response to an account
|
|
|
|
// range request. It contains the subtrie for the requested account range and
|
|
|
|
// the database that's going to be filled with the internal nodes on commit.
|
|
|
|
type accountResponse struct {
|
|
|
|
task *accountTask // Task which this request is filling
|
|
|
|
|
2021-09-28 11:48:07 +03:00
|
|
|
hashes []common.Hash // Account hashes in the returned range
|
|
|
|
accounts []*types.StateAccount // Expanded accounts in the returned range
|
2020-12-14 12:27:15 +03:00
|
|
|
|
|
|
|
cont bool // Whether the account range has a continuation
|
|
|
|
}
|
|
|
|
|
|
|
|
// bytecodeRequest tracks a pending bytecode request to ensure responses are to
|
|
|
|
// actual requests and to validate any security constraints.
|
|
|
|
//
|
|
|
|
// Concurrency note: bytecode requests and responses are handled concurrently from
|
|
|
|
// the main runloop to allow Keccak256 hash verifications on the peer's thread and
|
|
|
|
// to drop on invalid response. The request struct must contain all the data to
|
|
|
|
// construct the response without accessing runloop internals (i.e. task). That
|
|
|
|
// is only included to allow the runloop to match a response to the task being
|
|
|
|
// synced without having yet another set of maps.
|
|
|
|
type bytecodeRequest struct {
|
2021-05-19 15:09:03 +03:00
|
|
|
peer string // Peer to which this request is assigned
|
|
|
|
id uint64 // Request ID of this request
|
|
|
|
time time.Time // Timestamp when the request was sent
|
2020-12-14 12:27:15 +03:00
|
|
|
|
2021-04-15 21:01:16 +03:00
|
|
|
deliver chan *bytecodeResponse // Channel to deliver successful response on
|
|
|
|
revert chan *bytecodeRequest // Channel to deliver request failure on
|
|
|
|
cancel chan struct{} // Channel to track sync cancellation
|
|
|
|
timeout *time.Timer // Timer to track delivery timeout
|
|
|
|
stale chan struct{} // Channel to signal the request was dropped
|
2020-12-14 12:27:15 +03:00
|
|
|
|
|
|
|
hashes []common.Hash // Bytecode hashes to validate responses
|
|
|
|
task *accountTask // Task which this request is filling (only access fields through the runloop!!)
|
|
|
|
}
|
|
|
|
|
|
|
|
// bytecodeResponse is an already verified remote response to a bytecode request.
|
|
|
|
type bytecodeResponse struct {
|
|
|
|
task *accountTask // Task which this request is filling
|
|
|
|
|
|
|
|
hashes []common.Hash // Hashes of the bytecode to avoid double hashing
|
|
|
|
codes [][]byte // Actual bytecodes to store into the database (nil = missing)
|
|
|
|
}
|
|
|
|
|
|
|
|
// storageRequest tracks a pending storage ranges request to ensure responses are
|
|
|
|
// to actual requests and to validate any security constraints.
|
|
|
|
//
|
|
|
|
// Concurrency note: storage requests and responses are handled concurrently from
|
2022-03-11 11:32:08 +03:00
|
|
|
// the main runloop to allow Merkle proof verifications on the peer's thread and
|
2020-12-14 12:27:15 +03:00
|
|
|
// to drop on invalid response. The request struct must contain all the data to
|
|
|
|
// construct the response without accessing runloop internals (i.e. tasks). That
|
|
|
|
// is only included to allow the runloop to match a response to the task being
|
|
|
|
// synced without having yet another set of maps.
|
|
|
|
type storageRequest struct {
|
2021-05-19 15:09:03 +03:00
|
|
|
peer string // Peer to which this request is assigned
|
|
|
|
id uint64 // Request ID of this request
|
|
|
|
time time.Time // Timestamp when the request was sent
|
2020-12-14 12:27:15 +03:00
|
|
|
|
2021-04-15 21:01:16 +03:00
|
|
|
deliver chan *storageResponse // Channel to deliver successful response on
|
|
|
|
revert chan *storageRequest // Channel to deliver request failure on
|
|
|
|
cancel chan struct{} // Channel to track sync cancellation
|
|
|
|
timeout *time.Timer // Timer to track delivery timeout
|
|
|
|
stale chan struct{} // Channel to signal the request was dropped
|
2020-12-14 12:27:15 +03:00
|
|
|
|
|
|
|
accounts []common.Hash // Account hashes to validate responses
|
|
|
|
roots []common.Hash // Storage roots to validate responses
|
|
|
|
|
|
|
|
origin common.Hash // First storage slot requested to allow continuation checks
|
|
|
|
limit common.Hash // Last storage slot requested to allow non-overlapping chunking
|
|
|
|
|
|
|
|
mainTask *accountTask // Task which this response belongs to (only access fields through the runloop!!)
|
|
|
|
subTask *storageTask // Task which this response is filling (only access fields through the runloop!!)
|
|
|
|
}
|
|
|
|
|
|
|
|
// storageResponse is an already Merkle-verified remote response to a storage
|
|
|
|
// range request. It contains the subtries for the requested storage ranges and
|
|
|
|
// the databases that's going to be filled with the internal nodes on commit.
|
|
|
|
type storageResponse struct {
|
|
|
|
mainTask *accountTask // Task which this response belongs to
|
|
|
|
subTask *storageTask // Task which this response is filling
|
|
|
|
|
|
|
|
accounts []common.Hash // Account hashes requested, may be only partially filled
|
|
|
|
roots []common.Hash // Storage roots requested, may be only partially filled
|
|
|
|
|
2021-04-28 23:09:15 +03:00
|
|
|
hashes [][]common.Hash // Storage slot hashes in the returned range
|
|
|
|
slots [][][]byte // Storage slot values in the returned range
|
2020-12-14 12:27:15 +03:00
|
|
|
|
2021-04-27 17:19:59 +03:00
|
|
|
cont bool // Whether the last storage range has a continuation
|
2020-12-14 12:27:15 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// trienodeHealRequest tracks a pending state trie request to ensure responses
|
|
|
|
// are to actual requests and to validate any security constraints.
|
|
|
|
//
|
|
|
|
// Concurrency note: trie node requests and responses are handled concurrently from
|
|
|
|
// the main runloop to allow Keccak256 hash verifications on the peer's thread and
|
|
|
|
// to drop on invalid response. The request struct must contain all the data to
|
|
|
|
// construct the response without accessing runloop internals (i.e. task). That
|
|
|
|
// is only included to allow the runloop to match a response to the task being
|
|
|
|
// synced without having yet another set of maps.
|
|
|
|
type trienodeHealRequest struct {
|
2021-05-19 15:09:03 +03:00
|
|
|
peer string // Peer to which this request is assigned
|
|
|
|
id uint64 // Request ID of this request
|
|
|
|
time time.Time // Timestamp when the request was sent
|
2020-12-14 12:27:15 +03:00
|
|
|
|
2021-04-15 21:01:16 +03:00
|
|
|
deliver chan *trienodeHealResponse // Channel to deliver successful response on
|
|
|
|
revert chan *trienodeHealRequest // Channel to deliver request failure on
|
|
|
|
cancel chan struct{} // Channel to track sync cancellation
|
|
|
|
timeout *time.Timer // Timer to track delivery timeout
|
|
|
|
stale chan struct{} // Channel to signal the request was dropped
|
2020-12-14 12:27:15 +03:00
|
|
|
|
2022-07-15 14:55:51 +03:00
|
|
|
paths []string // Trie node paths for identifying trie node
|
|
|
|
hashes []common.Hash // Trie node hashes to validate responses
|
2020-12-14 12:27:15 +03:00
|
|
|
|
|
|
|
task *healTask // Task which this request is filling (only access fields through the runloop!!)
|
|
|
|
}
|
|
|
|
|
|
|
|
// trienodeHealResponse is an already verified remote response to a trie node request.
|
|
|
|
type trienodeHealResponse struct {
|
|
|
|
task *healTask // Task which this request is filling
|
|
|
|
|
2022-07-15 14:55:51 +03:00
|
|
|
paths []string // Paths of the trie nodes
|
|
|
|
hashes []common.Hash // Hashes of the trie nodes to avoid double hashing
|
|
|
|
nodes [][]byte // Actual trie nodes to store into the database (nil = missing)
|
2020-12-14 12:27:15 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// bytecodeHealRequest tracks a pending bytecode request to ensure responses are to
|
|
|
|
// actual requests and to validate any security constraints.
|
|
|
|
//
|
|
|
|
// Concurrency note: bytecode requests and responses are handled concurrently from
|
|
|
|
// the main runloop to allow Keccak256 hash verifications on the peer's thread and
|
|
|
|
// to drop on invalid response. The request struct must contain all the data to
|
|
|
|
// construct the response without accessing runloop internals (i.e. task). That
|
|
|
|
// is only included to allow the runloop to match a response to the task being
|
|
|
|
// synced without having yet another set of maps.
|
|
|
|
type bytecodeHealRequest struct {
|
2021-05-19 15:09:03 +03:00
|
|
|
peer string // Peer to which this request is assigned
|
|
|
|
id uint64 // Request ID of this request
|
|
|
|
time time.Time // Timestamp when the request was sent
|
2020-12-14 12:27:15 +03:00
|
|
|
|
2021-04-15 21:01:16 +03:00
|
|
|
deliver chan *bytecodeHealResponse // Channel to deliver successful response on
|
|
|
|
revert chan *bytecodeHealRequest // Channel to deliver request failure on
|
|
|
|
cancel chan struct{} // Channel to track sync cancellation
|
|
|
|
timeout *time.Timer // Timer to track delivery timeout
|
|
|
|
stale chan struct{} // Channel to signal the request was dropped
|
2020-12-14 12:27:15 +03:00
|
|
|
|
|
|
|
hashes []common.Hash // Bytecode hashes to validate responses
|
|
|
|
task *healTask // Task which this request is filling (only access fields through the runloop!!)
|
|
|
|
}
|
|
|
|
|
|
|
|
// bytecodeHealResponse is an already verified remote response to a bytecode request.
|
|
|
|
type bytecodeHealResponse struct {
|
|
|
|
task *healTask // Task which this request is filling
|
|
|
|
|
|
|
|
hashes []common.Hash // Hashes of the bytecode to avoid double hashing
|
|
|
|
codes [][]byte // Actual bytecodes to store into the database (nil = missing)
|
|
|
|
}
|
|
|
|
|
|
|
|
// accountTask represents the sync task for a chunk of the account snapshot.
|
|
|
|
type accountTask struct {
|
|
|
|
// These fields get serialized to leveldb on shutdown
|
|
|
|
Next common.Hash // Next account to sync in this interval
|
|
|
|
Last common.Hash // Last account to sync in this interval
|
|
|
|
SubTasks map[common.Hash][]*storageTask // Storage intervals needing fetching for large contracts
|
|
|
|
|
|
|
|
// These fields are internals used during runtime
|
|
|
|
req *accountRequest // Pending request to fill this task
|
|
|
|
res *accountResponse // Validate response filling this task
|
|
|
|
pend int // Number of pending subtasks for this round
|
|
|
|
|
|
|
|
needCode []bool // Flags whether the filling accounts need code retrieval
|
|
|
|
needState []bool // Flags whether the filling accounts need storage retrieval
|
|
|
|
needHeal []bool // Flags whether the filling accounts's state was chunked and need healing
|
|
|
|
|
|
|
|
codeTasks map[common.Hash]struct{} // Code hashes that need retrieval
|
|
|
|
stateTasks map[common.Hash]common.Hash // Account hashes->roots that need full state retrieval
|
|
|
|
|
2021-04-27 17:19:59 +03:00
|
|
|
genBatch ethdb.Batch // Batch used by the node generator
|
|
|
|
genTrie *trie.StackTrie // Node generator from storage slots
|
|
|
|
|
2020-12-14 12:27:15 +03:00
|
|
|
done bool // Flag whether the task can be removed
|
|
|
|
}
|
|
|
|
|
|
|
|
// storageTask represents the sync task for a chunk of the storage snapshot.
|
|
|
|
type storageTask struct {
|
|
|
|
Next common.Hash // Next account to sync in this interval
|
|
|
|
Last common.Hash // Last account to sync in this interval
|
|
|
|
|
|
|
|
// These fields are internals used during runtime
|
|
|
|
root common.Hash // Storage root hash for this instance
|
|
|
|
req *storageRequest // Pending request to fill this task
|
2021-04-27 17:19:59 +03:00
|
|
|
|
|
|
|
genBatch ethdb.Batch // Batch used by the node generator
|
|
|
|
genTrie *trie.StackTrie // Node generator from storage slots
|
|
|
|
|
|
|
|
done bool // Flag whether the task can be removed
|
2020-12-14 12:27:15 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// healTask represents the sync task for healing the snap-synced chunk boundaries.
|
|
|
|
type healTask struct {
|
|
|
|
scheduler *trie.Sync // State trie sync scheduler defining the tasks
|
|
|
|
|
2022-07-15 14:55:51 +03:00
|
|
|
trieTasks map[string]common.Hash // Set of trie node tasks currently queued for retrieval, indexed by node path
|
|
|
|
codeTasks map[common.Hash]struct{} // Set of byte code tasks currently queued for retrieval, indexed by code hash
|
2020-12-14 12:27:15 +03:00
|
|
|
}
|
|
|
|
|
2021-11-26 14:26:03 +03:00
|
|
|
// SyncProgress is a database entry to allow suspending and resuming a snapshot state
|
2020-12-14 12:27:15 +03:00
|
|
|
// sync. Opposed to full and fast sync, there is no way to restart a suspended
|
|
|
|
// snap sync without prior knowledge of the suspension point.
|
2021-11-26 14:26:03 +03:00
|
|
|
type SyncProgress struct {
|
2020-12-14 12:27:15 +03:00
|
|
|
Tasks []*accountTask // The suspended account tasks (contract tasks within)
|
|
|
|
|
|
|
|
// Status report during syncing phase
|
|
|
|
AccountSynced uint64 // Number of accounts downloaded
|
|
|
|
AccountBytes common.StorageSize // Number of account trie bytes persisted to disk
|
|
|
|
BytecodeSynced uint64 // Number of bytecodes downloaded
|
|
|
|
BytecodeBytes common.StorageSize // Number of bytecode bytes downloaded
|
|
|
|
StorageSynced uint64 // Number of storage slots downloaded
|
|
|
|
StorageBytes common.StorageSize // Number of storage trie bytes persisted to disk
|
|
|
|
|
|
|
|
// Status report during healing phase
|
|
|
|
TrienodeHealSynced uint64 // Number of state trie nodes downloaded
|
|
|
|
TrienodeHealBytes common.StorageSize // Number of state trie bytes persisted to disk
|
|
|
|
BytecodeHealSynced uint64 // Number of bytecodes downloaded
|
|
|
|
BytecodeHealBytes common.StorageSize // Number of bytecodes persisted to disk
|
2021-11-26 14:26:03 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// SyncPending is analogous to SyncProgress, but it's used to report on pending
|
|
|
|
// ephemeral sync progress that doesn't get persisted into the database.
|
|
|
|
type SyncPending struct {
|
|
|
|
TrienodeHeal uint64 // Number of state trie nodes pending
|
|
|
|
BytecodeHeal uint64 // Number of bytecodes pending
|
2020-12-14 12:27:15 +03:00
|
|
|
}
|
|
|
|
|
2021-01-25 09:17:05 +03:00
|
|
|
// SyncPeer abstracts out the methods required for a peer to be synced against
|
|
|
|
// with the goal of allowing the construction of mock peers without the full
|
|
|
|
// blown networking.
|
|
|
|
type SyncPeer interface {
|
|
|
|
// ID retrieves the peer's unique identifier.
|
|
|
|
ID() string
|
|
|
|
|
|
|
|
// RequestAccountRange fetches a batch of accounts rooted in a specific account
|
|
|
|
// trie, starting with the origin.
|
|
|
|
RequestAccountRange(id uint64, root, origin, limit common.Hash, bytes uint64) error
|
|
|
|
|
2021-04-27 17:19:59 +03:00
|
|
|
// RequestStorageRanges fetches a batch of storage slots belonging to one or
|
2022-08-19 09:00:21 +03:00
|
|
|
// more accounts. If slots from only one account is requested, an origin marker
|
2021-01-25 09:17:05 +03:00
|
|
|
// may also be used to retrieve from there.
|
|
|
|
RequestStorageRanges(id uint64, root common.Hash, accounts []common.Hash, origin, limit []byte, bytes uint64) error
|
|
|
|
|
|
|
|
// RequestByteCodes fetches a batch of bytecodes by hash.
|
|
|
|
RequestByteCodes(id uint64, hashes []common.Hash, bytes uint64) error
|
|
|
|
|
|
|
|
// RequestTrieNodes fetches a batch of account or storage trie nodes rooted in
|
2022-08-19 09:00:21 +03:00
|
|
|
// a specific state trie.
|
2021-01-25 09:17:05 +03:00
|
|
|
RequestTrieNodes(id uint64, root common.Hash, paths []TrieNodePathSet, bytes uint64) error
|
|
|
|
|
|
|
|
// Log retrieves the peer's own contextual logger.
|
|
|
|
Log() log.Logger
|
|
|
|
}
|
|
|
|
|
2020-12-14 12:27:15 +03:00
|
|
|
// Syncer is an Ethereum account and storage trie syncer based on snapshots and
|
|
|
|
// the snap protocol. It's purpose is to download all the accounts and storage
|
|
|
|
// slots from remote peers and reassemble chunks of the state trie, on top of
|
|
|
|
// which a state sync can be run to fix any gaps / overlaps.
|
|
|
|
//
|
|
|
|
// Every network request has a variety of failure events:
|
|
|
|
// - The peer disconnects after task assignment, failing to send the request
|
|
|
|
// - The peer disconnects after sending the request, before delivering on it
|
|
|
|
// - The peer remains connected, but does not deliver a response in time
|
|
|
|
// - The peer delivers a stale response after a previous timeout
|
|
|
|
// - The peer delivers a refusal to serve the requested state
|
|
|
|
type Syncer struct {
|
2022-11-28 16:31:28 +03:00
|
|
|
db ethdb.KeyValueStore // Database to store the trie nodes into (and dedup)
|
2023-02-06 18:28:40 +03:00
|
|
|
scheme string // Node scheme used in node database
|
2020-12-14 12:27:15 +03:00
|
|
|
|
2021-01-25 09:17:05 +03:00
|
|
|
root common.Hash // Current state trie root being synced
|
|
|
|
tasks []*accountTask // Current account task set being synced
|
|
|
|
snapped bool // Flag to signal that snap phase is done
|
|
|
|
healer *healTask // Current state healing task being executed
|
|
|
|
update chan struct{} // Notification channel for possible sync progression
|
2020-12-14 12:27:15 +03:00
|
|
|
|
2021-01-25 09:17:05 +03:00
|
|
|
peers map[string]SyncPeer // Currently active peers to download from
|
|
|
|
peerJoin *event.Feed // Event feed to react to peers joining
|
|
|
|
peerDrop *event.Feed // Event feed to react to peers dropping
|
2021-05-19 15:09:03 +03:00
|
|
|
rates *msgrate.Trackers // Message throughput rates for peers
|
2020-12-14 12:27:15 +03:00
|
|
|
|
|
|
|
// Request tracking during syncing phase
|
|
|
|
statelessPeers map[string]struct{} // Peers that failed to deliver state data
|
|
|
|
accountIdlers map[string]struct{} // Peers that aren't serving account requests
|
|
|
|
bytecodeIdlers map[string]struct{} // Peers that aren't serving bytecode requests
|
|
|
|
storageIdlers map[string]struct{} // Peers that aren't serving storage requests
|
|
|
|
|
|
|
|
accountReqs map[uint64]*accountRequest // Account requests currently running
|
|
|
|
bytecodeReqs map[uint64]*bytecodeRequest // Bytecode requests currently running
|
|
|
|
storageReqs map[uint64]*storageRequest // Storage requests currently running
|
|
|
|
|
|
|
|
accountSynced uint64 // Number of accounts downloaded
|
|
|
|
accountBytes common.StorageSize // Number of account trie bytes persisted to disk
|
|
|
|
bytecodeSynced uint64 // Number of bytecodes downloaded
|
|
|
|
bytecodeBytes common.StorageSize // Number of bytecode bytes downloaded
|
|
|
|
storageSynced uint64 // Number of storage slots downloaded
|
|
|
|
storageBytes common.StorageSize // Number of storage trie bytes persisted to disk
|
|
|
|
|
2022-05-06 18:20:41 +03:00
|
|
|
extProgress *SyncProgress // progress that can be exposed to external caller.
|
|
|
|
|
2020-12-14 12:27:15 +03:00
|
|
|
// Request tracking during healing phase
|
|
|
|
trienodeHealIdlers map[string]struct{} // Peers that aren't serving trie node requests
|
|
|
|
bytecodeHealIdlers map[string]struct{} // Peers that aren't serving bytecode requests
|
|
|
|
|
|
|
|
trienodeHealReqs map[uint64]*trienodeHealRequest // Trie node requests currently running
|
|
|
|
bytecodeHealReqs map[uint64]*bytecodeHealRequest // Bytecode requests currently running
|
|
|
|
|
2023-04-25 13:06:50 +03:00
|
|
|
trienodeHealRate float64 // Average heal rate for processing trie node data
|
|
|
|
trienodeHealPend atomic.Uint64 // Number of trie nodes currently pending for processing
|
|
|
|
trienodeHealThrottle float64 // Divisor for throttling the amount of trienode heal data requested
|
|
|
|
trienodeHealThrottled time.Time // Timestamp the last time the throttle was updated
|
2022-09-09 11:42:57 +03:00
|
|
|
|
2020-12-14 12:27:15 +03:00
|
|
|
trienodeHealSynced uint64 // Number of state trie nodes downloaded
|
|
|
|
trienodeHealBytes common.StorageSize // Number of state trie bytes persisted to disk
|
|
|
|
trienodeHealDups uint64 // Number of state trie nodes already processed
|
|
|
|
trienodeHealNops uint64 // Number of state trie nodes not requested
|
|
|
|
bytecodeHealSynced uint64 // Number of bytecodes downloaded
|
|
|
|
bytecodeHealBytes common.StorageSize // Number of bytecodes persisted to disk
|
|
|
|
bytecodeHealDups uint64 // Number of bytecodes already processed
|
|
|
|
bytecodeHealNops uint64 // Number of bytecodes not requested
|
|
|
|
|
core, eth: faster snapshot generation (#22504)
* eth/protocols: persist received state segments
* core: initial implementation
* core/state/snapshot: add tests
* core, eth: updates
* eth/protocols/snapshot: count flat state size
* core/state: add metrics
* core/state/snapshot: skip unnecessary deletion
* core/state/snapshot: rename
* core/state/snapshot: use the global batch
* core/state/snapshot: add logs and fix wiping
* core/state/snapshot: fix
* core/state/snapshot: save generation progress even if the batch is empty
* core/state/snapshot: fixes
* core/state/snapshot: fix initial account range length
* core/state/snapshot: fix initial account range
* eth/protocols/snap: store flat states during the healing
* eth/protocols/snap: print logs
* core/state/snapshot: refactor (#4)
* core/state/snapshot: refactor
* core/state/snapshot: tiny fix and polish
Co-authored-by: rjl493456442 <garyrong0905@gmail.com>
* core, eth: fixes
* core, eth: fix healing writer
* core, trie, eth: fix paths
* eth/protocols/snap: fix encoding
* eth, core: add debug log
* core/state/generate: release iterator asap (#5)
core/state/snapshot: less copy
core/state/snapshot: revert split loop
core/state/snapshot: handle storage becoming empty, improve test robustness
core/state: test modified codehash
core/state/snapshot: polish
* core/state/snapshot: optimize stats counter
* core, eth: add metric
* core/state/snapshot: update comments
* core/state/snapshot: improve tests
* core/state/snapshot: replace secure trie with standard trie
* core/state/snapshot: wrap return as the struct
* core/state/snapshot: skip wiping correct states
* core/state/snapshot: updates
* core/state/snapshot: fixes
* core/state/snapshot: fix panic due to reference flaw in closure
* core/state/snapshot: fix errors in state generation logic + fix log output
* core/state/snapshot: remove an error case
* core/state/snapshot: fix condition-check for exhausted snap state
* core/state/snapshot: use stackTrie for small tries
* core/state/snapshot: don't resolve small storage tries in vain
* core/state/snapshot: properly clean up storage of deleted accounts
* core/state/snapshot: avoid RLP-encoding in some cases + minor nitpicks
* core/state/snapshot: fix error (+testcase)
* core/state/snapshot: clean up tests a bit
* core/state/snapshot: work in progress on better tests
* core/state/snapshot: polish code
* core/state/snapshot: fix trie iteration abortion trigger
* core/state/snapshot: fixes flaws
* core/state/snapshot: remove panic
* core/state/snapshot: fix abort
* core/state/snapshot: more tests (plus failing testcase)
* core/state/snapshot: more testcases + fix for failing test
* core/state/snapshot: testcase for malformed data
* core/state/snapshot: some test nitpicks
* core/state/snapshot: improvements to logging
* core/state/snapshot: testcase to demo error in abortion
* core/state/snapshot: fix abortion
* cmd/geth: make verify-state report the root
* trie: fix failing test
* core/state/snapshot: add timer metrics
* core/state/snapshot: fix metrics
* core/state/snapshot: udpate tests
* eth/protocols/snap: write snapshot account even if code or state is needed
* core/state/snapshot: fix diskmore check
* core/state/snapshot: review fixes
* core/state/snapshot: improve error message
* cmd/geth: rename 'error' to 'err' in logs
* core/state/snapshot: fix some review concerns
* core/state/snapshot, eth/protocols/snap: clear snapshot marker when starting/resuming snap sync
* core: add error log
* core/state/snapshot: use proper timers for metrics collection
* core/state/snapshot: address some review concerns
* eth/protocols/snap: improved log message
* eth/protocols/snap: fix heal logs to condense infos
* core/state/snapshot: wait for generator termination before restarting
* core/state/snapshot: revert timers to counters to track total time
Co-authored-by: Martin Holst Swende <martin@swende.se>
Co-authored-by: Péter Szilágyi <peterke@gmail.com>
2021-04-14 23:23:11 +03:00
|
|
|
stateWriter ethdb.Batch // Shared batch writer used for persisting raw states
|
|
|
|
accountHealed uint64 // Number of accounts downloaded during the healing stage
|
|
|
|
accountHealedBytes common.StorageSize // Number of raw account bytes persisted to disk during the healing stage
|
|
|
|
storageHealed uint64 // Number of storage slots downloaded during the healing stage
|
|
|
|
storageHealedBytes common.StorageSize // Number of raw storage bytes persisted to disk during the healing stage
|
|
|
|
|
|
|
|
startTime time.Time // Time instance when snapshot sync started
|
|
|
|
logTime time.Time // Time instance when status was last reported
|
2020-12-14 12:27:15 +03:00
|
|
|
|
|
|
|
pend sync.WaitGroup // Tracks network request goroutines for graceful shutdown
|
|
|
|
lock sync.RWMutex // Protects fields that can change outside of sync (peers, reqs, root)
|
|
|
|
}
|
|
|
|
|
2021-01-25 09:17:05 +03:00
|
|
|
// NewSyncer creates a new snapshot syncer to download the Ethereum state over the
|
|
|
|
// snap protocol.
|
2023-02-06 18:28:40 +03:00
|
|
|
func NewSyncer(db ethdb.KeyValueStore, scheme string) *Syncer {
|
2020-12-14 12:27:15 +03:00
|
|
|
return &Syncer{
|
2022-11-28 16:31:28 +03:00
|
|
|
db: db,
|
|
|
|
scheme: scheme,
|
2020-12-14 12:27:15 +03:00
|
|
|
|
2021-01-25 09:17:05 +03:00
|
|
|
peers: make(map[string]SyncPeer),
|
2020-12-14 12:27:15 +03:00
|
|
|
peerJoin: new(event.Feed),
|
|
|
|
peerDrop: new(event.Feed),
|
2021-05-19 15:09:03 +03:00
|
|
|
rates: msgrate.NewTrackers(log.New("proto", "snap")),
|
2020-12-14 12:27:15 +03:00
|
|
|
update: make(chan struct{}, 1),
|
|
|
|
|
|
|
|
accountIdlers: make(map[string]struct{}),
|
|
|
|
storageIdlers: make(map[string]struct{}),
|
|
|
|
bytecodeIdlers: make(map[string]struct{}),
|
|
|
|
|
2021-04-15 21:01:16 +03:00
|
|
|
accountReqs: make(map[uint64]*accountRequest),
|
|
|
|
storageReqs: make(map[uint64]*storageRequest),
|
|
|
|
bytecodeReqs: make(map[uint64]*bytecodeRequest),
|
2020-12-14 12:27:15 +03:00
|
|
|
|
|
|
|
trienodeHealIdlers: make(map[string]struct{}),
|
|
|
|
bytecodeHealIdlers: make(map[string]struct{}),
|
|
|
|
|
2022-09-09 11:42:57 +03:00
|
|
|
trienodeHealReqs: make(map[uint64]*trienodeHealRequest),
|
|
|
|
bytecodeHealReqs: make(map[uint64]*bytecodeHealRequest),
|
|
|
|
trienodeHealThrottle: maxTrienodeHealThrottle, // Tune downward instead of insta-filling with junk
|
|
|
|
stateWriter: db.NewBatch(),
|
2022-05-06 18:20:41 +03:00
|
|
|
|
|
|
|
extProgress: new(SyncProgress),
|
2020-12-14 12:27:15 +03:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Register injects a new data source into the syncer's peerset.
|
2021-01-25 09:17:05 +03:00
|
|
|
func (s *Syncer) Register(peer SyncPeer) error {
|
2020-12-14 12:27:15 +03:00
|
|
|
// Make sure the peer is not registered yet
|
2021-01-25 09:17:05 +03:00
|
|
|
id := peer.ID()
|
|
|
|
|
2020-12-14 12:27:15 +03:00
|
|
|
s.lock.Lock()
|
2021-01-25 09:17:05 +03:00
|
|
|
if _, ok := s.peers[id]; ok {
|
|
|
|
log.Error("Snap peer already registered", "id", id)
|
2020-12-14 12:27:15 +03:00
|
|
|
|
|
|
|
s.lock.Unlock()
|
|
|
|
return errors.New("already registered")
|
|
|
|
}
|
2021-01-25 09:17:05 +03:00
|
|
|
s.peers[id] = peer
|
2021-05-19 15:09:03 +03:00
|
|
|
s.rates.Track(id, msgrate.NewTracker(s.rates.MeanCapacities(), s.rates.MedianRoundTrip()))
|
2020-12-14 12:27:15 +03:00
|
|
|
|
|
|
|
// Mark the peer as idle, even if no sync is running
|
2021-01-25 09:17:05 +03:00
|
|
|
s.accountIdlers[id] = struct{}{}
|
|
|
|
s.storageIdlers[id] = struct{}{}
|
|
|
|
s.bytecodeIdlers[id] = struct{}{}
|
|
|
|
s.trienodeHealIdlers[id] = struct{}{}
|
|
|
|
s.bytecodeHealIdlers[id] = struct{}{}
|
2020-12-14 12:27:15 +03:00
|
|
|
s.lock.Unlock()
|
|
|
|
|
|
|
|
// Notify any active syncs that a new peer can be assigned data
|
2021-01-25 09:17:05 +03:00
|
|
|
s.peerJoin.Send(id)
|
2020-12-14 12:27:15 +03:00
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// Unregister injects a new data source into the syncer's peerset.
|
|
|
|
func (s *Syncer) Unregister(id string) error {
|
|
|
|
// Remove all traces of the peer from the registry
|
|
|
|
s.lock.Lock()
|
|
|
|
if _, ok := s.peers[id]; !ok {
|
|
|
|
log.Error("Snap peer not registered", "id", id)
|
|
|
|
|
|
|
|
s.lock.Unlock()
|
|
|
|
return errors.New("not registered")
|
|
|
|
}
|
|
|
|
delete(s.peers, id)
|
2021-05-19 15:09:03 +03:00
|
|
|
s.rates.Untrack(id)
|
2020-12-14 12:27:15 +03:00
|
|
|
|
|
|
|
// Remove status markers, even if no sync is running
|
|
|
|
delete(s.statelessPeers, id)
|
|
|
|
|
|
|
|
delete(s.accountIdlers, id)
|
|
|
|
delete(s.storageIdlers, id)
|
|
|
|
delete(s.bytecodeIdlers, id)
|
|
|
|
delete(s.trienodeHealIdlers, id)
|
|
|
|
delete(s.bytecodeHealIdlers, id)
|
|
|
|
s.lock.Unlock()
|
|
|
|
|
|
|
|
// Notify any active syncs that pending requests need to be reverted
|
|
|
|
s.peerDrop.Send(id)
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2022-07-15 14:55:51 +03:00
|
|
|
// Sync starts (or resumes a previous) sync cycle to iterate over a state trie
|
2020-12-14 12:27:15 +03:00
|
|
|
// with the given root and reconstruct the nodes based on the snapshot leaves.
|
|
|
|
// Previously downloaded segments will not be redownloaded of fixed, rather any
|
|
|
|
// errors will be healed after the leaves are fully accumulated.
|
|
|
|
func (s *Syncer) Sync(root common.Hash, cancel chan struct{}) error {
|
|
|
|
// Move the trie root from any previous value, revert stateless markers for
|
|
|
|
// any peers and initialize the syncer if it was not yet run
|
|
|
|
s.lock.Lock()
|
|
|
|
s.root = root
|
|
|
|
s.healer = &healTask{
|
2022-11-28 16:31:28 +03:00
|
|
|
scheduler: state.NewStateSync(root, s.db, s.onHealState, s.scheme),
|
2022-07-15 14:55:51 +03:00
|
|
|
trieTasks: make(map[string]common.Hash),
|
2020-12-14 12:27:15 +03:00
|
|
|
codeTasks: make(map[common.Hash]struct{}),
|
|
|
|
}
|
|
|
|
s.statelessPeers = make(map[string]struct{})
|
|
|
|
s.lock.Unlock()
|
|
|
|
|
|
|
|
if s.startTime == (time.Time{}) {
|
|
|
|
s.startTime = time.Now()
|
|
|
|
}
|
|
|
|
// Retrieve the previous sync status from LevelDB and abort if already synced
|
|
|
|
s.loadSyncStatus()
|
|
|
|
if len(s.tasks) == 0 && s.healer.scheduler.Pending() == 0 {
|
|
|
|
log.Debug("Snapshot sync already completed")
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
defer func() { // Persist any progress, independent of failure
|
|
|
|
for _, task := range s.tasks {
|
|
|
|
s.forwardAccountTask(task)
|
|
|
|
}
|
|
|
|
s.cleanAccountTasks()
|
|
|
|
s.saveSyncStatus()
|
|
|
|
}()
|
|
|
|
|
|
|
|
log.Debug("Starting snapshot sync cycle", "root", root)
|
core, eth: faster snapshot generation (#22504)
* eth/protocols: persist received state segments
* core: initial implementation
* core/state/snapshot: add tests
* core, eth: updates
* eth/protocols/snapshot: count flat state size
* core/state: add metrics
* core/state/snapshot: skip unnecessary deletion
* core/state/snapshot: rename
* core/state/snapshot: use the global batch
* core/state/snapshot: add logs and fix wiping
* core/state/snapshot: fix
* core/state/snapshot: save generation progress even if the batch is empty
* core/state/snapshot: fixes
* core/state/snapshot: fix initial account range length
* core/state/snapshot: fix initial account range
* eth/protocols/snap: store flat states during the healing
* eth/protocols/snap: print logs
* core/state/snapshot: refactor (#4)
* core/state/snapshot: refactor
* core/state/snapshot: tiny fix and polish
Co-authored-by: rjl493456442 <garyrong0905@gmail.com>
* core, eth: fixes
* core, eth: fix healing writer
* core, trie, eth: fix paths
* eth/protocols/snap: fix encoding
* eth, core: add debug log
* core/state/generate: release iterator asap (#5)
core/state/snapshot: less copy
core/state/snapshot: revert split loop
core/state/snapshot: handle storage becoming empty, improve test robustness
core/state: test modified codehash
core/state/snapshot: polish
* core/state/snapshot: optimize stats counter
* core, eth: add metric
* core/state/snapshot: update comments
* core/state/snapshot: improve tests
* core/state/snapshot: replace secure trie with standard trie
* core/state/snapshot: wrap return as the struct
* core/state/snapshot: skip wiping correct states
* core/state/snapshot: updates
* core/state/snapshot: fixes
* core/state/snapshot: fix panic due to reference flaw in closure
* core/state/snapshot: fix errors in state generation logic + fix log output
* core/state/snapshot: remove an error case
* core/state/snapshot: fix condition-check for exhausted snap state
* core/state/snapshot: use stackTrie for small tries
* core/state/snapshot: don't resolve small storage tries in vain
* core/state/snapshot: properly clean up storage of deleted accounts
* core/state/snapshot: avoid RLP-encoding in some cases + minor nitpicks
* core/state/snapshot: fix error (+testcase)
* core/state/snapshot: clean up tests a bit
* core/state/snapshot: work in progress on better tests
* core/state/snapshot: polish code
* core/state/snapshot: fix trie iteration abortion trigger
* core/state/snapshot: fixes flaws
* core/state/snapshot: remove panic
* core/state/snapshot: fix abort
* core/state/snapshot: more tests (plus failing testcase)
* core/state/snapshot: more testcases + fix for failing test
* core/state/snapshot: testcase for malformed data
* core/state/snapshot: some test nitpicks
* core/state/snapshot: improvements to logging
* core/state/snapshot: testcase to demo error in abortion
* core/state/snapshot: fix abortion
* cmd/geth: make verify-state report the root
* trie: fix failing test
* core/state/snapshot: add timer metrics
* core/state/snapshot: fix metrics
* core/state/snapshot: udpate tests
* eth/protocols/snap: write snapshot account even if code or state is needed
* core/state/snapshot: fix diskmore check
* core/state/snapshot: review fixes
* core/state/snapshot: improve error message
* cmd/geth: rename 'error' to 'err' in logs
* core/state/snapshot: fix some review concerns
* core/state/snapshot, eth/protocols/snap: clear snapshot marker when starting/resuming snap sync
* core: add error log
* core/state/snapshot: use proper timers for metrics collection
* core/state/snapshot: address some review concerns
* eth/protocols/snap: improved log message
* eth/protocols/snap: fix heal logs to condense infos
* core/state/snapshot: wait for generator termination before restarting
* core/state/snapshot: revert timers to counters to track total time
Co-authored-by: Martin Holst Swende <martin@swende.se>
Co-authored-by: Péter Szilágyi <peterke@gmail.com>
2021-04-14 23:23:11 +03:00
|
|
|
|
|
|
|
// Flush out the last committed raw states
|
|
|
|
defer func() {
|
|
|
|
if s.stateWriter.ValueSize() > 0 {
|
|
|
|
s.stateWriter.Write()
|
|
|
|
s.stateWriter.Reset()
|
|
|
|
}
|
|
|
|
}()
|
2020-12-14 12:27:15 +03:00
|
|
|
defer s.report(true)
|
2022-09-28 09:08:18 +03:00
|
|
|
// commit any trie- and bytecode-healing data.
|
|
|
|
defer s.commitHealer(true)
|
2020-12-14 12:27:15 +03:00
|
|
|
|
|
|
|
// Whether sync completed or not, disregard any future packets
|
|
|
|
defer func() {
|
|
|
|
log.Debug("Terminating snapshot sync cycle", "root", root)
|
|
|
|
s.lock.Lock()
|
|
|
|
s.accountReqs = make(map[uint64]*accountRequest)
|
|
|
|
s.storageReqs = make(map[uint64]*storageRequest)
|
|
|
|
s.bytecodeReqs = make(map[uint64]*bytecodeRequest)
|
|
|
|
s.trienodeHealReqs = make(map[uint64]*trienodeHealRequest)
|
|
|
|
s.bytecodeHealReqs = make(map[uint64]*bytecodeHealRequest)
|
|
|
|
s.lock.Unlock()
|
|
|
|
}()
|
|
|
|
// Keep scheduling sync tasks
|
|
|
|
peerJoin := make(chan string, 16)
|
|
|
|
peerJoinSub := s.peerJoin.Subscribe(peerJoin)
|
|
|
|
defer peerJoinSub.Unsubscribe()
|
|
|
|
|
|
|
|
peerDrop := make(chan string, 16)
|
|
|
|
peerDropSub := s.peerDrop.Subscribe(peerDrop)
|
|
|
|
defer peerDropSub.Unsubscribe()
|
|
|
|
|
2021-04-15 21:01:16 +03:00
|
|
|
// Create a set of unique channels for this sync cycle. We need these to be
|
|
|
|
// ephemeral so a data race doesn't accidentally deliver something stale on
|
|
|
|
// a persistent channel across syncs (yup, this happened)
|
|
|
|
var (
|
|
|
|
accountReqFails = make(chan *accountRequest)
|
|
|
|
storageReqFails = make(chan *storageRequest)
|
|
|
|
bytecodeReqFails = make(chan *bytecodeRequest)
|
|
|
|
accountResps = make(chan *accountResponse)
|
|
|
|
storageResps = make(chan *storageResponse)
|
|
|
|
bytecodeResps = make(chan *bytecodeResponse)
|
|
|
|
trienodeHealReqFails = make(chan *trienodeHealRequest)
|
|
|
|
bytecodeHealReqFails = make(chan *bytecodeHealRequest)
|
|
|
|
trienodeHealResps = make(chan *trienodeHealResponse)
|
|
|
|
bytecodeHealResps = make(chan *bytecodeHealResponse)
|
|
|
|
)
|
2020-12-14 12:27:15 +03:00
|
|
|
for {
|
|
|
|
// Remove all completed tasks and terminate sync if everything's done
|
|
|
|
s.cleanStorageTasks()
|
|
|
|
s.cleanAccountTasks()
|
|
|
|
if len(s.tasks) == 0 && s.healer.scheduler.Pending() == 0 {
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
// Assign all the data retrieval tasks to any free peers
|
2021-04-15 21:01:16 +03:00
|
|
|
s.assignAccountTasks(accountResps, accountReqFails, cancel)
|
|
|
|
s.assignBytecodeTasks(bytecodeResps, bytecodeReqFails, cancel)
|
|
|
|
s.assignStorageTasks(storageResps, storageReqFails, cancel)
|
2021-01-25 09:17:05 +03:00
|
|
|
|
2020-12-14 12:27:15 +03:00
|
|
|
if len(s.tasks) == 0 {
|
|
|
|
// Sync phase done, run heal phase
|
2021-04-15 21:01:16 +03:00
|
|
|
s.assignTrienodeHealTasks(trienodeHealResps, trienodeHealReqFails, cancel)
|
|
|
|
s.assignBytecodeHealTasks(bytecodeHealResps, bytecodeHealReqFails, cancel)
|
2020-12-14 12:27:15 +03:00
|
|
|
}
|
2022-05-06 18:20:41 +03:00
|
|
|
// Update sync progress
|
|
|
|
s.lock.Lock()
|
|
|
|
s.extProgress = &SyncProgress{
|
|
|
|
AccountSynced: s.accountSynced,
|
|
|
|
AccountBytes: s.accountBytes,
|
|
|
|
BytecodeSynced: s.bytecodeSynced,
|
|
|
|
BytecodeBytes: s.bytecodeBytes,
|
|
|
|
StorageSynced: s.storageSynced,
|
|
|
|
StorageBytes: s.storageBytes,
|
|
|
|
TrienodeHealSynced: s.trienodeHealSynced,
|
|
|
|
TrienodeHealBytes: s.trienodeHealBytes,
|
|
|
|
BytecodeHealSynced: s.bytecodeHealSynced,
|
|
|
|
BytecodeHealBytes: s.bytecodeHealBytes,
|
|
|
|
}
|
|
|
|
s.lock.Unlock()
|
2020-12-14 12:27:15 +03:00
|
|
|
// Wait for something to happen
|
|
|
|
select {
|
|
|
|
case <-s.update:
|
|
|
|
// Something happened (new peer, delivery, timeout), recheck tasks
|
|
|
|
case <-peerJoin:
|
|
|
|
// A new peer joined, try to schedule it new tasks
|
|
|
|
case id := <-peerDrop:
|
|
|
|
s.revertRequests(id)
|
|
|
|
case <-cancel:
|
2021-02-16 17:11:33 +03:00
|
|
|
return ErrCancelled
|
2020-12-14 12:27:15 +03:00
|
|
|
|
2021-04-15 21:01:16 +03:00
|
|
|
case req := <-accountReqFails:
|
2020-12-14 12:27:15 +03:00
|
|
|
s.revertAccountRequest(req)
|
2021-04-15 21:01:16 +03:00
|
|
|
case req := <-bytecodeReqFails:
|
2020-12-14 12:27:15 +03:00
|
|
|
s.revertBytecodeRequest(req)
|
2021-04-15 21:01:16 +03:00
|
|
|
case req := <-storageReqFails:
|
2020-12-14 12:27:15 +03:00
|
|
|
s.revertStorageRequest(req)
|
2021-04-15 21:01:16 +03:00
|
|
|
case req := <-trienodeHealReqFails:
|
2020-12-14 12:27:15 +03:00
|
|
|
s.revertTrienodeHealRequest(req)
|
2021-04-15 21:01:16 +03:00
|
|
|
case req := <-bytecodeHealReqFails:
|
2020-12-14 12:27:15 +03:00
|
|
|
s.revertBytecodeHealRequest(req)
|
|
|
|
|
2021-04-15 21:01:16 +03:00
|
|
|
case res := <-accountResps:
|
2020-12-14 12:27:15 +03:00
|
|
|
s.processAccountResponse(res)
|
2021-04-15 21:01:16 +03:00
|
|
|
case res := <-bytecodeResps:
|
2020-12-14 12:27:15 +03:00
|
|
|
s.processBytecodeResponse(res)
|
2021-04-15 21:01:16 +03:00
|
|
|
case res := <-storageResps:
|
2020-12-14 12:27:15 +03:00
|
|
|
s.processStorageResponse(res)
|
2021-04-15 21:01:16 +03:00
|
|
|
case res := <-trienodeHealResps:
|
2020-12-14 12:27:15 +03:00
|
|
|
s.processTrienodeHealResponse(res)
|
2021-04-15 21:01:16 +03:00
|
|
|
case res := <-bytecodeHealResps:
|
2020-12-14 12:27:15 +03:00
|
|
|
s.processBytecodeHealResponse(res)
|
|
|
|
}
|
|
|
|
// Report stats if something meaningful happened
|
|
|
|
s.report(false)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-10-23 18:31:56 +03:00
|
|
|
// cleanPath is used to remove the dangling nodes in the stackTrie.
|
|
|
|
func (s *Syncer) cleanPath(batch ethdb.Batch, owner common.Hash, path []byte) {
|
|
|
|
if owner == (common.Hash{}) && rawdb.ExistsAccountTrieNode(s.db, path) {
|
|
|
|
rawdb.DeleteAccountTrieNode(batch, path)
|
|
|
|
deletionGauge.Inc(1)
|
|
|
|
}
|
|
|
|
if owner != (common.Hash{}) && rawdb.ExistsStorageTrieNode(s.db, owner, path) {
|
|
|
|
rawdb.DeleteStorageTrieNode(batch, owner, path)
|
|
|
|
deletionGauge.Inc(1)
|
|
|
|
}
|
|
|
|
lookupGauge.Inc(1)
|
|
|
|
}
|
|
|
|
|
2020-12-14 12:27:15 +03:00
|
|
|
// loadSyncStatus retrieves a previously aborted sync status from the database,
|
|
|
|
// or generates a fresh one if none is available.
|
|
|
|
func (s *Syncer) loadSyncStatus() {
|
2021-11-26 14:26:03 +03:00
|
|
|
var progress SyncProgress
|
2020-12-14 12:27:15 +03:00
|
|
|
|
2020-12-28 00:38:16 +03:00
|
|
|
if status := rawdb.ReadSnapshotSyncStatus(s.db); status != nil {
|
2020-12-14 12:27:15 +03:00
|
|
|
if err := json.Unmarshal(status, &progress); err != nil {
|
|
|
|
log.Error("Failed to decode snap sync status", "err", err)
|
|
|
|
} else {
|
|
|
|
for _, task := range progress.Tasks {
|
|
|
|
log.Debug("Scheduled account sync task", "from", task.Next, "last", task.Last)
|
|
|
|
}
|
|
|
|
s.tasks = progress.Tasks
|
2021-04-27 17:19:59 +03:00
|
|
|
for _, task := range s.tasks {
|
2023-08-03 14:51:02 +03:00
|
|
|
task := task // closure for task.genBatch in the stacktrie writer callback
|
|
|
|
|
2021-04-28 23:09:15 +03:00
|
|
|
task.genBatch = ethdb.HookedBatch{
|
|
|
|
Batch: s.db.NewBatch(),
|
|
|
|
OnPut: func(key []byte, value []byte) {
|
|
|
|
s.accountBytes += common.StorageSize(len(key) + len(value))
|
|
|
|
},
|
|
|
|
}
|
2023-10-17 15:09:25 +03:00
|
|
|
options := trie.NewStackTrieOptions()
|
|
|
|
options = options.WithWriter(func(path []byte, hash common.Hash, blob []byte) {
|
|
|
|
rawdb.WriteTrieNode(task.genBatch, common.Hash{}, path, hash, blob, s.scheme)
|
2022-11-28 16:31:28 +03:00
|
|
|
})
|
2023-10-23 18:31:56 +03:00
|
|
|
if s.scheme == rawdb.PathScheme {
|
|
|
|
// Configure the dangling node cleaner and also filter out boundary nodes
|
|
|
|
// only in the context of the path scheme. Deletion is forbidden in the
|
|
|
|
// hash scheme, as it can disrupt state completeness.
|
|
|
|
options = options.WithCleaner(func(path []byte) {
|
|
|
|
s.cleanPath(task.genBatch, common.Hash{}, path)
|
|
|
|
})
|
|
|
|
// Skip the left boundary if it's not the first range.
|
|
|
|
// Skip the right boundary if it's not the last range.
|
|
|
|
options = options.WithSkipBoundary(task.Next != (common.Hash{}), task.Last != common.MaxHash, boundaryAccountNodesGauge)
|
|
|
|
}
|
2023-10-17 15:09:25 +03:00
|
|
|
task.genTrie = trie.NewStackTrie(options)
|
2022-06-06 18:14:55 +03:00
|
|
|
for accountHash, subtasks := range task.SubTasks {
|
2021-04-27 17:19:59 +03:00
|
|
|
for _, subtask := range subtasks {
|
2023-08-03 14:51:02 +03:00
|
|
|
subtask := subtask // closure for subtask.genBatch in the stacktrie writer callback
|
|
|
|
|
2021-04-28 23:09:15 +03:00
|
|
|
subtask.genBatch = ethdb.HookedBatch{
|
|
|
|
Batch: s.db.NewBatch(),
|
|
|
|
OnPut: func(key []byte, value []byte) {
|
|
|
|
s.storageBytes += common.StorageSize(len(key) + len(value))
|
|
|
|
},
|
|
|
|
}
|
2023-10-11 07:12:45 +03:00
|
|
|
owner := accountHash // local assignment for stacktrie writer closure
|
2023-10-17 15:09:25 +03:00
|
|
|
options := trie.NewStackTrieOptions()
|
|
|
|
options = options.WithWriter(func(path []byte, hash common.Hash, blob []byte) {
|
|
|
|
rawdb.WriteTrieNode(subtask.genBatch, owner, path, hash, blob, s.scheme)
|
2023-10-11 07:12:45 +03:00
|
|
|
})
|
2023-10-23 18:31:56 +03:00
|
|
|
if s.scheme == rawdb.PathScheme {
|
|
|
|
// Configure the dangling node cleaner and also filter out boundary nodes
|
|
|
|
// only in the context of the path scheme. Deletion is forbidden in the
|
|
|
|
// hash scheme, as it can disrupt state completeness.
|
|
|
|
options = options.WithCleaner(func(path []byte) {
|
|
|
|
s.cleanPath(subtask.genBatch, owner, path)
|
|
|
|
})
|
|
|
|
// Skip the left boundary if it's not the first range.
|
|
|
|
// Skip the right boundary if it's not the last range.
|
|
|
|
options = options.WithSkipBoundary(subtask.Next != common.Hash{}, subtask.Last != common.MaxHash, boundaryStorageNodesGauge)
|
|
|
|
}
|
2023-10-17 15:09:25 +03:00
|
|
|
subtask.genTrie = trie.NewStackTrie(options)
|
2021-04-27 17:19:59 +03:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2022-05-06 18:20:41 +03:00
|
|
|
s.lock.Lock()
|
|
|
|
defer s.lock.Unlock()
|
|
|
|
|
2021-01-25 09:17:05 +03:00
|
|
|
s.snapped = len(s.tasks) == 0
|
2020-12-14 12:27:15 +03:00
|
|
|
|
|
|
|
s.accountSynced = progress.AccountSynced
|
|
|
|
s.accountBytes = progress.AccountBytes
|
|
|
|
s.bytecodeSynced = progress.BytecodeSynced
|
|
|
|
s.bytecodeBytes = progress.BytecodeBytes
|
|
|
|
s.storageSynced = progress.StorageSynced
|
|
|
|
s.storageBytes = progress.StorageBytes
|
|
|
|
|
|
|
|
s.trienodeHealSynced = progress.TrienodeHealSynced
|
|
|
|
s.trienodeHealBytes = progress.TrienodeHealBytes
|
|
|
|
s.bytecodeHealSynced = progress.BytecodeHealSynced
|
|
|
|
s.bytecodeHealBytes = progress.BytecodeHealBytes
|
|
|
|
return
|
|
|
|
}
|
|
|
|
}
|
2022-07-15 14:55:51 +03:00
|
|
|
// Either we've failed to decode the previous state, or there was none.
|
2020-12-14 12:27:15 +03:00
|
|
|
// Start a fresh sync by chunking up the account range and scheduling
|
|
|
|
// them for retrieval.
|
|
|
|
s.tasks = nil
|
|
|
|
s.accountSynced, s.accountBytes = 0, 0
|
|
|
|
s.bytecodeSynced, s.bytecodeBytes = 0, 0
|
|
|
|
s.storageSynced, s.storageBytes = 0, 0
|
|
|
|
s.trienodeHealSynced, s.trienodeHealBytes = 0, 0
|
|
|
|
s.bytecodeHealSynced, s.bytecodeHealBytes = 0, 0
|
|
|
|
|
|
|
|
var next common.Hash
|
|
|
|
step := new(big.Int).Sub(
|
|
|
|
new(big.Int).Div(
|
|
|
|
new(big.Int).Exp(common.Big2, common.Big256, nil),
|
2021-04-27 17:19:59 +03:00
|
|
|
big.NewInt(int64(accountConcurrency)),
|
2020-12-14 12:27:15 +03:00
|
|
|
), common.Big1,
|
|
|
|
)
|
|
|
|
for i := 0; i < accountConcurrency; i++ {
|
|
|
|
last := common.BigToHash(new(big.Int).Add(next.Big(), step))
|
|
|
|
if i == accountConcurrency-1 {
|
|
|
|
// Make sure we don't overflow if the step is not a proper divisor
|
eth/protocols/snap: fix snap sync failure on empty storage range (#28306)
This change addresses an issue in snap sync, specifically when the entire sync process can be halted due to an encountered empty storage range.
Currently, on the snap sync client side, the response to an empty (partial) storage range is discarded as a non-delivery. However, this response can be a valid response, when the particular range requested does not contain any slots.
For instance, consider a large contract where the entire key space is divided into 16 chunks, and there are no available slots in the last chunk [0xf] -> [end]. When the node receives a request for this particular range, the response includes:
The proof with origin [0xf]
A nil storage slot set
If we simply discard this response, the finalization of the last range will be skipped, halting the entire sync process indefinitely. The test case TestSyncWithUnevenStorage can reproduce the scenario described above.
In addition, this change also defines the common variables MaxAddress and MaxHash.
2023-10-13 10:08:26 +03:00
|
|
|
last = common.MaxHash
|
2020-12-14 12:27:15 +03:00
|
|
|
}
|
2021-04-28 23:09:15 +03:00
|
|
|
batch := ethdb.HookedBatch{
|
|
|
|
Batch: s.db.NewBatch(),
|
|
|
|
OnPut: func(key []byte, value []byte) {
|
|
|
|
s.accountBytes += common.StorageSize(len(key) + len(value))
|
|
|
|
},
|
|
|
|
}
|
2023-10-17 15:09:25 +03:00
|
|
|
options := trie.NewStackTrieOptions()
|
|
|
|
options = options.WithWriter(func(path []byte, hash common.Hash, blob []byte) {
|
|
|
|
rawdb.WriteTrieNode(batch, common.Hash{}, path, hash, blob, s.scheme)
|
|
|
|
})
|
2023-10-23 18:31:56 +03:00
|
|
|
if s.scheme == rawdb.PathScheme {
|
|
|
|
// Configure the dangling node cleaner and also filter out boundary nodes
|
|
|
|
// only in the context of the path scheme. Deletion is forbidden in the
|
|
|
|
// hash scheme, as it can disrupt state completeness.
|
|
|
|
options = options.WithCleaner(func(path []byte) {
|
|
|
|
s.cleanPath(batch, common.Hash{}, path)
|
|
|
|
})
|
|
|
|
// Skip the left boundary if it's not the first range.
|
|
|
|
// Skip the right boundary if it's not the last range.
|
|
|
|
options = options.WithSkipBoundary(next != common.Hash{}, last != common.MaxHash, boundaryAccountNodesGauge)
|
|
|
|
}
|
2020-12-14 12:27:15 +03:00
|
|
|
s.tasks = append(s.tasks, &accountTask{
|
|
|
|
Next: next,
|
|
|
|
Last: last,
|
|
|
|
SubTasks: make(map[common.Hash][]*storageTask),
|
2021-04-27 17:19:59 +03:00
|
|
|
genBatch: batch,
|
2023-10-17 15:09:25 +03:00
|
|
|
genTrie: trie.NewStackTrie(options),
|
2020-12-14 12:27:15 +03:00
|
|
|
})
|
|
|
|
log.Debug("Created account sync task", "from", next, "last", last)
|
|
|
|
next = common.BigToHash(new(big.Int).Add(last.Big(), common.Big1))
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// saveSyncStatus marshals the remaining sync tasks into leveldb.
|
|
|
|
func (s *Syncer) saveSyncStatus() {
|
2021-04-27 17:19:59 +03:00
|
|
|
// Serialize any partial progress to disk before spinning down
|
|
|
|
for _, task := range s.tasks {
|
|
|
|
if err := task.genBatch.Write(); err != nil {
|
|
|
|
log.Error("Failed to persist account slots", "err", err)
|
|
|
|
}
|
|
|
|
for _, subtasks := range task.SubTasks {
|
|
|
|
for _, subtask := range subtasks {
|
|
|
|
if err := subtask.genBatch.Write(); err != nil {
|
|
|
|
log.Error("Failed to persist storage slots", "err", err)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// Store the actual progress markers
|
2021-11-26 14:26:03 +03:00
|
|
|
progress := &SyncProgress{
|
2020-12-14 12:27:15 +03:00
|
|
|
Tasks: s.tasks,
|
|
|
|
AccountSynced: s.accountSynced,
|
|
|
|
AccountBytes: s.accountBytes,
|
|
|
|
BytecodeSynced: s.bytecodeSynced,
|
|
|
|
BytecodeBytes: s.bytecodeBytes,
|
|
|
|
StorageSynced: s.storageSynced,
|
|
|
|
StorageBytes: s.storageBytes,
|
|
|
|
TrienodeHealSynced: s.trienodeHealSynced,
|
|
|
|
TrienodeHealBytes: s.trienodeHealBytes,
|
|
|
|
BytecodeHealSynced: s.bytecodeHealSynced,
|
|
|
|
BytecodeHealBytes: s.bytecodeHealBytes,
|
|
|
|
}
|
|
|
|
status, err := json.Marshal(progress)
|
|
|
|
if err != nil {
|
|
|
|
panic(err) // This can only fail during implementation
|
|
|
|
}
|
|
|
|
rawdb.WriteSnapshotSyncStatus(s.db, status)
|
|
|
|
}
|
|
|
|
|
2021-11-26 14:26:03 +03:00
|
|
|
// Progress returns the snap sync status statistics.
|
|
|
|
func (s *Syncer) Progress() (*SyncProgress, *SyncPending) {
|
|
|
|
s.lock.Lock()
|
|
|
|
defer s.lock.Unlock()
|
|
|
|
pending := new(SyncPending)
|
|
|
|
if s.healer != nil {
|
|
|
|
pending.TrienodeHeal = uint64(len(s.healer.trieTasks))
|
|
|
|
pending.BytecodeHeal = uint64(len(s.healer.codeTasks))
|
|
|
|
}
|
2022-05-06 18:20:41 +03:00
|
|
|
return s.extProgress, pending
|
2021-11-26 14:26:03 +03:00
|
|
|
}
|
|
|
|
|
2020-12-14 12:27:15 +03:00
|
|
|
// cleanAccountTasks removes account range retrieval tasks that have already been
|
|
|
|
// completed.
|
|
|
|
func (s *Syncer) cleanAccountTasks() {
|
2021-04-27 17:19:59 +03:00
|
|
|
// If the sync was already done before, don't even bother
|
|
|
|
if len(s.tasks) == 0 {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
// Sync wasn't finished previously, check for any task that can be finalized
|
2020-12-14 12:27:15 +03:00
|
|
|
for i := 0; i < len(s.tasks); i++ {
|
|
|
|
if s.tasks[i].done {
|
|
|
|
s.tasks = append(s.tasks[:i], s.tasks[i+1:]...)
|
|
|
|
i--
|
|
|
|
}
|
|
|
|
}
|
2021-04-27 17:19:59 +03:00
|
|
|
// If everything was just finalized just, generate the account trie and start heal
|
2021-01-25 09:17:05 +03:00
|
|
|
if len(s.tasks) == 0 {
|
|
|
|
s.lock.Lock()
|
|
|
|
s.snapped = true
|
|
|
|
s.lock.Unlock()
|
2021-04-27 17:19:59 +03:00
|
|
|
|
|
|
|
// Push the final sync report
|
|
|
|
s.reportSyncProgress(true)
|
2021-01-25 09:17:05 +03:00
|
|
|
}
|
2020-12-14 12:27:15 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// cleanStorageTasks iterates over all the account tasks and storage sub-tasks
|
|
|
|
// within, cleaning any that have been completed.
|
|
|
|
func (s *Syncer) cleanStorageTasks() {
|
|
|
|
for _, task := range s.tasks {
|
|
|
|
for account, subtasks := range task.SubTasks {
|
|
|
|
// Remove storage range retrieval tasks that completed
|
|
|
|
for j := 0; j < len(subtasks); j++ {
|
|
|
|
if subtasks[j].done {
|
|
|
|
subtasks = append(subtasks[:j], subtasks[j+1:]...)
|
|
|
|
j--
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if len(subtasks) > 0 {
|
|
|
|
task.SubTasks[account] = subtasks
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
// If all storage chunks are done, mark the account as done too
|
|
|
|
for j, hash := range task.res.hashes {
|
|
|
|
if hash == account {
|
|
|
|
task.needState[j] = false
|
|
|
|
}
|
|
|
|
}
|
|
|
|
delete(task.SubTasks, account)
|
|
|
|
task.pend--
|
|
|
|
|
|
|
|
// If this was the last pending task, forward the account task
|
|
|
|
if task.pend == 0 {
|
|
|
|
s.forwardAccountTask(task)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// assignAccountTasks attempts to match idle peers to pending account range
|
|
|
|
// retrievals.
|
2021-04-15 21:01:16 +03:00
|
|
|
func (s *Syncer) assignAccountTasks(success chan *accountResponse, fail chan *accountRequest, cancel chan struct{}) {
|
2020-12-14 12:27:15 +03:00
|
|
|
s.lock.Lock()
|
|
|
|
defer s.lock.Unlock()
|
|
|
|
|
2021-05-19 15:09:03 +03:00
|
|
|
// Sort the peers by download capacity to use faster ones if many available
|
|
|
|
idlers := &capacitySort{
|
|
|
|
ids: make([]string, 0, len(s.accountIdlers)),
|
2021-05-27 19:43:55 +03:00
|
|
|
caps: make([]int, 0, len(s.accountIdlers)),
|
2021-05-19 15:09:03 +03:00
|
|
|
}
|
|
|
|
targetTTL := s.rates.TargetTimeout()
|
|
|
|
for id := range s.accountIdlers {
|
|
|
|
if _, ok := s.statelessPeers[id]; ok {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
idlers.ids = append(idlers.ids, id)
|
|
|
|
idlers.caps = append(idlers.caps, s.rates.Capacity(id, AccountRangeMsg, targetTTL))
|
|
|
|
}
|
|
|
|
if len(idlers.ids) == 0 {
|
2020-12-14 12:27:15 +03:00
|
|
|
return
|
|
|
|
}
|
2021-05-19 15:09:03 +03:00
|
|
|
sort.Sort(sort.Reverse(idlers))
|
|
|
|
|
2020-12-14 12:27:15 +03:00
|
|
|
// Iterate over all the tasks and try to find a pending one
|
|
|
|
for _, task := range s.tasks {
|
|
|
|
// Skip any tasks already filling
|
|
|
|
if task.req != nil || task.res != nil {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
// Task pending retrieval, try to find an idle peer. If no such peer
|
|
|
|
// exists, we probably assigned tasks for all (or they are stateless).
|
|
|
|
// Abort the entire assignment mechanism.
|
2021-05-19 15:09:03 +03:00
|
|
|
if len(idlers.ids) == 0 {
|
2020-12-14 12:27:15 +03:00
|
|
|
return
|
|
|
|
}
|
2021-05-19 15:09:03 +03:00
|
|
|
var (
|
|
|
|
idle = idlers.ids[0]
|
|
|
|
peer = s.peers[idle]
|
|
|
|
cap = idlers.caps[0]
|
|
|
|
)
|
|
|
|
idlers.ids, idlers.caps = idlers.ids[1:], idlers.caps[1:]
|
2021-03-26 23:29:22 +03:00
|
|
|
|
2020-12-14 12:27:15 +03:00
|
|
|
// Matched a pending task to an idle peer, allocate a unique request id
|
|
|
|
var reqid uint64
|
|
|
|
for {
|
|
|
|
reqid = uint64(rand.Int63())
|
|
|
|
if reqid == 0 {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
if _, ok := s.accountReqs[reqid]; ok {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
break
|
|
|
|
}
|
|
|
|
// Generate the network query and send it to the peer
|
|
|
|
req := &accountRequest{
|
2021-04-15 21:01:16 +03:00
|
|
|
peer: idle,
|
|
|
|
id: reqid,
|
2021-05-19 15:09:03 +03:00
|
|
|
time: time.Now(),
|
2021-04-15 21:01:16 +03:00
|
|
|
deliver: success,
|
|
|
|
revert: fail,
|
|
|
|
cancel: cancel,
|
|
|
|
stale: make(chan struct{}),
|
|
|
|
origin: task.Next,
|
|
|
|
limit: task.Last,
|
|
|
|
task: task,
|
2020-12-14 12:27:15 +03:00
|
|
|
}
|
2021-05-19 15:09:03 +03:00
|
|
|
req.timeout = time.AfterFunc(s.rates.TargetTimeout(), func() {
|
2021-03-26 23:29:22 +03:00
|
|
|
peer.Log().Debug("Account range request timed out", "reqid", reqid)
|
2021-05-19 15:09:03 +03:00
|
|
|
s.rates.Update(idle, AccountRangeMsg, 0, 0)
|
2021-01-07 13:58:07 +03:00
|
|
|
s.scheduleRevertAccountRequest(req)
|
2020-12-14 12:27:15 +03:00
|
|
|
})
|
|
|
|
s.accountReqs[reqid] = req
|
|
|
|
delete(s.accountIdlers, idle)
|
|
|
|
|
|
|
|
s.pend.Add(1)
|
2021-03-26 23:29:22 +03:00
|
|
|
go func(root common.Hash) {
|
2020-12-14 12:27:15 +03:00
|
|
|
defer s.pend.Done()
|
|
|
|
|
|
|
|
// Attempt to send the remote request and revert if it fails
|
2021-05-19 15:09:03 +03:00
|
|
|
if cap > maxRequestSize {
|
|
|
|
cap = maxRequestSize
|
|
|
|
}
|
|
|
|
if cap < minRequestSize { // Don't bother with peers below a bare minimum performance
|
|
|
|
cap = minRequestSize
|
|
|
|
}
|
|
|
|
if err := peer.RequestAccountRange(reqid, root, req.origin, req.limit, uint64(cap)); err != nil {
|
2020-12-14 12:27:15 +03:00
|
|
|
peer.Log().Debug("Failed to request account range", "err", err)
|
2021-01-07 13:58:07 +03:00
|
|
|
s.scheduleRevertAccountRequest(req)
|
2020-12-14 12:27:15 +03:00
|
|
|
}
|
2021-03-26 23:29:22 +03:00
|
|
|
}(s.root)
|
2020-12-14 12:27:15 +03:00
|
|
|
|
|
|
|
// Inject the request into the task to block further assignments
|
|
|
|
task.req = req
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// assignBytecodeTasks attempts to match idle peers to pending code retrievals.
|
2021-04-15 21:01:16 +03:00
|
|
|
func (s *Syncer) assignBytecodeTasks(success chan *bytecodeResponse, fail chan *bytecodeRequest, cancel chan struct{}) {
|
2020-12-14 12:27:15 +03:00
|
|
|
s.lock.Lock()
|
|
|
|
defer s.lock.Unlock()
|
|
|
|
|
2021-05-19 15:09:03 +03:00
|
|
|
// Sort the peers by download capacity to use faster ones if many available
|
|
|
|
idlers := &capacitySort{
|
|
|
|
ids: make([]string, 0, len(s.bytecodeIdlers)),
|
2021-05-27 19:43:55 +03:00
|
|
|
caps: make([]int, 0, len(s.bytecodeIdlers)),
|
2021-05-19 15:09:03 +03:00
|
|
|
}
|
|
|
|
targetTTL := s.rates.TargetTimeout()
|
|
|
|
for id := range s.bytecodeIdlers {
|
|
|
|
if _, ok := s.statelessPeers[id]; ok {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
idlers.ids = append(idlers.ids, id)
|
|
|
|
idlers.caps = append(idlers.caps, s.rates.Capacity(id, ByteCodesMsg, targetTTL))
|
|
|
|
}
|
|
|
|
if len(idlers.ids) == 0 {
|
2020-12-14 12:27:15 +03:00
|
|
|
return
|
|
|
|
}
|
2021-05-19 15:09:03 +03:00
|
|
|
sort.Sort(sort.Reverse(idlers))
|
|
|
|
|
2020-12-14 12:27:15 +03:00
|
|
|
// Iterate over all the tasks and try to find a pending one
|
|
|
|
for _, task := range s.tasks {
|
|
|
|
// Skip any tasks not in the bytecode retrieval phase
|
|
|
|
if task.res == nil {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
// Skip tasks that are already retrieving (or done with) all codes
|
|
|
|
if len(task.codeTasks) == 0 {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
// Task pending retrieval, try to find an idle peer. If no such peer
|
|
|
|
// exists, we probably assigned tasks for all (or they are stateless).
|
|
|
|
// Abort the entire assignment mechanism.
|
2021-05-19 15:09:03 +03:00
|
|
|
if len(idlers.ids) == 0 {
|
2020-12-14 12:27:15 +03:00
|
|
|
return
|
|
|
|
}
|
2021-05-19 15:09:03 +03:00
|
|
|
var (
|
|
|
|
idle = idlers.ids[0]
|
|
|
|
peer = s.peers[idle]
|
|
|
|
cap = idlers.caps[0]
|
|
|
|
)
|
|
|
|
idlers.ids, idlers.caps = idlers.ids[1:], idlers.caps[1:]
|
2021-03-26 23:29:22 +03:00
|
|
|
|
2020-12-14 12:27:15 +03:00
|
|
|
// Matched a pending task to an idle peer, allocate a unique request id
|
|
|
|
var reqid uint64
|
|
|
|
for {
|
|
|
|
reqid = uint64(rand.Int63())
|
|
|
|
if reqid == 0 {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
if _, ok := s.bytecodeReqs[reqid]; ok {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
break
|
|
|
|
}
|
|
|
|
// Generate the network query and send it to the peer
|
2021-05-19 15:09:03 +03:00
|
|
|
if cap > maxCodeRequestCount {
|
|
|
|
cap = maxCodeRequestCount
|
|
|
|
}
|
2021-05-27 19:43:55 +03:00
|
|
|
hashes := make([]common.Hash, 0, cap)
|
2020-12-14 12:27:15 +03:00
|
|
|
for hash := range task.codeTasks {
|
|
|
|
delete(task.codeTasks, hash)
|
|
|
|
hashes = append(hashes, hash)
|
2021-05-27 19:43:55 +03:00
|
|
|
if len(hashes) >= cap {
|
2020-12-14 12:27:15 +03:00
|
|
|
break
|
|
|
|
}
|
|
|
|
}
|
|
|
|
req := &bytecodeRequest{
|
2021-04-15 21:01:16 +03:00
|
|
|
peer: idle,
|
|
|
|
id: reqid,
|
2021-05-19 15:09:03 +03:00
|
|
|
time: time.Now(),
|
2021-04-15 21:01:16 +03:00
|
|
|
deliver: success,
|
|
|
|
revert: fail,
|
|
|
|
cancel: cancel,
|
|
|
|
stale: make(chan struct{}),
|
|
|
|
hashes: hashes,
|
|
|
|
task: task,
|
2020-12-14 12:27:15 +03:00
|
|
|
}
|
2021-05-19 15:09:03 +03:00
|
|
|
req.timeout = time.AfterFunc(s.rates.TargetTimeout(), func() {
|
2021-03-26 23:29:22 +03:00
|
|
|
peer.Log().Debug("Bytecode request timed out", "reqid", reqid)
|
2021-05-19 15:09:03 +03:00
|
|
|
s.rates.Update(idle, ByteCodesMsg, 0, 0)
|
2021-01-07 13:58:07 +03:00
|
|
|
s.scheduleRevertBytecodeRequest(req)
|
2020-12-14 12:27:15 +03:00
|
|
|
})
|
|
|
|
s.bytecodeReqs[reqid] = req
|
|
|
|
delete(s.bytecodeIdlers, idle)
|
|
|
|
|
|
|
|
s.pend.Add(1)
|
2021-03-26 23:29:22 +03:00
|
|
|
go func() {
|
2020-12-14 12:27:15 +03:00
|
|
|
defer s.pend.Done()
|
|
|
|
|
|
|
|
// Attempt to send the remote request and revert if it fails
|
|
|
|
if err := peer.RequestByteCodes(reqid, hashes, maxRequestSize); err != nil {
|
|
|
|
log.Debug("Failed to request bytecodes", "err", err)
|
2021-01-07 13:58:07 +03:00
|
|
|
s.scheduleRevertBytecodeRequest(req)
|
2020-12-14 12:27:15 +03:00
|
|
|
}
|
2021-03-26 23:29:22 +03:00
|
|
|
}()
|
2020-12-14 12:27:15 +03:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// assignStorageTasks attempts to match idle peers to pending storage range
|
|
|
|
// retrievals.
|
2021-04-15 21:01:16 +03:00
|
|
|
func (s *Syncer) assignStorageTasks(success chan *storageResponse, fail chan *storageRequest, cancel chan struct{}) {
|
2020-12-14 12:27:15 +03:00
|
|
|
s.lock.Lock()
|
|
|
|
defer s.lock.Unlock()
|
|
|
|
|
2021-05-19 15:09:03 +03:00
|
|
|
// Sort the peers by download capacity to use faster ones if many available
|
|
|
|
idlers := &capacitySort{
|
|
|
|
ids: make([]string, 0, len(s.storageIdlers)),
|
2021-05-27 19:43:55 +03:00
|
|
|
caps: make([]int, 0, len(s.storageIdlers)),
|
2021-05-19 15:09:03 +03:00
|
|
|
}
|
|
|
|
targetTTL := s.rates.TargetTimeout()
|
|
|
|
for id := range s.storageIdlers {
|
|
|
|
if _, ok := s.statelessPeers[id]; ok {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
idlers.ids = append(idlers.ids, id)
|
|
|
|
idlers.caps = append(idlers.caps, s.rates.Capacity(id, StorageRangesMsg, targetTTL))
|
|
|
|
}
|
|
|
|
if len(idlers.ids) == 0 {
|
2020-12-14 12:27:15 +03:00
|
|
|
return
|
|
|
|
}
|
2021-05-19 15:09:03 +03:00
|
|
|
sort.Sort(sort.Reverse(idlers))
|
|
|
|
|
2020-12-14 12:27:15 +03:00
|
|
|
// Iterate over all the tasks and try to find a pending one
|
|
|
|
for _, task := range s.tasks {
|
|
|
|
// Skip any tasks not in the storage retrieval phase
|
|
|
|
if task.res == nil {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
// Skip tasks that are already retrieving (or done with) all small states
|
|
|
|
if len(task.SubTasks) == 0 && len(task.stateTasks) == 0 {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
// Task pending retrieval, try to find an idle peer. If no such peer
|
|
|
|
// exists, we probably assigned tasks for all (or they are stateless).
|
|
|
|
// Abort the entire assignment mechanism.
|
2021-05-19 15:09:03 +03:00
|
|
|
if len(idlers.ids) == 0 {
|
2020-12-14 12:27:15 +03:00
|
|
|
return
|
|
|
|
}
|
2021-05-19 15:09:03 +03:00
|
|
|
var (
|
|
|
|
idle = idlers.ids[0]
|
|
|
|
peer = s.peers[idle]
|
|
|
|
cap = idlers.caps[0]
|
|
|
|
)
|
|
|
|
idlers.ids, idlers.caps = idlers.ids[1:], idlers.caps[1:]
|
2021-03-26 23:29:22 +03:00
|
|
|
|
2020-12-14 12:27:15 +03:00
|
|
|
// Matched a pending task to an idle peer, allocate a unique request id
|
|
|
|
var reqid uint64
|
|
|
|
for {
|
|
|
|
reqid = uint64(rand.Int63())
|
|
|
|
if reqid == 0 {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
if _, ok := s.storageReqs[reqid]; ok {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
break
|
|
|
|
}
|
|
|
|
// Generate the network query and send it to the peer. If there are
|
|
|
|
// large contract tasks pending, complete those before diving into
|
|
|
|
// even more new contracts.
|
2021-05-19 15:09:03 +03:00
|
|
|
if cap > maxRequestSize {
|
|
|
|
cap = maxRequestSize
|
|
|
|
}
|
|
|
|
if cap < minRequestSize { // Don't bother with peers below a bare minimum performance
|
|
|
|
cap = minRequestSize
|
|
|
|
}
|
2021-05-27 19:43:55 +03:00
|
|
|
storageSets := cap / 1024
|
2021-05-19 15:09:03 +03:00
|
|
|
|
2020-12-14 12:27:15 +03:00
|
|
|
var (
|
2021-05-19 15:09:03 +03:00
|
|
|
accounts = make([]common.Hash, 0, storageSets)
|
|
|
|
roots = make([]common.Hash, 0, storageSets)
|
2020-12-14 12:27:15 +03:00
|
|
|
subtask *storageTask
|
|
|
|
)
|
|
|
|
for account, subtasks := range task.SubTasks {
|
|
|
|
for _, st := range subtasks {
|
|
|
|
// Skip any subtasks already filling
|
|
|
|
if st.req != nil {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
// Found an incomplete storage chunk, schedule it
|
|
|
|
accounts = append(accounts, account)
|
|
|
|
roots = append(roots, st.root)
|
|
|
|
subtask = st
|
|
|
|
break // Large contract chunks are downloaded individually
|
|
|
|
}
|
|
|
|
if subtask != nil {
|
|
|
|
break // Large contract chunks are downloaded individually
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if subtask == nil {
|
|
|
|
// No large contract required retrieval, but small ones available
|
2022-08-19 09:00:21 +03:00
|
|
|
for account, root := range task.stateTasks {
|
|
|
|
delete(task.stateTasks, account)
|
2020-12-14 12:27:15 +03:00
|
|
|
|
2022-08-19 09:00:21 +03:00
|
|
|
accounts = append(accounts, account)
|
2020-12-14 12:27:15 +03:00
|
|
|
roots = append(roots, root)
|
|
|
|
|
2021-05-19 15:09:03 +03:00
|
|
|
if len(accounts) >= storageSets {
|
2020-12-14 12:27:15 +03:00
|
|
|
break
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// If nothing was found, it means this task is actually already fully
|
|
|
|
// retrieving, but large contracts are hard to detect. Skip to the next.
|
|
|
|
if len(accounts) == 0 {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
req := &storageRequest{
|
|
|
|
peer: idle,
|
|
|
|
id: reqid,
|
2021-05-19 15:09:03 +03:00
|
|
|
time: time.Now(),
|
2021-04-15 21:01:16 +03:00
|
|
|
deliver: success,
|
|
|
|
revert: fail,
|
2020-12-14 12:27:15 +03:00
|
|
|
cancel: cancel,
|
|
|
|
stale: make(chan struct{}),
|
|
|
|
accounts: accounts,
|
|
|
|
roots: roots,
|
|
|
|
mainTask: task,
|
|
|
|
subTask: subtask,
|
|
|
|
}
|
|
|
|
if subtask != nil {
|
|
|
|
req.origin = subtask.Next
|
|
|
|
req.limit = subtask.Last
|
|
|
|
}
|
2021-05-19 15:09:03 +03:00
|
|
|
req.timeout = time.AfterFunc(s.rates.TargetTimeout(), func() {
|
2021-03-26 23:29:22 +03:00
|
|
|
peer.Log().Debug("Storage request timed out", "reqid", reqid)
|
2021-05-19 15:09:03 +03:00
|
|
|
s.rates.Update(idle, StorageRangesMsg, 0, 0)
|
2021-01-07 13:58:07 +03:00
|
|
|
s.scheduleRevertStorageRequest(req)
|
2020-12-14 12:27:15 +03:00
|
|
|
})
|
|
|
|
s.storageReqs[reqid] = req
|
|
|
|
delete(s.storageIdlers, idle)
|
|
|
|
|
|
|
|
s.pend.Add(1)
|
2021-03-26 23:29:22 +03:00
|
|
|
go func(root common.Hash) {
|
2020-12-14 12:27:15 +03:00
|
|
|
defer s.pend.Done()
|
|
|
|
|
|
|
|
// Attempt to send the remote request and revert if it fails
|
|
|
|
var origin, limit []byte
|
|
|
|
if subtask != nil {
|
|
|
|
origin, limit = req.origin[:], req.limit[:]
|
|
|
|
}
|
2021-05-19 15:09:03 +03:00
|
|
|
if err := peer.RequestStorageRanges(reqid, root, accounts, origin, limit, uint64(cap)); err != nil {
|
2020-12-14 12:27:15 +03:00
|
|
|
log.Debug("Failed to request storage", "err", err)
|
2021-01-07 13:58:07 +03:00
|
|
|
s.scheduleRevertStorageRequest(req)
|
2020-12-14 12:27:15 +03:00
|
|
|
}
|
2021-03-26 23:29:22 +03:00
|
|
|
}(s.root)
|
2020-12-14 12:27:15 +03:00
|
|
|
|
|
|
|
// Inject the request into the subtask to block further assignments
|
|
|
|
if subtask != nil {
|
|
|
|
subtask.req = req
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// assignTrienodeHealTasks attempts to match idle peers to trie node requests to
|
|
|
|
// heal any trie errors caused by the snap sync's chunked retrieval model.
|
2021-04-15 21:01:16 +03:00
|
|
|
func (s *Syncer) assignTrienodeHealTasks(success chan *trienodeHealResponse, fail chan *trienodeHealRequest, cancel chan struct{}) {
|
2020-12-14 12:27:15 +03:00
|
|
|
s.lock.Lock()
|
|
|
|
defer s.lock.Unlock()
|
|
|
|
|
2021-05-19 15:09:03 +03:00
|
|
|
// Sort the peers by download capacity to use faster ones if many available
|
|
|
|
idlers := &capacitySort{
|
|
|
|
ids: make([]string, 0, len(s.trienodeHealIdlers)),
|
2021-05-27 19:43:55 +03:00
|
|
|
caps: make([]int, 0, len(s.trienodeHealIdlers)),
|
2021-05-19 15:09:03 +03:00
|
|
|
}
|
|
|
|
targetTTL := s.rates.TargetTimeout()
|
|
|
|
for id := range s.trienodeHealIdlers {
|
|
|
|
if _, ok := s.statelessPeers[id]; ok {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
idlers.ids = append(idlers.ids, id)
|
|
|
|
idlers.caps = append(idlers.caps, s.rates.Capacity(id, TrieNodesMsg, targetTTL))
|
|
|
|
}
|
|
|
|
if len(idlers.ids) == 0 {
|
2020-12-14 12:27:15 +03:00
|
|
|
return
|
|
|
|
}
|
2021-05-19 15:09:03 +03:00
|
|
|
sort.Sort(sort.Reverse(idlers))
|
|
|
|
|
2020-12-14 12:27:15 +03:00
|
|
|
// Iterate over pending tasks and try to find a peer to retrieve with
|
|
|
|
for len(s.healer.trieTasks) > 0 || s.healer.scheduler.Pending() > 0 {
|
|
|
|
// If there are not enough trie tasks queued to fully assign, fill the
|
|
|
|
// queue from the state sync scheduler. The trie synced schedules these
|
|
|
|
// together with bytecodes, so we need to queue them combined.
|
|
|
|
var (
|
|
|
|
have = len(s.healer.trieTasks) + len(s.healer.codeTasks)
|
|
|
|
want = maxTrieRequestCount + maxCodeRequestCount
|
|
|
|
)
|
|
|
|
if have < want {
|
2022-07-15 14:55:51 +03:00
|
|
|
paths, hashes, codes := s.healer.scheduler.Missing(want - have)
|
|
|
|
for i, path := range paths {
|
|
|
|
s.healer.trieTasks[path] = hashes[i]
|
2020-12-14 12:27:15 +03:00
|
|
|
}
|
|
|
|
for _, hash := range codes {
|
|
|
|
s.healer.codeTasks[hash] = struct{}{}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// If all the heal tasks are bytecodes or already downloading, bail
|
|
|
|
if len(s.healer.trieTasks) == 0 {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
// Task pending retrieval, try to find an idle peer. If no such peer
|
|
|
|
// exists, we probably assigned tasks for all (or they are stateless).
|
|
|
|
// Abort the entire assignment mechanism.
|
2021-05-19 15:09:03 +03:00
|
|
|
if len(idlers.ids) == 0 {
|
2020-12-14 12:27:15 +03:00
|
|
|
return
|
|
|
|
}
|
2021-05-19 15:09:03 +03:00
|
|
|
var (
|
|
|
|
idle = idlers.ids[0]
|
|
|
|
peer = s.peers[idle]
|
|
|
|
cap = idlers.caps[0]
|
|
|
|
)
|
|
|
|
idlers.ids, idlers.caps = idlers.ids[1:], idlers.caps[1:]
|
2021-03-26 23:29:22 +03:00
|
|
|
|
2020-12-14 12:27:15 +03:00
|
|
|
// Matched a pending task to an idle peer, allocate a unique request id
|
|
|
|
var reqid uint64
|
|
|
|
for {
|
|
|
|
reqid = uint64(rand.Int63())
|
|
|
|
if reqid == 0 {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
if _, ok := s.trienodeHealReqs[reqid]; ok {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
break
|
|
|
|
}
|
|
|
|
// Generate the network query and send it to the peer
|
2021-05-19 15:09:03 +03:00
|
|
|
if cap > maxTrieRequestCount {
|
|
|
|
cap = maxTrieRequestCount
|
|
|
|
}
|
2022-09-09 11:42:57 +03:00
|
|
|
cap = int(float64(cap) / s.trienodeHealThrottle)
|
|
|
|
if cap <= 0 {
|
|
|
|
cap = 1
|
|
|
|
}
|
2020-12-14 12:27:15 +03:00
|
|
|
var (
|
2021-05-27 19:43:55 +03:00
|
|
|
hashes = make([]common.Hash, 0, cap)
|
2022-07-15 14:55:51 +03:00
|
|
|
paths = make([]string, 0, cap)
|
2021-05-27 19:43:55 +03:00
|
|
|
pathsets = make([]TrieNodePathSet, 0, cap)
|
2020-12-14 12:27:15 +03:00
|
|
|
)
|
2022-07-15 14:55:51 +03:00
|
|
|
for path, hash := range s.healer.trieTasks {
|
|
|
|
delete(s.healer.trieTasks, path)
|
2020-12-14 12:27:15 +03:00
|
|
|
|
2022-07-15 14:55:51 +03:00
|
|
|
paths = append(paths, path)
|
2020-12-14 12:27:15 +03:00
|
|
|
hashes = append(hashes, hash)
|
2022-07-15 14:55:51 +03:00
|
|
|
if len(paths) >= cap {
|
2020-12-14 12:27:15 +03:00
|
|
|
break
|
|
|
|
}
|
|
|
|
}
|
2022-05-10 17:37:24 +03:00
|
|
|
// Group requests by account hash
|
2022-07-15 14:55:51 +03:00
|
|
|
paths, hashes, _, pathsets = sortByAccountPath(paths, hashes)
|
2020-12-14 12:27:15 +03:00
|
|
|
req := &trienodeHealRequest{
|
2021-04-15 21:01:16 +03:00
|
|
|
peer: idle,
|
|
|
|
id: reqid,
|
2021-05-19 15:09:03 +03:00
|
|
|
time: time.Now(),
|
2021-04-15 21:01:16 +03:00
|
|
|
deliver: success,
|
|
|
|
revert: fail,
|
|
|
|
cancel: cancel,
|
|
|
|
stale: make(chan struct{}),
|
|
|
|
paths: paths,
|
2022-07-15 14:55:51 +03:00
|
|
|
hashes: hashes,
|
2021-04-15 21:01:16 +03:00
|
|
|
task: s.healer,
|
2020-12-14 12:27:15 +03:00
|
|
|
}
|
2021-05-19 15:09:03 +03:00
|
|
|
req.timeout = time.AfterFunc(s.rates.TargetTimeout(), func() {
|
2021-03-26 23:29:22 +03:00
|
|
|
peer.Log().Debug("Trienode heal request timed out", "reqid", reqid)
|
2021-05-19 15:09:03 +03:00
|
|
|
s.rates.Update(idle, TrieNodesMsg, 0, 0)
|
2021-01-07 13:58:07 +03:00
|
|
|
s.scheduleRevertTrienodeHealRequest(req)
|
2020-12-14 12:27:15 +03:00
|
|
|
})
|
|
|
|
s.trienodeHealReqs[reqid] = req
|
|
|
|
delete(s.trienodeHealIdlers, idle)
|
|
|
|
|
|
|
|
s.pend.Add(1)
|
2021-03-26 23:29:22 +03:00
|
|
|
go func(root common.Hash) {
|
2020-12-14 12:27:15 +03:00
|
|
|
defer s.pend.Done()
|
|
|
|
|
|
|
|
// Attempt to send the remote request and revert if it fails
|
|
|
|
if err := peer.RequestTrieNodes(reqid, root, pathsets, maxRequestSize); err != nil {
|
|
|
|
log.Debug("Failed to request trienode healers", "err", err)
|
2021-01-07 13:58:07 +03:00
|
|
|
s.scheduleRevertTrienodeHealRequest(req)
|
2020-12-14 12:27:15 +03:00
|
|
|
}
|
2021-03-26 23:29:22 +03:00
|
|
|
}(s.root)
|
2020-12-14 12:27:15 +03:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// assignBytecodeHealTasks attempts to match idle peers to bytecode requests to
|
|
|
|
// heal any trie errors caused by the snap sync's chunked retrieval model.
|
2021-04-15 21:01:16 +03:00
|
|
|
func (s *Syncer) assignBytecodeHealTasks(success chan *bytecodeHealResponse, fail chan *bytecodeHealRequest, cancel chan struct{}) {
|
2020-12-14 12:27:15 +03:00
|
|
|
s.lock.Lock()
|
|
|
|
defer s.lock.Unlock()
|
|
|
|
|
2021-05-19 15:09:03 +03:00
|
|
|
// Sort the peers by download capacity to use faster ones if many available
|
|
|
|
idlers := &capacitySort{
|
|
|
|
ids: make([]string, 0, len(s.bytecodeHealIdlers)),
|
2021-05-27 19:43:55 +03:00
|
|
|
caps: make([]int, 0, len(s.bytecodeHealIdlers)),
|
2021-05-19 15:09:03 +03:00
|
|
|
}
|
|
|
|
targetTTL := s.rates.TargetTimeout()
|
|
|
|
for id := range s.bytecodeHealIdlers {
|
|
|
|
if _, ok := s.statelessPeers[id]; ok {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
idlers.ids = append(idlers.ids, id)
|
|
|
|
idlers.caps = append(idlers.caps, s.rates.Capacity(id, ByteCodesMsg, targetTTL))
|
|
|
|
}
|
|
|
|
if len(idlers.ids) == 0 {
|
2020-12-14 12:27:15 +03:00
|
|
|
return
|
|
|
|
}
|
2021-05-19 15:09:03 +03:00
|
|
|
sort.Sort(sort.Reverse(idlers))
|
|
|
|
|
2020-12-14 12:27:15 +03:00
|
|
|
// Iterate over pending tasks and try to find a peer to retrieve with
|
|
|
|
for len(s.healer.codeTasks) > 0 || s.healer.scheduler.Pending() > 0 {
|
|
|
|
// If there are not enough trie tasks queued to fully assign, fill the
|
|
|
|
// queue from the state sync scheduler. The trie synced schedules these
|
|
|
|
// together with trie nodes, so we need to queue them combined.
|
|
|
|
var (
|
|
|
|
have = len(s.healer.trieTasks) + len(s.healer.codeTasks)
|
|
|
|
want = maxTrieRequestCount + maxCodeRequestCount
|
|
|
|
)
|
|
|
|
if have < want {
|
2022-07-15 14:55:51 +03:00
|
|
|
paths, hashes, codes := s.healer.scheduler.Missing(want - have)
|
|
|
|
for i, path := range paths {
|
|
|
|
s.healer.trieTasks[path] = hashes[i]
|
2020-12-14 12:27:15 +03:00
|
|
|
}
|
|
|
|
for _, hash := range codes {
|
|
|
|
s.healer.codeTasks[hash] = struct{}{}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// If all the heal tasks are trienodes or already downloading, bail
|
|
|
|
if len(s.healer.codeTasks) == 0 {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
// Task pending retrieval, try to find an idle peer. If no such peer
|
|
|
|
// exists, we probably assigned tasks for all (or they are stateless).
|
|
|
|
// Abort the entire assignment mechanism.
|
2021-05-19 15:09:03 +03:00
|
|
|
if len(idlers.ids) == 0 {
|
2020-12-14 12:27:15 +03:00
|
|
|
return
|
|
|
|
}
|
2021-05-19 15:09:03 +03:00
|
|
|
var (
|
|
|
|
idle = idlers.ids[0]
|
|
|
|
peer = s.peers[idle]
|
|
|
|
cap = idlers.caps[0]
|
|
|
|
)
|
|
|
|
idlers.ids, idlers.caps = idlers.ids[1:], idlers.caps[1:]
|
2021-03-26 23:29:22 +03:00
|
|
|
|
2020-12-14 12:27:15 +03:00
|
|
|
// Matched a pending task to an idle peer, allocate a unique request id
|
|
|
|
var reqid uint64
|
|
|
|
for {
|
|
|
|
reqid = uint64(rand.Int63())
|
|
|
|
if reqid == 0 {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
if _, ok := s.bytecodeHealReqs[reqid]; ok {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
break
|
|
|
|
}
|
|
|
|
// Generate the network query and send it to the peer
|
2021-05-19 15:09:03 +03:00
|
|
|
if cap > maxCodeRequestCount {
|
|
|
|
cap = maxCodeRequestCount
|
|
|
|
}
|
2021-05-27 19:43:55 +03:00
|
|
|
hashes := make([]common.Hash, 0, cap)
|
2020-12-14 12:27:15 +03:00
|
|
|
for hash := range s.healer.codeTasks {
|
|
|
|
delete(s.healer.codeTasks, hash)
|
|
|
|
|
|
|
|
hashes = append(hashes, hash)
|
2021-05-27 19:43:55 +03:00
|
|
|
if len(hashes) >= cap {
|
2020-12-14 12:27:15 +03:00
|
|
|
break
|
|
|
|
}
|
|
|
|
}
|
|
|
|
req := &bytecodeHealRequest{
|
2021-04-15 21:01:16 +03:00
|
|
|
peer: idle,
|
|
|
|
id: reqid,
|
2021-05-19 15:09:03 +03:00
|
|
|
time: time.Now(),
|
2021-04-15 21:01:16 +03:00
|
|
|
deliver: success,
|
|
|
|
revert: fail,
|
|
|
|
cancel: cancel,
|
|
|
|
stale: make(chan struct{}),
|
|
|
|
hashes: hashes,
|
|
|
|
task: s.healer,
|
2020-12-14 12:27:15 +03:00
|
|
|
}
|
2021-05-19 15:09:03 +03:00
|
|
|
req.timeout = time.AfterFunc(s.rates.TargetTimeout(), func() {
|
2021-03-26 23:29:22 +03:00
|
|
|
peer.Log().Debug("Bytecode heal request timed out", "reqid", reqid)
|
2021-05-19 15:09:03 +03:00
|
|
|
s.rates.Update(idle, ByteCodesMsg, 0, 0)
|
2021-01-07 13:58:07 +03:00
|
|
|
s.scheduleRevertBytecodeHealRequest(req)
|
2020-12-14 12:27:15 +03:00
|
|
|
})
|
|
|
|
s.bytecodeHealReqs[reqid] = req
|
|
|
|
delete(s.bytecodeHealIdlers, idle)
|
|
|
|
|
|
|
|
s.pend.Add(1)
|
2021-03-26 23:29:22 +03:00
|
|
|
go func() {
|
2020-12-14 12:27:15 +03:00
|
|
|
defer s.pend.Done()
|
|
|
|
|
|
|
|
// Attempt to send the remote request and revert if it fails
|
|
|
|
if err := peer.RequestByteCodes(reqid, hashes, maxRequestSize); err != nil {
|
|
|
|
log.Debug("Failed to request bytecode healers", "err", err)
|
2021-01-07 13:58:07 +03:00
|
|
|
s.scheduleRevertBytecodeHealRequest(req)
|
2020-12-14 12:27:15 +03:00
|
|
|
}
|
2021-03-26 23:29:22 +03:00
|
|
|
}()
|
2020-12-14 12:27:15 +03:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2022-08-19 09:00:21 +03:00
|
|
|
// revertRequests locates all the currently pending requests from a particular
|
2020-12-14 12:27:15 +03:00
|
|
|
// peer and reverts them, rescheduling for others to fulfill.
|
|
|
|
func (s *Syncer) revertRequests(peer string) {
|
|
|
|
// Gather the requests first, revertals need the lock too
|
|
|
|
s.lock.Lock()
|
|
|
|
var accountReqs []*accountRequest
|
|
|
|
for _, req := range s.accountReqs {
|
|
|
|
if req.peer == peer {
|
|
|
|
accountReqs = append(accountReqs, req)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
var bytecodeReqs []*bytecodeRequest
|
|
|
|
for _, req := range s.bytecodeReqs {
|
|
|
|
if req.peer == peer {
|
|
|
|
bytecodeReqs = append(bytecodeReqs, req)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
var storageReqs []*storageRequest
|
|
|
|
for _, req := range s.storageReqs {
|
|
|
|
if req.peer == peer {
|
|
|
|
storageReqs = append(storageReqs, req)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
var trienodeHealReqs []*trienodeHealRequest
|
|
|
|
for _, req := range s.trienodeHealReqs {
|
|
|
|
if req.peer == peer {
|
|
|
|
trienodeHealReqs = append(trienodeHealReqs, req)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
var bytecodeHealReqs []*bytecodeHealRequest
|
|
|
|
for _, req := range s.bytecodeHealReqs {
|
|
|
|
if req.peer == peer {
|
|
|
|
bytecodeHealReqs = append(bytecodeHealReqs, req)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
s.lock.Unlock()
|
|
|
|
|
|
|
|
// Revert all the requests matching the peer
|
|
|
|
for _, req := range accountReqs {
|
|
|
|
s.revertAccountRequest(req)
|
|
|
|
}
|
|
|
|
for _, req := range bytecodeReqs {
|
|
|
|
s.revertBytecodeRequest(req)
|
|
|
|
}
|
|
|
|
for _, req := range storageReqs {
|
|
|
|
s.revertStorageRequest(req)
|
|
|
|
}
|
|
|
|
for _, req := range trienodeHealReqs {
|
|
|
|
s.revertTrienodeHealRequest(req)
|
|
|
|
}
|
|
|
|
for _, req := range bytecodeHealReqs {
|
|
|
|
s.revertBytecodeHealRequest(req)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2021-01-07 13:58:07 +03:00
|
|
|
// scheduleRevertAccountRequest asks the event loop to clean up an account range
|
|
|
|
// request and return all failed retrieval tasks to the scheduler for reassignment.
|
|
|
|
func (s *Syncer) scheduleRevertAccountRequest(req *accountRequest) {
|
|
|
|
select {
|
2021-04-15 21:01:16 +03:00
|
|
|
case req.revert <- req:
|
2021-01-07 13:58:07 +03:00
|
|
|
// Sync event loop notified
|
|
|
|
case <-req.cancel:
|
|
|
|
// Sync cycle got cancelled
|
|
|
|
case <-req.stale:
|
|
|
|
// Request already reverted
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-12-14 12:27:15 +03:00
|
|
|
// revertAccountRequest cleans up an account range request and returns all failed
|
|
|
|
// retrieval tasks to the scheduler for reassignment.
|
2021-01-07 13:58:07 +03:00
|
|
|
//
|
|
|
|
// Note, this needs to run on the event runloop thread to reschedule to idle peers.
|
|
|
|
// On peer threads, use scheduleRevertAccountRequest.
|
2020-12-14 12:27:15 +03:00
|
|
|
func (s *Syncer) revertAccountRequest(req *accountRequest) {
|
2021-01-07 13:58:07 +03:00
|
|
|
log.Debug("Reverting account request", "peer", req.peer, "reqid", req.id)
|
2020-12-14 12:27:15 +03:00
|
|
|
select {
|
|
|
|
case <-req.stale:
|
|
|
|
log.Trace("Account request already reverted", "peer", req.peer, "reqid", req.id)
|
|
|
|
return
|
|
|
|
default:
|
|
|
|
}
|
|
|
|
close(req.stale)
|
|
|
|
|
|
|
|
// Remove the request from the tracked set
|
|
|
|
s.lock.Lock()
|
|
|
|
delete(s.accountReqs, req.id)
|
|
|
|
s.lock.Unlock()
|
|
|
|
|
|
|
|
// If there's a timeout timer still running, abort it and mark the account
|
2022-08-19 09:00:21 +03:00
|
|
|
// task as not-pending, ready for rescheduling
|
2020-12-14 12:27:15 +03:00
|
|
|
req.timeout.Stop()
|
|
|
|
if req.task.req == req {
|
|
|
|
req.task.req = nil
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2021-01-07 13:58:07 +03:00
|
|
|
// scheduleRevertBytecodeRequest asks the event loop to clean up a bytecode request
|
|
|
|
// and return all failed retrieval tasks to the scheduler for reassignment.
|
|
|
|
func (s *Syncer) scheduleRevertBytecodeRequest(req *bytecodeRequest) {
|
|
|
|
select {
|
2021-04-15 21:01:16 +03:00
|
|
|
case req.revert <- req:
|
2021-01-07 13:58:07 +03:00
|
|
|
// Sync event loop notified
|
|
|
|
case <-req.cancel:
|
|
|
|
// Sync cycle got cancelled
|
|
|
|
case <-req.stale:
|
|
|
|
// Request already reverted
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// revertBytecodeRequest cleans up a bytecode request and returns all failed
|
2020-12-14 12:27:15 +03:00
|
|
|
// retrieval tasks to the scheduler for reassignment.
|
2021-01-07 13:58:07 +03:00
|
|
|
//
|
|
|
|
// Note, this needs to run on the event runloop thread to reschedule to idle peers.
|
|
|
|
// On peer threads, use scheduleRevertBytecodeRequest.
|
2020-12-14 12:27:15 +03:00
|
|
|
func (s *Syncer) revertBytecodeRequest(req *bytecodeRequest) {
|
2021-01-07 13:58:07 +03:00
|
|
|
log.Debug("Reverting bytecode request", "peer", req.peer)
|
2020-12-14 12:27:15 +03:00
|
|
|
select {
|
|
|
|
case <-req.stale:
|
|
|
|
log.Trace("Bytecode request already reverted", "peer", req.peer, "reqid", req.id)
|
|
|
|
return
|
|
|
|
default:
|
|
|
|
}
|
|
|
|
close(req.stale)
|
|
|
|
|
|
|
|
// Remove the request from the tracked set
|
|
|
|
s.lock.Lock()
|
|
|
|
delete(s.bytecodeReqs, req.id)
|
|
|
|
s.lock.Unlock()
|
|
|
|
|
|
|
|
// If there's a timeout timer still running, abort it and mark the code
|
2022-08-19 09:00:21 +03:00
|
|
|
// retrievals as not-pending, ready for rescheduling
|
2020-12-14 12:27:15 +03:00
|
|
|
req.timeout.Stop()
|
|
|
|
for _, hash := range req.hashes {
|
|
|
|
req.task.codeTasks[hash] = struct{}{}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2021-01-07 13:58:07 +03:00
|
|
|
// scheduleRevertStorageRequest asks the event loop to clean up a storage range
|
|
|
|
// request and return all failed retrieval tasks to the scheduler for reassignment.
|
|
|
|
func (s *Syncer) scheduleRevertStorageRequest(req *storageRequest) {
|
|
|
|
select {
|
2021-04-15 21:01:16 +03:00
|
|
|
case req.revert <- req:
|
2021-01-07 13:58:07 +03:00
|
|
|
// Sync event loop notified
|
|
|
|
case <-req.cancel:
|
|
|
|
// Sync cycle got cancelled
|
|
|
|
case <-req.stale:
|
|
|
|
// Request already reverted
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-12-14 12:27:15 +03:00
|
|
|
// revertStorageRequest cleans up a storage range request and returns all failed
|
|
|
|
// retrieval tasks to the scheduler for reassignment.
|
2021-01-07 13:58:07 +03:00
|
|
|
//
|
|
|
|
// Note, this needs to run on the event runloop thread to reschedule to idle peers.
|
|
|
|
// On peer threads, use scheduleRevertStorageRequest.
|
2020-12-14 12:27:15 +03:00
|
|
|
func (s *Syncer) revertStorageRequest(req *storageRequest) {
|
2021-01-07 13:58:07 +03:00
|
|
|
log.Debug("Reverting storage request", "peer", req.peer)
|
2020-12-14 12:27:15 +03:00
|
|
|
select {
|
|
|
|
case <-req.stale:
|
|
|
|
log.Trace("Storage request already reverted", "peer", req.peer, "reqid", req.id)
|
|
|
|
return
|
|
|
|
default:
|
|
|
|
}
|
|
|
|
close(req.stale)
|
|
|
|
|
|
|
|
// Remove the request from the tracked set
|
|
|
|
s.lock.Lock()
|
|
|
|
delete(s.storageReqs, req.id)
|
|
|
|
s.lock.Unlock()
|
|
|
|
|
|
|
|
// If there's a timeout timer still running, abort it and mark the storage
|
2022-08-19 09:00:21 +03:00
|
|
|
// task as not-pending, ready for rescheduling
|
2020-12-14 12:27:15 +03:00
|
|
|
req.timeout.Stop()
|
|
|
|
if req.subTask != nil {
|
|
|
|
req.subTask.req = nil
|
|
|
|
} else {
|
|
|
|
for i, account := range req.accounts {
|
|
|
|
req.mainTask.stateTasks[account] = req.roots[i]
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2021-01-07 13:58:07 +03:00
|
|
|
// scheduleRevertTrienodeHealRequest asks the event loop to clean up a trienode heal
|
|
|
|
// request and return all failed retrieval tasks to the scheduler for reassignment.
|
|
|
|
func (s *Syncer) scheduleRevertTrienodeHealRequest(req *trienodeHealRequest) {
|
|
|
|
select {
|
2021-04-15 21:01:16 +03:00
|
|
|
case req.revert <- req:
|
2021-01-07 13:58:07 +03:00
|
|
|
// Sync event loop notified
|
|
|
|
case <-req.cancel:
|
|
|
|
// Sync cycle got cancelled
|
|
|
|
case <-req.stale:
|
|
|
|
// Request already reverted
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// revertTrienodeHealRequest cleans up a trienode heal request and returns all
|
2020-12-14 12:27:15 +03:00
|
|
|
// failed retrieval tasks to the scheduler for reassignment.
|
2021-01-07 13:58:07 +03:00
|
|
|
//
|
|
|
|
// Note, this needs to run on the event runloop thread to reschedule to idle peers.
|
|
|
|
// On peer threads, use scheduleRevertTrienodeHealRequest.
|
2020-12-14 12:27:15 +03:00
|
|
|
func (s *Syncer) revertTrienodeHealRequest(req *trienodeHealRequest) {
|
2021-01-07 13:58:07 +03:00
|
|
|
log.Debug("Reverting trienode heal request", "peer", req.peer)
|
2020-12-14 12:27:15 +03:00
|
|
|
select {
|
|
|
|
case <-req.stale:
|
|
|
|
log.Trace("Trienode heal request already reverted", "peer", req.peer, "reqid", req.id)
|
|
|
|
return
|
|
|
|
default:
|
|
|
|
}
|
|
|
|
close(req.stale)
|
|
|
|
|
|
|
|
// Remove the request from the tracked set
|
|
|
|
s.lock.Lock()
|
|
|
|
delete(s.trienodeHealReqs, req.id)
|
|
|
|
s.lock.Unlock()
|
|
|
|
|
|
|
|
// If there's a timeout timer still running, abort it and mark the trie node
|
2022-07-15 14:55:51 +03:00
|
|
|
// retrievals as not-pending, ready for rescheduling
|
2020-12-14 12:27:15 +03:00
|
|
|
req.timeout.Stop()
|
2022-07-15 14:55:51 +03:00
|
|
|
for i, path := range req.paths {
|
|
|
|
req.task.trieTasks[path] = req.hashes[i]
|
2020-12-14 12:27:15 +03:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2021-01-07 13:58:07 +03:00
|
|
|
// scheduleRevertBytecodeHealRequest asks the event loop to clean up a bytecode heal
|
|
|
|
// request and return all failed retrieval tasks to the scheduler for reassignment.
|
|
|
|
func (s *Syncer) scheduleRevertBytecodeHealRequest(req *bytecodeHealRequest) {
|
|
|
|
select {
|
2021-04-15 21:01:16 +03:00
|
|
|
case req.revert <- req:
|
2021-01-07 13:58:07 +03:00
|
|
|
// Sync event loop notified
|
|
|
|
case <-req.cancel:
|
|
|
|
// Sync cycle got cancelled
|
|
|
|
case <-req.stale:
|
|
|
|
// Request already reverted
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// revertBytecodeHealRequest cleans up a bytecode heal request and returns all
|
|
|
|
// failed retrieval tasks to the scheduler for reassignment.
|
|
|
|
//
|
|
|
|
// Note, this needs to run on the event runloop thread to reschedule to idle peers.
|
|
|
|
// On peer threads, use scheduleRevertBytecodeHealRequest.
|
2020-12-14 12:27:15 +03:00
|
|
|
func (s *Syncer) revertBytecodeHealRequest(req *bytecodeHealRequest) {
|
2021-01-07 13:58:07 +03:00
|
|
|
log.Debug("Reverting bytecode heal request", "peer", req.peer)
|
2020-12-14 12:27:15 +03:00
|
|
|
select {
|
|
|
|
case <-req.stale:
|
|
|
|
log.Trace("Bytecode heal request already reverted", "peer", req.peer, "reqid", req.id)
|
|
|
|
return
|
|
|
|
default:
|
|
|
|
}
|
|
|
|
close(req.stale)
|
|
|
|
|
|
|
|
// Remove the request from the tracked set
|
|
|
|
s.lock.Lock()
|
|
|
|
delete(s.bytecodeHealReqs, req.id)
|
|
|
|
s.lock.Unlock()
|
|
|
|
|
|
|
|
// If there's a timeout timer still running, abort it and mark the code
|
2022-08-19 09:00:21 +03:00
|
|
|
// retrievals as not-pending, ready for rescheduling
|
2020-12-14 12:27:15 +03:00
|
|
|
req.timeout.Stop()
|
|
|
|
for _, hash := range req.hashes {
|
|
|
|
req.task.codeTasks[hash] = struct{}{}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// processAccountResponse integrates an already validated account range response
|
|
|
|
// into the account tasks.
|
|
|
|
func (s *Syncer) processAccountResponse(res *accountResponse) {
|
|
|
|
// Switch the task from pending to filling
|
|
|
|
res.task.req = nil
|
|
|
|
res.task.res = res
|
|
|
|
|
|
|
|
// Ensure that the response doesn't overflow into the subsequent task
|
|
|
|
last := res.task.Last.Big()
|
|
|
|
for i, hash := range res.hashes {
|
2021-03-24 17:33:34 +03:00
|
|
|
// Mark the range complete if the last is already included.
|
|
|
|
// Keep iteration to delete the extra states if exists.
|
|
|
|
cmp := hash.Big().Cmp(last)
|
|
|
|
if cmp == 0 {
|
|
|
|
res.cont = false
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
if cmp > 0 {
|
2021-04-27 17:19:59 +03:00
|
|
|
// Chunk overflown, cut off excess
|
2020-12-14 12:27:15 +03:00
|
|
|
res.hashes = res.hashes[:i]
|
|
|
|
res.accounts = res.accounts[:i]
|
|
|
|
res.cont = false // Mark range completed
|
|
|
|
break
|
|
|
|
}
|
|
|
|
}
|
2021-01-25 09:17:05 +03:00
|
|
|
// Iterate over all the accounts and assemble which ones need further sub-
|
2020-12-14 12:27:15 +03:00
|
|
|
// filling before the entire account range can be persisted.
|
|
|
|
res.task.needCode = make([]bool, len(res.accounts))
|
|
|
|
res.task.needState = make([]bool, len(res.accounts))
|
|
|
|
res.task.needHeal = make([]bool, len(res.accounts))
|
|
|
|
|
|
|
|
res.task.codeTasks = make(map[common.Hash]struct{})
|
|
|
|
res.task.stateTasks = make(map[common.Hash]common.Hash)
|
|
|
|
|
|
|
|
resumed := make(map[common.Hash]struct{})
|
|
|
|
|
|
|
|
res.task.pend = 0
|
|
|
|
for i, account := range res.accounts {
|
|
|
|
// Check if the account is a contract with an unknown code
|
2023-02-21 14:12:27 +03:00
|
|
|
if !bytes.Equal(account.CodeHash, types.EmptyCodeHash.Bytes()) {
|
2021-12-15 18:16:45 +03:00
|
|
|
if !rawdb.HasCodeWithPrefix(s.db, common.BytesToHash(account.CodeHash)) {
|
2020-12-14 12:27:15 +03:00
|
|
|
res.task.codeTasks[common.BytesToHash(account.CodeHash)] = struct{}{}
|
|
|
|
res.task.needCode[i] = true
|
|
|
|
res.task.pend++
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// Check if the account is a contract with an unknown storage trie
|
2023-02-21 14:12:27 +03:00
|
|
|
if account.Root != types.EmptyRootHash {
|
2023-02-06 18:28:40 +03:00
|
|
|
if !rawdb.HasTrieNode(s.db, res.hashes[i], nil, account.Root, s.scheme) {
|
2020-12-14 12:27:15 +03:00
|
|
|
// If there was a previous large state retrieval in progress,
|
|
|
|
// don't restart it from scratch. This happens if a sync cycle
|
|
|
|
// is interrupted and resumed later. However, *do* update the
|
|
|
|
// previous root hash.
|
|
|
|
if subtasks, ok := res.task.SubTasks[res.hashes[i]]; ok {
|
2021-02-25 13:56:18 +03:00
|
|
|
log.Debug("Resuming large storage retrieval", "account", res.hashes[i], "root", account.Root)
|
2020-12-14 12:27:15 +03:00
|
|
|
for _, subtask := range subtasks {
|
|
|
|
subtask.root = account.Root
|
|
|
|
}
|
|
|
|
res.task.needHeal[i] = true
|
|
|
|
resumed[res.hashes[i]] = struct{}{}
|
|
|
|
} else {
|
|
|
|
res.task.stateTasks[res.hashes[i]] = account.Root
|
|
|
|
}
|
|
|
|
res.task.needState[i] = true
|
|
|
|
res.task.pend++
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// Delete any subtasks that have been aborted but not resumed. This may undo
|
2021-01-25 09:17:05 +03:00
|
|
|
// some progress if a new peer gives us less accounts than an old one, but for
|
2020-12-14 12:27:15 +03:00
|
|
|
// now we have to live with that.
|
|
|
|
for hash := range res.task.SubTasks {
|
|
|
|
if _, ok := resumed[hash]; !ok {
|
2021-02-25 13:56:18 +03:00
|
|
|
log.Debug("Aborting suspended storage retrieval", "account", hash)
|
2020-12-14 12:27:15 +03:00
|
|
|
delete(res.task.SubTasks, hash)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// If the account range contained no contracts, or all have been fully filled
|
|
|
|
// beforehand, short circuit storage filling and forward to the next task
|
|
|
|
if res.task.pend == 0 {
|
|
|
|
s.forwardAccountTask(res.task)
|
|
|
|
return
|
|
|
|
}
|
|
|
|
// Some accounts are incomplete, leave as is for the storage and contract
|
eth/protocols/snap: fix snap sync failure on empty storage range (#28306)
This change addresses an issue in snap sync, specifically when the entire sync process can be halted due to an encountered empty storage range.
Currently, on the snap sync client side, the response to an empty (partial) storage range is discarded as a non-delivery. However, this response can be a valid response, when the particular range requested does not contain any slots.
For instance, consider a large contract where the entire key space is divided into 16 chunks, and there are no available slots in the last chunk [0xf] -> [end]. When the node receives a request for this particular range, the response includes:
The proof with origin [0xf]
A nil storage slot set
If we simply discard this response, the finalization of the last range will be skipped, halting the entire sync process indefinitely. The test case TestSyncWithUnevenStorage can reproduce the scenario described above.
In addition, this change also defines the common variables MaxAddress and MaxHash.
2023-10-13 10:08:26 +03:00
|
|
|
// task assigners to pick up and fill
|
2020-12-14 12:27:15 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// processBytecodeResponse integrates an already validated bytecode response
|
|
|
|
// into the account tasks.
|
|
|
|
func (s *Syncer) processBytecodeResponse(res *bytecodeResponse) {
|
|
|
|
batch := s.db.NewBatch()
|
|
|
|
|
|
|
|
var (
|
|
|
|
codes uint64
|
|
|
|
)
|
|
|
|
for i, hash := range res.hashes {
|
|
|
|
code := res.codes[i]
|
|
|
|
|
|
|
|
// If the bytecode was not delivered, reschedule it
|
|
|
|
if code == nil {
|
|
|
|
res.task.codeTasks[hash] = struct{}{}
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
// Code was delivered, mark it not needed any more
|
|
|
|
for j, account := range res.task.res.accounts {
|
|
|
|
if res.task.needCode[j] && hash == common.BytesToHash(account.CodeHash) {
|
|
|
|
res.task.needCode[j] = false
|
|
|
|
res.task.pend--
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// Push the bytecode into a database batch
|
|
|
|
codes++
|
|
|
|
rawdb.WriteCode(batch, hash, code)
|
|
|
|
}
|
2021-04-27 17:19:59 +03:00
|
|
|
bytes := common.StorageSize(batch.ValueSize())
|
2020-12-14 12:27:15 +03:00
|
|
|
if err := batch.Write(); err != nil {
|
|
|
|
log.Crit("Failed to persist bytecodes", "err", err)
|
|
|
|
}
|
2021-04-27 17:19:59 +03:00
|
|
|
s.bytecodeSynced += codes
|
|
|
|
s.bytecodeBytes += bytes
|
|
|
|
|
2020-12-14 12:27:15 +03:00
|
|
|
log.Debug("Persisted set of bytecodes", "count", codes, "bytes", bytes)
|
|
|
|
|
|
|
|
// If this delivery completed the last pending task, forward the account task
|
|
|
|
// to the next chunk
|
|
|
|
if res.task.pend == 0 {
|
|
|
|
s.forwardAccountTask(res.task)
|
|
|
|
return
|
|
|
|
}
|
|
|
|
// Some accounts are still incomplete, leave as is for the storage and contract
|
|
|
|
// task assigners to pick up and fill.
|
|
|
|
}
|
|
|
|
|
|
|
|
// processStorageResponse integrates an already validated storage response
|
|
|
|
// into the account tasks.
|
|
|
|
func (s *Syncer) processStorageResponse(res *storageResponse) {
|
core, eth: faster snapshot generation (#22504)
* eth/protocols: persist received state segments
* core: initial implementation
* core/state/snapshot: add tests
* core, eth: updates
* eth/protocols/snapshot: count flat state size
* core/state: add metrics
* core/state/snapshot: skip unnecessary deletion
* core/state/snapshot: rename
* core/state/snapshot: use the global batch
* core/state/snapshot: add logs and fix wiping
* core/state/snapshot: fix
* core/state/snapshot: save generation progress even if the batch is empty
* core/state/snapshot: fixes
* core/state/snapshot: fix initial account range length
* core/state/snapshot: fix initial account range
* eth/protocols/snap: store flat states during the healing
* eth/protocols/snap: print logs
* core/state/snapshot: refactor (#4)
* core/state/snapshot: refactor
* core/state/snapshot: tiny fix and polish
Co-authored-by: rjl493456442 <garyrong0905@gmail.com>
* core, eth: fixes
* core, eth: fix healing writer
* core, trie, eth: fix paths
* eth/protocols/snap: fix encoding
* eth, core: add debug log
* core/state/generate: release iterator asap (#5)
core/state/snapshot: less copy
core/state/snapshot: revert split loop
core/state/snapshot: handle storage becoming empty, improve test robustness
core/state: test modified codehash
core/state/snapshot: polish
* core/state/snapshot: optimize stats counter
* core, eth: add metric
* core/state/snapshot: update comments
* core/state/snapshot: improve tests
* core/state/snapshot: replace secure trie with standard trie
* core/state/snapshot: wrap return as the struct
* core/state/snapshot: skip wiping correct states
* core/state/snapshot: updates
* core/state/snapshot: fixes
* core/state/snapshot: fix panic due to reference flaw in closure
* core/state/snapshot: fix errors in state generation logic + fix log output
* core/state/snapshot: remove an error case
* core/state/snapshot: fix condition-check for exhausted snap state
* core/state/snapshot: use stackTrie for small tries
* core/state/snapshot: don't resolve small storage tries in vain
* core/state/snapshot: properly clean up storage of deleted accounts
* core/state/snapshot: avoid RLP-encoding in some cases + minor nitpicks
* core/state/snapshot: fix error (+testcase)
* core/state/snapshot: clean up tests a bit
* core/state/snapshot: work in progress on better tests
* core/state/snapshot: polish code
* core/state/snapshot: fix trie iteration abortion trigger
* core/state/snapshot: fixes flaws
* core/state/snapshot: remove panic
* core/state/snapshot: fix abort
* core/state/snapshot: more tests (plus failing testcase)
* core/state/snapshot: more testcases + fix for failing test
* core/state/snapshot: testcase for malformed data
* core/state/snapshot: some test nitpicks
* core/state/snapshot: improvements to logging
* core/state/snapshot: testcase to demo error in abortion
* core/state/snapshot: fix abortion
* cmd/geth: make verify-state report the root
* trie: fix failing test
* core/state/snapshot: add timer metrics
* core/state/snapshot: fix metrics
* core/state/snapshot: udpate tests
* eth/protocols/snap: write snapshot account even if code or state is needed
* core/state/snapshot: fix diskmore check
* core/state/snapshot: review fixes
* core/state/snapshot: improve error message
* cmd/geth: rename 'error' to 'err' in logs
* core/state/snapshot: fix some review concerns
* core/state/snapshot, eth/protocols/snap: clear snapshot marker when starting/resuming snap sync
* core: add error log
* core/state/snapshot: use proper timers for metrics collection
* core/state/snapshot: address some review concerns
* eth/protocols/snap: improved log message
* eth/protocols/snap: fix heal logs to condense infos
* core/state/snapshot: wait for generator termination before restarting
* core/state/snapshot: revert timers to counters to track total time
Co-authored-by: Martin Holst Swende <martin@swende.se>
Co-authored-by: Péter Szilágyi <peterke@gmail.com>
2021-04-14 23:23:11 +03:00
|
|
|
// Switch the subtask from pending to idle
|
2020-12-14 12:27:15 +03:00
|
|
|
if res.subTask != nil {
|
|
|
|
res.subTask.req = nil
|
|
|
|
}
|
2021-04-28 23:09:15 +03:00
|
|
|
batch := ethdb.HookedBatch{
|
|
|
|
Batch: s.db.NewBatch(),
|
|
|
|
OnPut: func(key []byte, value []byte) {
|
|
|
|
s.storageBytes += common.StorageSize(len(key) + len(value))
|
|
|
|
},
|
|
|
|
}
|
2020-12-14 12:27:15 +03:00
|
|
|
var (
|
2021-04-28 23:09:15 +03:00
|
|
|
slots int
|
|
|
|
oldStorageBytes = s.storageBytes
|
2020-12-14 12:27:15 +03:00
|
|
|
)
|
|
|
|
// Iterate over all the accounts and reconstruct their storage tries from the
|
|
|
|
// delivered slots
|
|
|
|
for i, account := range res.accounts {
|
|
|
|
// If the account was not delivered, reschedule it
|
|
|
|
if i >= len(res.hashes) {
|
2021-01-25 09:17:05 +03:00
|
|
|
res.mainTask.stateTasks[account] = res.roots[i]
|
2020-12-14 12:27:15 +03:00
|
|
|
continue
|
|
|
|
}
|
|
|
|
// State was delivered, if complete mark as not needed any more, otherwise
|
|
|
|
// mark the account as needing healing
|
2021-01-25 09:17:05 +03:00
|
|
|
for j, hash := range res.mainTask.res.hashes {
|
|
|
|
if account != hash {
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
acc := res.mainTask.res.accounts[j]
|
|
|
|
|
|
|
|
// If the packet contains multiple contract storage slots, all
|
|
|
|
// but the last are surely complete. The last contract may be
|
|
|
|
// chunked, so check it's continuation flag.
|
|
|
|
if res.subTask == nil && res.mainTask.needState[j] && (i < len(res.hashes)-1 || !res.cont) {
|
|
|
|
res.mainTask.needState[j] = false
|
|
|
|
res.mainTask.pend--
|
2023-10-23 18:31:56 +03:00
|
|
|
smallStorageGauge.Inc(1)
|
2021-01-25 09:17:05 +03:00
|
|
|
}
|
|
|
|
// If the last contract was chunked, mark it as needing healing
|
|
|
|
// to avoid writing it out to disk prematurely.
|
|
|
|
if res.subTask == nil && !res.mainTask.needHeal[j] && i == len(res.hashes)-1 && res.cont {
|
|
|
|
res.mainTask.needHeal[j] = true
|
|
|
|
}
|
|
|
|
// If the last contract was chunked, we need to switch to large
|
|
|
|
// contract handling mode
|
|
|
|
if res.subTask == nil && i == len(res.hashes)-1 && res.cont {
|
|
|
|
// If we haven't yet started a large-contract retrieval, create
|
|
|
|
// the subtasks for it within the main account task
|
|
|
|
if tasks, ok := res.mainTask.SubTasks[account]; !ok {
|
|
|
|
var (
|
2021-04-27 17:19:59 +03:00
|
|
|
keys = res.hashes[i]
|
|
|
|
chunks = uint64(storageConcurrency)
|
|
|
|
lastKey common.Hash
|
2021-01-25 09:17:05 +03:00
|
|
|
)
|
2021-04-27 17:19:59 +03:00
|
|
|
if len(keys) > 0 {
|
|
|
|
lastKey = keys[len(keys)-1]
|
|
|
|
}
|
|
|
|
// If the number of slots remaining is low, decrease the
|
|
|
|
// number of chunks. Somewhere on the order of 10-15K slots
|
|
|
|
// fit into a packet of 500KB. A key/slot pair is maximum 64
|
|
|
|
// bytes, so pessimistically maxRequestSize/64 = 8K.
|
|
|
|
//
|
|
|
|
// Chunk so that at least 2 packets are needed to fill a task.
|
|
|
|
if estimate, err := estimateRemainingSlots(len(keys), lastKey); err == nil {
|
|
|
|
if n := estimate / (2 * (maxRequestSize / 64)); n+1 < chunks {
|
|
|
|
chunks = n + 1
|
2020-12-14 12:27:15 +03:00
|
|
|
}
|
2021-04-27 17:19:59 +03:00
|
|
|
log.Debug("Chunked large contract", "initiators", len(keys), "tail", lastKey, "remaining", estimate, "chunks", chunks)
|
|
|
|
} else {
|
|
|
|
log.Debug("Chunked large contract", "initiators", len(keys), "tail", lastKey, "chunks", chunks)
|
|
|
|
}
|
|
|
|
r := newHashRange(lastKey, chunks)
|
2023-10-23 18:31:56 +03:00
|
|
|
if chunks == 1 {
|
|
|
|
smallStorageGauge.Inc(1)
|
|
|
|
} else {
|
|
|
|
largeStorageGauge.Inc(1)
|
|
|
|
}
|
2021-04-27 17:19:59 +03:00
|
|
|
// Our first task is the one that was just filled by this response.
|
2021-04-28 23:09:15 +03:00
|
|
|
batch := ethdb.HookedBatch{
|
|
|
|
Batch: s.db.NewBatch(),
|
|
|
|
OnPut: func(key []byte, value []byte) {
|
|
|
|
s.storageBytes += common.StorageSize(len(key) + len(value))
|
|
|
|
},
|
|
|
|
}
|
2023-10-11 07:12:45 +03:00
|
|
|
owner := account // local assignment for stacktrie writer closure
|
2023-10-17 15:09:25 +03:00
|
|
|
options := trie.NewStackTrieOptions()
|
|
|
|
options = options.WithWriter(func(path []byte, hash common.Hash, blob []byte) {
|
|
|
|
rawdb.WriteTrieNode(batch, owner, path, hash, blob, s.scheme)
|
|
|
|
})
|
2023-10-23 18:31:56 +03:00
|
|
|
if s.scheme == rawdb.PathScheme {
|
|
|
|
options = options.WithCleaner(func(path []byte) {
|
|
|
|
s.cleanPath(batch, owner, path)
|
|
|
|
})
|
|
|
|
// Keep the left boundary as it's the first range.
|
|
|
|
// Skip the right boundary if it's not the last range.
|
|
|
|
options = options.WithSkipBoundary(false, r.End() != common.MaxHash, boundaryStorageNodesGauge)
|
|
|
|
}
|
2021-04-27 17:19:59 +03:00
|
|
|
tasks = append(tasks, &storageTask{
|
|
|
|
Next: common.Hash{},
|
|
|
|
Last: r.End(),
|
|
|
|
root: acc.Root,
|
|
|
|
genBatch: batch,
|
2023-10-17 15:09:25 +03:00
|
|
|
genTrie: trie.NewStackTrie(options),
|
2021-04-27 17:19:59 +03:00
|
|
|
})
|
|
|
|
for r.Next() {
|
2021-04-28 23:09:15 +03:00
|
|
|
batch := ethdb.HookedBatch{
|
|
|
|
Batch: s.db.NewBatch(),
|
|
|
|
OnPut: func(key []byte, value []byte) {
|
|
|
|
s.storageBytes += common.StorageSize(len(key) + len(value))
|
|
|
|
},
|
|
|
|
}
|
2023-10-17 15:09:25 +03:00
|
|
|
options := trie.NewStackTrieOptions()
|
|
|
|
options = options.WithWriter(func(path []byte, hash common.Hash, blob []byte) {
|
|
|
|
rawdb.WriteTrieNode(batch, owner, path, hash, blob, s.scheme)
|
|
|
|
})
|
2023-10-23 18:31:56 +03:00
|
|
|
if s.scheme == rawdb.PathScheme {
|
|
|
|
// Configure the dangling node cleaner and also filter out boundary nodes
|
|
|
|
// only in the context of the path scheme. Deletion is forbidden in the
|
|
|
|
// hash scheme, as it can disrupt state completeness.
|
|
|
|
options = options.WithCleaner(func(path []byte) {
|
|
|
|
s.cleanPath(batch, owner, path)
|
|
|
|
})
|
|
|
|
// Skip the left boundary as it's not the first range
|
|
|
|
// Skip the right boundary if it's not the last range.
|
|
|
|
options = options.WithSkipBoundary(true, r.End() != common.MaxHash, boundaryStorageNodesGauge)
|
|
|
|
}
|
2021-01-25 09:17:05 +03:00
|
|
|
tasks = append(tasks, &storageTask{
|
2021-04-27 17:19:59 +03:00
|
|
|
Next: r.Start(),
|
|
|
|
Last: r.End(),
|
|
|
|
root: acc.Root,
|
|
|
|
genBatch: batch,
|
2023-10-17 15:09:25 +03:00
|
|
|
genTrie: trie.NewStackTrie(options),
|
2021-01-25 09:17:05 +03:00
|
|
|
})
|
2021-04-27 17:19:59 +03:00
|
|
|
}
|
|
|
|
for _, task := range tasks {
|
|
|
|
log.Debug("Created storage sync task", "account", account, "root", acc.Root, "from", task.Next, "last", task.Last)
|
2020-12-14 12:27:15 +03:00
|
|
|
}
|
2021-01-25 09:17:05 +03:00
|
|
|
res.mainTask.SubTasks[account] = tasks
|
|
|
|
|
|
|
|
// Since we've just created the sub-tasks, this response
|
|
|
|
// is surely for the first one (zero origin)
|
|
|
|
res.subTask = tasks[0]
|
2020-12-14 12:27:15 +03:00
|
|
|
}
|
2021-01-25 09:17:05 +03:00
|
|
|
}
|
|
|
|
// If we're in large contract delivery mode, forward the subtask
|
|
|
|
if res.subTask != nil {
|
|
|
|
// Ensure the response doesn't overflow into the subsequent task
|
|
|
|
last := res.subTask.Last.Big()
|
2021-04-27 17:19:59 +03:00
|
|
|
// Find the first overflowing key. While at it, mark res as complete
|
|
|
|
// if we find the range to include or pass the 'last'
|
|
|
|
index := sort.Search(len(res.hashes[i]), func(k int) bool {
|
|
|
|
cmp := res.hashes[i][k].Big().Cmp(last)
|
|
|
|
if cmp >= 0 {
|
2021-03-24 17:33:34 +03:00
|
|
|
res.cont = false
|
2020-12-14 12:27:15 +03:00
|
|
|
}
|
2021-04-27 17:19:59 +03:00
|
|
|
return cmp > 0
|
|
|
|
})
|
|
|
|
if index >= 0 {
|
|
|
|
// cut off excess
|
|
|
|
res.hashes[i] = res.hashes[i][:index]
|
|
|
|
res.slots[i] = res.slots[i][:index]
|
2020-12-14 12:27:15 +03:00
|
|
|
}
|
2021-01-25 09:17:05 +03:00
|
|
|
// Forward the relevant storage chunk (even if created just now)
|
|
|
|
if res.cont {
|
2021-04-27 17:19:59 +03:00
|
|
|
res.subTask.Next = incHash(res.hashes[i][len(res.hashes[i])-1])
|
2021-01-25 09:17:05 +03:00
|
|
|
} else {
|
|
|
|
res.subTask.done = true
|
|
|
|
}
|
2020-12-14 12:27:15 +03:00
|
|
|
}
|
|
|
|
}
|
2021-04-28 23:09:15 +03:00
|
|
|
// Iterate over all the complete contracts, reconstruct the trie nodes and
|
|
|
|
// push them to disk. If the contract is chunked, the trie nodes will be
|
|
|
|
// reconstructed later.
|
2020-12-14 12:27:15 +03:00
|
|
|
slots += len(res.hashes[i])
|
|
|
|
|
2021-04-27 17:19:59 +03:00
|
|
|
if i < len(res.hashes)-1 || res.subTask == nil {
|
2023-10-11 07:12:45 +03:00
|
|
|
// no need to make local reassignment of account: this closure does not outlive the loop
|
2023-10-17 15:09:25 +03:00
|
|
|
options := trie.NewStackTrieOptions()
|
|
|
|
options = options.WithWriter(func(path []byte, hash common.Hash, blob []byte) {
|
|
|
|
rawdb.WriteTrieNode(batch, account, path, hash, blob, s.scheme)
|
2023-10-11 07:12:45 +03:00
|
|
|
})
|
2023-10-23 18:31:56 +03:00
|
|
|
if s.scheme == rawdb.PathScheme {
|
|
|
|
// Configure the dangling node cleaner only in the context of the
|
|
|
|
// path scheme. Deletion is forbidden in the hash scheme, as it can
|
|
|
|
// disrupt state completeness.
|
|
|
|
//
|
|
|
|
// Notably, boundary nodes can be also kept because the whole storage
|
|
|
|
// trie is complete.
|
|
|
|
options = options.WithCleaner(func(path []byte) {
|
|
|
|
s.cleanPath(batch, account, path)
|
|
|
|
})
|
|
|
|
}
|
2023-10-17 15:09:25 +03:00
|
|
|
tr := trie.NewStackTrie(options)
|
2021-04-28 23:09:15 +03:00
|
|
|
for j := 0; j < len(res.hashes[i]); j++ {
|
|
|
|
tr.Update(res.hashes[i][j][:], res.slots[i][j])
|
2021-04-27 17:19:59 +03:00
|
|
|
}
|
2021-04-28 23:09:15 +03:00
|
|
|
tr.Commit()
|
2020-12-14 12:27:15 +03:00
|
|
|
}
|
2022-08-19 09:00:21 +03:00
|
|
|
// Persist the received storage segments. These flat state maybe
|
core, eth: faster snapshot generation (#22504)
* eth/protocols: persist received state segments
* core: initial implementation
* core/state/snapshot: add tests
* core, eth: updates
* eth/protocols/snapshot: count flat state size
* core/state: add metrics
* core/state/snapshot: skip unnecessary deletion
* core/state/snapshot: rename
* core/state/snapshot: use the global batch
* core/state/snapshot: add logs and fix wiping
* core/state/snapshot: fix
* core/state/snapshot: save generation progress even if the batch is empty
* core/state/snapshot: fixes
* core/state/snapshot: fix initial account range length
* core/state/snapshot: fix initial account range
* eth/protocols/snap: store flat states during the healing
* eth/protocols/snap: print logs
* core/state/snapshot: refactor (#4)
* core/state/snapshot: refactor
* core/state/snapshot: tiny fix and polish
Co-authored-by: rjl493456442 <garyrong0905@gmail.com>
* core, eth: fixes
* core, eth: fix healing writer
* core, trie, eth: fix paths
* eth/protocols/snap: fix encoding
* eth, core: add debug log
* core/state/generate: release iterator asap (#5)
core/state/snapshot: less copy
core/state/snapshot: revert split loop
core/state/snapshot: handle storage becoming empty, improve test robustness
core/state: test modified codehash
core/state/snapshot: polish
* core/state/snapshot: optimize stats counter
* core, eth: add metric
* core/state/snapshot: update comments
* core/state/snapshot: improve tests
* core/state/snapshot: replace secure trie with standard trie
* core/state/snapshot: wrap return as the struct
* core/state/snapshot: skip wiping correct states
* core/state/snapshot: updates
* core/state/snapshot: fixes
* core/state/snapshot: fix panic due to reference flaw in closure
* core/state/snapshot: fix errors in state generation logic + fix log output
* core/state/snapshot: remove an error case
* core/state/snapshot: fix condition-check for exhausted snap state
* core/state/snapshot: use stackTrie for small tries
* core/state/snapshot: don't resolve small storage tries in vain
* core/state/snapshot: properly clean up storage of deleted accounts
* core/state/snapshot: avoid RLP-encoding in some cases + minor nitpicks
* core/state/snapshot: fix error (+testcase)
* core/state/snapshot: clean up tests a bit
* core/state/snapshot: work in progress on better tests
* core/state/snapshot: polish code
* core/state/snapshot: fix trie iteration abortion trigger
* core/state/snapshot: fixes flaws
* core/state/snapshot: remove panic
* core/state/snapshot: fix abort
* core/state/snapshot: more tests (plus failing testcase)
* core/state/snapshot: more testcases + fix for failing test
* core/state/snapshot: testcase for malformed data
* core/state/snapshot: some test nitpicks
* core/state/snapshot: improvements to logging
* core/state/snapshot: testcase to demo error in abortion
* core/state/snapshot: fix abortion
* cmd/geth: make verify-state report the root
* trie: fix failing test
* core/state/snapshot: add timer metrics
* core/state/snapshot: fix metrics
* core/state/snapshot: udpate tests
* eth/protocols/snap: write snapshot account even if code or state is needed
* core/state/snapshot: fix diskmore check
* core/state/snapshot: review fixes
* core/state/snapshot: improve error message
* cmd/geth: rename 'error' to 'err' in logs
* core/state/snapshot: fix some review concerns
* core/state/snapshot, eth/protocols/snap: clear snapshot marker when starting/resuming snap sync
* core: add error log
* core/state/snapshot: use proper timers for metrics collection
* core/state/snapshot: address some review concerns
* eth/protocols/snap: improved log message
* eth/protocols/snap: fix heal logs to condense infos
* core/state/snapshot: wait for generator termination before restarting
* core/state/snapshot: revert timers to counters to track total time
Co-authored-by: Martin Holst Swende <martin@swende.se>
Co-authored-by: Péter Szilágyi <peterke@gmail.com>
2021-04-14 23:23:11 +03:00
|
|
|
// outdated during the sync, but it can be fixed later during the
|
|
|
|
// snapshot generation.
|
|
|
|
for j := 0; j < len(res.hashes[i]); j++ {
|
|
|
|
rawdb.WriteStorageSnapshot(batch, account, res.hashes[i][j], res.slots[i][j])
|
2021-04-27 17:19:59 +03:00
|
|
|
|
|
|
|
// If we're storing large contracts, generate the trie nodes
|
|
|
|
// on the fly to not trash the gluing points
|
|
|
|
if i == len(res.hashes)-1 && res.subTask != nil {
|
|
|
|
res.subTask.genTrie.Update(res.hashes[i][j][:], res.slots[i][j])
|
|
|
|
}
|
core, eth: faster snapshot generation (#22504)
* eth/protocols: persist received state segments
* core: initial implementation
* core/state/snapshot: add tests
* core, eth: updates
* eth/protocols/snapshot: count flat state size
* core/state: add metrics
* core/state/snapshot: skip unnecessary deletion
* core/state/snapshot: rename
* core/state/snapshot: use the global batch
* core/state/snapshot: add logs and fix wiping
* core/state/snapshot: fix
* core/state/snapshot: save generation progress even if the batch is empty
* core/state/snapshot: fixes
* core/state/snapshot: fix initial account range length
* core/state/snapshot: fix initial account range
* eth/protocols/snap: store flat states during the healing
* eth/protocols/snap: print logs
* core/state/snapshot: refactor (#4)
* core/state/snapshot: refactor
* core/state/snapshot: tiny fix and polish
Co-authored-by: rjl493456442 <garyrong0905@gmail.com>
* core, eth: fixes
* core, eth: fix healing writer
* core, trie, eth: fix paths
* eth/protocols/snap: fix encoding
* eth, core: add debug log
* core/state/generate: release iterator asap (#5)
core/state/snapshot: less copy
core/state/snapshot: revert split loop
core/state/snapshot: handle storage becoming empty, improve test robustness
core/state: test modified codehash
core/state/snapshot: polish
* core/state/snapshot: optimize stats counter
* core, eth: add metric
* core/state/snapshot: update comments
* core/state/snapshot: improve tests
* core/state/snapshot: replace secure trie with standard trie
* core/state/snapshot: wrap return as the struct
* core/state/snapshot: skip wiping correct states
* core/state/snapshot: updates
* core/state/snapshot: fixes
* core/state/snapshot: fix panic due to reference flaw in closure
* core/state/snapshot: fix errors in state generation logic + fix log output
* core/state/snapshot: remove an error case
* core/state/snapshot: fix condition-check for exhausted snap state
* core/state/snapshot: use stackTrie for small tries
* core/state/snapshot: don't resolve small storage tries in vain
* core/state/snapshot: properly clean up storage of deleted accounts
* core/state/snapshot: avoid RLP-encoding in some cases + minor nitpicks
* core/state/snapshot: fix error (+testcase)
* core/state/snapshot: clean up tests a bit
* core/state/snapshot: work in progress on better tests
* core/state/snapshot: polish code
* core/state/snapshot: fix trie iteration abortion trigger
* core/state/snapshot: fixes flaws
* core/state/snapshot: remove panic
* core/state/snapshot: fix abort
* core/state/snapshot: more tests (plus failing testcase)
* core/state/snapshot: more testcases + fix for failing test
* core/state/snapshot: testcase for malformed data
* core/state/snapshot: some test nitpicks
* core/state/snapshot: improvements to logging
* core/state/snapshot: testcase to demo error in abortion
* core/state/snapshot: fix abortion
* cmd/geth: make verify-state report the root
* trie: fix failing test
* core/state/snapshot: add timer metrics
* core/state/snapshot: fix metrics
* core/state/snapshot: udpate tests
* eth/protocols/snap: write snapshot account even if code or state is needed
* core/state/snapshot: fix diskmore check
* core/state/snapshot: review fixes
* core/state/snapshot: improve error message
* cmd/geth: rename 'error' to 'err' in logs
* core/state/snapshot: fix some review concerns
* core/state/snapshot, eth/protocols/snap: clear snapshot marker when starting/resuming snap sync
* core: add error log
* core/state/snapshot: use proper timers for metrics collection
* core/state/snapshot: address some review concerns
* eth/protocols/snap: improved log message
* eth/protocols/snap: fix heal logs to condense infos
* core/state/snapshot: wait for generator termination before restarting
* core/state/snapshot: revert timers to counters to track total time
Co-authored-by: Martin Holst Swende <martin@swende.se>
Co-authored-by: Péter Szilágyi <peterke@gmail.com>
2021-04-14 23:23:11 +03:00
|
|
|
}
|
2020-12-14 12:27:15 +03:00
|
|
|
}
|
2021-04-27 17:19:59 +03:00
|
|
|
// Large contracts could have generated new trie nodes, flush them to disk
|
|
|
|
if res.subTask != nil {
|
|
|
|
if res.subTask.done {
|
2023-10-17 15:09:25 +03:00
|
|
|
root := res.subTask.genTrie.Commit()
|
2023-10-23 18:31:56 +03:00
|
|
|
if err := res.subTask.genBatch.Write(); err != nil {
|
|
|
|
log.Error("Failed to persist stack slots", "err", err)
|
|
|
|
}
|
|
|
|
res.subTask.genBatch.Reset()
|
|
|
|
|
|
|
|
// If the chunk's root is an overflown but full delivery,
|
|
|
|
// clear the heal request.
|
|
|
|
accountHash := res.accounts[len(res.accounts)-1]
|
|
|
|
if root == res.subTask.root && rawdb.HasStorageTrieNode(s.db, accountHash, nil, root) {
|
2021-04-27 17:19:59 +03:00
|
|
|
for i, account := range res.mainTask.res.hashes {
|
2023-10-23 18:31:56 +03:00
|
|
|
if account == accountHash {
|
2021-04-27 17:19:59 +03:00
|
|
|
res.mainTask.needHeal[i] = false
|
2023-10-23 18:31:56 +03:00
|
|
|
skipStorageHealingGauge.Inc(1)
|
2021-04-27 17:19:59 +03:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2023-10-23 18:31:56 +03:00
|
|
|
if res.subTask.genBatch.ValueSize() > ethdb.IdealBatchSize {
|
2021-04-27 17:19:59 +03:00
|
|
|
if err := res.subTask.genBatch.Write(); err != nil {
|
|
|
|
log.Error("Failed to persist stack slots", "err", err)
|
|
|
|
}
|
|
|
|
res.subTask.genBatch.Reset()
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// Flush anything written just now and update the stats
|
2020-12-14 12:27:15 +03:00
|
|
|
if err := batch.Write(); err != nil {
|
|
|
|
log.Crit("Failed to persist storage slots", "err", err)
|
|
|
|
}
|
|
|
|
s.storageSynced += uint64(slots)
|
|
|
|
|
2021-04-28 23:09:15 +03:00
|
|
|
log.Debug("Persisted set of storage slots", "accounts", len(res.hashes), "slots", slots, "bytes", s.storageBytes-oldStorageBytes)
|
2020-12-14 12:27:15 +03:00
|
|
|
|
|
|
|
// If this delivery completed the last pending task, forward the account task
|
|
|
|
// to the next chunk
|
|
|
|
if res.mainTask.pend == 0 {
|
|
|
|
s.forwardAccountTask(res.mainTask)
|
|
|
|
return
|
|
|
|
}
|
|
|
|
// Some accounts are still incomplete, leave as is for the storage and contract
|
|
|
|
// task assigners to pick up and fill.
|
|
|
|
}
|
|
|
|
|
|
|
|
// processTrienodeHealResponse integrates an already validated trienode response
|
|
|
|
// into the healer tasks.
|
|
|
|
func (s *Syncer) processTrienodeHealResponse(res *trienodeHealResponse) {
|
2022-09-09 11:42:57 +03:00
|
|
|
var (
|
|
|
|
start = time.Now()
|
|
|
|
fills int
|
|
|
|
)
|
2020-12-14 12:27:15 +03:00
|
|
|
for i, hash := range res.hashes {
|
|
|
|
node := res.nodes[i]
|
|
|
|
|
|
|
|
// If the trie node was not delivered, reschedule it
|
|
|
|
if node == nil {
|
2022-07-15 14:55:51 +03:00
|
|
|
res.task.trieTasks[res.paths[i]] = res.hashes[i]
|
2020-12-14 12:27:15 +03:00
|
|
|
continue
|
|
|
|
}
|
2022-09-09 11:42:57 +03:00
|
|
|
fills++
|
|
|
|
|
2020-12-14 12:27:15 +03:00
|
|
|
// Push the trie node into the state syncer
|
|
|
|
s.trienodeHealSynced++
|
|
|
|
s.trienodeHealBytes += common.StorageSize(len(node))
|
|
|
|
|
2022-07-15 14:55:51 +03:00
|
|
|
err := s.healer.scheduler.ProcessNode(trie.NodeSyncResult{Path: res.paths[i], Data: node})
|
2020-12-14 12:27:15 +03:00
|
|
|
switch err {
|
|
|
|
case nil:
|
|
|
|
case trie.ErrAlreadyProcessed:
|
|
|
|
s.trienodeHealDups++
|
|
|
|
case trie.ErrNotRequested:
|
|
|
|
s.trienodeHealNops++
|
|
|
|
default:
|
|
|
|
log.Error("Invalid trienode processed", "hash", hash, "err", err)
|
|
|
|
}
|
|
|
|
}
|
2022-09-28 09:08:18 +03:00
|
|
|
s.commitHealer(false)
|
2022-09-09 11:42:57 +03:00
|
|
|
|
|
|
|
// Calculate the processing rate of one filled trie node
|
|
|
|
rate := float64(fills) / (float64(time.Since(start)) / float64(time.Second))
|
|
|
|
|
|
|
|
// Update the currently measured trienode queueing and processing throughput.
|
|
|
|
//
|
|
|
|
// The processing rate needs to be updated uniformly independent if we've
|
|
|
|
// processed 1x100 trie nodes or 100x1 to keep the rate consistent even in
|
|
|
|
// the face of varying network packets. As such, we cannot just measure the
|
|
|
|
// time it took to process N trie nodes and update once, we need one update
|
|
|
|
// per trie node.
|
|
|
|
//
|
|
|
|
// Naively, that would be:
|
|
|
|
//
|
|
|
|
// for i:=0; i<fills; i++ {
|
|
|
|
// healRate = (1-measurementImpact)*oldRate + measurementImpact*newRate
|
|
|
|
// }
|
|
|
|
//
|
|
|
|
// Essentially, a recursive expansion of HR = (1-MI)*HR + MI*NR.
|
|
|
|
//
|
|
|
|
// We can expand that formula for the Nth item as:
|
|
|
|
// HR(N) = (1-MI)^N*OR + (1-MI)^(N-1)*MI*NR + (1-MI)^(N-2)*MI*NR + ... + (1-MI)^0*MI*NR
|
|
|
|
//
|
|
|
|
// The above is a geometric sequence that can be summed to:
|
|
|
|
// HR(N) = (1-MI)^N*(OR-NR) + NR
|
|
|
|
s.trienodeHealRate = gomath.Pow(1-trienodeHealRateMeasurementImpact, float64(fills))*(s.trienodeHealRate-rate) + rate
|
|
|
|
|
2023-04-25 13:06:50 +03:00
|
|
|
pending := s.trienodeHealPend.Load()
|
2022-09-09 11:42:57 +03:00
|
|
|
if time.Since(s.trienodeHealThrottled) > time.Second {
|
|
|
|
// Periodically adjust the trie node throttler
|
|
|
|
if float64(pending) > 2*s.trienodeHealRate {
|
|
|
|
s.trienodeHealThrottle *= trienodeHealThrottleIncrease
|
|
|
|
} else {
|
|
|
|
s.trienodeHealThrottle /= trienodeHealThrottleDecrease
|
|
|
|
}
|
|
|
|
if s.trienodeHealThrottle > maxTrienodeHealThrottle {
|
|
|
|
s.trienodeHealThrottle = maxTrienodeHealThrottle
|
|
|
|
} else if s.trienodeHealThrottle < minTrienodeHealThrottle {
|
|
|
|
s.trienodeHealThrottle = minTrienodeHealThrottle
|
|
|
|
}
|
|
|
|
s.trienodeHealThrottled = time.Now()
|
|
|
|
|
|
|
|
log.Debug("Updated trie node heal throttler", "rate", s.trienodeHealRate, "pending", pending, "throttle", s.trienodeHealThrottle)
|
|
|
|
}
|
2020-12-14 12:27:15 +03:00
|
|
|
}
|
|
|
|
|
2022-09-28 09:08:18 +03:00
|
|
|
func (s *Syncer) commitHealer(force bool) {
|
|
|
|
if !force && s.healer.scheduler.MemSize() < ethdb.IdealBatchSize {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
batch := s.db.NewBatch()
|
|
|
|
if err := s.healer.scheduler.Commit(batch); err != nil {
|
|
|
|
log.Error("Failed to commit healing data", "err", err)
|
|
|
|
}
|
|
|
|
if err := batch.Write(); err != nil {
|
|
|
|
log.Crit("Failed to persist healing data", "err", err)
|
|
|
|
}
|
|
|
|
log.Debug("Persisted set of healing data", "type", "trienodes", "bytes", common.StorageSize(batch.ValueSize()))
|
|
|
|
}
|
|
|
|
|
2020-12-14 12:27:15 +03:00
|
|
|
// processBytecodeHealResponse integrates an already validated bytecode response
|
|
|
|
// into the healer tasks.
|
|
|
|
func (s *Syncer) processBytecodeHealResponse(res *bytecodeHealResponse) {
|
|
|
|
for i, hash := range res.hashes {
|
|
|
|
node := res.codes[i]
|
|
|
|
|
|
|
|
// If the trie node was not delivered, reschedule it
|
|
|
|
if node == nil {
|
|
|
|
res.task.codeTasks[hash] = struct{}{}
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
// Push the trie node into the state syncer
|
|
|
|
s.bytecodeHealSynced++
|
|
|
|
s.bytecodeHealBytes += common.StorageSize(len(node))
|
|
|
|
|
2022-07-15 14:55:51 +03:00
|
|
|
err := s.healer.scheduler.ProcessCode(trie.CodeSyncResult{Hash: hash, Data: node})
|
2020-12-14 12:27:15 +03:00
|
|
|
switch err {
|
|
|
|
case nil:
|
|
|
|
case trie.ErrAlreadyProcessed:
|
|
|
|
s.bytecodeHealDups++
|
|
|
|
case trie.ErrNotRequested:
|
|
|
|
s.bytecodeHealNops++
|
|
|
|
default:
|
|
|
|
log.Error("Invalid bytecode processed", "hash", hash, "err", err)
|
|
|
|
}
|
|
|
|
}
|
2022-09-28 09:08:18 +03:00
|
|
|
s.commitHealer(false)
|
2020-12-14 12:27:15 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// forwardAccountTask takes a filled account task and persists anything available
|
|
|
|
// into the database, after which it forwards the next account marker so that the
|
|
|
|
// task's next chunk may be filled.
|
|
|
|
func (s *Syncer) forwardAccountTask(task *accountTask) {
|
|
|
|
// Remove any pending delivery
|
|
|
|
res := task.res
|
|
|
|
if res == nil {
|
|
|
|
return // nothing to forward
|
|
|
|
}
|
|
|
|
task.res = nil
|
|
|
|
|
2022-08-19 09:00:21 +03:00
|
|
|
// Persist the received account segments. These flat state maybe
|
2021-04-27 17:19:59 +03:00
|
|
|
// outdated during the sync, but it can be fixed later during the
|
|
|
|
// snapshot generation.
|
2021-04-28 23:09:15 +03:00
|
|
|
oldAccountBytes := s.accountBytes
|
|
|
|
|
|
|
|
batch := ethdb.HookedBatch{
|
|
|
|
Batch: s.db.NewBatch(),
|
|
|
|
OnPut: func(key []byte, value []byte) {
|
|
|
|
s.accountBytes += common.StorageSize(len(key) + len(value))
|
|
|
|
},
|
|
|
|
}
|
2021-04-27 17:19:59 +03:00
|
|
|
for i, hash := range res.hashes {
|
2020-12-14 12:27:15 +03:00
|
|
|
if task.needCode[i] || task.needState[i] {
|
|
|
|
break
|
|
|
|
}
|
2023-06-06 11:17:39 +03:00
|
|
|
slim := types.SlimAccountRLP(*res.accounts[i])
|
2021-04-27 17:19:59 +03:00
|
|
|
rawdb.WriteAccountSnapshot(batch, hash, slim)
|
2020-12-14 12:27:15 +03:00
|
|
|
|
2021-04-27 17:19:59 +03:00
|
|
|
// If the task is complete, drop it into the stack trie to generate
|
|
|
|
// account trie nodes for it
|
|
|
|
if !task.needHeal[i] {
|
2023-06-06 11:17:39 +03:00
|
|
|
full, err := types.FullAccountRLP(slim) // TODO(karalabe): Slim parsing can be omitted
|
2021-04-27 17:19:59 +03:00
|
|
|
if err != nil {
|
|
|
|
panic(err) // Really shouldn't ever happen
|
|
|
|
}
|
|
|
|
task.genTrie.Update(hash[:], full)
|
2020-12-14 12:27:15 +03:00
|
|
|
}
|
core, eth: faster snapshot generation (#22504)
* eth/protocols: persist received state segments
* core: initial implementation
* core/state/snapshot: add tests
* core, eth: updates
* eth/protocols/snapshot: count flat state size
* core/state: add metrics
* core/state/snapshot: skip unnecessary deletion
* core/state/snapshot: rename
* core/state/snapshot: use the global batch
* core/state/snapshot: add logs and fix wiping
* core/state/snapshot: fix
* core/state/snapshot: save generation progress even if the batch is empty
* core/state/snapshot: fixes
* core/state/snapshot: fix initial account range length
* core/state/snapshot: fix initial account range
* eth/protocols/snap: store flat states during the healing
* eth/protocols/snap: print logs
* core/state/snapshot: refactor (#4)
* core/state/snapshot: refactor
* core/state/snapshot: tiny fix and polish
Co-authored-by: rjl493456442 <garyrong0905@gmail.com>
* core, eth: fixes
* core, eth: fix healing writer
* core, trie, eth: fix paths
* eth/protocols/snap: fix encoding
* eth, core: add debug log
* core/state/generate: release iterator asap (#5)
core/state/snapshot: less copy
core/state/snapshot: revert split loop
core/state/snapshot: handle storage becoming empty, improve test robustness
core/state: test modified codehash
core/state/snapshot: polish
* core/state/snapshot: optimize stats counter
* core, eth: add metric
* core/state/snapshot: update comments
* core/state/snapshot: improve tests
* core/state/snapshot: replace secure trie with standard trie
* core/state/snapshot: wrap return as the struct
* core/state/snapshot: skip wiping correct states
* core/state/snapshot: updates
* core/state/snapshot: fixes
* core/state/snapshot: fix panic due to reference flaw in closure
* core/state/snapshot: fix errors in state generation logic + fix log output
* core/state/snapshot: remove an error case
* core/state/snapshot: fix condition-check for exhausted snap state
* core/state/snapshot: use stackTrie for small tries
* core/state/snapshot: don't resolve small storage tries in vain
* core/state/snapshot: properly clean up storage of deleted accounts
* core/state/snapshot: avoid RLP-encoding in some cases + minor nitpicks
* core/state/snapshot: fix error (+testcase)
* core/state/snapshot: clean up tests a bit
* core/state/snapshot: work in progress on better tests
* core/state/snapshot: polish code
* core/state/snapshot: fix trie iteration abortion trigger
* core/state/snapshot: fixes flaws
* core/state/snapshot: remove panic
* core/state/snapshot: fix abort
* core/state/snapshot: more tests (plus failing testcase)
* core/state/snapshot: more testcases + fix for failing test
* core/state/snapshot: testcase for malformed data
* core/state/snapshot: some test nitpicks
* core/state/snapshot: improvements to logging
* core/state/snapshot: testcase to demo error in abortion
* core/state/snapshot: fix abortion
* cmd/geth: make verify-state report the root
* trie: fix failing test
* core/state/snapshot: add timer metrics
* core/state/snapshot: fix metrics
* core/state/snapshot: udpate tests
* eth/protocols/snap: write snapshot account even if code or state is needed
* core/state/snapshot: fix diskmore check
* core/state/snapshot: review fixes
* core/state/snapshot: improve error message
* cmd/geth: rename 'error' to 'err' in logs
* core/state/snapshot: fix some review concerns
* core/state/snapshot, eth/protocols/snap: clear snapshot marker when starting/resuming snap sync
* core: add error log
* core/state/snapshot: use proper timers for metrics collection
* core/state/snapshot: address some review concerns
* eth/protocols/snap: improved log message
* eth/protocols/snap: fix heal logs to condense infos
* core/state/snapshot: wait for generator termination before restarting
* core/state/snapshot: revert timers to counters to track total time
Co-authored-by: Martin Holst Swende <martin@swende.se>
Co-authored-by: Péter Szilágyi <peterke@gmail.com>
2021-04-14 23:23:11 +03:00
|
|
|
}
|
2021-04-27 17:19:59 +03:00
|
|
|
// Flush anything written just now and update the stats
|
2020-12-14 12:27:15 +03:00
|
|
|
if err := batch.Write(); err != nil {
|
|
|
|
log.Crit("Failed to persist accounts", "err", err)
|
|
|
|
}
|
|
|
|
s.accountSynced += uint64(len(res.accounts))
|
|
|
|
|
|
|
|
// Task filling persisted, push it the chunk marker forward to the first
|
|
|
|
// account still missing data.
|
|
|
|
for i, hash := range res.hashes {
|
|
|
|
if task.needCode[i] || task.needState[i] {
|
|
|
|
return
|
|
|
|
}
|
2021-04-27 17:19:59 +03:00
|
|
|
task.Next = incHash(hash)
|
2020-12-14 12:27:15 +03:00
|
|
|
}
|
|
|
|
// All accounts marked as complete, track if the entire task is done
|
|
|
|
task.done = !res.cont
|
2021-04-27 17:19:59 +03:00
|
|
|
|
|
|
|
// Stack trie could have generated trie nodes, push them to disk (we need to
|
|
|
|
// flush after finalizing task.done. It's fine even if we crash and lose this
|
|
|
|
// write as it will only cause more data to be downloaded during heal.
|
|
|
|
if task.done {
|
2023-10-17 15:09:25 +03:00
|
|
|
task.genTrie.Commit()
|
2021-04-27 17:19:59 +03:00
|
|
|
}
|
2021-04-28 23:09:15 +03:00
|
|
|
if task.genBatch.ValueSize() > ethdb.IdealBatchSize || task.done {
|
2021-04-27 17:19:59 +03:00
|
|
|
if err := task.genBatch.Write(); err != nil {
|
|
|
|
log.Error("Failed to persist stack account", "err", err)
|
|
|
|
}
|
|
|
|
task.genBatch.Reset()
|
|
|
|
}
|
2021-04-28 23:09:15 +03:00
|
|
|
log.Debug("Persisted range of accounts", "accounts", len(res.accounts), "bytes", s.accountBytes-oldAccountBytes)
|
2020-12-14 12:27:15 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// OnAccounts is a callback method to invoke when a range of accounts are
|
|
|
|
// received from a remote peer.
|
2021-01-25 09:17:05 +03:00
|
|
|
func (s *Syncer) OnAccounts(peer SyncPeer, id uint64, hashes []common.Hash, accounts [][]byte, proof [][]byte) error {
|
2020-12-14 12:27:15 +03:00
|
|
|
size := common.StorageSize(len(hashes) * common.HashLength)
|
|
|
|
for _, account := range accounts {
|
|
|
|
size += common.StorageSize(len(account))
|
|
|
|
}
|
|
|
|
for _, node := range proof {
|
|
|
|
size += common.StorageSize(len(node))
|
|
|
|
}
|
2021-01-25 09:17:05 +03:00
|
|
|
logger := peer.Log().New("reqid", id)
|
2020-12-14 12:27:15 +03:00
|
|
|
logger.Trace("Delivering range of accounts", "hashes", len(hashes), "accounts", len(accounts), "proofs", len(proof), "bytes", size)
|
|
|
|
|
|
|
|
// Whether or not the response is valid, we can mark the peer as idle and
|
|
|
|
// notify the scheduler to assign a new task. If the response is invalid,
|
|
|
|
// we'll drop the peer in a bit.
|
2022-08-31 18:58:18 +03:00
|
|
|
defer func() {
|
|
|
|
s.lock.Lock()
|
|
|
|
defer s.lock.Unlock()
|
|
|
|
if _, ok := s.peers[peer.ID()]; ok {
|
|
|
|
s.accountIdlers[peer.ID()] = struct{}{}
|
|
|
|
}
|
|
|
|
select {
|
|
|
|
case s.update <- struct{}{}:
|
|
|
|
default:
|
|
|
|
}
|
|
|
|
}()
|
2020-12-14 12:27:15 +03:00
|
|
|
s.lock.Lock()
|
|
|
|
// Ensure the response is for a valid request
|
|
|
|
req, ok := s.accountReqs[id]
|
|
|
|
if !ok {
|
|
|
|
// Request stale, perhaps the peer timed out but came through in the end
|
|
|
|
logger.Warn("Unexpected account range packet")
|
|
|
|
s.lock.Unlock()
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
delete(s.accountReqs, id)
|
2021-05-19 15:09:03 +03:00
|
|
|
s.rates.Update(peer.ID(), AccountRangeMsg, time.Since(req.time), int(size))
|
2020-12-14 12:27:15 +03:00
|
|
|
|
|
|
|
// Clean up the request timeout timer, we'll see how to proceed further based
|
|
|
|
// on the actual delivered content
|
2021-01-25 09:17:05 +03:00
|
|
|
if !req.timeout.Stop() {
|
|
|
|
// The timeout is already triggered, and this request will be reverted+rescheduled
|
|
|
|
s.lock.Unlock()
|
|
|
|
return nil
|
|
|
|
}
|
2020-12-14 12:27:15 +03:00
|
|
|
// Response is valid, but check if peer is signalling that it does not have
|
|
|
|
// the requested data. For account range queries that means the state being
|
|
|
|
// retrieved was either already pruned remotely, or the peer is not yet
|
|
|
|
// synced to our head.
|
|
|
|
if len(hashes) == 0 && len(accounts) == 0 && len(proof) == 0 {
|
|
|
|
logger.Debug("Peer rejected account range request", "root", s.root)
|
2021-01-25 09:17:05 +03:00
|
|
|
s.statelessPeers[peer.ID()] = struct{}{}
|
2020-12-14 12:27:15 +03:00
|
|
|
s.lock.Unlock()
|
2021-01-07 13:58:07 +03:00
|
|
|
|
|
|
|
// Signal this request as failed, and ready for rescheduling
|
|
|
|
s.scheduleRevertAccountRequest(req)
|
2020-12-14 12:27:15 +03:00
|
|
|
return nil
|
|
|
|
}
|
|
|
|
root := s.root
|
|
|
|
s.lock.Unlock()
|
|
|
|
|
|
|
|
// Reconstruct a partial trie from the response and verify it
|
|
|
|
keys := make([][]byte, len(hashes))
|
|
|
|
for i, key := range hashes {
|
|
|
|
keys[i] = common.CopyBytes(key[:])
|
|
|
|
}
|
2023-10-10 11:30:47 +03:00
|
|
|
nodes := make(trienode.ProofList, len(proof))
|
2020-12-14 12:27:15 +03:00
|
|
|
for i, node := range proof {
|
|
|
|
nodes[i] = node
|
|
|
|
}
|
trie: make rhs-proof align with last key in range proofs (#28311)
During snap-sync, we request ranges of values: either a range of accounts or a range of storage values. For any large trie, e.g. the main account trie or a large storage trie, we cannot fetch everything at once.
Short version; we split it up and request in multiple stages. To do so, we use an origin field, to say "Give me all storage key/values where key > 0x20000000000000000". When the server fulfils this, the server provides the first key after origin, let's say 0x2e030000000000000 -- never providing the exact origin. However, the client-side needs to be able to verify that the 0x2e03.. indeed is the first one after 0x2000.., and therefore the attached proof concerns the origin, not the first key.
So, short-short version: the left-hand side of the proof relates to the origin, and is free-standing from the first leaf.
On the other hand, (pun intended), the right-hand side, there's no such 'gap' between "along what path does the proof walk" and the last provided leaf. The proof must prove the last element (unless there are no elements).
Therefore, we can simplify the semantics for trie.VerifyRangeProof by removing an argument. This doesn't make much difference in practice, but makes it so that we can remove some tests. The reason I am raising this is that the upcoming stacktrie-based verifier does not support such fancy features as standalone right-hand borders.
2023-10-13 17:05:29 +03:00
|
|
|
cont, err := trie.VerifyRangeProof(root, req.origin[:], keys, accounts, nodes.Set())
|
2020-12-14 12:27:15 +03:00
|
|
|
if err != nil {
|
|
|
|
logger.Warn("Account range failed proof", "err", err)
|
2021-01-25 09:17:05 +03:00
|
|
|
// Signal this request as failed, and ready for rescheduling
|
|
|
|
s.scheduleRevertAccountRequest(req)
|
2020-12-14 12:27:15 +03:00
|
|
|
return err
|
|
|
|
}
|
2021-09-28 11:48:07 +03:00
|
|
|
accs := make([]*types.StateAccount, len(accounts))
|
2020-12-14 12:27:15 +03:00
|
|
|
for i, account := range accounts {
|
2021-09-28 11:48:07 +03:00
|
|
|
acc := new(types.StateAccount)
|
2020-12-14 12:27:15 +03:00
|
|
|
if err := rlp.DecodeBytes(account, acc); err != nil {
|
|
|
|
panic(err) // We created these blobs, we must be able to decode them
|
|
|
|
}
|
|
|
|
accs[i] = acc
|
|
|
|
}
|
|
|
|
response := &accountResponse{
|
|
|
|
task: req.task,
|
|
|
|
hashes: hashes,
|
|
|
|
accounts: accs,
|
|
|
|
cont: cont,
|
|
|
|
}
|
|
|
|
select {
|
2021-04-15 21:01:16 +03:00
|
|
|
case req.deliver <- response:
|
2020-12-14 12:27:15 +03:00
|
|
|
case <-req.cancel:
|
|
|
|
case <-req.stale:
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// OnByteCodes is a callback method to invoke when a batch of contract
|
|
|
|
// bytes codes are received from a remote peer.
|
2021-01-25 09:17:05 +03:00
|
|
|
func (s *Syncer) OnByteCodes(peer SyncPeer, id uint64, bytecodes [][]byte) error {
|
2020-12-14 12:27:15 +03:00
|
|
|
s.lock.RLock()
|
2021-01-25 09:17:05 +03:00
|
|
|
syncing := !s.snapped
|
2020-12-14 12:27:15 +03:00
|
|
|
s.lock.RUnlock()
|
|
|
|
|
|
|
|
if syncing {
|
|
|
|
return s.onByteCodes(peer, id, bytecodes)
|
|
|
|
}
|
|
|
|
return s.onHealByteCodes(peer, id, bytecodes)
|
|
|
|
}
|
|
|
|
|
|
|
|
// onByteCodes is a callback method to invoke when a batch of contract
|
|
|
|
// bytes codes are received from a remote peer in the syncing phase.
|
2021-01-25 09:17:05 +03:00
|
|
|
func (s *Syncer) onByteCodes(peer SyncPeer, id uint64, bytecodes [][]byte) error {
|
2020-12-14 12:27:15 +03:00
|
|
|
var size common.StorageSize
|
|
|
|
for _, code := range bytecodes {
|
|
|
|
size += common.StorageSize(len(code))
|
|
|
|
}
|
2021-01-25 09:17:05 +03:00
|
|
|
logger := peer.Log().New("reqid", id)
|
2020-12-14 12:27:15 +03:00
|
|
|
logger.Trace("Delivering set of bytecodes", "bytecodes", len(bytecodes), "bytes", size)
|
|
|
|
|
|
|
|
// Whether or not the response is valid, we can mark the peer as idle and
|
|
|
|
// notify the scheduler to assign a new task. If the response is invalid,
|
|
|
|
// we'll drop the peer in a bit.
|
2022-08-31 18:58:18 +03:00
|
|
|
defer func() {
|
|
|
|
s.lock.Lock()
|
|
|
|
defer s.lock.Unlock()
|
|
|
|
if _, ok := s.peers[peer.ID()]; ok {
|
|
|
|
s.bytecodeIdlers[peer.ID()] = struct{}{}
|
|
|
|
}
|
|
|
|
select {
|
|
|
|
case s.update <- struct{}{}:
|
|
|
|
default:
|
|
|
|
}
|
|
|
|
}()
|
2020-12-14 12:27:15 +03:00
|
|
|
s.lock.Lock()
|
|
|
|
// Ensure the response is for a valid request
|
|
|
|
req, ok := s.bytecodeReqs[id]
|
|
|
|
if !ok {
|
|
|
|
// Request stale, perhaps the peer timed out but came through in the end
|
|
|
|
logger.Warn("Unexpected bytecode packet")
|
|
|
|
s.lock.Unlock()
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
delete(s.bytecodeReqs, id)
|
2021-05-19 15:09:03 +03:00
|
|
|
s.rates.Update(peer.ID(), ByteCodesMsg, time.Since(req.time), len(bytecodes))
|
2020-12-14 12:27:15 +03:00
|
|
|
|
|
|
|
// Clean up the request timeout timer, we'll see how to proceed further based
|
|
|
|
// on the actual delivered content
|
2021-01-25 09:17:05 +03:00
|
|
|
if !req.timeout.Stop() {
|
|
|
|
// The timeout is already triggered, and this request will be reverted+rescheduled
|
|
|
|
s.lock.Unlock()
|
|
|
|
return nil
|
|
|
|
}
|
2020-12-14 12:27:15 +03:00
|
|
|
|
|
|
|
// Response is valid, but check if peer is signalling that it does not have
|
|
|
|
// the requested data. For bytecode range queries that means the peer is not
|
|
|
|
// yet synced.
|
|
|
|
if len(bytecodes) == 0 {
|
|
|
|
logger.Debug("Peer rejected bytecode request")
|
2021-01-25 09:17:05 +03:00
|
|
|
s.statelessPeers[peer.ID()] = struct{}{}
|
2020-12-14 12:27:15 +03:00
|
|
|
s.lock.Unlock()
|
2021-01-07 13:58:07 +03:00
|
|
|
|
|
|
|
// Signal this request as failed, and ready for rescheduling
|
|
|
|
s.scheduleRevertBytecodeRequest(req)
|
2020-12-14 12:27:15 +03:00
|
|
|
return nil
|
|
|
|
}
|
|
|
|
s.lock.Unlock()
|
|
|
|
|
|
|
|
// Cross reference the requested bytecodes with the response to find gaps
|
|
|
|
// that the serving node is missing
|
2021-01-07 19:12:41 +03:00
|
|
|
hasher := sha3.NewLegacyKeccak256().(crypto.KeccakState)
|
|
|
|
hash := make([]byte, 32)
|
2020-12-14 12:27:15 +03:00
|
|
|
|
|
|
|
codes := make([][]byte, len(req.hashes))
|
|
|
|
for i, j := 0, 0; i < len(bytecodes); i++ {
|
|
|
|
// Find the next hash that we've been served, leaving misses with nils
|
|
|
|
hasher.Reset()
|
|
|
|
hasher.Write(bytecodes[i])
|
2021-01-07 19:12:41 +03:00
|
|
|
hasher.Read(hash)
|
2020-12-14 12:27:15 +03:00
|
|
|
|
|
|
|
for j < len(req.hashes) && !bytes.Equal(hash, req.hashes[j][:]) {
|
|
|
|
j++
|
|
|
|
}
|
|
|
|
if j < len(req.hashes) {
|
|
|
|
codes[j] = bytecodes[i]
|
|
|
|
j++
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
// We've either ran out of hashes, or got unrequested data
|
|
|
|
logger.Warn("Unexpected bytecodes", "count", len(bytecodes)-i)
|
2021-01-25 09:17:05 +03:00
|
|
|
// Signal this request as failed, and ready for rescheduling
|
|
|
|
s.scheduleRevertBytecodeRequest(req)
|
2020-12-14 12:27:15 +03:00
|
|
|
return errors.New("unexpected bytecode")
|
|
|
|
}
|
|
|
|
// Response validated, send it to the scheduler for filling
|
|
|
|
response := &bytecodeResponse{
|
|
|
|
task: req.task,
|
|
|
|
hashes: req.hashes,
|
|
|
|
codes: codes,
|
|
|
|
}
|
|
|
|
select {
|
2021-04-15 21:01:16 +03:00
|
|
|
case req.deliver <- response:
|
2020-12-14 12:27:15 +03:00
|
|
|
case <-req.cancel:
|
|
|
|
case <-req.stale:
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// OnStorage is a callback method to invoke when ranges of storage slots
|
|
|
|
// are received from a remote peer.
|
2021-01-25 09:17:05 +03:00
|
|
|
func (s *Syncer) OnStorage(peer SyncPeer, id uint64, hashes [][]common.Hash, slots [][][]byte, proof [][]byte) error {
|
2020-12-14 12:27:15 +03:00
|
|
|
// Gather some trace stats to aid in debugging issues
|
|
|
|
var (
|
|
|
|
hashCount int
|
|
|
|
slotCount int
|
|
|
|
size common.StorageSize
|
|
|
|
)
|
|
|
|
for _, hashset := range hashes {
|
|
|
|
size += common.StorageSize(common.HashLength * len(hashset))
|
|
|
|
hashCount += len(hashset)
|
|
|
|
}
|
|
|
|
for _, slotset := range slots {
|
|
|
|
for _, slot := range slotset {
|
|
|
|
size += common.StorageSize(len(slot))
|
|
|
|
}
|
|
|
|
slotCount += len(slotset)
|
|
|
|
}
|
|
|
|
for _, node := range proof {
|
|
|
|
size += common.StorageSize(len(node))
|
|
|
|
}
|
2021-01-25 09:17:05 +03:00
|
|
|
logger := peer.Log().New("reqid", id)
|
2020-12-14 12:27:15 +03:00
|
|
|
logger.Trace("Delivering ranges of storage slots", "accounts", len(hashes), "hashes", hashCount, "slots", slotCount, "proofs", len(proof), "size", size)
|
|
|
|
|
|
|
|
// Whether or not the response is valid, we can mark the peer as idle and
|
|
|
|
// notify the scheduler to assign a new task. If the response is invalid,
|
|
|
|
// we'll drop the peer in a bit.
|
2022-08-31 18:58:18 +03:00
|
|
|
defer func() {
|
|
|
|
s.lock.Lock()
|
|
|
|
defer s.lock.Unlock()
|
|
|
|
if _, ok := s.peers[peer.ID()]; ok {
|
|
|
|
s.storageIdlers[peer.ID()] = struct{}{}
|
|
|
|
}
|
|
|
|
select {
|
|
|
|
case s.update <- struct{}{}:
|
|
|
|
default:
|
|
|
|
}
|
|
|
|
}()
|
2020-12-14 12:27:15 +03:00
|
|
|
s.lock.Lock()
|
|
|
|
// Ensure the response is for a valid request
|
|
|
|
req, ok := s.storageReqs[id]
|
|
|
|
if !ok {
|
|
|
|
// Request stale, perhaps the peer timed out but came through in the end
|
|
|
|
logger.Warn("Unexpected storage ranges packet")
|
|
|
|
s.lock.Unlock()
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
delete(s.storageReqs, id)
|
2021-05-19 15:09:03 +03:00
|
|
|
s.rates.Update(peer.ID(), StorageRangesMsg, time.Since(req.time), int(size))
|
2020-12-14 12:27:15 +03:00
|
|
|
|
|
|
|
// Clean up the request timeout timer, we'll see how to proceed further based
|
|
|
|
// on the actual delivered content
|
2021-01-25 09:17:05 +03:00
|
|
|
if !req.timeout.Stop() {
|
|
|
|
// The timeout is already triggered, and this request will be reverted+rescheduled
|
|
|
|
s.lock.Unlock()
|
|
|
|
return nil
|
|
|
|
}
|
2020-12-14 12:27:15 +03:00
|
|
|
|
|
|
|
// Reject the response if the hash sets and slot sets don't match, or if the
|
|
|
|
// peer sent more data than requested.
|
|
|
|
if len(hashes) != len(slots) {
|
|
|
|
s.lock.Unlock()
|
2021-01-25 09:17:05 +03:00
|
|
|
s.scheduleRevertStorageRequest(req) // reschedule request
|
2020-12-14 12:27:15 +03:00
|
|
|
logger.Warn("Hash and slot set size mismatch", "hashset", len(hashes), "slotset", len(slots))
|
|
|
|
return errors.New("hash and slot set size mismatch")
|
|
|
|
}
|
|
|
|
if len(hashes) > len(req.accounts) {
|
|
|
|
s.lock.Unlock()
|
2021-01-25 09:17:05 +03:00
|
|
|
s.scheduleRevertStorageRequest(req) // reschedule request
|
2020-12-14 12:27:15 +03:00
|
|
|
logger.Warn("Hash set larger than requested", "hashset", len(hashes), "requested", len(req.accounts))
|
|
|
|
return errors.New("hash set larger than requested")
|
|
|
|
}
|
|
|
|
// Response is valid, but check if peer is signalling that it does not have
|
|
|
|
// the requested data. For storage range queries that means the state being
|
|
|
|
// retrieved was either already pruned remotely, or the peer is not yet
|
|
|
|
// synced to our head.
|
eth/protocols/snap: fix snap sync failure on empty storage range (#28306)
This change addresses an issue in snap sync, specifically when the entire sync process can be halted due to an encountered empty storage range.
Currently, on the snap sync client side, the response to an empty (partial) storage range is discarded as a non-delivery. However, this response can be a valid response, when the particular range requested does not contain any slots.
For instance, consider a large contract where the entire key space is divided into 16 chunks, and there are no available slots in the last chunk [0xf] -> [end]. When the node receives a request for this particular range, the response includes:
The proof with origin [0xf]
A nil storage slot set
If we simply discard this response, the finalization of the last range will be skipped, halting the entire sync process indefinitely. The test case TestSyncWithUnevenStorage can reproduce the scenario described above.
In addition, this change also defines the common variables MaxAddress and MaxHash.
2023-10-13 10:08:26 +03:00
|
|
|
if len(hashes) == 0 && len(proof) == 0 {
|
2020-12-14 12:27:15 +03:00
|
|
|
logger.Debug("Peer rejected storage request")
|
2021-01-25 09:17:05 +03:00
|
|
|
s.statelessPeers[peer.ID()] = struct{}{}
|
2020-12-14 12:27:15 +03:00
|
|
|
s.lock.Unlock()
|
2021-01-25 09:17:05 +03:00
|
|
|
s.scheduleRevertStorageRequest(req) // reschedule request
|
2020-12-14 12:27:15 +03:00
|
|
|
return nil
|
|
|
|
}
|
|
|
|
s.lock.Unlock()
|
|
|
|
|
|
|
|
// Reconstruct the partial tries from the response and verify them
|
2021-04-28 23:09:15 +03:00
|
|
|
var cont bool
|
|
|
|
|
eth/protocols/snap: fix snap sync failure on empty storage range (#28306)
This change addresses an issue in snap sync, specifically when the entire sync process can be halted due to an encountered empty storage range.
Currently, on the snap sync client side, the response to an empty (partial) storage range is discarded as a non-delivery. However, this response can be a valid response, when the particular range requested does not contain any slots.
For instance, consider a large contract where the entire key space is divided into 16 chunks, and there are no available slots in the last chunk [0xf] -> [end]. When the node receives a request for this particular range, the response includes:
The proof with origin [0xf]
A nil storage slot set
If we simply discard this response, the finalization of the last range will be skipped, halting the entire sync process indefinitely. The test case TestSyncWithUnevenStorage can reproduce the scenario described above.
In addition, this change also defines the common variables MaxAddress and MaxHash.
2023-10-13 10:08:26 +03:00
|
|
|
// If a proof was attached while the response is empty, it indicates that the
|
|
|
|
// requested range specified with 'origin' is empty. Construct an empty state
|
|
|
|
// response locally to finalize the range.
|
|
|
|
if len(hashes) == 0 && len(proof) > 0 {
|
|
|
|
hashes = append(hashes, []common.Hash{})
|
|
|
|
slots = append(slots, [][]byte{})
|
|
|
|
}
|
2020-12-14 12:27:15 +03:00
|
|
|
for i := 0; i < len(hashes); i++ {
|
|
|
|
// Convert the keys and proofs into an internal format
|
|
|
|
keys := make([][]byte, len(hashes[i]))
|
|
|
|
for j, key := range hashes[i] {
|
|
|
|
keys[j] = common.CopyBytes(key[:])
|
|
|
|
}
|
2023-10-10 11:30:47 +03:00
|
|
|
nodes := make(trienode.ProofList, 0, len(proof))
|
2020-12-14 12:27:15 +03:00
|
|
|
if i == len(hashes)-1 {
|
|
|
|
for _, node := range proof {
|
|
|
|
nodes = append(nodes, node)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
var err error
|
|
|
|
if len(nodes) == 0 {
|
|
|
|
// No proof has been attached, the response must cover the entire key
|
|
|
|
// space and hash to the origin root.
|
trie: make rhs-proof align with last key in range proofs (#28311)
During snap-sync, we request ranges of values: either a range of accounts or a range of storage values. For any large trie, e.g. the main account trie or a large storage trie, we cannot fetch everything at once.
Short version; we split it up and request in multiple stages. To do so, we use an origin field, to say "Give me all storage key/values where key > 0x20000000000000000". When the server fulfils this, the server provides the first key after origin, let's say 0x2e030000000000000 -- never providing the exact origin. However, the client-side needs to be able to verify that the 0x2e03.. indeed is the first one after 0x2000.., and therefore the attached proof concerns the origin, not the first key.
So, short-short version: the left-hand side of the proof relates to the origin, and is free-standing from the first leaf.
On the other hand, (pun intended), the right-hand side, there's no such 'gap' between "along what path does the proof walk" and the last provided leaf. The proof must prove the last element (unless there are no elements).
Therefore, we can simplify the semantics for trie.VerifyRangeProof by removing an argument. This doesn't make much difference in practice, but makes it so that we can remove some tests. The reason I am raising this is that the upcoming stacktrie-based verifier does not support such fancy features as standalone right-hand borders.
2023-10-13 17:05:29 +03:00
|
|
|
_, err = trie.VerifyRangeProof(req.roots[i], nil, keys, slots[i], nil)
|
2020-12-14 12:27:15 +03:00
|
|
|
if err != nil {
|
2021-01-25 09:17:05 +03:00
|
|
|
s.scheduleRevertStorageRequest(req) // reschedule request
|
2020-12-14 12:27:15 +03:00
|
|
|
logger.Warn("Storage slots failed proof", "err", err)
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
// A proof was attached, the response is only partial, check that the
|
|
|
|
// returned data is indeed part of the storage trie
|
2023-10-10 11:30:47 +03:00
|
|
|
proofdb := nodes.Set()
|
2020-12-14 12:27:15 +03:00
|
|
|
|
trie: make rhs-proof align with last key in range proofs (#28311)
During snap-sync, we request ranges of values: either a range of accounts or a range of storage values. For any large trie, e.g. the main account trie or a large storage trie, we cannot fetch everything at once.
Short version; we split it up and request in multiple stages. To do so, we use an origin field, to say "Give me all storage key/values where key > 0x20000000000000000". When the server fulfils this, the server provides the first key after origin, let's say 0x2e030000000000000 -- never providing the exact origin. However, the client-side needs to be able to verify that the 0x2e03.. indeed is the first one after 0x2000.., and therefore the attached proof concerns the origin, not the first key.
So, short-short version: the left-hand side of the proof relates to the origin, and is free-standing from the first leaf.
On the other hand, (pun intended), the right-hand side, there's no such 'gap' between "along what path does the proof walk" and the last provided leaf. The proof must prove the last element (unless there are no elements).
Therefore, we can simplify the semantics for trie.VerifyRangeProof by removing an argument. This doesn't make much difference in practice, but makes it so that we can remove some tests. The reason I am raising this is that the upcoming stacktrie-based verifier does not support such fancy features as standalone right-hand borders.
2023-10-13 17:05:29 +03:00
|
|
|
cont, err = trie.VerifyRangeProof(req.roots[i], req.origin[:], keys, slots[i], proofdb)
|
2020-12-14 12:27:15 +03:00
|
|
|
if err != nil {
|
2021-01-25 09:17:05 +03:00
|
|
|
s.scheduleRevertStorageRequest(req) // reschedule request
|
2020-12-14 12:27:15 +03:00
|
|
|
logger.Warn("Storage range failed proof", "err", err)
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
// Partial tries reconstructed, send them to the scheduler for storage filling
|
|
|
|
response := &storageResponse{
|
|
|
|
mainTask: req.mainTask,
|
|
|
|
subTask: req.subTask,
|
|
|
|
accounts: req.accounts,
|
|
|
|
roots: req.roots,
|
|
|
|
hashes: hashes,
|
|
|
|
slots: slots,
|
|
|
|
cont: cont,
|
|
|
|
}
|
|
|
|
select {
|
2021-04-15 21:01:16 +03:00
|
|
|
case req.deliver <- response:
|
2020-12-14 12:27:15 +03:00
|
|
|
case <-req.cancel:
|
|
|
|
case <-req.stale:
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// OnTrieNodes is a callback method to invoke when a batch of trie nodes
|
|
|
|
// are received from a remote peer.
|
2021-01-25 09:17:05 +03:00
|
|
|
func (s *Syncer) OnTrieNodes(peer SyncPeer, id uint64, trienodes [][]byte) error {
|
2020-12-14 12:27:15 +03:00
|
|
|
var size common.StorageSize
|
|
|
|
for _, node := range trienodes {
|
|
|
|
size += common.StorageSize(len(node))
|
|
|
|
}
|
2021-01-25 09:17:05 +03:00
|
|
|
logger := peer.Log().New("reqid", id)
|
2020-12-14 12:27:15 +03:00
|
|
|
logger.Trace("Delivering set of healing trienodes", "trienodes", len(trienodes), "bytes", size)
|
|
|
|
|
|
|
|
// Whether or not the response is valid, we can mark the peer as idle and
|
|
|
|
// notify the scheduler to assign a new task. If the response is invalid,
|
|
|
|
// we'll drop the peer in a bit.
|
2022-08-31 18:58:18 +03:00
|
|
|
defer func() {
|
|
|
|
s.lock.Lock()
|
|
|
|
defer s.lock.Unlock()
|
|
|
|
if _, ok := s.peers[peer.ID()]; ok {
|
|
|
|
s.trienodeHealIdlers[peer.ID()] = struct{}{}
|
|
|
|
}
|
|
|
|
select {
|
|
|
|
case s.update <- struct{}{}:
|
|
|
|
default:
|
|
|
|
}
|
|
|
|
}()
|
2020-12-14 12:27:15 +03:00
|
|
|
s.lock.Lock()
|
|
|
|
// Ensure the response is for a valid request
|
|
|
|
req, ok := s.trienodeHealReqs[id]
|
|
|
|
if !ok {
|
|
|
|
// Request stale, perhaps the peer timed out but came through in the end
|
|
|
|
logger.Warn("Unexpected trienode heal packet")
|
|
|
|
s.lock.Unlock()
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
delete(s.trienodeHealReqs, id)
|
2021-05-19 15:09:03 +03:00
|
|
|
s.rates.Update(peer.ID(), TrieNodesMsg, time.Since(req.time), len(trienodes))
|
2020-12-14 12:27:15 +03:00
|
|
|
|
|
|
|
// Clean up the request timeout timer, we'll see how to proceed further based
|
|
|
|
// on the actual delivered content
|
2021-01-25 09:17:05 +03:00
|
|
|
if !req.timeout.Stop() {
|
|
|
|
// The timeout is already triggered, and this request will be reverted+rescheduled
|
|
|
|
s.lock.Unlock()
|
|
|
|
return nil
|
|
|
|
}
|
2020-12-14 12:27:15 +03:00
|
|
|
|
|
|
|
// Response is valid, but check if peer is signalling that it does not have
|
|
|
|
// the requested data. For bytecode range queries that means the peer is not
|
|
|
|
// yet synced.
|
|
|
|
if len(trienodes) == 0 {
|
|
|
|
logger.Debug("Peer rejected trienode heal request")
|
2021-01-25 09:17:05 +03:00
|
|
|
s.statelessPeers[peer.ID()] = struct{}{}
|
2020-12-14 12:27:15 +03:00
|
|
|
s.lock.Unlock()
|
2021-01-07 13:58:07 +03:00
|
|
|
|
|
|
|
// Signal this request as failed, and ready for rescheduling
|
|
|
|
s.scheduleRevertTrienodeHealRequest(req)
|
2020-12-14 12:27:15 +03:00
|
|
|
return nil
|
|
|
|
}
|
|
|
|
s.lock.Unlock()
|
|
|
|
|
|
|
|
// Cross reference the requested trienodes with the response to find gaps
|
|
|
|
// that the serving node is missing
|
2022-09-09 11:42:57 +03:00
|
|
|
var (
|
|
|
|
hasher = sha3.NewLegacyKeccak256().(crypto.KeccakState)
|
|
|
|
hash = make([]byte, 32)
|
|
|
|
nodes = make([][]byte, len(req.hashes))
|
|
|
|
fills uint64
|
|
|
|
)
|
2020-12-14 12:27:15 +03:00
|
|
|
for i, j := 0, 0; i < len(trienodes); i++ {
|
|
|
|
// Find the next hash that we've been served, leaving misses with nils
|
|
|
|
hasher.Reset()
|
|
|
|
hasher.Write(trienodes[i])
|
2021-01-07 19:12:41 +03:00
|
|
|
hasher.Read(hash)
|
2020-12-14 12:27:15 +03:00
|
|
|
|
|
|
|
for j < len(req.hashes) && !bytes.Equal(hash, req.hashes[j][:]) {
|
|
|
|
j++
|
|
|
|
}
|
|
|
|
if j < len(req.hashes) {
|
|
|
|
nodes[j] = trienodes[i]
|
2022-09-09 11:42:57 +03:00
|
|
|
fills++
|
2020-12-14 12:27:15 +03:00
|
|
|
j++
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
// We've either ran out of hashes, or got unrequested data
|
|
|
|
logger.Warn("Unexpected healing trienodes", "count", len(trienodes)-i)
|
2022-09-09 11:42:57 +03:00
|
|
|
|
2021-01-25 09:17:05 +03:00
|
|
|
// Signal this request as failed, and ready for rescheduling
|
|
|
|
s.scheduleRevertTrienodeHealRequest(req)
|
2020-12-14 12:27:15 +03:00
|
|
|
return errors.New("unexpected healing trienode")
|
|
|
|
}
|
|
|
|
// Response validated, send it to the scheduler for filling
|
2023-04-25 13:06:50 +03:00
|
|
|
s.trienodeHealPend.Add(fills)
|
2022-09-09 11:42:57 +03:00
|
|
|
defer func() {
|
2023-04-25 13:06:50 +03:00
|
|
|
s.trienodeHealPend.Add(^(fills - 1))
|
2022-09-09 11:42:57 +03:00
|
|
|
}()
|
2020-12-14 12:27:15 +03:00
|
|
|
response := &trienodeHealResponse{
|
2022-07-15 14:55:51 +03:00
|
|
|
paths: req.paths,
|
2020-12-14 12:27:15 +03:00
|
|
|
task: req.task,
|
|
|
|
hashes: req.hashes,
|
|
|
|
nodes: nodes,
|
|
|
|
}
|
|
|
|
select {
|
2021-04-15 21:01:16 +03:00
|
|
|
case req.deliver <- response:
|
2020-12-14 12:27:15 +03:00
|
|
|
case <-req.cancel:
|
|
|
|
case <-req.stale:
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
// onHealByteCodes is a callback method to invoke when a batch of contract
|
|
|
|
// bytes codes are received from a remote peer in the healing phase.
|
2021-01-25 09:17:05 +03:00
|
|
|
func (s *Syncer) onHealByteCodes(peer SyncPeer, id uint64, bytecodes [][]byte) error {
|
2020-12-14 12:27:15 +03:00
|
|
|
var size common.StorageSize
|
|
|
|
for _, code := range bytecodes {
|
|
|
|
size += common.StorageSize(len(code))
|
|
|
|
}
|
2021-01-25 09:17:05 +03:00
|
|
|
logger := peer.Log().New("reqid", id)
|
2020-12-14 12:27:15 +03:00
|
|
|
logger.Trace("Delivering set of healing bytecodes", "bytecodes", len(bytecodes), "bytes", size)
|
|
|
|
|
|
|
|
// Whether or not the response is valid, we can mark the peer as idle and
|
|
|
|
// notify the scheduler to assign a new task. If the response is invalid,
|
|
|
|
// we'll drop the peer in a bit.
|
2022-08-31 18:58:18 +03:00
|
|
|
defer func() {
|
|
|
|
s.lock.Lock()
|
|
|
|
defer s.lock.Unlock()
|
|
|
|
if _, ok := s.peers[peer.ID()]; ok {
|
|
|
|
s.bytecodeHealIdlers[peer.ID()] = struct{}{}
|
|
|
|
}
|
|
|
|
select {
|
|
|
|
case s.update <- struct{}{}:
|
|
|
|
default:
|
|
|
|
}
|
|
|
|
}()
|
2020-12-14 12:27:15 +03:00
|
|
|
s.lock.Lock()
|
|
|
|
// Ensure the response is for a valid request
|
|
|
|
req, ok := s.bytecodeHealReqs[id]
|
|
|
|
if !ok {
|
|
|
|
// Request stale, perhaps the peer timed out but came through in the end
|
|
|
|
logger.Warn("Unexpected bytecode heal packet")
|
|
|
|
s.lock.Unlock()
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
delete(s.bytecodeHealReqs, id)
|
2021-05-19 15:09:03 +03:00
|
|
|
s.rates.Update(peer.ID(), ByteCodesMsg, time.Since(req.time), len(bytecodes))
|
2020-12-14 12:27:15 +03:00
|
|
|
|
|
|
|
// Clean up the request timeout timer, we'll see how to proceed further based
|
|
|
|
// on the actual delivered content
|
2021-01-25 09:17:05 +03:00
|
|
|
if !req.timeout.Stop() {
|
|
|
|
// The timeout is already triggered, and this request will be reverted+rescheduled
|
|
|
|
s.lock.Unlock()
|
|
|
|
return nil
|
|
|
|
}
|
2020-12-14 12:27:15 +03:00
|
|
|
|
|
|
|
// Response is valid, but check if peer is signalling that it does not have
|
|
|
|
// the requested data. For bytecode range queries that means the peer is not
|
|
|
|
// yet synced.
|
|
|
|
if len(bytecodes) == 0 {
|
|
|
|
logger.Debug("Peer rejected bytecode heal request")
|
2021-01-25 09:17:05 +03:00
|
|
|
s.statelessPeers[peer.ID()] = struct{}{}
|
2020-12-14 12:27:15 +03:00
|
|
|
s.lock.Unlock()
|
2021-01-07 13:58:07 +03:00
|
|
|
|
|
|
|
// Signal this request as failed, and ready for rescheduling
|
|
|
|
s.scheduleRevertBytecodeHealRequest(req)
|
2020-12-14 12:27:15 +03:00
|
|
|
return nil
|
|
|
|
}
|
|
|
|
s.lock.Unlock()
|
|
|
|
|
|
|
|
// Cross reference the requested bytecodes with the response to find gaps
|
|
|
|
// that the serving node is missing
|
2021-01-07 19:12:41 +03:00
|
|
|
hasher := sha3.NewLegacyKeccak256().(crypto.KeccakState)
|
|
|
|
hash := make([]byte, 32)
|
2020-12-14 12:27:15 +03:00
|
|
|
|
|
|
|
codes := make([][]byte, len(req.hashes))
|
|
|
|
for i, j := 0, 0; i < len(bytecodes); i++ {
|
|
|
|
// Find the next hash that we've been served, leaving misses with nils
|
|
|
|
hasher.Reset()
|
|
|
|
hasher.Write(bytecodes[i])
|
2021-01-07 19:12:41 +03:00
|
|
|
hasher.Read(hash)
|
2020-12-14 12:27:15 +03:00
|
|
|
|
|
|
|
for j < len(req.hashes) && !bytes.Equal(hash, req.hashes[j][:]) {
|
|
|
|
j++
|
|
|
|
}
|
|
|
|
if j < len(req.hashes) {
|
|
|
|
codes[j] = bytecodes[i]
|
|
|
|
j++
|
|
|
|
continue
|
|
|
|
}
|
|
|
|
// We've either ran out of hashes, or got unrequested data
|
|
|
|
logger.Warn("Unexpected healing bytecodes", "count", len(bytecodes)-i)
|
2021-01-25 09:17:05 +03:00
|
|
|
// Signal this request as failed, and ready for rescheduling
|
|
|
|
s.scheduleRevertBytecodeHealRequest(req)
|
2020-12-14 12:27:15 +03:00
|
|
|
return errors.New("unexpected healing bytecode")
|
|
|
|
}
|
|
|
|
// Response validated, send it to the scheduler for filling
|
|
|
|
response := &bytecodeHealResponse{
|
|
|
|
task: req.task,
|
|
|
|
hashes: req.hashes,
|
|
|
|
codes: codes,
|
|
|
|
}
|
|
|
|
select {
|
2021-04-15 21:01:16 +03:00
|
|
|
case req.deliver <- response:
|
2020-12-14 12:27:15 +03:00
|
|
|
case <-req.cancel:
|
|
|
|
case <-req.stale:
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
core, eth: faster snapshot generation (#22504)
* eth/protocols: persist received state segments
* core: initial implementation
* core/state/snapshot: add tests
* core, eth: updates
* eth/protocols/snapshot: count flat state size
* core/state: add metrics
* core/state/snapshot: skip unnecessary deletion
* core/state/snapshot: rename
* core/state/snapshot: use the global batch
* core/state/snapshot: add logs and fix wiping
* core/state/snapshot: fix
* core/state/snapshot: save generation progress even if the batch is empty
* core/state/snapshot: fixes
* core/state/snapshot: fix initial account range length
* core/state/snapshot: fix initial account range
* eth/protocols/snap: store flat states during the healing
* eth/protocols/snap: print logs
* core/state/snapshot: refactor (#4)
* core/state/snapshot: refactor
* core/state/snapshot: tiny fix and polish
Co-authored-by: rjl493456442 <garyrong0905@gmail.com>
* core, eth: fixes
* core, eth: fix healing writer
* core, trie, eth: fix paths
* eth/protocols/snap: fix encoding
* eth, core: add debug log
* core/state/generate: release iterator asap (#5)
core/state/snapshot: less copy
core/state/snapshot: revert split loop
core/state/snapshot: handle storage becoming empty, improve test robustness
core/state: test modified codehash
core/state/snapshot: polish
* core/state/snapshot: optimize stats counter
* core, eth: add metric
* core/state/snapshot: update comments
* core/state/snapshot: improve tests
* core/state/snapshot: replace secure trie with standard trie
* core/state/snapshot: wrap return as the struct
* core/state/snapshot: skip wiping correct states
* core/state/snapshot: updates
* core/state/snapshot: fixes
* core/state/snapshot: fix panic due to reference flaw in closure
* core/state/snapshot: fix errors in state generation logic + fix log output
* core/state/snapshot: remove an error case
* core/state/snapshot: fix condition-check for exhausted snap state
* core/state/snapshot: use stackTrie for small tries
* core/state/snapshot: don't resolve small storage tries in vain
* core/state/snapshot: properly clean up storage of deleted accounts
* core/state/snapshot: avoid RLP-encoding in some cases + minor nitpicks
* core/state/snapshot: fix error (+testcase)
* core/state/snapshot: clean up tests a bit
* core/state/snapshot: work in progress on better tests
* core/state/snapshot: polish code
* core/state/snapshot: fix trie iteration abortion trigger
* core/state/snapshot: fixes flaws
* core/state/snapshot: remove panic
* core/state/snapshot: fix abort
* core/state/snapshot: more tests (plus failing testcase)
* core/state/snapshot: more testcases + fix for failing test
* core/state/snapshot: testcase for malformed data
* core/state/snapshot: some test nitpicks
* core/state/snapshot: improvements to logging
* core/state/snapshot: testcase to demo error in abortion
* core/state/snapshot: fix abortion
* cmd/geth: make verify-state report the root
* trie: fix failing test
* core/state/snapshot: add timer metrics
* core/state/snapshot: fix metrics
* core/state/snapshot: udpate tests
* eth/protocols/snap: write snapshot account even if code or state is needed
* core/state/snapshot: fix diskmore check
* core/state/snapshot: review fixes
* core/state/snapshot: improve error message
* cmd/geth: rename 'error' to 'err' in logs
* core/state/snapshot: fix some review concerns
* core/state/snapshot, eth/protocols/snap: clear snapshot marker when starting/resuming snap sync
* core: add error log
* core/state/snapshot: use proper timers for metrics collection
* core/state/snapshot: address some review concerns
* eth/protocols/snap: improved log message
* eth/protocols/snap: fix heal logs to condense infos
* core/state/snapshot: wait for generator termination before restarting
* core/state/snapshot: revert timers to counters to track total time
Co-authored-by: Martin Holst Swende <martin@swende.se>
Co-authored-by: Péter Szilágyi <peterke@gmail.com>
2021-04-14 23:23:11 +03:00
|
|
|
// onHealState is a callback method to invoke when a flat state(account
|
2022-08-19 09:00:21 +03:00
|
|
|
// or storage slot) is downloaded during the healing stage. The flat states
|
core, eth: faster snapshot generation (#22504)
* eth/protocols: persist received state segments
* core: initial implementation
* core/state/snapshot: add tests
* core, eth: updates
* eth/protocols/snapshot: count flat state size
* core/state: add metrics
* core/state/snapshot: skip unnecessary deletion
* core/state/snapshot: rename
* core/state/snapshot: use the global batch
* core/state/snapshot: add logs and fix wiping
* core/state/snapshot: fix
* core/state/snapshot: save generation progress even if the batch is empty
* core/state/snapshot: fixes
* core/state/snapshot: fix initial account range length
* core/state/snapshot: fix initial account range
* eth/protocols/snap: store flat states during the healing
* eth/protocols/snap: print logs
* core/state/snapshot: refactor (#4)
* core/state/snapshot: refactor
* core/state/snapshot: tiny fix and polish
Co-authored-by: rjl493456442 <garyrong0905@gmail.com>
* core, eth: fixes
* core, eth: fix healing writer
* core, trie, eth: fix paths
* eth/protocols/snap: fix encoding
* eth, core: add debug log
* core/state/generate: release iterator asap (#5)
core/state/snapshot: less copy
core/state/snapshot: revert split loop
core/state/snapshot: handle storage becoming empty, improve test robustness
core/state: test modified codehash
core/state/snapshot: polish
* core/state/snapshot: optimize stats counter
* core, eth: add metric
* core/state/snapshot: update comments
* core/state/snapshot: improve tests
* core/state/snapshot: replace secure trie with standard trie
* core/state/snapshot: wrap return as the struct
* core/state/snapshot: skip wiping correct states
* core/state/snapshot: updates
* core/state/snapshot: fixes
* core/state/snapshot: fix panic due to reference flaw in closure
* core/state/snapshot: fix errors in state generation logic + fix log output
* core/state/snapshot: remove an error case
* core/state/snapshot: fix condition-check for exhausted snap state
* core/state/snapshot: use stackTrie for small tries
* core/state/snapshot: don't resolve small storage tries in vain
* core/state/snapshot: properly clean up storage of deleted accounts
* core/state/snapshot: avoid RLP-encoding in some cases + minor nitpicks
* core/state/snapshot: fix error (+testcase)
* core/state/snapshot: clean up tests a bit
* core/state/snapshot: work in progress on better tests
* core/state/snapshot: polish code
* core/state/snapshot: fix trie iteration abortion trigger
* core/state/snapshot: fixes flaws
* core/state/snapshot: remove panic
* core/state/snapshot: fix abort
* core/state/snapshot: more tests (plus failing testcase)
* core/state/snapshot: more testcases + fix for failing test
* core/state/snapshot: testcase for malformed data
* core/state/snapshot: some test nitpicks
* core/state/snapshot: improvements to logging
* core/state/snapshot: testcase to demo error in abortion
* core/state/snapshot: fix abortion
* cmd/geth: make verify-state report the root
* trie: fix failing test
* core/state/snapshot: add timer metrics
* core/state/snapshot: fix metrics
* core/state/snapshot: udpate tests
* eth/protocols/snap: write snapshot account even if code or state is needed
* core/state/snapshot: fix diskmore check
* core/state/snapshot: review fixes
* core/state/snapshot: improve error message
* cmd/geth: rename 'error' to 'err' in logs
* core/state/snapshot: fix some review concerns
* core/state/snapshot, eth/protocols/snap: clear snapshot marker when starting/resuming snap sync
* core: add error log
* core/state/snapshot: use proper timers for metrics collection
* core/state/snapshot: address some review concerns
* eth/protocols/snap: improved log message
* eth/protocols/snap: fix heal logs to condense infos
* core/state/snapshot: wait for generator termination before restarting
* core/state/snapshot: revert timers to counters to track total time
Co-authored-by: Martin Holst Swende <martin@swende.se>
Co-authored-by: Péter Szilágyi <peterke@gmail.com>
2021-04-14 23:23:11 +03:00
|
|
|
// can be persisted blindly and can be fixed later in the generation stage.
|
|
|
|
// Note it's not concurrent safe, please handle the concurrent issue outside.
|
|
|
|
func (s *Syncer) onHealState(paths [][]byte, value []byte) error {
|
|
|
|
if len(paths) == 1 {
|
2021-09-28 11:48:07 +03:00
|
|
|
var account types.StateAccount
|
core, eth: faster snapshot generation (#22504)
* eth/protocols: persist received state segments
* core: initial implementation
* core/state/snapshot: add tests
* core, eth: updates
* eth/protocols/snapshot: count flat state size
* core/state: add metrics
* core/state/snapshot: skip unnecessary deletion
* core/state/snapshot: rename
* core/state/snapshot: use the global batch
* core/state/snapshot: add logs and fix wiping
* core/state/snapshot: fix
* core/state/snapshot: save generation progress even if the batch is empty
* core/state/snapshot: fixes
* core/state/snapshot: fix initial account range length
* core/state/snapshot: fix initial account range
* eth/protocols/snap: store flat states during the healing
* eth/protocols/snap: print logs
* core/state/snapshot: refactor (#4)
* core/state/snapshot: refactor
* core/state/snapshot: tiny fix and polish
Co-authored-by: rjl493456442 <garyrong0905@gmail.com>
* core, eth: fixes
* core, eth: fix healing writer
* core, trie, eth: fix paths
* eth/protocols/snap: fix encoding
* eth, core: add debug log
* core/state/generate: release iterator asap (#5)
core/state/snapshot: less copy
core/state/snapshot: revert split loop
core/state/snapshot: handle storage becoming empty, improve test robustness
core/state: test modified codehash
core/state/snapshot: polish
* core/state/snapshot: optimize stats counter
* core, eth: add metric
* core/state/snapshot: update comments
* core/state/snapshot: improve tests
* core/state/snapshot: replace secure trie with standard trie
* core/state/snapshot: wrap return as the struct
* core/state/snapshot: skip wiping correct states
* core/state/snapshot: updates
* core/state/snapshot: fixes
* core/state/snapshot: fix panic due to reference flaw in closure
* core/state/snapshot: fix errors in state generation logic + fix log output
* core/state/snapshot: remove an error case
* core/state/snapshot: fix condition-check for exhausted snap state
* core/state/snapshot: use stackTrie for small tries
* core/state/snapshot: don't resolve small storage tries in vain
* core/state/snapshot: properly clean up storage of deleted accounts
* core/state/snapshot: avoid RLP-encoding in some cases + minor nitpicks
* core/state/snapshot: fix error (+testcase)
* core/state/snapshot: clean up tests a bit
* core/state/snapshot: work in progress on better tests
* core/state/snapshot: polish code
* core/state/snapshot: fix trie iteration abortion trigger
* core/state/snapshot: fixes flaws
* core/state/snapshot: remove panic
* core/state/snapshot: fix abort
* core/state/snapshot: more tests (plus failing testcase)
* core/state/snapshot: more testcases + fix for failing test
* core/state/snapshot: testcase for malformed data
* core/state/snapshot: some test nitpicks
* core/state/snapshot: improvements to logging
* core/state/snapshot: testcase to demo error in abortion
* core/state/snapshot: fix abortion
* cmd/geth: make verify-state report the root
* trie: fix failing test
* core/state/snapshot: add timer metrics
* core/state/snapshot: fix metrics
* core/state/snapshot: udpate tests
* eth/protocols/snap: write snapshot account even if code or state is needed
* core/state/snapshot: fix diskmore check
* core/state/snapshot: review fixes
* core/state/snapshot: improve error message
* cmd/geth: rename 'error' to 'err' in logs
* core/state/snapshot: fix some review concerns
* core/state/snapshot, eth/protocols/snap: clear snapshot marker when starting/resuming snap sync
* core: add error log
* core/state/snapshot: use proper timers for metrics collection
* core/state/snapshot: address some review concerns
* eth/protocols/snap: improved log message
* eth/protocols/snap: fix heal logs to condense infos
* core/state/snapshot: wait for generator termination before restarting
* core/state/snapshot: revert timers to counters to track total time
Co-authored-by: Martin Holst Swende <martin@swende.se>
Co-authored-by: Péter Szilágyi <peterke@gmail.com>
2021-04-14 23:23:11 +03:00
|
|
|
if err := rlp.DecodeBytes(value, &account); err != nil {
|
2022-06-13 17:24:45 +03:00
|
|
|
return nil // Returning the error here would drop the remote peer
|
core, eth: faster snapshot generation (#22504)
* eth/protocols: persist received state segments
* core: initial implementation
* core/state/snapshot: add tests
* core, eth: updates
* eth/protocols/snapshot: count flat state size
* core/state: add metrics
* core/state/snapshot: skip unnecessary deletion
* core/state/snapshot: rename
* core/state/snapshot: use the global batch
* core/state/snapshot: add logs and fix wiping
* core/state/snapshot: fix
* core/state/snapshot: save generation progress even if the batch is empty
* core/state/snapshot: fixes
* core/state/snapshot: fix initial account range length
* core/state/snapshot: fix initial account range
* eth/protocols/snap: store flat states during the healing
* eth/protocols/snap: print logs
* core/state/snapshot: refactor (#4)
* core/state/snapshot: refactor
* core/state/snapshot: tiny fix and polish
Co-authored-by: rjl493456442 <garyrong0905@gmail.com>
* core, eth: fixes
* core, eth: fix healing writer
* core, trie, eth: fix paths
* eth/protocols/snap: fix encoding
* eth, core: add debug log
* core/state/generate: release iterator asap (#5)
core/state/snapshot: less copy
core/state/snapshot: revert split loop
core/state/snapshot: handle storage becoming empty, improve test robustness
core/state: test modified codehash
core/state/snapshot: polish
* core/state/snapshot: optimize stats counter
* core, eth: add metric
* core/state/snapshot: update comments
* core/state/snapshot: improve tests
* core/state/snapshot: replace secure trie with standard trie
* core/state/snapshot: wrap return as the struct
* core/state/snapshot: skip wiping correct states
* core/state/snapshot: updates
* core/state/snapshot: fixes
* core/state/snapshot: fix panic due to reference flaw in closure
* core/state/snapshot: fix errors in state generation logic + fix log output
* core/state/snapshot: remove an error case
* core/state/snapshot: fix condition-check for exhausted snap state
* core/state/snapshot: use stackTrie for small tries
* core/state/snapshot: don't resolve small storage tries in vain
* core/state/snapshot: properly clean up storage of deleted accounts
* core/state/snapshot: avoid RLP-encoding in some cases + minor nitpicks
* core/state/snapshot: fix error (+testcase)
* core/state/snapshot: clean up tests a bit
* core/state/snapshot: work in progress on better tests
* core/state/snapshot: polish code
* core/state/snapshot: fix trie iteration abortion trigger
* core/state/snapshot: fixes flaws
* core/state/snapshot: remove panic
* core/state/snapshot: fix abort
* core/state/snapshot: more tests (plus failing testcase)
* core/state/snapshot: more testcases + fix for failing test
* core/state/snapshot: testcase for malformed data
* core/state/snapshot: some test nitpicks
* core/state/snapshot: improvements to logging
* core/state/snapshot: testcase to demo error in abortion
* core/state/snapshot: fix abortion
* cmd/geth: make verify-state report the root
* trie: fix failing test
* core/state/snapshot: add timer metrics
* core/state/snapshot: fix metrics
* core/state/snapshot: udpate tests
* eth/protocols/snap: write snapshot account even if code or state is needed
* core/state/snapshot: fix diskmore check
* core/state/snapshot: review fixes
* core/state/snapshot: improve error message
* cmd/geth: rename 'error' to 'err' in logs
* core/state/snapshot: fix some review concerns
* core/state/snapshot, eth/protocols/snap: clear snapshot marker when starting/resuming snap sync
* core: add error log
* core/state/snapshot: use proper timers for metrics collection
* core/state/snapshot: address some review concerns
* eth/protocols/snap: improved log message
* eth/protocols/snap: fix heal logs to condense infos
* core/state/snapshot: wait for generator termination before restarting
* core/state/snapshot: revert timers to counters to track total time
Co-authored-by: Martin Holst Swende <martin@swende.se>
Co-authored-by: Péter Szilágyi <peterke@gmail.com>
2021-04-14 23:23:11 +03:00
|
|
|
}
|
2023-06-06 11:17:39 +03:00
|
|
|
blob := types.SlimAccountRLP(account)
|
core, eth: faster snapshot generation (#22504)
* eth/protocols: persist received state segments
* core: initial implementation
* core/state/snapshot: add tests
* core, eth: updates
* eth/protocols/snapshot: count flat state size
* core/state: add metrics
* core/state/snapshot: skip unnecessary deletion
* core/state/snapshot: rename
* core/state/snapshot: use the global batch
* core/state/snapshot: add logs and fix wiping
* core/state/snapshot: fix
* core/state/snapshot: save generation progress even if the batch is empty
* core/state/snapshot: fixes
* core/state/snapshot: fix initial account range length
* core/state/snapshot: fix initial account range
* eth/protocols/snap: store flat states during the healing
* eth/protocols/snap: print logs
* core/state/snapshot: refactor (#4)
* core/state/snapshot: refactor
* core/state/snapshot: tiny fix and polish
Co-authored-by: rjl493456442 <garyrong0905@gmail.com>
* core, eth: fixes
* core, eth: fix healing writer
* core, trie, eth: fix paths
* eth/protocols/snap: fix encoding
* eth, core: add debug log
* core/state/generate: release iterator asap (#5)
core/state/snapshot: less copy
core/state/snapshot: revert split loop
core/state/snapshot: handle storage becoming empty, improve test robustness
core/state: test modified codehash
core/state/snapshot: polish
* core/state/snapshot: optimize stats counter
* core, eth: add metric
* core/state/snapshot: update comments
* core/state/snapshot: improve tests
* core/state/snapshot: replace secure trie with standard trie
* core/state/snapshot: wrap return as the struct
* core/state/snapshot: skip wiping correct states
* core/state/snapshot: updates
* core/state/snapshot: fixes
* core/state/snapshot: fix panic due to reference flaw in closure
* core/state/snapshot: fix errors in state generation logic + fix log output
* core/state/snapshot: remove an error case
* core/state/snapshot: fix condition-check for exhausted snap state
* core/state/snapshot: use stackTrie for small tries
* core/state/snapshot: don't resolve small storage tries in vain
* core/state/snapshot: properly clean up storage of deleted accounts
* core/state/snapshot: avoid RLP-encoding in some cases + minor nitpicks
* core/state/snapshot: fix error (+testcase)
* core/state/snapshot: clean up tests a bit
* core/state/snapshot: work in progress on better tests
* core/state/snapshot: polish code
* core/state/snapshot: fix trie iteration abortion trigger
* core/state/snapshot: fixes flaws
* core/state/snapshot: remove panic
* core/state/snapshot: fix abort
* core/state/snapshot: more tests (plus failing testcase)
* core/state/snapshot: more testcases + fix for failing test
* core/state/snapshot: testcase for malformed data
* core/state/snapshot: some test nitpicks
* core/state/snapshot: improvements to logging
* core/state/snapshot: testcase to demo error in abortion
* core/state/snapshot: fix abortion
* cmd/geth: make verify-state report the root
* trie: fix failing test
* core/state/snapshot: add timer metrics
* core/state/snapshot: fix metrics
* core/state/snapshot: udpate tests
* eth/protocols/snap: write snapshot account even if code or state is needed
* core/state/snapshot: fix diskmore check
* core/state/snapshot: review fixes
* core/state/snapshot: improve error message
* cmd/geth: rename 'error' to 'err' in logs
* core/state/snapshot: fix some review concerns
* core/state/snapshot, eth/protocols/snap: clear snapshot marker when starting/resuming snap sync
* core: add error log
* core/state/snapshot: use proper timers for metrics collection
* core/state/snapshot: address some review concerns
* eth/protocols/snap: improved log message
* eth/protocols/snap: fix heal logs to condense infos
* core/state/snapshot: wait for generator termination before restarting
* core/state/snapshot: revert timers to counters to track total time
Co-authored-by: Martin Holst Swende <martin@swende.se>
Co-authored-by: Péter Szilágyi <peterke@gmail.com>
2021-04-14 23:23:11 +03:00
|
|
|
rawdb.WriteAccountSnapshot(s.stateWriter, common.BytesToHash(paths[0]), blob)
|
|
|
|
s.accountHealed += 1
|
|
|
|
s.accountHealedBytes += common.StorageSize(1 + common.HashLength + len(blob))
|
|
|
|
}
|
|
|
|
if len(paths) == 2 {
|
|
|
|
rawdb.WriteStorageSnapshot(s.stateWriter, common.BytesToHash(paths[0]), common.BytesToHash(paths[1]), value)
|
|
|
|
s.storageHealed += 1
|
|
|
|
s.storageHealedBytes += common.StorageSize(1 + 2*common.HashLength + len(value))
|
|
|
|
}
|
|
|
|
if s.stateWriter.ValueSize() > ethdb.IdealBatchSize {
|
|
|
|
s.stateWriter.Write() // It's fine to ignore the error here
|
|
|
|
s.stateWriter.Reset()
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
2020-12-14 12:27:15 +03:00
|
|
|
// hashSpace is the total size of the 256 bit hash space for accounts.
|
|
|
|
var hashSpace = new(big.Int).Exp(common.Big2, common.Big256, nil)
|
|
|
|
|
|
|
|
// report calculates various status reports and provides it to the user.
|
|
|
|
func (s *Syncer) report(force bool) {
|
|
|
|
if len(s.tasks) > 0 {
|
|
|
|
s.reportSyncProgress(force)
|
|
|
|
return
|
|
|
|
}
|
|
|
|
s.reportHealProgress(force)
|
|
|
|
}
|
|
|
|
|
|
|
|
// reportSyncProgress calculates various status reports and provides it to the user.
|
|
|
|
func (s *Syncer) reportSyncProgress(force bool) {
|
|
|
|
// Don't report all the events, just occasionally
|
2021-04-27 17:19:59 +03:00
|
|
|
if !force && time.Since(s.logTime) < 8*time.Second {
|
2020-12-14 12:27:15 +03:00
|
|
|
return
|
|
|
|
}
|
|
|
|
// Don't report anything until we have a meaningful progress
|
|
|
|
synced := s.accountBytes + s.bytecodeBytes + s.storageBytes
|
|
|
|
if synced == 0 {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
accountGaps := new(big.Int)
|
|
|
|
for _, task := range s.tasks {
|
|
|
|
accountGaps.Add(accountGaps, new(big.Int).Sub(task.Last.Big(), task.Next.Big()))
|
|
|
|
}
|
|
|
|
accountFills := new(big.Int).Sub(hashSpace, accountGaps)
|
|
|
|
if accountFills.BitLen() == 0 {
|
|
|
|
return
|
|
|
|
}
|
|
|
|
s.logTime = time.Now()
|
|
|
|
estBytes := float64(new(big.Int).Div(
|
|
|
|
new(big.Int).Mul(new(big.Int).SetUint64(uint64(synced)), hashSpace),
|
|
|
|
accountFills,
|
|
|
|
).Uint64())
|
2022-03-10 14:46:48 +03:00
|
|
|
// Don't report anything until we have a meaningful progress
|
|
|
|
if estBytes < 1.0 {
|
|
|
|
return
|
|
|
|
}
|
2020-12-14 12:27:15 +03:00
|
|
|
elapsed := time.Since(s.startTime)
|
|
|
|
estTime := elapsed / time.Duration(synced) * time.Duration(estBytes)
|
|
|
|
|
|
|
|
// Create a mega progress report
|
|
|
|
var (
|
|
|
|
progress = fmt.Sprintf("%.2f%%", float64(synced)*100/estBytes)
|
2021-04-15 20:35:00 +03:00
|
|
|
accounts = fmt.Sprintf("%v@%v", log.FormatLogfmtUint64(s.accountSynced), s.accountBytes.TerminalString())
|
|
|
|
storage = fmt.Sprintf("%v@%v", log.FormatLogfmtUint64(s.storageSynced), s.storageBytes.TerminalString())
|
|
|
|
bytecode = fmt.Sprintf("%v@%v", log.FormatLogfmtUint64(s.bytecodeSynced), s.bytecodeBytes.TerminalString())
|
2020-12-14 12:27:15 +03:00
|
|
|
)
|
2022-09-23 11:11:41 +03:00
|
|
|
log.Info("Syncing: state download in progress", "synced", progress, "state", synced,
|
2020-12-14 12:27:15 +03:00
|
|
|
"accounts", accounts, "slots", storage, "codes", bytecode, "eta", common.PrettyDuration(estTime-elapsed))
|
|
|
|
}
|
|
|
|
|
|
|
|
// reportHealProgress calculates various status reports and provides it to the user.
|
|
|
|
func (s *Syncer) reportHealProgress(force bool) {
|
|
|
|
// Don't report all the events, just occasionally
|
2021-04-27 17:19:59 +03:00
|
|
|
if !force && time.Since(s.logTime) < 8*time.Second {
|
2020-12-14 12:27:15 +03:00
|
|
|
return
|
|
|
|
}
|
|
|
|
s.logTime = time.Now()
|
|
|
|
|
|
|
|
// Create a mega progress report
|
|
|
|
var (
|
2021-04-15 20:35:00 +03:00
|
|
|
trienode = fmt.Sprintf("%v@%v", log.FormatLogfmtUint64(s.trienodeHealSynced), s.trienodeHealBytes.TerminalString())
|
|
|
|
bytecode = fmt.Sprintf("%v@%v", log.FormatLogfmtUint64(s.bytecodeHealSynced), s.bytecodeHealBytes.TerminalString())
|
|
|
|
accounts = fmt.Sprintf("%v@%v", log.FormatLogfmtUint64(s.accountHealed), s.accountHealedBytes.TerminalString())
|
|
|
|
storage = fmt.Sprintf("%v@%v", log.FormatLogfmtUint64(s.storageHealed), s.storageHealedBytes.TerminalString())
|
2020-12-14 12:27:15 +03:00
|
|
|
)
|
2022-09-23 11:11:41 +03:00
|
|
|
log.Info("Syncing: state healing in progress", "accounts", accounts, "slots", storage,
|
core, eth: faster snapshot generation (#22504)
* eth/protocols: persist received state segments
* core: initial implementation
* core/state/snapshot: add tests
* core, eth: updates
* eth/protocols/snapshot: count flat state size
* core/state: add metrics
* core/state/snapshot: skip unnecessary deletion
* core/state/snapshot: rename
* core/state/snapshot: use the global batch
* core/state/snapshot: add logs and fix wiping
* core/state/snapshot: fix
* core/state/snapshot: save generation progress even if the batch is empty
* core/state/snapshot: fixes
* core/state/snapshot: fix initial account range length
* core/state/snapshot: fix initial account range
* eth/protocols/snap: store flat states during the healing
* eth/protocols/snap: print logs
* core/state/snapshot: refactor (#4)
* core/state/snapshot: refactor
* core/state/snapshot: tiny fix and polish
Co-authored-by: rjl493456442 <garyrong0905@gmail.com>
* core, eth: fixes
* core, eth: fix healing writer
* core, trie, eth: fix paths
* eth/protocols/snap: fix encoding
* eth, core: add debug log
* core/state/generate: release iterator asap (#5)
core/state/snapshot: less copy
core/state/snapshot: revert split loop
core/state/snapshot: handle storage becoming empty, improve test robustness
core/state: test modified codehash
core/state/snapshot: polish
* core/state/snapshot: optimize stats counter
* core, eth: add metric
* core/state/snapshot: update comments
* core/state/snapshot: improve tests
* core/state/snapshot: replace secure trie with standard trie
* core/state/snapshot: wrap return as the struct
* core/state/snapshot: skip wiping correct states
* core/state/snapshot: updates
* core/state/snapshot: fixes
* core/state/snapshot: fix panic due to reference flaw in closure
* core/state/snapshot: fix errors in state generation logic + fix log output
* core/state/snapshot: remove an error case
* core/state/snapshot: fix condition-check for exhausted snap state
* core/state/snapshot: use stackTrie for small tries
* core/state/snapshot: don't resolve small storage tries in vain
* core/state/snapshot: properly clean up storage of deleted accounts
* core/state/snapshot: avoid RLP-encoding in some cases + minor nitpicks
* core/state/snapshot: fix error (+testcase)
* core/state/snapshot: clean up tests a bit
* core/state/snapshot: work in progress on better tests
* core/state/snapshot: polish code
* core/state/snapshot: fix trie iteration abortion trigger
* core/state/snapshot: fixes flaws
* core/state/snapshot: remove panic
* core/state/snapshot: fix abort
* core/state/snapshot: more tests (plus failing testcase)
* core/state/snapshot: more testcases + fix for failing test
* core/state/snapshot: testcase for malformed data
* core/state/snapshot: some test nitpicks
* core/state/snapshot: improvements to logging
* core/state/snapshot: testcase to demo error in abortion
* core/state/snapshot: fix abortion
* cmd/geth: make verify-state report the root
* trie: fix failing test
* core/state/snapshot: add timer metrics
* core/state/snapshot: fix metrics
* core/state/snapshot: udpate tests
* eth/protocols/snap: write snapshot account even if code or state is needed
* core/state/snapshot: fix diskmore check
* core/state/snapshot: review fixes
* core/state/snapshot: improve error message
* cmd/geth: rename 'error' to 'err' in logs
* core/state/snapshot: fix some review concerns
* core/state/snapshot, eth/protocols/snap: clear snapshot marker when starting/resuming snap sync
* core: add error log
* core/state/snapshot: use proper timers for metrics collection
* core/state/snapshot: address some review concerns
* eth/protocols/snap: improved log message
* eth/protocols/snap: fix heal logs to condense infos
* core/state/snapshot: wait for generator termination before restarting
* core/state/snapshot: revert timers to counters to track total time
Co-authored-by: Martin Holst Swende <martin@swende.se>
Co-authored-by: Péter Szilágyi <peterke@gmail.com>
2021-04-14 23:23:11 +03:00
|
|
|
"codes", bytecode, "nodes", trienode, "pending", s.healer.scheduler.Pending())
|
2020-12-14 12:27:15 +03:00
|
|
|
}
|
2021-04-27 17:19:59 +03:00
|
|
|
|
|
|
|
// estimateRemainingSlots tries to determine roughly how many slots are left in
|
|
|
|
// a contract storage, based on the number of keys and the last hash. This method
|
|
|
|
// assumes that the hashes are lexicographically ordered and evenly distributed.
|
|
|
|
func estimateRemainingSlots(hashes int, last common.Hash) (uint64, error) {
|
|
|
|
if last == (common.Hash{}) {
|
|
|
|
return 0, errors.New("last hash empty")
|
|
|
|
}
|
|
|
|
space := new(big.Int).Mul(math.MaxBig256, big.NewInt(int64(hashes)))
|
|
|
|
space.Div(space, last.Big())
|
|
|
|
if !space.IsUint64() {
|
|
|
|
// Gigantic address space probably due to too few or malicious slots
|
|
|
|
return 0, errors.New("too few slots for estimation")
|
|
|
|
}
|
|
|
|
return space.Uint64() - uint64(hashes), nil
|
|
|
|
}
|
2021-05-19 15:09:03 +03:00
|
|
|
|
|
|
|
// capacitySort implements the Sort interface, allowing sorting by peer message
|
|
|
|
// throughput. Note, callers should use sort.Reverse to get the desired effect
|
|
|
|
// of highest capacity being at the front.
|
|
|
|
type capacitySort struct {
|
|
|
|
ids []string
|
2021-05-27 19:43:55 +03:00
|
|
|
caps []int
|
2021-05-19 15:09:03 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
func (s *capacitySort) Len() int {
|
|
|
|
return len(s.ids)
|
|
|
|
}
|
|
|
|
|
|
|
|
func (s *capacitySort) Less(i, j int) bool {
|
|
|
|
return s.caps[i] < s.caps[j]
|
|
|
|
}
|
|
|
|
|
|
|
|
func (s *capacitySort) Swap(i, j int) {
|
|
|
|
s.ids[i], s.ids[j] = s.ids[j], s.ids[i]
|
|
|
|
s.caps[i], s.caps[j] = s.caps[j], s.caps[i]
|
|
|
|
}
|
2022-05-10 17:37:24 +03:00
|
|
|
|
|
|
|
// healRequestSort implements the Sort interface, allowing sorting trienode
|
|
|
|
// heal requests, which is a prerequisite for merging storage-requests.
|
|
|
|
type healRequestSort struct {
|
2022-07-15 14:55:51 +03:00
|
|
|
paths []string
|
|
|
|
hashes []common.Hash
|
|
|
|
syncPaths []trie.SyncPath
|
2022-05-10 17:37:24 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
func (t *healRequestSort) Len() int {
|
|
|
|
return len(t.hashes)
|
|
|
|
}
|
|
|
|
|
|
|
|
func (t *healRequestSort) Less(i, j int) bool {
|
2022-07-15 14:55:51 +03:00
|
|
|
a := t.syncPaths[i]
|
|
|
|
b := t.syncPaths[j]
|
2022-05-10 17:37:24 +03:00
|
|
|
switch bytes.Compare(a[0], b[0]) {
|
|
|
|
case -1:
|
|
|
|
return true
|
|
|
|
case 1:
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
// identical first part
|
|
|
|
if len(a) < len(b) {
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
if len(b) < len(a) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
if len(a) == 2 {
|
|
|
|
return bytes.Compare(a[1], b[1]) < 0
|
|
|
|
}
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
func (t *healRequestSort) Swap(i, j int) {
|
|
|
|
t.paths[i], t.paths[j] = t.paths[j], t.paths[i]
|
2022-07-15 14:55:51 +03:00
|
|
|
t.hashes[i], t.hashes[j] = t.hashes[j], t.hashes[i]
|
|
|
|
t.syncPaths[i], t.syncPaths[j] = t.syncPaths[j], t.syncPaths[i]
|
2022-05-10 17:37:24 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// Merge merges the pathsets, so that several storage requests concerning the
|
|
|
|
// same account are merged into one, to reduce bandwidth.
|
|
|
|
// OBS: This operation is moot if t has not first been sorted.
|
|
|
|
func (t *healRequestSort) Merge() []TrieNodePathSet {
|
|
|
|
var result []TrieNodePathSet
|
2022-07-15 14:55:51 +03:00
|
|
|
for _, path := range t.syncPaths {
|
2023-01-03 16:41:40 +03:00
|
|
|
pathset := TrieNodePathSet(path)
|
2022-05-10 17:37:24 +03:00
|
|
|
if len(path) == 1 {
|
|
|
|
// It's an account reference.
|
|
|
|
result = append(result, pathset)
|
|
|
|
} else {
|
|
|
|
// It's a storage reference.
|
|
|
|
end := len(result) - 1
|
|
|
|
if len(result) == 0 || !bytes.Equal(pathset[0], result[end][0]) {
|
2022-07-15 14:55:51 +03:00
|
|
|
// The account doesn't match last, create a new entry.
|
2022-05-10 17:37:24 +03:00
|
|
|
result = append(result, pathset)
|
|
|
|
} else {
|
|
|
|
// It's the same account as the previous one, add to the storage
|
|
|
|
// paths of that request.
|
|
|
|
result[end] = append(result[end], pathset[1])
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return result
|
|
|
|
}
|
|
|
|
|
|
|
|
// sortByAccountPath takes hashes and paths, and sorts them. After that, it generates
|
|
|
|
// the TrieNodePaths and merges paths which belongs to the same account path.
|
2022-07-15 14:55:51 +03:00
|
|
|
func sortByAccountPath(paths []string, hashes []common.Hash) ([]string, []common.Hash, []trie.SyncPath, []TrieNodePathSet) {
|
|
|
|
var syncPaths []trie.SyncPath
|
|
|
|
for _, path := range paths {
|
|
|
|
syncPaths = append(syncPaths, trie.NewSyncPath([]byte(path)))
|
|
|
|
}
|
|
|
|
n := &healRequestSort{paths, hashes, syncPaths}
|
2022-05-10 17:37:24 +03:00
|
|
|
sort.Sort(n)
|
|
|
|
pathsets := n.Merge()
|
2022-07-15 14:55:51 +03:00
|
|
|
return n.paths, n.hashes, n.syncPaths, pathsets
|
2022-05-10 17:37:24 +03:00
|
|
|
}
|