2015-07-07 03:54:22 +03:00
// Copyright 2014 The go-ethereum Authors
2015-07-22 19:48:40 +03:00
// This file is part of the go-ethereum library.
2015-07-07 03:54:22 +03:00
//
2015-07-23 19:35:11 +03:00
// The go-ethereum library is free software: you can redistribute it and/or modify
2015-07-07 03:54:22 +03:00
// it under the terms of the GNU Lesser General Public License as published by
// the Free Software Foundation, either version 3 of the License, or
// (at your option) any later version.
//
2015-07-22 19:48:40 +03:00
// The go-ethereum library is distributed in the hope that it will be useful,
2015-07-07 03:54:22 +03:00
// but WITHOUT ANY WARRANTY; without even the implied warranty of
2015-07-22 19:48:40 +03:00
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
2015-07-07 03:54:22 +03:00
// GNU Lesser General Public License for more details.
//
// You should have received a copy of the GNU Lesser General Public License
2015-07-22 19:48:40 +03:00
// along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
2015-07-07 03:54:22 +03:00
2015-07-07 06:08:16 +03:00
// Package core implements the Ethereum consensus protocol.
2014-12-04 11:28:02 +02:00
package core
2014-02-15 00:56:09 +02:00
import (
2015-07-10 15:29:40 +03:00
"errors"
2014-09-24 12:39:17 +03:00
"fmt"
2015-03-18 14:36:48 +02:00
"io"
2014-07-30 01:31:15 +03:00
"math/big"
2015-10-13 12:04:25 +03:00
mrand "math/rand"
2019-11-21 17:35:22 +03:00
"sort"
2014-12-18 14:12:54 +02:00
"sync"
2015-06-12 17:45:53 +03:00
"sync/atomic"
2015-04-04 17:35:23 +03:00
"time"
2014-07-30 01:31:15 +03:00
2015-03-16 12:27:38 +02:00
"github.com/ethereum/go-ethereum/common"
2017-01-11 00:55:54 +03:00
"github.com/ethereum/go-ethereum/common/mclock"
2018-09-03 18:33:21 +03:00
"github.com/ethereum/go-ethereum/common/prque"
2017-04-05 01:16:29 +03:00
"github.com/ethereum/go-ethereum/consensus"
2018-05-07 14:35:06 +03:00
"github.com/ethereum/go-ethereum/core/rawdb"
2015-03-23 23:59:19 +02:00
"github.com/ethereum/go-ethereum/core/state"
2019-08-06 13:40:28 +03:00
"github.com/ethereum/go-ethereum/core/state/snapshot"
2015-03-17 00:48:18 +02:00
"github.com/ethereum/go-ethereum/core/types"
2015-10-19 17:08:17 +03:00
"github.com/ethereum/go-ethereum/core/vm"
2015-09-14 10:35:57 +03:00
"github.com/ethereum/go-ethereum/ethdb"
2014-12-03 15:05:19 +02:00
"github.com/ethereum/go-ethereum/event"
2021-10-07 16:47:50 +03:00
"github.com/ethereum/go-ethereum/internal/syncx"
2017-02-22 15:10:07 +03:00
"github.com/ethereum/go-ethereum/log"
2015-06-29 16:11:01 +03:00
"github.com/ethereum/go-ethereum/metrics"
2016-10-20 14:36:29 +03:00
"github.com/ethereum/go-ethereum/params"
2015-10-05 19:37:56 +03:00
"github.com/ethereum/go-ethereum/trie"
2019-11-18 11:49:18 +03:00
lru "github.com/hashicorp/golang-lru"
2014-02-15 00:56:09 +02:00
)
2015-03-03 10:44:41 +02:00
var (
2019-06-10 14:21:02 +03:00
headBlockGauge = metrics . NewRegisteredGauge ( "chain/head/block" , nil )
headHeaderGauge = metrics . NewRegisteredGauge ( "chain/head/header" , nil )
headFastBlockGauge = metrics . NewRegisteredGauge ( "chain/head/receipt" , nil )
2019-03-25 11:01:18 +03:00
accountReadTimer = metrics . NewRegisteredTimer ( "chain/account/reads" , nil )
accountHashTimer = metrics . NewRegisteredTimer ( "chain/account/hashes" , nil )
accountUpdateTimer = metrics . NewRegisteredTimer ( "chain/account/updates" , nil )
accountCommitTimer = metrics . NewRegisteredTimer ( "chain/account/commits" , nil )
storageReadTimer = metrics . NewRegisteredTimer ( "chain/storage/reads" , nil )
storageHashTimer = metrics . NewRegisteredTimer ( "chain/storage/hashes" , nil )
storageUpdateTimer = metrics . NewRegisteredTimer ( "chain/storage/updates" , nil )
storageCommitTimer = metrics . NewRegisteredTimer ( "chain/storage/commits" , nil )
2019-11-26 10:48:29 +03:00
snapshotAccountReadTimer = metrics . NewRegisteredTimer ( "chain/snapshot/account/reads" , nil )
snapshotStorageReadTimer = metrics . NewRegisteredTimer ( "chain/snapshot/storage/reads" , nil )
2019-08-06 13:40:28 +03:00
snapshotCommitTimer = metrics . NewRegisteredTimer ( "chain/snapshot/commits" , nil )
2018-11-28 11:29:05 +03:00
blockInsertTimer = metrics . NewRegisteredTimer ( "chain/inserts" , nil )
blockValidationTimer = metrics . NewRegisteredTimer ( "chain/validation" , nil )
blockExecutionTimer = metrics . NewRegisteredTimer ( "chain/execution" , nil )
blockWriteTimer = metrics . NewRegisteredTimer ( "chain/write" , nil )
2020-08-20 10:49:35 +03:00
blockReorgMeter = metrics . NewRegisteredMeter ( "chain/reorg/executes" , nil )
blockReorgAddMeter = metrics . NewRegisteredMeter ( "chain/reorg/add" , nil )
blockReorgDropMeter = metrics . NewRegisteredMeter ( "chain/reorg/drop" , nil )
blockReorgInvalidatedTx = metrics . NewRegisteredMeter ( "chain/reorg/invalidTx" , nil )
2015-07-10 15:29:40 +03:00
2019-03-25 13:41:50 +03:00
blockPrefetchExecuteTimer = metrics . NewRegisteredTimer ( "chain/prefetch/executes" , nil )
2019-05-13 14:07:55 +03:00
blockPrefetchInterruptMeter = metrics . NewRegisteredMeter ( "chain/prefetch/interrupts" , nil )
all: integrate the freezer with fast sync
* all: freezer style syncing
core, eth, les, light: clean up freezer relative APIs
core, eth, les, trie, ethdb, light: clean a bit
core, eth, les, light: add unit tests
core, light: rewrite setHead function
core, eth: fix downloader unit tests
core: add receipt chain insertion test
core: use constant instead of hardcoding table name
core: fix rollback
core: fix setHead
core/rawdb: remove canonical block first and then iterate side chain
core/rawdb, ethdb: add hasAncient interface
eth/downloader: calculate ancient limit via cht first
core, eth, ethdb: lots of fixes
* eth/downloader: print ancient disable log only for fast sync
2019-04-25 17:59:48 +03:00
errInsertionInterrupted = errors . New ( "insertion is interrupted" )
2021-10-07 16:47:50 +03:00
errChainStopped = errors . New ( "blockchain is stopped" )
2015-03-03 10:44:41 +02:00
)
2014-06-23 14:54:10 +03:00
2015-04-20 21:37:40 +03:00
const (
2015-08-31 20:21:02 +03:00
bodyCacheLimit = 256
2015-06-19 19:16:09 +03:00
blockCacheLimit = 256
2018-09-29 23:53:31 +03:00
receiptsCacheLimit = 32
2019-08-21 12:29:34 +03:00
txLookupCacheLimit = 1024
2015-06-05 15:07:49 +03:00
maxFutureBlocks = 256
maxTimeFutureBlocks = 30
2019-05-30 21:51:13 +03:00
TriesInMemory = 128
2017-05-25 17:21:20 +03:00
// BlockChainVersion ensures that an incompatible database forces a resync from scratch.
2019-02-21 16:14:35 +03:00
//
2019-03-27 19:11:24 +03:00
// Changelog:
//
// - Version 4
// The following incompatible database changes were added:
// * the `BlockNumber`, `TxHash`, `TxIndex`, `BlockHash` and `Index` fields of log are deleted
// * the `Bloom` field of receipt is deleted
// * the `BlockIndex` and `TxIndex` fields of txlookup are deleted
// - Version 5
// The following incompatible database changes were added:
// * the `TxHash`, `GasCost`, and `ContractAddress` fields are no longer stored for a receipt
// * the `TxHash`, `GasCost`, and `ContractAddress` fields are computed by looking up the
// receipts' corresponding block
2019-04-25 17:24:56 +03:00
// - Version 6
// The following incompatible database changes were added:
// * Transaction lookup information stores the corresponding block number instead of block hash
core, cmd, vendor: fixes and database inspection tool (#15)
* core, eth: some fixes for freezer
* vendor, core/rawdb, cmd/geth: add db inspector
* core, cmd/utils: check ancient store path forceily
* cmd/geth, common, core/rawdb: a few fixes
* cmd/geth: support windows file rename and fix rename error
* core: support ancient plugin
* core, cmd: streaming file copy
* cmd, consensus, core, tests: keep genesis in leveldb
* core: write txlookup during ancient init
* core: bump database version
2019-05-14 17:07:44 +03:00
// - Version 7
// The following incompatible database changes were added:
// * Use freezer as the ancient database to maintain all ancient data
2020-08-21 15:10:40 +03:00
// - Version 8
// The following incompatible database changes were added:
// * New scheme for contract code in order to separate the codes and trie nodes
BlockChainVersion uint64 = 8
2015-04-20 21:37:40 +03:00
)
2015-01-02 13:07:54 +02:00
2018-02-05 19:40:32 +03:00
// CacheConfig contains the configuration values for the trie caching/pruning
// that's resident in a blockchain.
type CacheConfig struct {
2019-04-01 11:52:11 +03:00
TrieCleanLimit int // Memory allowance (MB) to use for caching trie nodes in memory
2020-07-28 16:30:31 +03:00
TrieCleanJournal string // Disk journal for saving clean cache entries.
TrieCleanRejournal time . Duration // Time interval to dump clean cache to disk periodically
2019-04-01 11:52:11 +03:00
TrieCleanNoPrefetch bool // Whether to disable heuristic state prefetching for followup blocks
TrieDirtyLimit int // Memory limit (MB) at which to start flushing dirty trie nodes to disk
TrieDirtyDisabled bool // Whether to disable trie write caching and GC altogether (archive node)
TrieTimeLimit time . Duration // Time limit after which to flush the current in-memory trie to disk
2019-11-26 10:48:29 +03:00
SnapshotLimit int // Memory allowance (MB) to use for caching snapshot entries in memory
2020-11-18 12:51:33 +03:00
Preimages bool // Whether to store preimage of trie key to the disk
2020-03-03 10:10:23 +03:00
SnapshotWait bool // Wait for snapshot construction on startup. TODO(karalabe): This is a dirty hack for testing, nuke it
2018-02-05 19:40:32 +03:00
}
2020-08-20 13:01:24 +03:00
// defaultCacheConfig are the default caching values if none are specified by the
// user (also used during testing).
var defaultCacheConfig = & CacheConfig {
TrieCleanLimit : 256 ,
TrieDirtyLimit : 256 ,
TrieTimeLimit : 5 * time . Minute ,
SnapshotLimit : 256 ,
SnapshotWait : true ,
}
2015-10-19 17:08:17 +03:00
// BlockChain represents the canonical chain given a database with a genesis
// block. The Blockchain manages chain imports, reverts, chain reorganisations.
//
// Importing blocks in to the block chain happens according to the set of rules
// defined by the two stage Validator. Processing of blocks is done using the
// Processor which processes the included transaction. The validation of the state
// is done in the second part of the Validator. Failing results in aborting of
// the import.
//
// The BlockChain also helps in returning blocks from **any** chain included
// in the database as well as blocks that represents the canonical chain. It's
// important to note that GetBlock can return any block and does not need to be
// included in the canonical one where as GetBlockByNumber always represents the
// canonical chain.
2015-08-31 18:09:50 +03:00
type BlockChain struct {
2018-02-05 19:40:32 +03:00
chainConfig * params . ChainConfig // Chain & network configuration
cacheConfig * CacheConfig // Cache configuration for pruning
2019-11-22 14:23:49 +03:00
db ethdb . Database // Low level persistent database to store final content in
snaps * snapshot . Tree // Snapshot tree for fast trie leaf access
triegc * prque . Prque // Priority queue mapping block numbers to tries to gc
gcproc time . Duration // Accumulates canonical block processing for trie dumping
2016-03-02 00:32:43 +02:00
2020-05-11 18:58:43 +03:00
// txLookupLimit is the maximum number of blocks from head whose tx indices
// are reserved:
// * 0: means no limit and regenerate any missing indexes
// * N: means N block limit [HEAD-N+1, HEAD] and delete extra indexes
// * nil: disable tx reindexer/deleter, but still index new blocks
txLookupLimit uint64
2017-08-18 13:58:36 +03:00
hc * HeaderChain
rmLogsFeed event . Feed
chainFeed event . Feed
chainSideFeed event . Feed
chainHeadFeed event . Feed
logsFeed event . Feed
2019-02-26 14:32:48 +03:00
blockProcFeed event . Feed
2017-08-18 13:58:36 +03:00
scope event . SubscriptionScope
genesisBlock * types . Block
2015-12-16 05:26:23 +02:00
2021-10-07 16:47:50 +03:00
// This mutex synchronizes chain write operations.
// Readers don't need to take it, they can just read the database.
chainmu * syncx . ClosableMutex
2015-04-28 18:48:46 +03:00
2018-02-26 12:53:10 +03:00
currentBlock atomic . Value // Current head of the block chain
currentFastBlock atomic . Value // Current head of the fast-sync chain (may be above the block chain!)
2014-02-15 00:56:09 +02:00
2018-09-29 23:53:31 +03:00
stateCache state . Database // State database to reuse between imports (contains state cache)
bodyCache * lru . Cache // Cache for the most recent block bodies
bodyRLPCache * lru . Cache // Cache for the most recent block bodies in RLP encoded format
receiptsCache * lru . Cache // Cache for the most recent receipts per block
blockCache * lru . Cache // Cache for the most recent entire blocks
2019-08-21 12:29:34 +03:00
txLookupCache * lru . Cache // Cache for the most recent transaction lookup data.
2018-09-29 23:53:31 +03:00
futureBlocks * lru . Cache // future blocks are blocks added for later processing
2015-03-20 16:54:42 +02:00
2021-10-07 16:47:50 +03:00
wg sync . WaitGroup //
quit chan struct { } // shutdown signal, closed in Stop.
2020-05-26 22:37:37 +03:00
running int32 // 0 if chain is running, 1 when stopped
procInterrupt int32 // interrupt signaler for block processing
2015-05-17 02:42:30 +03:00
2021-01-08 16:01:49 +03:00
engine consensus . Engine
validator Validator // Block and state validator interface
prefetcher Prefetcher
processor Processor // Block transaction processor interface
vmConfig vm . Config
2017-02-13 23:44:06 +03:00
2021-09-07 13:31:17 +03:00
shouldPreserve func ( * types . Block ) bool // Function used to determine whether should preserve the given block.
2014-12-18 14:12:54 +02:00
}
2014-02-15 00:56:09 +02:00
2015-10-19 17:08:17 +03:00
// NewBlockChain returns a fully initialised block chain using information
2017-04-10 16:01:31 +03:00
// available in the database. It initialises the default Ethereum Validator and
2015-10-19 17:08:17 +03:00
// Processor.
2020-05-11 18:58:43 +03:00
func NewBlockChain ( db ethdb . Database , cacheConfig * CacheConfig , chainConfig * params . ChainConfig , engine consensus . Engine , vmConfig vm . Config , shouldPreserve func ( block * types . Block ) bool , txLookupLimit * uint64 ) ( * BlockChain , error ) {
2018-02-05 19:40:32 +03:00
if cacheConfig == nil {
2020-08-20 13:01:24 +03:00
cacheConfig = defaultCacheConfig
2018-02-05 19:40:32 +03:00
}
2015-08-31 20:21:02 +03:00
bodyCache , _ := lru . New ( bodyCacheLimit )
bodyRLPCache , _ := lru . New ( bodyCacheLimit )
2018-09-29 23:53:31 +03:00
receiptsCache , _ := lru . New ( receiptsCacheLimit )
2015-08-31 20:21:02 +03:00
blockCache , _ := lru . New ( blockCacheLimit )
2019-08-21 12:29:34 +03:00
txLookupCache , _ := lru . New ( txLookupCacheLimit )
2015-08-31 20:21:02 +03:00
futureBlocks , _ := lru . New ( maxFutureBlocks )
2015-08-31 18:09:50 +03:00
bc := & BlockChain {
2020-11-18 12:51:33 +03:00
chainConfig : chainConfig ,
cacheConfig : cacheConfig ,
db : db ,
triegc : prque . New ( nil ) ,
stateCache : state . NewDatabaseWithConfig ( db , & trie . Config {
Cache : cacheConfig . TrieCleanLimit ,
Journal : cacheConfig . TrieCleanJournal ,
Preimages : cacheConfig . Preimages ,
} ) ,
2018-09-20 20:02:15 +03:00
quit : make ( chan struct { } ) ,
2021-10-07 16:47:50 +03:00
chainmu : syncx . NewClosableMutex ( ) ,
2018-09-20 20:02:15 +03:00
shouldPreserve : shouldPreserve ,
bodyCache : bodyCache ,
bodyRLPCache : bodyRLPCache ,
2018-09-29 23:53:31 +03:00
receiptsCache : receiptsCache ,
2018-09-20 20:02:15 +03:00
blockCache : blockCache ,
2019-08-21 12:29:34 +03:00
txLookupCache : txLookupCache ,
2018-09-20 20:02:15 +03:00
futureBlocks : futureBlocks ,
engine : engine ,
vmConfig : vmConfig ,
2015-04-20 21:37:40 +03:00
}
2019-03-25 13:41:50 +03:00
bc . validator = NewBlockValidator ( chainConfig , bc , engine )
bc . prefetcher = newStatePrefetcher ( chainConfig , bc , engine )
bc . processor = NewStateProcessor ( chainConfig , bc , engine )
2015-12-16 05:26:23 +02:00
var err error
2020-05-26 22:37:37 +03:00
bc . hc , err = NewHeaderChain ( db , chainConfig , engine , bc . insertStopped )
2015-10-13 12:04:25 +03:00
if err != nil {
return nil , err
}
2015-07-10 15:29:40 +03:00
bc . genesisBlock = bc . GetBlockByNumber ( 0 )
if bc . genesisBlock == nil {
2016-03-02 00:32:43 +02:00
return nil , ErrNoGenesis
2015-07-10 15:29:40 +03:00
}
2019-09-26 12:10:35 +03:00
var nilBlock * types . Block
bc . currentBlock . Store ( nilBlock )
bc . currentFastBlock . Store ( nilBlock )
core, cmd, vendor: fixes and database inspection tool (#15)
* core, eth: some fixes for freezer
* vendor, core/rawdb, cmd/geth: add db inspector
* core, cmd/utils: check ancient store path forceily
* cmd/geth, common, core/rawdb: a few fixes
* cmd/geth: support windows file rename and fix rename error
* core: support ancient plugin
* core, cmd: streaming file copy
* cmd, consensus, core, tests: keep genesis in leveldb
* core: write txlookup during ancient init
* core: bump database version
2019-05-14 17:07:44 +03:00
// Initialize the chain with ancient data if it isn't empty.
2020-05-11 18:58:43 +03:00
var txIndexBlock uint64
core, cmd, vendor: fixes and database inspection tool (#15)
* core, eth: some fixes for freezer
* vendor, core/rawdb, cmd/geth: add db inspector
* core, cmd/utils: check ancient store path forceily
* cmd/geth, common, core/rawdb: a few fixes
* cmd/geth: support windows file rename and fix rename error
* core: support ancient plugin
* core, cmd: streaming file copy
* cmd, consensus, core, tests: keep genesis in leveldb
* core: write txlookup during ancient init
* core: bump database version
2019-05-14 17:07:44 +03:00
if bc . empty ( ) {
2019-05-27 15:48:30 +03:00
rawdb . InitDatabaseFromFreezer ( bc . db )
2020-05-11 18:58:43 +03:00
// If ancient database is not empty, reconstruct all missing
// indices in the background.
frozen , _ := bc . db . Ancients ( )
if frozen > 0 {
txIndexBlock = frozen
}
core, cmd, vendor: fixes and database inspection tool (#15)
* core, eth: some fixes for freezer
* vendor, core/rawdb, cmd/geth: add db inspector
* core, cmd/utils: check ancient store path forceily
* cmd/geth, common, core/rawdb: a few fixes
* cmd/geth: support windows file rename and fix rename error
* core: support ancient plugin
* core, cmd: streaming file copy
* cmd, consensus, core, tests: keep genesis in leveldb
* core: write txlookup during ancient init
* core: bump database version
2019-05-14 17:07:44 +03:00
}
2015-09-21 15:36:29 +03:00
if err := bc . loadLastState ( ) ; err != nil {
2015-07-10 15:29:40 +03:00
return nil , err
2015-06-08 13:12:13 +03:00
}
2021-10-07 16:47:50 +03:00
2020-08-20 13:01:24 +03:00
// Make sure the state associated with the block is available
head := bc . CurrentBlock ( )
if _ , err := state . New ( head . Root ( ) , bc . stateCache , bc . snaps ) ; err != nil {
2020-10-29 22:01:58 +03:00
// Head state is missing, before the state recovery, find out the
// disk layer point of snapshot(if it's enabled). Make sure the
// rewound point is lower than disk layer.
var diskRoot common . Hash
if bc . cacheConfig . SnapshotLimit > 0 {
diskRoot = rawdb . ReadSnapshotRoot ( bc . db )
}
if diskRoot != ( common . Hash { } ) {
log . Warn ( "Head state missing, repairing" , "number" , head . Number ( ) , "hash" , head . Hash ( ) , "snaproot" , diskRoot )
snapDisk , err := bc . SetHeadBeyondRoot ( head . NumberU64 ( ) , diskRoot )
if err != nil {
return nil , err
}
// Chain rewound, persist old snapshot number to indicate recovery procedure
if snapDisk != 0 {
rawdb . WriteSnapshotRecoveryNumber ( bc . db , snapDisk )
}
} else {
log . Warn ( "Head state missing, repairing" , "number" , head . Number ( ) , "hash" , head . Hash ( ) )
if err := bc . SetHead ( head . NumberU64 ( ) ) ; err != nil {
return nil , err
}
2020-08-20 13:01:24 +03:00
}
}
2021-10-07 16:47:50 +03:00
2020-08-20 13:01:24 +03:00
// Ensure that a previous crash in SetHead doesn't leave extra ancients
core, cmd, vendor: fixes and database inspection tool (#15)
* core, eth: some fixes for freezer
* vendor, core/rawdb, cmd/geth: add db inspector
* core, cmd/utils: check ancient store path forceily
* cmd/geth, common, core/rawdb: a few fixes
* cmd/geth: support windows file rename and fix rename error
* core: support ancient plugin
* core, cmd: streaming file copy
* cmd, consensus, core, tests: keep genesis in leveldb
* core: write txlookup during ancient init
* core: bump database version
2019-05-14 17:07:44 +03:00
if frozen , err := bc . db . Ancients ( ) ; err == nil && frozen > 0 {
all: integrate the freezer with fast sync
* all: freezer style syncing
core, eth, les, light: clean up freezer relative APIs
core, eth, les, trie, ethdb, light: clean a bit
core, eth, les, light: add unit tests
core, light: rewrite setHead function
core, eth: fix downloader unit tests
core: add receipt chain insertion test
core: use constant instead of hardcoding table name
core: fix rollback
core: fix setHead
core/rawdb: remove canonical block first and then iterate side chain
core/rawdb, ethdb: add hasAncient interface
eth/downloader: calculate ancient limit via cht first
core, eth, ethdb: lots of fixes
* eth/downloader: print ancient disable log only for fast sync
2019-04-25 17:59:48 +03:00
var (
needRewind bool
low uint64
)
// The head full block may be rolled back to a very low height due to
// blockchain repair. If the head full block is even lower than the ancient
// chain, truncate the ancient store.
fullBlock := bc . CurrentBlock ( )
2020-08-20 13:01:24 +03:00
if fullBlock != nil && fullBlock . Hash ( ) != bc . genesisBlock . Hash ( ) && fullBlock . NumberU64 ( ) < frozen - 1 {
all: integrate the freezer with fast sync
* all: freezer style syncing
core, eth, les, light: clean up freezer relative APIs
core, eth, les, trie, ethdb, light: clean a bit
core, eth, les, light: add unit tests
core, light: rewrite setHead function
core, eth: fix downloader unit tests
core: add receipt chain insertion test
core: use constant instead of hardcoding table name
core: fix rollback
core: fix setHead
core/rawdb: remove canonical block first and then iterate side chain
core/rawdb, ethdb: add hasAncient interface
eth/downloader: calculate ancient limit via cht first
core, eth, ethdb: lots of fixes
* eth/downloader: print ancient disable log only for fast sync
2019-04-25 17:59:48 +03:00
needRewind = true
low = fullBlock . NumberU64 ( )
}
// In fast sync, it may happen that ancient data has been written to the
// ancient store, but the LastFastBlock has not been updated, truncate the
// extra data here.
fastBlock := bc . CurrentFastBlock ( )
if fastBlock != nil && fastBlock . NumberU64 ( ) < frozen - 1 {
needRewind = true
if fastBlock . NumberU64 ( ) < low || low == 0 {
low = fastBlock . NumberU64 ( )
}
}
if needRewind {
2020-08-20 13:01:24 +03:00
log . Error ( "Truncating ancient chain" , "from" , bc . CurrentHeader ( ) . Number . Uint64 ( ) , "to" , low )
if err := bc . SetHead ( low ) ; err != nil {
return nil , err
all: integrate the freezer with fast sync
* all: freezer style syncing
core, eth, les, light: clean up freezer relative APIs
core, eth, les, trie, ethdb, light: clean a bit
core, eth, les, light: add unit tests
core, light: rewrite setHead function
core, eth: fix downloader unit tests
core: add receipt chain insertion test
core: use constant instead of hardcoding table name
core: fix rollback
core: fix setHead
core/rawdb: remove canonical block first and then iterate side chain
core/rawdb, ethdb: add hasAncient interface
eth/downloader: calculate ancient limit via cht first
core, eth, ethdb: lots of fixes
* eth/downloader: print ancient disable log only for fast sync
2019-04-25 17:59:48 +03:00
}
}
}
2020-08-20 13:01:24 +03:00
// The first thing the node will do is reconstruct the verification data for
// the head block (ethash cache or clique voting snapshot). Might as well do
// it in advance.
bc . engine . VerifyHeader ( bc , bc . CurrentHeader ( ) , true )
2015-04-20 13:58:17 +03:00
// Check the current state of the block hashes and make sure that we do not have any of the bad blocks in our chain
2017-01-06 17:52:03 +03:00
for hash := range BadHashes {
2016-04-05 16:22:04 +03:00
if header := bc . GetHeaderByHash ( hash ) ; header != nil {
2016-11-28 12:37:42 +03:00
// get the canonical block corresponding to the offending header's number
headerByNumber := bc . GetHeaderByNumber ( header . Number . Uint64 ( ) )
// make sure the headerByNumber (if present) is in our current canonical chain
if headerByNumber != nil && headerByNumber . Hash ( ) == header . Hash ( ) {
2017-02-28 14:35:17 +03:00
log . Error ( "Found bad hash, rewinding chain" , "number" , header . Number , "hash" , header . ParentHash )
2020-08-20 13:01:24 +03:00
if err := bc . SetHead ( header . Number . Uint64 ( ) - 1 ) ; err != nil {
return nil , err
}
2017-02-28 14:35:17 +03:00
log . Error ( "Chain rewind was successful, resuming normal operation" )
2016-11-28 12:37:42 +03:00
}
2015-04-20 13:58:17 +03:00
}
}
2021-10-07 16:47:50 +03:00
2019-08-06 13:40:28 +03:00
// Load any existing snapshot, regenerating it if loading failed
2020-01-19 22:57:56 +03:00
if bc . cacheConfig . SnapshotLimit > 0 {
2020-10-29 22:01:58 +03:00
// If the chain was rewound past the snapshot persistent layer (causing
// a recovery block number to be persisted to disk), check if we're still
// in recovery mode and in that case, don't invalidate the snapshot on a
// head mismatch.
var recover bool
head := bc . CurrentBlock ( )
if layer := rawdb . ReadSnapshotRecoveryNumber ( bc . db ) ; layer != nil && * layer > head . NumberU64 ( ) {
log . Warn ( "Enabling snapshot recovery" , "chainhead" , head . NumberU64 ( ) , "diskbase" , * layer )
recover = true
}
all: bloom-filter based pruning mechanism (#21724)
* cmd, core, tests: initial state pruner
core: fix db inspector
cmd/geth: add verify-state
cmd/geth: add verification tool
core/rawdb: implement flatdb
cmd, core: fix rebase
core/state: use new contract code layout
core/state/pruner: avoid deleting genesis state
cmd/geth: add helper function
core, cmd: fix extract genesis
core: minor fixes
contracts: remove useless
core/state/snapshot: plugin stacktrie
core: polish
core/state/snapshot: iterate storage concurrently
core/state/snapshot: fix iteration
core: add comments
core/state/snapshot: polish code
core/state: polish
core/state/snapshot: rebase
core/rawdb: add comments
core/rawdb: fix tests
core/rawdb: improve tests
core/state/snapshot: fix concurrent iteration
core/state: run pruning during the recovery
core, trie: implement martin's idea
core, eth: delete flatdb and polish pruner
trie: fix import
core/state/pruner: add log
core/state/pruner: fix issues
core/state/pruner: don't read back
core/state/pruner: fix contract code write
core/state/pruner: check root node presence
cmd, core: polish log
core/state: use HEAD-127 as the target
core/state/snapshot: improve tests
cmd/geth: fix verification tool
cmd/geth: use HEAD as the verification default target
all: replace the bloomfilter with martin's fork
cmd, core: polish code
core, cmd: forcibly delete state root
core/state/pruner: add hash64
core/state/pruner: fix blacklist
core/state: remove blacklist
cmd, core: delete trie clean cache before pruning
cmd, core: fix lint
cmd, core: fix rebase
core/state: fix the special case for clique networks
core/state/snapshot: remove useless code
core/state/pruner: capping the snapshot after pruning
cmd, core, eth: fixes
core/rawdb: update db inspector
cmd/geth: polish code
core/state/pruner: fsync bloom filter
cmd, core: print warning log
core/state/pruner: adjust the parameters for bloom filter
cmd, core: create the bloom filter by size
core: polish
core/state/pruner: sanitize invalid bloomfilter size
cmd: address comments
cmd/geth: address comments
cmd/geth: address comment
core/state/pruner: address comments
core/state/pruner: rename homedir to datadir
cmd, core: address comments
core/state/pruner: address comment
core/state: address comments
core, cmd, tests: address comments
core: address comments
core/state/pruner: release the iterator after each commit
core/state/pruner: improve pruner
cmd, core: adjust bloom paramters
core/state/pruner: fix lint
core/state/pruner: fix tests
core: fix rebase
core/state/pruner: remove atomic rename
core/state/pruner: address comments
all: run go mod tidy
core/state/pruner: avoid false-positive for the middle state roots
core/state/pruner: add checks for middle roots
cmd/geth: replace crit with error
* core/state/pruner: fix lint
* core: drop legacy bloom filter
* core/state/snapshot: improve pruner
* core/state/snapshot: polish concurrent logs to report ETA vs. hashes
* core/state/pruner: add progress report for pruning and compaction too
* core: fix snapshot test API
* core/state: fix some pruning logs
* core/state/pruner: support recovering from bloom flush fail
Co-authored-by: Péter Szilágyi <peterke@gmail.com>
2021-02-08 14:16:30 +03:00
bc . snaps , _ = snapshot . New ( bc . db , bc . stateCache . TrieDB ( ) , bc . cacheConfig . SnapshotLimit , head . Root ( ) , ! bc . cacheConfig . SnapshotWait , true , recover )
2020-01-19 22:57:56 +03:00
}
2021-10-07 16:47:50 +03:00
// Start future block processor.
bc . wg . Add ( 1 )
go bc . futureBlocksLoop ( )
// Start tx indexer/unindexer.
2020-05-11 18:58:43 +03:00
if txLookupLimit != nil {
bc . txLookupLimit = * txLookupLimit
2020-10-20 17:34:50 +03:00
bc . wg . Add ( 1 )
2020-05-11 18:58:43 +03:00
go bc . maintainTxIndex ( txIndexBlock )
}
2021-10-07 16:47:50 +03:00
2020-07-28 16:30:31 +03:00
// If periodic cache journal is required, spin it up.
if bc . cacheConfig . TrieCleanRejournal > 0 {
if bc . cacheConfig . TrieCleanRejournal < time . Minute {
log . Warn ( "Sanitizing invalid trie cache journal time" , "provided" , bc . cacheConfig . TrieCleanRejournal , "updated" , time . Minute )
bc . cacheConfig . TrieCleanRejournal = time . Minute
}
triedb := bc . stateCache . TrieDB ( )
bc . wg . Add ( 1 )
go func ( ) {
defer bc . wg . Done ( )
triedb . SaveCachePeriodically ( bc . cacheConfig . TrieCleanJournal , bc . cacheConfig . TrieCleanRejournal , bc . quit )
} ( )
}
2015-06-08 13:12:13 +03:00
return bc , nil
2015-02-18 14:14:21 +02:00
}
core, cmd, vendor: fixes and database inspection tool (#15)
* core, eth: some fixes for freezer
* vendor, core/rawdb, cmd/geth: add db inspector
* core, cmd/utils: check ancient store path forceily
* cmd/geth, common, core/rawdb: a few fixes
* cmd/geth: support windows file rename and fix rename error
* core: support ancient plugin
* core, cmd: streaming file copy
* cmd, consensus, core, tests: keep genesis in leveldb
* core: write txlookup during ancient init
* core: bump database version
2019-05-14 17:07:44 +03:00
// empty returns an indicator whether the blockchain is empty.
// Note, it's a special case that we connect a non-empty ancient
// database with an empty node, so that we can plugin the ancient
// into node seamlessly.
func ( bc * BlockChain ) empty ( ) bool {
genesis := bc . genesisBlock . Hash ( )
for _ , hash := range [ ] common . Hash { rawdb . ReadHeadBlockHash ( bc . db ) , rawdb . ReadHeadHeaderHash ( bc . db ) , rawdb . ReadHeadFastBlockHash ( bc . db ) } {
if hash != genesis {
return false
}
}
return true
}
2015-09-21 15:36:29 +03:00
// loadLastState loads the last known chain state from the database. This method
// assumes that the chain manager mutex is held.
2017-05-11 04:55:48 +03:00
func ( bc * BlockChain ) loadLastState ( ) error {
2015-09-21 15:36:29 +03:00
// Restore the last known head block
2018-05-07 14:35:06 +03:00
head := rawdb . ReadHeadBlockHash ( bc . db )
2015-09-21 15:36:29 +03:00
if head == ( common . Hash { } ) {
// Corrupt or empty database, init from scratch
2017-03-22 03:37:24 +03:00
log . Warn ( "Empty database, resetting chain" )
2017-05-11 04:55:48 +03:00
return bc . Reset ( )
2017-03-22 03:37:24 +03:00
}
// Make sure the entire head block is available
2017-05-11 04:55:48 +03:00
currentBlock := bc . GetBlockByHash ( head )
2017-03-22 03:37:24 +03:00
if currentBlock == nil {
// Corrupt or empty database, init from scratch
log . Warn ( "Head block missing, resetting chain" , "hash" , head )
2017-05-11 04:55:48 +03:00
return bc . Reset ( )
2017-03-22 03:37:24 +03:00
}
// Everything seems to be fine, set as the head block
2018-02-26 12:53:10 +03:00
bc . currentBlock . Store ( currentBlock )
2019-06-10 14:21:02 +03:00
headBlockGauge . Update ( int64 ( currentBlock . NumberU64 ( ) ) )
2017-03-22 03:37:24 +03:00
2015-09-21 15:36:29 +03:00
// Restore the last known head header
2018-02-26 12:53:10 +03:00
currentHeader := currentBlock . Header ( )
2018-05-07 14:35:06 +03:00
if head := rawdb . ReadHeadHeaderHash ( bc . db ) ; head != ( common . Hash { } ) {
2017-05-11 04:55:48 +03:00
if header := bc . GetHeaderByHash ( head ) ; header != nil {
2015-12-16 05:26:23 +02:00
currentHeader = header
2015-09-21 15:36:29 +03:00
}
}
2017-05-11 04:55:48 +03:00
bc . hc . SetCurrentHeader ( currentHeader )
2017-03-22 03:37:24 +03:00
2015-09-30 19:23:31 +03:00
// Restore the last known head fast block
2018-02-26 12:53:10 +03:00
bc . currentFastBlock . Store ( currentBlock )
2019-06-10 14:21:02 +03:00
headFastBlockGauge . Update ( int64 ( currentBlock . NumberU64 ( ) ) )
2018-05-07 14:35:06 +03:00
if head := rawdb . ReadHeadFastBlockHash ( bc . db ) ; head != ( common . Hash { } ) {
2017-05-11 04:55:48 +03:00
if block := bc . GetBlockByHash ( head ) ; block != nil {
2018-02-26 12:53:10 +03:00
bc . currentFastBlock . Store ( block )
2019-06-10 14:21:02 +03:00
headFastBlockGauge . Update ( int64 ( block . NumberU64 ( ) ) )
2015-09-30 19:23:31 +03:00
}
}
2016-09-22 22:04:58 +03:00
// Issue a status log for the user
2018-02-26 12:53:10 +03:00
currentFastBlock := bc . CurrentFastBlock ( )
2017-05-11 04:55:48 +03:00
headerTd := bc . GetTd ( currentHeader . Hash ( ) , currentHeader . Number . Uint64 ( ) )
2018-02-26 12:53:10 +03:00
blockTd := bc . GetTd ( currentBlock . Hash ( ) , currentBlock . NumberU64 ( ) )
fastTd := bc . GetTd ( currentFastBlock . Hash ( ) , currentFastBlock . NumberU64 ( ) )
2015-09-21 15:36:29 +03:00
2019-04-02 23:28:48 +03:00
log . Info ( "Loaded most recent local header" , "number" , currentHeader . Number , "hash" , currentHeader . Hash ( ) , "td" , headerTd , "age" , common . PrettyAge ( time . Unix ( int64 ( currentHeader . Time ) , 0 ) ) )
log . Info ( "Loaded most recent local full block" , "number" , currentBlock . Number ( ) , "hash" , currentBlock . Hash ( ) , "td" , blockTd , "age" , common . PrettyAge ( time . Unix ( int64 ( currentBlock . Time ( ) ) , 0 ) ) )
log . Info ( "Loaded most recent local fast block" , "number" , currentFastBlock . Number ( ) , "hash" , currentFastBlock . Hash ( ) , "td" , fastTd , "age" , common . PrettyAge ( time . Unix ( int64 ( currentFastBlock . Time ( ) ) , 0 ) ) )
2020-08-20 13:01:24 +03:00
if pivot := rawdb . ReadLastPivotNumber ( bc . db ) ; pivot != nil {
log . Info ( "Loaded last fast-sync pivot marker" , "number" , * pivot )
}
2015-09-21 15:36:29 +03:00
return nil
}
2020-08-20 13:01:24 +03:00
// SetHead rewinds the local chain to a new head. Depending on whether the node
// was fast synced or full synced and in which state, the method will try to
// delete minimal data from disk whilst retaining chain consistency.
2017-03-22 03:37:24 +03:00
func ( bc * BlockChain ) SetHead ( head uint64 ) error {
2020-10-29 22:01:58 +03:00
_ , err := bc . SetHeadBeyondRoot ( head , common . Hash { } )
return err
}
// SetHeadBeyondRoot rewinds the local chain to a new head with the extra condition
// that the rewind must pass the specified state root. This method is meant to be
2021-04-30 13:50:02 +03:00
// used when rewinding with snapshots enabled to ensure that we go back further than
2020-10-29 22:01:58 +03:00
// persistent disk layer. Depending on whether the node was fast synced or full, and
// in which state, the method will try to delete minimal data from disk whilst
// retaining chain consistency.
//
// The method returns the block number where the requested root cap was found.
func ( bc * BlockChain ) SetHeadBeyondRoot ( head uint64 , root common . Hash ) ( uint64 , error ) {
2021-10-07 16:47:50 +03:00
if ! bc . chainmu . TryLock ( ) {
return 0 , errChainStopped
}
2019-01-11 16:27:47 +03:00
defer bc . chainmu . Unlock ( )
2015-04-20 13:29:02 +03:00
2020-10-29 22:01:58 +03:00
// Track the block number of the requested root hash
var rootNumber uint64 // (no root == always 0)
2020-08-20 13:01:24 +03:00
// Retrieve the last pivot block to short circuit rollbacks beyond it and the
// current freezer limit to start nuking id underflown
pivot := rawdb . ReadLastPivotNumber ( bc . db )
frozen , _ := bc . db . Ancients ( )
updateFn := func ( db ethdb . KeyValueWriter , header * types . Header ) ( uint64 , bool ) {
// Rewind the block chain, ensuring we don't end up with a stateless head
// block. Note, depth equality is permitted to allow using SetHead as a
// chain reparation mechanism without deleting any data!
if currentBlock := bc . CurrentBlock ( ) ; currentBlock != nil && header . Number . Uint64 ( ) <= currentBlock . NumberU64 ( ) {
all: integrate the freezer with fast sync
* all: freezer style syncing
core, eth, les, light: clean up freezer relative APIs
core, eth, les, trie, ethdb, light: clean a bit
core, eth, les, light: add unit tests
core, light: rewrite setHead function
core, eth: fix downloader unit tests
core: add receipt chain insertion test
core: use constant instead of hardcoding table name
core: fix rollback
core: fix setHead
core/rawdb: remove canonical block first and then iterate side chain
core/rawdb, ethdb: add hasAncient interface
eth/downloader: calculate ancient limit via cht first
core, eth, ethdb: lots of fixes
* eth/downloader: print ancient disable log only for fast sync
2019-04-25 17:59:48 +03:00
newHeadBlock := bc . GetBlock ( header . Hash ( ) , header . Number . Uint64 ( ) )
if newHeadBlock == nil {
2020-08-20 13:01:24 +03:00
log . Error ( "Gap in the chain, rewinding to genesis" , "number" , header . Number , "hash" , header . Hash ( ) )
all: integrate the freezer with fast sync
* all: freezer style syncing
core, eth, les, light: clean up freezer relative APIs
core, eth, les, trie, ethdb, light: clean a bit
core, eth, les, light: add unit tests
core, light: rewrite setHead function
core, eth: fix downloader unit tests
core: add receipt chain insertion test
core: use constant instead of hardcoding table name
core: fix rollback
core: fix setHead
core/rawdb: remove canonical block first and then iterate side chain
core/rawdb, ethdb: add hasAncient interface
eth/downloader: calculate ancient limit via cht first
core, eth, ethdb: lots of fixes
* eth/downloader: print ancient disable log only for fast sync
2019-04-25 17:59:48 +03:00
newHeadBlock = bc . genesisBlock
} else {
2020-10-29 22:01:58 +03:00
// Block exists, keep rewinding until we find one with state,
// keeping rewinding until we exceed the optional threshold
// root hash
beyondRoot := ( root == common . Hash { } ) // Flag whether we're beyond the requested root (no root, always true)
2020-08-20 13:01:24 +03:00
for {
2020-10-29 22:01:58 +03:00
// If a root threshold was requested but not yet crossed, check
if root != ( common . Hash { } ) && ! beyondRoot && newHeadBlock . Root ( ) == root {
beyondRoot , rootNumber = true , newHeadBlock . NumberU64 ( )
}
2020-08-20 13:01:24 +03:00
if _ , err := state . New ( newHeadBlock . Root ( ) , bc . stateCache , bc . snaps ) ; err != nil {
log . Trace ( "Block state missing, rewinding further" , "number" , newHeadBlock . NumberU64 ( ) , "hash" , newHeadBlock . Hash ( ) )
if pivot == nil || newHeadBlock . NumberU64 ( ) > * pivot {
2021-01-25 16:29:45 +03:00
parent := bc . GetBlock ( newHeadBlock . ParentHash ( ) , newHeadBlock . NumberU64 ( ) - 1 )
if parent != nil {
newHeadBlock = parent
continue
}
log . Error ( "Missing block in the middle, aiming genesis" , "number" , newHeadBlock . NumberU64 ( ) - 1 , "hash" , newHeadBlock . ParentHash ( ) )
newHeadBlock = bc . genesisBlock
2020-08-20 13:01:24 +03:00
} else {
log . Trace ( "Rewind passed pivot, aiming genesis" , "number" , newHeadBlock . NumberU64 ( ) , "hash" , newHeadBlock . Hash ( ) , "pivot" , * pivot )
newHeadBlock = bc . genesisBlock
}
}
2020-10-29 22:01:58 +03:00
if beyondRoot || newHeadBlock . NumberU64 ( ) == 0 {
log . Debug ( "Rewound to block with state" , "number" , newHeadBlock . NumberU64 ( ) , "hash" , newHeadBlock . Hash ( ) )
break
}
log . Debug ( "Skipping block with threshold state" , "number" , newHeadBlock . NumberU64 ( ) , "hash" , newHeadBlock . Hash ( ) , "root" , newHeadBlock . Root ( ) )
newHeadBlock = bc . GetBlock ( newHeadBlock . ParentHash ( ) , newHeadBlock . NumberU64 ( ) - 1 ) // Keep rewinding
all: integrate the freezer with fast sync
* all: freezer style syncing
core, eth, les, light: clean up freezer relative APIs
core, eth, les, trie, ethdb, light: clean a bit
core, eth, les, light: add unit tests
core, light: rewrite setHead function
core, eth: fix downloader unit tests
core: add receipt chain insertion test
core: use constant instead of hardcoding table name
core: fix rollback
core: fix setHead
core/rawdb: remove canonical block first and then iterate side chain
core/rawdb, ethdb: add hasAncient interface
eth/downloader: calculate ancient limit via cht first
core, eth, ethdb: lots of fixes
* eth/downloader: print ancient disable log only for fast sync
2019-04-25 17:59:48 +03:00
}
}
rawdb . WriteHeadBlockHash ( db , newHeadBlock . Hash ( ) )
2020-01-17 13:49:32 +03:00
// Degrade the chain markers if they are explicitly reverted.
// In theory we should update all in-memory markers in the
// last step, however the direction of SetHead is from high
// to low, so it's safe the update in-memory markers directly.
all: integrate the freezer with fast sync
* all: freezer style syncing
core, eth, les, light: clean up freezer relative APIs
core, eth, les, trie, ethdb, light: clean a bit
core, eth, les, light: add unit tests
core, light: rewrite setHead function
core, eth: fix downloader unit tests
core: add receipt chain insertion test
core: use constant instead of hardcoding table name
core: fix rollback
core: fix setHead
core/rawdb: remove canonical block first and then iterate side chain
core/rawdb, ethdb: add hasAncient interface
eth/downloader: calculate ancient limit via cht first
core, eth, ethdb: lots of fixes
* eth/downloader: print ancient disable log only for fast sync
2019-04-25 17:59:48 +03:00
bc . currentBlock . Store ( newHeadBlock )
2019-06-10 14:21:02 +03:00
headBlockGauge . Update ( int64 ( newHeadBlock . NumberU64 ( ) ) )
all: integrate the freezer with fast sync
* all: freezer style syncing
core, eth, les, light: clean up freezer relative APIs
core, eth, les, trie, ethdb, light: clean a bit
core, eth, les, light: add unit tests
core, light: rewrite setHead function
core, eth: fix downloader unit tests
core: add receipt chain insertion test
core: use constant instead of hardcoding table name
core: fix rollback
core: fix setHead
core/rawdb: remove canonical block first and then iterate side chain
core/rawdb, ethdb: add hasAncient interface
eth/downloader: calculate ancient limit via cht first
core, eth, ethdb: lots of fixes
* eth/downloader: print ancient disable log only for fast sync
2019-04-25 17:59:48 +03:00
}
// Rewind the fast block in a simpleton way to the target head
if currentFastBlock := bc . CurrentFastBlock ( ) ; currentFastBlock != nil && header . Number . Uint64 ( ) < currentFastBlock . NumberU64 ( ) {
newHeadFastBlock := bc . GetBlock ( header . Hash ( ) , header . Number . Uint64 ( ) )
// If either blocks reached nil, reset to the genesis state
if newHeadFastBlock == nil {
newHeadFastBlock = bc . genesisBlock
}
rawdb . WriteHeadFastBlockHash ( db , newHeadFastBlock . Hash ( ) )
2020-01-17 13:49:32 +03:00
// Degrade the chain markers if they are explicitly reverted.
// In theory we should update all in-memory markers in the
// last step, however the direction of SetHead is from high
// to low, so it's safe the update in-memory markers directly.
all: integrate the freezer with fast sync
* all: freezer style syncing
core, eth, les, light: clean up freezer relative APIs
core, eth, les, trie, ethdb, light: clean a bit
core, eth, les, light: add unit tests
core, light: rewrite setHead function
core, eth: fix downloader unit tests
core: add receipt chain insertion test
core: use constant instead of hardcoding table name
core: fix rollback
core: fix setHead
core/rawdb: remove canonical block first and then iterate side chain
core/rawdb, ethdb: add hasAncient interface
eth/downloader: calculate ancient limit via cht first
core, eth, ethdb: lots of fixes
* eth/downloader: print ancient disable log only for fast sync
2019-04-25 17:59:48 +03:00
bc . currentFastBlock . Store ( newHeadFastBlock )
2019-06-10 14:21:02 +03:00
headFastBlockGauge . Update ( int64 ( newHeadFastBlock . NumberU64 ( ) ) )
all: integrate the freezer with fast sync
* all: freezer style syncing
core, eth, les, light: clean up freezer relative APIs
core, eth, les, trie, ethdb, light: clean a bit
core, eth, les, light: add unit tests
core, light: rewrite setHead function
core, eth: fix downloader unit tests
core: add receipt chain insertion test
core: use constant instead of hardcoding table name
core: fix rollback
core: fix setHead
core/rawdb: remove canonical block first and then iterate side chain
core/rawdb, ethdb: add hasAncient interface
eth/downloader: calculate ancient limit via cht first
core, eth, ethdb: lots of fixes
* eth/downloader: print ancient disable log only for fast sync
2019-04-25 17:59:48 +03:00
}
2020-08-20 13:01:24 +03:00
head := bc . CurrentBlock ( ) . NumberU64 ( )
// If setHead underflown the freezer threshold and the block processing
// intent afterwards is full block importing, delete the chain segment
// between the stateful-block and the sethead target.
var wipe bool
if head + 1 < frozen {
wipe = pivot == nil || head >= * pivot
}
return head , wipe // Only force wipe if full synced
all: integrate the freezer with fast sync
* all: freezer style syncing
core, eth, les, light: clean up freezer relative APIs
core, eth, les, trie, ethdb, light: clean a bit
core, eth, les, light: add unit tests
core, light: rewrite setHead function
core, eth: fix downloader unit tests
core: add receipt chain insertion test
core: use constant instead of hardcoding table name
core: fix rollback
core: fix setHead
core/rawdb: remove canonical block first and then iterate side chain
core/rawdb, ethdb: add hasAncient interface
eth/downloader: calculate ancient limit via cht first
core, eth, ethdb: lots of fixes
* eth/downloader: print ancient disable log only for fast sync
2019-04-25 17:59:48 +03:00
}
2017-03-22 03:37:24 +03:00
// Rewind the header chain, deleting all block bodies until then
all: integrate the freezer with fast sync
* all: freezer style syncing
core, eth, les, light: clean up freezer relative APIs
core, eth, les, trie, ethdb, light: clean a bit
core, eth, les, light: add unit tests
core, light: rewrite setHead function
core, eth: fix downloader unit tests
core: add receipt chain insertion test
core: use constant instead of hardcoding table name
core: fix rollback
core: fix setHead
core/rawdb: remove canonical block first and then iterate side chain
core/rawdb, ethdb: add hasAncient interface
eth/downloader: calculate ancient limit via cht first
core, eth, ethdb: lots of fixes
* eth/downloader: print ancient disable log only for fast sync
2019-04-25 17:59:48 +03:00
delFn := func ( db ethdb . KeyValueWriter , hash common . Hash , num uint64 ) {
// Ignore the error here since light client won't hit this path
frozen , _ := bc . db . Ancients ( )
if num + 1 <= frozen {
// Truncate all relative data(header, total difficulty, body, receipt
// and canonical hash) from ancient store.
2020-08-20 13:01:24 +03:00
if err := bc . db . TruncateAncients ( num ) ; err != nil {
core, cmd, vendor: fixes and database inspection tool (#15)
* core, eth: some fixes for freezer
* vendor, core/rawdb, cmd/geth: add db inspector
* core, cmd/utils: check ancient store path forceily
* cmd/geth, common, core/rawdb: a few fixes
* cmd/geth: support windows file rename and fix rename error
* core: support ancient plugin
* core, cmd: streaming file copy
* cmd, consensus, core, tests: keep genesis in leveldb
* core: write txlookup during ancient init
* core: bump database version
2019-05-14 17:07:44 +03:00
log . Crit ( "Failed to truncate ancient data" , "number" , num , "err" , err )
}
all: integrate the freezer with fast sync
* all: freezer style syncing
core, eth, les, light: clean up freezer relative APIs
core, eth, les, trie, ethdb, light: clean a bit
core, eth, les, light: add unit tests
core, light: rewrite setHead function
core, eth: fix downloader unit tests
core: add receipt chain insertion test
core: use constant instead of hardcoding table name
core: fix rollback
core: fix setHead
core/rawdb: remove canonical block first and then iterate side chain
core/rawdb, ethdb: add hasAncient interface
eth/downloader: calculate ancient limit via cht first
core, eth, ethdb: lots of fixes
* eth/downloader: print ancient disable log only for fast sync
2019-04-25 17:59:48 +03:00
// Remove the hash <-> number mapping from the active store.
rawdb . DeleteHeaderNumber ( db , hash )
} else {
// Remove relative body and receipts from the active store.
// The header, total difficulty and canonical hash will be
// removed in the hc.SetHead function.
rawdb . DeleteBody ( db , hash , num )
rawdb . DeleteReceipts ( db , hash , num )
}
// Todo(rjl493456442) txlookup, bloombits, etc
2015-09-30 19:23:31 +03:00
}
2020-08-20 13:01:24 +03:00
// If SetHead was only called as a chain reparation method, try to skip
// touching the header chain altogether, unless the freezer is broken
if block := bc . CurrentBlock ( ) ; block . NumberU64 ( ) == head {
if target , force := updateFn ( bc . db , block . Header ( ) ) ; force {
bc . hc . SetHead ( target , updateFn , delFn )
}
} else {
// Rewind the chain to the requested head and keep going backwards until a
// block with a state is found or fast sync pivot is passed
log . Warn ( "Rewinding blockchain" , "target" , head )
bc . hc . SetHead ( head , updateFn , delFn )
}
2015-09-21 15:36:29 +03:00
// Clear out any stale content from the caches
2015-08-31 20:21:02 +03:00
bc . bodyCache . Purge ( )
bc . bodyRLPCache . Purge ( )
2018-09-29 23:53:31 +03:00
bc . receiptsCache . Purge ( )
2015-08-31 20:21:02 +03:00
bc . blockCache . Purge ( )
2019-08-21 12:29:34 +03:00
bc . txLookupCache . Purge ( )
2015-08-31 20:21:02 +03:00
bc . futureBlocks . Purge ( )
2015-04-20 13:29:02 +03:00
2020-10-29 22:01:58 +03:00
return rootNumber , bc . loadLastState ( )
2015-04-20 13:29:02 +03:00
}
2015-10-05 19:37:56 +03:00
// FastSyncCommitHead sets the current head block to the one defined by the hash
// irrelevant what the chain contents were prior.
2017-05-11 04:55:48 +03:00
func ( bc * BlockChain ) FastSyncCommitHead ( hash common . Hash ) error {
2015-10-13 12:04:25 +03:00
// Make sure that both the block as well at its state trie exists
2017-05-11 04:55:48 +03:00
block := bc . GetBlockByHash ( hash )
2015-10-05 19:37:56 +03:00
if block == nil {
2021-04-15 20:35:00 +03:00
return fmt . Errorf ( "non existent block [%x..]" , hash [ : 4 ] )
2015-10-05 19:37:56 +03:00
}
2019-03-14 16:25:12 +03:00
if _ , err := trie . NewSecure ( block . Root ( ) , bc . stateCache . TrieDB ( ) ) ; err != nil {
2015-10-05 19:37:56 +03:00
return err
}
2021-10-07 16:47:50 +03:00
// If all checks out, manually set the head block.
if ! bc . chainmu . TryLock ( ) {
return errChainStopped
}
2018-02-26 12:53:10 +03:00
bc . currentBlock . Store ( block )
2019-06-10 14:21:02 +03:00
headBlockGauge . Update ( int64 ( block . NumberU64 ( ) ) )
2019-01-11 16:27:47 +03:00
bc . chainmu . Unlock ( )
2015-10-05 19:37:56 +03:00
2021-04-29 17:33:45 +03:00
// Destroy any existing state snapshot and regenerate it in the background,
// also resuming the normal maintenance of any previously paused snapshot.
2020-01-19 22:57:56 +03:00
if bc . snaps != nil {
bc . snaps . Rebuild ( block . Root ( ) )
}
2017-02-28 14:35:17 +03:00
log . Info ( "Committed new head block" , "number" , block . Number ( ) , "hash" , hash )
2015-10-05 19:37:56 +03:00
return nil
}
2015-08-31 20:21:02 +03:00
// Reset purges the entire blockchain, restoring it to its genesis state.
2017-03-22 03:37:24 +03:00
func ( bc * BlockChain ) Reset ( ) error {
return bc . ResetWithGenesisBlock ( bc . genesisBlock )
2015-03-13 19:29:42 +02:00
}
2015-08-31 20:21:02 +03:00
// ResetWithGenesisBlock purges the entire blockchain, restoring it to the
// specified genesis state.
2017-03-22 03:37:24 +03:00
func ( bc * BlockChain ) ResetWithGenesisBlock ( genesis * types . Block ) error {
2015-09-30 19:23:31 +03:00
// Dump the entire block chain and purge the caches
2017-03-22 03:37:24 +03:00
if err := bc . SetHead ( 0 ) ; err != nil {
return err
}
2021-10-07 16:47:50 +03:00
if ! bc . chainmu . TryLock ( ) {
return errChainStopped
}
2019-01-11 16:27:47 +03:00
defer bc . chainmu . Unlock ( )
2015-03-03 19:41:51 +02:00
2015-10-13 12:04:25 +03:00
// Prepare the genesis block and reinitialise the chain
2020-01-17 13:49:32 +03:00
batch := bc . db . NewBatch ( )
rawdb . WriteTd ( batch , genesis . Hash ( ) , genesis . NumberU64 ( ) , genesis . Difficulty ( ) )
rawdb . WriteBlock ( batch , genesis )
if err := batch . Write ( ) ; err != nil {
log . Crit ( "Failed to write genesis block" , "err" , err )
2015-09-07 20:43:01 +03:00
}
2020-01-17 13:49:32 +03:00
bc . writeHeadBlock ( genesis )
2018-05-07 14:35:06 +03:00
2020-01-17 13:49:32 +03:00
// Last update all in-memory chain markers
2015-09-23 00:55:31 +03:00
bc . genesisBlock = genesis
2018-02-26 12:53:10 +03:00
bc . currentBlock . Store ( bc . genesisBlock )
2019-06-10 14:21:02 +03:00
headBlockGauge . Update ( int64 ( bc . genesisBlock . NumberU64 ( ) ) )
2015-12-16 05:26:23 +02:00
bc . hc . SetGenesis ( bc . genesisBlock . Header ( ) )
bc . hc . SetCurrentHeader ( bc . genesisBlock . Header ( ) )
2018-02-26 12:53:10 +03:00
bc . currentFastBlock . Store ( bc . genesisBlock )
2019-06-10 14:21:02 +03:00
headFastBlockGauge . Update ( int64 ( bc . genesisBlock . NumberU64 ( ) ) )
2017-03-22 03:37:24 +03:00
return nil
2015-03-03 19:41:51 +02:00
}
2015-03-18 14:36:48 +02:00
// Export writes the active chain to the given writer.
2017-05-11 04:55:48 +03:00
func ( bc * BlockChain ) Export ( w io . Writer ) error {
2018-02-26 12:53:10 +03:00
return bc . ExportN ( w , uint64 ( 0 ) , bc . CurrentBlock ( ) . NumberU64 ( ) )
2015-06-06 06:01:54 +03:00
}
// ExportN writes a subset of the active chain to the given writer.
2017-05-11 04:55:48 +03:00
func ( bc * BlockChain ) ExportN ( w io . Writer , first uint64 , last uint64 ) error {
2021-10-07 16:47:50 +03:00
if ! bc . chainmu . TryLock ( ) {
return errChainStopped
}
defer bc . chainmu . Unlock ( )
2015-04-05 00:04:19 +03:00
2015-06-06 06:01:54 +03:00
if first > last {
return fmt . Errorf ( "export failed: first (%d) is greater than last (%d)" , first , last )
}
2017-02-28 14:35:17 +03:00
log . Info ( "Exporting batch of blocks" , "count" , last - first + 1 )
2015-04-13 11:13:52 +03:00
2018-07-26 14:26:24 +03:00
start , reported := time . Now ( ) , time . Now ( )
2015-06-06 06:01:54 +03:00
for nr := first ; nr <= last ; nr ++ {
2017-05-11 04:55:48 +03:00
block := bc . GetBlockByNumber ( nr )
2015-04-20 17:02:50 +03:00
if block == nil {
return fmt . Errorf ( "export failed on #%d: not found" , nr )
}
if err := block . EncodeRLP ( w ) ; err != nil {
2015-03-18 14:36:48 +02:00
return err
}
2018-07-26 14:26:24 +03:00
if time . Since ( reported ) >= statsReportLimit {
log . Info ( "Exporting blocks" , "exported" , block . NumberU64 ( ) - first , "elapsed" , common . PrettyDuration ( time . Since ( start ) ) )
reported = time . Now ( )
}
2014-12-17 13:57:35 +02:00
}
2015-03-18 14:36:48 +02:00
return nil
2014-12-17 13:57:35 +02:00
}
2020-01-17 13:49:32 +03:00
// writeHeadBlock injects a new head block into the current block chain. This method
2015-09-21 15:36:29 +03:00
// assumes that the block is indeed a true head. It will also reset the head
2015-10-28 18:40:24 +03:00
// header and the head fast sync block to this very same block if they are older
// or if they are on a different side chain.
2015-09-21 15:36:29 +03:00
//
// Note, this function assumes that the `mu` mutex is held!
2020-01-17 13:49:32 +03:00
func ( bc * BlockChain ) writeHeadBlock ( block * types . Block ) {
2015-10-28 18:40:24 +03:00
// If the block is on a side chain or an unknown one, force other heads onto it too
2018-05-07 14:35:06 +03:00
updateHeads := rawdb . ReadCanonicalHash ( bc . db , block . NumberU64 ( ) ) != block . Hash ( )
2015-10-28 18:40:24 +03:00
2015-09-07 20:43:01 +03:00
// Add the block to the canonical chain number scheme and mark as the head
2020-01-17 13:49:32 +03:00
batch := bc . db . NewBatch ( )
rawdb . WriteCanonicalHash ( batch , block . Hash ( ) , block . NumberU64 ( ) )
2020-09-15 11:37:01 +03:00
rawdb . WriteTxLookupEntriesByBlock ( batch , block )
2020-01-17 13:49:32 +03:00
rawdb . WriteHeadBlockHash ( batch , block . Hash ( ) )
2015-10-28 18:40:24 +03:00
2018-01-22 15:07:47 +03:00
// If the block is better than our head or is on a different chain, force update heads
2020-01-17 13:49:32 +03:00
if updateHeads {
rawdb . WriteHeadHeaderHash ( batch , block . Hash ( ) )
rawdb . WriteHeadFastBlockHash ( batch , block . Hash ( ) )
}
// Flush the whole batch into the disk, exit the node if failed
if err := batch . Write ( ) ; err != nil {
log . Crit ( "Failed to update chain indexes and markers" , "err" , err )
}
// Update all in-memory chain markers in the last step
2015-10-28 18:40:24 +03:00
if updateHeads {
2015-12-16 05:26:23 +02:00
bc . hc . SetCurrentHeader ( block . Header ( ) )
2018-02-26 12:53:10 +03:00
bc . currentFastBlock . Store ( block )
2019-06-10 14:21:02 +03:00
headFastBlockGauge . Update ( int64 ( block . NumberU64 ( ) ) )
2015-10-28 18:40:24 +03:00
}
2020-01-17 13:49:32 +03:00
bc . currentBlock . Store ( block )
headBlockGauge . Update ( int64 ( block . NumberU64 ( ) ) )
2014-12-05 17:26:39 +02:00
}
2015-10-19 17:08:17 +03:00
// Stop stops the blockchain service. If any imports are currently in progress
// it will abort them using the procInterrupt.
2015-08-31 18:09:50 +03:00
func ( bc * BlockChain ) Stop ( ) {
2015-08-01 13:32:28 +03:00
if ! atomic . CompareAndSwapInt32 ( & bc . running , 0 , 1 ) {
return
}
2021-10-07 16:47:50 +03:00
// Unsubscribe all subscriptions registered from blockchain.
2017-08-18 13:58:36 +03:00
bc . scope . Close ( )
2021-10-07 16:47:50 +03:00
// Signal shutdown to all goroutines.
2015-03-06 16:50:44 +02:00
close ( bc . quit )
2020-05-26 22:37:37 +03:00
bc . StopInsert ( )
2021-10-07 16:47:50 +03:00
// Now wait for all chain modifications to end and persistent goroutines to exit.
//
// Note: Close waits for the mutex to become available, i.e. any running chain
// modification will have exited when Close returns. Since we also called StopInsert,
// the mutex should become available quickly. It cannot be taken again after Close has
// returned.
bc . chainmu . Close ( )
2015-04-30 18:50:23 +03:00
bc . wg . Wait ( )
2018-02-05 19:40:32 +03:00
2019-08-06 13:40:28 +03:00
// Ensure that the entirety of the state snapshot is journalled to disk.
2020-01-19 22:57:56 +03:00
var snapBase common . Hash
if bc . snaps != nil {
var err error
2021-04-20 08:27:46 +03:00
if snapBase , err = bc . snaps . Journal ( bc . CurrentBlock ( ) . Root ( ) ) ; err != nil {
log . Error ( "Failed to journal state snapshot" , "err" , err )
2020-01-19 22:57:56 +03:00
}
2019-08-06 13:40:28 +03:00
}
2021-10-07 16:47:50 +03:00
2018-02-05 19:40:32 +03:00
// Ensure the state of a recent block is also stored to disk before exiting.
2018-02-23 15:02:33 +03:00
// We're writing three different states to catch different restart scenarios:
// - HEAD: So we don't need to reprocess any blocks in the general case
// - HEAD-1: So we don't do large reorgs if our HEAD becomes an uncle
// - HEAD-127: So we have a hard limit on the number of blocks reexecuted
2019-04-01 11:52:11 +03:00
if ! bc . cacheConfig . TrieDirtyDisabled {
2018-02-05 19:40:32 +03:00
triedb := bc . stateCache . TrieDB ( )
2019-05-30 21:51:13 +03:00
for _ , offset := range [ ] uint64 { 0 , 1 , TriesInMemory - 1 } {
2018-02-23 15:02:33 +03:00
if number := bc . CurrentBlock ( ) . NumberU64 ( ) ; number > offset {
recent := bc . GetBlockByNumber ( number - offset )
log . Info ( "Writing cached state to disk" , "block" , recent . Number ( ) , "hash" , recent . Hash ( ) , "root" , recent . Root ( ) )
2020-07-13 12:02:54 +03:00
if err := triedb . Commit ( recent . Root ( ) , true , nil ) ; err != nil {
2018-02-23 15:02:33 +03:00
log . Error ( "Failed to commit recent state trie" , "err" , err )
}
2018-02-05 19:40:32 +03:00
}
}
2019-11-26 10:48:29 +03:00
if snapBase != ( common . Hash { } ) {
log . Info ( "Writing snapshot state to disk" , "root" , snapBase )
2020-07-13 12:02:54 +03:00
if err := triedb . Commit ( snapBase , true , nil ) ; err != nil {
2019-11-26 10:48:29 +03:00
log . Error ( "Failed to commit recent state trie" , "err" , err )
}
}
2018-02-05 19:40:32 +03:00
for ! bc . triegc . Empty ( ) {
2018-06-21 12:28:05 +03:00
triedb . Dereference ( bc . triegc . PopItem ( ) . ( common . Hash ) )
2018-02-05 19:40:32 +03:00
}
2018-06-04 10:47:43 +03:00
if size , _ := triedb . Size ( ) ; size != 0 {
2018-02-05 19:40:32 +03:00
log . Error ( "Dangling trie nodes after full cleanup" )
}
}
2020-07-28 16:30:31 +03:00
// Ensure all live cached entries be saved into disk, so that we can skip
// cache warmup when node restarts.
if bc . cacheConfig . TrieCleanJournal != "" {
triedb := bc . stateCache . TrieDB ( )
triedb . SaveCache ( bc . cacheConfig . TrieCleanJournal )
}
2020-03-27 16:03:20 +03:00
log . Info ( "Blockchain stopped" )
2015-03-06 16:50:44 +02:00
}
2020-05-26 22:37:37 +03:00
// StopInsert interrupts all insertion methods, causing them to return
// errInsertionInterrupted as soon as possible. Insertion is permanently disabled after
// calling this method.
func ( bc * BlockChain ) StopInsert ( ) {
atomic . StoreInt32 ( & bc . procInterrupt , 1 )
}
// insertStopped returns true after StopInsert has been called.
func ( bc * BlockChain ) insertStopped ( ) bool {
return atomic . LoadInt32 ( & bc . procInterrupt ) == 1
}
2017-05-11 04:55:48 +03:00
func ( bc * BlockChain ) procFutureBlocks ( ) {
blocks := make ( [ ] * types . Block , 0 , bc . futureBlocks . Len ( ) )
for _ , hash := range bc . futureBlocks . Keys ( ) {
if block , exist := bc . futureBlocks . Peek ( hash ) ; exist {
2016-03-08 16:55:27 +02:00
blocks = append ( blocks , block . ( * types . Block ) )
}
2015-06-29 23:42:13 +03:00
}
2015-05-29 19:55:42 +03:00
if len ( blocks ) > 0 {
2019-11-21 17:35:22 +03:00
sort . Slice ( blocks , func ( i , j int ) bool {
return blocks [ i ] . NumberU64 ( ) < blocks [ j ] . NumberU64 ( )
} )
2016-12-13 17:14:33 +03:00
// Insert one by one as chain insertion needs contiguous ancestry between blocks
for i := range blocks {
2017-05-11 04:55:48 +03:00
bc . InsertChain ( blocks [ i : i + 1 ] )
2016-12-13 17:14:33 +03:00
}
2015-05-29 19:55:42 +03:00
}
2015-04-04 17:35:23 +03:00
}
2017-05-11 04:55:48 +03:00
// WriteStatus status of write
2015-12-16 05:26:23 +02:00
type WriteStatus byte
2015-06-29 13:12:30 +03:00
const (
2015-12-16 05:26:23 +02:00
NonStatTy WriteStatus = iota
2015-07-03 12:24:42 +03:00
CanonStatTy
SideStatTy
2015-06-29 13:12:30 +03:00
)
2019-07-08 11:24:16 +03:00
// numberHash is just a container for a number and a hash, to represent a block
type numberHash struct {
number uint64
hash common . Hash
}
2015-09-30 19:23:31 +03:00
// InsertReceiptChain attempts to complete an already existing header chain with
// transaction and receipt data.
all: integrate the freezer with fast sync
* all: freezer style syncing
core, eth, les, light: clean up freezer relative APIs
core, eth, les, trie, ethdb, light: clean a bit
core, eth, les, light: add unit tests
core, light: rewrite setHead function
core, eth: fix downloader unit tests
core: add receipt chain insertion test
core: use constant instead of hardcoding table name
core: fix rollback
core: fix setHead
core/rawdb: remove canonical block first and then iterate side chain
core/rawdb, ethdb: add hasAncient interface
eth/downloader: calculate ancient limit via cht first
core, eth, ethdb: lots of fixes
* eth/downloader: print ancient disable log only for fast sync
2019-04-25 17:59:48 +03:00
func ( bc * BlockChain ) InsertReceiptChain ( blockChain types . Blocks , receiptChain [ ] types . Receipts , ancientLimit uint64 ) ( int , error ) {
2019-07-03 11:19:15 +03:00
// We don't require the chainMu here since we want to maximize the
// concurrency of header insertion and receipt insertion.
2017-09-09 19:03:07 +03:00
bc . wg . Add ( 1 )
defer bc . wg . Done ( )
all: integrate the freezer with fast sync
* all: freezer style syncing
core, eth, les, light: clean up freezer relative APIs
core, eth, les, trie, ethdb, light: clean a bit
core, eth, les, light: add unit tests
core, light: rewrite setHead function
core, eth: fix downloader unit tests
core: add receipt chain insertion test
core: use constant instead of hardcoding table name
core: fix rollback
core: fix setHead
core/rawdb: remove canonical block first and then iterate side chain
core/rawdb, ethdb: add hasAncient interface
eth/downloader: calculate ancient limit via cht first
core, eth, ethdb: lots of fixes
* eth/downloader: print ancient disable log only for fast sync
2019-04-25 17:59:48 +03:00
var (
ancientBlocks , liveBlocks types . Blocks
ancientReceipts , liveReceipts [ ] types . Receipts
)
2016-12-13 17:14:33 +03:00
// Do a sanity check that the provided chain is actually ordered and linked
all: integrate the freezer with fast sync
* all: freezer style syncing
core, eth, les, light: clean up freezer relative APIs
core, eth, les, trie, ethdb, light: clean a bit
core, eth, les, light: add unit tests
core, light: rewrite setHead function
core, eth: fix downloader unit tests
core: add receipt chain insertion test
core: use constant instead of hardcoding table name
core: fix rollback
core: fix setHead
core/rawdb: remove canonical block first and then iterate side chain
core/rawdb, ethdb: add hasAncient interface
eth/downloader: calculate ancient limit via cht first
core, eth, ethdb: lots of fixes
* eth/downloader: print ancient disable log only for fast sync
2019-04-25 17:59:48 +03:00
for i := 0 ; i < len ( blockChain ) ; i ++ {
if i != 0 {
if blockChain [ i ] . NumberU64 ( ) != blockChain [ i - 1 ] . NumberU64 ( ) + 1 || blockChain [ i ] . ParentHash ( ) != blockChain [ i - 1 ] . Hash ( ) {
log . Error ( "Non contiguous receipt insert" , "number" , blockChain [ i ] . Number ( ) , "hash" , blockChain [ i ] . Hash ( ) , "parent" , blockChain [ i ] . ParentHash ( ) ,
"prevnumber" , blockChain [ i - 1 ] . Number ( ) , "prevhash" , blockChain [ i - 1 ] . Hash ( ) )
2021-04-15 20:35:00 +03:00
return 0 , fmt . Errorf ( "non contiguous insert: item %d is #%d [%x..], item %d is #%d [%x..] (parent [%x..])" , i - 1 , blockChain [ i - 1 ] . NumberU64 ( ) ,
all: integrate the freezer with fast sync
* all: freezer style syncing
core, eth, les, light: clean up freezer relative APIs
core, eth, les, trie, ethdb, light: clean a bit
core, eth, les, light: add unit tests
core, light: rewrite setHead function
core, eth: fix downloader unit tests
core: add receipt chain insertion test
core: use constant instead of hardcoding table name
core: fix rollback
core: fix setHead
core/rawdb: remove canonical block first and then iterate side chain
core/rawdb, ethdb: add hasAncient interface
eth/downloader: calculate ancient limit via cht first
core, eth, ethdb: lots of fixes
* eth/downloader: print ancient disable log only for fast sync
2019-04-25 17:59:48 +03:00
blockChain [ i - 1 ] . Hash ( ) . Bytes ( ) [ : 4 ] , i , blockChain [ i ] . NumberU64 ( ) , blockChain [ i ] . Hash ( ) . Bytes ( ) [ : 4 ] , blockChain [ i ] . ParentHash ( ) . Bytes ( ) [ : 4 ] )
}
}
if blockChain [ i ] . NumberU64 ( ) <= ancientLimit {
ancientBlocks , ancientReceipts = append ( ancientBlocks , blockChain [ i ] ) , append ( ancientReceipts , receiptChain [ i ] )
} else {
liveBlocks , liveReceipts = append ( liveBlocks , blockChain [ i ] ) , append ( liveReceipts , receiptChain [ i ] )
2016-12-13 17:14:33 +03:00
}
}
2015-09-30 19:23:31 +03:00
2017-09-09 19:03:07 +03:00
var (
stats = struct { processed , ignored int32 } { }
start = time . Now ( )
2021-09-07 13:31:17 +03:00
size = int64 ( 0 )
2017-09-09 19:03:07 +03:00
)
2021-09-07 13:31:17 +03:00
all: integrate the freezer with fast sync
* all: freezer style syncing
core, eth, les, light: clean up freezer relative APIs
core, eth, les, trie, ethdb, light: clean a bit
core, eth, les, light: add unit tests
core, light: rewrite setHead function
core, eth: fix downloader unit tests
core: add receipt chain insertion test
core: use constant instead of hardcoding table name
core: fix rollback
core: fix setHead
core/rawdb: remove canonical block first and then iterate side chain
core/rawdb, ethdb: add hasAncient interface
eth/downloader: calculate ancient limit via cht first
core, eth, ethdb: lots of fixes
* eth/downloader: print ancient disable log only for fast sync
2019-04-25 17:59:48 +03:00
// updateHead updates the head fast sync block if the inserted blocks are better
2020-05-25 11:21:28 +03:00
// and returns an indicator whether the inserted blocks are canonical.
all: integrate the freezer with fast sync
* all: freezer style syncing
core, eth, les, light: clean up freezer relative APIs
core, eth, les, trie, ethdb, light: clean a bit
core, eth, les, light: add unit tests
core, light: rewrite setHead function
core, eth: fix downloader unit tests
core: add receipt chain insertion test
core: use constant instead of hardcoding table name
core: fix rollback
core: fix setHead
core/rawdb: remove canonical block first and then iterate side chain
core/rawdb, ethdb: add hasAncient interface
eth/downloader: calculate ancient limit via cht first
core, eth, ethdb: lots of fixes
* eth/downloader: print ancient disable log only for fast sync
2019-04-25 17:59:48 +03:00
updateHead := func ( head * types . Block ) bool {
2021-10-07 16:47:50 +03:00
if ! bc . chainmu . TryLock ( ) {
return false
}
2021-09-07 13:31:17 +03:00
defer bc . chainmu . Unlock ( )
2019-07-03 11:19:15 +03:00
// Rewind may have occurred, skip in that case.
if bc . CurrentHeader ( ) . Number . Cmp ( head . Number ( ) ) >= 0 {
currentFastBlock , td := bc . CurrentFastBlock ( ) , bc . GetTd ( head . Hash ( ) , head . NumberU64 ( ) )
all: integrate the freezer with fast sync
* all: freezer style syncing
core, eth, les, light: clean up freezer relative APIs
core, eth, les, trie, ethdb, light: clean a bit
core, eth, les, light: add unit tests
core, light: rewrite setHead function
core, eth: fix downloader unit tests
core: add receipt chain insertion test
core: use constant instead of hardcoding table name
core: fix rollback
core: fix setHead
core/rawdb: remove canonical block first and then iterate side chain
core/rawdb, ethdb: add hasAncient interface
eth/downloader: calculate ancient limit via cht first
core, eth, ethdb: lots of fixes
* eth/downloader: print ancient disable log only for fast sync
2019-04-25 17:59:48 +03:00
if bc . GetTd ( currentFastBlock . Hash ( ) , currentFastBlock . NumberU64 ( ) ) . Cmp ( td ) < 0 {
rawdb . WriteHeadFastBlockHash ( bc . db , head . Hash ( ) )
bc . currentFastBlock . Store ( head )
2019-06-10 14:21:02 +03:00
headFastBlockGauge . Update ( int64 ( head . NumberU64 ( ) ) )
2019-07-03 11:19:15 +03:00
return true
all: integrate the freezer with fast sync
* all: freezer style syncing
core, eth, les, light: clean up freezer relative APIs
core, eth, les, trie, ethdb, light: clean a bit
core, eth, les, light: add unit tests
core, light: rewrite setHead function
core, eth: fix downloader unit tests
core: add receipt chain insertion test
core: use constant instead of hardcoding table name
core: fix rollback
core: fix setHead
core/rawdb: remove canonical block first and then iterate side chain
core/rawdb, ethdb: add hasAncient interface
eth/downloader: calculate ancient limit via cht first
core, eth, ethdb: lots of fixes
* eth/downloader: print ancient disable log only for fast sync
2019-04-25 17:59:48 +03:00
}
2017-09-09 19:03:07 +03:00
}
2019-07-03 11:19:15 +03:00
return false
all: integrate the freezer with fast sync
* all: freezer style syncing
core, eth, les, light: clean up freezer relative APIs
core, eth, les, trie, ethdb, light: clean a bit
core, eth, les, light: add unit tests
core, light: rewrite setHead function
core, eth: fix downloader unit tests
core: add receipt chain insertion test
core: use constant instead of hardcoding table name
core: fix rollback
core: fix setHead
core/rawdb: remove canonical block first and then iterate side chain
core/rawdb, ethdb: add hasAncient interface
eth/downloader: calculate ancient limit via cht first
core, eth, ethdb: lots of fixes
* eth/downloader: print ancient disable log only for fast sync
2019-04-25 17:59:48 +03:00
}
2021-09-07 13:31:17 +03:00
all: integrate the freezer with fast sync
* all: freezer style syncing
core, eth, les, light: clean up freezer relative APIs
core, eth, les, trie, ethdb, light: clean a bit
core, eth, les, light: add unit tests
core, light: rewrite setHead function
core, eth: fix downloader unit tests
core: add receipt chain insertion test
core: use constant instead of hardcoding table name
core: fix rollback
core: fix setHead
core/rawdb: remove canonical block first and then iterate side chain
core/rawdb, ethdb: add hasAncient interface
eth/downloader: calculate ancient limit via cht first
core, eth, ethdb: lots of fixes
* eth/downloader: print ancient disable log only for fast sync
2019-04-25 17:59:48 +03:00
// writeAncient writes blockchain and corresponding receipt chain into ancient store.
//
// this function only accepts canonical chain data. All side chain will be reverted
// eventually.
writeAncient := func ( blockChain types . Blocks , receiptChain [ ] types . Receipts ) ( int , error ) {
2021-09-07 13:31:17 +03:00
first := blockChain [ 0 ]
last := blockChain [ len ( blockChain ) - 1 ]
// Ensure genesis is in ancients.
if first . NumberU64 ( ) == 1 {
if frozen , _ := bc . db . Ancients ( ) ; frozen == 0 {
b := bc . genesisBlock
td := bc . genesisBlock . Difficulty ( )
writeSize , err := rawdb . WriteAncientBlocks ( bc . db , [ ] * types . Block { b } , [ ] types . Receipts { nil } , td )
size += writeSize
if err != nil {
log . Error ( "Error writing genesis to ancients" , "err" , err )
return 0 , err
core, cmd, vendor: fixes and database inspection tool (#15)
* core, eth: some fixes for freezer
* vendor, core/rawdb, cmd/geth: add db inspector
* core, cmd/utils: check ancient store path forceily
* cmd/geth, common, core/rawdb: a few fixes
* cmd/geth: support windows file rename and fix rename error
* core: support ancient plugin
* core, cmd: streaming file copy
* cmd, consensus, core, tests: keep genesis in leveldb
* core: write txlookup during ancient init
* core: bump database version
2019-05-14 17:07:44 +03:00
}
2021-09-07 13:31:17 +03:00
log . Info ( "Wrote genesis to ancients" )
all: integrate the freezer with fast sync
* all: freezer style syncing
core, eth, les, light: clean up freezer relative APIs
core, eth, les, trie, ethdb, light: clean a bit
core, eth, les, light: add unit tests
core, light: rewrite setHead function
core, eth: fix downloader unit tests
core: add receipt chain insertion test
core: use constant instead of hardcoding table name
core: fix rollback
core: fix setHead
core/rawdb: remove canonical block first and then iterate side chain
core/rawdb, ethdb: add hasAncient interface
eth/downloader: calculate ancient limit via cht first
core, eth, ethdb: lots of fixes
* eth/downloader: print ancient disable log only for fast sync
2019-04-25 17:59:48 +03:00
}
2021-09-07 13:31:17 +03:00
}
// Before writing the blocks to the ancients, we need to ensure that
// they correspond to the what the headerchain 'expects'.
// We only check the last block/header, since it's a contiguous chain.
if ! bc . HasHeader ( last . Hash ( ) , last . NumberU64 ( ) ) {
return 0 , fmt . Errorf ( "containing header #%d [%x..] unknown" , last . Number ( ) , last . Hash ( ) . Bytes ( ) [ : 4 ] )
}
// Write all chain data to ancients.
td := bc . GetTd ( first . Hash ( ) , first . NumberU64 ( ) )
writeSize , err := rawdb . WriteAncientBlocks ( bc . db , blockChain , receiptChain , td )
size += writeSize
if err != nil {
log . Error ( "Error importing chain data to ancients" , "err" , err )
return 0 , err
}
// Write tx indices if any condition is satisfied:
// * If user requires to reserve all tx indices(txlookuplimit=0)
// * If all ancient tx indices are required to be reserved(txlookuplimit is even higher than ancientlimit)
// * If block number is large enough to be regarded as a recent block
// It means blocks below the ancientLimit-txlookupLimit won't be indexed.
//
// But if the `TxIndexTail` is not nil, e.g. Geth is initialized with
// an external ancient database, during the setup, blockchain will start
// a background routine to re-indexed all indices in [ancients - txlookupLimit, ancients)
// range. In this case, all tx indices of newly imported blocks should be
// generated.
var batch = bc . db . NewBatch ( )
for _ , block := range blockChain {
2020-05-11 18:58:43 +03:00
if bc . txLookupLimit == 0 || ancientLimit <= bc . txLookupLimit || block . NumberU64 ( ) >= ancientLimit - bc . txLookupLimit {
2020-09-15 11:37:01 +03:00
rawdb . WriteTxLookupEntriesByBlock ( batch , block )
2020-05-11 18:58:43 +03:00
} else if rawdb . ReadTxIndexTail ( bc . db ) != nil {
2020-09-15 11:37:01 +03:00
rawdb . WriteTxLookupEntriesByBlock ( batch , block )
2020-05-11 18:58:43 +03:00
}
all: integrate the freezer with fast sync
* all: freezer style syncing
core, eth, les, light: clean up freezer relative APIs
core, eth, les, trie, ethdb, light: clean a bit
core, eth, les, light: add unit tests
core, light: rewrite setHead function
core, eth: fix downloader unit tests
core: add receipt chain insertion test
core: use constant instead of hardcoding table name
core: fix rollback
core: fix setHead
core/rawdb: remove canonical block first and then iterate side chain
core/rawdb, ethdb: add hasAncient interface
eth/downloader: calculate ancient limit via cht first
core, eth, ethdb: lots of fixes
* eth/downloader: print ancient disable log only for fast sync
2019-04-25 17:59:48 +03:00
stats . processed ++
2017-09-09 19:03:07 +03:00
}
2021-09-07 13:31:17 +03:00
all: integrate the freezer with fast sync
* all: freezer style syncing
core, eth, les, light: clean up freezer relative APIs
core, eth, les, trie, ethdb, light: clean a bit
core, eth, les, light: add unit tests
core, light: rewrite setHead function
core, eth: fix downloader unit tests
core: add receipt chain insertion test
core: use constant instead of hardcoding table name
core: fix rollback
core: fix setHead
core/rawdb: remove canonical block first and then iterate side chain
core/rawdb, ethdb: add hasAncient interface
eth/downloader: calculate ancient limit via cht first
core, eth, ethdb: lots of fixes
* eth/downloader: print ancient disable log only for fast sync
2019-04-25 17:59:48 +03:00
// Flush all tx-lookup index data.
2021-09-07 13:31:17 +03:00
size += int64 ( batch . ValueSize ( ) )
all: integrate the freezer with fast sync
* all: freezer style syncing
core, eth, les, light: clean up freezer relative APIs
core, eth, les, trie, ethdb, light: clean a bit
core, eth, les, light: add unit tests
core, light: rewrite setHead function
core, eth: fix downloader unit tests
core: add receipt chain insertion test
core: use constant instead of hardcoding table name
core: fix rollback
core: fix setHead
core/rawdb: remove canonical block first and then iterate side chain
core/rawdb, ethdb: add hasAncient interface
eth/downloader: calculate ancient limit via cht first
core, eth, ethdb: lots of fixes
* eth/downloader: print ancient disable log only for fast sync
2019-04-25 17:59:48 +03:00
if err := batch . Write ( ) ; err != nil {
2021-09-07 13:31:17 +03:00
// The tx index data could not be written.
// Roll back the ancient store update.
fastBlock := bc . CurrentFastBlock ( ) . NumberU64 ( )
if err := bc . db . TruncateAncients ( fastBlock + 1 ) ; err != nil {
log . Error ( "Can't truncate ancient store after failed insert" , "err" , err )
}
all: integrate the freezer with fast sync
* all: freezer style syncing
core, eth, les, light: clean up freezer relative APIs
core, eth, les, trie, ethdb, light: clean a bit
core, eth, les, light: add unit tests
core, light: rewrite setHead function
core, eth: fix downloader unit tests
core: add receipt chain insertion test
core: use constant instead of hardcoding table name
core: fix rollback
core: fix setHead
core/rawdb: remove canonical block first and then iterate side chain
core/rawdb, ethdb: add hasAncient interface
eth/downloader: calculate ancient limit via cht first
core, eth, ethdb: lots of fixes
* eth/downloader: print ancient disable log only for fast sync
2019-04-25 17:59:48 +03:00
return 0 , err
2018-03-07 13:05:14 +03:00
}
2015-10-07 12:14:30 +03:00
all: integrate the freezer with fast sync
* all: freezer style syncing
core, eth, les, light: clean up freezer relative APIs
core, eth, les, trie, ethdb, light: clean a bit
core, eth, les, light: add unit tests
core, light: rewrite setHead function
core, eth: fix downloader unit tests
core: add receipt chain insertion test
core: use constant instead of hardcoding table name
core: fix rollback
core: fix setHead
core/rawdb: remove canonical block first and then iterate side chain
core/rawdb, ethdb: add hasAncient interface
eth/downloader: calculate ancient limit via cht first
core, eth, ethdb: lots of fixes
* eth/downloader: print ancient disable log only for fast sync
2019-04-25 17:59:48 +03:00
// Sync the ancient store explicitly to ensure all data has been flushed to disk.
if err := bc . db . Sync ( ) ; err != nil {
return 0 , err
}
2021-09-07 13:31:17 +03:00
// Update the current fast block because all block data is now present in DB.
previousFastBlock := bc . CurrentFastBlock ( ) . NumberU64 ( )
if ! updateHead ( blockChain [ len ( blockChain ) - 1 ] ) {
// We end up here if the header chain has reorg'ed, and the blocks/receipts
// don't match the canonical chain.
if err := bc . db . TruncateAncients ( previousFastBlock + 1 ) ; err != nil {
log . Error ( "Can't truncate ancient store after failed insert" , "err" , err )
2019-05-27 12:05:45 +03:00
}
2021-09-07 13:31:17 +03:00
return 0 , errSideChainReceipts
all: integrate the freezer with fast sync
* all: freezer style syncing
core, eth, les, light: clean up freezer relative APIs
core, eth, les, trie, ethdb, light: clean a bit
core, eth, les, light: add unit tests
core, light: rewrite setHead function
core, eth: fix downloader unit tests
core: add receipt chain insertion test
core: use constant instead of hardcoding table name
core: fix rollback
core: fix setHead
core/rawdb: remove canonical block first and then iterate side chain
core/rawdb, ethdb: add hasAncient interface
eth/downloader: calculate ancient limit via cht first
core, eth, ethdb: lots of fixes
* eth/downloader: print ancient disable log only for fast sync
2019-04-25 17:59:48 +03:00
}
core, cmd, vendor: fixes and database inspection tool (#15)
* core, eth: some fixes for freezer
* vendor, core/rawdb, cmd/geth: add db inspector
* core, cmd/utils: check ancient store path forceily
* cmd/geth, common, core/rawdb: a few fixes
* cmd/geth: support windows file rename and fix rename error
* core: support ancient plugin
* core, cmd: streaming file copy
* cmd, consensus, core, tests: keep genesis in leveldb
* core: write txlookup during ancient init
* core: bump database version
2019-05-14 17:07:44 +03:00
2021-09-07 13:31:17 +03:00
// Delete block data from the main database.
batch . Reset ( )
canonHashes := make ( map [ common . Hash ] struct { } )
for _ , block := range blockChain {
canonHashes [ block . Hash ( ) ] = struct { } { }
if block . NumberU64 ( ) == 0 {
continue
2019-07-08 11:24:16 +03:00
}
2021-09-07 13:31:17 +03:00
rawdb . DeleteCanonicalHash ( batch , block . NumberU64 ( ) )
rawdb . DeleteBlockWithoutNumber ( batch , block . Hash ( ) , block . NumberU64 ( ) )
2019-07-08 11:24:16 +03:00
}
2021-09-07 13:31:17 +03:00
// Delete side chain hash-to-number mappings.
for _ , nh := range rawdb . ReadAllHashesInRange ( bc . db , first . NumberU64 ( ) , last . NumberU64 ( ) ) {
if _ , canon := canonHashes [ nh . Hash ] ; ! canon {
rawdb . DeleteHeader ( batch , nh . Hash , nh . Number )
all: integrate the freezer with fast sync
* all: freezer style syncing
core, eth, les, light: clean up freezer relative APIs
core, eth, les, trie, ethdb, light: clean a bit
core, eth, les, light: add unit tests
core, light: rewrite setHead function
core, eth: fix downloader unit tests
core: add receipt chain insertion test
core: use constant instead of hardcoding table name
core: fix rollback
core: fix setHead
core/rawdb: remove canonical block first and then iterate side chain
core/rawdb, ethdb: add hasAncient interface
eth/downloader: calculate ancient limit via cht first
core, eth, ethdb: lots of fixes
* eth/downloader: print ancient disable log only for fast sync
2019-04-25 17:59:48 +03:00
}
}
if err := batch . Write ( ) ; err != nil {
return 0 , err
}
return 0 , nil
}
2021-09-07 13:31:17 +03:00
all: integrate the freezer with fast sync
* all: freezer style syncing
core, eth, les, light: clean up freezer relative APIs
core, eth, les, trie, ethdb, light: clean a bit
core, eth, les, light: add unit tests
core, light: rewrite setHead function
core, eth: fix downloader unit tests
core: add receipt chain insertion test
core: use constant instead of hardcoding table name
core: fix rollback
core: fix setHead
core/rawdb: remove canonical block first and then iterate side chain
core/rawdb, ethdb: add hasAncient interface
eth/downloader: calculate ancient limit via cht first
core, eth, ethdb: lots of fixes
* eth/downloader: print ancient disable log only for fast sync
2019-04-25 17:59:48 +03:00
// writeLive writes blockchain and corresponding receipt chain into active store.
writeLive := func ( blockChain types . Blocks , receiptChain [ ] types . Receipts ) ( int , error ) {
2020-08-22 19:12:04 +03:00
skipPresenceCheck := false
all: integrate the freezer with fast sync
* all: freezer style syncing
core, eth, les, light: clean up freezer relative APIs
core, eth, les, trie, ethdb, light: clean a bit
core, eth, les, light: add unit tests
core, light: rewrite setHead function
core, eth: fix downloader unit tests
core: add receipt chain insertion test
core: use constant instead of hardcoding table name
core: fix rollback
core: fix setHead
core/rawdb: remove canonical block first and then iterate side chain
core/rawdb, ethdb: add hasAncient interface
eth/downloader: calculate ancient limit via cht first
core, eth, ethdb: lots of fixes
* eth/downloader: print ancient disable log only for fast sync
2019-04-25 17:59:48 +03:00
batch := bc . db . NewBatch ( )
for i , block := range blockChain {
// Short circuit insertion if shutting down or processing failed
2020-05-26 22:37:37 +03:00
if bc . insertStopped ( ) {
all: integrate the freezer with fast sync
* all: freezer style syncing
core, eth, les, light: clean up freezer relative APIs
core, eth, les, trie, ethdb, light: clean a bit
core, eth, les, light: add unit tests
core, light: rewrite setHead function
core, eth: fix downloader unit tests
core: add receipt chain insertion test
core: use constant instead of hardcoding table name
core: fix rollback
core: fix setHead
core/rawdb: remove canonical block first and then iterate side chain
core/rawdb, ethdb: add hasAncient interface
eth/downloader: calculate ancient limit via cht first
core, eth, ethdb: lots of fixes
* eth/downloader: print ancient disable log only for fast sync
2019-04-25 17:59:48 +03:00
return 0 , errInsertionInterrupted
}
// Short circuit if the owner header is unknown
if ! bc . HasHeader ( block . Hash ( ) , block . NumberU64 ( ) ) {
2021-04-15 20:35:00 +03:00
return i , fmt . Errorf ( "containing header #%d [%x..] unknown" , block . Number ( ) , block . Hash ( ) . Bytes ( ) [ : 4 ] )
all: integrate the freezer with fast sync
* all: freezer style syncing
core, eth, les, light: clean up freezer relative APIs
core, eth, les, trie, ethdb, light: clean a bit
core, eth, les, light: add unit tests
core, light: rewrite setHead function
core, eth: fix downloader unit tests
core: add receipt chain insertion test
core: use constant instead of hardcoding table name
core: fix rollback
core: fix setHead
core/rawdb: remove canonical block first and then iterate side chain
core/rawdb, ethdb: add hasAncient interface
eth/downloader: calculate ancient limit via cht first
core, eth, ethdb: lots of fixes
* eth/downloader: print ancient disable log only for fast sync
2019-04-25 17:59:48 +03:00
}
2020-08-22 19:12:04 +03:00
if ! skipPresenceCheck {
// Ignore if the entire data is already known
if bc . HasBlock ( block . Hash ( ) , block . NumberU64 ( ) ) {
stats . ignored ++
continue
} else {
// If block N is not present, neither are the later blocks.
// This should be true, but if we are mistaken, the shortcut
// here will only cause overwriting of some existing data
skipPresenceCheck = true
}
all: integrate the freezer with fast sync
* all: freezer style syncing
core, eth, les, light: clean up freezer relative APIs
core, eth, les, trie, ethdb, light: clean a bit
core, eth, les, light: add unit tests
core, light: rewrite setHead function
core, eth: fix downloader unit tests
core: add receipt chain insertion test
core: use constant instead of hardcoding table name
core: fix rollback
core: fix setHead
core/rawdb: remove canonical block first and then iterate side chain
core/rawdb, ethdb: add hasAncient interface
eth/downloader: calculate ancient limit via cht first
core, eth, ethdb: lots of fixes
* eth/downloader: print ancient disable log only for fast sync
2019-04-25 17:59:48 +03:00
}
// Write all the data out into the database
rawdb . WriteBody ( batch , block . Hash ( ) , block . NumberU64 ( ) , block . Body ( ) )
rawdb . WriteReceipts ( batch , block . Hash ( ) , block . NumberU64 ( ) , receiptChain [ i ] )
2020-09-15 11:37:01 +03:00
rawdb . WriteTxLookupEntriesByBlock ( batch , block ) // Always write tx indices for live blocks, we assume they are needed
all: integrate the freezer with fast sync
* all: freezer style syncing
core, eth, les, light: clean up freezer relative APIs
core, eth, les, trie, ethdb, light: clean a bit
core, eth, les, light: add unit tests
core, light: rewrite setHead function
core, eth: fix downloader unit tests
core: add receipt chain insertion test
core: use constant instead of hardcoding table name
core: fix rollback
core: fix setHead
core/rawdb: remove canonical block first and then iterate side chain
core/rawdb, ethdb: add hasAncient interface
eth/downloader: calculate ancient limit via cht first
core, eth, ethdb: lots of fixes
* eth/downloader: print ancient disable log only for fast sync
2019-04-25 17:59:48 +03:00
2020-01-17 13:49:32 +03:00
// Write everything belongs to the blocks into the database. So that
// we can ensure all components of body is completed(body, receipts,
// tx indexes)
all: integrate the freezer with fast sync
* all: freezer style syncing
core, eth, les, light: clean up freezer relative APIs
core, eth, les, trie, ethdb, light: clean a bit
core, eth, les, light: add unit tests
core, light: rewrite setHead function
core, eth: fix downloader unit tests
core: add receipt chain insertion test
core: use constant instead of hardcoding table name
core: fix rollback
core: fix setHead
core/rawdb: remove canonical block first and then iterate side chain
core/rawdb, ethdb: add hasAncient interface
eth/downloader: calculate ancient limit via cht first
core, eth, ethdb: lots of fixes
* eth/downloader: print ancient disable log only for fast sync
2019-04-25 17:59:48 +03:00
if batch . ValueSize ( ) >= ethdb . IdealBatchSize {
if err := batch . Write ( ) ; err != nil {
return 0 , err
}
2021-09-07 13:31:17 +03:00
size += int64 ( batch . ValueSize ( ) )
all: integrate the freezer with fast sync
* all: freezer style syncing
core, eth, les, light: clean up freezer relative APIs
core, eth, les, trie, ethdb, light: clean a bit
core, eth, les, light: add unit tests
core, light: rewrite setHead function
core, eth: fix downloader unit tests
core: add receipt chain insertion test
core: use constant instead of hardcoding table name
core: fix rollback
core: fix setHead
core/rawdb: remove canonical block first and then iterate side chain
core/rawdb, ethdb: add hasAncient interface
eth/downloader: calculate ancient limit via cht first
core, eth, ethdb: lots of fixes
* eth/downloader: print ancient disable log only for fast sync
2019-04-25 17:59:48 +03:00
batch . Reset ( )
}
2020-01-17 13:49:32 +03:00
stats . processed ++
all: integrate the freezer with fast sync
* all: freezer style syncing
core, eth, les, light: clean up freezer relative APIs
core, eth, les, trie, ethdb, light: clean a bit
core, eth, les, light: add unit tests
core, light: rewrite setHead function
core, eth: fix downloader unit tests
core: add receipt chain insertion test
core: use constant instead of hardcoding table name
core: fix rollback
core: fix setHead
core/rawdb: remove canonical block first and then iterate side chain
core/rawdb, ethdb: add hasAncient interface
eth/downloader: calculate ancient limit via cht first
core, eth, ethdb: lots of fixes
* eth/downloader: print ancient disable log only for fast sync
2019-04-25 17:59:48 +03:00
}
2020-01-17 13:49:32 +03:00
// Write everything belongs to the blocks into the database. So that
// we can ensure all components of body is completed(body, receipts,
// tx indexes)
all: integrate the freezer with fast sync
* all: freezer style syncing
core, eth, les, light: clean up freezer relative APIs
core, eth, les, trie, ethdb, light: clean a bit
core, eth, les, light: add unit tests
core, light: rewrite setHead function
core, eth: fix downloader unit tests
core: add receipt chain insertion test
core: use constant instead of hardcoding table name
core: fix rollback
core: fix setHead
core/rawdb: remove canonical block first and then iterate side chain
core/rawdb, ethdb: add hasAncient interface
eth/downloader: calculate ancient limit via cht first
core, eth, ethdb: lots of fixes
* eth/downloader: print ancient disable log only for fast sync
2019-04-25 17:59:48 +03:00
if batch . ValueSize ( ) > 0 {
2021-09-07 13:31:17 +03:00
size += int64 ( batch . ValueSize ( ) )
2017-09-09 19:03:07 +03:00
if err := batch . Write ( ) ; err != nil {
return 0 , err
2015-10-07 12:14:30 +03:00
}
2015-09-30 19:23:31 +03:00
}
all: integrate the freezer with fast sync
* all: freezer style syncing
core, eth, les, light: clean up freezer relative APIs
core, eth, les, trie, ethdb, light: clean a bit
core, eth, les, light: add unit tests
core, light: rewrite setHead function
core, eth: fix downloader unit tests
core: add receipt chain insertion test
core: use constant instead of hardcoding table name
core: fix rollback
core: fix setHead
core/rawdb: remove canonical block first and then iterate side chain
core/rawdb, ethdb: add hasAncient interface
eth/downloader: calculate ancient limit via cht first
core, eth, ethdb: lots of fixes
* eth/downloader: print ancient disable log only for fast sync
2019-04-25 17:59:48 +03:00
updateHead ( blockChain [ len ( blockChain ) - 1 ] )
return 0 , nil
2015-10-07 12:14:30 +03:00
}
2021-09-07 13:31:17 +03:00
2020-05-11 18:58:43 +03:00
// Write downloaded chain data and corresponding receipt chain data
all: integrate the freezer with fast sync
* all: freezer style syncing
core, eth, les, light: clean up freezer relative APIs
core, eth, les, trie, ethdb, light: clean a bit
core, eth, les, light: add unit tests
core, light: rewrite setHead function
core, eth: fix downloader unit tests
core: add receipt chain insertion test
core: use constant instead of hardcoding table name
core: fix rollback
core: fix setHead
core/rawdb: remove canonical block first and then iterate side chain
core/rawdb, ethdb: add hasAncient interface
eth/downloader: calculate ancient limit via cht first
core, eth, ethdb: lots of fixes
* eth/downloader: print ancient disable log only for fast sync
2019-04-25 17:59:48 +03:00
if len ( ancientBlocks ) > 0 {
if n , err := writeAncient ( ancientBlocks , ancientReceipts ) ; err != nil {
if err == errInsertionInterrupted {
return 0 , nil
}
return n , err
2015-09-30 19:23:31 +03:00
}
}
2020-05-11 18:58:43 +03:00
// Write the tx index tail (block number from where we index) before write any live blocks
if len ( liveBlocks ) > 0 && liveBlocks [ 0 ] . NumberU64 ( ) == ancientLimit + 1 {
// The tx index tail can only be one of the following two options:
// * 0: all ancient blocks have been indexed
// * ancient-limit: the indices of blocks before ancient-limit are ignored
if tail := rawdb . ReadTxIndexTail ( bc . db ) ; tail == nil {
if bc . txLookupLimit == 0 || ancientLimit <= bc . txLookupLimit {
rawdb . WriteTxIndexTail ( bc . db , 0 )
} else {
rawdb . WriteTxIndexTail ( bc . db , ancientLimit - bc . txLookupLimit )
}
}
}
all: integrate the freezer with fast sync
* all: freezer style syncing
core, eth, les, light: clean up freezer relative APIs
core, eth, les, trie, ethdb, light: clean a bit
core, eth, les, light: add unit tests
core, light: rewrite setHead function
core, eth: fix downloader unit tests
core: add receipt chain insertion test
core: use constant instead of hardcoding table name
core: fix rollback
core: fix setHead
core/rawdb: remove canonical block first and then iterate side chain
core/rawdb, ethdb: add hasAncient interface
eth/downloader: calculate ancient limit via cht first
core, eth, ethdb: lots of fixes
* eth/downloader: print ancient disable log only for fast sync
2019-04-25 17:59:48 +03:00
if len ( liveBlocks ) > 0 {
if n , err := writeLive ( liveBlocks , liveReceipts ) ; err != nil {
if err == errInsertionInterrupted {
return 0 , nil
}
return n , err
2015-10-07 12:14:30 +03:00
}
}
all: integrate the freezer with fast sync
* all: freezer style syncing
core, eth, les, light: clean up freezer relative APIs
core, eth, les, trie, ethdb, light: clean a bit
core, eth, les, light: add unit tests
core, light: rewrite setHead function
core, eth: fix downloader unit tests
core: add receipt chain insertion test
core: use constant instead of hardcoding table name
core: fix rollback
core: fix setHead
core/rawdb: remove canonical block first and then iterate side chain
core/rawdb, ethdb: add hasAncient interface
eth/downloader: calculate ancient limit via cht first
core, eth, ethdb: lots of fixes
* eth/downloader: print ancient disable log only for fast sync
2019-04-25 17:59:48 +03:00
head := blockChain [ len ( blockChain ) - 1 ]
2018-09-20 11:41:59 +03:00
context := [ ] interface { } {
"count" , stats . processed , "elapsed" , common . PrettyDuration ( time . Since ( start ) ) ,
2019-04-02 23:28:48 +03:00
"number" , head . Number ( ) , "hash" , head . Hash ( ) , "age" , common . PrettyAge ( time . Unix ( int64 ( head . Time ( ) ) , 0 ) ) ,
all: integrate the freezer with fast sync
* all: freezer style syncing
core, eth, les, light: clean up freezer relative APIs
core, eth, les, trie, ethdb, light: clean a bit
core, eth, les, light: add unit tests
core, light: rewrite setHead function
core, eth: fix downloader unit tests
core: add receipt chain insertion test
core: use constant instead of hardcoding table name
core: fix rollback
core: fix setHead
core/rawdb: remove canonical block first and then iterate side chain
core/rawdb, ethdb: add hasAncient interface
eth/downloader: calculate ancient limit via cht first
core, eth, ethdb: lots of fixes
* eth/downloader: print ancient disable log only for fast sync
2019-04-25 17:59:48 +03:00
"size" , common . StorageSize ( size ) ,
2018-09-20 11:41:59 +03:00
}
if stats . ignored > 0 {
context = append ( context , [ ] interface { } { "ignored" , stats . ignored } ... )
}
log . Info ( "Imported new block receipts" , context ... )
2015-09-30 19:23:31 +03:00
return 0 , nil
}
2018-02-05 19:40:32 +03:00
var lastWrite uint64
2019-05-08 14:30:36 +03:00
// writeBlockWithoutState writes only the block and its metadata to the database,
2018-02-05 19:40:32 +03:00
// but does not write any state. This is used to construct competing side forks
// up to the point where they exceed the canonical total difficulty.
2019-05-08 14:30:36 +03:00
func ( bc * BlockChain ) writeBlockWithoutState ( block * types . Block , td * big . Int ) ( err error ) {
2021-10-07 16:47:50 +03:00
if bc . insertStopped ( ) {
return errInsertionInterrupted
}
2018-02-05 19:40:32 +03:00
2020-01-17 13:49:32 +03:00
batch := bc . db . NewBatch ( )
rawdb . WriteTd ( batch , block . Hash ( ) , block . NumberU64 ( ) , td )
rawdb . WriteBlock ( batch , block )
if err := batch . Write ( ) ; err != nil {
log . Crit ( "Failed to write block into disk" , "err" , err )
2018-02-05 19:40:32 +03:00
}
return nil
}
2019-05-08 14:30:36 +03:00
// writeKnownBlock updates the head block flag with a known block
// and introduces chain reorg if necessary.
func ( bc * BlockChain ) writeKnownBlock ( block * types . Block ) error {
current := bc . CurrentBlock ( )
if block . ParentHash ( ) != current . Hash ( ) {
if err := bc . reorg ( current , block ) ; err != nil {
return err
}
}
2020-01-17 13:49:32 +03:00
bc . writeHeadBlock ( block )
2019-05-08 14:30:36 +03:00
return nil
}
2018-02-05 19:40:32 +03:00
// WriteBlockWithState writes the block and all associated state to the database.
2019-11-29 16:22:08 +03:00
func ( bc * BlockChain ) WriteBlockWithState ( block * types . Block , receipts [ ] * types . Receipt , logs [ ] * types . Log , state * state . StateDB , emitHeadEvent bool ) ( status WriteStatus , err error ) {
2021-10-07 16:47:50 +03:00
if ! bc . chainmu . TryLock ( ) {
return NonStatTy , errInsertionInterrupted
}
2019-01-11 16:27:47 +03:00
defer bc . chainmu . Unlock ( )
2019-11-29 16:22:08 +03:00
return bc . writeBlockWithState ( block , receipts , logs , state , emitHeadEvent )
2019-01-11 16:27:47 +03:00
}
// writeBlockWithState writes the block and all associated state to the database,
// but is expects the chain mutex to be held.
2019-11-29 16:22:08 +03:00
func ( bc * BlockChain ) writeBlockWithState ( block * types . Block , receipts [ ] * types . Receipt , logs [ ] * types . Log , state * state . StateDB , emitHeadEvent bool ) ( status WriteStatus , err error ) {
2021-10-07 16:47:50 +03:00
if bc . insertStopped ( ) {
return NonStatTy , errInsertionInterrupted
}
2015-06-29 13:12:30 +03:00
2015-09-07 20:43:01 +03:00
// Calculate the total difficulty of the block
2017-05-11 04:55:48 +03:00
ptd := bc . GetTd ( block . ParentHash ( ) , block . NumberU64 ( ) - 1 )
2015-09-07 20:43:01 +03:00
if ptd == nil {
2017-04-06 14:58:03 +03:00
return NonStatTy , consensus . ErrUnknownAncestor
2015-09-07 20:43:01 +03:00
}
2015-09-29 14:24:28 +03:00
// Make sure no inconsistent state is leaked during insertion
2018-02-26 12:53:10 +03:00
currentBlock := bc . CurrentBlock ( )
localTd := bc . GetTd ( currentBlock . Hash ( ) , currentBlock . NumberU64 ( ) )
2016-07-08 16:59:19 +03:00
externTd := new ( big . Int ) . Add ( block . Difficulty ( ) , ptd )
2020-01-17 13:49:32 +03:00
// Irrelevant of the canonical status, write the block itself to the database.
//
// Note all the components of block(td, hash->number map, header, body, receipts)
// should be written atomically. BlockBatch is used for containing all components.
blockBatch := bc . db . NewBatch ( )
rawdb . WriteTd ( blockBatch , block . Hash ( ) , block . NumberU64 ( ) , externTd )
rawdb . WriteBlock ( blockBatch , block )
rawdb . WriteReceipts ( blockBatch , block . Hash ( ) , block . NumberU64 ( ) , receipts )
rawdb . WritePreimages ( blockBatch , state . Preimages ( ) )
if err := blockBatch . Write ( ) ; err != nil {
log . Crit ( "Failed to write block into disk" , "err" , err )
}
// Commit all cached state changes into underlying memory database.
2018-02-05 19:40:32 +03:00
root , err := state . Commit ( bc . chainConfig . IsEIP158 ( block . Number ( ) ) )
if err != nil {
2017-09-09 19:03:07 +03:00
return NonStatTy , err
}
2018-02-05 19:40:32 +03:00
triedb := bc . stateCache . TrieDB ( )
// If we're running an archive node, always flush
2019-04-01 11:52:11 +03:00
if bc . cacheConfig . TrieDirtyDisabled {
2020-07-13 12:02:54 +03:00
if err := triedb . Commit ( root , false , nil ) ; err != nil {
2018-02-05 19:40:32 +03:00
return NonStatTy , err
}
} else {
// Full but not archive node, do proper garbage collection
triedb . Reference ( root , common . Hash { } ) // metadata reference to keep trie alive
2018-09-03 18:33:21 +03:00
bc . triegc . Push ( root , - int64 ( block . NumberU64 ( ) ) )
2018-02-05 19:40:32 +03:00
2019-05-30 21:51:13 +03:00
if current := block . NumberU64 ( ) ; current > TriesInMemory {
2018-06-04 10:47:43 +03:00
// If we exceeded our memory allowance, flush matured singleton nodes to disk
var (
nodes , imgs = triedb . Size ( )
2018-11-12 19:47:34 +03:00
limit = common . StorageSize ( bc . cacheConfig . TrieDirtyLimit ) * 1024 * 1024
2018-06-04 10:47:43 +03:00
)
if nodes > limit || imgs > 4 * 1024 * 1024 {
triedb . Cap ( limit - ethdb . IdealBatchSize )
}
2018-02-05 19:40:32 +03:00
// Find the next state trie we need to commit
2019-05-30 21:51:13 +03:00
chosen := current - TriesInMemory
2018-02-05 19:40:32 +03:00
2018-06-04 10:47:43 +03:00
// If we exceeded out time allowance, flush an entire trie to disk
if bc . gcproc > bc . cacheConfig . TrieTimeLimit {
2019-02-08 12:11:31 +03:00
// If the header is missing (canonical chain behind), we're reorging a low
// diff sidechain. Suspend committing until this operation is completed.
header := bc . GetHeaderByNumber ( chosen )
if header == nil {
log . Warn ( "Reorg in progress, trie commit postponed" , "number" , chosen )
} else {
// If we're exceeding limits but haven't reached a large enough memory gap,
// warn the user that the system is becoming unstable.
2019-05-30 21:51:13 +03:00
if chosen < lastWrite + TriesInMemory && bc . gcproc >= 2 * bc . cacheConfig . TrieTimeLimit {
log . Info ( "State in memory for too long, committing" , "time" , bc . gcproc , "allowance" , bc . cacheConfig . TrieTimeLimit , "optimum" , float64 ( chosen - lastWrite ) / TriesInMemory )
2019-02-08 12:11:31 +03:00
}
// Flush an entire trie and restart the counters
2020-07-13 12:02:54 +03:00
triedb . Commit ( header . Root , true , nil )
2019-02-08 12:11:31 +03:00
lastWrite = chosen
bc . gcproc = 0
2018-02-05 19:40:32 +03:00
}
}
// Garbage collect anything below our required write retention
for ! bc . triegc . Empty ( ) {
root , number := bc . triegc . Pop ( )
if uint64 ( - number ) > chosen {
bc . triegc . Push ( root , number )
break
}
2018-06-21 12:28:05 +03:00
triedb . Dereference ( root . ( common . Hash ) )
2018-02-05 19:40:32 +03:00
}
}
}
2015-09-29 14:24:28 +03:00
// If the total difficulty is higher than our known, add it to the canonical chain
2016-01-12 12:54:29 +02:00
// Second clause in the if statement reduces the vulnerability to selfish mining.
// Please refer to http://www.cs.cornell.edu/~ie53/publications/btcProcFC.pdf
2017-11-13 18:07:05 +03:00
reorg := externTd . Cmp ( localTd ) > 0
2018-02-26 12:53:10 +03:00
currentBlock = bc . CurrentBlock ( )
2017-11-13 18:07:05 +03:00
if ! reorg && externTd . Cmp ( localTd ) == 0 {
2018-09-20 15:09:30 +03:00
// Split same-difficulty blocks by number, then preferentially select
// the block generated by the local miner as the canonical block.
if block . NumberU64 ( ) < currentBlock . NumberU64 ( ) {
reorg = true
} else if block . NumberU64 ( ) == currentBlock . NumberU64 ( ) {
2018-09-20 20:02:15 +03:00
var currentPreserve , blockPreserve bool
if bc . shouldPreserve != nil {
currentPreserve , blockPreserve = bc . shouldPreserve ( currentBlock ) , bc . shouldPreserve ( block )
2018-09-20 15:09:30 +03:00
}
2018-09-20 20:02:15 +03:00
reorg = ! currentPreserve && ( blockPreserve || mrand . Float64 ( ) < 0.5 )
2018-09-20 15:09:30 +03:00
}
2017-11-13 18:07:05 +03:00
}
if reorg {
2016-03-15 20:55:39 +02:00
// Reorganise the chain if the parent is not the head block
2018-02-26 12:53:10 +03:00
if block . ParentHash ( ) != currentBlock . Hash ( ) {
if err := bc . reorg ( currentBlock , block ) ; err != nil {
2015-07-03 12:24:42 +03:00
return NonStatTy , err
2015-06-29 13:12:30 +03:00
}
}
2015-09-29 14:24:28 +03:00
status = CanonStatTy
2015-06-29 13:12:30 +03:00
} else {
2015-07-03 12:24:42 +03:00
status = SideStatTy
2015-06-29 13:12:30 +03:00
}
2017-09-09 19:03:07 +03:00
// Set new head.
if status == CanonStatTy {
2020-01-17 13:49:32 +03:00
bc . writeHeadBlock ( block )
2017-09-09 19:03:07 +03:00
}
2017-05-11 04:55:48 +03:00
bc . futureBlocks . Remove ( block . Hash ( ) )
2019-11-29 16:22:08 +03:00
if status == CanonStatTy {
bc . chainFeed . Send ( ChainEvent { Block : block , Hash : block . Hash ( ) , Logs : logs } )
if len ( logs ) > 0 {
bc . logsFeed . Send ( logs )
}
// In theory we should fire a ChainHeadEvent when we inject
// a canonical block, but sometimes we can insert a batch of
// canonicial blocks. Avoid firing too much ChainHeadEvents,
// we will fire an accumulated ChainHeadEvent and disable fire
// event here.
if emitHeadEvent {
bc . chainHeadFeed . Send ( ChainHeadEvent { Block : block } )
}
} else {
bc . chainSideFeed . Send ( ChainSideEvent { Block : block } )
}
2017-09-09 19:03:07 +03:00
return status , nil
2015-06-29 13:12:30 +03:00
}
2018-11-20 15:15:26 +03:00
// addFutureBlock checks if the block is within the max allowed window to get
// accepted for future processing, and returns an error if the block is too far
// ahead and was not added.
2018-10-20 11:43:59 +03:00
func ( bc * BlockChain ) addFutureBlock ( block * types . Block ) error {
2019-04-02 23:28:48 +03:00
max := uint64 ( time . Now ( ) . Unix ( ) + maxTimeFutureBlocks )
if block . Time ( ) > max {
2018-11-20 15:15:26 +03:00
return fmt . Errorf ( "future block timestamp %v > allowed %v" , block . Time ( ) , max )
2018-10-20 11:43:59 +03:00
}
bc . futureBlocks . Add ( block . Hash ( ) , block )
return nil
}
2017-09-11 13:13:05 +03:00
// InsertChain attempts to insert the given batch of blocks in to the canonical
// chain or, otherwise, create a fork. If an error is returned it will return
// the index number of the failing block as well an error describing what went
// wrong.
//
// After insertion is done, all accumulated events will be fired.
2017-05-11 04:55:48 +03:00
func ( bc * BlockChain ) InsertChain ( chain types . Blocks ) ( int , error ) {
2018-06-04 14:09:16 +03:00
// Sanity check that we have something meaningful to import
if len ( chain ) == 0 {
2018-11-20 15:15:26 +03:00
return 0 , nil
2018-06-04 14:09:16 +03:00
}
2019-02-26 14:32:48 +03:00
bc . blockProcFeed . Send ( true )
defer bc . blockProcFeed . Send ( false )
2021-10-07 16:47:50 +03:00
// Do a sanity check that the provided chain is actually ordered and linked.
2016-12-13 17:14:33 +03:00
for i := 1 ; i < len ( chain ) ; i ++ {
2021-10-07 16:47:50 +03:00
block , prev := chain [ i ] , chain [ i - 1 ]
2019-02-21 13:36:49 +03:00
if block . NumberU64 ( ) != prev . NumberU64 ( ) + 1 || block . ParentHash ( ) != prev . Hash ( ) {
2021-10-07 16:47:50 +03:00
log . Error ( "Non contiguous block insert" ,
"number" , block . Number ( ) ,
"hash" , block . Hash ( ) ,
"parent" , block . ParentHash ( ) ,
"prevnumber" , prev . Number ( ) ,
"prevhash" , prev . Hash ( ) ,
)
2021-04-15 20:35:00 +03:00
return 0 , fmt . Errorf ( "non contiguous insert: item %d is #%d [%x..], item %d is #%d [%x..] (parent [%x..])" , i - 1 , prev . NumberU64 ( ) ,
2019-02-21 13:36:49 +03:00
prev . Hash ( ) . Bytes ( ) [ : 4 ] , i , block . NumberU64 ( ) , block . Hash ( ) . Bytes ( ) [ : 4 ] , block . ParentHash ( ) . Bytes ( ) [ : 4 ] )
2016-12-13 17:14:33 +03:00
}
}
2018-10-20 11:43:59 +03:00
2021-10-07 16:47:50 +03:00
// Pre-check passed, start the full block imports.
if ! bc . chainmu . TryLock ( ) {
return 0 , errChainStopped
}
defer bc . chainmu . Unlock ( )
return bc . insertChain ( chain , true )
2018-11-20 15:15:26 +03:00
}
2018-10-20 11:43:59 +03:00
2021-04-16 22:29:22 +03:00
// InsertChainWithoutSealVerification works exactly the same
// except for seal verification, seal verification is omitted
func ( bc * BlockChain ) InsertChainWithoutSealVerification ( block * types . Block ) ( int , error ) {
bc . blockProcFeed . Send ( true )
defer bc . blockProcFeed . Send ( false )
2021-10-07 16:47:50 +03:00
if ! bc . chainmu . TryLock ( ) {
return 0 , errChainStopped
}
defer bc . chainmu . Unlock ( )
return bc . insertChain ( types . Blocks ( [ ] * types . Block { block } ) , false )
2021-04-16 22:29:22 +03:00
}
2019-04-02 13:01:02 +03:00
// insertChain is the internal implementation of InsertChain, which assumes that
2018-11-20 15:15:26 +03:00
// 1) chains are contiguous, and 2) The chain mutex is held.
//
// This method is split out so that import batches that require re-injecting
// historical blocks can do so without releasing the lock, which could lead to
// racey behaviour. If a sidechain import is in progress, and the historic state
// is imported, but then new canon-head is added before the actual sidechain
// completes, then the historic state could be pruned again
2019-11-29 16:22:08 +03:00
func ( bc * BlockChain ) insertChain ( chain types . Blocks , verifySeals bool ) ( int , error ) {
2021-10-07 16:47:50 +03:00
// If the chain is terminating, don't even bother starting up.
if bc . insertStopped ( ) {
2019-11-29 16:22:08 +03:00
return 0 , nil
2018-11-20 15:15:26 +03:00
}
2021-10-07 16:47:50 +03:00
2018-10-20 11:43:59 +03:00
// Start a parallel signature recovery (signer will fluke on fork transition, minimal perf loss)
senderCacher . recoverFromBlocks ( types . MakeSigner ( bc . chainConfig , chain [ 0 ] . Number ( ) ) , chain )
2015-05-17 01:55:02 +03:00
2015-04-05 00:04:19 +03:00
var (
2019-11-29 16:22:08 +03:00
stats = insertStats { startTime : mclock . Now ( ) }
lastCanon * types . Block
2015-04-05 00:04:19 +03:00
)
2019-11-29 16:22:08 +03:00
// Fire a single chain head event if we've progressed the chain
defer func ( ) {
if lastCanon != nil && bc . CurrentBlock ( ) . Hash ( ) == lastCanon . Hash ( ) {
bc . chainHeadFeed . Send ( ChainHeadEvent { lastCanon } )
}
} ( )
2017-04-05 01:16:29 +03:00
// Start the parallel header verifier
headers := make ( [ ] * types . Header , len ( chain ) )
seals := make ( [ ] bool , len ( chain ) )
2015-05-17 02:42:30 +03:00
2017-04-05 01:16:29 +03:00
for i , block := range chain {
headers [ i ] = block . Header ( )
2018-10-20 11:43:59 +03:00
seals [ i ] = verifySeals
2017-04-05 01:16:29 +03:00
}
2017-05-11 04:55:48 +03:00
abort , results := bc . engine . VerifyHeaders ( bc , headers , seals )
2017-04-05 01:16:29 +03:00
defer close ( abort )
2015-06-19 17:21:20 +03:00
2018-11-20 15:15:26 +03:00
// Peek the error for the first block to decide the directing import logic
2019-03-25 13:41:50 +03:00
it := newInsertIterator ( chain , results , bc . validator )
2018-11-20 15:15:26 +03:00
block , err := it . next ( )
2019-02-21 13:36:49 +03:00
2021-11-01 16:09:36 +03:00
// Left-trim all the known blocks that don't need to build snapshot
if bc . skipBlock ( err , it ) {
2019-02-21 13:36:49 +03:00
// First block (and state) is known
// 1. We did a roll-back, and should now do a re-import
// 2. The block is stored as a sidechain, and is lying about it's stateroot, and passes a stateroot
2021-10-07 16:47:50 +03:00
// from the canonical chain, which has not been verified.
// Skip all known blocks that are behind us.
2019-05-08 14:30:36 +03:00
var (
current = bc . CurrentBlock ( )
localTd = bc . GetTd ( current . Hash ( ) , current . NumberU64 ( ) )
externTd = bc . GetTd ( block . ParentHash ( ) , block . NumberU64 ( ) - 1 ) // The first block can't be nil
)
2021-11-01 16:09:36 +03:00
for block != nil && bc . skipBlock ( err , it ) {
2019-05-08 14:30:36 +03:00
externTd = new ( big . Int ) . Add ( externTd , block . Difficulty ( ) )
if localTd . Cmp ( externTd ) < 0 {
break
}
2019-05-10 17:04:10 +03:00
log . Debug ( "Ignoring already known block" , "number" , block . Number ( ) , "hash" , block . Hash ( ) )
2019-02-21 13:36:49 +03:00
stats . ignored ++
2019-05-10 17:04:10 +03:00
2019-02-21 13:36:49 +03:00
block , err = it . next ( )
}
2019-05-08 14:30:36 +03:00
// The remaining blocks are still known blocks, the only scenario here is:
// During the fast sync, the pivot point is already submitted but rollback
// happens. Then node resets the head full block to a lower height via `rollback`
// and leaves a few known blocks in the database.
//
// When node runs a fast sync again, it can re-import a batch of known blocks via
// `insertChain` while a part of them have higher total difficulty than current
// head full block(new pivot point).
2021-11-01 16:09:36 +03:00
for block != nil && bc . skipBlock ( err , it ) {
2019-05-10 17:04:10 +03:00
log . Debug ( "Writing previously known block" , "number" , block . Number ( ) , "hash" , block . Hash ( ) )
2019-05-08 14:30:36 +03:00
if err := bc . writeKnownBlock ( block ) ; err != nil {
2019-11-29 16:22:08 +03:00
return it . index , err
2019-05-08 14:30:36 +03:00
}
lastCanon = block
block , err = it . next ( )
}
2019-02-21 13:36:49 +03:00
// Falls through to the block import
}
2018-11-20 15:15:26 +03:00
switch {
// First block is pruned, insert as sidechain and reorg only if TD grows enough
2020-08-05 10:52:54 +03:00
case errors . Is ( err , consensus . ErrPrunedAncestor ) :
2019-05-10 17:04:10 +03:00
log . Debug ( "Pruned ancestor, inserting as sidechain" , "number" , block . Number ( ) , "hash" , block . Hash ( ) )
2019-05-08 14:30:36 +03:00
return bc . insertSideChain ( block , it )
2018-11-20 15:15:26 +03:00
// First block is future, shove it (and all children) to the future queue (unknown ancestor)
2020-08-05 10:52:54 +03:00
case errors . Is ( err , consensus . ErrFutureBlock ) || ( errors . Is ( err , consensus . ErrUnknownAncestor ) && bc . futureBlocks . Contains ( it . first ( ) . ParentHash ( ) ) ) :
for block != nil && ( it . index == 0 || errors . Is ( err , consensus . ErrUnknownAncestor ) ) {
2019-05-10 17:04:10 +03:00
log . Debug ( "Future block, postponing import" , "number" , block . Number ( ) , "hash" , block . Hash ( ) )
2018-11-20 15:15:26 +03:00
if err := bc . addFutureBlock ( block ) ; err != nil {
2019-11-29 16:22:08 +03:00
return it . index , err
2018-10-20 11:43:59 +03:00
}
2018-11-20 15:15:26 +03:00
block , err = it . next ( )
}
stats . queued += it . processed ( )
stats . ignored += it . remaining ( )
// If there are any still remaining, mark as ignored
2019-11-29 16:22:08 +03:00
return it . index , err
2018-11-20 15:15:26 +03:00
2021-11-01 16:09:36 +03:00
// Some other error(except ErrKnownBlock) occurred, abort.
// ErrKnownBlock is allowed here since some known blocks
// still need re-execution to generate snapshots that are missing
case err != nil && ! errors . Is ( err , ErrKnownBlock ) :
2019-09-26 11:57:51 +03:00
bc . futureBlocks . Remove ( block . Hash ( ) )
2018-11-20 15:15:26 +03:00
stats . ignored += len ( it . chain )
bc . reportBlock ( block , nil , err )
2019-11-29 16:22:08 +03:00
return it . index , err
2018-10-20 11:43:59 +03:00
}
2018-11-20 15:15:26 +03:00
// No validation errors for the first block (or chain prefix skipped)
2021-02-12 13:45:34 +03:00
var activeState * state . StateDB
defer func ( ) {
// The chain importer is starting and stopping trie prefetchers. If a bad
// block or other error is hit however, an early return may not properly
// terminate the background threads. This defer ensures that we clean up
// and dangling prefetcher, without defering each and holding on live refs.
if activeState != nil {
activeState . StopPrefetcher ( )
}
} ( )
2021-11-01 16:09:36 +03:00
for ; block != nil && err == nil || errors . Is ( err , ErrKnownBlock ) ; block , err = it . next ( ) {
2017-04-05 01:16:29 +03:00
// If the chain is terminating, stop processing blocks
2020-05-26 22:37:37 +03:00
if bc . insertStopped ( ) {
log . Debug ( "Abort during block processing" )
2015-06-12 17:45:53 +03:00
break
}
2017-04-05 01:16:29 +03:00
// If the header is a banned one, straight out abort
2015-06-12 17:45:53 +03:00
if BadHashes [ block . Hash ( ) ] {
2021-07-29 11:17:40 +03:00
bc . reportBlock ( block , nil , ErrBannedHash )
return it . index , ErrBannedHash
2015-06-12 17:45:53 +03:00
}
2019-05-10 17:04:10 +03:00
// If the block is known (in the middle of the chain), it's a special case for
// Clique blocks where they can share state among each other, so importing an
// older block might complete the state of the subsequent one. In this case,
// just skip the block (we already validated it once fully (and crashed), since
2021-11-01 16:09:36 +03:00
// its header and body was already in the database). But if the corresponding
// snapshot layer is missing, forcibly rerun the execution to build it.
if bc . skipBlock ( err , it ) {
2019-05-10 17:04:10 +03:00
logger := log . Debug
if bc . chainConfig . Clique == nil {
logger = log . Warn
}
logger ( "Inserted known block" , "number" , block . Number ( ) , "hash" , block . Hash ( ) ,
"uncles" , len ( block . Uncles ( ) ) , "txs" , len ( block . Transactions ( ) ) , "gas" , block . GasUsed ( ) ,
"root" , block . Root ( ) )
2020-05-13 11:33:48 +03:00
// Special case. Commit the empty receipt slice if we meet the known
// block in the middle. It can only happen in the clique chain. Whenever
// we insert blocks via `insertSideChain`, we only commit `td`, `header`
// and `body` if it's non-existent. Since we don't have receipts without
// reexecution, so nothing to commit. But if the sidechain will be adpoted
// as the canonical chain eventually, it needs to be reexecuted for missing
// state, but if it's this special case here(skip reexecution) we will lose
// the empty receipt entry.
if len ( block . Transactions ( ) ) == 0 {
rawdb . WriteReceipts ( bc . db , block . Hash ( ) , block . NumberU64 ( ) , nil )
} else {
log . Error ( "Please file an issue, skip known block execution without receipt" ,
"hash" , block . Hash ( ) , "number" , block . NumberU64 ( ) )
}
2019-05-10 17:04:10 +03:00
if err := bc . writeKnownBlock ( block ) ; err != nil {
2019-11-29 16:22:08 +03:00
return it . index , err
2019-05-10 17:04:10 +03:00
}
stats . processed ++
// We can assume that logs are empty here, since the only way for consecutive
// Clique blocks to have the same state is if there are no transactions.
lastCanon = block
continue
}
2021-10-07 16:47:50 +03:00
2018-11-20 15:15:26 +03:00
// Retrieve the parent block and it's state to execute on top
start := time . Now ( )
parent := it . previous ( )
2018-10-20 11:43:59 +03:00
if parent == nil {
2019-03-13 13:31:35 +03:00
parent = bc . GetHeader ( block . ParentHash ( ) , block . NumberU64 ( ) - 1 )
2016-05-19 13:24:14 +03:00
}
2019-08-06 13:40:28 +03:00
statedb , err := state . New ( parent . Root , bc . stateCache , bc . snaps )
2015-10-19 17:08:17 +03:00
if err != nil {
2019-11-29 16:22:08 +03:00
return it . index , err
2015-06-12 17:45:53 +03:00
}
2021-10-07 16:47:50 +03:00
2021-01-08 16:01:49 +03:00
// Enable prefetching to pull in trie node paths while processing transactions
statedb . StartPrefetcher ( "chain" )
2021-02-12 13:45:34 +03:00
activeState = statedb
2021-01-08 16:01:49 +03:00
2019-03-25 13:41:50 +03:00
// If we have a followup block, run that against the current state to pre-cache
// transactions and probabilistically some of the account/storage trie nodes.
var followupInterrupt uint32
2019-04-01 11:52:11 +03:00
if ! bc . cacheConfig . TrieCleanNoPrefetch {
if followup , err := it . peek ( ) ; followup != nil && err == nil {
2019-08-06 13:40:28 +03:00
throwaway , _ := state . New ( parent . Root , bc . stateCache , bc . snaps )
2021-01-08 16:01:49 +03:00
2020-02-14 11:54:02 +03:00
go func ( start time . Time , followup * types . Block , throwaway * state . StateDB , interrupt * uint32 ) {
2019-08-06 13:40:28 +03:00
bc . prefetcher . Prefetch ( followup , throwaway , bc . vmConfig , & followupInterrupt )
2019-03-25 13:41:50 +03:00
2019-04-01 11:52:11 +03:00
blockPrefetchExecuteTimer . Update ( time . Since ( start ) )
2020-02-14 11:54:02 +03:00
if atomic . LoadUint32 ( interrupt ) == 1 {
2019-04-01 11:52:11 +03:00
blockPrefetchInterruptMeter . Mark ( 1 )
}
2020-02-14 11:54:02 +03:00
} ( time . Now ( ) , followup , throwaway , & followupInterrupt )
2019-04-01 11:52:11 +03:00
}
2019-03-25 13:41:50 +03:00
}
2021-10-07 16:47:50 +03:00
2019-03-25 13:41:50 +03:00
// Process block using the parent state as reference point
2019-03-27 14:02:04 +03:00
substart := time . Now ( )
2019-03-25 13:41:50 +03:00
receipts , logs , usedGas , err := bc . processor . Process ( block , statedb , bc . vmConfig )
2015-10-19 17:08:17 +03:00
if err != nil {
2017-05-11 04:55:48 +03:00
bc . reportBlock ( block , receipts , err )
2019-03-25 13:41:50 +03:00
atomic . StoreUint32 ( & followupInterrupt , 1 )
2019-11-29 16:22:08 +03:00
return it . index , err
2015-10-19 17:08:17 +03:00
}
2021-10-07 16:47:50 +03:00
2019-03-27 14:02:04 +03:00
// Update the metrics touched during block processing
2019-08-06 13:40:28 +03:00
accountReadTimer . Update ( statedb . AccountReads ) // Account reads are complete, we can mark them
storageReadTimer . Update ( statedb . StorageReads ) // Storage reads are complete, we can mark them
accountUpdateTimer . Update ( statedb . AccountUpdates ) // Account updates are complete, we can mark them
storageUpdateTimer . Update ( statedb . StorageUpdates ) // Storage updates are complete, we can mark them
snapshotAccountReadTimer . Update ( statedb . SnapshotAccountReads ) // Account reads are complete, we can mark them
snapshotStorageReadTimer . Update ( statedb . SnapshotStorageReads ) // Storage reads are complete, we can mark them
2020-02-05 15:12:09 +03:00
triehash := statedb . AccountHashes + statedb . StorageHashes // Save to not double count in validation
2019-08-06 13:40:28 +03:00
trieproc := statedb . SnapshotAccountReads + statedb . AccountReads + statedb . AccountUpdates
trieproc += statedb . SnapshotStorageReads + statedb . StorageReads + statedb . StorageUpdates
2019-03-27 14:02:04 +03:00
blockExecutionTimer . Update ( time . Since ( substart ) - trieproc - triehash )
2015-10-19 17:08:17 +03:00
// Validate the state using the default validator
2019-03-27 14:02:04 +03:00
substart = time . Now ( )
2019-03-25 13:41:50 +03:00
if err := bc . validator . ValidateState ( block , statedb , receipts , usedGas ) ; err != nil {
2017-05-11 04:55:48 +03:00
bc . reportBlock ( block , receipts , err )
2019-03-25 13:41:50 +03:00
atomic . StoreUint32 ( & followupInterrupt , 1 )
2019-11-29 16:22:08 +03:00
return it . index , err
2015-10-19 17:08:17 +03:00
}
2018-11-20 15:15:26 +03:00
proctime := time . Since ( start )
2018-02-05 19:40:32 +03:00
2019-03-27 14:02:04 +03:00
// Update the metrics touched during block validation
2019-03-25 13:41:50 +03:00
accountHashTimer . Update ( statedb . AccountHashes ) // Account hashes are complete, we can mark them
storageHashTimer . Update ( statedb . StorageHashes ) // Storage hashes are complete, we can mark them
2019-03-27 14:02:04 +03:00
2019-03-25 13:41:50 +03:00
blockValidationTimer . Update ( time . Since ( substart ) - ( statedb . AccountHashes + statedb . StorageHashes - triehash ) )
2019-03-27 14:02:04 +03:00
2017-09-09 19:03:07 +03:00
// Write the block to the chain and get the status.
2019-03-27 14:02:04 +03:00
substart = time . Now ( )
2019-11-29 16:22:08 +03:00
status , err := bc . writeBlockWithState ( block , receipts , logs , statedb , false )
2020-03-23 17:05:15 +03:00
atomic . StoreUint32 ( & followupInterrupt , 1 )
2015-06-29 13:12:30 +03:00
if err != nil {
2019-11-29 16:22:08 +03:00
return it . index , err
2015-06-29 13:12:30 +03:00
}
2019-03-27 14:02:04 +03:00
// Update the metrics touched during block commit
2019-08-06 13:40:28 +03:00
accountCommitTimer . Update ( statedb . AccountCommits ) // Account commits are complete, we can mark them
storageCommitTimer . Update ( statedb . StorageCommits ) // Storage commits are complete, we can mark them
snapshotCommitTimer . Update ( statedb . SnapshotCommits ) // Snapshot commits are complete, we can mark them
2019-03-25 11:01:18 +03:00
2019-08-06 13:40:28 +03:00
blockWriteTimer . Update ( time . Since ( substart ) - statedb . AccountCommits - statedb . StorageCommits - statedb . SnapshotCommits )
2018-11-28 11:29:05 +03:00
blockInsertTimer . UpdateSince ( start )
2019-03-25 11:01:18 +03:00
2015-06-29 13:12:30 +03:00
switch status {
2015-07-03 12:24:42 +03:00
case CanonStatTy :
2018-10-20 11:43:59 +03:00
log . Debug ( "Inserted new block" , "number" , block . Number ( ) , "hash" , block . Hash ( ) ,
"uncles" , len ( block . Uncles ( ) ) , "txs" , len ( block . Transactions ( ) ) , "gas" , block . GasUsed ( ) ,
2018-11-20 15:15:26 +03:00
"elapsed" , common . PrettyDuration ( time . Since ( start ) ) ,
"root" , block . Root ( ) )
2017-09-11 13:13:05 +03:00
lastCanon = block
2018-02-05 19:40:32 +03:00
// Only count canonical blocks for GC processing time
bc . gcproc += proctime
2015-07-03 12:24:42 +03:00
case SideStatTy :
2018-10-20 11:43:59 +03:00
log . Debug ( "Inserted forked block" , "number" , block . Number ( ) , "hash" , block . Hash ( ) ,
2018-11-20 15:15:26 +03:00
"diff" , block . Difficulty ( ) , "elapsed" , common . PrettyDuration ( time . Since ( start ) ) ,
2018-10-20 11:43:59 +03:00
"txs" , len ( block . Transactions ( ) ) , "gas" , block . GasUsed ( ) , "uncles" , len ( block . Uncles ( ) ) ,
2018-11-20 15:15:26 +03:00
"root" , block . Root ( ) )
2019-05-10 17:04:10 +03:00
default :
// This in theory is impossible, but lets be nice to our future selves and leave
// a log, instead of trying to track down blocks imports that don't emit logs.
log . Warn ( "Inserted block with unknown status" , "number" , block . Number ( ) , "hash" , block . Hash ( ) ,
"diff" , block . Difficulty ( ) , "elapsed" , common . PrettyDuration ( time . Since ( start ) ) ,
"txs" , len ( block . Transactions ( ) ) , "gas" , block . GasUsed ( ) , "uncles" , len ( block . Uncles ( ) ) ,
"root" , block . Root ( ) )
2015-06-12 14:36:38 +03:00
}
2015-06-12 17:45:53 +03:00
stats . processed ++
2017-11-13 14:47:27 +03:00
stats . usedGas += usedGas
2018-06-04 10:47:43 +03:00
2019-02-05 13:49:59 +03:00
dirty , _ := bc . stateCache . TrieDB ( ) . Size ( )
stats . report ( chain , it . index , dirty )
2014-11-04 13:46:33 +02:00
}
2021-10-07 16:47:50 +03:00
2018-11-20 15:15:26 +03:00
// Any blocks remaining here? The only ones we care about are the future ones
2020-08-05 10:52:54 +03:00
if block != nil && errors . Is ( err , consensus . ErrFutureBlock ) {
2018-11-20 15:15:26 +03:00
if err := bc . addFutureBlock ( block ) ; err != nil {
2019-11-29 16:22:08 +03:00
return it . index , err
2018-10-20 11:43:59 +03:00
}
2018-11-20 15:15:26 +03:00
block , err = it . next ( )
2020-08-05 10:52:54 +03:00
for ; block != nil && errors . Is ( err , consensus . ErrUnknownAncestor ) ; block , err = it . next ( ) {
2018-11-20 15:15:26 +03:00
if err := bc . addFutureBlock ( block ) ; err != nil {
2019-11-29 16:22:08 +03:00
return it . index , err
2018-10-20 11:43:59 +03:00
}
stats . queued ++
}
2014-11-04 13:46:33 +02:00
}
2018-11-20 15:15:26 +03:00
stats . ignored += it . remaining ( )
2019-11-29 16:22:08 +03:00
return it . index , err
2014-11-17 13:12:55 +02:00
}
2015-01-02 13:07:54 +02:00
2019-05-08 14:30:36 +03:00
// insertSideChain is called when an import batch hits upon a pruned ancestor
2018-11-20 15:15:26 +03:00
// error, which happens when a sidechain with a sufficiently old fork-block is
// found.
//
// The method writes all (header-and-body-valid) blocks to disk, then tries to
// switch over to the new chain if the TD exceeded the current chain.
2019-11-29 16:22:08 +03:00
func ( bc * BlockChain ) insertSideChain ( block * types . Block , it * insertIterator ) ( int , error ) {
2016-10-18 11:18:07 +03:00
var (
2018-11-20 15:15:26 +03:00
externTd * big . Int
2019-01-11 16:27:47 +03:00
current = bc . CurrentBlock ( )
2016-10-18 11:18:07 +03:00
)
2018-11-20 15:15:26 +03:00
// The first sidechain block error is already verified to be ErrPrunedAncestor.
// Since we don't import them here, we expect ErrUnknownAncestor for the remaining
// ones. Any other errors means that the block is invalid, and should not be written
// to disk.
2019-02-04 15:30:19 +03:00
err := consensus . ErrPrunedAncestor
2020-08-05 10:52:54 +03:00
for ; block != nil && errors . Is ( err , consensus . ErrPrunedAncestor ) ; block , err = it . next ( ) {
2018-11-20 15:15:26 +03:00
// Check the canonical state root for that number
2019-01-11 16:27:47 +03:00
if number := block . NumberU64 ( ) ; current . NumberU64 ( ) >= number {
2019-02-21 13:36:49 +03:00
canonical := bc . GetBlockByNumber ( number )
if canonical != nil && canonical . Hash ( ) == block . Hash ( ) {
// Not a sidechain block, this is a re-import of a canon block which has it's state pruned
2019-08-21 10:17:19 +03:00
// Collect the TD of the block. Since we know it's a canon one,
// we can get it directly, and not (like further below) use
// the parent and then add the block on top
externTd = bc . GetTd ( block . Hash ( ) , block . NumberU64 ( ) )
2019-02-21 13:36:49 +03:00
continue
}
if canonical != nil && canonical . Root ( ) == block . Root ( ) {
2018-11-20 15:15:26 +03:00
// This is most likely a shadow-state attack. When a fork is imported into the
// database, and it eventually reaches a block height which is not pruned, we
// just found that the state already exist! This means that the sidechain block
2018-10-20 11:43:59 +03:00
// refers to a state which already exists in our canon chain.
2018-11-20 15:15:26 +03:00
//
// If left unchecked, we would now proceed importing the blocks, without actually
// having verified the state of the previous blocks.
log . Warn ( "Sidechain ghost-state attack detected" , "number" , block . NumberU64 ( ) , "sideroot" , block . Root ( ) , "canonroot" , canonical . Root ( ) )
2018-10-20 11:43:59 +03:00
// If someone legitimately side-mines blocks, they would still be imported as usual. However,
// we cannot risk writing unverified blocks to disk when they obviously target the pruning
// mechanism.
2019-11-29 16:22:08 +03:00
return it . index , errors . New ( "sidechain ghost-state attack" )
2018-10-20 11:43:59 +03:00
}
2016-10-18 11:18:07 +03:00
}
2018-10-20 11:43:59 +03:00
if externTd == nil {
externTd = bc . GetTd ( block . ParentHash ( ) , block . NumberU64 ( ) - 1 )
2018-09-20 11:41:59 +03:00
}
2018-10-20 11:43:59 +03:00
externTd = new ( big . Int ) . Add ( externTd , block . Difficulty ( ) )
2018-09-20 11:41:59 +03:00
2018-10-20 11:43:59 +03:00
if ! bc . HasBlock ( block . Hash ( ) , block . NumberU64 ( ) ) {
2018-11-20 15:15:26 +03:00
start := time . Now ( )
2019-05-08 14:30:36 +03:00
if err := bc . writeBlockWithoutState ( block , externTd ) ; err != nil {
2019-11-29 16:22:08 +03:00
return it . index , err
2018-10-20 11:43:59 +03:00
}
2019-02-08 12:11:31 +03:00
log . Debug ( "Injected sidechain block" , "number" , block . Number ( ) , "hash" , block . Hash ( ) ,
2018-11-20 15:15:26 +03:00
"diff" , block . Difficulty ( ) , "elapsed" , common . PrettyDuration ( time . Since ( start ) ) ,
"txs" , len ( block . Transactions ( ) ) , "gas" , block . GasUsed ( ) , "uncles" , len ( block . Uncles ( ) ) ,
"root" , block . Root ( ) )
2017-02-28 14:35:17 +03:00
}
2018-10-20 11:43:59 +03:00
}
2018-11-20 15:15:26 +03:00
// At this point, we've written all sidechain blocks to database. Loop ended
// either on some other error or all were processed. If there was some other
// error, we can ignore the rest of those blocks.
2018-10-20 11:43:59 +03:00
//
// If the externTd was larger than our local TD, we now need to reimport the previous
// blocks to regenerate the required state
2019-01-11 16:27:47 +03:00
localTd := bc . GetTd ( current . Hash ( ) , current . NumberU64 ( ) )
2018-10-20 11:43:59 +03:00
if localTd . Cmp ( externTd ) > 0 {
2019-03-13 13:31:35 +03:00
log . Info ( "Sidechain written to disk" , "start" , it . first ( ) . NumberU64 ( ) , "end" , it . previous ( ) . Number , "sidetd" , externTd , "localtd" , localTd )
2019-11-29 16:22:08 +03:00
return it . index , err
2018-10-20 11:43:59 +03:00
}
2018-11-20 15:15:26 +03:00
// Gather all the sidechain hashes (full blocks may be memory heavy)
var (
hashes [ ] common . Hash
numbers [ ] uint64
)
2019-03-13 13:31:35 +03:00
parent := it . previous ( )
2018-11-20 15:15:26 +03:00
for parent != nil && ! bc . HasState ( parent . Root ) {
hashes = append ( hashes , parent . Hash ( ) )
numbers = append ( numbers , parent . Number . Uint64 ( ) )
2016-10-18 11:18:07 +03:00
2018-11-20 15:15:26 +03:00
parent = bc . GetHeader ( parent . ParentHash , parent . Number . Uint64 ( ) - 1 )
2016-10-07 15:25:01 +03:00
}
2018-11-20 15:15:26 +03:00
if parent == nil {
2019-11-29 16:22:08 +03:00
return it . index , errors . New ( "missing parent" )
2018-10-20 11:43:59 +03:00
}
// Import all the pruned blocks to make the state available
2016-10-18 11:18:07 +03:00
var (
2018-11-20 15:15:26 +03:00
blocks [ ] * types . Block
memory common . StorageSize
2016-10-18 11:18:07 +03:00
)
2018-11-20 15:15:26 +03:00
for i := len ( hashes ) - 1 ; i >= 0 ; i -- {
// Append the next block to our batch
block := bc . GetBlock ( hashes [ i ] , numbers [ i ] )
2018-09-20 11:41:59 +03:00
2018-11-20 15:15:26 +03:00
blocks = append ( blocks , block )
memory += block . Size ( )
// If memory use grew too large, import and continue. Sadly we need to discard
// all raised events and logs from notifications since we're too heavy on the
// memory here.
if len ( blocks ) >= 2048 || memory > 64 * 1024 * 1024 {
log . Info ( "Importing heavy sidechain segment" , "blocks" , len ( blocks ) , "start" , blocks [ 0 ] . NumberU64 ( ) , "end" , block . NumberU64 ( ) )
2019-11-29 16:22:08 +03:00
if _ , err := bc . insertChain ( blocks , false ) ; err != nil {
return 0 , err
2018-11-20 15:15:26 +03:00
}
blocks , memory = blocks [ : 0 ] , 0
2016-10-07 15:25:01 +03:00
2018-11-20 15:15:26 +03:00
// If the chain is terminating, stop processing blocks
2020-05-26 22:37:37 +03:00
if bc . insertStopped ( ) {
log . Debug ( "Abort during blocks processing" )
2019-11-29 16:22:08 +03:00
return 0 , nil
2018-11-20 15:15:26 +03:00
}
}
2016-10-07 15:25:01 +03:00
}
2018-11-20 15:15:26 +03:00
if len ( blocks ) > 0 {
log . Info ( "Importing sidechain segment" , "start" , blocks [ 0 ] . NumberU64 ( ) , "end" , blocks [ len ( blocks ) - 1 ] . NumberU64 ( ) )
return bc . insertChain ( blocks , false )
2016-10-07 15:25:01 +03:00
}
2019-11-29 16:22:08 +03:00
return 0 , nil
2016-10-07 15:25:01 +03:00
}
2019-04-04 14:39:11 +03:00
// reorg takes two blocks, an old chain and a new chain and will reconstruct the
// blocks and inserts them to be part of the new canonical chain and accumulates
// potential missing transactions and post an event about them.
2017-05-11 04:55:48 +03:00
func ( bc * BlockChain ) reorg ( oldBlock , newBlock * types . Block ) error {
2015-05-13 23:27:18 +03:00
var (
2016-12-04 21:07:24 +03:00
newChain types . Blocks
oldChain types . Blocks
commonBlock * types . Block
2019-04-04 14:39:11 +03:00
deletedTxs types . Transactions
addedTxs types . Transactions
2019-11-29 16:22:08 +03:00
deletedLogs [ ] [ ] * types . Log
rebirthLogs [ ] [ ] * types . Log
2019-04-04 14:39:11 +03:00
2019-11-29 16:22:08 +03:00
// collectLogs collects the logs that were generated or removed during
// the processing of the block that corresponds with the given hash.
2018-12-17 10:23:54 +03:00
// These logs are later announced as deleted or reborn
collectLogs = func ( hash common . Hash , removed bool ) {
2018-05-07 14:35:06 +03:00
number := bc . hc . GetBlockNumber ( hash )
if number == nil {
return
}
2019-04-15 12:36:27 +03:00
receipts := rawdb . ReadReceipts ( bc . db , hash , * number , bc . chainConfig )
2019-11-29 16:22:08 +03:00
var logs [ ] * types . Log
2015-12-01 01:11:24 +02:00
for _ , receipt := range receipts {
2016-12-04 21:07:24 +03:00
for _ , log := range receipt . Logs {
2018-12-17 10:23:54 +03:00
l := * log
if removed {
l . Removed = true
}
2019-11-29 16:22:08 +03:00
logs = append ( logs , & l )
}
}
if len ( logs ) > 0 {
if removed {
deletedLogs = append ( deletedLogs , logs )
} else {
rebirthLogs = append ( rebirthLogs , logs )
2016-12-04 21:07:24 +03:00
}
2015-12-01 01:11:24 +02:00
}
}
2019-11-29 16:22:08 +03:00
// mergeLogs returns a merged log slice with specified sort order.
mergeLogs = func ( logs [ ] [ ] * types . Log , reverse bool ) [ ] * types . Log {
var ret [ ] * types . Log
if reverse {
for i := len ( logs ) - 1 ; i >= 0 ; i -- {
ret = append ( ret , logs [ i ] ... )
}
} else {
for i := 0 ; i < len ( logs ) ; i ++ {
ret = append ( ret , logs [ i ] ... )
}
}
return ret
}
2015-05-13 23:27:18 +03:00
)
2019-04-04 14:39:11 +03:00
// Reduce the longer chain to the same number as the shorter one
2015-05-15 16:30:34 +03:00
if oldBlock . NumberU64 ( ) > newBlock . NumberU64 ( ) {
2019-04-04 14:39:11 +03:00
// Old chain is longer, gather all transactions and logs as deleted ones
2017-05-11 04:55:48 +03:00
for ; oldBlock != nil && oldBlock . NumberU64 ( ) != newBlock . NumberU64 ( ) ; oldBlock = bc . GetBlock ( oldBlock . ParentHash ( ) , oldBlock . NumberU64 ( ) - 1 ) {
2016-03-07 19:11:52 +02:00
oldChain = append ( oldChain , oldBlock )
2015-08-17 15:01:41 +03:00
deletedTxs = append ( deletedTxs , oldBlock . Transactions ( ) ... )
2018-12-17 10:23:54 +03:00
collectLogs ( oldBlock . Hash ( ) , true )
2015-05-15 16:30:34 +03:00
}
} else {
2019-04-04 14:39:11 +03:00
// New chain is longer, stash all blocks away for subsequent insertion
2017-05-11 04:55:48 +03:00
for ; newBlock != nil && newBlock . NumberU64 ( ) != oldBlock . NumberU64 ( ) ; newBlock = bc . GetBlock ( newBlock . ParentHash ( ) , newBlock . NumberU64 ( ) - 1 ) {
2015-05-15 16:30:34 +03:00
newChain = append ( newChain , newBlock )
}
2015-04-29 13:43:24 +03:00
}
2015-05-28 19:18:23 +03:00
if oldBlock == nil {
2019-04-04 14:39:11 +03:00
return fmt . Errorf ( "invalid old chain" )
2015-05-28 19:18:23 +03:00
}
if newBlock == nil {
2019-04-04 14:39:11 +03:00
return fmt . Errorf ( "invalid new chain" )
2015-05-28 19:18:23 +03:00
}
2019-04-04 14:39:11 +03:00
// Both sides of the reorg are at the same number, reduce both until the common
// ancestor is found
2015-04-14 01:18:38 +03:00
for {
2019-04-04 14:39:11 +03:00
// If the common ancestor was found, bail out
2015-04-14 01:18:38 +03:00
if oldBlock . Hash ( ) == newBlock . Hash ( ) {
2015-05-13 23:27:18 +03:00
commonBlock = oldBlock
2015-04-14 01:18:38 +03:00
break
}
2019-04-04 14:39:11 +03:00
// Remove an old block as well as stash away a new block
2016-03-07 19:11:52 +02:00
oldChain = append ( oldChain , oldBlock )
2015-08-17 15:01:41 +03:00
deletedTxs = append ( deletedTxs , oldBlock . Transactions ( ) ... )
2018-12-17 10:23:54 +03:00
collectLogs ( oldBlock . Hash ( ) , true )
2015-04-29 13:43:24 +03:00
2019-04-04 14:39:11 +03:00
newChain = append ( newChain , newBlock )
// Step back with both chains
oldBlock = bc . GetBlock ( oldBlock . ParentHash ( ) , oldBlock . NumberU64 ( ) - 1 )
2015-05-28 16:35:50 +03:00
if oldBlock == nil {
2019-04-04 14:39:11 +03:00
return fmt . Errorf ( "invalid old chain" )
2015-05-28 16:35:50 +03:00
}
2019-04-04 14:39:11 +03:00
newBlock = bc . GetBlock ( newBlock . ParentHash ( ) , newBlock . NumberU64 ( ) - 1 )
2015-05-28 16:35:50 +03:00
if newBlock == nil {
2019-04-04 14:39:11 +03:00
return fmt . Errorf ( "invalid new chain" )
2015-05-28 16:35:50 +03:00
}
2015-04-14 01:18:38 +03:00
}
2017-02-22 15:10:07 +03:00
// Ensure the user sees large reorgs
2017-03-03 10:54:13 +03:00
if len ( oldChain ) > 0 && len ( newChain ) > 0 {
2019-08-22 12:28:23 +03:00
logFn := log . Info
msg := "Chain reorg detected"
2017-03-03 10:54:13 +03:00
if len ( oldChain ) > 63 {
2019-08-22 12:28:23 +03:00
msg = "Large chain reorg detected"
2017-03-03 10:54:13 +03:00
logFn = log . Warn
}
2019-08-22 12:28:23 +03:00
logFn ( msg , "number" , commonBlock . Number ( ) , "hash" , commonBlock . Hash ( ) ,
2017-03-03 10:54:13 +03:00
"drop" , len ( oldChain ) , "dropfrom" , oldChain [ 0 ] . Hash ( ) , "add" , len ( newChain ) , "addfrom" , newChain [ 0 ] . Hash ( ) )
2019-08-22 12:28:23 +03:00
blockReorgAddMeter . Mark ( int64 ( len ( newChain ) ) )
blockReorgDropMeter . Mark ( int64 ( len ( oldChain ) ) )
2020-08-20 10:49:35 +03:00
blockReorgMeter . Mark ( 1 )
2017-03-03 10:54:13 +03:00
} else {
log . Error ( "Impossible reorg, please file an issue" , "oldnum" , oldBlock . Number ( ) , "oldhash" , oldBlock . Hash ( ) , "newnum" , newBlock . Number ( ) , "newhash" , newBlock . Hash ( ) )
2017-02-22 15:10:07 +03:00
}
2019-05-08 14:30:36 +03:00
// Insert the new chain(except the head block(reverse order)),
// taking care of the proper incremental order.
for i := len ( newChain ) - 1 ; i >= 1 ; i -- {
2019-04-04 14:39:11 +03:00
// Insert the block in the canonical way, re-writing history
2020-01-17 13:49:32 +03:00
bc . writeHeadBlock ( newChain [ i ] )
2019-04-04 14:39:11 +03:00
2019-05-08 14:30:36 +03:00
// Collect reborn logs due to chain reorg
collectLogs ( newChain [ i ] . Hash ( ) , false )
2020-01-17 13:49:32 +03:00
// Collect the new added transactions.
2018-01-22 15:07:47 +03:00
addedTxs = append ( addedTxs , newChain [ i ] . Transactions ( ) ... )
2015-08-17 15:01:41 +03:00
}
2020-01-17 13:49:32 +03:00
// Delete useless indexes right now which includes the non-canonical
// transaction indexes, canonical chain indexes which above the head.
indexesBatch := bc . db . NewBatch ( )
2019-04-04 14:39:11 +03:00
for _ , tx := range types . TxDifference ( deletedTxs , addedTxs ) {
2020-01-17 13:49:32 +03:00
rawdb . DeleteTxLookupEntry ( indexesBatch , tx . Hash ( ) )
2015-04-14 01:18:38 +03:00
}
2019-05-07 15:26:00 +03:00
// Delete any canonical number assignments above the new head
number := bc . CurrentBlock ( ) . NumberU64 ( )
for i := number + 1 ; ; i ++ {
hash := rawdb . ReadCanonicalHash ( bc . db , i )
if hash == ( common . Hash { } ) {
break
}
2020-01-17 13:49:32 +03:00
rawdb . DeleteCanonicalHash ( indexesBatch , i )
}
if err := indexesBatch . Write ( ) ; err != nil {
log . Crit ( "Failed to delete useless indexes" , "err" , err )
2019-05-07 15:26:00 +03:00
}
2019-04-04 14:39:11 +03:00
// If any logs need to be fired, do it now. In theory we could avoid creating
// this goroutine if there are no events to fire, but realistcally that only
// ever happens if we're reorging empty blocks, which will only happen on idle
// networks where performance is not an issue either way.
2019-11-29 16:22:08 +03:00
if len ( deletedLogs ) > 0 {
bc . rmLogsFeed . Send ( RemovedLogsEvent { mergeLogs ( deletedLogs , true ) } )
2017-08-18 13:58:36 +03:00
}
2019-11-29 16:22:08 +03:00
if len ( rebirthLogs ) > 0 {
bc . logsFeed . Send ( mergeLogs ( rebirthLogs , false ) )
}
if len ( oldChain ) > 0 {
for i := len ( oldChain ) - 1 ; i >= 0 ; i -- {
bc . chainSideFeed . Send ( ChainSideEvent { Block : oldChain [ i ] } )
2015-10-12 15:04:38 +03:00
}
}
2019-11-29 16:22:08 +03:00
return nil
2015-10-12 15:04:38 +03:00
}
2021-10-07 16:47:50 +03:00
// futureBlocksLoop processes the 'future block' queue.
func ( bc * BlockChain ) futureBlocksLoop ( ) {
defer bc . wg . Done ( )
2018-01-02 14:50:46 +03:00
futureTimer := time . NewTicker ( 5 * time . Second )
defer futureTimer . Stop ( )
2015-03-06 16:50:44 +02:00
for {
select {
2018-01-02 14:50:46 +03:00
case <- futureTimer . C :
2017-05-11 04:55:48 +03:00
bc . procFutureBlocks ( )
case <- bc . quit :
2015-10-12 15:04:38 +03:00
return
2015-03-06 16:50:44 +02:00
}
}
}
2015-05-17 02:42:30 +03:00
2021-11-01 16:09:36 +03:00
// skipBlock returns 'true', if the block being imported can be skipped over, meaning
// that the block does not need to be processed but can be considered already fully 'done'.
func ( bc * BlockChain ) skipBlock ( err error , it * insertIterator ) bool {
// We can only ever bypass processing if the only error returned by the validator
// is ErrKnownBlock, which means all checks passed, but we already have the block
// and state.
if ! errors . Is ( err , ErrKnownBlock ) {
return false
}
// If we're not using snapshots, we can skip this, since we have both block
// and (trie-) state
if bc . snaps == nil {
return true
}
var (
header = it . current ( ) // header can't be nil
parentRoot common . Hash
)
// If we also have the snapshot-state, we can skip the processing.
if bc . snaps . Snapshot ( header . Root ) != nil {
return true
}
// In this case, we have the trie-state but not snapshot-state. If the parent
// snapshot-state exists, we need to process this in order to not get a gap
// in the snapshot layers.
// Resolve parent block
if parent := it . previous ( ) ; parent != nil {
parentRoot = parent . Root
} else if parent = bc . GetHeaderByHash ( header . ParentHash ) ; parent != nil {
parentRoot = parent . Root
}
if parentRoot == ( common . Hash { } ) {
return false // Theoretically impossible case
}
// Parent is also missing snapshot: we can skip this. Otherwise process.
if bc . snaps . Snapshot ( parentRoot ) == nil {
return true
}
return false
}
2020-05-11 18:58:43 +03:00
// maintainTxIndex is responsible for the construction and deletion of the
// transaction index.
//
// User can use flag `txlookuplimit` to specify a "recentness" block, below
// which ancient tx indices get deleted. If `txlookuplimit` is 0, it means
// all tx indices will be reserved.
//
// The user can adjust the txlookuplimit value for each launch after fast
// sync, Geth will automatically construct the missing indices and delete
// the extra indices.
func ( bc * BlockChain ) maintainTxIndex ( ancients uint64 ) {
2020-10-20 17:34:50 +03:00
defer bc . wg . Done ( )
2020-05-11 18:58:43 +03:00
// Before starting the actual maintenance, we need to handle a special case,
// where user might init Geth with an external ancient database. If so, we
// need to reindex all necessary transactions before starting to process any
// pruning requests.
if ancients > 0 {
var from = uint64 ( 0 )
if bc . txLookupLimit != 0 && ancients > bc . txLookupLimit {
from = ancients - bc . txLookupLimit
}
2020-10-20 17:34:50 +03:00
rawdb . IndexTransactions ( bc . db , from , ancients , bc . quit )
2020-05-11 18:58:43 +03:00
}
2021-10-07 16:47:50 +03:00
2020-05-11 18:58:43 +03:00
// indexBlocks reindexes or unindexes transactions depending on user configuration
indexBlocks := func ( tail * uint64 , head uint64 , done chan struct { } ) {
defer func ( ) { done <- struct { } { } } ( )
// If the user just upgraded Geth to a new version which supports transaction
// index pruning, write the new tail and remove anything older.
if tail == nil {
if bc . txLookupLimit == 0 || head < bc . txLookupLimit {
// Nothing to delete, write the tail and return
rawdb . WriteTxIndexTail ( bc . db , 0 )
} else {
// Prune all stale tx indices and record the tx index tail
2020-10-20 17:34:50 +03:00
rawdb . UnindexTransactions ( bc . db , 0 , head - bc . txLookupLimit + 1 , bc . quit )
2020-05-11 18:58:43 +03:00
}
return
}
// If a previous indexing existed, make sure that we fill in any missing entries
if bc . txLookupLimit == 0 || head < bc . txLookupLimit {
if * tail > 0 {
2020-10-20 17:34:50 +03:00
rawdb . IndexTransactions ( bc . db , 0 , * tail , bc . quit )
2020-05-11 18:58:43 +03:00
}
return
}
// Update the transaction index to the new chain state
if head - bc . txLookupLimit + 1 < * tail {
// Reindex a part of missing indices and rewind index tail to HEAD-limit
2020-10-20 17:34:50 +03:00
rawdb . IndexTransactions ( bc . db , head - bc . txLookupLimit + 1 , * tail , bc . quit )
2020-05-11 18:58:43 +03:00
} else {
// Unindex a part of stale indices and forward index tail to HEAD-limit
2020-10-20 17:34:50 +03:00
rawdb . UnindexTransactions ( bc . db , * tail , head - bc . txLookupLimit + 1 , bc . quit )
2020-05-11 18:58:43 +03:00
}
}
2021-10-07 16:47:50 +03:00
2020-05-11 18:58:43 +03:00
// Any reindexing done, start listening to chain events and moving the index window
var (
done chan struct { } // Non-nil if background unindexing or reindexing routine is active.
headCh = make ( chan ChainHeadEvent , 1 ) // Buffered to avoid locking up the event feed
)
sub := bc . SubscribeChainHeadEvent ( headCh )
if sub == nil {
return
}
defer sub . Unsubscribe ( )
for {
select {
case head := <- headCh :
if done == nil {
done = make ( chan struct { } )
go indexBlocks ( rawdb . ReadTxIndexTail ( bc . db ) , head . Block . NumberU64 ( ) , done )
}
case <- done :
done = nil
case <- bc . quit :
2020-10-20 17:34:50 +03:00
if done != nil {
log . Info ( "Waiting background transaction indexer to exit" )
<- done
}
2020-05-11 18:58:43 +03:00
return
}
}
}
2016-05-24 19:49:54 +03:00
// reportBlock logs a bad block error.
2016-11-23 15:32:25 +03:00
func ( bc * BlockChain ) reportBlock ( block * types . Block , receipts types . Receipts , err error ) {
2021-01-10 14:54:15 +03:00
rawdb . WriteBadBlock ( bc . db , block )
2017-02-28 14:35:17 +03:00
var receiptString string
2018-11-22 12:00:16 +03:00
for i , receipt := range receipts {
receiptString += fmt . Sprintf ( "\t %d: cumulative: %v gas: %v contract: %v status: %v tx: %v logs: %v bloom: %x state: %x\n" ,
i , receipt . CumulativeGasUsed , receipt . GasUsed , receipt . ContractAddress . Hex ( ) ,
receipt . Status , receipt . TxHash . Hex ( ) , receipt . Logs , receipt . Bloom , receipt . PostState )
2017-02-28 14:35:17 +03:00
}
log . Error ( fmt . Sprintf ( `
2016-11-23 15:32:25 +03:00
# # # # # # # # # # BAD BLOCK # # # # # # # # #
Chain config : % v
Number : % v
Hash : 0 x % x
% v
Error : % v
# # # # # # # # # # # # # # # # # # # # # # # # # # # # # #
2018-02-05 19:40:32 +03:00
` , bc . chainConfig , block . Number ( ) , block . Hash ( ) , receiptString , err ) )
2015-05-29 19:07:23 +03:00
}
2015-12-16 05:26:23 +02:00
// InsertHeaderChain attempts to insert the given header chain in to the local
// chain, possibly creating a reorg. If an error is returned, it will return the
// index number of the failing header as well an error describing what went wrong.
//
// The verify parameter can be used to fine tune whether nonce verification
// should be done or not. The reason behind the optional check is because some
2016-03-15 18:12:03 +02:00
// of the header retrieval mechanisms already need to verify nonces, as well as
2015-12-16 05:26:23 +02:00
// because nonces can be verified sparsely, not needing to check each.
2017-05-11 04:55:48 +03:00
func ( bc * BlockChain ) InsertHeaderChain ( chain [ ] * types . Header , checkFreq int ) ( int , error ) {
2017-03-22 22:44:22 +03:00
start := time . Now ( )
2017-05-11 04:55:48 +03:00
if i , err := bc . hc . ValidateHeaderChain ( chain , checkFreq ) ; err != nil {
2017-03-22 22:44:22 +03:00
return i , err
}
2021-10-07 16:47:50 +03:00
if ! bc . chainmu . TryLock ( ) {
return 0 , errChainStopped
}
2017-05-11 04:55:48 +03:00
defer bc . chainmu . Unlock ( )
2020-12-09 13:13:02 +03:00
_ , err := bc . hc . InsertHeaderChain ( chain , start )
return 0 , err
2015-12-16 05:26:23 +02:00
}