2022-01-20 14:29:06 +03:00
|
|
|
// Copyright 2022 The go-ethereum Authors
|
|
|
|
// This file is part of the go-ethereum library.
|
|
|
|
//
|
|
|
|
// The go-ethereum library is free software: you can redistribute it and/or modify
|
|
|
|
// it under the terms of the GNU Lesser General Public License as published by
|
|
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
|
|
// (at your option) any later version.
|
|
|
|
//
|
|
|
|
// The go-ethereum library is distributed in the hope that it will be useful,
|
|
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
|
|
// GNU Lesser General Public License for more details.
|
|
|
|
//
|
|
|
|
// You should have received a copy of the GNU Lesser General Public License
|
|
|
|
// along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
|
|
|
|
package catalyst
|
|
|
|
|
2022-01-31 15:22:35 +03:00
|
|
|
import (
|
|
|
|
"sync"
|
2022-05-18 17:33:37 +03:00
|
|
|
"time"
|
2022-01-31 15:22:35 +03:00
|
|
|
|
2022-03-11 15:14:45 +03:00
|
|
|
"github.com/ethereum/go-ethereum/common"
|
2022-01-31 15:22:35 +03:00
|
|
|
"github.com/ethereum/go-ethereum/core/beacon"
|
2022-03-11 15:14:45 +03:00
|
|
|
"github.com/ethereum/go-ethereum/core/types"
|
2022-01-31 15:22:35 +03:00
|
|
|
)
|
2022-01-20 14:29:06 +03:00
|
|
|
|
|
|
|
// maxTrackedPayloads is the maximum number of prepared payloads the execution
|
|
|
|
// engine tracks before evicting old ones. Ideally we should only ever track the
|
|
|
|
// latest one; but have a slight wiggle room for non-ideal conditions.
|
|
|
|
const maxTrackedPayloads = 10
|
|
|
|
|
2022-03-11 15:14:45 +03:00
|
|
|
// maxTrackedHeaders is the maximum number of executed payloads the execution
|
|
|
|
// engine tracks before evicting old ones. Ideally we should only ever track the
|
|
|
|
// latest one; but have a slight wiggle room for non-ideal conditions.
|
|
|
|
const maxTrackedHeaders = 10
|
|
|
|
|
2022-05-18 17:33:37 +03:00
|
|
|
// payload wraps the miner's block production channel, allowing the mined block
|
|
|
|
// to be retrieved later upon the GetPayload engine API call.
|
|
|
|
type payload struct {
|
|
|
|
lock sync.Mutex
|
|
|
|
done bool
|
|
|
|
empty *types.Block
|
|
|
|
block *types.Block
|
|
|
|
result chan *types.Block
|
|
|
|
}
|
|
|
|
|
|
|
|
// resolve extracts the generated full block from the given channel if possible
|
|
|
|
// or fallback to empty block as an alternative.
|
|
|
|
func (req *payload) resolve() *beacon.ExecutableDataV1 {
|
|
|
|
// this function can be called concurrently, prevent any
|
|
|
|
// concurrency issue in the first place.
|
|
|
|
req.lock.Lock()
|
|
|
|
defer req.lock.Unlock()
|
|
|
|
|
|
|
|
// Try to resolve the full block first if it's not obtained
|
|
|
|
// yet. The returned block can be nil if the generation fails.
|
|
|
|
|
|
|
|
if !req.done {
|
|
|
|
timeout := time.NewTimer(500 * time.Millisecond)
|
|
|
|
defer timeout.Stop()
|
|
|
|
|
|
|
|
select {
|
|
|
|
case req.block = <-req.result:
|
|
|
|
req.done = true
|
|
|
|
case <-timeout.C:
|
|
|
|
// TODO(rjl49345642, Marius), should we keep this
|
|
|
|
// 100ms timeout allowance? Why not just use the
|
|
|
|
// default and then fallback to empty directly?
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if req.block != nil {
|
|
|
|
return beacon.BlockToExecutableData(req.block)
|
|
|
|
}
|
|
|
|
return beacon.BlockToExecutableData(req.empty)
|
|
|
|
}
|
|
|
|
|
2022-01-20 14:29:06 +03:00
|
|
|
// payloadQueueItem represents an id->payload tuple to store until it's retrieved
|
|
|
|
// or evicted.
|
|
|
|
type payloadQueueItem struct {
|
2022-05-18 17:33:37 +03:00
|
|
|
id beacon.PayloadID
|
|
|
|
data *payload
|
2022-01-20 14:29:06 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
// payloadQueue tracks the latest handful of constructed payloads to be retrieved
|
|
|
|
// by the beacon chain if block production is requested.
|
|
|
|
type payloadQueue struct {
|
|
|
|
payloads []*payloadQueueItem
|
|
|
|
lock sync.RWMutex
|
|
|
|
}
|
|
|
|
|
|
|
|
// newPayloadQueue creates a pre-initialized queue with a fixed number of slots
|
|
|
|
// all containing empty items.
|
|
|
|
func newPayloadQueue() *payloadQueue {
|
|
|
|
return &payloadQueue{
|
|
|
|
payloads: make([]*payloadQueueItem, maxTrackedPayloads),
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// put inserts a new payload into the queue at the given id.
|
2022-05-18 17:33:37 +03:00
|
|
|
func (q *payloadQueue) put(id beacon.PayloadID, data *payload) {
|
2022-01-20 14:29:06 +03:00
|
|
|
q.lock.Lock()
|
|
|
|
defer q.lock.Unlock()
|
|
|
|
|
|
|
|
copy(q.payloads[1:], q.payloads)
|
|
|
|
q.payloads[0] = &payloadQueueItem{
|
2022-05-18 17:33:37 +03:00
|
|
|
id: id,
|
|
|
|
data: data,
|
2022-01-20 14:29:06 +03:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// get retrieves a previously stored payload item or nil if it does not exist.
|
2022-01-31 15:22:35 +03:00
|
|
|
func (q *payloadQueue) get(id beacon.PayloadID) *beacon.ExecutableDataV1 {
|
2022-01-20 14:29:06 +03:00
|
|
|
q.lock.RLock()
|
|
|
|
defer q.lock.RUnlock()
|
|
|
|
|
|
|
|
for _, item := range q.payloads {
|
|
|
|
if item == nil {
|
|
|
|
return nil // no more items
|
|
|
|
}
|
|
|
|
if item.id == id {
|
2022-05-18 17:33:37 +03:00
|
|
|
return item.data.resolve()
|
2022-01-20 14:29:06 +03:00
|
|
|
}
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}
|
2022-03-11 15:14:45 +03:00
|
|
|
|
|
|
|
// headerQueueItem represents an hash->header tuple to store until it's retrieved
|
|
|
|
// or evicted.
|
|
|
|
type headerQueueItem struct {
|
|
|
|
hash common.Hash
|
|
|
|
header *types.Header
|
|
|
|
}
|
|
|
|
|
|
|
|
// headerQueue tracks the latest handful of constructed headers to be retrieved
|
|
|
|
// by the beacon chain if block production is requested.
|
|
|
|
type headerQueue struct {
|
|
|
|
headers []*headerQueueItem
|
|
|
|
lock sync.RWMutex
|
|
|
|
}
|
|
|
|
|
|
|
|
// newHeaderQueue creates a pre-initialized queue with a fixed number of slots
|
|
|
|
// all containing empty items.
|
|
|
|
func newHeaderQueue() *headerQueue {
|
|
|
|
return &headerQueue{
|
|
|
|
headers: make([]*headerQueueItem, maxTrackedHeaders),
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// put inserts a new header into the queue at the given hash.
|
|
|
|
func (q *headerQueue) put(hash common.Hash, data *types.Header) {
|
|
|
|
q.lock.Lock()
|
|
|
|
defer q.lock.Unlock()
|
|
|
|
|
|
|
|
copy(q.headers[1:], q.headers)
|
|
|
|
q.headers[0] = &headerQueueItem{
|
|
|
|
hash: hash,
|
|
|
|
header: data,
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// get retrieves a previously stored header item or nil if it does not exist.
|
|
|
|
func (q *headerQueue) get(hash common.Hash) *types.Header {
|
|
|
|
q.lock.RLock()
|
|
|
|
defer q.lock.RUnlock()
|
|
|
|
|
|
|
|
for _, item := range q.headers {
|
|
|
|
if item == nil {
|
|
|
|
return nil // no more items
|
|
|
|
}
|
|
|
|
if item.hash == hash {
|
|
|
|
return item.header
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return nil
|
|
|
|
}
|