2a6beb6a39
This PR removes the newly added txpool.Transaction wrapper type, and instead adds a way of keeping the blob sidecar within types.Transaction. It's better this way because most code in go-ethereum does not care about blob transactions, and probably never will. This will start mattering especially on the client side of RPC, where all APIs are based on types.Transaction. Users need to be able to use the same signing flows they already have. However, since blobs are only allowed in some places but not others, we will now need to add checks to avoid creating invalid blocks. I'm still trying to figure out the best place to do some of these. The way I have it currently is as follows: - In block validation (import), txs are verified not to have a blob sidecar. - In miner, we strip off the sidecar when committing the transaction into the block. - In TxPool validation, txs must have a sidecar to be added into the blobpool. - Note there is a special case here: when transactions are re-added because of a chain reorg, we cannot use the transactions gathered from the old chain blocks as-is, because they will be missing their blobs. This was previously handled by storing the blobs into the 'blobpool limbo'. The code has now changed to store the full transaction in the limbo instead, but it might be confusing for code readers why we're not simply adding the types.Transaction we already have. Code changes summary: - txpool.Transaction removed and all uses replaced by types.Transaction again - blobpool now stores types.Transaction instead of defining its own blobTx format for storage - the blobpool limbo now stores types.Transaction instead of storing only the blobs - checks to validate the presence/absence of the blob sidecar added in certain critical places
207 lines
6.5 KiB
Go
207 lines
6.5 KiB
Go
// Copyright 2020 The go-ethereum Authors
|
|
// This file is part of the go-ethereum library.
|
|
//
|
|
// The go-ethereum library is free software: you can redistribute it and/or modify
|
|
// it under the terms of the GNU Lesser General Public License as published by
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
// (at your option) any later version.
|
|
//
|
|
// The go-ethereum library is distributed in the hope that it will be useful,
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
// GNU Lesser General Public License for more details.
|
|
//
|
|
// You should have received a copy of the GNU Lesser General Public License
|
|
// along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
package eth
|
|
|
|
import (
|
|
"math/big"
|
|
|
|
"github.com/ethereum/go-ethereum/common"
|
|
"github.com/ethereum/go-ethereum/core/types"
|
|
)
|
|
|
|
const (
|
|
// This is the target size for the packs of transactions or announcements. A
|
|
// pack can get larger than this if a single transactions exceeds this size.
|
|
maxTxPacketSize = 100 * 1024
|
|
)
|
|
|
|
// blockPropagation is a block propagation event, waiting for its turn in the
|
|
// broadcast queue.
|
|
type blockPropagation struct {
|
|
block *types.Block
|
|
td *big.Int
|
|
}
|
|
|
|
// broadcastBlocks is a write loop that multiplexes blocks and block announcements
|
|
// to the remote peer. The goal is to have an async writer that does not lock up
|
|
// node internals and at the same time rate limits queued data.
|
|
func (p *Peer) broadcastBlocks() {
|
|
for {
|
|
select {
|
|
case prop := <-p.queuedBlocks:
|
|
if err := p.SendNewBlock(prop.block, prop.td); err != nil {
|
|
return
|
|
}
|
|
p.Log().Trace("Propagated block", "number", prop.block.Number(), "hash", prop.block.Hash(), "td", prop.td)
|
|
|
|
case block := <-p.queuedBlockAnns:
|
|
if err := p.SendNewBlockHashes([]common.Hash{block.Hash()}, []uint64{block.NumberU64()}); err != nil {
|
|
return
|
|
}
|
|
p.Log().Trace("Announced block", "number", block.Number(), "hash", block.Hash())
|
|
|
|
case <-p.term:
|
|
return
|
|
}
|
|
}
|
|
}
|
|
|
|
// broadcastTransactions is a write loop that schedules transaction broadcasts
|
|
// to the remote peer. The goal is to have an async writer that does not lock up
|
|
// node internals and at the same time rate limits queued data.
|
|
func (p *Peer) broadcastTransactions() {
|
|
var (
|
|
queue []common.Hash // Queue of hashes to broadcast as full transactions
|
|
done chan struct{} // Non-nil if background broadcaster is running
|
|
fail = make(chan error, 1) // Channel used to receive network error
|
|
failed bool // Flag whether a send failed, discard everything onward
|
|
)
|
|
for {
|
|
// If there's no in-flight broadcast running, check if a new one is needed
|
|
if done == nil && len(queue) > 0 {
|
|
// Pile transaction until we reach our allowed network limit
|
|
var (
|
|
hashesCount uint64
|
|
txs []*types.Transaction
|
|
size common.StorageSize
|
|
)
|
|
for i := 0; i < len(queue) && size < maxTxPacketSize; i++ {
|
|
if tx := p.txpool.Get(queue[i]); tx != nil {
|
|
txs = append(txs, tx)
|
|
size += common.StorageSize(tx.Size())
|
|
}
|
|
hashesCount++
|
|
}
|
|
queue = queue[:copy(queue, queue[hashesCount:])]
|
|
|
|
// If there's anything available to transfer, fire up an async writer
|
|
if len(txs) > 0 {
|
|
done = make(chan struct{})
|
|
go func() {
|
|
if err := p.SendTransactions(txs); err != nil {
|
|
fail <- err
|
|
return
|
|
}
|
|
close(done)
|
|
p.Log().Trace("Sent transactions", "count", len(txs))
|
|
}()
|
|
}
|
|
}
|
|
// Transfer goroutine may or may not have been started, listen for events
|
|
select {
|
|
case hashes := <-p.txBroadcast:
|
|
// If the connection failed, discard all transaction events
|
|
if failed {
|
|
continue
|
|
}
|
|
// New batch of transactions to be broadcast, queue them (with cap)
|
|
queue = append(queue, hashes...)
|
|
if len(queue) > maxQueuedTxs {
|
|
// Fancy copy and resize to ensure buffer doesn't grow indefinitely
|
|
queue = queue[:copy(queue, queue[len(queue)-maxQueuedTxs:])]
|
|
}
|
|
|
|
case <-done:
|
|
done = nil
|
|
|
|
case <-fail:
|
|
failed = true
|
|
|
|
case <-p.term:
|
|
return
|
|
}
|
|
}
|
|
}
|
|
|
|
// announceTransactions is a write loop that schedules transaction broadcasts
|
|
// to the remote peer. The goal is to have an async writer that does not lock up
|
|
// node internals and at the same time rate limits queued data.
|
|
func (p *Peer) announceTransactions() {
|
|
var (
|
|
queue []common.Hash // Queue of hashes to announce as transaction stubs
|
|
done chan struct{} // Non-nil if background announcer is running
|
|
fail = make(chan error, 1) // Channel used to receive network error
|
|
failed bool // Flag whether a send failed, discard everything onward
|
|
)
|
|
for {
|
|
// If there's no in-flight announce running, check if a new one is needed
|
|
if done == nil && len(queue) > 0 {
|
|
// Pile transaction hashes until we reach our allowed network limit
|
|
var (
|
|
count int
|
|
pending []common.Hash
|
|
pendingTypes []byte
|
|
pendingSizes []uint32
|
|
size common.StorageSize
|
|
)
|
|
for count = 0; count < len(queue) && size < maxTxPacketSize; count++ {
|
|
if tx := p.txpool.Get(queue[count]); tx != nil {
|
|
pending = append(pending, queue[count])
|
|
pendingTypes = append(pendingTypes, tx.Type())
|
|
pendingSizes = append(pendingSizes, uint32(tx.Size()))
|
|
size += common.HashLength
|
|
}
|
|
}
|
|
// Shift and trim queue
|
|
queue = queue[:copy(queue, queue[count:])]
|
|
|
|
// If there's anything available to transfer, fire up an async writer
|
|
if len(pending) > 0 {
|
|
done = make(chan struct{})
|
|
go func() {
|
|
if p.version >= ETH68 {
|
|
if err := p.sendPooledTransactionHashes68(pending, pendingTypes, pendingSizes); err != nil {
|
|
fail <- err
|
|
return
|
|
}
|
|
} else {
|
|
if err := p.sendPooledTransactionHashes66(pending); err != nil {
|
|
fail <- err
|
|
return
|
|
}
|
|
}
|
|
close(done)
|
|
p.Log().Trace("Sent transaction announcements", "count", len(pending))
|
|
}()
|
|
}
|
|
}
|
|
// Transfer goroutine may or may not have been started, listen for events
|
|
select {
|
|
case hashes := <-p.txAnnounce:
|
|
// If the connection failed, discard all transaction events
|
|
if failed {
|
|
continue
|
|
}
|
|
// New batch of transactions to be broadcast, queue them (with cap)
|
|
queue = append(queue, hashes...)
|
|
if len(queue) > maxQueuedTxAnns {
|
|
// Fancy copy and resize to ensure buffer doesn't grow indefinitely
|
|
queue = queue[:copy(queue, queue[len(queue)-maxQueuedTxAnns:])]
|
|
}
|
|
|
|
case <-done:
|
|
done = nil
|
|
|
|
case <-fail:
|
|
failed = true
|
|
|
|
case <-p.term:
|
|
return
|
|
}
|
|
}
|
|
}
|