2020-01-22 14:39:43 +00:00
|
|
|
// Copyright 2020 The go-ethereum Authors
|
|
|
|
// This file is part of the go-ethereum library.
|
|
|
|
//
|
|
|
|
// The go-ethereum library is free software: you can redistribute it and/or modify
|
|
|
|
// it under the terms of the GNU Lesser General Public License as published by
|
|
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
|
|
// (at your option) any later version.
|
|
|
|
//
|
|
|
|
// The go-ethereum library is distributed in the hope that it will be useful,
|
|
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
|
|
// GNU Lesser General Public License for more details.
|
|
|
|
//
|
|
|
|
// You should have received a copy of the GNU Lesser General Public License
|
|
|
|
// along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
|
|
|
|
package txfetcher
|
|
|
|
|
|
|
|
import (
|
|
|
|
"bytes"
|
|
|
|
"fmt"
|
|
|
|
"math/big"
|
|
|
|
"math/rand"
|
|
|
|
"time"
|
|
|
|
|
|
|
|
"github.com/ethereum/go-ethereum/common"
|
|
|
|
"github.com/ethereum/go-ethereum/common/mclock"
|
|
|
|
"github.com/ethereum/go-ethereum/core/types"
|
|
|
|
"github.com/ethereum/go-ethereum/eth/fetcher"
|
|
|
|
)
|
|
|
|
|
|
|
|
var (
|
|
|
|
peers []string
|
|
|
|
txs []*types.Transaction
|
|
|
|
)
|
|
|
|
|
|
|
|
func init() {
|
|
|
|
// Random is nice, but we need it deterministic
|
|
|
|
rand := rand.New(rand.NewSource(0x3a29))
|
|
|
|
|
|
|
|
peers = make([]string, 10)
|
|
|
|
for i := 0; i < len(peers); i++ {
|
|
|
|
peers[i] = fmt.Sprintf("Peer #%d", i)
|
|
|
|
}
|
|
|
|
txs = make([]*types.Transaction, 65536) // We need to bump enough to hit all the limits
|
|
|
|
for i := 0; i < len(txs); i++ {
|
|
|
|
txs[i] = types.NewTransaction(rand.Uint64(), common.Address{byte(rand.Intn(256))}, new(big.Int), 0, new(big.Int), nil)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
func Fuzz(input []byte) int {
|
|
|
|
// Don't generate insanely large test cases, not much value in them
|
|
|
|
if len(input) > 16*1024 {
|
2020-11-13 11:36:38 +00:00
|
|
|
return 0
|
2020-01-22 14:39:43 +00:00
|
|
|
}
|
2020-11-13 11:36:38 +00:00
|
|
|
verbose := false
|
2020-01-22 14:39:43 +00:00
|
|
|
r := bytes.NewReader(input)
|
|
|
|
|
|
|
|
// Reduce the problem space for certain fuzz runs. Small tx space is better
|
|
|
|
// for testing clashes and in general the fetcher, but we should still run
|
|
|
|
// some tests with large spaces to hit potential issues on limits.
|
|
|
|
limit, err := r.ReadByte()
|
|
|
|
if err != nil {
|
|
|
|
return 0
|
|
|
|
}
|
|
|
|
switch limit % 4 {
|
|
|
|
case 0:
|
|
|
|
txs = txs[:4]
|
|
|
|
case 1:
|
|
|
|
txs = txs[:256]
|
|
|
|
case 2:
|
|
|
|
txs = txs[:4096]
|
|
|
|
case 3:
|
|
|
|
// Full run
|
|
|
|
}
|
|
|
|
// Create a fetcher and hook into it's simulated fields
|
|
|
|
clock := new(mclock.Simulated)
|
|
|
|
rand := rand.New(rand.NewSource(0x3a29)) // Same used in package tests!!!
|
|
|
|
|
|
|
|
f := fetcher.NewTxFetcherForTests(
|
|
|
|
func(common.Hash) bool { return false },
|
core/types: support for optional blob sidecar in BlobTx (#27841)
This PR removes the newly added txpool.Transaction wrapper type, and instead adds a way
of keeping the blob sidecar within types.Transaction. It's better this way because most
code in go-ethereum does not care about blob transactions, and probably never will. This
will start mattering especially on the client side of RPC, where all APIs are based on
types.Transaction. Users need to be able to use the same signing flows they already
have.
However, since blobs are only allowed in some places but not others, we will now need to
add checks to avoid creating invalid blocks. I'm still trying to figure out the best place
to do some of these. The way I have it currently is as follows:
- In block validation (import), txs are verified not to have a blob sidecar.
- In miner, we strip off the sidecar when committing the transaction into the block.
- In TxPool validation, txs must have a sidecar to be added into the blobpool.
- Note there is a special case here: when transactions are re-added because of a chain
reorg, we cannot use the transactions gathered from the old chain blocks as-is,
because they will be missing their blobs. This was previously handled by storing the
blobs into the 'blobpool limbo'. The code has now changed to store the full
transaction in the limbo instead, but it might be confusing for code readers why we're
not simply adding the types.Transaction we already have.
Code changes summary:
- txpool.Transaction removed and all uses replaced by types.Transaction again
- blobpool now stores types.Transaction instead of defining its own blobTx format for storage
- the blobpool limbo now stores types.Transaction instead of storing only the blobs
- checks to validate the presence/absence of the blob sidecar added in certain critical places
2023-08-14 08:13:34 +00:00
|
|
|
func(txs []*types.Transaction) []error {
|
2020-01-22 14:39:43 +00:00
|
|
|
return make([]error, len(txs))
|
|
|
|
},
|
|
|
|
func(string, []common.Hash) error { return nil },
|
2023-10-10 08:35:51 +00:00
|
|
|
nil,
|
2020-01-22 14:39:43 +00:00
|
|
|
clock, rand,
|
|
|
|
)
|
|
|
|
f.Start()
|
|
|
|
defer f.Stop()
|
|
|
|
|
|
|
|
// Try to throw random junk at the fetcher
|
|
|
|
for {
|
|
|
|
// Read the next command and abort if we're done
|
|
|
|
cmd, err := r.ReadByte()
|
|
|
|
if err != nil {
|
|
|
|
return 0
|
|
|
|
}
|
|
|
|
switch cmd % 4 {
|
|
|
|
case 0:
|
|
|
|
// Notify a new set of transactions:
|
|
|
|
// Byte 1: Peer index to announce with
|
|
|
|
// Byte 2: Number of hashes to announce
|
|
|
|
// Byte 3-4, 5-6, etc: Transaction indices (2 byte) to announce
|
|
|
|
peerIdx, err := r.ReadByte()
|
|
|
|
if err != nil {
|
|
|
|
return 0
|
|
|
|
}
|
|
|
|
peer := peers[int(peerIdx)%len(peers)]
|
|
|
|
|
|
|
|
announceCnt, err := r.ReadByte()
|
|
|
|
if err != nil {
|
|
|
|
return 0
|
|
|
|
}
|
|
|
|
announce := int(announceCnt) % (2 * len(txs)) // No point in generating too many duplicates
|
|
|
|
|
|
|
|
var (
|
|
|
|
announceIdxs = make([]int, announce)
|
|
|
|
announces = make([]common.Hash, announce)
|
2023-10-10 08:35:51 +00:00
|
|
|
types = make([]byte, announce)
|
|
|
|
sizes = make([]uint32, announce)
|
2020-01-22 14:39:43 +00:00
|
|
|
)
|
|
|
|
for i := 0; i < len(announces); i++ {
|
|
|
|
annBuf := make([]byte, 2)
|
|
|
|
if n, err := r.Read(annBuf); err != nil || n != 2 {
|
|
|
|
return 0
|
|
|
|
}
|
|
|
|
announceIdxs[i] = (int(annBuf[0])*256 + int(annBuf[1])) % len(txs)
|
|
|
|
announces[i] = txs[announceIdxs[i]].Hash()
|
2023-10-10 08:35:51 +00:00
|
|
|
types[i] = txs[announceIdxs[i]].Type()
|
|
|
|
sizes[i] = uint32(txs[announceIdxs[i]].Size())
|
2020-01-22 14:39:43 +00:00
|
|
|
}
|
2020-11-13 11:36:38 +00:00
|
|
|
if verbose {
|
|
|
|
fmt.Println("Notify", peer, announceIdxs)
|
|
|
|
}
|
2023-10-10 08:35:51 +00:00
|
|
|
if err := f.Notify(peer, types, sizes, announces); err != nil {
|
2020-01-22 14:39:43 +00:00
|
|
|
panic(err)
|
|
|
|
}
|
|
|
|
|
|
|
|
case 1:
|
|
|
|
// Deliver a new set of transactions:
|
|
|
|
// Byte 1: Peer index to announce with
|
|
|
|
// Byte 2: Number of hashes to announce
|
|
|
|
// Byte 3-4, 5-6, etc: Transaction indices (2 byte) to announce
|
|
|
|
peerIdx, err := r.ReadByte()
|
|
|
|
if err != nil {
|
|
|
|
return 0
|
|
|
|
}
|
|
|
|
peer := peers[int(peerIdx)%len(peers)]
|
|
|
|
|
|
|
|
deliverCnt, err := r.ReadByte()
|
|
|
|
if err != nil {
|
|
|
|
return 0
|
|
|
|
}
|
|
|
|
deliver := int(deliverCnt) % (2 * len(txs)) // No point in generating too many duplicates
|
|
|
|
|
|
|
|
var (
|
|
|
|
deliverIdxs = make([]int, deliver)
|
|
|
|
deliveries = make([]*types.Transaction, deliver)
|
|
|
|
)
|
|
|
|
for i := 0; i < len(deliveries); i++ {
|
|
|
|
deliverBuf := make([]byte, 2)
|
|
|
|
if n, err := r.Read(deliverBuf); err != nil || n != 2 {
|
|
|
|
return 0
|
|
|
|
}
|
|
|
|
deliverIdxs[i] = (int(deliverBuf[0])*256 + int(deliverBuf[1])) % len(txs)
|
|
|
|
deliveries[i] = txs[deliverIdxs[i]]
|
|
|
|
}
|
|
|
|
directFlag, err := r.ReadByte()
|
|
|
|
if err != nil {
|
|
|
|
return 0
|
|
|
|
}
|
|
|
|
direct := (directFlag % 2) == 0
|
2020-11-13 11:36:38 +00:00
|
|
|
if verbose {
|
|
|
|
fmt.Println("Enqueue", peer, deliverIdxs, direct)
|
|
|
|
}
|
2020-01-22 14:39:43 +00:00
|
|
|
if err := f.Enqueue(peer, deliveries, direct); err != nil {
|
|
|
|
panic(err)
|
|
|
|
}
|
|
|
|
|
|
|
|
case 2:
|
|
|
|
// Drop a peer:
|
|
|
|
// Byte 1: Peer index to drop
|
|
|
|
peerIdx, err := r.ReadByte()
|
|
|
|
if err != nil {
|
|
|
|
return 0
|
|
|
|
}
|
|
|
|
peer := peers[int(peerIdx)%len(peers)]
|
2020-11-13 11:36:38 +00:00
|
|
|
if verbose {
|
|
|
|
fmt.Println("Drop", peer)
|
|
|
|
}
|
2020-01-22 14:39:43 +00:00
|
|
|
if err := f.Drop(peer); err != nil {
|
|
|
|
panic(err)
|
|
|
|
}
|
|
|
|
|
|
|
|
case 3:
|
|
|
|
// Move the simulated clock forward
|
|
|
|
// Byte 1: 100ms increment to move forward
|
|
|
|
tickCnt, err := r.ReadByte()
|
|
|
|
if err != nil {
|
|
|
|
return 0
|
|
|
|
}
|
|
|
|
tick := time.Duration(tickCnt) * 100 * time.Millisecond
|
2020-11-13 11:36:38 +00:00
|
|
|
if verbose {
|
|
|
|
fmt.Println("Sleep", tick)
|
|
|
|
}
|
2020-01-22 14:39:43 +00:00
|
|
|
clock.Run(tick)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|