Merge pull request #73 from openrelayxyz/merge/geth-v1.11.5
Merge/geth v1.11.5
This commit is contained in:
commit
fc577b80aa
2
.github/CONTRIBUTING.md
vendored
2
.github/CONTRIBUTING.md
vendored
@ -35,6 +35,6 @@ and help.
|
|||||||
|
|
||||||
## Configuration, dependencies, and tests
|
## Configuration, dependencies, and tests
|
||||||
|
|
||||||
Please see the [Developers' Guide](https://geth.ethereum.org/docs/developers/devguide)
|
Please see the [Developers' Guide](https://geth.ethereum.org/docs/developers/geth-developer/dev-guide)
|
||||||
for more details on configuring your environment, managing project dependencies
|
for more details on configuring your environment, managing project dependencies
|
||||||
and testing procedures.
|
and testing procedures.
|
||||||
|
48
.travis.yml
48
.travis.yml
@ -57,30 +57,6 @@ jobs:
|
|||||||
script:
|
script:
|
||||||
- go run build/ci.go docker -image -manifest amd64,arm64 -upload ethereum/client-go
|
- go run build/ci.go docker -image -manifest amd64,arm64 -upload ethereum/client-go
|
||||||
|
|
||||||
# This builder does the Ubuntu PPA upload
|
|
||||||
- stage: build
|
|
||||||
if: type = push
|
|
||||||
os: linux
|
|
||||||
dist: bionic
|
|
||||||
go: 1.20.x
|
|
||||||
env:
|
|
||||||
- ubuntu-ppa
|
|
||||||
- GO111MODULE=on
|
|
||||||
git:
|
|
||||||
submodules: false # avoid cloning ethereum/tests
|
|
||||||
addons:
|
|
||||||
apt:
|
|
||||||
packages:
|
|
||||||
- devscripts
|
|
||||||
- debhelper
|
|
||||||
- dput
|
|
||||||
- fakeroot
|
|
||||||
- python-bzrlib
|
|
||||||
- python-paramiko
|
|
||||||
script:
|
|
||||||
- echo '|1|7SiYPr9xl3uctzovOTj4gMwAC1M=|t6ReES75Bo/PxlOPJ6/GsGbTrM0= ssh-rsa AAAAB3NzaC1yc2EAAAABIwAAAQEA0aKz5UTUndYgIGG7dQBV+HaeuEZJ2xPHo2DS2iSKvUL4xNMSAY4UguNW+pX56nAQmZKIZZ8MaEvSj6zMEDiq6HFfn5JcTlM80UwlnyKe8B8p7Nk06PPQLrnmQt5fh0HmEcZx+JU9TZsfCHPnX7MNz4ELfZE6cFsclClrKim3BHUIGq//t93DllB+h4O9LHjEUsQ1Sr63irDLSutkLJD6RXchjROXkNirlcNVHH/jwLWR5RcYilNX7S5bIkK8NlWPjsn/8Ua5O7I9/YoE97PpO6i73DTGLh5H9JN/SITwCKBkgSDWUt61uPK3Y11Gty7o2lWsBjhBUm2Y38CBsoGmBw==' >> ~/.ssh/known_hosts
|
|
||||||
- go run build/ci.go debsrc -upload ethereum/ethereum -sftp-user geth-ci -signer "Go Ethereum Linux Builder <geth-ci@ethereum.org>"
|
|
||||||
|
|
||||||
# This builder does the Linux Azure uploads
|
# This builder does the Linux Azure uploads
|
||||||
- stage: build
|
- stage: build
|
||||||
if: type = push
|
if: type = push
|
||||||
@ -162,6 +138,30 @@ jobs:
|
|||||||
script:
|
script:
|
||||||
- go run build/ci.go test $TEST_PACKAGES
|
- go run build/ci.go test $TEST_PACKAGES
|
||||||
|
|
||||||
|
# This builder does the Ubuntu PPA nightly uploads
|
||||||
|
- stage: build
|
||||||
|
if: type = cron || (type = push && tag ~= /^v[0-9]/)
|
||||||
|
os: linux
|
||||||
|
dist: bionic
|
||||||
|
go: 1.20.x
|
||||||
|
env:
|
||||||
|
- ubuntu-ppa
|
||||||
|
- GO111MODULE=on
|
||||||
|
git:
|
||||||
|
submodules: false # avoid cloning ethereum/tests
|
||||||
|
addons:
|
||||||
|
apt:
|
||||||
|
packages:
|
||||||
|
- devscripts
|
||||||
|
- debhelper
|
||||||
|
- dput
|
||||||
|
- fakeroot
|
||||||
|
- python-bzrlib
|
||||||
|
- python-paramiko
|
||||||
|
script:
|
||||||
|
- echo '|1|7SiYPr9xl3uctzovOTj4gMwAC1M=|t6ReES75Bo/PxlOPJ6/GsGbTrM0= ssh-rsa AAAAB3NzaC1yc2EAAAABIwAAAQEA0aKz5UTUndYgIGG7dQBV+HaeuEZJ2xPHo2DS2iSKvUL4xNMSAY4UguNW+pX56nAQmZKIZZ8MaEvSj6zMEDiq6HFfn5JcTlM80UwlnyKe8B8p7Nk06PPQLrnmQt5fh0HmEcZx+JU9TZsfCHPnX7MNz4ELfZE6cFsclClrKim3BHUIGq//t93DllB+h4O9LHjEUsQ1Sr63irDLSutkLJD6RXchjROXkNirlcNVHH/jwLWR5RcYilNX7S5bIkK8NlWPjsn/8Ua5O7I9/YoE97PpO6i73DTGLh5H9JN/SITwCKBkgSDWUt61uPK3Y11Gty7o2lWsBjhBUm2Y38CBsoGmBw==' >> ~/.ssh/known_hosts
|
||||||
|
- go run build/ci.go debsrc -upload ethereum/ethereum -sftp-user geth-ci -signer "Go Ethereum Linux Builder <geth-ci@ethereum.org>"
|
||||||
|
|
||||||
# This builder does the Azure archive purges to avoid accumulating junk
|
# This builder does the Azure archive purges to avoid accumulating junk
|
||||||
- stage: build
|
- stage: build
|
||||||
if: type = cron
|
if: type = cron
|
||||||
|
12
README.md
12
README.md
@ -36,10 +36,10 @@ directory.
|
|||||||
|
|
||||||
| Command | Description |
|
| Command | Description |
|
||||||
| :--------: | ---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- |
|
| :--------: | ---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- |
|
||||||
| **`geth`** | Our main Ethereum CLI client. It is the entry point into the Ethereum network (main-, test- or private net), capable of running as a full node (default), archive node (retaining all historical state) or a light node (retrieving data live). It can be used by other processes as a gateway into the Ethereum network via JSON RPC endpoints exposed on top of HTTP, WebSocket and/or IPC transports. `geth --help` and the [CLI page](https://geth.ethereum.org/docs/interface/command-line-options) for command line options. |
|
| **`geth`** | Our main Ethereum CLI client. It is the entry point into the Ethereum network (main-, test- or private net), capable of running as a full node (default), archive node (retaining all historical state) or a light node (retrieving data live). It can be used by other processes as a gateway into the Ethereum network via JSON RPC endpoints exposed on top of HTTP, WebSocket and/or IPC transports. `geth --help` and the [CLI page](https://geth.ethereum.org/docs/fundamentals/command-line-options) for command line options. |
|
||||||
| `clef` | Stand-alone signing tool, which can be used as a backend signer for `geth`. |
|
| `clef` | Stand-alone signing tool, which can be used as a backend signer for `geth`. |
|
||||||
| `devp2p` | Utilities to interact with nodes on the networking layer, without running a full blockchain. |
|
| `devp2p` | Utilities to interact with nodes on the networking layer, without running a full blockchain. |
|
||||||
| `abigen` | Source code generator to convert Ethereum contract definitions into easy-to-use, compile-time type-safe Go packages. It operates on plain [Ethereum contract ABIs](https://docs.soliditylang.org/en/develop/abi-spec.html) with expanded functionality if the contract bytecode is also available. However, it also accepts Solidity source files, making development much more streamlined. Please see our [Native DApps](https://geth.ethereum.org/docs/dapp/native-bindings) page for details. |
|
| `abigen` | Source code generator to convert Ethereum contract definitions into easy-to-use, compile-time type-safe Go packages. It operates on plain [Ethereum contract ABIs](https://docs.soliditylang.org/en/develop/abi-spec.html) with expanded functionality if the contract bytecode is also available. However, it also accepts Solidity source files, making development much more streamlined. Please see our [Native DApps](https://geth.ethereum.org/docs/developers/dapp-developer/native-bindings) page for details. |
|
||||||
| `bootnode` | Stripped down version of our Ethereum client implementation that only takes part in the network node discovery protocol, but does not run any of the higher level application protocols. It can be used as a lightweight bootstrap node to aid in finding peers in private networks. |
|
| `bootnode` | Stripped down version of our Ethereum client implementation that only takes part in the network node discovery protocol, but does not run any of the higher level application protocols. It can be used as a lightweight bootstrap node to aid in finding peers in private networks. |
|
||||||
| `evm` | Developer utility version of the EVM (Ethereum Virtual Machine) that is capable of running bytecode snippets within a configurable environment and execution mode. Its purpose is to allow isolated, fine-grained debugging of EVM opcodes (e.g. `evm --code 60ff60ff --debug run`). |
|
| `evm` | Developer utility version of the EVM (Ethereum Virtual Machine) that is capable of running bytecode snippets within a configurable environment and execution mode. Its purpose is to allow isolated, fine-grained debugging of EVM opcodes (e.g. `evm --code 60ff60ff --debug run`). |
|
||||||
| `rlpdump` | Developer utility tool to convert binary RLP ([Recursive Length Prefix](https://ethereum.org/en/developers/docs/data-structures-and-encoding/rlp)) dumps (data encoding used by the Ethereum protocol both network as well as consensus wise) to user-friendlier hierarchical representation (e.g. `rlpdump --hex CE0183FFFFFFC4C304050583616263`). |
|
| `rlpdump` | Developer utility tool to convert binary RLP ([Recursive Length Prefix](https://ethereum.org/en/developers/docs/data-structures-and-encoding/rlp)) dumps (data encoding used by the Ethereum protocol both network as well as consensus wise) to user-friendlier hierarchical representation (e.g. `rlpdump --hex CE0183FFFFFFC4C304050583616263`). |
|
||||||
@ -47,7 +47,7 @@ directory.
|
|||||||
## Running `geth`
|
## Running `geth`
|
||||||
|
|
||||||
Going through all the possible command line flags is out of scope here (please consult our
|
Going through all the possible command line flags is out of scope here (please consult our
|
||||||
[CLI Wiki page](https://geth.ethereum.org/docs/interface/command-line-options)),
|
[CLI Wiki page](https://geth.ethereum.org/docs/fundamentals/command-line-options)),
|
||||||
but we've enumerated a few common parameter combos to get you up to speed quickly
|
but we've enumerated a few common parameter combos to get you up to speed quickly
|
||||||
on how you can run your own `geth` instance.
|
on how you can run your own `geth` instance.
|
||||||
|
|
||||||
@ -82,10 +82,10 @@ This command will:
|
|||||||
* Start `geth` in snap sync mode (default, can be changed with the `--syncmode` flag),
|
* Start `geth` in snap sync mode (default, can be changed with the `--syncmode` flag),
|
||||||
causing it to download more data in exchange for avoiding processing the entire history
|
causing it to download more data in exchange for avoiding processing the entire history
|
||||||
of the Ethereum network, which is very CPU intensive.
|
of the Ethereum network, which is very CPU intensive.
|
||||||
* Start the built-in interactive [JavaScript console](https://geth.ethereum.org/docs/interface/javascript-console),
|
* Start the built-in interactive [JavaScript console](https://geth.ethereum.org/docs/interacting-with-geth/javascript-console),
|
||||||
(via the trailing `console` subcommand) through which you can interact using [`web3` methods](https://github.com/ChainSafe/web3.js/blob/0.20.7/DOCUMENTATION.md)
|
(via the trailing `console` subcommand) through which you can interact using [`web3` methods](https://github.com/ChainSafe/web3.js/blob/0.20.7/DOCUMENTATION.md)
|
||||||
(note: the `web3` version bundled within `geth` is very old, and not up to date with official docs),
|
(note: the `web3` version bundled within `geth` is very old, and not up to date with official docs),
|
||||||
as well as `geth`'s own [management APIs](https://geth.ethereum.org/docs/rpc/server).
|
as well as `geth`'s own [management APIs](https://geth.ethereum.org/docs/interacting-with-geth/rpc).
|
||||||
This tool is optional and if you leave it out you can always attach it to an already running
|
This tool is optional and if you leave it out you can always attach it to an already running
|
||||||
`geth` instance with `geth attach`.
|
`geth` instance with `geth attach`.
|
||||||
|
|
||||||
@ -175,7 +175,7 @@ accessible from the outside.
|
|||||||
As a developer, sooner rather than later you'll want to start interacting with `geth` and the
|
As a developer, sooner rather than later you'll want to start interacting with `geth` and the
|
||||||
Ethereum network via your own programs and not manually through the console. To aid
|
Ethereum network via your own programs and not manually through the console. To aid
|
||||||
this, `geth` has built-in support for a JSON-RPC based APIs ([standard APIs](https://ethereum.github.io/execution-apis/api-documentation/)
|
this, `geth` has built-in support for a JSON-RPC based APIs ([standard APIs](https://ethereum.github.io/execution-apis/api-documentation/)
|
||||||
and [`geth` specific APIs](https://geth.ethereum.org/docs/rpc/server)).
|
and [`geth` specific APIs](https://geth.ethereum.org/docs/interacting-with-geth/rpc)).
|
||||||
These can be exposed via HTTP, WebSockets and IPC (UNIX sockets on UNIX based
|
These can be exposed via HTTP, WebSockets and IPC (UNIX sockets on UNIX based
|
||||||
platforms, and named pipes on Windows).
|
platforms, and named pipes on Windows).
|
||||||
|
|
||||||
|
@ -246,7 +246,10 @@ func UnpackRevert(data []byte) (string, error) {
|
|||||||
if !bytes.Equal(data[:4], revertSelector) {
|
if !bytes.Equal(data[:4], revertSelector) {
|
||||||
return "", errors.New("invalid data for unpacking")
|
return "", errors.New("invalid data for unpacking")
|
||||||
}
|
}
|
||||||
typ, _ := NewType("string", "", nil)
|
typ, err := NewType("string", "", nil)
|
||||||
|
if err != nil {
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
unpacked, err := (Arguments{{Type: typ}}).Unpack(data[4:])
|
unpacked, err := (Arguments{{Type: typ}}).Unpack(data[4:])
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return "", err
|
return "", err
|
||||||
|
@ -644,20 +644,33 @@ func (b *SimulatedBackend) callContract(ctx context.Context, call ethereum.CallM
|
|||||||
if call.Value == nil {
|
if call.Value == nil {
|
||||||
call.Value = new(big.Int)
|
call.Value = new(big.Int)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Set infinite balance to the fake caller account.
|
// Set infinite balance to the fake caller account.
|
||||||
from := stateDB.GetOrNewStateObject(call.From)
|
from := stateDB.GetOrNewStateObject(call.From)
|
||||||
from.SetBalance(math.MaxBig256)
|
from.SetBalance(math.MaxBig256)
|
||||||
// Execute the call.
|
|
||||||
msg := callMsg{call}
|
|
||||||
|
|
||||||
txContext := core.NewEVMTxContext(msg)
|
// Execute the call.
|
||||||
evmContext := core.NewEVMBlockContext(header, b.blockchain, nil)
|
msg := &core.Message{
|
||||||
|
From: call.From,
|
||||||
|
To: call.To,
|
||||||
|
Value: call.Value,
|
||||||
|
GasLimit: call.Gas,
|
||||||
|
GasPrice: call.GasPrice,
|
||||||
|
GasFeeCap: call.GasFeeCap,
|
||||||
|
GasTipCap: call.GasTipCap,
|
||||||
|
Data: call.Data,
|
||||||
|
AccessList: call.AccessList,
|
||||||
|
SkipAccountChecks: true,
|
||||||
|
}
|
||||||
|
|
||||||
// Create a new environment which holds all relevant information
|
// Create a new environment which holds all relevant information
|
||||||
// about the transaction and calling mechanisms.
|
// about the transaction and calling mechanisms.
|
||||||
|
txContext := core.NewEVMTxContext(msg)
|
||||||
|
evmContext := core.NewEVMBlockContext(header, b.blockchain, nil)
|
||||||
vmEnv := vm.NewEVM(evmContext, txContext, stateDB, b.config, vm.Config{NoBaseFee: true})
|
vmEnv := vm.NewEVM(evmContext, txContext, stateDB, b.config, vm.Config{NoBaseFee: true})
|
||||||
gasPool := new(core.GasPool).AddGas(math.MaxUint64)
|
gasPool := new(core.GasPool).AddGas(math.MaxUint64)
|
||||||
|
|
||||||
return core.NewStateTransition(vmEnv, msg, gasPool).TransitionDb()
|
return core.ApplyMessage(vmEnv, msg, gasPool)
|
||||||
}
|
}
|
||||||
|
|
||||||
// SendTransaction updates the pending block to include the given transaction.
|
// SendTransaction updates the pending block to include the given transaction.
|
||||||
@ -821,23 +834,6 @@ func (b *SimulatedBackend) Blockchain() *core.BlockChain {
|
|||||||
return b.blockchain
|
return b.blockchain
|
||||||
}
|
}
|
||||||
|
|
||||||
// callMsg implements core.Message to allow passing it as a transaction simulator.
|
|
||||||
type callMsg struct {
|
|
||||||
ethereum.CallMsg
|
|
||||||
}
|
|
||||||
|
|
||||||
func (m callMsg) From() common.Address { return m.CallMsg.From }
|
|
||||||
func (m callMsg) Nonce() uint64 { return 0 }
|
|
||||||
func (m callMsg) IsFake() bool { return true }
|
|
||||||
func (m callMsg) To() *common.Address { return m.CallMsg.To }
|
|
||||||
func (m callMsg) GasPrice() *big.Int { return m.CallMsg.GasPrice }
|
|
||||||
func (m callMsg) GasFeeCap() *big.Int { return m.CallMsg.GasFeeCap }
|
|
||||||
func (m callMsg) GasTipCap() *big.Int { return m.CallMsg.GasTipCap }
|
|
||||||
func (m callMsg) Gas() uint64 { return m.CallMsg.Gas }
|
|
||||||
func (m callMsg) Value() *big.Int { return m.CallMsg.Value }
|
|
||||||
func (m callMsg) Data() []byte { return m.CallMsg.Data }
|
|
||||||
func (m callMsg) AccessList() types.AccessList { return m.CallMsg.AccessList }
|
|
||||||
|
|
||||||
// filterBackend implements filters.Backend to support filtering for logs without
|
// filterBackend implements filters.Backend to support filtering for logs without
|
||||||
// taking bloom-bits acceleration structures into account.
|
// taking bloom-bits acceleration structures into account.
|
||||||
type filterBackend struct {
|
type filterBackend struct {
|
||||||
|
@ -46,7 +46,7 @@ Deploy checkpoint oracle contract. `--signers` indicates the specified trusted s
|
|||||||
checkpoint-admin deploy --rpc <NODE_RPC_ENDPOINT> --clef <CLEF_ENDPOINT> --signer <SIGNER_TO_SIGN_TX> --signers <TRUSTED_SIGNER_LIST> --threshold 1
|
checkpoint-admin deploy --rpc <NODE_RPC_ENDPOINT> --clef <CLEF_ENDPOINT> --signer <SIGNER_TO_SIGN_TX> --signers <TRUSTED_SIGNER_LIST> --threshold 1
|
||||||
```
|
```
|
||||||
|
|
||||||
It is worth noting that checkpoint-admin only supports clef as a signer for transactions and plain text(checkpoint). For more clef usage, please see the clef [tutorial](https://geth.ethereum.org/docs/clef/tutorial) .
|
It is worth noting that checkpoint-admin only supports clef as a signer for transactions and plain text(checkpoint). For more clef usage, please see the clef [tutorial](https://geth.ethereum.org/docs/tools/clef/tutorial) .
|
||||||
|
|
||||||
#### Sign
|
#### Sign
|
||||||
|
|
||||||
|
@ -29,7 +29,6 @@ import (
|
|||||||
var app = flags.NewApp("go-ethereum devp2p tool")
|
var app = flags.NewApp("go-ethereum devp2p tool")
|
||||||
|
|
||||||
func init() {
|
func init() {
|
||||||
app.HideVersion = true
|
|
||||||
app.Flags = append(app.Flags, debug.Flags...)
|
app.Flags = append(app.Flags, debug.Flags...)
|
||||||
app.Before = func(ctx *cli.Context) error {
|
app.Before = func(ctx *cli.Context) error {
|
||||||
flags.MigrateGlobalFlags(ctx)
|
flags.MigrateGlobalFlags(ctx)
|
||||||
|
@ -163,7 +163,7 @@ func (pre *Prestate) Apply(vmConfig vm.Config, chainConfig *params.ChainConfig,
|
|||||||
}
|
}
|
||||||
|
|
||||||
for i, tx := range txs {
|
for i, tx := range txs {
|
||||||
msg, err := tx.AsMessage(signer, pre.Env.BaseFee)
|
msg, err := core.TransactionToMessage(tx, signer, pre.Env.BaseFee)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
log.Warn("rejected tx", "index", i, "hash", tx.Hash(), "error", err)
|
log.Warn("rejected tx", "index", i, "hash", tx.Hash(), "error", err)
|
||||||
rejectedTxs = append(rejectedTxs, &rejectedTx{i, err.Error()})
|
rejectedTxs = append(rejectedTxs, &rejectedTx{i, err.Error()})
|
||||||
@ -188,7 +188,7 @@ func (pre *Prestate) Apply(vmConfig vm.Config, chainConfig *params.ChainConfig,
|
|||||||
msgResult, err := core.ApplyMessage(evm, msg, gaspool)
|
msgResult, err := core.ApplyMessage(evm, msg, gaspool)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
statedb.RevertToSnapshot(snapshot)
|
statedb.RevertToSnapshot(snapshot)
|
||||||
log.Info("rejected tx", "index", i, "hash", tx.Hash(), "from", msg.From(), "error", err)
|
log.Info("rejected tx", "index", i, "hash", tx.Hash(), "from", msg.From, "error", err)
|
||||||
rejectedTxs = append(rejectedTxs, &rejectedTx{i, err.Error()})
|
rejectedTxs = append(rejectedTxs, &rejectedTx{i, err.Error()})
|
||||||
gaspool.SetGas(prevGas)
|
gaspool.SetGas(prevGas)
|
||||||
continue
|
continue
|
||||||
@ -220,7 +220,7 @@ func (pre *Prestate) Apply(vmConfig vm.Config, chainConfig *params.ChainConfig,
|
|||||||
receipt.GasUsed = msgResult.UsedGas
|
receipt.GasUsed = msgResult.UsedGas
|
||||||
|
|
||||||
// If the transaction created a contract, store the creation address in the receipt.
|
// If the transaction created a contract, store the creation address in the receipt.
|
||||||
if msg.To() == nil {
|
if msg.To == nil {
|
||||||
receipt.ContractAddress = crypto.CreateAddress(evm.TxContext.Origin, tx.Nonce())
|
receipt.ContractAddress = crypto.CreateAddress(evm.TxContext.Origin, tx.Nonce())
|
||||||
}
|
}
|
||||||
|
|
||||||
|
134
cmd/evm/testdata/10/readme.md
vendored
134
cmd/evm/testdata/10/readme.md
vendored
@ -9,71 +9,77 @@ INFO [05-09|22:11:59.436] rejected tx index=3 hash=
|
|||||||
Output:
|
Output:
|
||||||
```json
|
```json
|
||||||
{
|
{
|
||||||
"alloc": {
|
"alloc": {
|
||||||
"0x1111111111111111111111111111111111111111": {
|
"0x1111111111111111111111111111111111111111": {
|
||||||
"code": "0xfe",
|
"code": "0xfe",
|
||||||
"balance": "0x10000000000",
|
"balance": "0x10000000000",
|
||||||
"nonce": "0x1"
|
"nonce": "0x1"
|
||||||
|
},
|
||||||
|
"0xa94f5374fce5edbc8e2a8697c15331677e6ebf0b": {
|
||||||
|
"balance": "0x10000000000",
|
||||||
|
"nonce": "0x1"
|
||||||
|
},
|
||||||
|
"0xd02d72e067e77158444ef2020ff2d325f929b363": {
|
||||||
|
"balance": "0xff5beffffc95",
|
||||||
|
"nonce": "0x4"
|
||||||
|
}
|
||||||
},
|
},
|
||||||
"0xa94f5374fce5edbc8e2a8697c15331677e6ebf0b": {
|
"result": {
|
||||||
"balance": "0x10000000000",
|
"stateRoot": "0xf91a7ec08e4bfea88719aab34deabb000c86902360532b52afa9599d41f2bb8b",
|
||||||
"nonce": "0x1"
|
"txRoot": "0xda925f2306a52fa24c15d5cd212d736ee016415fd8dd0c45fd368de7917d64bb",
|
||||||
},
|
"receiptsRoot": "0x439a25f7fc424c10fb1f89800e4aa1df74156b137239d9ac3eaa7c911c353cd5",
|
||||||
"0xd02d72e067e77158444ef2020ff2d325f929b363": {
|
"logsHash": "0x1dcc4de8dec75d7aab85b567b6ccd41ad312451b948a7413f0a142fd40d49347",
|
||||||
"balance": "0xff5beffffc95",
|
"logsBloom": "0x00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000",
|
||||||
"nonce": "0x4"
|
"receipts": [
|
||||||
|
{
|
||||||
|
"type": "0x2",
|
||||||
|
"root": "0x",
|
||||||
|
"status": "0x0",
|
||||||
|
"cumulativeGasUsed": "0x10000001",
|
||||||
|
"logsBloom": "0x00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000",
|
||||||
|
"logs": null,
|
||||||
|
"transactionHash": "0x88980f6efcc5358d9c359663e7b9414722d430497637340ea056b076bc206701",
|
||||||
|
"contractAddress": "0x0000000000000000000000000000000000000000",
|
||||||
|
"gasUsed": "0x10000001",
|
||||||
|
"blockHash": "0x0000000000000000000000000000000000000000000000000000000000000000",
|
||||||
|
"transactionIndex": "0x0"
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"type": "0x2",
|
||||||
|
"root": "0x",
|
||||||
|
"status": "0x0",
|
||||||
|
"cumulativeGasUsed": "0x20000001",
|
||||||
|
"logsBloom": "0x00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000",
|
||||||
|
"logs": null,
|
||||||
|
"transactionHash": "0xd7bf3886f4e2aef74d525ae072c680f3846f550254401b67cbfda4a233757582",
|
||||||
|
"contractAddress": "0x0000000000000000000000000000000000000000",
|
||||||
|
"gasUsed": "0x10000000",
|
||||||
|
"blockHash": "0x0000000000000000000000000000000000000000000000000000000000000000",
|
||||||
|
"transactionIndex": "0x1"
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"type": "0x2",
|
||||||
|
"root": "0x",
|
||||||
|
"status": "0x0",
|
||||||
|
"cumulativeGasUsed": "0x30000001",
|
||||||
|
"logsBloom": "0x00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000",
|
||||||
|
"logs": null,
|
||||||
|
"transactionHash": "0x50308296760f01f1eeec7500e9e73cad67469249b1f59e9a9f55e6625a4923db",
|
||||||
|
"contractAddress": "0x0000000000000000000000000000000000000000",
|
||||||
|
"gasUsed": "0x10000000",
|
||||||
|
"blockHash": "0x0000000000000000000000000000000000000000000000000000000000000000",
|
||||||
|
"transactionIndex": "0x2"
|
||||||
|
}
|
||||||
|
],
|
||||||
|
"rejected": [
|
||||||
|
{
|
||||||
|
"index": 3,
|
||||||
|
"error": "gas limit reached"
|
||||||
|
}
|
||||||
|
],
|
||||||
|
"currentDifficulty": "0x20000",
|
||||||
|
"gasUsed": "0x30000001",
|
||||||
|
"currentBaseFee": "0x36b"
|
||||||
}
|
}
|
||||||
},
|
|
||||||
"result": {
|
|
||||||
"stateRoot": "0xf91a7ec08e4bfea88719aab34deabb000c86902360532b52afa9599d41f2bb8b",
|
|
||||||
"txRoot": "0xda925f2306a52fa24c15d5cd212d736ee016415fd8dd0c45fd368de7917d64bb",
|
|
||||||
"receiptRoot": "0x439a25f7fc424c10fb1f89800e4aa1df74156b137239d9ac3eaa7c911c353cd5",
|
|
||||||
"logsHash": "0x1dcc4de8dec75d7aab85b567b6ccd41ad312451b948a7413f0a142fd40d49347",
|
|
||||||
"logsBloom": "0x00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000",
|
|
||||||
"receipts": [
|
|
||||||
{
|
|
||||||
"type": "0x2",
|
|
||||||
"root": "0x",
|
|
||||||
"status": "0x0",
|
|
||||||
"cumulativeGasUsed": "0x10000001",
|
|
||||||
"logsBloom": "0x00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000",
|
|
||||||
"logs": null,
|
|
||||||
"transactionHash": "0x88980f6efcc5358d9c359663e7b9414722d430497637340ea056b076bc206701",
|
|
||||||
"contractAddress": "0x0000000000000000000000000000000000000000",
|
|
||||||
"gasUsed": "0x10000001",
|
|
||||||
"blockHash": "0x0000000000000000000000000000000000000000000000000000000000000000",
|
|
||||||
"transactionIndex": "0x0"
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"type": "0x2",
|
|
||||||
"root": "0x",
|
|
||||||
"status": "0x0",
|
|
||||||
"cumulativeGasUsed": "0x20000001",
|
|
||||||
"logsBloom": "0x00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000",
|
|
||||||
"logs": null,
|
|
||||||
"transactionHash": "0xd7bf3886f4e2aef74d525ae072c680f3846f550254401b67cbfda4a233757582",
|
|
||||||
"contractAddress": "0x0000000000000000000000000000000000000000",
|
|
||||||
"gasUsed": "0x10000000",
|
|
||||||
"blockHash": "0x0000000000000000000000000000000000000000000000000000000000000000",
|
|
||||||
"transactionIndex": "0x1"
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"type": "0x2",
|
|
||||||
"root": "0x",
|
|
||||||
"status": "0x0",
|
|
||||||
"cumulativeGasUsed": "0x30000001",
|
|
||||||
"logsBloom": "0x00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000",
|
|
||||||
"logs": null,
|
|
||||||
"transactionHash": "0x50308296760f01f1eeec7500e9e73cad67469249b1f59e9a9f55e6625a4923db",
|
|
||||||
"contractAddress": "0x0000000000000000000000000000000000000000",
|
|
||||||
"gasUsed": "0x10000000",
|
|
||||||
"blockHash": "0x0000000000000000000000000000000000000000000000000000000000000000",
|
|
||||||
"transactionIndex": "0x2"
|
|
||||||
}
|
|
||||||
],
|
|
||||||
"rejected": [
|
|
||||||
3
|
|
||||||
]
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
```
|
```
|
||||||
|
45
cmd/evm/testdata/12/readme.md
vendored
45
cmd/evm/testdata/12/readme.md
vendored
@ -11,29 +11,32 @@ dir=./testdata/12 && ./evm t8n --state.fork=London --input.alloc=$dir/alloc.json
|
|||||||
With the fix applied, the result is:
|
With the fix applied, the result is:
|
||||||
```
|
```
|
||||||
dir=./testdata/12 && ./evm t8n --state.fork=London --input.alloc=$dir/alloc.json --input.txs=$dir/txs.json --input.env=$dir/env.json --output.alloc=stdout --output.result=stdout
|
dir=./testdata/12 && ./evm t8n --state.fork=London --input.alloc=$dir/alloc.json --input.txs=$dir/txs.json --input.env=$dir/env.json --output.alloc=stdout --output.result=stdout
|
||||||
INFO [07-21|19:03:50.276] rejected tx index=0 hash=ccc996..d83435 from=0xa94f5374Fce5edBC8E2a8697C15331677e6EbF0B error="insufficient funds for gas * price + value: address 0xa94f5374Fce5edBC8E2a8697C15331677e6EbF0B have 84000000 want 84000032"
|
INFO [03-09|10:43:12.649] rejected tx index=0 hash=ccc996..d83435 from=0xa94f5374Fce5edBC8E2a8697C15331677e6EbF0B error="insufficient funds for gas * price + value: address 0xa94f5374Fce5edBC8E2a8697C15331677e6EbF0B have 84000000 want 84000032"
|
||||||
INFO [07-21|19:03:50.276] Trie dumping started root=e05f81..6597a5
|
INFO [03-09|10:43:12.650] Trie dumping started root=e05f81..6597a5
|
||||||
INFO [07-21|19:03:50.276] Trie dumping complete accounts=1 elapsed="39.549µs"
|
INFO [03-09|10:43:12.650] Trie dumping complete accounts=1 elapsed="46.393µs"
|
||||||
{
|
{
|
||||||
"alloc": {
|
"alloc": {
|
||||||
"0xa94f5374fce5edbc8e2a8697c15331677e6ebf0b": {
|
"0xa94f5374fce5edbc8e2a8697c15331677e6ebf0b": {
|
||||||
"balance": "0x501bd00"
|
"balance": "0x501bd00"
|
||||||
|
}
|
||||||
|
},
|
||||||
|
"result": {
|
||||||
|
"stateRoot": "0xe05f81f8244a76503ceec6f88abfcd03047a612a1001217f37d30984536597a5",
|
||||||
|
"txRoot": "0x56e81f171bcc55a6ff8345e692c0f86e5b48e01b996cadc001622fb5e363b421",
|
||||||
|
"receiptsRoot": "0x56e81f171bcc55a6ff8345e692c0f86e5b48e01b996cadc001622fb5e363b421",
|
||||||
|
"logsHash": "0x1dcc4de8dec75d7aab85b567b6ccd41ad312451b948a7413f0a142fd40d49347",
|
||||||
|
"logsBloom": "0x00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000",
|
||||||
|
"receipts": [],
|
||||||
|
"rejected": [
|
||||||
|
{
|
||||||
|
"index": 0,
|
||||||
|
"error": "insufficient funds for gas * price + value: address 0xa94f5374Fce5edBC8E2a8697C15331677e6EbF0B have 84000000 want 84000032"
|
||||||
|
}
|
||||||
|
],
|
||||||
|
"currentDifficulty": "0x20000",
|
||||||
|
"gasUsed": "0x0",
|
||||||
|
"currentBaseFee": "0x20"
|
||||||
}
|
}
|
||||||
},
|
|
||||||
"result": {
|
|
||||||
"stateRoot": "0xe05f81f8244a76503ceec6f88abfcd03047a612a1001217f37d30984536597a5",
|
|
||||||
"txRoot": "0x56e81f171bcc55a6ff8345e692c0f86e5b48e01b996cadc001622fb5e363b421",
|
|
||||||
"receiptRoot": "0x56e81f171bcc55a6ff8345e692c0f86e5b48e01b996cadc001622fb5e363b421",
|
|
||||||
"logsHash": "0x1dcc4de8dec75d7aab85b567b6ccd41ad312451b948a7413f0a142fd40d49347",
|
|
||||||
"logsBloom": "0x00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000",
|
|
||||||
"receipts": [],
|
|
||||||
"rejected": [
|
|
||||||
{
|
|
||||||
"index": 0,
|
|
||||||
"error": "insufficient funds for gas * price + value: address 0xa94f5374Fce5edBC8E2a8697C15331677e6EbF0B have 84000000 want 84000032"
|
|
||||||
}
|
|
||||||
]
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
```
|
```
|
||||||
|
|
||||||
|
24
cmd/evm/testdata/14/readme.md
vendored
24
cmd/evm/testdata/14/readme.md
vendored
@ -5,36 +5,40 @@ This test shows how the `evm t8n` can be used to calculate the (ethash) difficul
|
|||||||
Calculating it (with an empty set of txs) using `London` rules (and no provided unclehash for the parent block):
|
Calculating it (with an empty set of txs) using `London` rules (and no provided unclehash for the parent block):
|
||||||
```
|
```
|
||||||
[user@work evm]$ ./evm t8n --input.alloc=./testdata/14/alloc.json --input.txs=./testdata/14/txs.json --input.env=./testdata/14/env.json --output.result=stdout --state.fork=London
|
[user@work evm]$ ./evm t8n --input.alloc=./testdata/14/alloc.json --input.txs=./testdata/14/txs.json --input.env=./testdata/14/env.json --output.result=stdout --state.fork=London
|
||||||
INFO [08-30|20:43:09.352] Trie dumping started root=6f0588..7f4bdc
|
INFO [03-09|10:43:57.070] Trie dumping started root=6f0588..7f4bdc
|
||||||
INFO [08-30|20:43:09.352] Trie dumping complete accounts=2 elapsed="82.533µs"
|
INFO [03-09|10:43:57.070] Trie dumping complete accounts=2 elapsed="214.663µs"
|
||||||
INFO [08-30|20:43:09.352] Wrote file file=alloc.json
|
INFO [03-09|10:43:57.071] Wrote file file=alloc.json
|
||||||
{
|
{
|
||||||
"result": {
|
"result": {
|
||||||
"stateRoot": "0x6f058887ca01549716789c380ede95aecc510e6d1fdc4dbf67d053c7c07f4bdc",
|
"stateRoot": "0x6f058887ca01549716789c380ede95aecc510e6d1fdc4dbf67d053c7c07f4bdc",
|
||||||
"txRoot": "0x56e81f171bcc55a6ff8345e692c0f86e5b48e01b996cadc001622fb5e363b421",
|
"txRoot": "0x56e81f171bcc55a6ff8345e692c0f86e5b48e01b996cadc001622fb5e363b421",
|
||||||
"receiptRoot": "0x56e81f171bcc55a6ff8345e692c0f86e5b48e01b996cadc001622fb5e363b421",
|
"receiptsRoot": "0x56e81f171bcc55a6ff8345e692c0f86e5b48e01b996cadc001622fb5e363b421",
|
||||||
"logsHash": "0x1dcc4de8dec75d7aab85b567b6ccd41ad312451b948a7413f0a142fd40d49347",
|
"logsHash": "0x1dcc4de8dec75d7aab85b567b6ccd41ad312451b948a7413f0a142fd40d49347",
|
||||||
"logsBloom": "0x00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000",
|
"logsBloom": "0x00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000",
|
||||||
"receipts": [],
|
"receipts": [],
|
||||||
"currentDifficulty": "0x2000020000000"
|
"currentDifficulty": "0x2000020000000",
|
||||||
|
"gasUsed": "0x0",
|
||||||
|
"currentBaseFee": "0x500"
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
```
|
```
|
||||||
Same thing, but this time providing a non-empty (and non-`emptyKeccak`) unclehash, which leads to a slightly different result:
|
Same thing, but this time providing a non-empty (and non-`emptyKeccak`) unclehash, which leads to a slightly different result:
|
||||||
```
|
```
|
||||||
[user@work evm]$ ./evm t8n --input.alloc=./testdata/14/alloc.json --input.txs=./testdata/14/txs.json --input.env=./testdata/14/env.uncles.json --output.result=stdout --state.fork=London
|
[user@work evm]$ ./evm t8n --input.alloc=./testdata/14/alloc.json --input.txs=./testdata/14/txs.json --input.env=./testdata/14/env.uncles.json --output.result=stdout --state.fork=London
|
||||||
INFO [08-30|20:44:33.102] Trie dumping started root=6f0588..7f4bdc
|
INFO [03-09|10:44:20.511] Trie dumping started root=6f0588..7f4bdc
|
||||||
INFO [08-30|20:44:33.102] Trie dumping complete accounts=2 elapsed="72.91µs"
|
INFO [03-09|10:44:20.511] Trie dumping complete accounts=2 elapsed="184.319µs"
|
||||||
INFO [08-30|20:44:33.102] Wrote file file=alloc.json
|
INFO [03-09|10:44:20.512] Wrote file file=alloc.json
|
||||||
{
|
{
|
||||||
"result": {
|
"result": {
|
||||||
"stateRoot": "0x6f058887ca01549716789c380ede95aecc510e6d1fdc4dbf67d053c7c07f4bdc",
|
"stateRoot": "0x6f058887ca01549716789c380ede95aecc510e6d1fdc4dbf67d053c7c07f4bdc",
|
||||||
"txRoot": "0x56e81f171bcc55a6ff8345e692c0f86e5b48e01b996cadc001622fb5e363b421",
|
"txRoot": "0x56e81f171bcc55a6ff8345e692c0f86e5b48e01b996cadc001622fb5e363b421",
|
||||||
"receiptRoot": "0x56e81f171bcc55a6ff8345e692c0f86e5b48e01b996cadc001622fb5e363b421",
|
"receiptsRoot": "0x56e81f171bcc55a6ff8345e692c0f86e5b48e01b996cadc001622fb5e363b421",
|
||||||
"logsHash": "0x1dcc4de8dec75d7aab85b567b6ccd41ad312451b948a7413f0a142fd40d49347",
|
"logsHash": "0x1dcc4de8dec75d7aab85b567b6ccd41ad312451b948a7413f0a142fd40d49347",
|
||||||
"logsBloom": "0x00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000",
|
"logsBloom": "0x00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000",
|
||||||
"receipts": [],
|
"receipts": [],
|
||||||
"currentDifficulty": "0x1ff8020000000"
|
"currentDifficulty": "0x1ff8020000000",
|
||||||
|
"gasUsed": "0x0",
|
||||||
|
"currentBaseFee": "0x500"
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
```
|
```
|
||||||
|
16
cmd/evm/testdata/19/readme.md
vendored
16
cmd/evm/testdata/19/readme.md
vendored
@ -6,4 +6,20 @@ this time on `GrayGlacier` (Eip 5133).
|
|||||||
Calculating it (with an empty set of txs) using `GrayGlacier` rules (and no provided unclehash for the parent block):
|
Calculating it (with an empty set of txs) using `GrayGlacier` rules (and no provided unclehash for the parent block):
|
||||||
```
|
```
|
||||||
[user@work evm]$ ./evm t8n --input.alloc=./testdata/19/alloc.json --input.txs=./testdata/19/txs.json --input.env=./testdata/19/env.json --output.result=stdout --state.fork=GrayGlacier
|
[user@work evm]$ ./evm t8n --input.alloc=./testdata/19/alloc.json --input.txs=./testdata/19/txs.json --input.env=./testdata/19/env.json --output.result=stdout --state.fork=GrayGlacier
|
||||||
|
INFO [03-09|10:45:26.777] Trie dumping started root=6f0588..7f4bdc
|
||||||
|
INFO [03-09|10:45:26.777] Trie dumping complete accounts=2 elapsed="176.471µs"
|
||||||
|
INFO [03-09|10:45:26.777] Wrote file file=alloc.json
|
||||||
|
{
|
||||||
|
"result": {
|
||||||
|
"stateRoot": "0x6f058887ca01549716789c380ede95aecc510e6d1fdc4dbf67d053c7c07f4bdc",
|
||||||
|
"txRoot": "0x56e81f171bcc55a6ff8345e692c0f86e5b48e01b996cadc001622fb5e363b421",
|
||||||
|
"receiptsRoot": "0x56e81f171bcc55a6ff8345e692c0f86e5b48e01b996cadc001622fb5e363b421",
|
||||||
|
"logsHash": "0x1dcc4de8dec75d7aab85b567b6ccd41ad312451b948a7413f0a142fd40d49347",
|
||||||
|
"logsBloom": "0x00000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000000",
|
||||||
|
"receipts": [],
|
||||||
|
"currentDifficulty": "0x2000000004000",
|
||||||
|
"gasUsed": "0x0",
|
||||||
|
"currentBaseFee": "0x500"
|
||||||
|
}
|
||||||
|
}
|
||||||
```
|
```
|
370
cmd/evm/testdata/7/readme.md
vendored
370
cmd/evm/testdata/7/readme.md
vendored
@ -3,5 +3,373 @@ DAO-transition works
|
|||||||
|
|
||||||
Example:
|
Example:
|
||||||
```
|
```
|
||||||
./statet8n --input.alloc=./testdata/7/alloc.json --input.txs=./testdata/7/txs.json --input.env=./testdata/7/env.json --output.alloc=stdout --state.fork=HomesteadToDaoAt5
|
./evm t8n --input.alloc=./testdata/7/alloc.json --input.txs=./testdata/7/txs.json --input.env=./testdata/7/env.json --output.alloc=stdout --state.fork=HomesteadToDaoAt5
|
||||||
|
INFO [03-09|10:47:37.255] Trie dumping started root=157847..2891b7
|
||||||
|
INFO [03-09|10:47:37.256] Trie dumping complete accounts=120 elapsed="715.635µs"
|
||||||
|
INFO [03-09|10:47:37.256] Wrote file file=result.json
|
||||||
|
{
|
||||||
|
"alloc": {
|
||||||
|
"0x005f5cee7a43331d5a3d3eec71305925a62f34b6": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x0101f3be8ebb4bbd39a2e3b9a3639d4259832fd9": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x057b56736d32b86616a10f619859c6cd6f59092a": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x06706dd3f2c9abf0a21ddcc6941d9b86f0596936": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x0737a6b837f97f46ebade41b9bc3e1c509c85c53": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x07f5c1e1bc2c93e0402f23341973a0e043f7bf8a": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x0e0da70933f4c7849fc0d203f5d1d43b9ae4532d": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x0ff30d6de14a8224aa97b78aea5388d1c51c1f00": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x12e626b0eebfe86a56d633b9864e389b45dcb260": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x1591fc0f688c81fbeb17f5426a162a7024d430c2": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x17802f43a0137c506ba92291391a8a8f207f487d": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x1975bd06d486162d5dc297798dfc41edd5d160a7": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x1ca6abd14d30affe533b24d7a21bff4c2d5e1f3b": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x1cba23d343a983e9b5cfd19496b9a9701ada385f": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x200450f06520bdd6c527622a273333384d870efb": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x21c7fdb9ed8d291d79ffd82eb2c4356ec0d81241": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x23b75c2f6791eef49c69684db4c6c1f93bf49a50": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x24c4d950dfd4dd1902bbed3508144a54542bba94": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x253488078a4edf4d6f42f113d1e62836a942cf1a": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x27b137a85656544b1ccb5a0f2e561a5703c6a68f": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x2a5ed960395e2a49b1c758cef4aa15213cfd874c": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x2b3455ec7fedf16e646268bf88846bd7a2319bb2": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x2c19c7f9ae8b751e37aeb2d93a699722395ae18f": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x304a554a310c7e546dfe434669c62820b7d83490": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x319f70bab6845585f412ec7724b744fec6095c85": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x35a051a0010aba705c9008d7a7eff6fb88f6ea7b": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x3ba4d81db016dc2890c81f3acec2454bff5aada5": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x3c02a7bc0391e86d91b7d144e61c2c01a25a79c5": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x40b803a9abce16f50f36a77ba41180eb90023925": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x440c59b325d2997a134c2c7c60a8c61611212bad": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x4486a3d68fac6967006d7a517b889fd3f98c102b": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x4613f3bca5c44ea06337a9e439fbc6d42e501d0a": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x47e7aa56d6bdf3f36be34619660de61275420af8": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x4863226780fe7c0356454236d3b1c8792785748d": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x492ea3bb0f3315521c31f273e565b868fc090f17": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x4cb31628079fb14e4bc3cd5e30c2f7489b00960c": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x4deb0033bb26bc534b197e61d19e0733e5679784": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x4fa802324e929786dbda3b8820dc7834e9134a2a": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x4fd6ace747f06ece9c49699c7cabc62d02211f75": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x51e0ddd9998364a2eb38588679f0d2c42653e4a6": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x52c5317c848ba20c7504cb2c8052abd1fde29d03": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x542a9515200d14b68e934e9830d91645a980dd7a": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x5524c55fb03cf21f549444ccbecb664d0acad706": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x579a80d909f346fbfb1189493f521d7f48d52238": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x58b95c9a9d5d26825e70a82b6adb139d3fd829eb": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x5c6e67ccd5849c0d29219c4f95f1a7a93b3f5dc5": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x5c8536898fbb74fc7445814902fd08422eac56d0": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x5d2b2e6fcbe3b11d26b525e085ff818dae332479": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x5dc28b15dffed94048d73806ce4b7a4612a1d48f": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x5f9f3392e9f62f63b8eac0beb55541fc8627f42c": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x6131c42fa982e56929107413a9d526fd99405560": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x6231b6d0d5e77fe001c2a460bd9584fee60d409b": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x627a0a960c079c21c34f7612d5d230e01b4ad4c7": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x63ed5a272de2f6d968408b4acb9024f4cc208ebf": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x6966ab0d485353095148a2155858910e0965b6f9": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x6b0c4d41ba9ab8d8cfb5d379c69a612f2ced8ecb": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x6d87578288b6cb5549d5076a207456a1f6a63dc0": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x6f6704e5a10332af6672e50b3d9754dc460dfa4d": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x7602b46df5390e432ef1c307d4f2c9ff6d65cc97": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x779543a0491a837ca36ce8c635d6154e3c4911a6": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x77ca7b50b6cd7e2f3fa008e24ab793fd56cb15f6": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x782495b7b3355efb2833d56ecb34dc22ad7dfcc4": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x807640a13483f8ac783c557fcdf27be11ea4ac7a": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x8163e7fb499e90f8544ea62bbf80d21cd26d9efd": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x84ef4b2357079cd7a7c69fd7a37cd0609a679106": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x86af3e9626fce1957c82e88cbf04ddf3a2ed7915": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x8a8eafb1cf62bfbeb1741769dae1a9dd47996192": {
|
||||||
|
"balance": "0xfeedbead"
|
||||||
|
},
|
||||||
|
"0x8d9edb3054ce5c5774a420ac37ebae0ac02343c6": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x914d1b8b43e92723e64fd0a06f5bdb8dd9b10c79": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x97f43a37f595ab5dd318fb46e7a155eae057317a": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x9aa008f65de0b923a2a4f02012ad034a5e2e2192": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x9c15b54878ba618f494b38f0ae7443db6af648ba": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x9c50426be05db97f5d64fc54bf89eff947f0a321": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x9da397b9e80755301a3b32173283a91c0ef6c87e": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x9ea779f907f0b315b364b0cfc39a0fde5b02a416": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x9f27daea7aca0aa0446220b98d028715e3bc803d": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0x9fcd2deaff372a39cc679d5c5e4de7bafb0b1339": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0xa2f1ccba9395d7fcb155bba8bc92db9bafaeade7": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0xa3acf3a1e16b1d7c315e23510fdd7847b48234f6": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0xa5dc5acd6a7968a4554d89d65e59b7fd3bff0f90": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0xa82f360a8d3455c5c41366975bde739c37bfeb8a": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0xa94f5374fce5edbc8e2a8697c15331677e6ebf0b": {
|
||||||
|
"balance": "0x5ffd4878be161d74",
|
||||||
|
"nonce": "0xac"
|
||||||
|
},
|
||||||
|
"0xac1ecab32727358dba8962a0f3b261731aad9723": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0xaccc230e8a6e5be9160b8cdf2864dd2a001c28b6": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0xacd87e28b0c9d1254e868b81cba4cc20d9a32225": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0xadf80daec7ba8dcf15392f1ac611fff65d94f880": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0xaeeb8ff27288bdabc0fa5ebb731b6f409507516c": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0xb136707642a4ea12fb4bae820f03d2562ebff487": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0xb2c6f0dfbb716ac562e2d85d6cb2f8d5ee87603e": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0xb3fb0e5aba0e20e5c49d252dfd30e102b171a425": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0xb52042c8ca3f8aa246fa79c3feaa3d959347c0ab": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0xb9637156d330c0d605a791f1c31ba5890582fe1c": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0xbb9bc244d798123fde783fcc1c72d3bb8c189413": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0xbc07118b9ac290e4622f5e77a0853539789effbe": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0xbcf899e6c7d9d5a215ab1e3444c86806fa854c76": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0xbe8539bfe837b67d1282b2b1d61c3f723966f049": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0xbf4ed7b27f1d666546e30d74d50d173d20bca754": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0xc4bbd073882dd2add2424cf47d35213405b01324": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0xc94f5374fce5edbc8e2a8697c15331677e6ebf0b": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0xca544e5c4687d109611d0f8f928b53a25af72448": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0xcbb9d3703e651b0d496cdefb8b92c25aeb2171f7": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0xcc34673c6c40e791051898567a1222daf90be287": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0xceaeb481747ca6c540a000c1f3641f8cef161fa7": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0xd131637d5275fd1a68a3200f4ad25c71a2a9522e": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0xd164b088bd9108b60d0ca3751da4bceb207b0782": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0xd1ac8b1ef1b69ff51d1d401a476e7e612414f091": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0xd343b217de44030afaa275f54d31a9317c7f441e": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0xd4fe7bc31cedb7bfb8a345f31e668033056b2728": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0xd9aef3a1e38a39c16b31d1ace71bca8ef58d315b": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0xda2fef9e4a3230988ff17df2165440f37e8b1708": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0xdbe9b615a3ae8709af8b93336ce9b477e4ac0940": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0xe308bd1ac5fda103967359b2712dd89deffb7973": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0xe4ae1efdfc53b73893af49113d8694a057b9c0d1": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0xec8e57756626fdc07c63ad2eafbd28d08e7b0ca5": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0xecd135fa4f61a655311e86238c92adcd779555d2": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0xf0b1aa0eb660754448a7937c022e30aa692fe0c5": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0xf1385fb24aad0cd7432824085e42aff90886fef5": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0xf14c14075d6c4ed84b86798af0956deef67365b5": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0xf4c64518ea10f995918a454158c6b61407ea345c": {
|
||||||
|
"balance": "0x0"
|
||||||
|
},
|
||||||
|
"0xfe24cdd8648121a43a7c86d289be4dd2951ed49f": {
|
||||||
|
"balance": "0x0"
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
```
|
```
|
38
cmd/evm/testdata/8/readme.md
vendored
38
cmd/evm/testdata/8/readme.md
vendored
@ -23,41 +23,37 @@ There are three transactions, each invokes the contract above.
|
|||||||
|
|
||||||
Running it yields:
|
Running it yields:
|
||||||
```
|
```
|
||||||
dir=./testdata/8 && ./evm t8n --state.fork=Berlin --input.alloc=$dir/alloc.json --input.txs=$dir/txs.json --input.env=$dir/env.json --trace && cat trace-* | grep SLOAD
|
dir=./testdata/8 && ./evm t8n --state.fork=Berlin --input.alloc=$dir/alloc.json --input.txs=$dir/txs.json --input.env=$dir/env.json --trace 2>/dev/null && cat trace-* | grep SLOAD
|
||||||
{"pc":1,"op":84,"gas":"0x484be","gasCost":"0x834","memory":"0x","memSize":0,"stack":["0x0"],"returnStack":[],"returnData":"0x","depth":1,"refund":0,"opName":"SLOAD","error":""}
|
{"pc":1,"op":84,"gas":"0x484be","gasCost":"0x834","memSize":0,"stack":["0x0"],"depth":1,"refund":0,"opName":"SLOAD"}
|
||||||
{"pc":4,"op":84,"gas":"0x47c86","gasCost":"0x834","memory":"0x","memSize":0,"stack":["0x3"],"returnStack":[],"returnData":"0x","depth":1,"refund":0,"opName":"SLOAD","error":""}
|
{"pc":4,"op":84,"gas":"0x47c86","gasCost":"0x834","memSize":0,"stack":["0x3"],"depth":1,"refund":0,"opName":"SLOAD"}
|
||||||
{"pc":1,"op":84,"gas":"0x49cf6","gasCost":"0x834","memory":"0x","memSize":0,"stack":["0x0"],"returnStack":[],"returnData":"0x","depth":1,"refund":0,"opName":"SLOAD","error":""}
|
{"pc":1,"op":84,"gas":"0x49cf6","gasCost":"0x834","memSize":0,"stack":["0x0"],"depth":1,"refund":0,"opName":"SLOAD"}
|
||||||
{"pc":4,"op":84,"gas":"0x494be","gasCost":"0x834","memory":"0x","memSize":0,"stack":["0x3"],"returnStack":[],"returnData":"0x","depth":1,"refund":0,"opName":"SLOAD","error":""}
|
{"pc":4,"op":84,"gas":"0x494be","gasCost":"0x834","memSize":0,"stack":["0x3"],"depth":1,"refund":0,"opName":"SLOAD"}
|
||||||
{"pc":1,"op":84,"gas":"0x484be","gasCost":"0x64","memory":"0x","memSize":0,"stack":["0x0"],"returnStack":[],"returnData":"0x","depth":1,"refund":0,"opName":"SLOAD","error":""}
|
{"pc":1,"op":84,"gas":"0x484be","gasCost":"0x64","memSize":0,"stack":["0x0"],"depth":1,"refund":0,"opName":"SLOAD"}
|
||||||
{"pc":4,"op":84,"gas":"0x48456","gasCost":"0x64","memory":"0x","memSize":0,"stack":["0x3"],"returnStack":[],"returnData":"0x","depth":1,"refund":0,"opName":"SLOAD","error":""}
|
{"pc":4,"op":84,"gas":"0x48456","gasCost":"0x64","memSize":0,"stack":["0x3"],"depth":1,"refund":0,"opName":"SLOAD"}
|
||||||
|
|
||||||
```
|
```
|
||||||
|
|
||||||
Simlarly, we can provide the input transactions via `stdin` instead of as file:
|
Simlarly, we can provide the input transactions via `stdin` instead of as file:
|
||||||
|
|
||||||
```
|
```
|
||||||
dir=./testdata/8 \
|
$ dir=./testdata/8 \
|
||||||
&& cat $dir/txs.json | jq "{txs: .}" \
|
&& cat $dir/txs.json | jq "{txs: .}" \
|
||||||
| ./evm t8n --state.fork=Berlin \
|
| ./evm t8n --state.fork=Berlin \
|
||||||
--input.alloc=$dir/alloc.json \
|
--input.alloc=$dir/alloc.json \
|
||||||
--input.txs=stdin \
|
--input.txs=stdin \
|
||||||
--input.env=$dir/env.json \
|
--input.env=$dir/env.json \
|
||||||
--trace \
|
--trace \
|
||||||
|
2>/dev/null \
|
||||||
&& cat trace-* | grep SLOAD
|
&& cat trace-* | grep SLOAD
|
||||||
|
{"pc":1,"op":84,"gas":"0x484be","gasCost":"0x834","memSize":0,"stack":["0x0"],"depth":1,"refund":0,"opName":"SLOAD"}
|
||||||
{"pc":1,"op":84,"gas":"0x484be","gasCost":"0x834","memory":"0x","memSize":0,"stack":["0x0"],"returnStack":[],"returnData":"0x","depth":1,"refund":0,"opName":"SLOAD","error":""}
|
{"pc":4,"op":84,"gas":"0x47c86","gasCost":"0x834","memSize":0,"stack":["0x3"],"depth":1,"refund":0,"opName":"SLOAD"}
|
||||||
{"pc":4,"op":84,"gas":"0x47c86","gasCost":"0x834","memory":"0x","memSize":0,"stack":["0x3"],"returnStack":[],"returnData":"0x","depth":1,"refund":0,"opName":"SLOAD","error":""}
|
{"pc":1,"op":84,"gas":"0x49cf6","gasCost":"0x834","memSize":0,"stack":["0x0"],"depth":1,"refund":0,"opName":"SLOAD"}
|
||||||
{"pc":1,"op":84,"gas":"0x49cf6","gasCost":"0x834","memory":"0x","memSize":0,"stack":["0x0"],"returnStack":[],"returnData":"0x","depth":1,"refund":0,"opName":"SLOAD","error":""}
|
{"pc":4,"op":84,"gas":"0x494be","gasCost":"0x834","memSize":0,"stack":["0x3"],"depth":1,"refund":0,"opName":"SLOAD"}
|
||||||
{"pc":4,"op":84,"gas":"0x494be","gasCost":"0x834","memory":"0x","memSize":0,"stack":["0x3"],"returnStack":[],"returnData":"0x","depth":1,"refund":0,"opName":"SLOAD","error":""}
|
{"pc":1,"op":84,"gas":"0x484be","gasCost":"0x64","memSize":0,"stack":["0x0"],"depth":1,"refund":0,"opName":"SLOAD"}
|
||||||
{"pc":1,"op":84,"gas":"0x484be","gasCost":"0x64","memory":"0x","memSize":0,"stack":["0x0"],"returnStack":[],"returnData":"0x","depth":1,"refund":0,"opName":"SLOAD","error":""}
|
{"pc":4,"op":84,"gas":"0x48456","gasCost":"0x64","memSize":0,"stack":["0x3"],"depth":1,"refund":0,"opName":"SLOAD"}
|
||||||
{"pc":4,"op":84,"gas":"0x48456","gasCost":"0x64","memory":"0x","memSize":0,"stack":["0x3"],"returnStack":[],"returnData":"0x","depth":1,"refund":0,"opName":"SLOAD","error":""}
|
|
||||||
```
|
```
|
||||||
|
|
||||||
If we try to execute it on older rules:
|
If we try to execute it on older rules:
|
||||||
```
|
```
|
||||||
dir=./testdata/8 && ./evm t8n --state.fork=Istanbul --input.alloc=$dir/alloc.json --input.txs=$dir/txs.json --input.env=$dir/env.json
|
$ dir=./testdata/8 && ./evm t8n --state.fork=Istanbul --input.alloc=$dir/alloc.json --input.txs=$dir/txs.json --input.env=$dir/env.json
|
||||||
INFO [01-21|23:21:51.265] rejected tx index=0 hash=d2818d..6ab3da error="tx type not supported"
|
ERROR(10): failed signing transactions: ERROR(10): tx 0: failed to sign tx: transaction type not supported
|
||||||
INFO [01-21|23:21:51.265] rejected tx index=1 hash=26ea00..81c01b from=0xa94f5374Fce5edBC8E2a8697C15331677e6EbF0B error="nonce too high: address 0xa94f5374Fce5edBC8E2a8697C15331677e6EbF0B, tx: 1 state: 0"
|
|
||||||
INFO [01-21|23:21:51.265] rejected tx index=2 hash=698d01..369cee error="tx type not supported"
|
|
||||||
```
|
```
|
||||||
Number `1` and `3` are not applicable, and therefore number `2` has wrong nonce.
|
|
58
cmd/evm/testdata/9/readme.md
vendored
58
cmd/evm/testdata/9/readme.md
vendored
@ -22,35 +22,37 @@ There are two transactions, each invokes the contract above.
|
|||||||
|
|
||||||
Running it yields:
|
Running it yields:
|
||||||
```
|
```
|
||||||
$ dir=./testdata/9 && ./evm t8n --state.fork=London --input.alloc=$dir/alloc.json --input.txs=$dir/txs.json --input.env=$dir/env.json --trace && cat trace-* | grep SLOAD
|
$ dir=./testdata/9 && ./evm t8n --state.fork=London --input.alloc=$dir/alloc.json --input.txs=$dir/txs.json --input.env=$dir/env.json --trace 2>/dev/null && cat trace-* | grep SLOAD
|
||||||
{"pc":2,"op":84,"gas":"0x48c28","gasCost":"0x834","memory":"0x","memSize":0,"stack":["0x0","0x1"],"returnStack":null,"returnD
|
{"pc":1,"op":84,"gas":"0x484be","gasCost":"0x834","memSize":0,"stack":["0x0"],"depth":1,"refund":0,"opName":"SLOAD"}
|
||||||
ata":"0x","depth":1,"refund":0,"opName":"SLOAD","error":""}
|
{"pc":4,"op":84,"gas":"0x47c86","gasCost":"0x834","memSize":0,"stack":["0x3"],"depth":1,"refund":0,"opName":"SLOAD"}
|
||||||
{"pc":3,"op":84,"gas":"0x483f4","gasCost":"0x64","memory":"0x","memSize":0,"stack":["0x0","0x0"],"returnStack":null,"returnDa
|
{"pc":2,"op":84,"gas":"0x48c28","gasCost":"0x834","memSize":0,"stack":["0x0","0x1"],"depth":1,"refund":0,"opName":"SLOAD"}
|
||||||
ta":"0x","depth":1,"refund":0,"opName":"SLOAD","error":""}
|
{"pc":3,"op":84,"gas":"0x483f4","gasCost":"0x64","memSize":0,"stack":["0x0","0x0"],"depth":1,"refund":0,"opName":"SLOAD"}
|
||||||
{"pc":2,"op":84,"gas":"0x49cf4","gasCost":"0x834","memory":"0x","memSize":0,"stack":["0x0","0x1"],"returnStack":null,"returnD
|
{"pc":1,"op":84,"gas":"0x49cf6","gasCost":"0x834","memSize":0,"stack":["0x0"],"depth":1,"refund":0,"opName":"SLOAD"}
|
||||||
ata":"0x","depth":1,"refund":0,"opName":"SLOAD","error":""}
|
{"pc":4,"op":84,"gas":"0x494be","gasCost":"0x834","memSize":0,"stack":["0x3"],"depth":1,"refund":0,"opName":"SLOAD"}
|
||||||
{"pc":3,"op":84,"gas":"0x494c0","gasCost":"0x834","memory":"0x","memSize":0,"stack":["0x0","0x0"],"returnStack":null,"returnD
|
{"pc":2,"op":84,"gas":"0x49cf4","gasCost":"0x834","memSize":0,"stack":["0x0","0x1"],"depth":1,"refund":0,"opName":"SLOAD"}
|
||||||
ata":"0x","depth":1,"refund":0,"opName":"SLOAD","error":""}
|
{"pc":3,"op":84,"gas":"0x494c0","gasCost":"0x834","memSize":0,"stack":["0x0","0x0"],"depth":1,"refund":0,"opName":"SLOAD"}
|
||||||
|
{"pc":1,"op":84,"gas":"0x484be","gasCost":"0x64","memSize":0,"stack":["0x0"],"depth":1,"refund":0,"opName":"SLOAD"}
|
||||||
|
{"pc":4,"op":84,"gas":"0x48456","gasCost":"0x64","memSize":0,"stack":["0x3"],"depth":1,"refund":0,"opName":"SLOAD"}
|
||||||
```
|
```
|
||||||
|
|
||||||
We can also get the post-alloc:
|
We can also get the post-alloc:
|
||||||
```
|
```
|
||||||
$ dir=./testdata/9 && ./evm t8n --state.fork=London --input.alloc=$dir/alloc.json --input.txs=$dir/txs.json --input.env=$dir/env.json --output.alloc=stdout
|
$ dir=./testdata/9 && ./evm t8n --state.fork=London --input.alloc=$dir/alloc.json --input.txs=$dir/txs.json --input.env=$dir/env.json --output.alloc=stdout 2>/dev/null
|
||||||
{
|
{
|
||||||
"alloc": {
|
"alloc": {
|
||||||
"0x000000000000000000000000000000000000aaaa": {
|
"0x000000000000000000000000000000000000aaaa": {
|
||||||
"code": "0x58585454",
|
"code": "0x58585454",
|
||||||
"balance": "0x3",
|
"balance": "0x3",
|
||||||
"nonce": "0x1"
|
"nonce": "0x1"
|
||||||
},
|
},
|
||||||
"0x2adc25665018aa1fe0e6bc666dac8fc2697ff9ba": {
|
"0x2adc25665018aa1fe0e6bc666dac8fc2697ff9ba": {
|
||||||
"balance": "0x5bb10ddef6e0"
|
"balance": "0x5bb10ddef6e0"
|
||||||
},
|
},
|
||||||
"0xa94f5374fce5edbc8e2a8697c15331677e6ebf0b": {
|
"0xa94f5374fce5edbc8e2a8697c15331677e6ebf0b": {
|
||||||
"balance": "0xff745ee8832120",
|
"balance": "0xff745ee8832120",
|
||||||
"nonce": "0x2"
|
"nonce": "0x2"
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
|
||||||
}
|
}
|
||||||
```
|
```
|
||||||
|
|
||||||
@ -65,10 +67,12 @@ by feeding it presigned transactions, located in `txs_signed.json`.
|
|||||||
|
|
||||||
```
|
```
|
||||||
dir=./testdata/9 && ./evm t8n --state.fork=Berlin --input.alloc=$dir/alloc.json --input.txs=$dir/txs_signed.json --input.env=$dir/env.json
|
dir=./testdata/9 && ./evm t8n --state.fork=Berlin --input.alloc=$dir/alloc.json --input.txs=$dir/txs_signed.json --input.env=$dir/env.json
|
||||||
INFO [05-07|12:28:42.072] rejected tx index=0 hash=b4821e..536819 error="transaction type not supported"
|
WARN [03-09|11:06:22.065] rejected tx index=0 hash=334e09..f8dce5 error="transaction type not supported"
|
||||||
INFO [05-07|12:28:42.072] rejected tx index=1 hash=a9c6c6..fa4036 from=0xa94f5374Fce5edBC8E2a8697C15331677e6EbF0B error="nonce too high: address 0xa94f5374Fce5edBC8E2a8697C15331677e6EbF0B, tx: 1 state: 0"
|
INFO [03-09|11:06:22.066] rejected tx index=1 hash=a9c6c6..fa4036 from=0xa94f5374Fce5edBC8E2a8697C15331677e6EbF0B error="nonce too high: address 0xa94f5374Fce5edBC8E2a8697C15331677e6EbF0B, tx: 1 state: 0"
|
||||||
INFO [05-07|12:28:42.073] Wrote file file=alloc.json
|
INFO [03-09|11:06:22.066] Trie dumping started root=6eebe9..a0fda5
|
||||||
INFO [05-07|12:28:42.073] Wrote file file=result.json
|
INFO [03-09|11:06:22.066] Trie dumping complete accounts=2 elapsed="55.844µs"
|
||||||
|
INFO [03-09|11:06:22.066] Wrote file file=alloc.json
|
||||||
|
INFO [03-09|11:06:22.066] Wrote file file=result.json
|
||||||
```
|
```
|
||||||
|
|
||||||
Number `0` is not applicable, and therefore number `1` has wrong nonce, and both are rejected.
|
Number `0` is not applicable, and therefore number `1` has wrong nonce, and both are rejected.
|
||||||
|
37
cmd/evm/testdata/9/txs_signed.json
vendored
Normal file
37
cmd/evm/testdata/9/txs_signed.json
vendored
Normal file
@ -0,0 +1,37 @@
|
|||||||
|
[
|
||||||
|
{
|
||||||
|
"gas": "0x4ef00",
|
||||||
|
"maxFeePerGas": "0x2",
|
||||||
|
"maxPriorityFeePerGas": "0x12A05F200",
|
||||||
|
"chainId": "0x1",
|
||||||
|
"input": "0x",
|
||||||
|
"nonce": "0x0",
|
||||||
|
"to": "0x000000000000000000000000000000000000aaaa",
|
||||||
|
"value": "0x0",
|
||||||
|
"type" : "0x2",
|
||||||
|
"accessList": [
|
||||||
|
{"address": "0x000000000000000000000000000000000000aaaa",
|
||||||
|
"storageKeys": [
|
||||||
|
"0x0000000000000000000000000000000000000000000000000000000000000000"
|
||||||
|
]
|
||||||
|
}
|
||||||
|
],
|
||||||
|
"v": "0x1",
|
||||||
|
"r": "0xd77c8ff989789b5d9d99254cbae2e2996dc7e6215cba4d55254c14e6d6b9f314",
|
||||||
|
"s": "0x5cc021481e7e6bb444bbb87ab32071e8fd0a8d1e125c7bb352d2879bd7ff5c0a",
|
||||||
|
"secretKey": "0x45a915e4d060149eb4365960e6a7a45f334393093061116b197e3240065ff2d8"
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"gas": "0x4ef00",
|
||||||
|
"gasPrice": "0x12A05F200",
|
||||||
|
"chainId": "0x1",
|
||||||
|
"input": "0x",
|
||||||
|
"nonce": "0x1",
|
||||||
|
"to": "0x000000000000000000000000000000000000aaaa",
|
||||||
|
"value": "0x0",
|
||||||
|
"v": "0x25",
|
||||||
|
"r": "0xbee5ec9f6650020266bf3455a852eece2b073a2fa918c4d1836a1af69c2aa50c",
|
||||||
|
"s": "0x556c897a58dbc007a6b09814e1fba7502adb76effd2146da4365816926f387ce",
|
||||||
|
"secretKey": "0x45a915e4d060149eb4365960e6a7a45f334393093061116b197e3240065ff2d8"
|
||||||
|
}
|
||||||
|
]
|
@ -37,7 +37,7 @@ var (
|
|||||||
Description: `
|
Description: `
|
||||||
The Geth console is an interactive shell for the JavaScript runtime environment
|
The Geth console is an interactive shell for the JavaScript runtime environment
|
||||||
which exposes a node admin interface as well as the Ðapp JavaScript API.
|
which exposes a node admin interface as well as the Ðapp JavaScript API.
|
||||||
See https://geth.ethereum.org/docs/interface/javascript-console.`,
|
See https://geth.ethereum.org/docs/interacting-with-geth/javascript-console.`,
|
||||||
}
|
}
|
||||||
|
|
||||||
attachCommand = &cli.Command{
|
attachCommand = &cli.Command{
|
||||||
@ -49,7 +49,7 @@ See https://geth.ethereum.org/docs/interface/javascript-console.`,
|
|||||||
Description: `
|
Description: `
|
||||||
The Geth console is an interactive shell for the JavaScript runtime environment
|
The Geth console is an interactive shell for the JavaScript runtime environment
|
||||||
which exposes a node admin interface as well as the Ðapp JavaScript API.
|
which exposes a node admin interface as well as the Ðapp JavaScript API.
|
||||||
See https://geth.ethereum.org/docs/interface/javascript-console.
|
See https://geth.ethereum.org/docs/interacting-with-geth/javascript-console.
|
||||||
This command allows to open a console on a running geth node.`,
|
This command allows to open a console on a running geth node.`,
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -61,7 +61,7 @@ This command allows to open a console on a running geth node.`,
|
|||||||
Flags: flags.Merge(nodeFlags, consoleFlags),
|
Flags: flags.Merge(nodeFlags, consoleFlags),
|
||||||
Description: `
|
Description: `
|
||||||
The JavaScript VM exposes a node admin interface as well as the Ðapp
|
The JavaScript VM exposes a node admin interface as well as the Ðapp
|
||||||
JavaScript API. See https://geth.ethereum.org/docs/interface/javascript-console`,
|
JavaScript API. See https://geth.ethereum.org/docs/interacting-with-geth/javascript-console`,
|
||||||
}
|
}
|
||||||
)
|
)
|
||||||
|
|
||||||
|
@ -694,41 +694,19 @@ func showMetaData(ctx *cli.Context) error {
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
fmt.Fprintf(os.Stderr, "Error accessing ancients: %v", err)
|
fmt.Fprintf(os.Stderr, "Error accessing ancients: %v", err)
|
||||||
}
|
}
|
||||||
pp := func(val *uint64) string {
|
data := rawdb.ReadChainMetadata(db)
|
||||||
if val == nil {
|
data = append(data, []string{"frozen", fmt.Sprintf("%d items", ancients)})
|
||||||
return "<nil>"
|
data = append(data, []string{"snapshotGenerator", snapshot.ParseGeneratorStatus(rawdb.ReadSnapshotGenerator(db))})
|
||||||
}
|
|
||||||
return fmt.Sprintf("%d (%#x)", *val, *val)
|
|
||||||
}
|
|
||||||
data := [][]string{
|
|
||||||
{"databaseVersion", pp(rawdb.ReadDatabaseVersion(db))},
|
|
||||||
{"headBlockHash", fmt.Sprintf("%v", rawdb.ReadHeadBlockHash(db))},
|
|
||||||
{"headFastBlockHash", fmt.Sprintf("%v", rawdb.ReadHeadFastBlockHash(db))},
|
|
||||||
{"headHeaderHash", fmt.Sprintf("%v", rawdb.ReadHeadHeaderHash(db))}}
|
|
||||||
if b := rawdb.ReadHeadBlock(db); b != nil {
|
if b := rawdb.ReadHeadBlock(db); b != nil {
|
||||||
data = append(data, []string{"headBlock.Hash", fmt.Sprintf("%v", b.Hash())})
|
data = append(data, []string{"headBlock.Hash", fmt.Sprintf("%v", b.Hash())})
|
||||||
data = append(data, []string{"headBlock.Root", fmt.Sprintf("%v", b.Root())})
|
data = append(data, []string{"headBlock.Root", fmt.Sprintf("%v", b.Root())})
|
||||||
data = append(data, []string{"headBlock.Number", fmt.Sprintf("%d (%#x)", b.Number(), b.Number())})
|
data = append(data, []string{"headBlock.Number", fmt.Sprintf("%d (%#x)", b.Number(), b.Number())})
|
||||||
}
|
}
|
||||||
if b := rawdb.ReadSkeletonSyncStatus(db); b != nil {
|
|
||||||
data = append(data, []string{"SkeletonSyncStatus", string(b)})
|
|
||||||
}
|
|
||||||
if h := rawdb.ReadHeadHeader(db); h != nil {
|
if h := rawdb.ReadHeadHeader(db); h != nil {
|
||||||
data = append(data, []string{"headHeader.Hash", fmt.Sprintf("%v", h.Hash())})
|
data = append(data, []string{"headHeader.Hash", fmt.Sprintf("%v", h.Hash())})
|
||||||
data = append(data, []string{"headHeader.Root", fmt.Sprintf("%v", h.Root)})
|
data = append(data, []string{"headHeader.Root", fmt.Sprintf("%v", h.Root)})
|
||||||
data = append(data, []string{"headHeader.Number", fmt.Sprintf("%d (%#x)", h.Number, h.Number)})
|
data = append(data, []string{"headHeader.Number", fmt.Sprintf("%d (%#x)", h.Number, h.Number)})
|
||||||
}
|
}
|
||||||
data = append(data, [][]string{{"frozen", fmt.Sprintf("%d items", ancients)},
|
|
||||||
{"lastPivotNumber", pp(rawdb.ReadLastPivotNumber(db))},
|
|
||||||
{"len(snapshotSyncStatus)", fmt.Sprintf("%d bytes", len(rawdb.ReadSnapshotSyncStatus(db)))},
|
|
||||||
{"snapshotGenerator", snapshot.ParseGeneratorStatus(rawdb.ReadSnapshotGenerator(db))},
|
|
||||||
{"snapshotDisabled", fmt.Sprintf("%v", rawdb.ReadSnapshotDisabled(db))},
|
|
||||||
{"snapshotJournal", fmt.Sprintf("%d bytes", len(rawdb.ReadSnapshotJournal(db)))},
|
|
||||||
{"snapshotRecoveryNumber", pp(rawdb.ReadSnapshotRecoveryNumber(db))},
|
|
||||||
{"snapshotRoot", fmt.Sprintf("%v", rawdb.ReadSnapshotRoot(db))},
|
|
||||||
{"txIndexTail", pp(rawdb.ReadTxIndexTail(db))},
|
|
||||||
{"fastTxLookupLimit", pp(rawdb.ReadFastTxLookupLimit(db))},
|
|
||||||
}...)
|
|
||||||
table := tablewriter.NewWriter(os.Stdout)
|
table := tablewriter.NewWriter(os.Stdout)
|
||||||
table.SetHeader([]string{"Field", "Value"})
|
table.SetHeader([]string{"Field", "Value"})
|
||||||
table.AppendBulk(data)
|
table.AppendBulk(data)
|
||||||
|
@ -208,7 +208,6 @@ var app = flags.NewApp("the go-ethereum command line interface")
|
|||||||
func init() {
|
func init() {
|
||||||
// Initialize the CLI app and start Geth
|
// Initialize the CLI app and start Geth
|
||||||
app.Action = geth
|
app.Action = geth
|
||||||
app.HideVersion = true // we have a command to print the version
|
|
||||||
app.Copyright = "Copyright 2013-2023 The go-ethereum Authors"
|
app.Copyright = "Copyright 2013-2023 The go-ethereum Authors"
|
||||||
app.Commands = []*cli.Command{
|
app.Commands = []*cli.Command{
|
||||||
// See chaincmd.go:
|
// See chaincmd.go:
|
||||||
|
File diff suppressed because one or more lines are too long
@ -326,10 +326,8 @@ func (beacon *Beacon) Prepare(chain consensus.ChainHeaderReader, header *types.H
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// Finalize implements consensus.Engine, setting the final state on the header
|
// Finalize implements consensus.Engine and processes withdrawals on top.
|
||||||
func (beacon *Beacon) Finalize(chain consensus.ChainHeaderReader, header *types.Header, state *state.StateDB, txs []*types.Transaction, uncles []*types.Header, withdrawals []*types.Withdrawal) {
|
func (beacon *Beacon) Finalize(chain consensus.ChainHeaderReader, header *types.Header, state *state.StateDB, txs []*types.Transaction, uncles []*types.Header, withdrawals []*types.Withdrawal) {
|
||||||
// Finalize is different with Prepare, it can be used in both block generation
|
|
||||||
// and verification. So determine the consensus rules by header type.
|
|
||||||
if !beacon.IsPoSHeader(header) {
|
if !beacon.IsPoSHeader(header) {
|
||||||
beacon.ethone.Finalize(chain, header, state, txs, uncles, nil)
|
beacon.ethone.Finalize(chain, header, state, txs, uncles, nil)
|
||||||
return
|
return
|
||||||
@ -341,16 +339,12 @@ func (beacon *Beacon) Finalize(chain consensus.ChainHeaderReader, header *types.
|
|||||||
amount = amount.Mul(amount, big.NewInt(params.GWei))
|
amount = amount.Mul(amount, big.NewInt(params.GWei))
|
||||||
state.AddBalance(w.Address, amount)
|
state.AddBalance(w.Address, amount)
|
||||||
}
|
}
|
||||||
// The block reward is no longer handled here. It's done by the
|
// No block reward which is issued by consensus layer instead.
|
||||||
// external consensus engine.
|
|
||||||
header.Root = state.IntermediateRoot(true)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// FinalizeAndAssemble implements consensus.Engine, setting the final state and
|
// FinalizeAndAssemble implements consensus.Engine, setting the final state and
|
||||||
// assembling the block.
|
// assembling the block.
|
||||||
func (beacon *Beacon) FinalizeAndAssemble(chain consensus.ChainHeaderReader, header *types.Header, state *state.StateDB, txs []*types.Transaction, uncles []*types.Header, receipts []*types.Receipt, withdrawals []*types.Withdrawal) (*types.Block, error) {
|
func (beacon *Beacon) FinalizeAndAssemble(chain consensus.ChainHeaderReader, header *types.Header, state *state.StateDB, txs []*types.Transaction, uncles []*types.Header, receipts []*types.Receipt, withdrawals []*types.Withdrawal) (*types.Block, error) {
|
||||||
// FinalizeAndAssemble is different with Prepare, it can be used in both block
|
|
||||||
// generation and verification. So determine the consensus rules by header type.
|
|
||||||
if !beacon.IsPoSHeader(header) {
|
if !beacon.IsPoSHeader(header) {
|
||||||
return beacon.ethone.FinalizeAndAssemble(chain, header, state, txs, uncles, receipts, nil)
|
return beacon.ethone.FinalizeAndAssemble(chain, header, state, txs, uncles, receipts, nil)
|
||||||
}
|
}
|
||||||
@ -367,6 +361,11 @@ func (beacon *Beacon) FinalizeAndAssemble(chain consensus.ChainHeaderReader, hea
|
|||||||
}
|
}
|
||||||
// Finalize and assemble the block.
|
// Finalize and assemble the block.
|
||||||
beacon.Finalize(chain, header, state, txs, uncles, withdrawals)
|
beacon.Finalize(chain, header, state, txs, uncles, withdrawals)
|
||||||
|
|
||||||
|
// Assign the final state root to header.
|
||||||
|
header.Root = state.IntermediateRoot(true)
|
||||||
|
|
||||||
|
// Assemble and return the final block.
|
||||||
return types.NewBlockWithWithdrawals(header, txs, uncles, receipts, withdrawals, trie.NewStackTrie(nil)), nil
|
return types.NewBlockWithWithdrawals(header, txs, uncles, receipts, withdrawals, trie.NewStackTrie(nil)), nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -565,12 +565,10 @@ func (c *Clique) Prepare(chain consensus.ChainHeaderReader, header *types.Header
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// Finalize implements consensus.Engine, ensuring no uncles are set, nor block
|
// Finalize implements consensus.Engine. There is no post-transaction
|
||||||
// rewards given.
|
// consensus rules in clique, do nothing here.
|
||||||
func (c *Clique) Finalize(chain consensus.ChainHeaderReader, header *types.Header, state *state.StateDB, txs []*types.Transaction, uncles []*types.Header, withdrawals []*types.Withdrawal) {
|
func (c *Clique) Finalize(chain consensus.ChainHeaderReader, header *types.Header, state *state.StateDB, txs []*types.Transaction, uncles []*types.Header, withdrawals []*types.Withdrawal) {
|
||||||
// No block rewards in PoA, so the state remains as is and uncles are dropped
|
// No block rewards in PoA, so the state remains as is
|
||||||
header.Root = state.IntermediateRoot(chain.Config().IsEIP158(header.Number))
|
|
||||||
header.UncleHash = types.CalcUncleHash(nil)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// FinalizeAndAssemble implements consensus.Engine, ensuring no uncles are set,
|
// FinalizeAndAssemble implements consensus.Engine, ensuring no uncles are set,
|
||||||
@ -579,11 +577,13 @@ func (c *Clique) FinalizeAndAssemble(chain consensus.ChainHeaderReader, header *
|
|||||||
if len(withdrawals) > 0 {
|
if len(withdrawals) > 0 {
|
||||||
return nil, errors.New("clique does not support withdrawals")
|
return nil, errors.New("clique does not support withdrawals")
|
||||||
}
|
}
|
||||||
|
|
||||||
// Finalize block
|
// Finalize block
|
||||||
c.Finalize(chain, header, state, txs, uncles, nil)
|
c.Finalize(chain, header, state, txs, uncles, nil)
|
||||||
|
|
||||||
// Assemble and return the final block for sealing
|
// Assign the final state root to header.
|
||||||
|
header.Root = state.IntermediateRoot(chain.Config().IsEIP158(header.Number))
|
||||||
|
|
||||||
|
// Assemble and return the final block for sealing.
|
||||||
return types.NewBlock(header, txs, nil, receipts, trie.NewStackTrie(nil)), nil
|
return types.NewBlock(header, txs, nil, receipts, trie.NewStackTrie(nil)), nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -84,16 +84,16 @@ type Engine interface {
|
|||||||
// rules of a particular engine. The changes are executed inline.
|
// rules of a particular engine. The changes are executed inline.
|
||||||
Prepare(chain ChainHeaderReader, header *types.Header) error
|
Prepare(chain ChainHeaderReader, header *types.Header) error
|
||||||
|
|
||||||
// Finalize runs any post-transaction state modifications (e.g. block rewards)
|
// Finalize runs any post-transaction state modifications (e.g. block rewards
|
||||||
// but does not assemble the block.
|
// or process withdrawals) but does not assemble the block.
|
||||||
//
|
//
|
||||||
// Note: The block header and state database might be updated to reflect any
|
// Note: The state database might be updated to reflect any consensus rules
|
||||||
// consensus rules that happen at finalization (e.g. block rewards).
|
// that happen at finalization (e.g. block rewards).
|
||||||
Finalize(chain ChainHeaderReader, header *types.Header, state *state.StateDB, txs []*types.Transaction,
|
Finalize(chain ChainHeaderReader, header *types.Header, state *state.StateDB, txs []*types.Transaction,
|
||||||
uncles []*types.Header, withdrawals []*types.Withdrawal)
|
uncles []*types.Header, withdrawals []*types.Withdrawal)
|
||||||
|
|
||||||
// FinalizeAndAssemble runs any post-transaction state modifications (e.g. block
|
// FinalizeAndAssemble runs any post-transaction state modifications (e.g. block
|
||||||
// rewards) and assembles the final block.
|
// rewards or process withdrawals) and assembles the final block.
|
||||||
//
|
//
|
||||||
// Note: The block header and state database might be updated to reflect any
|
// Note: The block header and state database might be updated to reflect any
|
||||||
// consensus rules that happen at finalization (e.g. block rewards).
|
// consensus rules that happen at finalization (e.g. block rewards).
|
||||||
|
@ -598,12 +598,10 @@ func (ethash *Ethash) Prepare(chain consensus.ChainHeaderReader, header *types.H
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// Finalize implements consensus.Engine, accumulating the block and uncle rewards,
|
// Finalize implements consensus.Engine, accumulating the block and uncle rewards.
|
||||||
// setting the final state on the header
|
|
||||||
func (ethash *Ethash) Finalize(chain consensus.ChainHeaderReader, header *types.Header, state *state.StateDB, txs []*types.Transaction, uncles []*types.Header, withdrawals []*types.Withdrawal) {
|
func (ethash *Ethash) Finalize(chain consensus.ChainHeaderReader, header *types.Header, state *state.StateDB, txs []*types.Transaction, uncles []*types.Header, withdrawals []*types.Withdrawal) {
|
||||||
// Accumulate any block and uncle rewards and commit the final state root
|
// Accumulate any block and uncle rewards
|
||||||
accumulateRewards(chain.Config(), state, header, uncles)
|
accumulateRewards(chain.Config(), state, header, uncles)
|
||||||
header.Root = state.IntermediateRoot(chain.Config().IsEIP158(header.Number))
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// FinalizeAndAssemble implements consensus.Engine, accumulating the block and
|
// FinalizeAndAssemble implements consensus.Engine, accumulating the block and
|
||||||
@ -612,9 +610,12 @@ func (ethash *Ethash) FinalizeAndAssemble(chain consensus.ChainHeaderReader, hea
|
|||||||
if len(withdrawals) > 0 {
|
if len(withdrawals) > 0 {
|
||||||
return nil, errors.New("ethash does not support withdrawals")
|
return nil, errors.New("ethash does not support withdrawals")
|
||||||
}
|
}
|
||||||
|
|
||||||
// Finalize block
|
// Finalize block
|
||||||
ethash.Finalize(chain, header, state, txs, uncles, nil)
|
ethash.Finalize(chain, header, state, txs, uncles, nil)
|
||||||
|
|
||||||
|
// Assign the final state root to header.
|
||||||
|
header.Root = state.IntermediateRoot(chain.Config().IsEIP158(header.Number))
|
||||||
|
|
||||||
// Header seems complete, assemble into a block and return
|
// Header seems complete, assemble into a block and return
|
||||||
return types.NewBlock(header, txs, uncles, receipts, trie.NewStackTrie(nil)), nil
|
return types.NewBlock(header, txs, uncles, receipts, trie.NewStackTrie(nil)), nil
|
||||||
}
|
}
|
||||||
|
@ -90,10 +90,8 @@ func (v *BlockValidator) ValidateBody(block *types.Block) error {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// ValidateState validates the various changes that happen after a state
|
// ValidateState validates the various changes that happen after a state transition,
|
||||||
// transition, such as amount of used gas, the receipt roots and the state root
|
// such as amount of used gas, the receipt roots and the state root itself.
|
||||||
// itself. ValidateState returns a database batch if the validation was a success
|
|
||||||
// otherwise nil and an error is returned.
|
|
||||||
func (v *BlockValidator) ValidateState(block *types.Block, statedb *state.StateDB, receipts types.Receipts, usedGas uint64) error {
|
func (v *BlockValidator) ValidateState(block *types.Block, statedb *state.StateDB, receipts types.Receipts, usedGas uint64) error {
|
||||||
header := block.Header()
|
header := block.Header()
|
||||||
if block.GasUsed() != usedGas {
|
if block.GasUsed() != usedGas {
|
||||||
@ -113,7 +111,7 @@ func (v *BlockValidator) ValidateState(block *types.Block, statedb *state.StateD
|
|||||||
// Validate the state root against the received state root and throw
|
// Validate the state root against the received state root and throw
|
||||||
// an error if they don't match.
|
// an error if they don't match.
|
||||||
if root := statedb.IntermediateRoot(v.config.IsEIP158(header.Number)); header.Root != root {
|
if root := statedb.IntermediateRoot(v.config.IsEIP158(header.Number)); header.Root != root {
|
||||||
return fmt.Errorf("invalid merkle root (remote: %x local: %x)", header.Root, root)
|
return fmt.Errorf("invalid merkle root (remote: %x local: %x) dberr: %w", header.Root, root, statedb.Error())
|
||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
@ -156,6 +156,11 @@ func (b *BlockGen) Number() *big.Int {
|
|||||||
return new(big.Int).Set(b.header.Number)
|
return new(big.Int).Set(b.header.Number)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Timestamp returns the timestamp of the block being generated.
|
||||||
|
func (b *BlockGen) Timestamp() uint64 {
|
||||||
|
return b.header.Time
|
||||||
|
}
|
||||||
|
|
||||||
// BaseFee returns the EIP-1559 base fee of the block being generated.
|
// BaseFee returns the EIP-1559 base fee of the block being generated.
|
||||||
func (b *BlockGen) BaseFee() *big.Int {
|
func (b *BlockGen) BaseFee() *big.Int {
|
||||||
return new(big.Int).Set(b.header.BaseFee)
|
return new(big.Int).Set(b.header.BaseFee)
|
||||||
|
@ -70,10 +70,10 @@ func NewEVMBlockContext(header *types.Header, chain ChainContext, author *common
|
|||||||
}
|
}
|
||||||
|
|
||||||
// NewEVMTxContext creates a new transaction context for a single transaction.
|
// NewEVMTxContext creates a new transaction context for a single transaction.
|
||||||
func NewEVMTxContext(msg Message) vm.TxContext {
|
func NewEVMTxContext(msg *Message) vm.TxContext {
|
||||||
return vm.TxContext{
|
return vm.TxContext{
|
||||||
Origin: msg.From(),
|
Origin: msg.From,
|
||||||
GasPrice: new(big.Int).Set(msg.GasPrice()),
|
GasPrice: new(big.Int).Set(msg.GasPrice),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -26,15 +26,9 @@ import (
|
|||||||
"github.com/ethereum/go-ethereum/rlp"
|
"github.com/ethereum/go-ethereum/rlp"
|
||||||
)
|
)
|
||||||
|
|
||||||
func u64(val uint64) *uint64 { return &val }
|
|
||||||
|
|
||||||
// TestCreation tests that different genesis and fork rule combinations result in
|
// TestCreation tests that different genesis and fork rule combinations result in
|
||||||
// the correct fork ID.
|
// the correct fork ID.
|
||||||
func TestCreation(t *testing.T) {
|
func TestCreation(t *testing.T) {
|
||||||
// Temporary non-existent scenario TODO(karalabe): delete when Shanghai is enabled
|
|
||||||
timestampedConfig := *params.MainnetChainConfig
|
|
||||||
timestampedConfig.ShanghaiTime = u64(1668000000)
|
|
||||||
|
|
||||||
type testcase struct {
|
type testcase struct {
|
||||||
head uint64
|
head uint64
|
||||||
time uint64
|
time uint64
|
||||||
@ -50,32 +44,34 @@ func TestCreation(t *testing.T) {
|
|||||||
params.MainnetChainConfig,
|
params.MainnetChainConfig,
|
||||||
params.MainnetGenesisHash,
|
params.MainnetGenesisHash,
|
||||||
[]testcase{
|
[]testcase{
|
||||||
{0, 0, ID{Hash: checksumToBytes(0xfc64ec04), Next: 1150000}}, // Unsynced
|
{0, 0, ID{Hash: checksumToBytes(0xfc64ec04), Next: 1150000}}, // Unsynced
|
||||||
{1149999, 0, ID{Hash: checksumToBytes(0xfc64ec04), Next: 1150000}}, // Last Frontier block
|
{1149999, 0, ID{Hash: checksumToBytes(0xfc64ec04), Next: 1150000}}, // Last Frontier block
|
||||||
{1150000, 0, ID{Hash: checksumToBytes(0x97c2c34c), Next: 1920000}}, // First Homestead block
|
{1150000, 0, ID{Hash: checksumToBytes(0x97c2c34c), Next: 1920000}}, // First Homestead block
|
||||||
{1919999, 0, ID{Hash: checksumToBytes(0x97c2c34c), Next: 1920000}}, // Last Homestead block
|
{1919999, 0, ID{Hash: checksumToBytes(0x97c2c34c), Next: 1920000}}, // Last Homestead block
|
||||||
{1920000, 0, ID{Hash: checksumToBytes(0x91d1f948), Next: 2463000}}, // First DAO block
|
{1920000, 0, ID{Hash: checksumToBytes(0x91d1f948), Next: 2463000}}, // First DAO block
|
||||||
{2462999, 0, ID{Hash: checksumToBytes(0x91d1f948), Next: 2463000}}, // Last DAO block
|
{2462999, 0, ID{Hash: checksumToBytes(0x91d1f948), Next: 2463000}}, // Last DAO block
|
||||||
{2463000, 0, ID{Hash: checksumToBytes(0x7a64da13), Next: 2675000}}, // First Tangerine block
|
{2463000, 0, ID{Hash: checksumToBytes(0x7a64da13), Next: 2675000}}, // First Tangerine block
|
||||||
{2674999, 0, ID{Hash: checksumToBytes(0x7a64da13), Next: 2675000}}, // Last Tangerine block
|
{2674999, 0, ID{Hash: checksumToBytes(0x7a64da13), Next: 2675000}}, // Last Tangerine block
|
||||||
{2675000, 0, ID{Hash: checksumToBytes(0x3edd5b10), Next: 4370000}}, // First Spurious block
|
{2675000, 0, ID{Hash: checksumToBytes(0x3edd5b10), Next: 4370000}}, // First Spurious block
|
||||||
{4369999, 0, ID{Hash: checksumToBytes(0x3edd5b10), Next: 4370000}}, // Last Spurious block
|
{4369999, 0, ID{Hash: checksumToBytes(0x3edd5b10), Next: 4370000}}, // Last Spurious block
|
||||||
{4370000, 0, ID{Hash: checksumToBytes(0xa00bc324), Next: 7280000}}, // First Byzantium block
|
{4370000, 0, ID{Hash: checksumToBytes(0xa00bc324), Next: 7280000}}, // First Byzantium block
|
||||||
{7279999, 0, ID{Hash: checksumToBytes(0xa00bc324), Next: 7280000}}, // Last Byzantium block
|
{7279999, 0, ID{Hash: checksumToBytes(0xa00bc324), Next: 7280000}}, // Last Byzantium block
|
||||||
{7280000, 0, ID{Hash: checksumToBytes(0x668db0af), Next: 9069000}}, // First and last Constantinople, first Petersburg block
|
{7280000, 0, ID{Hash: checksumToBytes(0x668db0af), Next: 9069000}}, // First and last Constantinople, first Petersburg block
|
||||||
{9068999, 0, ID{Hash: checksumToBytes(0x668db0af), Next: 9069000}}, // Last Petersburg block
|
{9068999, 0, ID{Hash: checksumToBytes(0x668db0af), Next: 9069000}}, // Last Petersburg block
|
||||||
{9069000, 0, ID{Hash: checksumToBytes(0x879d6e30), Next: 9200000}}, // First Istanbul and first Muir Glacier block
|
{9069000, 0, ID{Hash: checksumToBytes(0x879d6e30), Next: 9200000}}, // First Istanbul and first Muir Glacier block
|
||||||
{9199999, 0, ID{Hash: checksumToBytes(0x879d6e30), Next: 9200000}}, // Last Istanbul and first Muir Glacier block
|
{9199999, 0, ID{Hash: checksumToBytes(0x879d6e30), Next: 9200000}}, // Last Istanbul and first Muir Glacier block
|
||||||
{9200000, 0, ID{Hash: checksumToBytes(0xe029e991), Next: 12244000}}, // First Muir Glacier block
|
{9200000, 0, ID{Hash: checksumToBytes(0xe029e991), Next: 12244000}}, // First Muir Glacier block
|
||||||
{12243999, 0, ID{Hash: checksumToBytes(0xe029e991), Next: 12244000}}, // Last Muir Glacier block
|
{12243999, 0, ID{Hash: checksumToBytes(0xe029e991), Next: 12244000}}, // Last Muir Glacier block
|
||||||
{12244000, 0, ID{Hash: checksumToBytes(0x0eb440f6), Next: 12965000}}, // First Berlin block
|
{12244000, 0, ID{Hash: checksumToBytes(0x0eb440f6), Next: 12965000}}, // First Berlin block
|
||||||
{12964999, 0, ID{Hash: checksumToBytes(0x0eb440f6), Next: 12965000}}, // Last Berlin block
|
{12964999, 0, ID{Hash: checksumToBytes(0x0eb440f6), Next: 12965000}}, // Last Berlin block
|
||||||
{12965000, 0, ID{Hash: checksumToBytes(0xb715077d), Next: 13773000}}, // First London block
|
{12965000, 0, ID{Hash: checksumToBytes(0xb715077d), Next: 13773000}}, // First London block
|
||||||
{13772999, 0, ID{Hash: checksumToBytes(0xb715077d), Next: 13773000}}, // Last London block
|
{13772999, 0, ID{Hash: checksumToBytes(0xb715077d), Next: 13773000}}, // Last London block
|
||||||
{13773000, 0, ID{Hash: checksumToBytes(0x20c327fc), Next: 15050000}}, // First Arrow Glacier block
|
{13773000, 0, ID{Hash: checksumToBytes(0x20c327fc), Next: 15050000}}, // First Arrow Glacier block
|
||||||
{15049999, 0, ID{Hash: checksumToBytes(0x20c327fc), Next: 15050000}}, // Last Arrow Glacier block
|
{15049999, 0, ID{Hash: checksumToBytes(0x20c327fc), Next: 15050000}}, // Last Arrow Glacier block
|
||||||
{15050000, 0, ID{Hash: checksumToBytes(0xf0afd0e3), Next: 0}}, // First Gray Glacier block
|
{15050000, 0, ID{Hash: checksumToBytes(0xf0afd0e3), Next: 1681338455}}, // First Gray Glacier block
|
||||||
{20000000, 0, ID{Hash: checksumToBytes(0xf0afd0e3), Next: 0}}, // Future Gray Glacier block
|
{20000000, 1681338454, ID{Hash: checksumToBytes(0xf0afd0e3), Next: 1681338455}}, // Last Gray Glacier block
|
||||||
|
{20000000, 1681338455, ID{Hash: checksumToBytes(0xdce96c2d), Next: 0}}, // First Shanghai block
|
||||||
|
{30000000, 2000000000, ID{Hash: checksumToBytes(0xdce96c2d), Next: 0}}, // Future Shanghai block
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
// Rinkeby test cases
|
// Rinkeby test cases
|
||||||
@ -131,41 +127,6 @@ func TestCreation(t *testing.T) {
|
|||||||
{1735372, 1677557088, ID{Hash: checksumToBytes(0xf7f9bc08), Next: 0}}, // First Shanghai block
|
{1735372, 1677557088, ID{Hash: checksumToBytes(0xf7f9bc08), Next: 0}}, // First Shanghai block
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
// Temporary timestamped test cases
|
|
||||||
{
|
|
||||||
×tampedConfig,
|
|
||||||
params.MainnetGenesisHash,
|
|
||||||
[]testcase{
|
|
||||||
{0, 0, ID{Hash: checksumToBytes(0xfc64ec04), Next: 1150000}}, // Unsynced
|
|
||||||
{1149999, 0, ID{Hash: checksumToBytes(0xfc64ec04), Next: 1150000}}, // Last Frontier block
|
|
||||||
{1150000, 0, ID{Hash: checksumToBytes(0x97c2c34c), Next: 1920000}}, // First Homestead block
|
|
||||||
{1919999, 0, ID{Hash: checksumToBytes(0x97c2c34c), Next: 1920000}}, // Last Homestead block
|
|
||||||
{1920000, 0, ID{Hash: checksumToBytes(0x91d1f948), Next: 2463000}}, // First DAO block
|
|
||||||
{2462999, 0, ID{Hash: checksumToBytes(0x91d1f948), Next: 2463000}}, // Last DAO block
|
|
||||||
{2463000, 0, ID{Hash: checksumToBytes(0x7a64da13), Next: 2675000}}, // First Tangerine block
|
|
||||||
{2674999, 0, ID{Hash: checksumToBytes(0x7a64da13), Next: 2675000}}, // Last Tangerine block
|
|
||||||
{2675000, 0, ID{Hash: checksumToBytes(0x3edd5b10), Next: 4370000}}, // First Spurious block
|
|
||||||
{4369999, 0, ID{Hash: checksumToBytes(0x3edd5b10), Next: 4370000}}, // Last Spurious block
|
|
||||||
{4370000, 0, ID{Hash: checksumToBytes(0xa00bc324), Next: 7280000}}, // First Byzantium block
|
|
||||||
{7279999, 0, ID{Hash: checksumToBytes(0xa00bc324), Next: 7280000}}, // Last Byzantium block
|
|
||||||
{7280000, 0, ID{Hash: checksumToBytes(0x668db0af), Next: 9069000}}, // First and last Constantinople, first Petersburg block
|
|
||||||
{9068999, 0, ID{Hash: checksumToBytes(0x668db0af), Next: 9069000}}, // Last Petersburg block
|
|
||||||
{9069000, 0, ID{Hash: checksumToBytes(0x879d6e30), Next: 9200000}}, // First Istanbul and first Muir Glacier block
|
|
||||||
{9199999, 0, ID{Hash: checksumToBytes(0x879d6e30), Next: 9200000}}, // Last Istanbul and first Muir Glacier block
|
|
||||||
{9200000, 0, ID{Hash: checksumToBytes(0xe029e991), Next: 12244000}}, // First Muir Glacier block
|
|
||||||
{12243999, 0, ID{Hash: checksumToBytes(0xe029e991), Next: 12244000}}, // Last Muir Glacier block
|
|
||||||
{12244000, 0, ID{Hash: checksumToBytes(0x0eb440f6), Next: 12965000}}, // First Berlin block
|
|
||||||
{12964999, 0, ID{Hash: checksumToBytes(0x0eb440f6), Next: 12965000}}, // Last Berlin block
|
|
||||||
{12965000, 0, ID{Hash: checksumToBytes(0xb715077d), Next: 13773000}}, // First London block
|
|
||||||
{13772999, 0, ID{Hash: checksumToBytes(0xb715077d), Next: 13773000}}, // Last London block
|
|
||||||
{13773000, 0, ID{Hash: checksumToBytes(0x20c327fc), Next: 15050000}}, // First Arrow Glacier block
|
|
||||||
{15049999, 0, ID{Hash: checksumToBytes(0x20c327fc), Next: 15050000}}, // Last Arrow Glacier block
|
|
||||||
{15050000, 0, ID{Hash: checksumToBytes(0xf0afd0e3), Next: 1668000000}}, // First Gray Glacier block
|
|
||||||
{19999999, 1667999999, ID{Hash: checksumToBytes(0xf0afd0e3), Next: 1668000000}}, // Last Gray Glacier block
|
|
||||||
{20000000, 1668000000, ID{Hash: checksumToBytes(0x71147644), Next: 0}}, // First Shanghai block
|
|
||||||
{20000000, 2668000000, ID{Hash: checksumToBytes(0x71147644), Next: 0}}, // Future Shanghai block
|
|
||||||
},
|
|
||||||
},
|
|
||||||
}
|
}
|
||||||
for i, tt := range tests {
|
for i, tt := range tests {
|
||||||
for j, ttt := range tt.cases {
|
for j, ttt := range tt.cases {
|
||||||
@ -179,9 +140,9 @@ func TestCreation(t *testing.T) {
|
|||||||
// TestValidation tests that a local peer correctly validates and accepts a remote
|
// TestValidation tests that a local peer correctly validates and accepts a remote
|
||||||
// fork ID.
|
// fork ID.
|
||||||
func TestValidation(t *testing.T) {
|
func TestValidation(t *testing.T) {
|
||||||
// Temporary non-existent scenario TODO(karalabe): delete when Shanghai is enabled
|
// Config that has not timestamp enabled
|
||||||
timestampedConfig := *params.MainnetChainConfig
|
legacyConfig := *params.MainnetChainConfig
|
||||||
timestampedConfig.ShanghaiTime = u64(1668000000)
|
legacyConfig.ShanghaiTime = nil
|
||||||
|
|
||||||
tests := []struct {
|
tests := []struct {
|
||||||
config *params.ChainConfig
|
config *params.ChainConfig
|
||||||
@ -195,60 +156,60 @@ func TestValidation(t *testing.T) {
|
|||||||
//------------------
|
//------------------
|
||||||
|
|
||||||
// Local is mainnet Gray Glacier, remote announces the same. No future fork is announced.
|
// Local is mainnet Gray Glacier, remote announces the same. No future fork is announced.
|
||||||
{params.MainnetChainConfig, 15050000, 0, ID{Hash: checksumToBytes(0xf0afd0e3), Next: 0}, nil},
|
{&legacyConfig, 15050000, 0, ID{Hash: checksumToBytes(0xf0afd0e3), Next: 0}, nil},
|
||||||
|
|
||||||
// Local is mainnet Gray Glacier, remote announces the same. Remote also announces a next fork
|
// Local is mainnet Gray Glacier, remote announces the same. Remote also announces a next fork
|
||||||
// at block 0xffffffff, but that is uncertain.
|
// at block 0xffffffff, but that is uncertain.
|
||||||
{params.MainnetChainConfig, 15050000, 0, ID{Hash: checksumToBytes(0xf0afd0e3), Next: math.MaxUint64}, nil},
|
{&legacyConfig, 15050000, 0, ID{Hash: checksumToBytes(0xf0afd0e3), Next: math.MaxUint64}, nil},
|
||||||
|
|
||||||
// Local is mainnet currently in Byzantium only (so it's aware of Petersburg), remote announces
|
// Local is mainnet currently in Byzantium only (so it's aware of Petersburg), remote announces
|
||||||
// also Byzantium, but it's not yet aware of Petersburg (e.g. non updated node before the fork).
|
// also Byzantium, but it's not yet aware of Petersburg (e.g. non updated node before the fork).
|
||||||
// In this case we don't know if Petersburg passed yet or not.
|
// In this case we don't know if Petersburg passed yet or not.
|
||||||
{params.MainnetChainConfig, 7279999, 0, ID{Hash: checksumToBytes(0xa00bc324), Next: 0}, nil},
|
{&legacyConfig, 7279999, 0, ID{Hash: checksumToBytes(0xa00bc324), Next: 0}, nil},
|
||||||
|
|
||||||
// Local is mainnet currently in Byzantium only (so it's aware of Petersburg), remote announces
|
// Local is mainnet currently in Byzantium only (so it's aware of Petersburg), remote announces
|
||||||
// also Byzantium, and it's also aware of Petersburg (e.g. updated node before the fork). We
|
// also Byzantium, and it's also aware of Petersburg (e.g. updated node before the fork). We
|
||||||
// don't know if Petersburg passed yet (will pass) or not.
|
// don't know if Petersburg passed yet (will pass) or not.
|
||||||
{params.MainnetChainConfig, 7279999, 0, ID{Hash: checksumToBytes(0xa00bc324), Next: 7280000}, nil},
|
{&legacyConfig, 7279999, 0, ID{Hash: checksumToBytes(0xa00bc324), Next: 7280000}, nil},
|
||||||
|
|
||||||
// Local is mainnet currently in Byzantium only (so it's aware of Petersburg), remote announces
|
// Local is mainnet currently in Byzantium only (so it's aware of Petersburg), remote announces
|
||||||
// also Byzantium, and it's also aware of some random fork (e.g. misconfigured Petersburg). As
|
// also Byzantium, and it's also aware of some random fork (e.g. misconfigured Petersburg). As
|
||||||
// neither forks passed at neither nodes, they may mismatch, but we still connect for now.
|
// neither forks passed at neither nodes, they may mismatch, but we still connect for now.
|
||||||
{params.MainnetChainConfig, 7279999, 0, ID{Hash: checksumToBytes(0xa00bc324), Next: math.MaxUint64}, nil},
|
{&legacyConfig, 7279999, 0, ID{Hash: checksumToBytes(0xa00bc324), Next: math.MaxUint64}, nil},
|
||||||
|
|
||||||
// Local is mainnet exactly on Petersburg, remote announces Byzantium + knowledge about Petersburg. Remote
|
// Local is mainnet exactly on Petersburg, remote announces Byzantium + knowledge about Petersburg. Remote
|
||||||
// is simply out of sync, accept.
|
// is simply out of sync, accept.
|
||||||
{params.MainnetChainConfig, 7280000, 0, ID{Hash: checksumToBytes(0xa00bc324), Next: 7280000}, nil},
|
{&legacyConfig, 7280000, 0, ID{Hash: checksumToBytes(0xa00bc324), Next: 7280000}, nil},
|
||||||
|
|
||||||
// Local is mainnet Petersburg, remote announces Byzantium + knowledge about Petersburg. Remote
|
// Local is mainnet Petersburg, remote announces Byzantium + knowledge about Petersburg. Remote
|
||||||
// is simply out of sync, accept.
|
// is simply out of sync, accept.
|
||||||
{params.MainnetChainConfig, 7987396, 0, ID{Hash: checksumToBytes(0xa00bc324), Next: 7280000}, nil},
|
{&legacyConfig, 7987396, 0, ID{Hash: checksumToBytes(0xa00bc324), Next: 7280000}, nil},
|
||||||
|
|
||||||
// Local is mainnet Petersburg, remote announces Spurious + knowledge about Byzantium. Remote
|
// Local is mainnet Petersburg, remote announces Spurious + knowledge about Byzantium. Remote
|
||||||
// is definitely out of sync. It may or may not need the Petersburg update, we don't know yet.
|
// is definitely out of sync. It may or may not need the Petersburg update, we don't know yet.
|
||||||
{params.MainnetChainConfig, 7987396, 0, ID{Hash: checksumToBytes(0x3edd5b10), Next: 4370000}, nil},
|
{&legacyConfig, 7987396, 0, ID{Hash: checksumToBytes(0x3edd5b10), Next: 4370000}, nil},
|
||||||
|
|
||||||
// Local is mainnet Byzantium, remote announces Petersburg. Local is out of sync, accept.
|
// Local is mainnet Byzantium, remote announces Petersburg. Local is out of sync, accept.
|
||||||
{params.MainnetChainConfig, 7279999, 0, ID{Hash: checksumToBytes(0x668db0af), Next: 0}, nil},
|
{&legacyConfig, 7279999, 0, ID{Hash: checksumToBytes(0x668db0af), Next: 0}, nil},
|
||||||
|
|
||||||
// Local is mainnet Spurious, remote announces Byzantium, but is not aware of Petersburg. Local
|
// Local is mainnet Spurious, remote announces Byzantium, but is not aware of Petersburg. Local
|
||||||
// out of sync. Local also knows about a future fork, but that is uncertain yet.
|
// out of sync. Local also knows about a future fork, but that is uncertain yet.
|
||||||
{params.MainnetChainConfig, 4369999, 0, ID{Hash: checksumToBytes(0xa00bc324), Next: 0}, nil},
|
{&legacyConfig, 4369999, 0, ID{Hash: checksumToBytes(0xa00bc324), Next: 0}, nil},
|
||||||
|
|
||||||
// Local is mainnet Petersburg. remote announces Byzantium but is not aware of further forks.
|
// Local is mainnet Petersburg. remote announces Byzantium but is not aware of further forks.
|
||||||
// Remote needs software update.
|
// Remote needs software update.
|
||||||
{params.MainnetChainConfig, 7987396, 0, ID{Hash: checksumToBytes(0xa00bc324), Next: 0}, ErrRemoteStale},
|
{&legacyConfig, 7987396, 0, ID{Hash: checksumToBytes(0xa00bc324), Next: 0}, ErrRemoteStale},
|
||||||
|
|
||||||
// Local is mainnet Petersburg, and isn't aware of more forks. Remote announces Petersburg +
|
// Local is mainnet Petersburg, and isn't aware of more forks. Remote announces Petersburg +
|
||||||
// 0xffffffff. Local needs software update, reject.
|
// 0xffffffff. Local needs software update, reject.
|
||||||
{params.MainnetChainConfig, 7987396, 0, ID{Hash: checksumToBytes(0x5cddc0e1), Next: 0}, ErrLocalIncompatibleOrStale},
|
{&legacyConfig, 7987396, 0, ID{Hash: checksumToBytes(0x5cddc0e1), Next: 0}, ErrLocalIncompatibleOrStale},
|
||||||
|
|
||||||
// Local is mainnet Byzantium, and is aware of Petersburg. Remote announces Petersburg +
|
// Local is mainnet Byzantium, and is aware of Petersburg. Remote announces Petersburg +
|
||||||
// 0xffffffff. Local needs software update, reject.
|
// 0xffffffff. Local needs software update, reject.
|
||||||
{params.MainnetChainConfig, 7279999, 0, ID{Hash: checksumToBytes(0x5cddc0e1), Next: 0}, ErrLocalIncompatibleOrStale},
|
{&legacyConfig, 7279999, 0, ID{Hash: checksumToBytes(0x5cddc0e1), Next: 0}, ErrLocalIncompatibleOrStale},
|
||||||
|
|
||||||
// Local is mainnet Petersburg, remote is Rinkeby Petersburg.
|
// Local is mainnet Petersburg, remote is Rinkeby Petersburg.
|
||||||
{params.MainnetChainConfig, 7987396, 0, ID{Hash: checksumToBytes(0xafec6b27), Next: 0}, ErrLocalIncompatibleOrStale},
|
{&legacyConfig, 7987396, 0, ID{Hash: checksumToBytes(0xafec6b27), Next: 0}, ErrLocalIncompatibleOrStale},
|
||||||
|
|
||||||
// Local is mainnet Gray Glacier, far in the future. Remote announces Gopherium (non existing fork)
|
// Local is mainnet Gray Glacier, far in the future. Remote announces Gopherium (non existing fork)
|
||||||
// at some future block 88888888, for itself, but past block for local. Local is incompatible.
|
// at some future block 88888888, for itself, but past block for local. Local is incompatible.
|
||||||
@ -256,13 +217,13 @@ func TestValidation(t *testing.T) {
|
|||||||
// This case detects non-upgraded nodes with majority hash power (typical Ropsten mess).
|
// This case detects non-upgraded nodes with majority hash power (typical Ropsten mess).
|
||||||
//
|
//
|
||||||
// TODO(karalabe): This testcase will fail once mainnet gets timestamped forks, make legacy chain config
|
// TODO(karalabe): This testcase will fail once mainnet gets timestamped forks, make legacy chain config
|
||||||
{params.MainnetChainConfig, 88888888, 0, ID{Hash: checksumToBytes(0xf0afd0e3), Next: 88888888}, ErrLocalIncompatibleOrStale},
|
{&legacyConfig, 88888888, 0, ID{Hash: checksumToBytes(0xf0afd0e3), Next: 88888888}, ErrLocalIncompatibleOrStale},
|
||||||
|
|
||||||
// Local is mainnet Byzantium. Remote is also in Byzantium, but announces Gopherium (non existing
|
// Local is mainnet Byzantium. Remote is also in Byzantium, but announces Gopherium (non existing
|
||||||
// fork) at block 7279999, before Petersburg. Local is incompatible.
|
// fork) at block 7279999, before Petersburg. Local is incompatible.
|
||||||
//
|
//
|
||||||
// TODO(karalabe): This testcase will fail once mainnet gets timestamped forks, make legacy chain config
|
// TODO(karalabe): This testcase will fail once mainnet gets timestamped forks, make legacy chain config
|
||||||
{params.MainnetChainConfig, 7279999, 0, ID{Hash: checksumToBytes(0xa00bc324), Next: 7279999}, ErrLocalIncompatibleOrStale},
|
{&legacyConfig, 7279999, 0, ID{Hash: checksumToBytes(0xa00bc324), Next: 7279999}, ErrLocalIncompatibleOrStale},
|
||||||
|
|
||||||
//------------------------------------
|
//------------------------------------
|
||||||
// Block to timestamp transition tests
|
// Block to timestamp transition tests
|
||||||
@ -271,48 +232,48 @@ func TestValidation(t *testing.T) {
|
|||||||
// Local is mainnet currently in Gray Glacier only (so it's aware of Shanghai), remote announces
|
// Local is mainnet currently in Gray Glacier only (so it's aware of Shanghai), remote announces
|
||||||
// also Gray Glacier, but it's not yet aware of Shanghai (e.g. non updated node before the fork).
|
// also Gray Glacier, but it's not yet aware of Shanghai (e.g. non updated node before the fork).
|
||||||
// In this case we don't know if Shanghai passed yet or not.
|
// In this case we don't know if Shanghai passed yet or not.
|
||||||
{×tampedConfig, 15050000, 0, ID{Hash: checksumToBytes(0xf0afd0e3), Next: 0}, nil},
|
{params.MainnetChainConfig, 15050000, 0, ID{Hash: checksumToBytes(0xf0afd0e3), Next: 0}, nil},
|
||||||
|
|
||||||
// Local is mainnet currently in Gray Glacier only (so it's aware of Shanghai), remote announces
|
// Local is mainnet currently in Gray Glacier only (so it's aware of Shanghai), remote announces
|
||||||
// also Gray Glacier, and it's also aware of Shanghai (e.g. updated node before the fork). We
|
// also Gray Glacier, and it's also aware of Shanghai (e.g. updated node before the fork). We
|
||||||
// don't know if Shanghai passed yet (will pass) or not.
|
// don't know if Shanghai passed yet (will pass) or not.
|
||||||
{×tampedConfig, 15050000, 0, ID{Hash: checksumToBytes(0xf0afd0e3), Next: 1668000000}, nil},
|
{params.MainnetChainConfig, 15050000, 0, ID{Hash: checksumToBytes(0xf0afd0e3), Next: 1681338455}, nil},
|
||||||
|
|
||||||
// Local is mainnet currently in Gray Glacier only (so it's aware of Shanghai), remote announces
|
// Local is mainnet currently in Gray Glacier only (so it's aware of Shanghai), remote announces
|
||||||
// also Gray Glacier, and it's also aware of some random fork (e.g. misconfigured Shanghai). As
|
// also Gray Glacier, and it's also aware of some random fork (e.g. misconfigured Shanghai). As
|
||||||
// neither forks passed at neither nodes, they may mismatch, but we still connect for now.
|
// neither forks passed at neither nodes, they may mismatch, but we still connect for now.
|
||||||
{×tampedConfig, 15050000, 0, ID{Hash: checksumToBytes(0xf0afd0e3), Next: math.MaxUint64}, nil},
|
{params.MainnetChainConfig, 15050000, 0, ID{Hash: checksumToBytes(0xf0afd0e3), Next: math.MaxUint64}, nil},
|
||||||
|
|
||||||
// Local is mainnet exactly on Shanghai, remote announces Gray Glacier + knowledge about Shanghai. Remote
|
// Local is mainnet exactly on Shanghai, remote announces Gray Glacier + knowledge about Shanghai. Remote
|
||||||
// is simply out of sync, accept.
|
// is simply out of sync, accept.
|
||||||
{×tampedConfig, 20000000, 1668000000, ID{Hash: checksumToBytes(0xf0afd0e3), Next: 1668000000}, nil},
|
{params.MainnetChainConfig, 20000000, 1681338455, ID{Hash: checksumToBytes(0xf0afd0e3), Next: 1681338455}, nil},
|
||||||
|
|
||||||
// Local is mainnet Shanghai, remote announces Gray Glacier + knowledge about Shanghai. Remote
|
// Local is mainnet Shanghai, remote announces Gray Glacier + knowledge about Shanghai. Remote
|
||||||
// is simply out of sync, accept.
|
// is simply out of sync, accept.
|
||||||
{×tampedConfig, 20123456, 1668123456, ID{Hash: checksumToBytes(0xf0afd0e3), Next: 1668000000}, nil},
|
{params.MainnetChainConfig, 20123456, 1681338456, ID{Hash: checksumToBytes(0xf0afd0e3), Next: 1681338455}, nil},
|
||||||
|
|
||||||
// Local is mainnet Shanghai, remote announces Arrow Glacier + knowledge about Gray Glacier. Remote
|
// Local is mainnet Shanghai, remote announces Arrow Glacier + knowledge about Gray Glacier. Remote
|
||||||
// is definitely out of sync. It may or may not need the Shanghai update, we don't know yet.
|
// is definitely out of sync. It may or may not need the Shanghai update, we don't know yet.
|
||||||
{×tampedConfig, 20000000, 1668000000, ID{Hash: checksumToBytes(0x20c327fc), Next: 15050000}, nil},
|
{params.MainnetChainConfig, 20000000, 1681338455, ID{Hash: checksumToBytes(0x20c327fc), Next: 15050000}, nil},
|
||||||
|
|
||||||
// Local is mainnet Gray Glacier, remote announces Shanghai. Local is out of sync, accept.
|
// Local is mainnet Gray Glacier, remote announces Shanghai. Local is out of sync, accept.
|
||||||
{×tampedConfig, 15050000, 0, ID{Hash: checksumToBytes(0x71147644), Next: 0}, nil},
|
{params.MainnetChainConfig, 15050000, 0, ID{Hash: checksumToBytes(0xdce96c2d), Next: 0}, nil},
|
||||||
|
|
||||||
// Local is mainnet Arrow Glacier, remote announces Gray Glacier, but is not aware of Shanghai. Local
|
// Local is mainnet Arrow Glacier, remote announces Gray Glacier, but is not aware of Shanghai. Local
|
||||||
// out of sync. Local also knows about a future fork, but that is uncertain yet.
|
// out of sync. Local also knows about a future fork, but that is uncertain yet.
|
||||||
{×tampedConfig, 13773000, 0, ID{Hash: checksumToBytes(0xf0afd0e3), Next: 0}, nil},
|
{params.MainnetChainConfig, 13773000, 0, ID{Hash: checksumToBytes(0xf0afd0e3), Next: 0}, nil},
|
||||||
|
|
||||||
// Local is mainnet Shanghai. remote announces Gray Glacier but is not aware of further forks.
|
// Local is mainnet Shanghai. remote announces Gray Glacier but is not aware of further forks.
|
||||||
// Remote needs software update.
|
// Remote needs software update.
|
||||||
{×tampedConfig, 20000000, 1668000000, ID{Hash: checksumToBytes(0xf0afd0e3), Next: 0}, ErrRemoteStale},
|
{params.MainnetChainConfig, 20000000, 1681338455, ID{Hash: checksumToBytes(0xf0afd0e3), Next: 0}, ErrRemoteStale},
|
||||||
|
|
||||||
// Local is mainnet Gray Glacier, and isn't aware of more forks. Remote announces Gray Glacier +
|
// Local is mainnet Gray Glacier, and isn't aware of more forks. Remote announces Gray Glacier +
|
||||||
// 0xffffffff. Local needs software update, reject.
|
// 0xffffffff. Local needs software update, reject.
|
||||||
{×tampedConfig, 15050000, 0, ID{Hash: checksumToBytes(checksumUpdate(0xf0afd0e3, math.MaxUint64)), Next: 0}, ErrLocalIncompatibleOrStale},
|
{params.MainnetChainConfig, 15050000, 0, ID{Hash: checksumToBytes(checksumUpdate(0xf0afd0e3, math.MaxUint64)), Next: 0}, ErrLocalIncompatibleOrStale},
|
||||||
|
|
||||||
// Local is mainnet Gray Glacier, and is aware of Shanghai. Remote announces Shanghai +
|
// Local is mainnet Gray Glacier, and is aware of Shanghai. Remote announces Shanghai +
|
||||||
// 0xffffffff. Local needs software update, reject.
|
// 0xffffffff. Local needs software update, reject.
|
||||||
{×tampedConfig, 15050000, 0, ID{Hash: checksumToBytes(checksumUpdate(0x71147644, math.MaxUint64)), Next: 0}, ErrLocalIncompatibleOrStale},
|
{params.MainnetChainConfig, 15050000, 0, ID{Hash: checksumToBytes(checksumUpdate(0xdce96c2d, math.MaxUint64)), Next: 0}, ErrLocalIncompatibleOrStale},
|
||||||
|
|
||||||
// Local is mainnet Gray Glacier, far in the future. Remote announces Gopherium (non existing fork)
|
// Local is mainnet Gray Glacier, far in the future. Remote announces Gopherium (non existing fork)
|
||||||
// at some future timestamp 8888888888, for itself, but past block for local. Local is incompatible.
|
// at some future timestamp 8888888888, for itself, but past block for local. Local is incompatible.
|
||||||
@ -322,92 +283,92 @@ func TestValidation(t *testing.T) {
|
|||||||
|
|
||||||
// Local is mainnet Gray Glacier. Remote is also in Gray Glacier, but announces Gopherium (non existing
|
// Local is mainnet Gray Glacier. Remote is also in Gray Glacier, but announces Gopherium (non existing
|
||||||
// fork) at block 7279999, before Shanghai. Local is incompatible.
|
// fork) at block 7279999, before Shanghai. Local is incompatible.
|
||||||
{×tampedConfig, 19999999, 1667999999, ID{Hash: checksumToBytes(0xf0afd0e3), Next: 1667999999}, ErrLocalIncompatibleOrStale},
|
{params.MainnetChainConfig, 19999999, 1667999999, ID{Hash: checksumToBytes(0xf0afd0e3), Next: 1667999999}, ErrLocalIncompatibleOrStale},
|
||||||
|
|
||||||
//----------------------
|
//----------------------
|
||||||
// Timestamp based tests
|
// Timestamp based tests
|
||||||
//----------------------
|
//----------------------
|
||||||
|
|
||||||
// Local is mainnet Shanghai, remote announces the same. No future fork is announced.
|
// Local is mainnet Shanghai, remote announces the same. No future fork is announced.
|
||||||
{×tampedConfig, 20000000, 1668000000, ID{Hash: checksumToBytes(0x71147644), Next: 0}, nil},
|
{params.MainnetChainConfig, 20000000, 1681338455, ID{Hash: checksumToBytes(0xdce96c2d), Next: 0}, nil},
|
||||||
|
|
||||||
// Local is mainnet Shanghai, remote announces the same. Remote also announces a next fork
|
// Local is mainnet Shanghai, remote announces the same. Remote also announces a next fork
|
||||||
// at time 0xffffffff, but that is uncertain.
|
// at time 0xffffffff, but that is uncertain.
|
||||||
{×tampedConfig, 20000000, 1668000000, ID{Hash: checksumToBytes(0x71147644), Next: math.MaxUint64}, nil},
|
{params.MainnetChainConfig, 20000000, 1681338455, ID{Hash: checksumToBytes(0xdce96c2d), Next: math.MaxUint64}, nil},
|
||||||
|
|
||||||
// Local is mainnet currently in Shanghai only (so it's aware of Cancun), remote announces
|
// Local is mainnet currently in Shanghai only (so it's aware of Cancun), remote announces
|
||||||
// also Shanghai, but it's not yet aware of Cancun (e.g. non updated node before the fork).
|
// also Shanghai, but it's not yet aware of Cancun (e.g. non updated node before the fork).
|
||||||
// In this case we don't know if Cancun passed yet or not.
|
// In this case we don't know if Cancun passed yet or not.
|
||||||
//
|
//
|
||||||
// TODO(karalabe): Enable this when Cancun is specced
|
// TODO(karalabe): Enable this when Cancun is specced
|
||||||
//{×tampedConfig, 20000000, 1668000000, ID{Hash: checksumToBytes(0x71147644), Next: 0}, nil},
|
//{params.MainnetChainConfig, 20000000, 1668000000, ID{Hash: checksumToBytes(0x71147644), Next: 0}, nil},
|
||||||
|
|
||||||
// Local is mainnet currently in Shanghai only (so it's aware of Cancun), remote announces
|
// Local is mainnet currently in Shanghai only (so it's aware of Cancun), remote announces
|
||||||
// also Shanghai, and it's also aware of Cancun (e.g. updated node before the fork). We
|
// also Shanghai, and it's also aware of Cancun (e.g. updated node before the fork). We
|
||||||
// don't know if Cancun passed yet (will pass) or not.
|
// don't know if Cancun passed yet (will pass) or not.
|
||||||
//
|
//
|
||||||
// TODO(karalabe): Enable this when Cancun is specced and update next timestamp
|
// TODO(karalabe): Enable this when Cancun is specced and update next timestamp
|
||||||
//{×tampedConfig, 20000000, 1668000000, ID{Hash: checksumToBytes(0x71147644), Next: 1678000000}, nil},
|
//{params.MainnetChainConfig, 20000000, 1668000000, ID{Hash: checksumToBytes(0x71147644), Next: 1678000000}, nil},
|
||||||
|
|
||||||
// Local is mainnet currently in Shanghai only (so it's aware of Cancun), remote announces
|
// Local is mainnet currently in Shanghai only (so it's aware of Cancun), remote announces
|
||||||
// also Shanghai, and it's also aware of some random fork (e.g. misconfigured Cancun). As
|
// also Shanghai, and it's also aware of some random fork (e.g. misconfigured Cancun). As
|
||||||
// neither forks passed at neither nodes, they may mismatch, but we still connect for now.
|
// neither forks passed at neither nodes, they may mismatch, but we still connect for now.
|
||||||
//
|
//
|
||||||
// TODO(karalabe): Enable this when Cancun is specced
|
// TODO(karalabe): Enable this when Cancun is specced
|
||||||
//{×tampedConfig, 20000000, 1668000000, ID{Hash: checksumToBytes(0x71147644), Next: math.MaxUint64}, nil},
|
//{params.MainnetChainConfig, 20000000, 1668000000, ID{Hash: checksumToBytes(0x71147644), Next: math.MaxUint64}, nil},
|
||||||
|
|
||||||
// Local is mainnet exactly on Cancun, remote announces Shanghai + knowledge about Cancun. Remote
|
// Local is mainnet exactly on Cancun, remote announces Shanghai + knowledge about Cancun. Remote
|
||||||
// is simply out of sync, accept.
|
// is simply out of sync, accept.
|
||||||
//
|
//
|
||||||
// TODO(karalabe): Enable this when Cancun is specced, update local head and time, next timestamp
|
// TODO(karalabe): Enable this when Cancun is specced, update local head and time, next timestamp
|
||||||
// {×tampedConfig, 21000000, 1678000000, ID{Hash: checksumToBytes(0x71147644), Next: 1678000000}, nil},
|
// {params.MainnetChainConfig, 21000000, 1678000000, ID{Hash: checksumToBytes(0x71147644), Next: 1678000000}, nil},
|
||||||
|
|
||||||
// Local is mainnet Cancun, remote announces Shanghai + knowledge about Cancun. Remote
|
// Local is mainnet Cancun, remote announces Shanghai + knowledge about Cancun. Remote
|
||||||
// is simply out of sync, accept.
|
// is simply out of sync, accept.
|
||||||
// TODO(karalabe): Enable this when Cancun is specced, update local head and time, next timestamp
|
// TODO(karalabe): Enable this when Cancun is specced, update local head and time, next timestamp
|
||||||
//{×tampedConfig, 21123456, 1678123456, ID{Hash: checksumToBytes(0x71147644), Next: 1678000000}, nil},
|
//{params.MainnetChainConfig, 21123456, 1678123456, ID{Hash: checksumToBytes(0x71147644), Next: 1678000000}, nil},
|
||||||
|
|
||||||
// Local is mainnet Prague, remote announces Shanghai + knowledge about Cancun. Remote
|
// Local is mainnet Prague, remote announces Shanghai + knowledge about Cancun. Remote
|
||||||
// is definitely out of sync. It may or may not need the Prague update, we don't know yet.
|
// is definitely out of sync. It may or may not need the Prague update, we don't know yet.
|
||||||
//
|
//
|
||||||
// TODO(karalabe): Enable this when Cancun **and** Prague is specced, update all the numbers
|
// TODO(karalabe): Enable this when Cancun **and** Prague is specced, update all the numbers
|
||||||
//{×tampedConfig, 0, 0, ID{Hash: checksumToBytes(0x3edd5b10), Next: 4370000}, nil},
|
//{params.MainnetChainConfig, 0, 0, ID{Hash: checksumToBytes(0x3edd5b10), Next: 4370000}, nil},
|
||||||
|
|
||||||
// Local is mainnet Shanghai, remote announces Cancun. Local is out of sync, accept.
|
// Local is mainnet Shanghai, remote announces Cancun. Local is out of sync, accept.
|
||||||
//
|
//
|
||||||
// TODO(karalabe): Enable this when Cancun is specced, update remote checksum
|
// TODO(karalabe): Enable this when Cancun is specced, update remote checksum
|
||||||
//{×tampedConfig, 21000000, 1678000000, ID{Hash: checksumToBytes(0x00000000), Next: 0}, nil},
|
//{params.MainnetChainConfig, 21000000, 1678000000, ID{Hash: checksumToBytes(0x00000000), Next: 0}, nil},
|
||||||
|
|
||||||
// Local is mainnet Shanghai, remote announces Cancun, but is not aware of Prague. Local
|
// Local is mainnet Shanghai, remote announces Cancun, but is not aware of Prague. Local
|
||||||
// out of sync. Local also knows about a future fork, but that is uncertain yet.
|
// out of sync. Local also knows about a future fork, but that is uncertain yet.
|
||||||
//
|
//
|
||||||
// TODO(karalabe): Enable this when Cancun **and** Prague is specced, update remote checksum
|
// TODO(karalabe): Enable this when Cancun **and** Prague is specced, update remote checksum
|
||||||
//{×tampedConfig, 21000000, 1678000000, ID{Hash: checksumToBytes(0x00000000), Next: 0}, nil},
|
//{params.MainnetChainConfig, 21000000, 1678000000, ID{Hash: checksumToBytes(0x00000000), Next: 0}, nil},
|
||||||
|
|
||||||
// Local is mainnet Cancun. remote announces Shanghai but is not aware of further forks.
|
// Local is mainnet Cancun. remote announces Shanghai but is not aware of further forks.
|
||||||
// Remote needs software update.
|
// Remote needs software update.
|
||||||
//
|
//
|
||||||
// TODO(karalabe): Enable this when Cancun is specced, update local head and time
|
// TODO(karalabe): Enable this when Cancun is specced, update local head and time
|
||||||
//{×tampedConfig, 21000000, 1678000000, ID{Hash: checksumToBytes(0x71147644), Next: 0}, ErrRemoteStale},
|
//{params.MainnetChainConfig, 21000000, 1678000000, ID{Hash: checksumToBytes(0x71147644), Next: 0}, ErrRemoteStale},
|
||||||
|
|
||||||
// Local is mainnet Shanghai, and isn't aware of more forks. Remote announces Shanghai +
|
// Local is mainnet Shanghai, and isn't aware of more forks. Remote announces Shanghai +
|
||||||
// 0xffffffff. Local needs software update, reject.
|
// 0xffffffff. Local needs software update, reject.
|
||||||
{×tampedConfig, 20000000, 1668000000, ID{Hash: checksumToBytes(checksumUpdate(0x71147644, math.MaxUint64)), Next: 0}, ErrLocalIncompatibleOrStale},
|
{params.MainnetChainConfig, 20000000, 1681338455, ID{Hash: checksumToBytes(checksumUpdate(0xdce96c2d, math.MaxUint64)), Next: 0}, ErrLocalIncompatibleOrStale},
|
||||||
|
|
||||||
// Local is mainnet Shanghai, and is aware of Cancun. Remote announces Cancun +
|
// Local is mainnet Shanghai, and is aware of Cancun. Remote announces Cancun +
|
||||||
// 0xffffffff. Local needs software update, reject.
|
// 0xffffffff. Local needs software update, reject.
|
||||||
//
|
//
|
||||||
// TODO(karalabe): Enable this when Cancun is specced, update remote checksum
|
// TODO(karalabe): Enable this when Cancun is specced, update remote checksum
|
||||||
//{×tampedConfig, 20000000, 1668000000, ID{Hash: checksumToBytes(checksumUpdate(0x00000000, math.MaxUint64)), Next: 0}, ErrLocalIncompatibleOrStale},
|
//{params.MainnetChainConfig, 20000000, 1668000000, ID{Hash: checksumToBytes(checksumUpdate(0x00000000, math.MaxUint64)), Next: 0}, ErrLocalIncompatibleOrStale},
|
||||||
|
|
||||||
// Local is mainnet Shanghai, remote is random Shanghai.
|
// Local is mainnet Shanghai, remote is random Shanghai.
|
||||||
{×tampedConfig, 20000000, 1668000000, ID{Hash: checksumToBytes(0x12345678), Next: 0}, ErrLocalIncompatibleOrStale},
|
{params.MainnetChainConfig, 20000000, 1681338455, ID{Hash: checksumToBytes(0x12345678), Next: 0}, ErrLocalIncompatibleOrStale},
|
||||||
|
|
||||||
// Local is mainnet Shanghai, far in the future. Remote announces Gopherium (non existing fork)
|
// Local is mainnet Shanghai, far in the future. Remote announces Gopherium (non existing fork)
|
||||||
// at some future timestamp 8888888888, for itself, but past block for local. Local is incompatible.
|
// at some future timestamp 8888888888, for itself, but past block for local. Local is incompatible.
|
||||||
//
|
//
|
||||||
// This case detects non-upgraded nodes with majority hash power (typical Ropsten mess).
|
// This case detects non-upgraded nodes with majority hash power (typical Ropsten mess).
|
||||||
{×tampedConfig, 88888888, 8888888888, ID{Hash: checksumToBytes(0x71147644), Next: 8888888888}, ErrLocalIncompatibleOrStale},
|
{params.MainnetChainConfig, 88888888, 8888888888, ID{Hash: checksumToBytes(0xdce96c2d), Next: 8888888888}, ErrLocalIncompatibleOrStale},
|
||||||
|
|
||||||
// Local is mainnet Shanghai. Remote is also in Shanghai, but announces Gopherium (non existing
|
// Local is mainnet Shanghai. Remote is also in Shanghai, but announces Gopherium (non existing
|
||||||
// fork) at timestamp 1668000000, before Cancun. Local is incompatible.
|
// fork) at timestamp 1668000000, before Cancun. Local is incompatible.
|
||||||
|
@ -202,6 +202,7 @@ func NewDatabaseWithFreezer(db ethdb.KeyValueStore, ancient string, namespace st
|
|||||||
// Create the idle freezer instance
|
// Create the idle freezer instance
|
||||||
frdb, err := newChainFreezer(resolveChainFreezerDir(ancient), namespace, readonly)
|
frdb, err := newChainFreezer(resolveChainFreezerDir(ancient), namespace, readonly)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
printChainMetadata(db)
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
// Since the freezer can be stored separately from the user's key-value database,
|
// Since the freezer can be stored separately from the user's key-value database,
|
||||||
@ -233,8 +234,10 @@ func NewDatabaseWithFreezer(db ethdb.KeyValueStore, ancient string, namespace st
|
|||||||
// the freezer and the key-value store.
|
// the freezer and the key-value store.
|
||||||
frgenesis, err := frdb.Ancient(ChainFreezerHashTable, 0)
|
frgenesis, err := frdb.Ancient(ChainFreezerHashTable, 0)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
printChainMetadata(db)
|
||||||
return nil, fmt.Errorf("failed to retrieve genesis from ancient %v", err)
|
return nil, fmt.Errorf("failed to retrieve genesis from ancient %v", err)
|
||||||
} else if !bytes.Equal(kvgenesis, frgenesis) {
|
} else if !bytes.Equal(kvgenesis, frgenesis) {
|
||||||
|
printChainMetadata(db)
|
||||||
return nil, fmt.Errorf("genesis mismatch: %#x (leveldb) != %#x (ancients)", kvgenesis, frgenesis)
|
return nil, fmt.Errorf("genesis mismatch: %#x (leveldb) != %#x (ancients)", kvgenesis, frgenesis)
|
||||||
}
|
}
|
||||||
// Key-value store and freezer belong to the same network. Ensure that they
|
// Key-value store and freezer belong to the same network. Ensure that they
|
||||||
@ -242,8 +245,19 @@ func NewDatabaseWithFreezer(db ethdb.KeyValueStore, ancient string, namespace st
|
|||||||
if kvhash, _ := db.Get(headerHashKey(frozen)); len(kvhash) == 0 {
|
if kvhash, _ := db.Get(headerHashKey(frozen)); len(kvhash) == 0 {
|
||||||
// Subsequent header after the freezer limit is missing from the database.
|
// Subsequent header after the freezer limit is missing from the database.
|
||||||
// Reject startup if the database has a more recent head.
|
// Reject startup if the database has a more recent head.
|
||||||
if ldbNum := *ReadHeaderNumber(db, ReadHeadHeaderHash(db)); ldbNum > frozen-1 {
|
if head := *ReadHeaderNumber(db, ReadHeadHeaderHash(db)); head > frozen-1 {
|
||||||
return nil, fmt.Errorf("gap in the chain between ancients (#%d) and leveldb (#%d) ", frozen, ldbNum)
|
// Find the smallest block stored in the key-value store
|
||||||
|
// in range of [frozen, head]
|
||||||
|
var number uint64
|
||||||
|
for number = frozen; number <= head; number++ {
|
||||||
|
if present, _ := db.Has(headerHashKey(number)); present {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
// We are about to exit on error. Print database metdata beore exiting
|
||||||
|
printChainMetadata(db)
|
||||||
|
return nil, fmt.Errorf("gap in the chain between ancients [0 - #%d] and leveldb [#%d - #%d] ",
|
||||||
|
frozen-1, number, head)
|
||||||
}
|
}
|
||||||
// Database contains only older data than the freezer, this happens if the
|
// Database contains only older data than the freezer, this happens if the
|
||||||
// state was wiped and reinited from an existing freezer.
|
// state was wiped and reinited from an existing freezer.
|
||||||
@ -260,6 +274,7 @@ func NewDatabaseWithFreezer(db ethdb.KeyValueStore, ancient string, namespace st
|
|||||||
// Key-value store contains more data than the genesis block, make sure we
|
// Key-value store contains more data than the genesis block, make sure we
|
||||||
// didn't freeze anything yet.
|
// didn't freeze anything yet.
|
||||||
if kvblob, _ := db.Get(headerHashKey(1)); len(kvblob) == 0 {
|
if kvblob, _ := db.Get(headerHashKey(1)); len(kvblob) == 0 {
|
||||||
|
printChainMetadata(db)
|
||||||
return nil, errors.New("ancient chain segments already extracted, please set --datadir.ancient to the correct path")
|
return nil, errors.New("ancient chain segments already extracted, please set --datadir.ancient to the correct path")
|
||||||
}
|
}
|
||||||
// Block #1 is still in the database, we're allowed to init a new freezer
|
// Block #1 is still in the database, we're allowed to init a new freezer
|
||||||
@ -581,3 +596,42 @@ func InspectDatabase(db ethdb.Database, keyPrefix, keyStart []byte) error {
|
|||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// printChainMetadata prints out chain metadata to stderr.
|
||||||
|
func printChainMetadata(db ethdb.KeyValueStore) {
|
||||||
|
fmt.Fprintf(os.Stderr, "Chain metadata\n")
|
||||||
|
for _, v := range ReadChainMetadata(db) {
|
||||||
|
fmt.Fprintf(os.Stderr, " %s\n", strings.Join(v, ": "))
|
||||||
|
}
|
||||||
|
fmt.Fprintf(os.Stderr, "\n\n")
|
||||||
|
}
|
||||||
|
|
||||||
|
// ReadChainMetadata returns a set of key/value pairs that contains informatin
|
||||||
|
// about the database chain status. This can be used for diagnostic purposes
|
||||||
|
// when investigating the state of the node.
|
||||||
|
func ReadChainMetadata(db ethdb.KeyValueStore) [][]string {
|
||||||
|
pp := func(val *uint64) string {
|
||||||
|
if val == nil {
|
||||||
|
return "<nil>"
|
||||||
|
}
|
||||||
|
return fmt.Sprintf("%d (%#x)", *val, *val)
|
||||||
|
}
|
||||||
|
data := [][]string{
|
||||||
|
{"databaseVersion", pp(ReadDatabaseVersion(db))},
|
||||||
|
{"headBlockHash", fmt.Sprintf("%v", ReadHeadBlockHash(db))},
|
||||||
|
{"headFastBlockHash", fmt.Sprintf("%v", ReadHeadFastBlockHash(db))},
|
||||||
|
{"headHeaderHash", fmt.Sprintf("%v", ReadHeadHeaderHash(db))},
|
||||||
|
{"lastPivotNumber", pp(ReadLastPivotNumber(db))},
|
||||||
|
{"len(snapshotSyncStatus)", fmt.Sprintf("%d bytes", len(ReadSnapshotSyncStatus(db)))},
|
||||||
|
{"snapshotDisabled", fmt.Sprintf("%v", ReadSnapshotDisabled(db))},
|
||||||
|
{"snapshotJournal", fmt.Sprintf("%d bytes", len(ReadSnapshotJournal(db)))},
|
||||||
|
{"snapshotRecoveryNumber", pp(ReadSnapshotRecoveryNumber(db))},
|
||||||
|
{"snapshotRoot", fmt.Sprintf("%v", ReadSnapshotRoot(db))},
|
||||||
|
{"txIndexTail", pp(ReadTxIndexTail(db))},
|
||||||
|
{"fastTxLookupLimit", pp(ReadFastTxLookupLimit(db))},
|
||||||
|
}
|
||||||
|
if b := ReadSkeletonSyncStatus(db); b != nil {
|
||||||
|
data = append(data, []string{"SkeletonSyncStatus", string(b)})
|
||||||
|
}
|
||||||
|
return data
|
||||||
|
}
|
||||||
|
@ -43,7 +43,6 @@ func (s Storage) String() (str string) {
|
|||||||
for key, value := range s {
|
for key, value := range s {
|
||||||
str += fmt.Sprintf("%X : %X\n", key, value)
|
str += fmt.Sprintf("%X : %X\n", key, value)
|
||||||
}
|
}
|
||||||
|
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -52,7 +51,6 @@ func (s Storage) Copy() Storage {
|
|||||||
for key, value := range s {
|
for key, value := range s {
|
||||||
cpy[key] = value
|
cpy[key] = value
|
||||||
}
|
}
|
||||||
|
|
||||||
return cpy
|
return cpy
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -68,13 +66,6 @@ type stateObject struct {
|
|||||||
data types.StateAccount
|
data types.StateAccount
|
||||||
db *StateDB
|
db *StateDB
|
||||||
|
|
||||||
// DB error.
|
|
||||||
// State objects are used by the consensus core and VM which are
|
|
||||||
// unable to deal with database-level errors. Any error that occurs
|
|
||||||
// during a database read is memoized here and will eventually be returned
|
|
||||||
// by StateDB.Commit.
|
|
||||||
dbErr error
|
|
||||||
|
|
||||||
// Write caches.
|
// Write caches.
|
||||||
trie Trie // storage trie, which becomes non-nil on first access
|
trie Trie // storage trie, which becomes non-nil on first access
|
||||||
code Code // contract bytecode, which gets set when code is loaded
|
code Code // contract bytecode, which gets set when code is loaded
|
||||||
@ -84,7 +75,7 @@ type stateObject struct {
|
|||||||
dirtyStorage Storage // Storage entries that have been modified in the current transaction execution
|
dirtyStorage Storage // Storage entries that have been modified in the current transaction execution
|
||||||
|
|
||||||
// Cache flags.
|
// Cache flags.
|
||||||
// When an object is marked suicided it will be delete from the trie
|
// When an object is marked suicided it will be deleted from the trie
|
||||||
// during the "update" phase of the state transition.
|
// during the "update" phase of the state transition.
|
||||||
dirtyCode bool // true if the code was updated
|
dirtyCode bool // true if the code was updated
|
||||||
suicided bool
|
suicided bool
|
||||||
@ -123,13 +114,6 @@ func (s *stateObject) EncodeRLP(w io.Writer) error {
|
|||||||
return rlp.Encode(w, &s.data)
|
return rlp.Encode(w, &s.data)
|
||||||
}
|
}
|
||||||
|
|
||||||
// setError remembers the first non-nil error it is called with.
|
|
||||||
func (s *stateObject) setError(err error) {
|
|
||||||
if s.dbErr == nil {
|
|
||||||
s.dbErr = err
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func (s *stateObject) markSuicided() {
|
func (s *stateObject) markSuicided() {
|
||||||
s.suicided = true
|
s.suicided = true
|
||||||
}
|
}
|
||||||
@ -214,7 +198,7 @@ func (s *stateObject) GetCommittedState(db Database, key common.Hash) common.Has
|
|||||||
start := time.Now()
|
start := time.Now()
|
||||||
tr, err := s.getTrie(db)
|
tr, err := s.getTrie(db)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
s.setError(err)
|
s.db.setError(err)
|
||||||
return common.Hash{}
|
return common.Hash{}
|
||||||
}
|
}
|
||||||
enc, err = tr.TryGet(key.Bytes())
|
enc, err = tr.TryGet(key.Bytes())
|
||||||
@ -222,7 +206,7 @@ func (s *stateObject) GetCommittedState(db Database, key common.Hash) common.Has
|
|||||||
s.db.StorageReads += time.Since(start)
|
s.db.StorageReads += time.Since(start)
|
||||||
}
|
}
|
||||||
if err != nil {
|
if err != nil {
|
||||||
s.setError(err)
|
s.db.setError(err)
|
||||||
return common.Hash{}
|
return common.Hash{}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -230,7 +214,7 @@ func (s *stateObject) GetCommittedState(db Database, key common.Hash) common.Has
|
|||||||
if len(enc) > 0 {
|
if len(enc) > 0 {
|
||||||
_, content, _, err := rlp.Split(enc)
|
_, content, _, err := rlp.Split(enc)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
s.setError(err)
|
s.db.setError(err)
|
||||||
}
|
}
|
||||||
value.SetBytes(content)
|
value.SetBytes(content)
|
||||||
}
|
}
|
||||||
@ -296,7 +280,7 @@ func (s *stateObject) updateTrie(db Database) (Trie, error) {
|
|||||||
)
|
)
|
||||||
tr, err := s.getTrie(db)
|
tr, err := s.getTrie(db)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
s.setError(err)
|
s.db.setError(err)
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
// Insert all the pending updates into the trie
|
// Insert all the pending updates into the trie
|
||||||
@ -311,7 +295,7 @@ func (s *stateObject) updateTrie(db Database) (Trie, error) {
|
|||||||
var v []byte
|
var v []byte
|
||||||
if (value == common.Hash{}) {
|
if (value == common.Hash{}) {
|
||||||
if err := tr.TryDelete(key[:]); err != nil {
|
if err := tr.TryDelete(key[:]); err != nil {
|
||||||
s.setError(err)
|
s.db.setError(err)
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
s.db.StorageDeleted += 1
|
s.db.StorageDeleted += 1
|
||||||
@ -319,7 +303,7 @@ func (s *stateObject) updateTrie(db Database) (Trie, error) {
|
|||||||
// Encoding []byte cannot fail, ok to ignore the error.
|
// Encoding []byte cannot fail, ok to ignore the error.
|
||||||
v, _ = rlp.EncodeToBytes(common.TrimLeftZeroes(value[:]))
|
v, _ = rlp.EncodeToBytes(common.TrimLeftZeroes(value[:]))
|
||||||
if err := tr.TryUpdate(key[:], v); err != nil {
|
if err := tr.TryUpdate(key[:], v); err != nil {
|
||||||
s.setError(err)
|
s.db.setError(err)
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
s.db.StorageUpdated += 1
|
s.db.StorageUpdated += 1
|
||||||
@ -351,7 +335,6 @@ func (s *stateObject) updateTrie(db Database) (Trie, error) {
|
|||||||
func (s *stateObject) updateRoot(db Database) {
|
func (s *stateObject) updateRoot(db Database) {
|
||||||
tr, err := s.updateTrie(db)
|
tr, err := s.updateTrie(db)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
s.setError(fmt.Errorf("updateRoot (%x) error: %w", s.address, err))
|
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
// If nothing changed, don't bother with hashing anything
|
// If nothing changed, don't bother with hashing anything
|
||||||
@ -372,9 +355,6 @@ func (s *stateObject) commitTrie(db Database) (*trie.NodeSet, error) {
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
if s.dbErr != nil {
|
|
||||||
return nil, s.dbErr
|
|
||||||
}
|
|
||||||
// If nothing changed, don't bother with committing anything
|
// If nothing changed, don't bother with committing anything
|
||||||
if tr == nil {
|
if tr == nil {
|
||||||
return nil, nil
|
return nil, nil
|
||||||
@ -385,7 +365,7 @@ func (s *stateObject) commitTrie(db Database) (*trie.NodeSet, error) {
|
|||||||
}
|
}
|
||||||
root, nodes := tr.Commit(false)
|
root, nodes := tr.Commit(false)
|
||||||
s.data.Root = root
|
s.data.Root = root
|
||||||
return nodes, err
|
return nodes, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// AddBalance adds amount to s's balance.
|
// AddBalance adds amount to s's balance.
|
||||||
@ -457,7 +437,7 @@ func (s *stateObject) Code(db Database) []byte {
|
|||||||
}
|
}
|
||||||
code, err := db.ContractCode(s.addrHash, common.BytesToHash(s.CodeHash()))
|
code, err := db.ContractCode(s.addrHash, common.BytesToHash(s.CodeHash()))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
s.setError(fmt.Errorf("can't load code hash %x: %v", s.CodeHash(), err))
|
s.db.setError(fmt.Errorf("can't load code hash %x: %v", s.CodeHash(), err))
|
||||||
}
|
}
|
||||||
s.code = code
|
s.code = code
|
||||||
return code
|
return code
|
||||||
@ -475,7 +455,7 @@ func (s *stateObject) CodeSize(db Database) int {
|
|||||||
}
|
}
|
||||||
size, err := db.ContractCodeSize(s.addrHash, common.BytesToHash(s.CodeHash()))
|
size, err := db.ContractCodeSize(s.addrHash, common.BytesToHash(s.CodeHash()))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
s.setError(fmt.Errorf("can't load code size %x: %v", s.CodeHash(), err))
|
s.db.setError(fmt.Errorf("can't load code size %x: %v", s.CodeHash(), err))
|
||||||
}
|
}
|
||||||
return size
|
return size
|
||||||
}
|
}
|
||||||
@ -519,10 +499,3 @@ func (s *stateObject) Balance() *big.Int {
|
|||||||
func (s *stateObject) Nonce() uint64 {
|
func (s *stateObject) Nonce() uint64 {
|
||||||
return s.data.Nonce
|
return s.data.Nonce
|
||||||
}
|
}
|
||||||
|
|
||||||
// Value is never called, but must be present to allow stateObject to be used
|
|
||||||
// as a vm.Account interface that also satisfies the vm.ContractRef
|
|
||||||
// interface. Interfaces are awesome.
|
|
||||||
func (s *stateObject) Value() *big.Int {
|
|
||||||
panic("Value on stateObject should never be called")
|
|
||||||
}
|
|
||||||
|
@ -81,8 +81,10 @@ type StateDB struct {
|
|||||||
// DB error.
|
// DB error.
|
||||||
// State objects are used by the consensus core and VM which are
|
// State objects are used by the consensus core and VM which are
|
||||||
// unable to deal with database-level errors. Any error that occurs
|
// unable to deal with database-level errors. Any error that occurs
|
||||||
// during a database read is memoized here and will eventually be returned
|
// during a database read is memoized here and will eventually be
|
||||||
// by StateDB.Commit.
|
// returned by StateDB.Commit. Notably, this error is also shared
|
||||||
|
// by all cached state objects in case the database failure occurs
|
||||||
|
// when accessing state of accounts.
|
||||||
dbErr error
|
dbErr error
|
||||||
|
|
||||||
// The refund counter, also used by state transitioning.
|
// The refund counter, also used by state transitioning.
|
||||||
@ -196,6 +198,7 @@ func (s *StateDB) setError(err error) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Error returns the memorized database failure occurred earlier.
|
||||||
func (s *StateDB) Error() error {
|
func (s *StateDB) Error() error {
|
||||||
return s.dbErr
|
return s.dbErr
|
||||||
}
|
}
|
||||||
@ -487,13 +490,11 @@ func (s *StateDB) SetTransientState(addr common.Address, key, value common.Hash)
|
|||||||
if prev == value {
|
if prev == value {
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
s.journal.append(transientStorageChange{
|
s.journal.append(transientStorageChange{
|
||||||
account: &addr,
|
account: &addr,
|
||||||
key: key,
|
key: key,
|
||||||
prevalue: prev,
|
prevalue: prev,
|
||||||
})
|
})
|
||||||
|
|
||||||
s.setTransientState(addr, key, value)
|
s.setTransientState(addr, key, value)
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -966,6 +967,7 @@ func (s *StateDB) clearJournalAndRefund() {
|
|||||||
|
|
||||||
// Commit writes the state to the underlying in-memory trie database.
|
// Commit writes the state to the underlying in-memory trie database.
|
||||||
func (s *StateDB) Commit(deleteEmptyObjects bool) (common.Hash, error) {
|
func (s *StateDB) Commit(deleteEmptyObjects bool) (common.Hash, error) {
|
||||||
|
// Short circuit in case any database failure occurred earlier.
|
||||||
if s.dbErr != nil {
|
if s.dbErr != nil {
|
||||||
return common.Hash{}, fmt.Errorf("commit aborted due to earlier error: %v", s.dbErr)
|
return common.Hash{}, fmt.Errorf("commit aborted due to earlier error: %v", s.dbErr)
|
||||||
}
|
}
|
||||||
@ -979,11 +981,11 @@ func (s *StateDB) Commit(deleteEmptyObjects bool) (common.Hash, error) {
|
|||||||
storageTrieNodesUpdated int
|
storageTrieNodesUpdated int
|
||||||
storageTrieNodesDeleted int
|
storageTrieNodesDeleted int
|
||||||
nodes = trie.NewMergedNodeSet()
|
nodes = trie.NewMergedNodeSet()
|
||||||
|
codeWriter = s.db.DiskDB().NewBatch()
|
||||||
)
|
)
|
||||||
// begin PluGeth injection
|
// begin PluGeth injection
|
||||||
codeUpdates := make(map[common.Hash][]byte)
|
codeUpdates := make(map[common.Hash][]byte)
|
||||||
// end PluGeth injection
|
// end PluGeth injection
|
||||||
codeWriter := s.db.DiskDB().NewBatch()
|
|
||||||
for addr := range s.stateObjectsDirty {
|
for addr := range s.stateObjectsDirty {
|
||||||
if obj := s.stateObjects[addr]; !obj.deleted {
|
if obj := s.stateObjects[addr]; !obj.deleted {
|
||||||
// Write any contract code associated with the state object
|
// Write any contract code associated with the state object
|
||||||
|
@ -63,7 +63,7 @@ func (p *statePrefetcher) Prefetch(block *types.Block, statedb *state.StateDB, c
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
// Convert the transaction into an executable message and pre-cache its sender
|
// Convert the transaction into an executable message and pre-cache its sender
|
||||||
msg, err := tx.AsMessage(signer, header.BaseFee)
|
msg, err := TransactionToMessage(tx, signer, header.BaseFee)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return // Also invalid block, bail out
|
return // Also invalid block, bail out
|
||||||
}
|
}
|
||||||
@ -85,7 +85,7 @@ func (p *statePrefetcher) Prefetch(block *types.Block, statedb *state.StateDB, c
|
|||||||
// precacheTransaction attempts to apply a transaction to the given state database
|
// precacheTransaction attempts to apply a transaction to the given state database
|
||||||
// and uses the input parameters for its environment. The goal is not to execute
|
// and uses the input parameters for its environment. The goal is not to execute
|
||||||
// the transaction successfully, rather to warm up touched data slots.
|
// the transaction successfully, rather to warm up touched data slots.
|
||||||
func precacheTransaction(msg types.Message, config *params.ChainConfig, gaspool *GasPool, statedb *state.StateDB, header *types.Header, evm *vm.EVM) error {
|
func precacheTransaction(msg *Message, config *params.ChainConfig, gaspool *GasPool, statedb *state.StateDB, header *types.Header, evm *vm.EVM) error {
|
||||||
// Update the evm with the new transaction context.
|
// Update the evm with the new transaction context.
|
||||||
evm.Reset(NewEVMTxContext(msg), statedb)
|
evm.Reset(NewEVMTxContext(msg), statedb)
|
||||||
// Add addresses to access list if applicable
|
// Add addresses to access list if applicable
|
||||||
|
@ -85,7 +85,7 @@ func (p *StateProcessor) Process(block *types.Block, statedb *state.StateDB, cfg
|
|||||||
// end PluGeth code injection
|
// end PluGeth code injection
|
||||||
// Iterate over and process the individual transactions
|
// Iterate over and process the individual transactions
|
||||||
for i, tx := range block.Transactions() {
|
for i, tx := range block.Transactions() {
|
||||||
msg, err := tx.AsMessage(types.MakeSigner(p.config, header.Number), header.BaseFee)
|
msg, err := TransactionToMessage(tx, types.MakeSigner(p.config, header.Number), header.BaseFee)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
// begin PluGeth injection
|
// begin PluGeth injection
|
||||||
pluginBlockProcessingError(tx, block, err)
|
pluginBlockProcessingError(tx, block, err)
|
||||||
@ -127,7 +127,7 @@ func (p *StateProcessor) Process(block *types.Block, statedb *state.StateDB, cfg
|
|||||||
return receipts, allLogs, *usedGas, nil
|
return receipts, allLogs, *usedGas, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func applyTransaction(msg types.Message, config *params.ChainConfig, gp *GasPool, statedb *state.StateDB, blockNumber *big.Int, blockHash common.Hash, tx *types.Transaction, usedGas *uint64, evm *vm.EVM) (*types.Receipt, error) {
|
func applyTransaction(msg *Message, config *params.ChainConfig, gp *GasPool, statedb *state.StateDB, blockNumber *big.Int, blockHash common.Hash, tx *types.Transaction, usedGas *uint64, evm *vm.EVM) (*types.Receipt, error) {
|
||||||
// Create a new context to be used in the EVM environment.
|
// Create a new context to be used in the EVM environment.
|
||||||
txContext := NewEVMTxContext(msg)
|
txContext := NewEVMTxContext(msg)
|
||||||
evm.Reset(txContext, statedb)
|
evm.Reset(txContext, statedb)
|
||||||
@ -159,7 +159,7 @@ func applyTransaction(msg types.Message, config *params.ChainConfig, gp *GasPool
|
|||||||
receipt.GasUsed = result.UsedGas
|
receipt.GasUsed = result.UsedGas
|
||||||
|
|
||||||
// If the transaction created a contract, store the creation address in the receipt.
|
// If the transaction created a contract, store the creation address in the receipt.
|
||||||
if msg.To() == nil {
|
if msg.To == nil {
|
||||||
receipt.ContractAddress = crypto.CreateAddress(evm.TxContext.Origin, tx.Nonce())
|
receipt.ContractAddress = crypto.CreateAddress(evm.TxContext.Origin, tx.Nonce())
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -177,7 +177,7 @@ func applyTransaction(msg types.Message, config *params.ChainConfig, gp *GasPool
|
|||||||
// for the transaction, gas used and an error if the transaction failed,
|
// for the transaction, gas used and an error if the transaction failed,
|
||||||
// indicating the block was invalid.
|
// indicating the block was invalid.
|
||||||
func ApplyTransaction(config *params.ChainConfig, bc ChainContext, author *common.Address, gp *GasPool, statedb *state.StateDB, header *types.Header, tx *types.Transaction, usedGas *uint64, cfg vm.Config) (*types.Receipt, error) {
|
func ApplyTransaction(config *params.ChainConfig, bc ChainContext, author *common.Address, gp *GasPool, statedb *state.StateDB, header *types.Header, tx *types.Transaction, usedGas *uint64, cfg vm.Config) (*types.Receipt, error) {
|
||||||
msg, err := tx.AsMessage(types.MakeSigner(config, header.Number), header.BaseFee)
|
msg, err := TransactionToMessage(tx, types.MakeSigner(config, header.Number), header.BaseFee)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
@ -25,65 +25,9 @@ import (
|
|||||||
cmath "github.com/ethereum/go-ethereum/common/math"
|
cmath "github.com/ethereum/go-ethereum/common/math"
|
||||||
"github.com/ethereum/go-ethereum/core/types"
|
"github.com/ethereum/go-ethereum/core/types"
|
||||||
"github.com/ethereum/go-ethereum/core/vm"
|
"github.com/ethereum/go-ethereum/core/vm"
|
||||||
"github.com/ethereum/go-ethereum/crypto"
|
|
||||||
"github.com/ethereum/go-ethereum/params"
|
"github.com/ethereum/go-ethereum/params"
|
||||||
)
|
)
|
||||||
|
|
||||||
var emptyCodeHash = crypto.Keccak256Hash(nil)
|
|
||||||
|
|
||||||
// StateTransition represents a state transition.
|
|
||||||
//
|
|
||||||
// == The State Transitioning Model
|
|
||||||
//
|
|
||||||
// A state transition is a change made when a transaction is applied to the current world
|
|
||||||
// state. The state transitioning model does all the necessary work to work out a valid new
|
|
||||||
// state root.
|
|
||||||
//
|
|
||||||
// 1. Nonce handling
|
|
||||||
// 2. Pre pay gas
|
|
||||||
// 3. Create a new state object if the recipient is nil
|
|
||||||
// 4. Value transfer
|
|
||||||
//
|
|
||||||
// == If contract creation ==
|
|
||||||
//
|
|
||||||
// 4a. Attempt to run transaction data
|
|
||||||
// 4b. If valid, use result as code for the new state object
|
|
||||||
//
|
|
||||||
// == end ==
|
|
||||||
//
|
|
||||||
// 5. Run Script section
|
|
||||||
// 6. Derive new state root
|
|
||||||
type StateTransition struct {
|
|
||||||
gp *GasPool
|
|
||||||
msg Message
|
|
||||||
gas uint64
|
|
||||||
gasPrice *big.Int
|
|
||||||
gasFeeCap *big.Int
|
|
||||||
gasTipCap *big.Int
|
|
||||||
initialGas uint64
|
|
||||||
value *big.Int
|
|
||||||
data []byte
|
|
||||||
state vm.StateDB
|
|
||||||
evm *vm.EVM
|
|
||||||
}
|
|
||||||
|
|
||||||
// Message represents a message sent to a contract.
|
|
||||||
type Message interface {
|
|
||||||
From() common.Address
|
|
||||||
To() *common.Address
|
|
||||||
|
|
||||||
GasPrice() *big.Int
|
|
||||||
GasFeeCap() *big.Int
|
|
||||||
GasTipCap() *big.Int
|
|
||||||
Gas() uint64
|
|
||||||
Value() *big.Int
|
|
||||||
|
|
||||||
Nonce() uint64
|
|
||||||
IsFake() bool
|
|
||||||
Data() []byte
|
|
||||||
AccessList() types.AccessList
|
|
||||||
}
|
|
||||||
|
|
||||||
// ExecutionResult includes all output after executing given evm
|
// ExecutionResult includes all output after executing given evm
|
||||||
// message no matter the execution itself is successful or not.
|
// message no matter the execution itself is successful or not.
|
||||||
type ExecutionResult struct {
|
type ExecutionResult struct {
|
||||||
@ -178,19 +122,47 @@ func toWordSize(size uint64) uint64 {
|
|||||||
return (size + 31) / 32
|
return (size + 31) / 32
|
||||||
}
|
}
|
||||||
|
|
||||||
// NewStateTransition initialises and returns a new state transition object.
|
// A Message contains the data derived from a single transaction that is relevant to state
|
||||||
func NewStateTransition(evm *vm.EVM, msg Message, gp *GasPool) *StateTransition {
|
// processing.
|
||||||
return &StateTransition{
|
type Message struct {
|
||||||
gp: gp,
|
To *common.Address
|
||||||
evm: evm,
|
From common.Address
|
||||||
msg: msg,
|
Nonce uint64
|
||||||
gasPrice: msg.GasPrice(),
|
Value *big.Int
|
||||||
gasFeeCap: msg.GasFeeCap(),
|
GasLimit uint64
|
||||||
gasTipCap: msg.GasTipCap(),
|
GasPrice *big.Int
|
||||||
value: msg.Value(),
|
GasFeeCap *big.Int
|
||||||
data: msg.Data(),
|
GasTipCap *big.Int
|
||||||
state: evm.StateDB,
|
Data []byte
|
||||||
|
AccessList types.AccessList
|
||||||
|
|
||||||
|
// When SkipAccountCheckss is true, the message nonce is not checked against the
|
||||||
|
// account nonce in state. It also disables checking that the sender is an EOA.
|
||||||
|
// This field will be set to true for operations like RPC eth_call.
|
||||||
|
SkipAccountChecks bool
|
||||||
|
}
|
||||||
|
|
||||||
|
// TransactionToMessage converts a transaction into a Message.
|
||||||
|
func TransactionToMessage(tx *types.Transaction, s types.Signer, baseFee *big.Int) (*Message, error) {
|
||||||
|
msg := &Message{
|
||||||
|
Nonce: tx.Nonce(),
|
||||||
|
GasLimit: tx.Gas(),
|
||||||
|
GasPrice: new(big.Int).Set(tx.GasPrice()),
|
||||||
|
GasFeeCap: new(big.Int).Set(tx.GasFeeCap()),
|
||||||
|
GasTipCap: new(big.Int).Set(tx.GasTipCap()),
|
||||||
|
To: tx.To(),
|
||||||
|
Value: tx.Value(),
|
||||||
|
Data: tx.Data(),
|
||||||
|
AccessList: tx.AccessList(),
|
||||||
|
SkipAccountChecks: false,
|
||||||
}
|
}
|
||||||
|
// If baseFee provided, set gasPrice to effectiveGasPrice.
|
||||||
|
if baseFee != nil {
|
||||||
|
msg.GasPrice = cmath.BigMin(msg.GasPrice.Add(msg.GasTipCap, baseFee), msg.GasFeeCap)
|
||||||
|
}
|
||||||
|
var err error
|
||||||
|
msg.From, err = types.Sender(s, tx)
|
||||||
|
return msg, err
|
||||||
}
|
}
|
||||||
|
|
||||||
// ApplyMessage computes the new state by applying the given message
|
// ApplyMessage computes the new state by applying the given message
|
||||||
@ -200,82 +172,126 @@ func NewStateTransition(evm *vm.EVM, msg Message, gp *GasPool) *StateTransition
|
|||||||
// the gas used (which includes gas refunds) and an error if it failed. An error always
|
// the gas used (which includes gas refunds) and an error if it failed. An error always
|
||||||
// indicates a core error meaning that the message would always fail for that particular
|
// indicates a core error meaning that the message would always fail for that particular
|
||||||
// state and would never be accepted within a block.
|
// state and would never be accepted within a block.
|
||||||
func ApplyMessage(evm *vm.EVM, msg Message, gp *GasPool) (*ExecutionResult, error) {
|
func ApplyMessage(evm *vm.EVM, msg *Message, gp *GasPool) (*ExecutionResult, error) {
|
||||||
return NewStateTransition(evm, msg, gp).TransitionDb()
|
return NewStateTransition(evm, msg, gp).TransitionDb()
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// StateTransition represents a state transition.
|
||||||
|
//
|
||||||
|
// == The State Transitioning Model
|
||||||
|
//
|
||||||
|
// A state transition is a change made when a transaction is applied to the current world
|
||||||
|
// state. The state transitioning model does all the necessary work to work out a valid new
|
||||||
|
// state root.
|
||||||
|
//
|
||||||
|
// 1. Nonce handling
|
||||||
|
// 2. Pre pay gas
|
||||||
|
// 3. Create a new state object if the recipient is nil
|
||||||
|
// 4. Value transfer
|
||||||
|
//
|
||||||
|
// == If contract creation ==
|
||||||
|
//
|
||||||
|
// 4a. Attempt to run transaction data
|
||||||
|
// 4b. If valid, use result as code for the new state object
|
||||||
|
//
|
||||||
|
// == end ==
|
||||||
|
//
|
||||||
|
// 5. Run Script section
|
||||||
|
// 6. Derive new state root
|
||||||
|
type StateTransition struct {
|
||||||
|
gp *GasPool
|
||||||
|
msg *Message
|
||||||
|
gasRemaining uint64
|
||||||
|
initialGas uint64
|
||||||
|
state vm.StateDB
|
||||||
|
evm *vm.EVM
|
||||||
|
}
|
||||||
|
|
||||||
|
// NewStateTransition initialises and returns a new state transition object.
|
||||||
|
func NewStateTransition(evm *vm.EVM, msg *Message, gp *GasPool) *StateTransition {
|
||||||
|
return &StateTransition{
|
||||||
|
gp: gp,
|
||||||
|
evm: evm,
|
||||||
|
msg: msg,
|
||||||
|
state: evm.StateDB,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
// to returns the recipient of the message.
|
// to returns the recipient of the message.
|
||||||
func (st *StateTransition) to() common.Address {
|
func (st *StateTransition) to() common.Address {
|
||||||
if st.msg == nil || st.msg.To() == nil /* contract creation */ {
|
if st.msg == nil || st.msg.To == nil /* contract creation */ {
|
||||||
return common.Address{}
|
return common.Address{}
|
||||||
}
|
}
|
||||||
return *st.msg.To()
|
return *st.msg.To
|
||||||
}
|
}
|
||||||
|
|
||||||
func (st *StateTransition) buyGas() error {
|
func (st *StateTransition) buyGas() error {
|
||||||
mgval := new(big.Int).SetUint64(st.msg.Gas())
|
mgval := new(big.Int).SetUint64(st.msg.GasLimit)
|
||||||
mgval = mgval.Mul(mgval, st.gasPrice)
|
mgval = mgval.Mul(mgval, st.msg.GasPrice)
|
||||||
balanceCheck := mgval
|
balanceCheck := mgval
|
||||||
if st.gasFeeCap != nil {
|
if st.msg.GasFeeCap != nil {
|
||||||
balanceCheck = new(big.Int).SetUint64(st.msg.Gas())
|
balanceCheck = new(big.Int).SetUint64(st.msg.GasLimit)
|
||||||
balanceCheck = balanceCheck.Mul(balanceCheck, st.gasFeeCap)
|
balanceCheck = balanceCheck.Mul(balanceCheck, st.msg.GasFeeCap)
|
||||||
balanceCheck.Add(balanceCheck, st.value)
|
balanceCheck.Add(balanceCheck, st.msg.Value)
|
||||||
}
|
}
|
||||||
if have, want := st.state.GetBalance(st.msg.From()), balanceCheck; have.Cmp(want) < 0 {
|
if have, want := st.state.GetBalance(st.msg.From), balanceCheck; have.Cmp(want) < 0 {
|
||||||
return fmt.Errorf("%w: address %v have %v want %v", ErrInsufficientFunds, st.msg.From().Hex(), have, want)
|
return fmt.Errorf("%w: address %v have %v want %v", ErrInsufficientFunds, st.msg.From.Hex(), have, want)
|
||||||
}
|
}
|
||||||
if err := st.gp.SubGas(st.msg.Gas()); err != nil {
|
if err := st.gp.SubGas(st.msg.GasLimit); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
st.gas += st.msg.Gas()
|
st.gasRemaining += st.msg.GasLimit
|
||||||
|
|
||||||
st.initialGas = st.msg.Gas()
|
st.initialGas = st.msg.GasLimit
|
||||||
st.state.SubBalance(st.msg.From(), mgval)
|
st.state.SubBalance(st.msg.From, mgval)
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (st *StateTransition) preCheck() error {
|
func (st *StateTransition) preCheck() error {
|
||||||
// Only check transactions that are not fake
|
// Only check transactions that are not fake
|
||||||
if !st.msg.IsFake() {
|
msg := st.msg
|
||||||
|
if !msg.SkipAccountChecks {
|
||||||
// Make sure this transaction's nonce is correct.
|
// Make sure this transaction's nonce is correct.
|
||||||
stNonce := st.state.GetNonce(st.msg.From())
|
stNonce := st.state.GetNonce(msg.From)
|
||||||
if msgNonce := st.msg.Nonce(); stNonce < msgNonce {
|
if msgNonce := msg.Nonce; stNonce < msgNonce {
|
||||||
return fmt.Errorf("%w: address %v, tx: %d state: %d", ErrNonceTooHigh,
|
return fmt.Errorf("%w: address %v, tx: %d state: %d", ErrNonceTooHigh,
|
||||||
st.msg.From().Hex(), msgNonce, stNonce)
|
msg.From.Hex(), msgNonce, stNonce)
|
||||||
} else if stNonce > msgNonce {
|
} else if stNonce > msgNonce {
|
||||||
return fmt.Errorf("%w: address %v, tx: %d state: %d", ErrNonceTooLow,
|
return fmt.Errorf("%w: address %v, tx: %d state: %d", ErrNonceTooLow,
|
||||||
st.msg.From().Hex(), msgNonce, stNonce)
|
msg.From.Hex(), msgNonce, stNonce)
|
||||||
} else if stNonce+1 < stNonce {
|
} else if stNonce+1 < stNonce {
|
||||||
return fmt.Errorf("%w: address %v, nonce: %d", ErrNonceMax,
|
return fmt.Errorf("%w: address %v, nonce: %d", ErrNonceMax,
|
||||||
st.msg.From().Hex(), stNonce)
|
msg.From.Hex(), stNonce)
|
||||||
}
|
}
|
||||||
// Make sure the sender is an EOA
|
// Make sure the sender is an EOA
|
||||||
if codeHash := st.state.GetCodeHash(st.msg.From()); codeHash != emptyCodeHash && codeHash != (common.Hash{}) {
|
codeHash := st.state.GetCodeHash(msg.From)
|
||||||
|
if codeHash != (common.Hash{}) && codeHash != types.EmptyCodeHash {
|
||||||
return fmt.Errorf("%w: address %v, codehash: %s", ErrSenderNoEOA,
|
return fmt.Errorf("%w: address %v, codehash: %s", ErrSenderNoEOA,
|
||||||
st.msg.From().Hex(), codeHash)
|
msg.From.Hex(), codeHash)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Make sure that transaction gasFeeCap is greater than the baseFee (post london)
|
// Make sure that transaction gasFeeCap is greater than the baseFee (post london)
|
||||||
if st.evm.ChainConfig().IsLondon(st.evm.Context.BlockNumber) {
|
if st.evm.ChainConfig().IsLondon(st.evm.Context.BlockNumber) {
|
||||||
// Skip the checks if gas fields are zero and baseFee was explicitly disabled (eth_call)
|
// Skip the checks if gas fields are zero and baseFee was explicitly disabled (eth_call)
|
||||||
if !st.evm.Config.NoBaseFee || st.gasFeeCap.BitLen() > 0 || st.gasTipCap.BitLen() > 0 {
|
if !st.evm.Config.NoBaseFee || msg.GasFeeCap.BitLen() > 0 || msg.GasTipCap.BitLen() > 0 {
|
||||||
if l := st.gasFeeCap.BitLen(); l > 256 {
|
if l := msg.GasFeeCap.BitLen(); l > 256 {
|
||||||
return fmt.Errorf("%w: address %v, maxFeePerGas bit length: %d", ErrFeeCapVeryHigh,
|
return fmt.Errorf("%w: address %v, maxFeePerGas bit length: %d", ErrFeeCapVeryHigh,
|
||||||
st.msg.From().Hex(), l)
|
msg.From.Hex(), l)
|
||||||
}
|
}
|
||||||
if l := st.gasTipCap.BitLen(); l > 256 {
|
if l := msg.GasTipCap.BitLen(); l > 256 {
|
||||||
return fmt.Errorf("%w: address %v, maxPriorityFeePerGas bit length: %d", ErrTipVeryHigh,
|
return fmt.Errorf("%w: address %v, maxPriorityFeePerGas bit length: %d", ErrTipVeryHigh,
|
||||||
st.msg.From().Hex(), l)
|
msg.From.Hex(), l)
|
||||||
}
|
}
|
||||||
if st.gasFeeCap.Cmp(st.gasTipCap) < 0 {
|
if msg.GasFeeCap.Cmp(msg.GasTipCap) < 0 {
|
||||||
return fmt.Errorf("%w: address %v, maxPriorityFeePerGas: %s, maxFeePerGas: %s", ErrTipAboveFeeCap,
|
return fmt.Errorf("%w: address %v, maxPriorityFeePerGas: %s, maxFeePerGas: %s", ErrTipAboveFeeCap,
|
||||||
st.msg.From().Hex(), st.gasTipCap, st.gasFeeCap)
|
msg.From.Hex(), msg.GasTipCap, msg.GasFeeCap)
|
||||||
}
|
}
|
||||||
// This will panic if baseFee is nil, but basefee presence is verified
|
// This will panic if baseFee is nil, but basefee presence is verified
|
||||||
// as part of header validation.
|
// as part of header validation.
|
||||||
if st.gasFeeCap.Cmp(st.evm.Context.BaseFee) < 0 {
|
if msg.GasFeeCap.Cmp(st.evm.Context.BaseFee) < 0 {
|
||||||
return fmt.Errorf("%w: address %v, maxFeePerGas: %s baseFee: %s", ErrFeeCapTooLow,
|
return fmt.Errorf("%w: address %v, maxFeePerGas: %s baseFee: %s", ErrFeeCapTooLow,
|
||||||
st.msg.From().Hex(), st.gasFeeCap, st.evm.Context.BaseFee)
|
msg.From.Hex(), msg.GasFeeCap, st.evm.Context.BaseFee)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -304,9 +320,14 @@ func (st *StateTransition) TransitionDb() (*ExecutionResult, error) {
|
|||||||
// 6. caller has enough balance to cover asset transfer for **topmost** call
|
// 6. caller has enough balance to cover asset transfer for **topmost** call
|
||||||
|
|
||||||
// Check clauses 1-3, buy gas if everything is correct
|
// Check clauses 1-3, buy gas if everything is correct
|
||||||
|
|
||||||
|
// begin PluGeth injection
|
||||||
if v, ok := st.evm.Config.Tracer.(PreTracer); ok {
|
if v, ok := st.evm.Config.Tracer.(PreTracer); ok {
|
||||||
v.CapturePreStart(st.msg.From(), st.msg.To(), st.data, st.gas, st.msg.Value())
|
v.CapturePreStart(st.msg.From, st.msg.To, st.msg.Data, st.msg.GasLimit, st.msg.Value)
|
||||||
}
|
}
|
||||||
|
// end PluGeth injection
|
||||||
|
|
||||||
|
|
||||||
if err := st.preCheck(); err != nil {
|
if err := st.preCheck(); err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
@ -314,52 +335,52 @@ func (st *StateTransition) TransitionDb() (*ExecutionResult, error) {
|
|||||||
if st.evm.Config.Debug {
|
if st.evm.Config.Debug {
|
||||||
st.evm.Config.Tracer.CaptureTxStart(st.initialGas)
|
st.evm.Config.Tracer.CaptureTxStart(st.initialGas)
|
||||||
defer func() {
|
defer func() {
|
||||||
st.evm.Config.Tracer.CaptureTxEnd(st.gas)
|
st.evm.Config.Tracer.CaptureTxEnd(st.gasRemaining)
|
||||||
}()
|
}()
|
||||||
}
|
}
|
||||||
|
|
||||||
var (
|
var (
|
||||||
msg = st.msg
|
msg = st.msg
|
||||||
sender = vm.AccountRef(msg.From())
|
sender = vm.AccountRef(msg.From)
|
||||||
rules = st.evm.ChainConfig().Rules(st.evm.Context.BlockNumber, st.evm.Context.Random != nil, st.evm.Context.Time)
|
rules = st.evm.ChainConfig().Rules(st.evm.Context.BlockNumber, st.evm.Context.Random != nil, st.evm.Context.Time)
|
||||||
contractCreation = msg.To() == nil
|
contractCreation = msg.To == nil
|
||||||
)
|
)
|
||||||
|
|
||||||
// Check clauses 4-5, subtract intrinsic gas if everything is correct
|
// Check clauses 4-5, subtract intrinsic gas if everything is correct
|
||||||
gas, err := IntrinsicGas(st.data, st.msg.AccessList(), contractCreation, rules.IsHomestead, rules.IsIstanbul, rules.IsShanghai)
|
gas, err := IntrinsicGas(msg.Data, msg.AccessList, contractCreation, rules.IsHomestead, rules.IsIstanbul, rules.IsShanghai)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
if st.gas < gas {
|
if st.gasRemaining < gas {
|
||||||
return nil, fmt.Errorf("%w: have %d, want %d", ErrIntrinsicGas, st.gas, gas)
|
return nil, fmt.Errorf("%w: have %d, want %d", ErrIntrinsicGas, st.gasRemaining, gas)
|
||||||
}
|
}
|
||||||
st.gas -= gas
|
st.gasRemaining -= gas
|
||||||
|
|
||||||
// Check clause 6
|
// Check clause 6
|
||||||
if msg.Value().Sign() > 0 && !st.evm.Context.CanTransfer(st.state, msg.From(), msg.Value()) {
|
if msg.Value.Sign() > 0 && !st.evm.Context.CanTransfer(st.state, msg.From, msg.Value) {
|
||||||
return nil, fmt.Errorf("%w: address %v", ErrInsufficientFundsForTransfer, msg.From().Hex())
|
return nil, fmt.Errorf("%w: address %v", ErrInsufficientFundsForTransfer, msg.From.Hex())
|
||||||
}
|
}
|
||||||
|
|
||||||
// Check whether the init code size has been exceeded.
|
// Check whether the init code size has been exceeded.
|
||||||
if rules.IsShanghai && contractCreation && len(st.data) > params.MaxInitCodeSize {
|
if rules.IsShanghai && contractCreation && len(msg.Data) > params.MaxInitCodeSize {
|
||||||
return nil, fmt.Errorf("%w: code size %v limit %v", ErrMaxInitCodeSizeExceeded, len(st.data), params.MaxInitCodeSize)
|
return nil, fmt.Errorf("%w: code size %v limit %v", ErrMaxInitCodeSizeExceeded, len(msg.Data), params.MaxInitCodeSize)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Execute the preparatory steps for state transition which includes:
|
// Execute the preparatory steps for state transition which includes:
|
||||||
// - prepare accessList(post-berlin)
|
// - prepare accessList(post-berlin)
|
||||||
// - reset transient storage(eip 1153)
|
// - reset transient storage(eip 1153)
|
||||||
st.state.Prepare(rules, msg.From(), st.evm.Context.Coinbase, msg.To(), vm.ActivePrecompiles(rules), msg.AccessList())
|
st.state.Prepare(rules, msg.From, st.evm.Context.Coinbase, msg.To, vm.ActivePrecompiles(rules), msg.AccessList)
|
||||||
|
|
||||||
var (
|
var (
|
||||||
ret []byte
|
ret []byte
|
||||||
vmerr error // vm errors do not effect consensus and are therefore not assigned to err
|
vmerr error // vm errors do not effect consensus and are therefore not assigned to err
|
||||||
)
|
)
|
||||||
if contractCreation {
|
if contractCreation {
|
||||||
ret, _, st.gas, vmerr = st.evm.Create(sender, st.data, st.gas, st.value)
|
ret, _, st.gasRemaining, vmerr = st.evm.Create(sender, msg.Data, st.gasRemaining, msg.Value)
|
||||||
} else {
|
} else {
|
||||||
// Increment the nonce for the next transaction
|
// Increment the nonce for the next transaction
|
||||||
st.state.SetNonce(msg.From(), st.state.GetNonce(sender.Address())+1)
|
st.state.SetNonce(msg.From, st.state.GetNonce(sender.Address())+1)
|
||||||
ret, st.gas, vmerr = st.evm.Call(sender, st.to(), st.data, st.gas, st.value)
|
ret, st.gasRemaining, vmerr = st.evm.Call(sender, st.to(), msg.Data, st.gasRemaining, msg.Value)
|
||||||
}
|
}
|
||||||
|
|
||||||
if !rules.IsLondon {
|
if !rules.IsLondon {
|
||||||
@ -369,12 +390,12 @@ func (st *StateTransition) TransitionDb() (*ExecutionResult, error) {
|
|||||||
// After EIP-3529: refunds are capped to gasUsed / 5
|
// After EIP-3529: refunds are capped to gasUsed / 5
|
||||||
st.refundGas(params.RefundQuotientEIP3529)
|
st.refundGas(params.RefundQuotientEIP3529)
|
||||||
}
|
}
|
||||||
effectiveTip := st.gasPrice
|
effectiveTip := msg.GasPrice
|
||||||
if rules.IsLondon {
|
if rules.IsLondon {
|
||||||
effectiveTip = cmath.BigMin(st.gasTipCap, new(big.Int).Sub(st.gasFeeCap, st.evm.Context.BaseFee))
|
effectiveTip = cmath.BigMin(msg.GasTipCap, new(big.Int).Sub(msg.GasFeeCap, st.evm.Context.BaseFee))
|
||||||
}
|
}
|
||||||
|
|
||||||
if st.evm.Config.NoBaseFee && st.gasFeeCap.Sign() == 0 && st.gasTipCap.Sign() == 0 {
|
if st.evm.Config.NoBaseFee && msg.GasFeeCap.Sign() == 0 && msg.GasTipCap.Sign() == 0 {
|
||||||
// Skip fee payment when NoBaseFee is set and the fee fields
|
// Skip fee payment when NoBaseFee is set and the fee fields
|
||||||
// are 0. This avoids a negative effectiveTip being applied to
|
// are 0. This avoids a negative effectiveTip being applied to
|
||||||
// the coinbase when simulating calls.
|
// the coinbase when simulating calls.
|
||||||
@ -397,18 +418,18 @@ func (st *StateTransition) refundGas(refundQuotient uint64) {
|
|||||||
if refund > st.state.GetRefund() {
|
if refund > st.state.GetRefund() {
|
||||||
refund = st.state.GetRefund()
|
refund = st.state.GetRefund()
|
||||||
}
|
}
|
||||||
st.gas += refund
|
st.gasRemaining += refund
|
||||||
|
|
||||||
// Return ETH for remaining gas, exchanged at the original rate.
|
// Return ETH for remaining gas, exchanged at the original rate.
|
||||||
remaining := new(big.Int).Mul(new(big.Int).SetUint64(st.gas), st.gasPrice)
|
remaining := new(big.Int).Mul(new(big.Int).SetUint64(st.gasRemaining), st.msg.GasPrice)
|
||||||
st.state.AddBalance(st.msg.From(), remaining)
|
st.state.AddBalance(st.msg.From, remaining)
|
||||||
|
|
||||||
// Also return remaining gas to the block gas counter so it is
|
// Also return remaining gas to the block gas counter so it is
|
||||||
// available for the next transaction.
|
// available for the next transaction.
|
||||||
st.gp.AddGas(st.gas)
|
st.gp.AddGas(st.gasRemaining)
|
||||||
}
|
}
|
||||||
|
|
||||||
// gasUsed returns the amount of gas used up by the state transition.
|
// gasUsed returns the amount of gas used up by the state transition.
|
||||||
func (st *StateTransition) gasUsed() uint64 {
|
func (st *StateTransition) gasUsed() uint64 {
|
||||||
return st.initialGas - st.gas
|
return st.initialGas - st.gasRemaining
|
||||||
}
|
}
|
||||||
|
@ -254,17 +254,19 @@ type list struct {
|
|||||||
strict bool // Whether nonces are strictly continuous or not
|
strict bool // Whether nonces are strictly continuous or not
|
||||||
txs *sortedMap // Heap indexed sorted hash map of the transactions
|
txs *sortedMap // Heap indexed sorted hash map of the transactions
|
||||||
|
|
||||||
costcap *big.Int // Price of the highest costing transaction (reset only if exceeds balance)
|
costcap *big.Int // Price of the highest costing transaction (reset only if exceeds balance)
|
||||||
gascap uint64 // Gas limit of the highest spending transaction (reset only if exceeds block limit)
|
gascap uint64 // Gas limit of the highest spending transaction (reset only if exceeds block limit)
|
||||||
|
totalcost *big.Int // Total cost of all transactions in the list
|
||||||
}
|
}
|
||||||
|
|
||||||
// newList create a new transaction list for maintaining nonce-indexable fast,
|
// newList create a new transaction list for maintaining nonce-indexable fast,
|
||||||
// gapped, sortable transaction lists.
|
// gapped, sortable transaction lists.
|
||||||
func newList(strict bool) *list {
|
func newList(strict bool) *list {
|
||||||
return &list{
|
return &list{
|
||||||
strict: strict,
|
strict: strict,
|
||||||
txs: newSortedMap(),
|
txs: newSortedMap(),
|
||||||
costcap: new(big.Int),
|
costcap: new(big.Int),
|
||||||
|
totalcost: new(big.Int),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -302,7 +304,11 @@ func (l *list) Add(tx *types.Transaction, priceBump uint64) (bool, *types.Transa
|
|||||||
if tx.GasFeeCapIntCmp(thresholdFeeCap) < 0 || tx.GasTipCapIntCmp(thresholdTip) < 0 {
|
if tx.GasFeeCapIntCmp(thresholdFeeCap) < 0 || tx.GasTipCapIntCmp(thresholdTip) < 0 {
|
||||||
return false, nil
|
return false, nil
|
||||||
}
|
}
|
||||||
|
// Old is being replaced, subtract old cost
|
||||||
|
l.subTotalCost([]*types.Transaction{old})
|
||||||
}
|
}
|
||||||
|
// Add new tx cost to totalcost
|
||||||
|
l.totalcost.Add(l.totalcost, tx.Cost())
|
||||||
// Otherwise overwrite the old transaction with the current one
|
// Otherwise overwrite the old transaction with the current one
|
||||||
l.txs.Put(tx)
|
l.txs.Put(tx)
|
||||||
if cost := tx.Cost(); l.costcap.Cmp(cost) < 0 {
|
if cost := tx.Cost(); l.costcap.Cmp(cost) < 0 {
|
||||||
@ -318,7 +324,9 @@ func (l *list) Add(tx *types.Transaction, priceBump uint64) (bool, *types.Transa
|
|||||||
// provided threshold. Every removed transaction is returned for any post-removal
|
// provided threshold. Every removed transaction is returned for any post-removal
|
||||||
// maintenance.
|
// maintenance.
|
||||||
func (l *list) Forward(threshold uint64) types.Transactions {
|
func (l *list) Forward(threshold uint64) types.Transactions {
|
||||||
return l.txs.Forward(threshold)
|
txs := l.txs.Forward(threshold)
|
||||||
|
l.subTotalCost(txs)
|
||||||
|
return txs
|
||||||
}
|
}
|
||||||
|
|
||||||
// Filter removes all transactions from the list with a cost or gas limit higher
|
// Filter removes all transactions from the list with a cost or gas limit higher
|
||||||
@ -357,6 +365,9 @@ func (l *list) Filter(costLimit *big.Int, gasLimit uint64) (types.Transactions,
|
|||||||
}
|
}
|
||||||
invalids = l.txs.filter(func(tx *types.Transaction) bool { return tx.Nonce() > lowest })
|
invalids = l.txs.filter(func(tx *types.Transaction) bool { return tx.Nonce() > lowest })
|
||||||
}
|
}
|
||||||
|
// Reset total cost
|
||||||
|
l.subTotalCost(removed)
|
||||||
|
l.subTotalCost(invalids)
|
||||||
l.txs.reheap()
|
l.txs.reheap()
|
||||||
return removed, invalids
|
return removed, invalids
|
||||||
}
|
}
|
||||||
@ -364,7 +375,9 @@ func (l *list) Filter(costLimit *big.Int, gasLimit uint64) (types.Transactions,
|
|||||||
// Cap places a hard limit on the number of items, returning all transactions
|
// Cap places a hard limit on the number of items, returning all transactions
|
||||||
// exceeding that limit.
|
// exceeding that limit.
|
||||||
func (l *list) Cap(threshold int) types.Transactions {
|
func (l *list) Cap(threshold int) types.Transactions {
|
||||||
return l.txs.Cap(threshold)
|
txs := l.txs.Cap(threshold)
|
||||||
|
l.subTotalCost(txs)
|
||||||
|
return txs
|
||||||
}
|
}
|
||||||
|
|
||||||
// Remove deletes a transaction from the maintained list, returning whether the
|
// Remove deletes a transaction from the maintained list, returning whether the
|
||||||
@ -376,9 +389,12 @@ func (l *list) Remove(tx *types.Transaction) (bool, types.Transactions) {
|
|||||||
if removed := l.txs.Remove(nonce); !removed {
|
if removed := l.txs.Remove(nonce); !removed {
|
||||||
return false, nil
|
return false, nil
|
||||||
}
|
}
|
||||||
|
l.subTotalCost([]*types.Transaction{tx})
|
||||||
// In strict mode, filter out non-executable transactions
|
// In strict mode, filter out non-executable transactions
|
||||||
if l.strict {
|
if l.strict {
|
||||||
return true, l.txs.Filter(func(tx *types.Transaction) bool { return tx.Nonce() > nonce })
|
txs := l.txs.Filter(func(tx *types.Transaction) bool { return tx.Nonce() > nonce })
|
||||||
|
l.subTotalCost(txs)
|
||||||
|
return true, txs
|
||||||
}
|
}
|
||||||
return true, nil
|
return true, nil
|
||||||
}
|
}
|
||||||
@ -391,7 +407,9 @@ func (l *list) Remove(tx *types.Transaction) (bool, types.Transactions) {
|
|||||||
// prevent getting into and invalid state. This is not something that should ever
|
// prevent getting into and invalid state. This is not something that should ever
|
||||||
// happen but better to be self correcting than failing!
|
// happen but better to be self correcting than failing!
|
||||||
func (l *list) Ready(start uint64) types.Transactions {
|
func (l *list) Ready(start uint64) types.Transactions {
|
||||||
return l.txs.Ready(start)
|
txs := l.txs.Ready(start)
|
||||||
|
l.subTotalCost(txs)
|
||||||
|
return txs
|
||||||
}
|
}
|
||||||
|
|
||||||
// Len returns the length of the transaction list.
|
// Len returns the length of the transaction list.
|
||||||
@ -417,6 +435,14 @@ func (l *list) LastElement() *types.Transaction {
|
|||||||
return l.txs.LastElement()
|
return l.txs.LastElement()
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// subTotalCost subtracts the cost of the given transactions from the
|
||||||
|
// total cost of all transactions.
|
||||||
|
func (l *list) subTotalCost(txs []*types.Transaction) {
|
||||||
|
for _, tx := range txs {
|
||||||
|
l.totalcost.Sub(l.totalcost, tx.Cost())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
// priceHeap is a heap.Interface implementation over transactions for retrieving
|
// priceHeap is a heap.Interface implementation over transactions for retrieving
|
||||||
// price-sorted transactions to discard when the pool fills up. If baseFee is set
|
// price-sorted transactions to discard when the pool fills up. If baseFee is set
|
||||||
// then the heap is sorted based on the effective tip based on the given base fee.
|
// then the heap is sorted based on the effective tip based on the given base fee.
|
||||||
@ -482,10 +508,7 @@ func (h *priceHeap) Pop() interface{} {
|
|||||||
// the floating heap is better. When baseFee is decreasing they behave similarly.
|
// the floating heap is better. When baseFee is decreasing they behave similarly.
|
||||||
type pricedList struct {
|
type pricedList struct {
|
||||||
// Number of stale price points to (re-heap trigger).
|
// Number of stale price points to (re-heap trigger).
|
||||||
// This field is accessed atomically, and must be the first field
|
stales atomic.Int64
|
||||||
// to ensure it has correct alignment for atomic.AddInt64.
|
|
||||||
// See https://golang.org/pkg/sync/atomic/#pkg-note-BUG.
|
|
||||||
stales int64
|
|
||||||
|
|
||||||
all *lookup // Pointer to the map of all transactions
|
all *lookup // Pointer to the map of all transactions
|
||||||
urgent, floating priceHeap // Heaps of prices of all the stored **remote** transactions
|
urgent, floating priceHeap // Heaps of prices of all the stored **remote** transactions
|
||||||
@ -519,7 +542,7 @@ func (l *pricedList) Put(tx *types.Transaction, local bool) {
|
|||||||
// the heap if a large enough ratio of transactions go stale.
|
// the heap if a large enough ratio of transactions go stale.
|
||||||
func (l *pricedList) Removed(count int) {
|
func (l *pricedList) Removed(count int) {
|
||||||
// Bump the stale counter, but exit if still too low (< 25%)
|
// Bump the stale counter, but exit if still too low (< 25%)
|
||||||
stales := atomic.AddInt64(&l.stales, int64(count))
|
stales := l.stales.Add(int64(count))
|
||||||
if int(stales) <= (len(l.urgent.list)+len(l.floating.list))/4 {
|
if int(stales) <= (len(l.urgent.list)+len(l.floating.list))/4 {
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
@ -544,7 +567,7 @@ func (l *pricedList) underpricedFor(h *priceHeap, tx *types.Transaction) bool {
|
|||||||
for len(h.list) > 0 {
|
for len(h.list) > 0 {
|
||||||
head := h.list[0]
|
head := h.list[0]
|
||||||
if l.all.GetRemote(head.Hash()) == nil { // Removed or migrated
|
if l.all.GetRemote(head.Hash()) == nil { // Removed or migrated
|
||||||
atomic.AddInt64(&l.stales, -1)
|
l.stales.Add(-1)
|
||||||
heap.Pop(h)
|
heap.Pop(h)
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
@ -561,6 +584,7 @@ func (l *pricedList) underpricedFor(h *priceHeap, tx *types.Transaction) bool {
|
|||||||
|
|
||||||
// Discard finds a number of most underpriced transactions, removes them from the
|
// Discard finds a number of most underpriced transactions, removes them from the
|
||||||
// priced list and returns them for further removal from the entire pool.
|
// priced list and returns them for further removal from the entire pool.
|
||||||
|
// If noPending is set to true, we will only consider the floating list
|
||||||
//
|
//
|
||||||
// Note local transaction won't be considered for eviction.
|
// Note local transaction won't be considered for eviction.
|
||||||
func (l *pricedList) Discard(slots int, force bool) (types.Transactions, bool) {
|
func (l *pricedList) Discard(slots int, force bool) (types.Transactions, bool) {
|
||||||
@ -570,7 +594,7 @@ func (l *pricedList) Discard(slots int, force bool) (types.Transactions, bool) {
|
|||||||
// Discard stale transactions if found during cleanup
|
// Discard stale transactions if found during cleanup
|
||||||
tx := heap.Pop(&l.urgent).(*types.Transaction)
|
tx := heap.Pop(&l.urgent).(*types.Transaction)
|
||||||
if l.all.GetRemote(tx.Hash()) == nil { // Removed or migrated
|
if l.all.GetRemote(tx.Hash()) == nil { // Removed or migrated
|
||||||
atomic.AddInt64(&l.stales, -1)
|
l.stales.Add(-1)
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
// Non stale transaction found, move to floating heap
|
// Non stale transaction found, move to floating heap
|
||||||
@ -583,7 +607,7 @@ func (l *pricedList) Discard(slots int, force bool) (types.Transactions, bool) {
|
|||||||
// Discard stale transactions if found during cleanup
|
// Discard stale transactions if found during cleanup
|
||||||
tx := heap.Pop(&l.floating).(*types.Transaction)
|
tx := heap.Pop(&l.floating).(*types.Transaction)
|
||||||
if l.all.GetRemote(tx.Hash()) == nil { // Removed or migrated
|
if l.all.GetRemote(tx.Hash()) == nil { // Removed or migrated
|
||||||
atomic.AddInt64(&l.stales, -1)
|
l.stales.Add(-1)
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
// Non stale transaction found, discard it
|
// Non stale transaction found, discard it
|
||||||
@ -606,7 +630,7 @@ func (l *pricedList) Reheap() {
|
|||||||
l.reheapMu.Lock()
|
l.reheapMu.Lock()
|
||||||
defer l.reheapMu.Unlock()
|
defer l.reheapMu.Unlock()
|
||||||
start := time.Now()
|
start := time.Now()
|
||||||
atomic.StoreInt64(&l.stales, 0)
|
l.stales.Store(0)
|
||||||
l.urgent.list = make([]*types.Transaction, 0, l.all.RemoteCount())
|
l.urgent.list = make([]*types.Transaction, 0, l.all.RemoteCount())
|
||||||
l.all.Range(func(hash common.Hash, tx *types.Transaction, local bool) bool {
|
l.all.Range(func(hash common.Hash, tx *types.Transaction, local bool) bool {
|
||||||
l.urgent.list = append(l.urgent.list, tx)
|
l.urgent.list = append(l.urgent.list, tx)
|
||||||
|
@ -23,7 +23,6 @@ import (
|
|||||||
"math/big"
|
"math/big"
|
||||||
"sort"
|
"sort"
|
||||||
"sync"
|
"sync"
|
||||||
"sync/atomic"
|
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/ethereum/go-ethereum/common"
|
"github.com/ethereum/go-ethereum/common"
|
||||||
@ -87,6 +86,14 @@ var (
|
|||||||
// than some meaningful limit a user might use. This is not a consensus error
|
// than some meaningful limit a user might use. This is not a consensus error
|
||||||
// making the transaction invalid, rather a DOS protection.
|
// making the transaction invalid, rather a DOS protection.
|
||||||
ErrOversizedData = errors.New("oversized data")
|
ErrOversizedData = errors.New("oversized data")
|
||||||
|
|
||||||
|
// ErrFutureReplacePending is returned if a future transaction replaces a pending
|
||||||
|
// transaction. Future transactions should only be able to replace other future transactions.
|
||||||
|
ErrFutureReplacePending = errors.New("future transaction tries to replace pending")
|
||||||
|
|
||||||
|
// ErrOverdraft is returned if a transaction would cause the senders balance to go negative
|
||||||
|
// thus invalidating a potential large number of transactions.
|
||||||
|
ErrOverdraft = errors.New("transaction would cause overdraft")
|
||||||
)
|
)
|
||||||
|
|
||||||
var (
|
var (
|
||||||
@ -375,7 +382,7 @@ func (pool *TxPool) loop() {
|
|||||||
pool.mu.RLock()
|
pool.mu.RLock()
|
||||||
pending, queued := pool.stats()
|
pending, queued := pool.stats()
|
||||||
pool.mu.RUnlock()
|
pool.mu.RUnlock()
|
||||||
stales := int(atomic.LoadInt64(&pool.priced.stales))
|
stales := int(pool.priced.stales.Load())
|
||||||
|
|
||||||
if pending != prevPending || queued != prevQueued || stales != prevStales {
|
if pending != prevPending || queued != prevQueued || stales != prevStales {
|
||||||
log.Debug("Transaction pool status report", "executable", pending, "queued", queued, "stales", stales)
|
log.Debug("Transaction pool status report", "executable", pending, "queued", queued, "stales", stales)
|
||||||
@ -639,9 +646,25 @@ func (pool *TxPool) validateTx(tx *types.Transaction, local bool) error {
|
|||||||
}
|
}
|
||||||
// Transactor should have enough funds to cover the costs
|
// Transactor should have enough funds to cover the costs
|
||||||
// cost == V + GP * GL
|
// cost == V + GP * GL
|
||||||
if pool.currentState.GetBalance(from).Cmp(tx.Cost()) < 0 {
|
balance := pool.currentState.GetBalance(from)
|
||||||
|
if balance.Cmp(tx.Cost()) < 0 {
|
||||||
return core.ErrInsufficientFunds
|
return core.ErrInsufficientFunds
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Verify that replacing transactions will not result in overdraft
|
||||||
|
list := pool.pending[from]
|
||||||
|
if list != nil { // Sender already has pending txs
|
||||||
|
sum := new(big.Int).Add(tx.Cost(), list.totalcost)
|
||||||
|
if repl := list.txs.Get(tx.Nonce()); repl != nil {
|
||||||
|
// Deduct the cost of a transaction replaced by this
|
||||||
|
sum.Sub(sum, repl.Cost())
|
||||||
|
}
|
||||||
|
if balance.Cmp(sum) < 0 {
|
||||||
|
log.Trace("Replacing transactions would overdraft", "sender", from, "balance", pool.currentState.GetBalance(from), "required", sum)
|
||||||
|
return ErrOverdraft
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
// Ensure the transaction has more gas than the basic tx fee.
|
// Ensure the transaction has more gas than the basic tx fee.
|
||||||
intrGas, err := core.IntrinsicGas(tx.Data(), tx.AccessList(), tx.To() == nil, true, pool.istanbul, pool.shanghai)
|
intrGas, err := core.IntrinsicGas(tx.Data(), tx.AccessList(), tx.To() == nil, true, pool.istanbul, pool.shanghai)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@ -678,6 +701,10 @@ func (pool *TxPool) add(tx *types.Transaction, local bool) (replaced bool, err e
|
|||||||
invalidTxMeter.Mark(1)
|
invalidTxMeter.Mark(1)
|
||||||
return false, err
|
return false, err
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// already validated by this point
|
||||||
|
from, _ := types.Sender(pool.signer, tx)
|
||||||
|
|
||||||
// If the transaction pool is full, discard underpriced transactions
|
// If the transaction pool is full, discard underpriced transactions
|
||||||
if uint64(pool.all.Slots()+numSlots(tx)) > pool.config.GlobalSlots+pool.config.GlobalQueue {
|
if uint64(pool.all.Slots()+numSlots(tx)) > pool.config.GlobalSlots+pool.config.GlobalQueue {
|
||||||
// If the new transaction is underpriced, don't accept it
|
// If the new transaction is underpriced, don't accept it
|
||||||
@ -686,6 +713,7 @@ func (pool *TxPool) add(tx *types.Transaction, local bool) (replaced bool, err e
|
|||||||
underpricedTxMeter.Mark(1)
|
underpricedTxMeter.Mark(1)
|
||||||
return false, ErrUnderpriced
|
return false, ErrUnderpriced
|
||||||
}
|
}
|
||||||
|
|
||||||
// We're about to replace a transaction. The reorg does a more thorough
|
// We're about to replace a transaction. The reorg does a more thorough
|
||||||
// analysis of what to remove and how, but it runs async. We don't want to
|
// analysis of what to remove and how, but it runs async. We don't want to
|
||||||
// do too many replacements between reorg-runs, so we cap the number of
|
// do too many replacements between reorg-runs, so we cap the number of
|
||||||
@ -706,17 +734,37 @@ func (pool *TxPool) add(tx *types.Transaction, local bool) (replaced bool, err e
|
|||||||
overflowedTxMeter.Mark(1)
|
overflowedTxMeter.Mark(1)
|
||||||
return false, ErrTxPoolOverflow
|
return false, ErrTxPoolOverflow
|
||||||
}
|
}
|
||||||
// Bump the counter of rejections-since-reorg
|
|
||||||
pool.changesSinceReorg += len(drop)
|
// If the new transaction is a future transaction it should never churn pending transactions
|
||||||
|
if !isLocal && pool.isFuture(from, tx) {
|
||||||
|
var replacesPending bool
|
||||||
|
for _, dropTx := range drop {
|
||||||
|
dropSender, _ := types.Sender(pool.signer, dropTx)
|
||||||
|
if list := pool.pending[dropSender]; list != nil && list.Overlaps(dropTx) {
|
||||||
|
replacesPending = true
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
// Add all transactions back to the priced queue
|
||||||
|
if replacesPending {
|
||||||
|
for _, dropTx := range drop {
|
||||||
|
pool.priced.Put(dropTx, false)
|
||||||
|
}
|
||||||
|
log.Trace("Discarding future transaction replacing pending tx", "hash", hash)
|
||||||
|
return false, ErrFutureReplacePending
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
// Kick out the underpriced remote transactions.
|
// Kick out the underpriced remote transactions.
|
||||||
for _, tx := range drop {
|
for _, tx := range drop {
|
||||||
log.Trace("Discarding freshly underpriced transaction", "hash", tx.Hash(), "gasTipCap", tx.GasTipCap(), "gasFeeCap", tx.GasFeeCap())
|
log.Trace("Discarding freshly underpriced transaction", "hash", tx.Hash(), "gasTipCap", tx.GasTipCap(), "gasFeeCap", tx.GasFeeCap())
|
||||||
underpricedTxMeter.Mark(1)
|
underpricedTxMeter.Mark(1)
|
||||||
pool.removeTx(tx.Hash(), false)
|
dropped := pool.removeTx(tx.Hash(), false)
|
||||||
|
pool.changesSinceReorg += dropped
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Try to replace an existing transaction in the pending pool
|
// Try to replace an existing transaction in the pending pool
|
||||||
from, _ := types.Sender(pool.signer, tx) // already validated
|
|
||||||
if list := pool.pending[from]; list != nil && list.Overlaps(tx) {
|
if list := pool.pending[from]; list != nil && list.Overlaps(tx) {
|
||||||
// Nonce already pending, check if required price bump is met
|
// Nonce already pending, check if required price bump is met
|
||||||
inserted, old := list.Add(tx, pool.config.PriceBump)
|
inserted, old := list.Add(tx, pool.config.PriceBump)
|
||||||
@ -760,6 +808,20 @@ func (pool *TxPool) add(tx *types.Transaction, local bool) (replaced bool, err e
|
|||||||
return replaced, nil
|
return replaced, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// isFuture reports whether the given transaction is immediately executable.
|
||||||
|
func (pool *TxPool) isFuture(from common.Address, tx *types.Transaction) bool {
|
||||||
|
list := pool.pending[from]
|
||||||
|
if list == nil {
|
||||||
|
return pool.pendingNonces.get(from) != tx.Nonce()
|
||||||
|
}
|
||||||
|
// Sender has pending transactions.
|
||||||
|
if old := list.txs.Get(tx.Nonce()); old != nil {
|
||||||
|
return false // It replaces a pending transaction.
|
||||||
|
}
|
||||||
|
// Not replacing, check if parent nonce exists in pending.
|
||||||
|
return list.txs.Get(tx.Nonce()-1) == nil
|
||||||
|
}
|
||||||
|
|
||||||
// enqueueTx inserts a new transaction into the non-executable transaction queue.
|
// enqueueTx inserts a new transaction into the non-executable transaction queue.
|
||||||
//
|
//
|
||||||
// Note, this method assumes the pool lock is held!
|
// Note, this method assumes the pool lock is held!
|
||||||
@ -996,11 +1058,12 @@ func (pool *TxPool) Has(hash common.Hash) bool {
|
|||||||
|
|
||||||
// removeTx removes a single transaction from the queue, moving all subsequent
|
// removeTx removes a single transaction from the queue, moving all subsequent
|
||||||
// transactions back to the future queue.
|
// transactions back to the future queue.
|
||||||
func (pool *TxPool) removeTx(hash common.Hash, outofbound bool) {
|
// Returns the number of transactions removed from the pending queue.
|
||||||
|
func (pool *TxPool) removeTx(hash common.Hash, outofbound bool) int {
|
||||||
// Fetch the transaction we wish to delete
|
// Fetch the transaction we wish to delete
|
||||||
tx := pool.all.Get(hash)
|
tx := pool.all.Get(hash)
|
||||||
if tx == nil {
|
if tx == nil {
|
||||||
return
|
return 0
|
||||||
}
|
}
|
||||||
addr, _ := types.Sender(pool.signer, tx) // already validated during insertion
|
addr, _ := types.Sender(pool.signer, tx) // already validated during insertion
|
||||||
|
|
||||||
@ -1028,7 +1091,7 @@ func (pool *TxPool) removeTx(hash common.Hash, outofbound bool) {
|
|||||||
pool.pendingNonces.setIfLower(addr, tx.Nonce())
|
pool.pendingNonces.setIfLower(addr, tx.Nonce())
|
||||||
// Reduce the pending counter
|
// Reduce the pending counter
|
||||||
pendingGauge.Dec(int64(1 + len(invalids)))
|
pendingGauge.Dec(int64(1 + len(invalids)))
|
||||||
return
|
return 1 + len(invalids)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
// Transaction is in the future queue
|
// Transaction is in the future queue
|
||||||
@ -1042,6 +1105,7 @@ func (pool *TxPool) removeTx(hash common.Hash, outofbound bool) {
|
|||||||
delete(pool.beats, addr)
|
delete(pool.beats, addr)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
return 0
|
||||||
}
|
}
|
||||||
|
|
||||||
// requestReset requests a pool reset to the new head block.
|
// requestReset requests a pool reset to the new head block.
|
||||||
|
212
core/txpool/txpool2_test.go
Normal file
212
core/txpool/txpool2_test.go
Normal file
@ -0,0 +1,212 @@
|
|||||||
|
// Copyright 2023 The go-ethereum Authors
|
||||||
|
// This file is part of the go-ethereum library.
|
||||||
|
//
|
||||||
|
// The go-ethereum library is free software: you can redistribute it and/or modify
|
||||||
|
// it under the terms of the GNU Lesser General Public License as published by
|
||||||
|
// the Free Software Foundation, either version 3 of the License, or
|
||||||
|
// (at your option) any later version.
|
||||||
|
//
|
||||||
|
// The go-ethereum library is distributed in the hope that it will be useful,
|
||||||
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||||
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||||
|
// GNU Lesser General Public License for more details.
|
||||||
|
//
|
||||||
|
// You should have received a copy of the GNU Lesser General Public License
|
||||||
|
// along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
|
||||||
|
package txpool
|
||||||
|
|
||||||
|
import (
|
||||||
|
"crypto/ecdsa"
|
||||||
|
"math/big"
|
||||||
|
"testing"
|
||||||
|
|
||||||
|
"github.com/ethereum/go-ethereum/common"
|
||||||
|
"github.com/ethereum/go-ethereum/core/rawdb"
|
||||||
|
"github.com/ethereum/go-ethereum/core/state"
|
||||||
|
"github.com/ethereum/go-ethereum/core/types"
|
||||||
|
"github.com/ethereum/go-ethereum/crypto"
|
||||||
|
"github.com/ethereum/go-ethereum/event"
|
||||||
|
)
|
||||||
|
|
||||||
|
func pricedValuedTransaction(nonce uint64, value int64, gaslimit uint64, gasprice *big.Int, key *ecdsa.PrivateKey) *types.Transaction {
|
||||||
|
tx, _ := types.SignTx(types.NewTransaction(nonce, common.Address{}, big.NewInt(value), gaslimit, gasprice, nil), types.HomesteadSigner{}, key)
|
||||||
|
return tx
|
||||||
|
}
|
||||||
|
|
||||||
|
func count(t *testing.T, pool *TxPool) (pending int, queued int) {
|
||||||
|
t.Helper()
|
||||||
|
pending, queued = pool.stats()
|
||||||
|
if err := validatePoolInternals(pool); err != nil {
|
||||||
|
t.Fatalf("pool internal state corrupted: %v", err)
|
||||||
|
}
|
||||||
|
return pending, queued
|
||||||
|
}
|
||||||
|
|
||||||
|
func fillPool(t *testing.T, pool *TxPool) {
|
||||||
|
t.Helper()
|
||||||
|
// Create a number of test accounts, fund them and make transactions
|
||||||
|
executableTxs := types.Transactions{}
|
||||||
|
nonExecutableTxs := types.Transactions{}
|
||||||
|
for i := 0; i < 384; i++ {
|
||||||
|
key, _ := crypto.GenerateKey()
|
||||||
|
pool.currentState.AddBalance(crypto.PubkeyToAddress(key.PublicKey), big.NewInt(10000000000))
|
||||||
|
// Add executable ones
|
||||||
|
for j := 0; j < int(pool.config.AccountSlots); j++ {
|
||||||
|
executableTxs = append(executableTxs, pricedTransaction(uint64(j), 100000, big.NewInt(300), key))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
// Import the batch and verify that limits have been enforced
|
||||||
|
pool.AddRemotesSync(executableTxs)
|
||||||
|
pool.AddRemotesSync(nonExecutableTxs)
|
||||||
|
pending, queued := pool.Stats()
|
||||||
|
slots := pool.all.Slots()
|
||||||
|
// sanity-check that the test prerequisites are ok (pending full)
|
||||||
|
if have, want := pending, slots; have != want {
|
||||||
|
t.Fatalf("have %d, want %d", have, want)
|
||||||
|
}
|
||||||
|
if have, want := queued, 0; have != want {
|
||||||
|
t.Fatalf("have %d, want %d", have, want)
|
||||||
|
}
|
||||||
|
|
||||||
|
t.Logf("pool.config: GlobalSlots=%d, GlobalQueue=%d\n", pool.config.GlobalSlots, pool.config.GlobalQueue)
|
||||||
|
t.Logf("pending: %d queued: %d, all: %d\n", pending, queued, slots)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Tests that if a batch high-priced of non-executables arrive, they do not kick out
|
||||||
|
// executable transactions
|
||||||
|
func TestTransactionFutureAttack(t *testing.T) {
|
||||||
|
t.Parallel()
|
||||||
|
|
||||||
|
// Create the pool to test the limit enforcement with
|
||||||
|
statedb, _ := state.New(common.Hash{}, state.NewDatabase(rawdb.NewMemoryDatabase()), nil)
|
||||||
|
blockchain := newTestBlockChain(1000000, statedb, new(event.Feed))
|
||||||
|
config := testTxPoolConfig
|
||||||
|
config.GlobalQueue = 100
|
||||||
|
config.GlobalSlots = 100
|
||||||
|
pool := NewTxPool(config, eip1559Config, blockchain)
|
||||||
|
defer pool.Stop()
|
||||||
|
fillPool(t, pool)
|
||||||
|
pending, _ := pool.Stats()
|
||||||
|
// Now, future transaction attack starts, let's add a bunch of expensive non-executables, and see if the pending-count drops
|
||||||
|
{
|
||||||
|
key, _ := crypto.GenerateKey()
|
||||||
|
pool.currentState.AddBalance(crypto.PubkeyToAddress(key.PublicKey), big.NewInt(100000000000))
|
||||||
|
futureTxs := types.Transactions{}
|
||||||
|
for j := 0; j < int(pool.config.GlobalSlots+pool.config.GlobalQueue); j++ {
|
||||||
|
futureTxs = append(futureTxs, pricedTransaction(1000+uint64(j), 100000, big.NewInt(500), key))
|
||||||
|
}
|
||||||
|
for i := 0; i < 5; i++ {
|
||||||
|
pool.AddRemotesSync(futureTxs)
|
||||||
|
newPending, newQueued := count(t, pool)
|
||||||
|
t.Logf("pending: %d queued: %d, all: %d\n", newPending, newQueued, pool.all.Slots())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
newPending, _ := pool.Stats()
|
||||||
|
// Pending should not have been touched
|
||||||
|
if have, want := newPending, pending; have < want {
|
||||||
|
t.Errorf("wrong pending-count, have %d, want %d (GlobalSlots: %d)",
|
||||||
|
have, want, pool.config.GlobalSlots)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Tests that if a batch high-priced of non-executables arrive, they do not kick out
|
||||||
|
// executable transactions
|
||||||
|
func TestTransactionFuture1559(t *testing.T) {
|
||||||
|
t.Parallel()
|
||||||
|
// Create the pool to test the pricing enforcement with
|
||||||
|
statedb, _ := state.New(common.Hash{}, state.NewDatabase(rawdb.NewMemoryDatabase()), nil)
|
||||||
|
blockchain := newTestBlockChain(1000000, statedb, new(event.Feed))
|
||||||
|
pool := NewTxPool(testTxPoolConfig, eip1559Config, blockchain)
|
||||||
|
defer pool.Stop()
|
||||||
|
|
||||||
|
// Create a number of test accounts, fund them and make transactions
|
||||||
|
fillPool(t, pool)
|
||||||
|
pending, _ := pool.Stats()
|
||||||
|
|
||||||
|
// Now, future transaction attack starts, let's add a bunch of expensive non-executables, and see if the pending-count drops
|
||||||
|
{
|
||||||
|
key, _ := crypto.GenerateKey()
|
||||||
|
pool.currentState.AddBalance(crypto.PubkeyToAddress(key.PublicKey), big.NewInt(100000000000))
|
||||||
|
futureTxs := types.Transactions{}
|
||||||
|
for j := 0; j < int(pool.config.GlobalSlots+pool.config.GlobalQueue); j++ {
|
||||||
|
futureTxs = append(futureTxs, dynamicFeeTx(1000+uint64(j), 100000, big.NewInt(200), big.NewInt(101), key))
|
||||||
|
}
|
||||||
|
pool.AddRemotesSync(futureTxs)
|
||||||
|
}
|
||||||
|
newPending, _ := pool.Stats()
|
||||||
|
// Pending should not have been touched
|
||||||
|
if have, want := newPending, pending; have != want {
|
||||||
|
t.Errorf("Wrong pending-count, have %d, want %d (GlobalSlots: %d)",
|
||||||
|
have, want, pool.config.GlobalSlots)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Tests that if a batch of balance-overdraft txs arrive, they do not kick out
|
||||||
|
// executable transactions
|
||||||
|
func TestTransactionZAttack(t *testing.T) {
|
||||||
|
t.Parallel()
|
||||||
|
// Create the pool to test the pricing enforcement with
|
||||||
|
statedb, _ := state.New(common.Hash{}, state.NewDatabase(rawdb.NewMemoryDatabase()), nil)
|
||||||
|
blockchain := newTestBlockChain(1000000, statedb, new(event.Feed))
|
||||||
|
pool := NewTxPool(testTxPoolConfig, eip1559Config, blockchain)
|
||||||
|
defer pool.Stop()
|
||||||
|
// Create a number of test accounts, fund them and make transactions
|
||||||
|
fillPool(t, pool)
|
||||||
|
|
||||||
|
countInvalidPending := func() int {
|
||||||
|
t.Helper()
|
||||||
|
var ivpendingNum int
|
||||||
|
pendingtxs, _ := pool.Content()
|
||||||
|
for account, txs := range pendingtxs {
|
||||||
|
cur_balance := new(big.Int).Set(pool.currentState.GetBalance(account))
|
||||||
|
for _, tx := range txs {
|
||||||
|
if cur_balance.Cmp(tx.Value()) <= 0 {
|
||||||
|
ivpendingNum++
|
||||||
|
} else {
|
||||||
|
cur_balance.Sub(cur_balance, tx.Value())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if err := validatePoolInternals(pool); err != nil {
|
||||||
|
t.Fatalf("pool internal state corrupted: %v", err)
|
||||||
|
}
|
||||||
|
return ivpendingNum
|
||||||
|
}
|
||||||
|
ivPending := countInvalidPending()
|
||||||
|
t.Logf("invalid pending: %d\n", ivPending)
|
||||||
|
|
||||||
|
// Now, DETER-Z attack starts, let's add a bunch of expensive non-executables (from N accounts) along with balance-overdraft txs (from one account), and see if the pending-count drops
|
||||||
|
for j := 0; j < int(pool.config.GlobalQueue); j++ {
|
||||||
|
futureTxs := types.Transactions{}
|
||||||
|
key, _ := crypto.GenerateKey()
|
||||||
|
pool.currentState.AddBalance(crypto.PubkeyToAddress(key.PublicKey), big.NewInt(100000000000))
|
||||||
|
futureTxs = append(futureTxs, pricedTransaction(1000+uint64(j), 21000, big.NewInt(500), key))
|
||||||
|
pool.AddRemotesSync(futureTxs)
|
||||||
|
}
|
||||||
|
|
||||||
|
overDraftTxs := types.Transactions{}
|
||||||
|
{
|
||||||
|
key, _ := crypto.GenerateKey()
|
||||||
|
pool.currentState.AddBalance(crypto.PubkeyToAddress(key.PublicKey), big.NewInt(100000000000))
|
||||||
|
for j := 0; j < int(pool.config.GlobalSlots); j++ {
|
||||||
|
overDraftTxs = append(overDraftTxs, pricedValuedTransaction(uint64(j), 60000000000, 21000, big.NewInt(500), key))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
pool.AddRemotesSync(overDraftTxs)
|
||||||
|
pool.AddRemotesSync(overDraftTxs)
|
||||||
|
pool.AddRemotesSync(overDraftTxs)
|
||||||
|
pool.AddRemotesSync(overDraftTxs)
|
||||||
|
pool.AddRemotesSync(overDraftTxs)
|
||||||
|
|
||||||
|
newPending, newQueued := count(t, pool)
|
||||||
|
newIvPending := countInvalidPending()
|
||||||
|
t.Logf("pool.all.Slots(): %d\n", pool.all.Slots())
|
||||||
|
t.Logf("pending: %d queued: %d, all: %d\n", newPending, newQueued, pool.all.Slots())
|
||||||
|
t.Logf("invalid pending: %d\n", newIvPending)
|
||||||
|
|
||||||
|
// Pending should not have been touched
|
||||||
|
if newIvPending != ivPending {
|
||||||
|
t.Errorf("Wrong invalid pending-count, have %d, want %d (GlobalSlots: %d, queued: %d)",
|
||||||
|
newIvPending, ivPending, pool.config.GlobalSlots, newQueued)
|
||||||
|
}
|
||||||
|
}
|
@ -59,15 +59,21 @@ func init() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
type testBlockChain struct {
|
type testBlockChain struct {
|
||||||
gasLimit uint64 // must be first field for 64 bit alignment (atomic access)
|
gasLimit atomic.Uint64
|
||||||
statedb *state.StateDB
|
statedb *state.StateDB
|
||||||
chainHeadFeed *event.Feed
|
chainHeadFeed *event.Feed
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func newTestBlockChain(gasLimit uint64, statedb *state.StateDB, chainHeadFeed *event.Feed) *testBlockChain {
|
||||||
|
bc := testBlockChain{statedb: statedb, chainHeadFeed: new(event.Feed)}
|
||||||
|
bc.gasLimit.Store(gasLimit)
|
||||||
|
return &bc
|
||||||
|
}
|
||||||
|
|
||||||
func (bc *testBlockChain) CurrentBlock() *types.Header {
|
func (bc *testBlockChain) CurrentBlock() *types.Header {
|
||||||
return &types.Header{
|
return &types.Header{
|
||||||
Number: new(big.Int),
|
Number: new(big.Int),
|
||||||
GasLimit: atomic.LoadUint64(&bc.gasLimit),
|
GasLimit: bc.gasLimit.Load(),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -121,7 +127,7 @@ func setupPool() (*TxPool, *ecdsa.PrivateKey) {
|
|||||||
|
|
||||||
func setupPoolWithConfig(config *params.ChainConfig) (*TxPool, *ecdsa.PrivateKey) {
|
func setupPoolWithConfig(config *params.ChainConfig) (*TxPool, *ecdsa.PrivateKey) {
|
||||||
statedb, _ := state.New(common.Hash{}, state.NewDatabase(rawdb.NewMemoryDatabase()), nil)
|
statedb, _ := state.New(common.Hash{}, state.NewDatabase(rawdb.NewMemoryDatabase()), nil)
|
||||||
blockchain := &testBlockChain{10000000, statedb, new(event.Feed)}
|
blockchain := newTestBlockChain(10000000, statedb, new(event.Feed))
|
||||||
|
|
||||||
key, _ := crypto.GenerateKey()
|
key, _ := crypto.GenerateKey()
|
||||||
pool := NewTxPool(testTxPoolConfig, config, blockchain)
|
pool := NewTxPool(testTxPoolConfig, config, blockchain)
|
||||||
@ -158,6 +164,9 @@ func validatePoolInternals(pool *TxPool) error {
|
|||||||
if nonce := pool.pendingNonces.get(addr); nonce != last+1 {
|
if nonce := pool.pendingNonces.get(addr); nonce != last+1 {
|
||||||
return fmt.Errorf("pending nonce mismatch: have %v, want %v", nonce, last+1)
|
return fmt.Errorf("pending nonce mismatch: have %v, want %v", nonce, last+1)
|
||||||
}
|
}
|
||||||
|
if txs.totalcost.Cmp(common.Big0) < 0 {
|
||||||
|
return fmt.Errorf("totalcost went negative: %v", txs.totalcost)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
@ -233,7 +242,7 @@ func TestStateChangeDuringReset(t *testing.T) {
|
|||||||
|
|
||||||
// setup pool with 2 transaction in it
|
// setup pool with 2 transaction in it
|
||||||
statedb.SetBalance(address, new(big.Int).SetUint64(params.Ether))
|
statedb.SetBalance(address, new(big.Int).SetUint64(params.Ether))
|
||||||
blockchain := &testChain{&testBlockChain{1000000000, statedb, new(event.Feed)}, address, &trigger}
|
blockchain := &testChain{newTestBlockChain(1000000000, statedb, new(event.Feed)), address, &trigger}
|
||||||
|
|
||||||
tx0 := transaction(0, 100000, key)
|
tx0 := transaction(0, 100000, key)
|
||||||
tx1 := transaction(1, 100000, key)
|
tx1 := transaction(1, 100000, key)
|
||||||
@ -427,7 +436,7 @@ func TestChainFork(t *testing.T) {
|
|||||||
statedb, _ := state.New(common.Hash{}, state.NewDatabase(rawdb.NewMemoryDatabase()), nil)
|
statedb, _ := state.New(common.Hash{}, state.NewDatabase(rawdb.NewMemoryDatabase()), nil)
|
||||||
statedb.AddBalance(addr, big.NewInt(100000000000000))
|
statedb.AddBalance(addr, big.NewInt(100000000000000))
|
||||||
|
|
||||||
pool.chain = &testBlockChain{1000000, statedb, new(event.Feed)}
|
pool.chain = newTestBlockChain(1000000, statedb, new(event.Feed))
|
||||||
<-pool.requestReset(nil, nil)
|
<-pool.requestReset(nil, nil)
|
||||||
}
|
}
|
||||||
resetState()
|
resetState()
|
||||||
@ -456,7 +465,7 @@ func TestDoubleNonce(t *testing.T) {
|
|||||||
statedb, _ := state.New(common.Hash{}, state.NewDatabase(rawdb.NewMemoryDatabase()), nil)
|
statedb, _ := state.New(common.Hash{}, state.NewDatabase(rawdb.NewMemoryDatabase()), nil)
|
||||||
statedb.AddBalance(addr, big.NewInt(100000000000000))
|
statedb.AddBalance(addr, big.NewInt(100000000000000))
|
||||||
|
|
||||||
pool.chain = &testBlockChain{1000000, statedb, new(event.Feed)}
|
pool.chain = newTestBlockChain(1000000, statedb, new(event.Feed))
|
||||||
<-pool.requestReset(nil, nil)
|
<-pool.requestReset(nil, nil)
|
||||||
}
|
}
|
||||||
resetState()
|
resetState()
|
||||||
@ -626,7 +635,7 @@ func TestDropping(t *testing.T) {
|
|||||||
t.Errorf("total transaction mismatch: have %d, want %d", pool.all.Count(), 4)
|
t.Errorf("total transaction mismatch: have %d, want %d", pool.all.Count(), 4)
|
||||||
}
|
}
|
||||||
// Reduce the block gas limit, check that invalidated transactions are dropped
|
// Reduce the block gas limit, check that invalidated transactions are dropped
|
||||||
atomic.StoreUint64(&pool.chain.(*testBlockChain).gasLimit, 100)
|
pool.chain.(*testBlockChain).gasLimit.Store(100)
|
||||||
<-pool.requestReset(nil, nil)
|
<-pool.requestReset(nil, nil)
|
||||||
|
|
||||||
if _, ok := pool.pending[account].txs.items[tx0.Nonce()]; !ok {
|
if _, ok := pool.pending[account].txs.items[tx0.Nonce()]; !ok {
|
||||||
@ -654,7 +663,7 @@ func TestPostponing(t *testing.T) {
|
|||||||
|
|
||||||
// Create the pool to test the postponing with
|
// Create the pool to test the postponing with
|
||||||
statedb, _ := state.New(common.Hash{}, state.NewDatabase(rawdb.NewMemoryDatabase()), nil)
|
statedb, _ := state.New(common.Hash{}, state.NewDatabase(rawdb.NewMemoryDatabase()), nil)
|
||||||
blockchain := &testBlockChain{1000000, statedb, new(event.Feed)}
|
blockchain := newTestBlockChain(1000000, statedb, new(event.Feed))
|
||||||
|
|
||||||
pool := NewTxPool(testTxPoolConfig, params.TestChainConfig, blockchain)
|
pool := NewTxPool(testTxPoolConfig, params.TestChainConfig, blockchain)
|
||||||
defer pool.Stop()
|
defer pool.Stop()
|
||||||
@ -866,7 +875,7 @@ func testQueueGlobalLimiting(t *testing.T, nolocals bool) {
|
|||||||
|
|
||||||
// Create the pool to test the limit enforcement with
|
// Create the pool to test the limit enforcement with
|
||||||
statedb, _ := state.New(common.Hash{}, state.NewDatabase(rawdb.NewMemoryDatabase()), nil)
|
statedb, _ := state.New(common.Hash{}, state.NewDatabase(rawdb.NewMemoryDatabase()), nil)
|
||||||
blockchain := &testBlockChain{1000000, statedb, new(event.Feed)}
|
blockchain := newTestBlockChain(1000000, statedb, new(event.Feed))
|
||||||
|
|
||||||
config := testTxPoolConfig
|
config := testTxPoolConfig
|
||||||
config.NoLocals = nolocals
|
config.NoLocals = nolocals
|
||||||
@ -958,7 +967,7 @@ func testQueueTimeLimiting(t *testing.T, nolocals bool) {
|
|||||||
|
|
||||||
// Create the pool to test the non-expiration enforcement
|
// Create the pool to test the non-expiration enforcement
|
||||||
statedb, _ := state.New(common.Hash{}, state.NewDatabase(rawdb.NewMemoryDatabase()), nil)
|
statedb, _ := state.New(common.Hash{}, state.NewDatabase(rawdb.NewMemoryDatabase()), nil)
|
||||||
blockchain := &testBlockChain{1000000, statedb, new(event.Feed)}
|
blockchain := newTestBlockChain(1000000, statedb, new(event.Feed))
|
||||||
|
|
||||||
config := testTxPoolConfig
|
config := testTxPoolConfig
|
||||||
config.Lifetime = time.Second
|
config.Lifetime = time.Second
|
||||||
@ -1105,7 +1114,7 @@ func TestPendingLimiting(t *testing.T) {
|
|||||||
defer pool.Stop()
|
defer pool.Stop()
|
||||||
|
|
||||||
account := crypto.PubkeyToAddress(key.PublicKey)
|
account := crypto.PubkeyToAddress(key.PublicKey)
|
||||||
testAddBalance(pool, account, big.NewInt(1000000))
|
testAddBalance(pool, account, big.NewInt(1000000000000))
|
||||||
|
|
||||||
// Keep track of transaction events to ensure all executables get announced
|
// Keep track of transaction events to ensure all executables get announced
|
||||||
events := make(chan core.NewTxsEvent, testTxPoolConfig.AccountQueue+5)
|
events := make(chan core.NewTxsEvent, testTxPoolConfig.AccountQueue+5)
|
||||||
@ -1143,7 +1152,7 @@ func TestPendingGlobalLimiting(t *testing.T) {
|
|||||||
|
|
||||||
// Create the pool to test the limit enforcement with
|
// Create the pool to test the limit enforcement with
|
||||||
statedb, _ := state.New(common.Hash{}, state.NewDatabase(rawdb.NewMemoryDatabase()), nil)
|
statedb, _ := state.New(common.Hash{}, state.NewDatabase(rawdb.NewMemoryDatabase()), nil)
|
||||||
blockchain := &testBlockChain{1000000, statedb, new(event.Feed)}
|
blockchain := newTestBlockChain(1000000, statedb, new(event.Feed))
|
||||||
|
|
||||||
config := testTxPoolConfig
|
config := testTxPoolConfig
|
||||||
config.GlobalSlots = config.AccountSlots * 10
|
config.GlobalSlots = config.AccountSlots * 10
|
||||||
@ -1245,7 +1254,7 @@ func TestCapClearsFromAll(t *testing.T) {
|
|||||||
|
|
||||||
// Create the pool to test the limit enforcement with
|
// Create the pool to test the limit enforcement with
|
||||||
statedb, _ := state.New(common.Hash{}, state.NewDatabase(rawdb.NewMemoryDatabase()), nil)
|
statedb, _ := state.New(common.Hash{}, state.NewDatabase(rawdb.NewMemoryDatabase()), nil)
|
||||||
blockchain := &testBlockChain{1000000, statedb, new(event.Feed)}
|
blockchain := newTestBlockChain(1000000, statedb, new(event.Feed))
|
||||||
|
|
||||||
config := testTxPoolConfig
|
config := testTxPoolConfig
|
||||||
config.AccountSlots = 2
|
config.AccountSlots = 2
|
||||||
@ -1279,7 +1288,7 @@ func TestPendingMinimumAllowance(t *testing.T) {
|
|||||||
|
|
||||||
// Create the pool to test the limit enforcement with
|
// Create the pool to test the limit enforcement with
|
||||||
statedb, _ := state.New(common.Hash{}, state.NewDatabase(rawdb.NewMemoryDatabase()), nil)
|
statedb, _ := state.New(common.Hash{}, state.NewDatabase(rawdb.NewMemoryDatabase()), nil)
|
||||||
blockchain := &testBlockChain{1000000, statedb, new(event.Feed)}
|
blockchain := newTestBlockChain(1000000, statedb, new(event.Feed))
|
||||||
|
|
||||||
config := testTxPoolConfig
|
config := testTxPoolConfig
|
||||||
config.GlobalSlots = 1
|
config.GlobalSlots = 1
|
||||||
@ -1327,7 +1336,7 @@ func TestRepricing(t *testing.T) {
|
|||||||
|
|
||||||
// Create the pool to test the pricing enforcement with
|
// Create the pool to test the pricing enforcement with
|
||||||
statedb, _ := state.New(common.Hash{}, state.NewDatabase(rawdb.NewMemoryDatabase()), nil)
|
statedb, _ := state.New(common.Hash{}, state.NewDatabase(rawdb.NewMemoryDatabase()), nil)
|
||||||
blockchain := &testBlockChain{1000000, statedb, new(event.Feed)}
|
blockchain := newTestBlockChain(1000000, statedb, new(event.Feed))
|
||||||
|
|
||||||
pool := NewTxPool(testTxPoolConfig, params.TestChainConfig, blockchain)
|
pool := NewTxPool(testTxPoolConfig, params.TestChainConfig, blockchain)
|
||||||
defer pool.Stop()
|
defer pool.Stop()
|
||||||
@ -1575,7 +1584,7 @@ func TestRepricingKeepsLocals(t *testing.T) {
|
|||||||
|
|
||||||
// Create the pool to test the pricing enforcement with
|
// Create the pool to test the pricing enforcement with
|
||||||
statedb, _ := state.New(common.Hash{}, state.NewDatabase(rawdb.NewMemoryDatabase()), nil)
|
statedb, _ := state.New(common.Hash{}, state.NewDatabase(rawdb.NewMemoryDatabase()), nil)
|
||||||
blockchain := &testBlockChain{1000000, statedb, new(event.Feed)}
|
blockchain := newTestBlockChain(1000000, statedb, new(event.Feed))
|
||||||
|
|
||||||
pool := NewTxPool(testTxPoolConfig, eip1559Config, blockchain)
|
pool := NewTxPool(testTxPoolConfig, eip1559Config, blockchain)
|
||||||
defer pool.Stop()
|
defer pool.Stop()
|
||||||
@ -1584,7 +1593,7 @@ func TestRepricingKeepsLocals(t *testing.T) {
|
|||||||
keys := make([]*ecdsa.PrivateKey, 3)
|
keys := make([]*ecdsa.PrivateKey, 3)
|
||||||
for i := 0; i < len(keys); i++ {
|
for i := 0; i < len(keys); i++ {
|
||||||
keys[i], _ = crypto.GenerateKey()
|
keys[i], _ = crypto.GenerateKey()
|
||||||
testAddBalance(pool, crypto.PubkeyToAddress(keys[i].PublicKey), big.NewInt(1000*1000000))
|
testAddBalance(pool, crypto.PubkeyToAddress(keys[i].PublicKey), big.NewInt(100000*1000000))
|
||||||
}
|
}
|
||||||
// Create transaction (both pending and queued) with a linearly growing gasprice
|
// Create transaction (both pending and queued) with a linearly growing gasprice
|
||||||
for i := uint64(0); i < 500; i++ {
|
for i := uint64(0); i < 500; i++ {
|
||||||
@ -1648,7 +1657,7 @@ func TestUnderpricing(t *testing.T) {
|
|||||||
|
|
||||||
// Create the pool to test the pricing enforcement with
|
// Create the pool to test the pricing enforcement with
|
||||||
statedb, _ := state.New(common.Hash{}, state.NewDatabase(rawdb.NewMemoryDatabase()), nil)
|
statedb, _ := state.New(common.Hash{}, state.NewDatabase(rawdb.NewMemoryDatabase()), nil)
|
||||||
blockchain := &testBlockChain{1000000, statedb, new(event.Feed)}
|
blockchain := newTestBlockChain(1000000, statedb, new(event.Feed))
|
||||||
|
|
||||||
config := testTxPoolConfig
|
config := testTxPoolConfig
|
||||||
config.GlobalSlots = 2
|
config.GlobalSlots = 2
|
||||||
@ -1663,7 +1672,7 @@ func TestUnderpricing(t *testing.T) {
|
|||||||
defer sub.Unsubscribe()
|
defer sub.Unsubscribe()
|
||||||
|
|
||||||
// Create a number of test accounts and fund them
|
// Create a number of test accounts and fund them
|
||||||
keys := make([]*ecdsa.PrivateKey, 4)
|
keys := make([]*ecdsa.PrivateKey, 5)
|
||||||
for i := 0; i < len(keys); i++ {
|
for i := 0; i < len(keys); i++ {
|
||||||
keys[i], _ = crypto.GenerateKey()
|
keys[i], _ = crypto.GenerateKey()
|
||||||
testAddBalance(pool, crypto.PubkeyToAddress(keys[i].PublicKey), big.NewInt(1000000))
|
testAddBalance(pool, crypto.PubkeyToAddress(keys[i].PublicKey), big.NewInt(1000000))
|
||||||
@ -1699,6 +1708,10 @@ func TestUnderpricing(t *testing.T) {
|
|||||||
if err := pool.AddRemote(pricedTransaction(0, 100000, big.NewInt(1), keys[1])); err != ErrUnderpriced {
|
if err := pool.AddRemote(pricedTransaction(0, 100000, big.NewInt(1), keys[1])); err != ErrUnderpriced {
|
||||||
t.Fatalf("adding underpriced pending transaction error mismatch: have %v, want %v", err, ErrUnderpriced)
|
t.Fatalf("adding underpriced pending transaction error mismatch: have %v, want %v", err, ErrUnderpriced)
|
||||||
}
|
}
|
||||||
|
// Replace a future transaction with a future transaction
|
||||||
|
if err := pool.AddRemote(pricedTransaction(1, 100000, big.NewInt(2), keys[1])); err != nil { // +K1:1 => -K1:1 => Pend K0:0, K0:1, K2:0; Que K1:1
|
||||||
|
t.Fatalf("failed to add well priced transaction: %v", err)
|
||||||
|
}
|
||||||
// Ensure that adding high priced transactions drops cheap ones, but not own
|
// Ensure that adding high priced transactions drops cheap ones, but not own
|
||||||
if err := pool.AddRemote(pricedTransaction(0, 100000, big.NewInt(3), keys[1])); err != nil { // +K1:0 => -K1:1 => Pend K0:0, K0:1, K1:0, K2:0; Que -
|
if err := pool.AddRemote(pricedTransaction(0, 100000, big.NewInt(3), keys[1])); err != nil { // +K1:0 => -K1:1 => Pend K0:0, K0:1, K1:0, K2:0; Que -
|
||||||
t.Fatalf("failed to add well priced transaction: %v", err)
|
t.Fatalf("failed to add well priced transaction: %v", err)
|
||||||
@ -1709,6 +1722,10 @@ func TestUnderpricing(t *testing.T) {
|
|||||||
if err := pool.AddRemote(pricedTransaction(3, 100000, big.NewInt(5), keys[1])); err != nil { // +K1:3 => -K0:1 => Pend K1:0, K2:0; Que K1:2 K1:3
|
if err := pool.AddRemote(pricedTransaction(3, 100000, big.NewInt(5), keys[1])); err != nil { // +K1:3 => -K0:1 => Pend K1:0, K2:0; Que K1:2 K1:3
|
||||||
t.Fatalf("failed to add well priced transaction: %v", err)
|
t.Fatalf("failed to add well priced transaction: %v", err)
|
||||||
}
|
}
|
||||||
|
// Ensure that replacing a pending transaction with a future transaction fails
|
||||||
|
if err := pool.AddRemote(pricedTransaction(5, 100000, big.NewInt(6), keys[1])); err != ErrFutureReplacePending {
|
||||||
|
t.Fatalf("adding future replace transaction error mismatch: have %v, want %v", err, ErrFutureReplacePending)
|
||||||
|
}
|
||||||
pending, queued = pool.Stats()
|
pending, queued = pool.Stats()
|
||||||
if pending != 2 {
|
if pending != 2 {
|
||||||
t.Fatalf("pending transactions mismatched: have %d, want %d", pending, 2)
|
t.Fatalf("pending transactions mismatched: have %d, want %d", pending, 2)
|
||||||
@ -1716,7 +1733,7 @@ func TestUnderpricing(t *testing.T) {
|
|||||||
if queued != 2 {
|
if queued != 2 {
|
||||||
t.Fatalf("queued transactions mismatched: have %d, want %d", queued, 2)
|
t.Fatalf("queued transactions mismatched: have %d, want %d", queued, 2)
|
||||||
}
|
}
|
||||||
if err := validateEvents(events, 1); err != nil {
|
if err := validateEvents(events, 2); err != nil {
|
||||||
t.Fatalf("additional event firing failed: %v", err)
|
t.Fatalf("additional event firing failed: %v", err)
|
||||||
}
|
}
|
||||||
if err := validatePoolInternals(pool); err != nil {
|
if err := validatePoolInternals(pool); err != nil {
|
||||||
@ -1754,7 +1771,7 @@ func TestStableUnderpricing(t *testing.T) {
|
|||||||
|
|
||||||
// Create the pool to test the pricing enforcement with
|
// Create the pool to test the pricing enforcement with
|
||||||
statedb, _ := state.New(common.Hash{}, state.NewDatabase(rawdb.NewMemoryDatabase()), nil)
|
statedb, _ := state.New(common.Hash{}, state.NewDatabase(rawdb.NewMemoryDatabase()), nil)
|
||||||
blockchain := &testBlockChain{1000000, statedb, new(event.Feed)}
|
blockchain := newTestBlockChain(1000000, statedb, new(event.Feed))
|
||||||
|
|
||||||
config := testTxPoolConfig
|
config := testTxPoolConfig
|
||||||
config.GlobalSlots = 128
|
config.GlobalSlots = 128
|
||||||
@ -1878,11 +1895,11 @@ func TestUnderpricingDynamicFee(t *testing.T) {
|
|||||||
t.Fatalf("failed to add well priced transaction: %v", err)
|
t.Fatalf("failed to add well priced transaction: %v", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
tx = pricedTransaction(2, 100000, big.NewInt(3), keys[1])
|
tx = pricedTransaction(1, 100000, big.NewInt(3), keys[1])
|
||||||
if err := pool.AddRemote(tx); err != nil { // +K1:2, -K0:1 => Pend K0:0 K1:0, K2:0; Que K1:2
|
if err := pool.AddRemote(tx); err != nil { // +K1:2, -K0:1 => Pend K0:0 K1:0, K2:0; Que K1:2
|
||||||
t.Fatalf("failed to add well priced transaction: %v", err)
|
t.Fatalf("failed to add well priced transaction: %v", err)
|
||||||
}
|
}
|
||||||
tx = dynamicFeeTx(3, 100000, big.NewInt(4), big.NewInt(1), keys[1])
|
tx = dynamicFeeTx(2, 100000, big.NewInt(4), big.NewInt(1), keys[1])
|
||||||
if err := pool.AddRemote(tx); err != nil { // +K1:3, -K1:0 => Pend K0:0 K2:0; Que K1:2 K1:3
|
if err := pool.AddRemote(tx); err != nil { // +K1:3, -K1:0 => Pend K0:0 K2:0; Que K1:2 K1:3
|
||||||
t.Fatalf("failed to add well priced transaction: %v", err)
|
t.Fatalf("failed to add well priced transaction: %v", err)
|
||||||
}
|
}
|
||||||
@ -1893,7 +1910,7 @@ func TestUnderpricingDynamicFee(t *testing.T) {
|
|||||||
if queued != 2 {
|
if queued != 2 {
|
||||||
t.Fatalf("queued transactions mismatched: have %d, want %d", queued, 2)
|
t.Fatalf("queued transactions mismatched: have %d, want %d", queued, 2)
|
||||||
}
|
}
|
||||||
if err := validateEvents(events, 1); err != nil {
|
if err := validateEvents(events, 2); err != nil {
|
||||||
t.Fatalf("additional event firing failed: %v", err)
|
t.Fatalf("additional event firing failed: %v", err)
|
||||||
}
|
}
|
||||||
if err := validatePoolInternals(pool); err != nil {
|
if err := validatePoolInternals(pool); err != nil {
|
||||||
@ -1986,7 +2003,7 @@ func TestDeduplication(t *testing.T) {
|
|||||||
|
|
||||||
// Create the pool to test the pricing enforcement with
|
// Create the pool to test the pricing enforcement with
|
||||||
statedb, _ := state.New(common.Hash{}, state.NewDatabase(rawdb.NewMemoryDatabase()), nil)
|
statedb, _ := state.New(common.Hash{}, state.NewDatabase(rawdb.NewMemoryDatabase()), nil)
|
||||||
blockchain := &testBlockChain{1000000, statedb, new(event.Feed)}
|
blockchain := newTestBlockChain(1000000, statedb, new(event.Feed))
|
||||||
|
|
||||||
pool := NewTxPool(testTxPoolConfig, params.TestChainConfig, blockchain)
|
pool := NewTxPool(testTxPoolConfig, params.TestChainConfig, blockchain)
|
||||||
defer pool.Stop()
|
defer pool.Stop()
|
||||||
@ -2052,7 +2069,7 @@ func TestReplacement(t *testing.T) {
|
|||||||
|
|
||||||
// Create the pool to test the pricing enforcement with
|
// Create the pool to test the pricing enforcement with
|
||||||
statedb, _ := state.New(common.Hash{}, state.NewDatabase(rawdb.NewMemoryDatabase()), nil)
|
statedb, _ := state.New(common.Hash{}, state.NewDatabase(rawdb.NewMemoryDatabase()), nil)
|
||||||
blockchain := &testBlockChain{1000000, statedb, new(event.Feed)}
|
blockchain := newTestBlockChain(1000000, statedb, new(event.Feed))
|
||||||
|
|
||||||
pool := NewTxPool(testTxPoolConfig, params.TestChainConfig, blockchain)
|
pool := NewTxPool(testTxPoolConfig, params.TestChainConfig, blockchain)
|
||||||
defer pool.Stop()
|
defer pool.Stop()
|
||||||
@ -2257,7 +2274,7 @@ func testJournaling(t *testing.T, nolocals bool) {
|
|||||||
|
|
||||||
// Create the original pool to inject transaction into the journal
|
// Create the original pool to inject transaction into the journal
|
||||||
statedb, _ := state.New(common.Hash{}, state.NewDatabase(rawdb.NewMemoryDatabase()), nil)
|
statedb, _ := state.New(common.Hash{}, state.NewDatabase(rawdb.NewMemoryDatabase()), nil)
|
||||||
blockchain := &testBlockChain{1000000, statedb, new(event.Feed)}
|
blockchain := newTestBlockChain(1000000, statedb, new(event.Feed))
|
||||||
|
|
||||||
config := testTxPoolConfig
|
config := testTxPoolConfig
|
||||||
config.NoLocals = nolocals
|
config.NoLocals = nolocals
|
||||||
@ -2299,7 +2316,7 @@ func testJournaling(t *testing.T, nolocals bool) {
|
|||||||
// Terminate the old pool, bump the local nonce, create a new pool and ensure relevant transaction survive
|
// Terminate the old pool, bump the local nonce, create a new pool and ensure relevant transaction survive
|
||||||
pool.Stop()
|
pool.Stop()
|
||||||
statedb.SetNonce(crypto.PubkeyToAddress(local.PublicKey), 1)
|
statedb.SetNonce(crypto.PubkeyToAddress(local.PublicKey), 1)
|
||||||
blockchain = &testBlockChain{1000000, statedb, new(event.Feed)}
|
blockchain = newTestBlockChain(1000000, statedb, new(event.Feed))
|
||||||
|
|
||||||
pool = NewTxPool(config, params.TestChainConfig, blockchain)
|
pool = NewTxPool(config, params.TestChainConfig, blockchain)
|
||||||
|
|
||||||
@ -2326,7 +2343,7 @@ func testJournaling(t *testing.T, nolocals bool) {
|
|||||||
pool.Stop()
|
pool.Stop()
|
||||||
|
|
||||||
statedb.SetNonce(crypto.PubkeyToAddress(local.PublicKey), 1)
|
statedb.SetNonce(crypto.PubkeyToAddress(local.PublicKey), 1)
|
||||||
blockchain = &testBlockChain{1000000, statedb, new(event.Feed)}
|
blockchain = newTestBlockChain(1000000, statedb, new(event.Feed))
|
||||||
pool = NewTxPool(config, params.TestChainConfig, blockchain)
|
pool = NewTxPool(config, params.TestChainConfig, blockchain)
|
||||||
|
|
||||||
pending, queued = pool.Stats()
|
pending, queued = pool.Stats()
|
||||||
@ -2355,7 +2372,7 @@ func TestStatusCheck(t *testing.T) {
|
|||||||
|
|
||||||
// Create the pool to test the status retrievals with
|
// Create the pool to test the status retrievals with
|
||||||
statedb, _ := state.New(common.Hash{}, state.NewDatabase(rawdb.NewMemoryDatabase()), nil)
|
statedb, _ := state.New(common.Hash{}, state.NewDatabase(rawdb.NewMemoryDatabase()), nil)
|
||||||
blockchain := &testBlockChain{1000000, statedb, new(event.Feed)}
|
blockchain := newTestBlockChain(1000000, statedb, new(event.Feed))
|
||||||
|
|
||||||
pool := NewTxPool(testTxPoolConfig, params.TestChainConfig, blockchain)
|
pool := NewTxPool(testTxPoolConfig, params.TestChainConfig, blockchain)
|
||||||
defer pool.Stop()
|
defer pool.Stop()
|
||||||
@ -2487,7 +2504,7 @@ func benchmarkBatchInsert(b *testing.B, size int, local bool) {
|
|||||||
defer pool.Stop()
|
defer pool.Stop()
|
||||||
|
|
||||||
account := crypto.PubkeyToAddress(key.PublicKey)
|
account := crypto.PubkeyToAddress(key.PublicKey)
|
||||||
testAddBalance(pool, account, big.NewInt(1000000))
|
testAddBalance(pool, account, big.NewInt(1000000000000000000))
|
||||||
|
|
||||||
batches := make([]types.Transactions, b.N)
|
batches := make([]types.Transactions, b.N)
|
||||||
for i := 0; i < b.N; i++ {
|
for i := 0; i < b.N; i++ {
|
||||||
|
@ -589,74 +589,6 @@ func (t *TransactionsByPriceAndNonce) Pop() {
|
|||||||
heap.Pop(&t.heads)
|
heap.Pop(&t.heads)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Message is a fully derived transaction and implements core.Message
|
|
||||||
//
|
|
||||||
// NOTE: In a future PR this will be removed.
|
|
||||||
type Message struct {
|
|
||||||
to *common.Address
|
|
||||||
from common.Address
|
|
||||||
nonce uint64
|
|
||||||
amount *big.Int
|
|
||||||
gasLimit uint64
|
|
||||||
gasPrice *big.Int
|
|
||||||
gasFeeCap *big.Int
|
|
||||||
gasTipCap *big.Int
|
|
||||||
data []byte
|
|
||||||
accessList AccessList
|
|
||||||
isFake bool
|
|
||||||
}
|
|
||||||
|
|
||||||
func NewMessage(from common.Address, to *common.Address, nonce uint64, amount *big.Int, gasLimit uint64, gasPrice, gasFeeCap, gasTipCap *big.Int, data []byte, accessList AccessList, isFake bool) Message {
|
|
||||||
return Message{
|
|
||||||
from: from,
|
|
||||||
to: to,
|
|
||||||
nonce: nonce,
|
|
||||||
amount: amount,
|
|
||||||
gasLimit: gasLimit,
|
|
||||||
gasPrice: gasPrice,
|
|
||||||
gasFeeCap: gasFeeCap,
|
|
||||||
gasTipCap: gasTipCap,
|
|
||||||
data: data,
|
|
||||||
accessList: accessList,
|
|
||||||
isFake: isFake,
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// AsMessage returns the transaction as a core.Message.
|
|
||||||
func (tx *Transaction) AsMessage(s Signer, baseFee *big.Int) (Message, error) {
|
|
||||||
msg := Message{
|
|
||||||
nonce: tx.Nonce(),
|
|
||||||
gasLimit: tx.Gas(),
|
|
||||||
gasPrice: new(big.Int).Set(tx.GasPrice()),
|
|
||||||
gasFeeCap: new(big.Int).Set(tx.GasFeeCap()),
|
|
||||||
gasTipCap: new(big.Int).Set(tx.GasTipCap()),
|
|
||||||
to: tx.To(),
|
|
||||||
amount: tx.Value(),
|
|
||||||
data: tx.Data(),
|
|
||||||
accessList: tx.AccessList(),
|
|
||||||
isFake: false,
|
|
||||||
}
|
|
||||||
// If baseFee provided, set gasPrice to effectiveGasPrice.
|
|
||||||
if baseFee != nil {
|
|
||||||
msg.gasPrice = math.BigMin(msg.gasPrice.Add(msg.gasTipCap, baseFee), msg.gasFeeCap)
|
|
||||||
}
|
|
||||||
var err error
|
|
||||||
msg.from, err = Sender(s, tx)
|
|
||||||
return msg, err
|
|
||||||
}
|
|
||||||
|
|
||||||
func (m Message) From() common.Address { return m.from }
|
|
||||||
func (m Message) To() *common.Address { return m.to }
|
|
||||||
func (m Message) GasPrice() *big.Int { return m.gasPrice }
|
|
||||||
func (m Message) GasFeeCap() *big.Int { return m.gasFeeCap }
|
|
||||||
func (m Message) GasTipCap() *big.Int { return m.gasTipCap }
|
|
||||||
func (m Message) Value() *big.Int { return m.amount }
|
|
||||||
func (m Message) Gas() uint64 { return m.gasLimit }
|
|
||||||
func (m Message) Nonce() uint64 { return m.nonce }
|
|
||||||
func (m Message) Data() []byte { return m.data }
|
|
||||||
func (m Message) AccessList() AccessList { return m.accessList }
|
|
||||||
func (m Message) IsFake() bool { return m.isFake }
|
|
||||||
|
|
||||||
// copyAddressPtr copies an address.
|
// copyAddressPtr copies an address.
|
||||||
func copyAddressPtr(a *common.Address) *common.Address {
|
func copyAddressPtr(a *common.Address) *common.Address {
|
||||||
if a == nil {
|
if a == nil {
|
||||||
|
@ -29,7 +29,6 @@ import (
|
|||||||
"github.com/ethereum/go-ethereum/crypto/bls12381"
|
"github.com/ethereum/go-ethereum/crypto/bls12381"
|
||||||
"github.com/ethereum/go-ethereum/crypto/bn256"
|
"github.com/ethereum/go-ethereum/crypto/bn256"
|
||||||
"github.com/ethereum/go-ethereum/params"
|
"github.com/ethereum/go-ethereum/params"
|
||||||
big2 "github.com/holiman/big"
|
|
||||||
"golang.org/x/crypto/ripemd160"
|
"golang.org/x/crypto/ripemd160"
|
||||||
)
|
)
|
||||||
|
|
||||||
@ -378,9 +377,9 @@ func (c *bigModExp) Run(input []byte) ([]byte, error) {
|
|||||||
}
|
}
|
||||||
// Retrieve the operands and execute the exponentiation
|
// Retrieve the operands and execute the exponentiation
|
||||||
var (
|
var (
|
||||||
base = new(big2.Int).SetBytes(getData(input, 0, baseLen))
|
base = new(big.Int).SetBytes(getData(input, 0, baseLen))
|
||||||
exp = new(big2.Int).SetBytes(getData(input, baseLen, expLen))
|
exp = new(big.Int).SetBytes(getData(input, baseLen, expLen))
|
||||||
mod = new(big2.Int).SetBytes(getData(input, baseLen+expLen, modLen))
|
mod = new(big.Int).SetBytes(getData(input, baseLen+expLen, modLen))
|
||||||
v []byte
|
v []byte
|
||||||
)
|
)
|
||||||
switch {
|
switch {
|
||||||
|
@ -163,7 +163,7 @@ func gasSStore(evm *EVM, contract *Contract, stack *Stack, mem *Memory, memorySi
|
|||||||
return params.NetSstoreDirtyGas, nil
|
return params.NetSstoreDirtyGas, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// Here come the EIP220 rules:
|
// Here come the EIP2200 rules:
|
||||||
//
|
//
|
||||||
// (0.) If *gasleft* is less than or equal to 2300, fail the current call.
|
// (0.) If *gasleft* is less than or equal to 2300, fail the current call.
|
||||||
// (1.) If current value equals new value (this is a no-op), SLOAD_GAS is deducted.
|
// (1.) If current value equals new value (this is a no-op), SLOAD_GAS is deducted.
|
||||||
|
@ -76,8 +76,6 @@ type StateDB interface {
|
|||||||
|
|
||||||
AddLog(*types.Log)
|
AddLog(*types.Log)
|
||||||
AddPreimage(common.Hash, []byte)
|
AddPreimage(common.Hash, []byte)
|
||||||
|
|
||||||
ForEachStorage(common.Address, func(common.Hash, common.Hash) bool) error
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// CallContext provides a basic interface for the EVM calling conventions. The EVM
|
// CallContext provides a basic interface for the EVM calling conventions. The EVM
|
||||||
|
@ -98,7 +98,7 @@ func newMergeInstructionSet() JumpTable {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// newLondonInstructionSet returns the frontier, homestead, byzantium,
|
// newLondonInstructionSet returns the frontier, homestead, byzantium,
|
||||||
// contantinople, istanbul, petersburg, berlin and london instructions.
|
// constantinople, istanbul, petersburg, berlin and london instructions.
|
||||||
func newLondonInstructionSet() JumpTable {
|
func newLondonInstructionSet() JumpTable {
|
||||||
instructionSet := newBerlinInstructionSet()
|
instructionSet := newBerlinInstructionSet()
|
||||||
enable3529(&instructionSet) // EIP-3529: Reduction in refunds https://eips.ethereum.org/EIPS/eip-3529
|
enable3529(&instructionSet) // EIP-3529: Reduction in refunds https://eips.ethereum.org/EIPS/eip-3529
|
||||||
@ -107,7 +107,7 @@ func newLondonInstructionSet() JumpTable {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// newBerlinInstructionSet returns the frontier, homestead, byzantium,
|
// newBerlinInstructionSet returns the frontier, homestead, byzantium,
|
||||||
// contantinople, istanbul, petersburg and berlin instructions.
|
// constantinople, istanbul, petersburg and berlin instructions.
|
||||||
func newBerlinInstructionSet() JumpTable {
|
func newBerlinInstructionSet() JumpTable {
|
||||||
instructionSet := newIstanbulInstructionSet()
|
instructionSet := newIstanbulInstructionSet()
|
||||||
enable2929(&instructionSet) // Access lists for trie accesses https://eips.ethereum.org/EIPS/eip-2929
|
enable2929(&instructionSet) // Access lists for trie accesses https://eips.ethereum.org/EIPS/eip-2929
|
||||||
@ -115,7 +115,7 @@ func newBerlinInstructionSet() JumpTable {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// newIstanbulInstructionSet returns the frontier, homestead, byzantium,
|
// newIstanbulInstructionSet returns the frontier, homestead, byzantium,
|
||||||
// contantinople, istanbul and petersburg instructions.
|
// constantinople, istanbul and petersburg instructions.
|
||||||
func newIstanbulInstructionSet() JumpTable {
|
func newIstanbulInstructionSet() JumpTable {
|
||||||
instructionSet := newConstantinopleInstructionSet()
|
instructionSet := newConstantinopleInstructionSet()
|
||||||
|
|
||||||
@ -127,7 +127,7 @@ func newIstanbulInstructionSet() JumpTable {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// newConstantinopleInstructionSet returns the frontier, homestead,
|
// newConstantinopleInstructionSet returns the frontier, homestead,
|
||||||
// byzantium and contantinople instructions.
|
// byzantium and constantinople instructions.
|
||||||
func newConstantinopleInstructionSet() JumpTable {
|
func newConstantinopleInstructionSet() JumpTable {
|
||||||
instructionSet := newByzantiumInstructionSet()
|
instructionSet := newByzantiumInstructionSet()
|
||||||
instructionSet[SHL] = &operation{
|
instructionSet[SHL] = &operation{
|
||||||
|
@ -161,7 +161,7 @@ func (api *AdminAPI) ExportChain(file string, first *uint64, last *uint64) (bool
|
|||||||
return false, errors.New("location would overwrite an existing file")
|
return false, errors.New("location would overwrite an existing file")
|
||||||
}
|
}
|
||||||
// Make sure we can create the file to export into
|
// Make sure we can create the file to export into
|
||||||
out, err := os.OpenFile(file, os.O_CREATE|os.O_WRONLY|os.O_TRUNC, os.ModePerm)
|
out, err := os.OpenFile(file, os.O_CREATE|os.O_WRONLY|os.O_TRUNC, 0644)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return false, err
|
return false, err
|
||||||
}
|
}
|
||||||
|
@ -75,6 +75,9 @@ func (b *EthAPIBackend) HeaderByNumber(ctx context.Context, number rpc.BlockNumb
|
|||||||
return b.eth.blockchain.CurrentBlock(), nil
|
return b.eth.blockchain.CurrentBlock(), nil
|
||||||
}
|
}
|
||||||
if number == rpc.FinalizedBlockNumber {
|
if number == rpc.FinalizedBlockNumber {
|
||||||
|
if !b.eth.Merger().TDDReached() {
|
||||||
|
return nil, errors.New("'finalized' tag not supported on pre-merge network")
|
||||||
|
}
|
||||||
block := b.eth.blockchain.CurrentFinalBlock()
|
block := b.eth.blockchain.CurrentFinalBlock()
|
||||||
if block != nil {
|
if block != nil {
|
||||||
return block, nil
|
return block, nil
|
||||||
@ -82,6 +85,9 @@ func (b *EthAPIBackend) HeaderByNumber(ctx context.Context, number rpc.BlockNumb
|
|||||||
return nil, errors.New("finalized block not found")
|
return nil, errors.New("finalized block not found")
|
||||||
}
|
}
|
||||||
if number == rpc.SafeBlockNumber {
|
if number == rpc.SafeBlockNumber {
|
||||||
|
if !b.eth.Merger().TDDReached() {
|
||||||
|
return nil, errors.New("'safe' tag not supported on pre-merge network")
|
||||||
|
}
|
||||||
block := b.eth.blockchain.CurrentSafeBlock()
|
block := b.eth.blockchain.CurrentSafeBlock()
|
||||||
if block != nil {
|
if block != nil {
|
||||||
return block, nil
|
return block, nil
|
||||||
@ -124,10 +130,16 @@ func (b *EthAPIBackend) BlockByNumber(ctx context.Context, number rpc.BlockNumbe
|
|||||||
return b.eth.blockchain.GetBlock(header.Hash(), header.Number.Uint64()), nil
|
return b.eth.blockchain.GetBlock(header.Hash(), header.Number.Uint64()), nil
|
||||||
}
|
}
|
||||||
if number == rpc.FinalizedBlockNumber {
|
if number == rpc.FinalizedBlockNumber {
|
||||||
|
if !b.eth.Merger().TDDReached() {
|
||||||
|
return nil, errors.New("'finalized' tag not supported on pre-merge network")
|
||||||
|
}
|
||||||
header := b.eth.blockchain.CurrentFinalBlock()
|
header := b.eth.blockchain.CurrentFinalBlock()
|
||||||
return b.eth.blockchain.GetBlock(header.Hash(), header.Number.Uint64()), nil
|
return b.eth.blockchain.GetBlock(header.Hash(), header.Number.Uint64()), nil
|
||||||
}
|
}
|
||||||
if number == rpc.SafeBlockNumber {
|
if number == rpc.SafeBlockNumber {
|
||||||
|
if !b.eth.Merger().TDDReached() {
|
||||||
|
return nil, errors.New("'safe' tag not supported on pre-merge network")
|
||||||
|
}
|
||||||
header := b.eth.blockchain.CurrentSafeBlock()
|
header := b.eth.blockchain.CurrentSafeBlock()
|
||||||
return b.eth.blockchain.GetBlock(header.Hash(), header.Number.Uint64()), nil
|
return b.eth.blockchain.GetBlock(header.Hash(), header.Number.Uint64()), nil
|
||||||
}
|
}
|
||||||
@ -228,7 +240,7 @@ func (b *EthAPIBackend) GetTd(ctx context.Context, hash common.Hash) *big.Int {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (b *EthAPIBackend) GetEVM(ctx context.Context, msg core.Message, state *state.StateDB, header *types.Header, vmConfig *vm.Config) (*vm.EVM, func() error, error) {
|
func (b *EthAPIBackend) GetEVM(ctx context.Context, msg *core.Message, state *state.StateDB, header *types.Header, vmConfig *vm.Config) (*vm.EVM, func() error, error) {
|
||||||
if vmConfig == nil {
|
if vmConfig == nil {
|
||||||
vmConfig = b.eth.blockchain.GetVMConfig()
|
vmConfig = b.eth.blockchain.GetVMConfig()
|
||||||
}
|
}
|
||||||
@ -382,6 +394,6 @@ func (b *EthAPIBackend) StateAtBlock(ctx context.Context, block *types.Block, re
|
|||||||
return b.eth.StateAtBlock(ctx, block, reexec, base, readOnly, preferDisk)
|
return b.eth.StateAtBlock(ctx, block, reexec, base, readOnly, preferDisk)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (b *EthAPIBackend) StateAtTransaction(ctx context.Context, block *types.Block, txIndex int, reexec uint64) (core.Message, vm.BlockContext, *state.StateDB, tracers.StateReleaseFunc, error) {
|
func (b *EthAPIBackend) StateAtTransaction(ctx context.Context, block *types.Block, txIndex int, reexec uint64) (*core.Message, vm.BlockContext, *state.StateDB, tracers.StateReleaseFunc, error) {
|
||||||
return b.eth.stateAtTransaction(ctx, block, txIndex, reexec)
|
return b.eth.stateAtTransaction(ctx, block, txIndex, reexec)
|
||||||
}
|
}
|
||||||
|
@ -72,7 +72,7 @@ const (
|
|||||||
// beaconUpdateConsensusTimeout is the max time allowed for a beacon client
|
// beaconUpdateConsensusTimeout is the max time allowed for a beacon client
|
||||||
// to send a consensus update before it's considered offline and the user is
|
// to send a consensus update before it's considered offline and the user is
|
||||||
// warned.
|
// warned.
|
||||||
beaconUpdateConsensusTimeout = 30 * time.Second
|
beaconUpdateConsensusTimeout = 2 * time.Minute
|
||||||
|
|
||||||
// beaconUpdateWarnFrequency is the frequency at which to warn the user that
|
// beaconUpdateWarnFrequency is the frequency at which to warn the user that
|
||||||
// the beacon client is offline.
|
// the beacon client is offline.
|
||||||
|
@ -189,7 +189,7 @@ func (eth *Ethereum) StateAtBlock(ctx context.Context, block *types.Block, reexe
|
|||||||
}
|
}
|
||||||
|
|
||||||
// stateAtTransaction returns the execution environment of a certain transaction.
|
// stateAtTransaction returns the execution environment of a certain transaction.
|
||||||
func (eth *Ethereum) stateAtTransaction(ctx context.Context, block *types.Block, txIndex int, reexec uint64) (core.Message, vm.BlockContext, *state.StateDB, tracers.StateReleaseFunc, error) {
|
func (eth *Ethereum) stateAtTransaction(ctx context.Context, block *types.Block, txIndex int, reexec uint64) (*core.Message, vm.BlockContext, *state.StateDB, tracers.StateReleaseFunc, error) {
|
||||||
// Short circuit if it's genesis block.
|
// Short circuit if it's genesis block.
|
||||||
if block.NumberU64() == 0 {
|
if block.NumberU64() == 0 {
|
||||||
return nil, vm.BlockContext{}, nil, nil, errors.New("no transaction in genesis")
|
return nil, vm.BlockContext{}, nil, nil, errors.New("no transaction in genesis")
|
||||||
@ -212,7 +212,7 @@ func (eth *Ethereum) stateAtTransaction(ctx context.Context, block *types.Block,
|
|||||||
signer := types.MakeSigner(eth.blockchain.Config(), block.Number())
|
signer := types.MakeSigner(eth.blockchain.Config(), block.Number())
|
||||||
for idx, tx := range block.Transactions() {
|
for idx, tx := range block.Transactions() {
|
||||||
// Assemble the transaction call message and return if the requested offset
|
// Assemble the transaction call message and return if the requested offset
|
||||||
msg, _ := tx.AsMessage(signer, block.BaseFee())
|
msg, _ := core.TransactionToMessage(tx, signer, block.BaseFee())
|
||||||
txContext := core.NewEVMTxContext(msg)
|
txContext := core.NewEVMTxContext(msg)
|
||||||
context := core.NewEVMBlockContext(block.Header(), eth.blockchain, nil)
|
context := core.NewEVMBlockContext(block.Header(), eth.blockchain, nil)
|
||||||
if idx == txIndex {
|
if idx == txIndex {
|
||||||
|
@ -87,7 +87,7 @@ type Backend interface {
|
|||||||
Engine() consensus.Engine
|
Engine() consensus.Engine
|
||||||
ChainDb() ethdb.Database
|
ChainDb() ethdb.Database
|
||||||
StateAtBlock(ctx context.Context, block *types.Block, reexec uint64, base *state.StateDB, readOnly bool, preferDisk bool) (*state.StateDB, StateReleaseFunc, error)
|
StateAtBlock(ctx context.Context, block *types.Block, reexec uint64, base *state.StateDB, readOnly bool, preferDisk bool) (*state.StateDB, StateReleaseFunc, error)
|
||||||
StateAtTransaction(ctx context.Context, block *types.Block, txIndex int, reexec uint64) (core.Message, vm.BlockContext, *state.StateDB, StateReleaseFunc, error)
|
StateAtTransaction(ctx context.Context, block *types.Block, txIndex int, reexec uint64) (*core.Message, vm.BlockContext, *state.StateDB, StateReleaseFunc, error)
|
||||||
}
|
}
|
||||||
|
|
||||||
// API is the collection of tracing APIs exposed over the private debugging endpoint.
|
// API is the collection of tracing APIs exposed over the private debugging endpoint.
|
||||||
@ -293,7 +293,7 @@ func (api *API) traceChain(start, end *types.Block, config *TraceConfig, closed
|
|||||||
)
|
)
|
||||||
// Trace all the transactions contained within
|
// Trace all the transactions contained within
|
||||||
for i, tx := range task.block.Transactions() {
|
for i, tx := range task.block.Transactions() {
|
||||||
msg, _ := tx.AsMessage(signer, task.block.BaseFee())
|
msg, _ := core.TransactionToMessage(tx, signer, task.block.BaseFee())
|
||||||
txctx := &Context{
|
txctx := &Context{
|
||||||
BlockHash: task.block.Hash(),
|
BlockHash: task.block.Hash(),
|
||||||
BlockNumber: task.block.Number(),
|
BlockNumber: task.block.Number(),
|
||||||
@ -554,12 +554,12 @@ func (api *API) IntermediateRoots(ctx context.Context, hash common.Hash, config
|
|||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
var (
|
var (
|
||||||
msg, _ = tx.AsMessage(signer, block.BaseFee())
|
msg, _ = core.TransactionToMessage(tx, signer, block.BaseFee())
|
||||||
txContext = core.NewEVMTxContext(msg)
|
txContext = core.NewEVMTxContext(msg)
|
||||||
vmenv = vm.NewEVM(vmctx, txContext, statedb, chainConfig, vm.Config{})
|
vmenv = vm.NewEVM(vmctx, txContext, statedb, chainConfig, vm.Config{})
|
||||||
)
|
)
|
||||||
statedb.SetTxContext(tx.Hash(), i)
|
statedb.SetTxContext(tx.Hash(), i)
|
||||||
if _, err := core.ApplyMessage(vmenv, msg, new(core.GasPool).AddGas(msg.Gas())); err != nil {
|
if _, err := core.ApplyMessage(vmenv, msg, new(core.GasPool).AddGas(msg.GasLimit)); err != nil {
|
||||||
log.Warn("Tracing intermediate roots did not complete", "txindex", i, "txhash", tx.Hash(), "err", err)
|
log.Warn("Tracing intermediate roots did not complete", "txindex", i, "txhash", tx.Hash(), "err", err)
|
||||||
// We intentionally don't return the error here: if we do, then the RPC server will not
|
// We intentionally don't return the error here: if we do, then the RPC server will not
|
||||||
// return the roots. Most likely, the caller already knows that a certain transaction fails to
|
// return the roots. Most likely, the caller already knows that a certain transaction fails to
|
||||||
@ -628,7 +628,7 @@ func (api *API) traceBlock(ctx context.Context, block *types.Block, config *Trac
|
|||||||
)
|
)
|
||||||
for i, tx := range txs {
|
for i, tx := range txs {
|
||||||
// Generate the next state snapshot fast without tracing
|
// Generate the next state snapshot fast without tracing
|
||||||
msg, _ := tx.AsMessage(signer, block.BaseFee())
|
msg, _ := core.TransactionToMessage(tx, signer, block.BaseFee())
|
||||||
txctx := &Context{
|
txctx := &Context{
|
||||||
BlockHash: blockHash,
|
BlockHash: blockHash,
|
||||||
BlockNumber: block.Number(),
|
BlockNumber: block.Number(),
|
||||||
@ -671,7 +671,7 @@ func (api *API) traceBlockParallel(ctx context.Context, block *types.Block, stat
|
|||||||
defer pend.Done()
|
defer pend.Done()
|
||||||
// Fetch and execute the next transaction trace tasks
|
// Fetch and execute the next transaction trace tasks
|
||||||
for task := range jobs {
|
for task := range jobs {
|
||||||
msg, _ := txs[task.index].AsMessage(signer, block.BaseFee())
|
msg, _ := core.TransactionToMessage(txs[task.index], signer, block.BaseFee())
|
||||||
txctx := &Context{
|
txctx := &Context{
|
||||||
BlockHash: blockHash,
|
BlockHash: blockHash,
|
||||||
BlockNumber: block.Number(),
|
BlockNumber: block.Number(),
|
||||||
@ -702,10 +702,10 @@ txloop:
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Generate the next state snapshot fast without tracing
|
// Generate the next state snapshot fast without tracing
|
||||||
msg, _ := tx.AsMessage(signer, block.BaseFee())
|
msg, _ := core.TransactionToMessage(tx, signer, block.BaseFee())
|
||||||
statedb.SetTxContext(tx.Hash(), i)
|
statedb.SetTxContext(tx.Hash(), i)
|
||||||
vmenv := vm.NewEVM(blockCtx, core.NewEVMTxContext(msg), statedb, api.backend.ChainConfig(), vm.Config{})
|
vmenv := vm.NewEVM(blockCtx, core.NewEVMTxContext(msg), statedb, api.backend.ChainConfig(), vm.Config{})
|
||||||
if _, err := core.ApplyMessage(vmenv, msg, new(core.GasPool).AddGas(msg.Gas())); err != nil {
|
if _, err := core.ApplyMessage(vmenv, msg, new(core.GasPool).AddGas(msg.GasLimit)); err != nil {
|
||||||
failed = err
|
failed = err
|
||||||
break txloop
|
break txloop
|
||||||
}
|
}
|
||||||
@ -782,7 +782,7 @@ func (api *API) standardTraceBlockToFile(ctx context.Context, block *types.Block
|
|||||||
for i, tx := range block.Transactions() {
|
for i, tx := range block.Transactions() {
|
||||||
// Prepare the transaction for un-traced execution
|
// Prepare the transaction for un-traced execution
|
||||||
var (
|
var (
|
||||||
msg, _ = tx.AsMessage(signer, block.BaseFee())
|
msg, _ = core.TransactionToMessage(tx, signer, block.BaseFee())
|
||||||
txContext = core.NewEVMTxContext(msg)
|
txContext = core.NewEVMTxContext(msg)
|
||||||
vmConf vm.Config
|
vmConf vm.Config
|
||||||
dump *os.File
|
dump *os.File
|
||||||
@ -813,7 +813,7 @@ func (api *API) standardTraceBlockToFile(ctx context.Context, block *types.Block
|
|||||||
// Execute the transaction and flush any traces to disk
|
// Execute the transaction and flush any traces to disk
|
||||||
vmenv := vm.NewEVM(vmctx, txContext, statedb, chainConfig, vmConf)
|
vmenv := vm.NewEVM(vmctx, txContext, statedb, chainConfig, vmConf)
|
||||||
statedb.SetTxContext(tx.Hash(), i)
|
statedb.SetTxContext(tx.Hash(), i)
|
||||||
_, err = core.ApplyMessage(vmenv, msg, new(core.GasPool).AddGas(msg.Gas()))
|
_, err = core.ApplyMessage(vmenv, msg, new(core.GasPool).AddGas(msg.GasLimit))
|
||||||
if writer != nil {
|
if writer != nil {
|
||||||
writer.Flush()
|
writer.Flush()
|
||||||
}
|
}
|
||||||
@ -947,7 +947,7 @@ func (api *API) TraceCall(ctx context.Context, args ethapi.TransactionArgs, bloc
|
|||||||
// traceTx configures a new tracer according to the provided configuration, and
|
// traceTx configures a new tracer according to the provided configuration, and
|
||||||
// executes the given message in the provided environment. The return value will
|
// executes the given message in the provided environment. The return value will
|
||||||
// be tracer dependent.
|
// be tracer dependent.
|
||||||
func (api *API) traceTx(ctx context.Context, message core.Message, txctx *Context, vmctx vm.BlockContext, statedb *state.StateDB, config *TraceConfig) (interface{}, error) {
|
func (api *API) traceTx(ctx context.Context, message *core.Message, txctx *Context, vmctx vm.BlockContext, statedb *state.StateDB, config *TraceConfig) (interface{}, error) {
|
||||||
var (
|
var (
|
||||||
tracer Tracer
|
tracer Tracer
|
||||||
err error
|
err error
|
||||||
@ -993,7 +993,7 @@ func (api *API) traceTx(ctx context.Context, message core.Message, txctx *Contex
|
|||||||
|
|
||||||
// Call Prepare to clear out the statedb access list
|
// Call Prepare to clear out the statedb access list
|
||||||
statedb.SetTxContext(txctx.TxHash, txctx.TxIndex)
|
statedb.SetTxContext(txctx.TxHash, txctx.TxIndex)
|
||||||
if _, err = core.ApplyMessage(vmenv, message, new(core.GasPool).AddGas(message.Gas())); err != nil {
|
if _, err = core.ApplyMessage(vmenv, message, new(core.GasPool).AddGas(message.GasLimit)); err != nil {
|
||||||
return nil, fmt.Errorf("tracing failed: %w", err)
|
return nil, fmt.Errorf("tracing failed: %w", err)
|
||||||
}
|
}
|
||||||
return tracer.GetResult()
|
return tracer.GetResult()
|
||||||
|
@ -156,7 +156,7 @@ func (b *testBackend) StateAtBlock(ctx context.Context, block *types.Block, reex
|
|||||||
return statedb, release, nil
|
return statedb, release, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (b *testBackend) StateAtTransaction(ctx context.Context, block *types.Block, txIndex int, reexec uint64) (core.Message, vm.BlockContext, *state.StateDB, StateReleaseFunc, error) {
|
func (b *testBackend) StateAtTransaction(ctx context.Context, block *types.Block, txIndex int, reexec uint64) (*core.Message, vm.BlockContext, *state.StateDB, StateReleaseFunc, error) {
|
||||||
parent := b.chain.GetBlock(block.ParentHash(), block.NumberU64()-1)
|
parent := b.chain.GetBlock(block.ParentHash(), block.NumberU64()-1)
|
||||||
if parent == nil {
|
if parent == nil {
|
||||||
return nil, vm.BlockContext{}, nil, nil, errBlockNotFound
|
return nil, vm.BlockContext{}, nil, nil, errBlockNotFound
|
||||||
@ -171,7 +171,7 @@ func (b *testBackend) StateAtTransaction(ctx context.Context, block *types.Block
|
|||||||
// Recompute transactions up to the target index.
|
// Recompute transactions up to the target index.
|
||||||
signer := types.MakeSigner(b.chainConfig, block.Number())
|
signer := types.MakeSigner(b.chainConfig, block.Number())
|
||||||
for idx, tx := range block.Transactions() {
|
for idx, tx := range block.Transactions() {
|
||||||
msg, _ := tx.AsMessage(signer, block.BaseFee())
|
msg, _ := core.TransactionToMessage(tx, signer, block.BaseFee())
|
||||||
txContext := core.NewEVMTxContext(msg)
|
txContext := core.NewEVMTxContext(msg)
|
||||||
context := core.NewEVMBlockContext(block.Header(), b.chain, nil)
|
context := core.NewEVMBlockContext(block.Header(), b.chain, nil)
|
||||||
if idx == txIndex {
|
if idx == txIndex {
|
||||||
|
@ -145,7 +145,7 @@ func testCallTracer(tracerName string, dirPath string, t *testing.T) {
|
|||||||
t.Fatalf("failed to create call tracer: %v", err)
|
t.Fatalf("failed to create call tracer: %v", err)
|
||||||
}
|
}
|
||||||
evm := vm.NewEVM(context, txContext, statedb, test.Genesis.Config, vm.Config{Debug: true, Tracer: tracer})
|
evm := vm.NewEVM(context, txContext, statedb, test.Genesis.Config, vm.Config{Debug: true, Tracer: tracer})
|
||||||
msg, err := tx.AsMessage(signer, nil)
|
msg, err := core.TransactionToMessage(tx, signer, nil)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
t.Fatalf("failed to prepare transaction for tracing: %v", err)
|
t.Fatalf("failed to prepare transaction for tracing: %v", err)
|
||||||
}
|
}
|
||||||
@ -220,7 +220,7 @@ func benchTracer(tracerName string, test *callTracerTest, b *testing.B) {
|
|||||||
b.Fatalf("failed to parse testcase input: %v", err)
|
b.Fatalf("failed to parse testcase input: %v", err)
|
||||||
}
|
}
|
||||||
signer := types.MakeSigner(test.Genesis.Config, new(big.Int).SetUint64(uint64(test.Context.Number)))
|
signer := types.MakeSigner(test.Genesis.Config, new(big.Int).SetUint64(uint64(test.Context.Number)))
|
||||||
msg, err := tx.AsMessage(signer, nil)
|
msg, err := core.TransactionToMessage(tx, signer, nil)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
b.Fatalf("failed to prepare transaction for tracing: %v", err)
|
b.Fatalf("failed to prepare transaction for tracing: %v", err)
|
||||||
}
|
}
|
||||||
@ -314,7 +314,7 @@ func TestZeroValueToNotExitCall(t *testing.T) {
|
|||||||
t.Fatalf("failed to create call tracer: %v", err)
|
t.Fatalf("failed to create call tracer: %v", err)
|
||||||
}
|
}
|
||||||
evm := vm.NewEVM(context, txContext, statedb, params.MainnetChainConfig, vm.Config{Debug: true, Tracer: tracer})
|
evm := vm.NewEVM(context, txContext, statedb, params.MainnetChainConfig, vm.Config{Debug: true, Tracer: tracer})
|
||||||
msg, err := tx.AsMessage(signer, nil)
|
msg, err := core.TransactionToMessage(tx, signer, nil)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
t.Fatalf("failed to prepare transaction for tracing: %v", err)
|
t.Fatalf("failed to prepare transaction for tracing: %v", err)
|
||||||
}
|
}
|
||||||
|
@ -109,7 +109,7 @@ func flatCallTracerTestRunner(tracerName string, filename string, dirPath string
|
|||||||
}
|
}
|
||||||
evm := vm.NewEVM(context, txContext, statedb, test.Genesis.Config, vm.Config{Debug: true, Tracer: tracer})
|
evm := vm.NewEVM(context, txContext, statedb, test.Genesis.Config, vm.Config{Debug: true, Tracer: tracer})
|
||||||
|
|
||||||
msg, err := tx.AsMessage(signer, nil)
|
msg, err := core.TransactionToMessage(tx, signer, nil)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return fmt.Errorf("failed to prepare transaction for tracing: %v", err)
|
return fmt.Errorf("failed to prepare transaction for tracing: %v", err)
|
||||||
}
|
}
|
||||||
|
@ -115,7 +115,7 @@ func testPrestateDiffTracer(tracerName string, dirPath string, t *testing.T) {
|
|||||||
t.Fatalf("failed to create call tracer: %v", err)
|
t.Fatalf("failed to create call tracer: %v", err)
|
||||||
}
|
}
|
||||||
evm := vm.NewEVM(context, txContext, statedb, test.Genesis.Config, vm.Config{Debug: true, Tracer: tracer})
|
evm := vm.NewEVM(context, txContext, statedb, test.Genesis.Config, vm.Config{Debug: true, Tracer: tracer})
|
||||||
msg, err := tx.AsMessage(signer, nil)
|
msg, err := core.TransactionToMessage(tx, signer, nil)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
t.Fatalf("failed to prepare transaction for tracing: %v", err)
|
t.Fatalf("failed to prepare transaction for tracing: %v", err)
|
||||||
}
|
}
|
||||||
|
@ -88,7 +88,7 @@ func BenchmarkTransactionTrace(b *testing.B) {
|
|||||||
//EnableReturnData: false,
|
//EnableReturnData: false,
|
||||||
})
|
})
|
||||||
evm := vm.NewEVM(context, txContext, statedb, params.AllEthashProtocolChanges, vm.Config{Debug: true, Tracer: tracer})
|
evm := vm.NewEVM(context, txContext, statedb, params.AllEthashProtocolChanges, vm.Config{Debug: true, Tracer: tracer})
|
||||||
msg, err := tx.AsMessage(signer, nil)
|
msg, err := core.TransactionToMessage(tx, signer, nil)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
b.Fatalf("failed to prepare transaction for tracing: %v", err)
|
b.Fatalf("failed to prepare transaction for tracing: %v", err)
|
||||||
}
|
}
|
||||||
|
13
go.mod
13
go.mod
@ -36,7 +36,6 @@ require (
|
|||||||
github.com/gorilla/websocket v1.4.2
|
github.com/gorilla/websocket v1.4.2
|
||||||
github.com/graph-gophers/graphql-go v1.3.0
|
github.com/graph-gophers/graphql-go v1.3.0
|
||||||
github.com/hashicorp/go-bexpr v0.1.10
|
github.com/hashicorp/go-bexpr v0.1.10
|
||||||
github.com/holiman/big v0.0.0-20221017200358-a027dc42d04e
|
|
||||||
github.com/holiman/bloomfilter/v2 v2.0.3
|
github.com/holiman/bloomfilter/v2 v2.0.3
|
||||||
github.com/holiman/uint256 v1.2.0
|
github.com/holiman/uint256 v1.2.0
|
||||||
github.com/huin/goupnp v1.0.3
|
github.com/huin/goupnp v1.0.3
|
||||||
@ -47,6 +46,7 @@ require (
|
|||||||
github.com/jedisct1/go-minisign v0.0.0-20190909160543-45766022959e
|
github.com/jedisct1/go-minisign v0.0.0-20190909160543-45766022959e
|
||||||
github.com/julienschmidt/httprouter v1.3.0
|
github.com/julienschmidt/httprouter v1.3.0
|
||||||
github.com/karalabe/usb v0.0.2
|
github.com/karalabe/usb v0.0.2
|
||||||
|
github.com/kylelemons/godebug v1.1.0
|
||||||
github.com/mattn/go-colorable v0.1.13
|
github.com/mattn/go-colorable v0.1.13
|
||||||
github.com/mattn/go-isatty v0.0.16
|
github.com/mattn/go-isatty v0.0.16
|
||||||
github.com/naoina/toml v0.1.2-0.20170918210437-9fafd6967416
|
github.com/naoina/toml v0.1.2-0.20170918210437-9fafd6967416
|
||||||
@ -64,10 +64,10 @@ require (
|
|||||||
golang.org/x/crypto v0.1.0
|
golang.org/x/crypto v0.1.0
|
||||||
golang.org/x/exp v0.0.0-20230206171751-46f607a40771
|
golang.org/x/exp v0.0.0-20230206171751-46f607a40771
|
||||||
golang.org/x/sync v0.1.0
|
golang.org/x/sync v0.1.0
|
||||||
golang.org/x/sys v0.5.0
|
golang.org/x/sys v0.6.0
|
||||||
golang.org/x/text v0.7.0
|
golang.org/x/text v0.8.0
|
||||||
golang.org/x/time v0.0.0-20220922220347-f3bd1da661af
|
golang.org/x/time v0.0.0-20220922220347-f3bd1da661af
|
||||||
golang.org/x/tools v0.2.0
|
golang.org/x/tools v0.6.0
|
||||||
gopkg.in/natefinch/npipe.v2 v2.0.0-20160621034901-c1b8fa8bdcce
|
gopkg.in/natefinch/npipe.v2 v2.0.0-20160621034901-c1b8fa8bdcce
|
||||||
)
|
)
|
||||||
|
|
||||||
@ -99,7 +99,6 @@ require (
|
|||||||
github.com/klauspost/compress v1.15.15 // indirect
|
github.com/klauspost/compress v1.15.15 // indirect
|
||||||
github.com/kr/pretty v0.3.1 // indirect
|
github.com/kr/pretty v0.3.1 // indirect
|
||||||
github.com/kr/text v0.2.0 // indirect
|
github.com/kr/text v0.2.0 // indirect
|
||||||
github.com/kylelemons/godebug v1.1.0 // indirect
|
|
||||||
github.com/mattn/go-runewidth v0.0.9 // indirect
|
github.com/mattn/go-runewidth v0.0.9 // indirect
|
||||||
github.com/matttproud/golang_protobuf_extensions v1.0.4 // indirect
|
github.com/matttproud/golang_protobuf_extensions v1.0.4 // indirect
|
||||||
github.com/mitchellh/mapstructure v1.4.1 // indirect
|
github.com/mitchellh/mapstructure v1.4.1 // indirect
|
||||||
@ -118,8 +117,8 @@ require (
|
|||||||
github.com/tklauser/go-sysconf v0.3.5 // indirect
|
github.com/tklauser/go-sysconf v0.3.5 // indirect
|
||||||
github.com/tklauser/numcpus v0.2.2 // indirect
|
github.com/tklauser/numcpus v0.2.2 // indirect
|
||||||
github.com/xrash/smetrics v0.0.0-20201216005158-039620a65673 // indirect
|
github.com/xrash/smetrics v0.0.0-20201216005158-039620a65673 // indirect
|
||||||
golang.org/x/mod v0.6.0 // indirect
|
golang.org/x/mod v0.8.0 // indirect
|
||||||
golang.org/x/net v0.4.0 // indirect
|
golang.org/x/net v0.8.0 // indirect
|
||||||
golang.org/x/xerrors v0.0.0-20220517211312-f3a8303e98df // indirect
|
golang.org/x/xerrors v0.0.0-20220517211312-f3a8303e98df // indirect
|
||||||
google.golang.org/protobuf v1.28.1 // indirect
|
google.golang.org/protobuf v1.28.1 // indirect
|
||||||
gopkg.in/yaml.v2 v2.4.0 // indirect
|
gopkg.in/yaml.v2 v2.4.0 // indirect
|
||||||
|
23
go.sum
23
go.sum
@ -291,11 +291,8 @@ github.com/hashicorp/go-bexpr v0.1.10 h1:9kuI5PFotCboP3dkDYFr/wi0gg0QVbSNz5oFRpx
|
|||||||
github.com/hashicorp/go-bexpr v0.1.10/go.mod h1:oxlubA2vC/gFVfX1A6JGp7ls7uCDlfJn732ehYYg+g0=
|
github.com/hashicorp/go-bexpr v0.1.10/go.mod h1:oxlubA2vC/gFVfX1A6JGp7ls7uCDlfJn732ehYYg+g0=
|
||||||
github.com/hashicorp/go-version v1.2.0/go.mod h1:fltr4n8CU8Ke44wwGCBoEymUuxUHl09ZGVZPK5anwXA=
|
github.com/hashicorp/go-version v1.2.0/go.mod h1:fltr4n8CU8Ke44wwGCBoEymUuxUHl09ZGVZPK5anwXA=
|
||||||
github.com/hashicorp/golang-lru v0.5.0/go.mod h1:/m3WP610KZHVQ1SGc6re/UDhFvYD7pJ4Ao+sR/qLZy8=
|
github.com/hashicorp/golang-lru v0.5.0/go.mod h1:/m3WP610KZHVQ1SGc6re/UDhFvYD7pJ4Ao+sR/qLZy8=
|
||||||
github.com/hashicorp/golang-lru v0.5.1 h1:0hERBMJE1eitiLkihrMvRVBYAkpHzc/J3QdDN+dAcgU=
|
|
||||||
github.com/hashicorp/golang-lru v0.5.1/go.mod h1:/m3WP610KZHVQ1SGc6re/UDhFvYD7pJ4Ao+sR/qLZy8=
|
github.com/hashicorp/golang-lru v0.5.1/go.mod h1:/m3WP610KZHVQ1SGc6re/UDhFvYD7pJ4Ao+sR/qLZy8=
|
||||||
github.com/hashicorp/hcl v1.0.0/go.mod h1:E5yfLk+7swimpb2L/Alb/PJmXilQ/rhwaUYs4T20WEQ=
|
github.com/hashicorp/hcl v1.0.0/go.mod h1:E5yfLk+7swimpb2L/Alb/PJmXilQ/rhwaUYs4T20WEQ=
|
||||||
github.com/holiman/big v0.0.0-20221017200358-a027dc42d04e h1:pIYdhNkDh+YENVNi3gto8n9hAmRxKxoar0iE6BLucjw=
|
|
||||||
github.com/holiman/big v0.0.0-20221017200358-a027dc42d04e/go.mod h1:j9cQbcqHQujT0oKJ38PylVfqohClLr3CvDC+Qcg+lhU=
|
|
||||||
github.com/holiman/bloomfilter/v2 v2.0.3 h1:73e0e/V0tCydx14a0SCYS/EWCxgwLZ18CZcZKVu0fao=
|
github.com/holiman/bloomfilter/v2 v2.0.3 h1:73e0e/V0tCydx14a0SCYS/EWCxgwLZ18CZcZKVu0fao=
|
||||||
github.com/holiman/bloomfilter/v2 v2.0.3/go.mod h1:zpoh+gs7qcpqrHr3dB55AMiJwo0iURXE7ZOP9L9hSkA=
|
github.com/holiman/bloomfilter/v2 v2.0.3/go.mod h1:zpoh+gs7qcpqrHr3dB55AMiJwo0iURXE7ZOP9L9hSkA=
|
||||||
github.com/holiman/uint256 v1.2.0 h1:gpSYcPLWGv4sG43I2mVLiDZCNDh/EpGjSk8tmtxitHM=
|
github.com/holiman/uint256 v1.2.0 h1:gpSYcPLWGv4sG43I2mVLiDZCNDh/EpGjSk8tmtxitHM=
|
||||||
@ -628,8 +625,8 @@ golang.org/x/mod v0.1.1-0.20191105210325-c90efee705ee/go.mod h1:QqPTAvyqsEbceGzB
|
|||||||
golang.org/x/mod v0.2.0/go.mod h1:s0Qsj1ACt9ePp/hMypM3fl4fZqREWJwdYDEqhRiZZUA=
|
golang.org/x/mod v0.2.0/go.mod h1:s0Qsj1ACt9ePp/hMypM3fl4fZqREWJwdYDEqhRiZZUA=
|
||||||
golang.org/x/mod v0.3.0/go.mod h1:s0Qsj1ACt9ePp/hMypM3fl4fZqREWJwdYDEqhRiZZUA=
|
golang.org/x/mod v0.3.0/go.mod h1:s0Qsj1ACt9ePp/hMypM3fl4fZqREWJwdYDEqhRiZZUA=
|
||||||
golang.org/x/mod v0.4.2/go.mod h1:s0Qsj1ACt9ePp/hMypM3fl4fZqREWJwdYDEqhRiZZUA=
|
golang.org/x/mod v0.4.2/go.mod h1:s0Qsj1ACt9ePp/hMypM3fl4fZqREWJwdYDEqhRiZZUA=
|
||||||
golang.org/x/mod v0.6.0 h1:b9gGHsz9/HhJ3HF5DHQytPpuwocVTChQJK3AvoLRD5I=
|
golang.org/x/mod v0.8.0 h1:LUYupSeNrTNCGzR/hVBk2NHZO4hXcVaW1k4Qx7rjPx8=
|
||||||
golang.org/x/mod v0.6.0/go.mod h1:4mET923SAdbXp2ki8ey+zGs1SLqsuM2Y0uvdZR/fUNI=
|
golang.org/x/mod v0.8.0/go.mod h1:iBbtSCu2XBx23ZKBPSOrRkjjQPZFPuis4dIYUhu/chs=
|
||||||
golang.org/x/net v0.0.0-20180724234803-3673e40ba225/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4=
|
golang.org/x/net v0.0.0-20180724234803-3673e40ba225/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4=
|
||||||
golang.org/x/net v0.0.0-20180826012351-8a410e7b638d/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4=
|
golang.org/x/net v0.0.0-20180826012351-8a410e7b638d/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4=
|
||||||
golang.org/x/net v0.0.0-20180906233101-161cd47e91fd/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4=
|
golang.org/x/net v0.0.0-20180906233101-161cd47e91fd/go.mod h1:mL1N/T3taQHkDXs73rZJwtUhF3w3ftmwwsq0BUmARs4=
|
||||||
@ -661,8 +658,8 @@ golang.org/x/net v0.0.0-20210405180319-a5a99cb37ef4/go.mod h1:p54w0d4576C0XHj96b
|
|||||||
golang.org/x/net v0.0.0-20210610132358-84b48f89b13b/go.mod h1:9nx3DQGgdP8bBQD5qxJ1jj9UTztislL4KSBs9R2vV5Y=
|
golang.org/x/net v0.0.0-20210610132358-84b48f89b13b/go.mod h1:9nx3DQGgdP8bBQD5qxJ1jj9UTztislL4KSBs9R2vV5Y=
|
||||||
golang.org/x/net v0.0.0-20210805182204-aaa1db679c0d/go.mod h1:9nx3DQGgdP8bBQD5qxJ1jj9UTztislL4KSBs9R2vV5Y=
|
golang.org/x/net v0.0.0-20210805182204-aaa1db679c0d/go.mod h1:9nx3DQGgdP8bBQD5qxJ1jj9UTztislL4KSBs9R2vV5Y=
|
||||||
golang.org/x/net v0.0.0-20211008194852-3b03d305991f/go.mod h1:9nx3DQGgdP8bBQD5qxJ1jj9UTztislL4KSBs9R2vV5Y=
|
golang.org/x/net v0.0.0-20211008194852-3b03d305991f/go.mod h1:9nx3DQGgdP8bBQD5qxJ1jj9UTztislL4KSBs9R2vV5Y=
|
||||||
golang.org/x/net v0.4.0 h1:Q5QPcMlvfxFTAPV0+07Xz/MpK9NTXu2VDUuy0FeMfaU=
|
golang.org/x/net v0.8.0 h1:Zrh2ngAOFYneWTAIAPethzeaQLuHwhuBkuV6ZiRnUaQ=
|
||||||
golang.org/x/net v0.4.0/go.mod h1:MBQ8lrhLObU/6UmLb4fmbmk5OcyYmqtbGd/9yIeKjEE=
|
golang.org/x/net v0.8.0/go.mod h1:QVkue5JL9kW//ek3r6jTKnTFis1tRmNAW2P1shuFdJc=
|
||||||
golang.org/x/oauth2 v0.0.0-20180821212333-d2e6202438be/go.mod h1:N/0e6XlmueqKjAGxoOufVs8QHGRruUQn6yWY3a++T0U=
|
golang.org/x/oauth2 v0.0.0-20180821212333-d2e6202438be/go.mod h1:N/0e6XlmueqKjAGxoOufVs8QHGRruUQn6yWY3a++T0U=
|
||||||
golang.org/x/oauth2 v0.0.0-20190226205417-e64efc72b421/go.mod h1:gOpvHmFTYa4IltrdGE7lF6nIHvwfUNPOp7c8zoXwtLw=
|
golang.org/x/oauth2 v0.0.0-20190226205417-e64efc72b421/go.mod h1:gOpvHmFTYa4IltrdGE7lF6nIHvwfUNPOp7c8zoXwtLw=
|
||||||
golang.org/x/oauth2 v0.0.0-20190604053449-0f29369cfe45/go.mod h1:gOpvHmFTYa4IltrdGE7lF6nIHvwfUNPOp7c8zoXwtLw=
|
golang.org/x/oauth2 v0.0.0-20190604053449-0f29369cfe45/go.mod h1:gOpvHmFTYa4IltrdGE7lF6nIHvwfUNPOp7c8zoXwtLw=
|
||||||
@ -724,8 +721,8 @@ golang.org/x/sys v0.0.0-20211020174200-9d6173849985/go.mod h1:oPkhp1MJrh7nUepCBc
|
|||||||
golang.org/x/sys v0.0.0-20220209214540-3681064d5158/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg=
|
golang.org/x/sys v0.0.0-20220209214540-3681064d5158/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg=
|
||||||
golang.org/x/sys v0.0.0-20220811171246-fbc7d0a398ab/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg=
|
golang.org/x/sys v0.0.0-20220811171246-fbc7d0a398ab/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg=
|
||||||
golang.org/x/sys v0.0.0-20220908164124-27713097b956/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg=
|
golang.org/x/sys v0.0.0-20220908164124-27713097b956/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg=
|
||||||
golang.org/x/sys v0.5.0 h1:MUK/U/4lj1t1oPg0HfuXDN/Z1wv31ZJ/YcPiGccS4DU=
|
golang.org/x/sys v0.6.0 h1:MVltZSvRTcU2ljQOhs94SXPftV6DCNnZViHeQps87pQ=
|
||||||
golang.org/x/sys v0.5.0/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg=
|
golang.org/x/sys v0.6.0/go.mod h1:oPkhp1MJrh7nUepCBck5+mAzfO9JrbApNNgaTdGDITg=
|
||||||
golang.org/x/term v0.0.0-20201117132131-f5c789dd3221/go.mod h1:Nr5EML6q2oocZ2LXRh80K7BxOlk5/8JxuGnuhpl+muw=
|
golang.org/x/term v0.0.0-20201117132131-f5c789dd3221/go.mod h1:Nr5EML6q2oocZ2LXRh80K7BxOlk5/8JxuGnuhpl+muw=
|
||||||
golang.org/x/term v0.0.0-20201126162022-7de9c90e9dd1/go.mod h1:bj7SfCRtBDWHUb9snDiAeCFNEtKQo2Wmx5Cou7ajbmo=
|
golang.org/x/term v0.0.0-20201126162022-7de9c90e9dd1/go.mod h1:bj7SfCRtBDWHUb9snDiAeCFNEtKQo2Wmx5Cou7ajbmo=
|
||||||
golang.org/x/text v0.3.0/go.mod h1:NqM8EUOU14njkJ3fqMW+pc6Ldnwhi/IjpwHt7yyuwOQ=
|
golang.org/x/text v0.3.0/go.mod h1:NqM8EUOU14njkJ3fqMW+pc6Ldnwhi/IjpwHt7yyuwOQ=
|
||||||
@ -736,8 +733,8 @@ golang.org/x/text v0.3.4/go.mod h1:5Zoc/QRtKVWzQhOtBMvqHzDpF6irO9z98xDceosuGiQ=
|
|||||||
golang.org/x/text v0.3.5/go.mod h1:5Zoc/QRtKVWzQhOtBMvqHzDpF6irO9z98xDceosuGiQ=
|
golang.org/x/text v0.3.5/go.mod h1:5Zoc/QRtKVWzQhOtBMvqHzDpF6irO9z98xDceosuGiQ=
|
||||||
golang.org/x/text v0.3.6/go.mod h1:5Zoc/QRtKVWzQhOtBMvqHzDpF6irO9z98xDceosuGiQ=
|
golang.org/x/text v0.3.6/go.mod h1:5Zoc/QRtKVWzQhOtBMvqHzDpF6irO9z98xDceosuGiQ=
|
||||||
golang.org/x/text v0.3.7/go.mod h1:u+2+/6zg+i71rQMx5EYifcz6MCKuco9NR6JIITiCfzQ=
|
golang.org/x/text v0.3.7/go.mod h1:u+2+/6zg+i71rQMx5EYifcz6MCKuco9NR6JIITiCfzQ=
|
||||||
golang.org/x/text v0.7.0 h1:4BRB4x83lYWy72KwLD/qYDuTu7q9PjSagHvijDw7cLo=
|
golang.org/x/text v0.8.0 h1:57P1ETyNKtuIjB4SRd15iJxuhj8Gc416Y78H3qgMh68=
|
||||||
golang.org/x/text v0.7.0/go.mod h1:mrYo+phRRbMaCq/xk9113O4dZlRixOauAjOtrjsXDZ8=
|
golang.org/x/text v0.8.0/go.mod h1:e1OnstbJyHTd6l/uOt8jFFHp6TRDWZR/bV3emEE/zU8=
|
||||||
golang.org/x/time v0.0.0-20181108054448-85acf8d2951c/go.mod h1:tRJNPiyCQ0inRvYxbN9jk5I+vvW/OXSQhTDSoE431IQ=
|
golang.org/x/time v0.0.0-20181108054448-85acf8d2951c/go.mod h1:tRJNPiyCQ0inRvYxbN9jk5I+vvW/OXSQhTDSoE431IQ=
|
||||||
golang.org/x/time v0.0.0-20190308202827-9d24e82272b4/go.mod h1:tRJNPiyCQ0inRvYxbN9jk5I+vvW/OXSQhTDSoE431IQ=
|
golang.org/x/time v0.0.0-20190308202827-9d24e82272b4/go.mod h1:tRJNPiyCQ0inRvYxbN9jk5I+vvW/OXSQhTDSoE431IQ=
|
||||||
golang.org/x/time v0.0.0-20201208040808-7e3f01d25324/go.mod h1:tRJNPiyCQ0inRvYxbN9jk5I+vvW/OXSQhTDSoE431IQ=
|
golang.org/x/time v0.0.0-20201208040808-7e3f01d25324/go.mod h1:tRJNPiyCQ0inRvYxbN9jk5I+vvW/OXSQhTDSoE431IQ=
|
||||||
@ -776,8 +773,8 @@ golang.org/x/tools v0.0.0-20200130002326-2f3ba24bd6e7/go.mod h1:TB2adYChydJhpapK
|
|||||||
golang.org/x/tools v0.0.0-20200619180055-7c47624df98f/go.mod h1:EkVYQZoAsY45+roYkvgYkIh4xh/qjgUK9TdY2XT94GE=
|
golang.org/x/tools v0.0.0-20200619180055-7c47624df98f/go.mod h1:EkVYQZoAsY45+roYkvgYkIh4xh/qjgUK9TdY2XT94GE=
|
||||||
golang.org/x/tools v0.0.0-20210106214847-113979e3529a/go.mod h1:emZCQorbCU4vsT4fOWvOPXz4eW1wZW4PmDk9uLelYpA=
|
golang.org/x/tools v0.0.0-20210106214847-113979e3529a/go.mod h1:emZCQorbCU4vsT4fOWvOPXz4eW1wZW4PmDk9uLelYpA=
|
||||||
golang.org/x/tools v0.1.3/go.mod h1:o0xws9oXOQQZyjljx8fwUC0k7L1pTE6eaCbjGeHmOkk=
|
golang.org/x/tools v0.1.3/go.mod h1:o0xws9oXOQQZyjljx8fwUC0k7L1pTE6eaCbjGeHmOkk=
|
||||||
golang.org/x/tools v0.2.0 h1:G6AHpWxTMGY1KyEYoAQ5WTtIekUUvDNjan3ugu60JvE=
|
golang.org/x/tools v0.6.0 h1:BOw41kyTf3PuCW1pVQf8+Cyg8pMlkYB1oo9iJ6D/lKM=
|
||||||
golang.org/x/tools v0.2.0/go.mod h1:y4OqIKeOV/fWJetJ8bXPU1sEVniLMIyDAZWeHdV+NTA=
|
golang.org/x/tools v0.6.0/go.mod h1:Xwgl3UAJ/d3gWutnCtw505GrjyAbvKui8lOU390QaIU=
|
||||||
golang.org/x/xerrors v0.0.0-20190717185122-a985d3407aa7/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0=
|
golang.org/x/xerrors v0.0.0-20190717185122-a985d3407aa7/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0=
|
||||||
golang.org/x/xerrors v0.0.0-20191011141410-1b5146add898/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0=
|
golang.org/x/xerrors v0.0.0-20191011141410-1b5146add898/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0=
|
||||||
golang.org/x/xerrors v0.0.0-20191204190536-9bdfabe68543/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0=
|
golang.org/x/xerrors v0.0.0-20191204190536-9bdfabe68543/go.mod h1:I/5z698sn9Ka8TeJc9MKroUUfqBBauWjQqLJ2OPfmY0=
|
||||||
|
@ -1005,7 +1005,7 @@ func DoCall(ctx context.Context, b Backend, args TransactionArgs, blockNrOrHash
|
|||||||
return nil, fmt.Errorf("execution aborted (timeout = %v)", timeout)
|
return nil, fmt.Errorf("execution aborted (timeout = %v)", timeout)
|
||||||
}
|
}
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return result, fmt.Errorf("err: %w (supplied gas %d)", err, msg.Gas())
|
return result, fmt.Errorf("err: %w (supplied gas %d)", err, msg.GasLimit)
|
||||||
}
|
}
|
||||||
return result, nil
|
return result, nil
|
||||||
}
|
}
|
||||||
@ -1478,7 +1478,7 @@ func AccessList(ctx context.Context, b Backend, blockNrOrHash rpc.BlockNumberOrH
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, 0, nil, err
|
return nil, 0, nil, err
|
||||||
}
|
}
|
||||||
res, err := core.ApplyMessage(vmenv, msg, new(core.GasPool).AddGas(msg.Gas()))
|
res, err := core.ApplyMessage(vmenv, msg, new(core.GasPool).AddGas(msg.GasLimit))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, 0, nil, fmt.Errorf("failed to apply transaction: %v err: %v", args.toTransaction().Hash(), err)
|
return nil, 0, nil, fmt.Errorf("failed to apply transaction: %v err: %v", args.toTransaction().Hash(), err)
|
||||||
}
|
}
|
||||||
@ -1626,7 +1626,7 @@ func (s *TransactionAPI) GetTransactionReceipt(ctx context.Context, hash common.
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
if len(receipts) <= int(index) {
|
if uint64(len(receipts)) <= index {
|
||||||
return nil, nil
|
return nil, nil
|
||||||
}
|
}
|
||||||
receipt := receipts[index]
|
receipt := receipts[index]
|
||||||
|
@ -68,7 +68,7 @@ type Backend interface {
|
|||||||
PendingBlockAndReceipts() (*types.Block, types.Receipts)
|
PendingBlockAndReceipts() (*types.Block, types.Receipts)
|
||||||
GetReceipts(ctx context.Context, hash common.Hash) (types.Receipts, error)
|
GetReceipts(ctx context.Context, hash common.Hash) (types.Receipts, error)
|
||||||
GetTd(ctx context.Context, hash common.Hash) *big.Int
|
GetTd(ctx context.Context, hash common.Hash) *big.Int
|
||||||
GetEVM(ctx context.Context, msg core.Message, state *state.StateDB, header *types.Header, vmConfig *vm.Config) (*vm.EVM, func() error, error)
|
GetEVM(ctx context.Context, msg *core.Message, state *state.StateDB, header *types.Header, vmConfig *vm.Config) (*vm.EVM, func() error, error)
|
||||||
SubscribeChainEvent(ch chan<- core.ChainEvent) event.Subscription
|
SubscribeChainEvent(ch chan<- core.ChainEvent) event.Subscription
|
||||||
SubscribeChainHeadEvent(ch chan<- core.ChainHeadEvent) event.Subscription
|
SubscribeChainHeadEvent(ch chan<- core.ChainHeadEvent) event.Subscription
|
||||||
SubscribeChainSideEvent(ch chan<- core.ChainSideEvent) event.Subscription
|
SubscribeChainSideEvent(ch chan<- core.ChainSideEvent) event.Subscription
|
||||||
|
@ -26,6 +26,7 @@ import (
|
|||||||
"github.com/ethereum/go-ethereum/common"
|
"github.com/ethereum/go-ethereum/common"
|
||||||
"github.com/ethereum/go-ethereum/common/hexutil"
|
"github.com/ethereum/go-ethereum/common/hexutil"
|
||||||
"github.com/ethereum/go-ethereum/common/math"
|
"github.com/ethereum/go-ethereum/common/math"
|
||||||
|
"github.com/ethereum/go-ethereum/core"
|
||||||
"github.com/ethereum/go-ethereum/core/types"
|
"github.com/ethereum/go-ethereum/core/types"
|
||||||
"github.com/ethereum/go-ethereum/log"
|
"github.com/ethereum/go-ethereum/log"
|
||||||
"github.com/ethereum/go-ethereum/rpc"
|
"github.com/ethereum/go-ethereum/rpc"
|
||||||
@ -199,10 +200,10 @@ func (args *TransactionArgs) setLondonFeeDefaults(ctx context.Context, head *typ
|
|||||||
// ToMessage converts the transaction arguments to the Message type used by the
|
// ToMessage converts the transaction arguments to the Message type used by the
|
||||||
// core evm. This method is used in calls and traces that do not require a real
|
// core evm. This method is used in calls and traces that do not require a real
|
||||||
// live transaction.
|
// live transaction.
|
||||||
func (args *TransactionArgs) ToMessage(globalGasCap uint64, baseFee *big.Int) (types.Message, error) {
|
func (args *TransactionArgs) ToMessage(globalGasCap uint64, baseFee *big.Int) (*core.Message, error) {
|
||||||
// Reject invalid combinations of pre- and post-1559 fee styles
|
// Reject invalid combinations of pre- and post-1559 fee styles
|
||||||
if args.GasPrice != nil && (args.MaxFeePerGas != nil || args.MaxPriorityFeePerGas != nil) {
|
if args.GasPrice != nil && (args.MaxFeePerGas != nil || args.MaxPriorityFeePerGas != nil) {
|
||||||
return types.Message{}, errors.New("both gasPrice and (maxFeePerGas or maxPriorityFeePerGas) specified")
|
return nil, errors.New("both gasPrice and (maxFeePerGas or maxPriorityFeePerGas) specified")
|
||||||
}
|
}
|
||||||
// Set sender address or use zero address if none specified.
|
// Set sender address or use zero address if none specified.
|
||||||
addr := args.from()
|
addr := args.from()
|
||||||
@ -263,7 +264,18 @@ func (args *TransactionArgs) ToMessage(globalGasCap uint64, baseFee *big.Int) (t
|
|||||||
if args.AccessList != nil {
|
if args.AccessList != nil {
|
||||||
accessList = *args.AccessList
|
accessList = *args.AccessList
|
||||||
}
|
}
|
||||||
msg := types.NewMessage(addr, args.To, 0, value, gas, gasPrice, gasFeeCap, gasTipCap, data, accessList, true)
|
msg := &core.Message{
|
||||||
|
From: addr,
|
||||||
|
To: args.To,
|
||||||
|
Value: value,
|
||||||
|
GasLimit: gas,
|
||||||
|
GasPrice: gasPrice,
|
||||||
|
GasFeeCap: gasFeeCap,
|
||||||
|
GasTipCap: gasTipCap,
|
||||||
|
Data: data,
|
||||||
|
AccessList: accessList,
|
||||||
|
SkipAccountChecks: true,
|
||||||
|
}
|
||||||
return msg, nil
|
return msg, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -305,7 +305,7 @@ func (b *backendMock) GetLogs(ctx context.Context, blockHash common.Hash, number
|
|||||||
return nil, nil
|
return nil, nil
|
||||||
}
|
}
|
||||||
func (b *backendMock) GetTd(ctx context.Context, hash common.Hash) *big.Int { return nil }
|
func (b *backendMock) GetTd(ctx context.Context, hash common.Hash) *big.Int { return nil }
|
||||||
func (b *backendMock) GetEVM(ctx context.Context, msg core.Message, state *state.StateDB, header *types.Header, vmConfig *vm.Config) (*vm.EVM, func() error, error) {
|
func (b *backendMock) GetEVM(ctx context.Context, msg *core.Message, state *state.StateDB, header *types.Header, vmConfig *vm.Config) (*vm.EVM, func() error, error) {
|
||||||
return nil, nil, nil
|
return nil, nil, nil
|
||||||
}
|
}
|
||||||
func (b *backendMock) SubscribeChainEvent(ch chan<- core.ChainEvent) event.Subscription { return nil }
|
func (b *backendMock) SubscribeChainEvent(ch chan<- core.ChainEvent) event.Subscription { return nil }
|
||||||
|
@ -184,7 +184,7 @@ func (b *LesApiBackend) GetTd(ctx context.Context, hash common.Hash) *big.Int {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (b *LesApiBackend) GetEVM(ctx context.Context, msg core.Message, state *state.StateDB, header *types.Header, vmConfig *vm.Config) (*vm.EVM, func() error, error) {
|
func (b *LesApiBackend) GetEVM(ctx context.Context, msg *core.Message, state *state.StateDB, header *types.Header, vmConfig *vm.Config) (*vm.EVM, func() error, error) {
|
||||||
if vmConfig == nil {
|
if vmConfig == nil {
|
||||||
vmConfig = new(vm.Config)
|
vmConfig = new(vm.Config)
|
||||||
}
|
}
|
||||||
@ -330,6 +330,6 @@ func (b *LesApiBackend) StateAtBlock(ctx context.Context, block *types.Block, re
|
|||||||
return b.eth.stateAtBlock(ctx, block, reexec)
|
return b.eth.stateAtBlock(ctx, block, reexec)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (b *LesApiBackend) StateAtTransaction(ctx context.Context, block *types.Block, txIndex int, reexec uint64) (core.Message, vm.BlockContext, *state.StateDB, tracers.StateReleaseFunc, error) {
|
func (b *LesApiBackend) StateAtTransaction(ctx context.Context, block *types.Block, txIndex int, reexec uint64) (*core.Message, vm.BlockContext, *state.StateDB, tracers.StateReleaseFunc, error) {
|
||||||
return b.eth.stateAtTransaction(ctx, block, txIndex, reexec)
|
return b.eth.stateAtTransaction(ctx, block, txIndex, reexec)
|
||||||
}
|
}
|
||||||
|
@ -116,12 +116,6 @@ func TestOdrContractCallLes2(t *testing.T) { testOdr(t, 2, 2, true, odrContractC
|
|||||||
func TestOdrContractCallLes3(t *testing.T) { testOdr(t, 3, 2, true, odrContractCall) }
|
func TestOdrContractCallLes3(t *testing.T) { testOdr(t, 3, 2, true, odrContractCall) }
|
||||||
func TestOdrContractCallLes4(t *testing.T) { testOdr(t, 4, 2, true, odrContractCall) }
|
func TestOdrContractCallLes4(t *testing.T) { testOdr(t, 4, 2, true, odrContractCall) }
|
||||||
|
|
||||||
type callmsg struct {
|
|
||||||
types.Message
|
|
||||||
}
|
|
||||||
|
|
||||||
func (callmsg) CheckNonce() bool { return false }
|
|
||||||
|
|
||||||
func odrContractCall(ctx context.Context, db ethdb.Database, config *params.ChainConfig, bc *core.BlockChain, lc *light.LightChain, bhash common.Hash) []byte {
|
func odrContractCall(ctx context.Context, db ethdb.Database, config *params.ChainConfig, bc *core.BlockChain, lc *light.LightChain, bhash common.Hash) []byte {
|
||||||
data := common.Hex2Bytes("60CD26850000000000000000000000000000000000000000000000000000000000000000")
|
data := common.Hex2Bytes("60CD26850000000000000000000000000000000000000000000000000000000000000000")
|
||||||
|
|
||||||
@ -136,7 +130,17 @@ func odrContractCall(ctx context.Context, db ethdb.Database, config *params.Chai
|
|||||||
from := statedb.GetOrNewStateObject(bankAddr)
|
from := statedb.GetOrNewStateObject(bankAddr)
|
||||||
from.SetBalance(math.MaxBig256)
|
from.SetBalance(math.MaxBig256)
|
||||||
|
|
||||||
msg := callmsg{types.NewMessage(from.Address(), &testContractAddr, 0, new(big.Int), 100000, big.NewInt(params.InitialBaseFee), big.NewInt(params.InitialBaseFee), new(big.Int), data, nil, true)}
|
msg := &core.Message{
|
||||||
|
From: from.Address(),
|
||||||
|
To: &testContractAddr,
|
||||||
|
Value: new(big.Int),
|
||||||
|
GasLimit: 100000,
|
||||||
|
GasPrice: big.NewInt(params.InitialBaseFee),
|
||||||
|
GasFeeCap: big.NewInt(params.InitialBaseFee),
|
||||||
|
GasTipCap: new(big.Int),
|
||||||
|
Data: data,
|
||||||
|
SkipAccountChecks: true,
|
||||||
|
}
|
||||||
|
|
||||||
context := core.NewEVMBlockContext(header, bc, nil)
|
context := core.NewEVMBlockContext(header, bc, nil)
|
||||||
txContext := core.NewEVMTxContext(msg)
|
txContext := core.NewEVMTxContext(msg)
|
||||||
@ -151,7 +155,17 @@ func odrContractCall(ctx context.Context, db ethdb.Database, config *params.Chai
|
|||||||
header := lc.GetHeaderByHash(bhash)
|
header := lc.GetHeaderByHash(bhash)
|
||||||
state := light.NewState(ctx, header, lc.Odr())
|
state := light.NewState(ctx, header, lc.Odr())
|
||||||
state.SetBalance(bankAddr, math.MaxBig256)
|
state.SetBalance(bankAddr, math.MaxBig256)
|
||||||
msg := callmsg{types.NewMessage(bankAddr, &testContractAddr, 0, new(big.Int), 100000, big.NewInt(params.InitialBaseFee), big.NewInt(params.InitialBaseFee), new(big.Int), data, nil, true)}
|
msg := &core.Message{
|
||||||
|
From: bankAddr,
|
||||||
|
To: &testContractAddr,
|
||||||
|
Value: new(big.Int),
|
||||||
|
GasLimit: 100000,
|
||||||
|
GasPrice: big.NewInt(params.InitialBaseFee),
|
||||||
|
GasFeeCap: big.NewInt(params.InitialBaseFee),
|
||||||
|
GasTipCap: new(big.Int),
|
||||||
|
Data: data,
|
||||||
|
SkipAccountChecks: true,
|
||||||
|
}
|
||||||
context := core.NewEVMBlockContext(header, lc, nil)
|
context := core.NewEVMBlockContext(header, lc, nil)
|
||||||
txContext := core.NewEVMTxContext(msg)
|
txContext := core.NewEVMTxContext(msg)
|
||||||
vmenv := vm.NewEVM(context, txContext, state, config, vm.Config{NoBaseFee: true})
|
vmenv := vm.NewEVM(context, txContext, state, config, vm.Config{NoBaseFee: true})
|
||||||
|
@ -39,7 +39,7 @@ func (leth *LightEthereum) stateAtBlock(ctx context.Context, block *types.Block,
|
|||||||
}
|
}
|
||||||
|
|
||||||
// stateAtTransaction returns the execution environment of a certain transaction.
|
// stateAtTransaction returns the execution environment of a certain transaction.
|
||||||
func (leth *LightEthereum) stateAtTransaction(ctx context.Context, block *types.Block, txIndex int, reexec uint64) (core.Message, vm.BlockContext, *state.StateDB, tracers.StateReleaseFunc, error) {
|
func (leth *LightEthereum) stateAtTransaction(ctx context.Context, block *types.Block, txIndex int, reexec uint64) (*core.Message, vm.BlockContext, *state.StateDB, tracers.StateReleaseFunc, error) {
|
||||||
// Short circuit if it's genesis block.
|
// Short circuit if it's genesis block.
|
||||||
if block.NumberU64() == 0 {
|
if block.NumberU64() == 0 {
|
||||||
return nil, vm.BlockContext{}, nil, nil, errors.New("no transaction in genesis")
|
return nil, vm.BlockContext{}, nil, nil, errors.New("no transaction in genesis")
|
||||||
@ -60,7 +60,7 @@ func (leth *LightEthereum) stateAtTransaction(ctx context.Context, block *types.
|
|||||||
signer := types.MakeSigner(leth.blockchain.Config(), block.Number())
|
signer := types.MakeSigner(leth.blockchain.Config(), block.Number())
|
||||||
for idx, tx := range block.Transactions() {
|
for idx, tx := range block.Transactions() {
|
||||||
// Assemble the transaction call message and return if the requested offset
|
// Assemble the transaction call message and return if the requested offset
|
||||||
msg, _ := tx.AsMessage(signer, block.BaseFee())
|
msg, _ := core.TransactionToMessage(tx, signer, block.BaseFee())
|
||||||
txContext := core.NewEVMTxContext(msg)
|
txContext := core.NewEVMTxContext(msg)
|
||||||
context := core.NewEVMBlockContext(block.Header(), leth.blockchain, nil)
|
context := core.NewEVMBlockContext(block.Header(), leth.blockchain, nil)
|
||||||
statedb.SetTxContext(tx.Hash(), idx)
|
statedb.SetTxContext(tx.Hash(), idx)
|
||||||
|
@ -174,12 +174,6 @@ func odrAccounts(ctx context.Context, db ethdb.Database, bc *core.BlockChain, lc
|
|||||||
|
|
||||||
func TestOdrContractCallLes2(t *testing.T) { testChainOdr(t, 1, odrContractCall) }
|
func TestOdrContractCallLes2(t *testing.T) { testChainOdr(t, 1, odrContractCall) }
|
||||||
|
|
||||||
type callmsg struct {
|
|
||||||
types.Message
|
|
||||||
}
|
|
||||||
|
|
||||||
func (callmsg) CheckNonce() bool { return false }
|
|
||||||
|
|
||||||
func odrContractCall(ctx context.Context, db ethdb.Database, bc *core.BlockChain, lc *LightChain, bhash common.Hash) ([]byte, error) {
|
func odrContractCall(ctx context.Context, db ethdb.Database, bc *core.BlockChain, lc *LightChain, bhash common.Hash) ([]byte, error) {
|
||||||
data := common.Hex2Bytes("60CD26850000000000000000000000000000000000000000000000000000000000000000")
|
data := common.Hex2Bytes("60CD26850000000000000000000000000000000000000000000000000000000000000000")
|
||||||
config := params.TestChainConfig
|
config := params.TestChainConfig
|
||||||
@ -205,7 +199,17 @@ func odrContractCall(ctx context.Context, db ethdb.Database, bc *core.BlockChain
|
|||||||
|
|
||||||
// Perform read-only call.
|
// Perform read-only call.
|
||||||
st.SetBalance(testBankAddress, math.MaxBig256)
|
st.SetBalance(testBankAddress, math.MaxBig256)
|
||||||
msg := callmsg{types.NewMessage(testBankAddress, &testContractAddr, 0, new(big.Int), 1000000, big.NewInt(params.InitialBaseFee), big.NewInt(params.InitialBaseFee), new(big.Int), data, nil, true)}
|
msg := &core.Message{
|
||||||
|
From: testBankAddress,
|
||||||
|
To: &testContractAddr,
|
||||||
|
Value: new(big.Int),
|
||||||
|
GasLimit: 1000000,
|
||||||
|
GasPrice: big.NewInt(params.InitialBaseFee),
|
||||||
|
GasFeeCap: big.NewInt(params.InitialBaseFee),
|
||||||
|
GasTipCap: new(big.Int),
|
||||||
|
Data: data,
|
||||||
|
SkipAccountChecks: true,
|
||||||
|
}
|
||||||
txContext := core.NewEVMTxContext(msg)
|
txContext := core.NewEVMTxContext(msg)
|
||||||
context := core.NewEVMBlockContext(header, chain, nil)
|
context := core.NewEVMBlockContext(header, chain, nil)
|
||||||
vmenv := vm.NewEVM(context, txContext, st, config, vm.Config{NoBaseFee: true})
|
vmenv := vm.NewEVM(context, txContext, st, config, vm.Config{NoBaseFee: true})
|
||||||
|
@ -125,7 +125,5 @@ compile_fuzzer tests/fuzzers/snap FuzzSRange fuzz_storage_range
|
|||||||
compile_fuzzer tests/fuzzers/snap FuzzByteCodes fuzz_byte_codes
|
compile_fuzzer tests/fuzzers/snap FuzzByteCodes fuzz_byte_codes
|
||||||
compile_fuzzer tests/fuzzers/snap FuzzTrieNodes fuzz_trie_nodes
|
compile_fuzzer tests/fuzzers/snap FuzzTrieNodes fuzz_trie_nodes
|
||||||
|
|
||||||
compile_fuzzer tests/fuzzers/modexp Fuzz fuzzModexp
|
|
||||||
|
|
||||||
#TODO: move this to tests/fuzzers, if possible
|
#TODO: move this to tests/fuzzers, if possible
|
||||||
compile_fuzzer crypto/blake2b Fuzz fuzzBlake2b
|
compile_fuzzer crypto/blake2b Fuzz fuzzBlake2b
|
||||||
|
@ -83,6 +83,7 @@ type UDPv5 struct {
|
|||||||
callCh chan *callV5
|
callCh chan *callV5
|
||||||
callDoneCh chan *callV5
|
callDoneCh chan *callV5
|
||||||
respTimeoutCh chan *callTimeout
|
respTimeoutCh chan *callTimeout
|
||||||
|
unhandled chan<- ReadPacket
|
||||||
|
|
||||||
// state of dispatch
|
// state of dispatch
|
||||||
codec codecV5
|
codec codecV5
|
||||||
@ -156,6 +157,7 @@ func newUDPv5(conn UDPConn, ln *enode.LocalNode, cfg Config) (*UDPv5, error) {
|
|||||||
callCh: make(chan *callV5),
|
callCh: make(chan *callV5),
|
||||||
callDoneCh: make(chan *callV5),
|
callDoneCh: make(chan *callV5),
|
||||||
respTimeoutCh: make(chan *callTimeout),
|
respTimeoutCh: make(chan *callTimeout),
|
||||||
|
unhandled: cfg.Unhandled,
|
||||||
// state of dispatch
|
// state of dispatch
|
||||||
codec: v5wire.NewCodec(ln, cfg.PrivateKey, cfg.Clock, cfg.V5ProtocolID),
|
codec: v5wire.NewCodec(ln, cfg.PrivateKey, cfg.Clock, cfg.V5ProtocolID),
|
||||||
activeCallByNode: make(map[enode.ID]*callV5),
|
activeCallByNode: make(map[enode.ID]*callV5),
|
||||||
@ -657,6 +659,14 @@ func (t *UDPv5) handlePacket(rawpacket []byte, fromAddr *net.UDPAddr) error {
|
|||||||
addr := fromAddr.String()
|
addr := fromAddr.String()
|
||||||
fromID, fromNode, packet, err := t.codec.Decode(rawpacket, addr)
|
fromID, fromNode, packet, err := t.codec.Decode(rawpacket, addr)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
if t.unhandled != nil && v5wire.IsInvalidHeader(err) {
|
||||||
|
// The packet seems unrelated to discv5, send it to the next protocol.
|
||||||
|
// t.log.Trace("Unhandled discv5 packet", "id", fromID, "addr", addr, "err", err)
|
||||||
|
up := ReadPacket{Data: make([]byte, len(rawpacket)), Addr: fromAddr}
|
||||||
|
copy(up.Data, rawpacket)
|
||||||
|
t.unhandled <- up
|
||||||
|
return nil
|
||||||
|
}
|
||||||
t.log.Debug("Bad discv5 packet", "id", fromID, "addr", addr, "err", err)
|
t.log.Debug("Bad discv5 packet", "id", fromID, "addr", addr, "err", err)
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
@ -94,6 +94,8 @@ const (
|
|||||||
// Should reject packets smaller than minPacketSize.
|
// Should reject packets smaller than minPacketSize.
|
||||||
minPacketSize = 63
|
minPacketSize = 63
|
||||||
|
|
||||||
|
maxPacketSize = 1280
|
||||||
|
|
||||||
minMessageSize = 48 // this refers to data after static headers
|
minMessageSize = 48 // this refers to data after static headers
|
||||||
randomPacketMsgSize = 20
|
randomPacketMsgSize = 20
|
||||||
)
|
)
|
||||||
@ -122,6 +124,13 @@ var (
|
|||||||
ErrInvalidReqID = errors.New("request ID larger than 8 bytes")
|
ErrInvalidReqID = errors.New("request ID larger than 8 bytes")
|
||||||
)
|
)
|
||||||
|
|
||||||
|
// IsInvalidHeader reports whether 'err' is related to an invalid packet header. When it
|
||||||
|
// returns false, it is pretty certain that the packet causing the error does not belong
|
||||||
|
// to discv5.
|
||||||
|
func IsInvalidHeader(err error) bool {
|
||||||
|
return err == errTooShort || err == errInvalidHeader || err == errMsgTooShort
|
||||||
|
}
|
||||||
|
|
||||||
// Packet sizes.
|
// Packet sizes.
|
||||||
var (
|
var (
|
||||||
sizeofStaticHeader = binary.Size(StaticHeader{})
|
sizeofStaticHeader = binary.Size(StaticHeader{})
|
||||||
@ -147,6 +156,7 @@ type Codec struct {
|
|||||||
msgctbuf []byte // message data ciphertext
|
msgctbuf []byte // message data ciphertext
|
||||||
|
|
||||||
// decoder buffer
|
// decoder buffer
|
||||||
|
decbuf []byte
|
||||||
reader bytes.Reader
|
reader bytes.Reader
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -158,6 +168,7 @@ func NewCodec(ln *enode.LocalNode, key *ecdsa.PrivateKey, clock mclock.Clock, pr
|
|||||||
privkey: key,
|
privkey: key,
|
||||||
sc: NewSessionCache(1024, clock),
|
sc: NewSessionCache(1024, clock),
|
||||||
protocolID: DefaultProtocolID,
|
protocolID: DefaultProtocolID,
|
||||||
|
decbuf: make([]byte, maxPacketSize),
|
||||||
}
|
}
|
||||||
if protocolID != nil {
|
if protocolID != nil {
|
||||||
c.protocolID = *protocolID
|
c.protocolID = *protocolID
|
||||||
@ -424,10 +435,13 @@ func (c *Codec) encryptMessage(s *session, p Packet, head *Header, headerData []
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Decode decodes a discovery packet.
|
// Decode decodes a discovery packet.
|
||||||
func (c *Codec) Decode(input []byte, addr string) (src enode.ID, n *enode.Node, p Packet, err error) {
|
func (c *Codec) Decode(inputData []byte, addr string) (src enode.ID, n *enode.Node, p Packet, err error) {
|
||||||
if len(input) < minPacketSize {
|
if len(inputData) < minPacketSize {
|
||||||
return enode.ID{}, nil, nil, errTooShort
|
return enode.ID{}, nil, nil, errTooShort
|
||||||
}
|
}
|
||||||
|
// Copy the packet to a tmp buffer to avoid modifying it.
|
||||||
|
c.decbuf = append(c.decbuf[:0], inputData...)
|
||||||
|
input := c.decbuf
|
||||||
// Unmask the static header.
|
// Unmask the static header.
|
||||||
var head Header
|
var head Header
|
||||||
copy(head.IV[:], input[:sizeofMaskingIV])
|
copy(head.IV[:], input[:sizeofMaskingIV])
|
||||||
|
@ -1057,7 +1057,7 @@ func (srv *Server) runPeer(p *Peer) {
|
|||||||
// Broadcast peer drop to external subscribers. This needs to be
|
// Broadcast peer drop to external subscribers. This needs to be
|
||||||
// after the send to delpeer so subscribers have a consistent view of
|
// after the send to delpeer so subscribers have a consistent view of
|
||||||
// the peer set (i.e. Server.Peers() doesn't include the peer when the
|
// the peer set (i.e. Server.Peers() doesn't include the peer when the
|
||||||
// event is received.
|
// event is received).
|
||||||
srv.peerFeed.Send(&PeerEvent{
|
srv.peerFeed.Send(&PeerEvent{
|
||||||
Type: PeerEventTypeDrop,
|
Type: PeerEventTypeDrop,
|
||||||
Peer: p.ID(),
|
Peer: p.ID(),
|
||||||
|
@ -22,14 +22,10 @@ import "github.com/ethereum/go-ethereum/common"
|
|||||||
// the main Ethereum network.
|
// the main Ethereum network.
|
||||||
var MainnetBootnodes = []string{
|
var MainnetBootnodes = []string{
|
||||||
// Ethereum Foundation Go Bootnodes
|
// Ethereum Foundation Go Bootnodes
|
||||||
"enode://d860a01f9722d78051619d1e2351aba3f43f943f6f00718d1b9baa4101932a1f5011f16bb2b1bb35db20d6fe28fa0bf09636d26a87d31de9ec6203eeedb1f666@18.138.108.67:30303", // bootnode-aws-ap-southeast-1-001
|
"enode://d860a01f9722d78051619d1e2351aba3f43f943f6f00718d1b9baa4101932a1f5011f16bb2b1bb35db20d6fe28fa0bf09636d26a87d31de9ec6203eeedb1f666@18.138.108.67:30303", // bootnode-aws-ap-southeast-1-001
|
||||||
"enode://22a8232c3abc76a16ae9d6c3b164f98775fe226f0917b0ca871128a74a8e9630b458460865bab457221f1d448dd9791d24c4e5d88786180ac185df813a68d4de@3.209.45.79:30303", // bootnode-aws-us-east-1-001
|
"enode://22a8232c3abc76a16ae9d6c3b164f98775fe226f0917b0ca871128a74a8e9630b458460865bab457221f1d448dd9791d24c4e5d88786180ac185df813a68d4de@3.209.45.79:30303", // bootnode-aws-us-east-1-001
|
||||||
"enode://8499da03c47d637b20eee24eec3c356c9a2e6148d6fe25ca195c7949ab8ec2c03e3556126b0d7ed644675e78c4318b08691b7b57de10e5f0d40d05b09238fa0a@52.187.207.27:30303", // bootnode-azure-australiaeast-001
|
"enode://2b252ab6a1d0f971d9722cb839a42cb81db019ba44c08754628ab4a823487071b5695317c8ccd085219c3a03af063495b2f1da8d18218da2d6a82981b45e6ffc@65.108.70.101:30303", // bootnode-hetzner-hel
|
||||||
"enode://103858bdb88756c71f15e9b5e09b56dc1be52f0a5021d46301dbbfb7e130029cc9d0d6f73f693bc29b665770fff7da4d34f3c6379fe12721b5d7a0bcb5ca1fc1@191.234.162.198:30303", // bootnode-azure-brazilsouth-001
|
"enode://4aeb4ab6c14b23e2c4cfdce879c04b0748a20d8e9b59e25ded2a08143e265c6c25936e74cbc8e641e3312ca288673d91f2f93f8e277de3cfa444ecdaaf982052@157.90.35.166:30303", // bootnode-hetzner-fsn
|
||||||
"enode://715171f50508aba88aecd1250af392a45a330af91d7b90701c436b618c86aaa1589c9184561907bebbb56439b8f8787bc01f49a7c77276c58c1b09822d75e8e8@52.231.165.108:30303", // bootnode-azure-koreasouth-001
|
|
||||||
"enode://5d6d7cd20d6da4bb83a1d28cadb5d409b64edf314c0335df658c1a54e32c7c4a7ab7823d57c39b6a757556e68ff1df17c748b698544a55cb488b52479a92b60f@104.42.217.25:30303", // bootnode-azure-westus-001
|
|
||||||
"enode://2b252ab6a1d0f971d9722cb839a42cb81db019ba44c08754628ab4a823487071b5695317c8ccd085219c3a03af063495b2f1da8d18218da2d6a82981b45e6ffc@65.108.70.101:30303", // bootnode-hetzner-hel
|
|
||||||
"enode://4aeb4ab6c14b23e2c4cfdce879c04b0748a20d8e9b59e25ded2a08143e265c6c25936e74cbc8e641e3312ca288673d91f2f93f8e277de3cfa444ecdaaf982052@157.90.35.166:30303", // bootnode-hetzner-fsn
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// SepoliaBootnodes are the enode URLs of the P2P bootstrap nodes running on the
|
// SepoliaBootnodes are the enode URLs of the P2P bootstrap nodes running on the
|
||||||
|
@ -76,6 +76,7 @@ var (
|
|||||||
GrayGlacierBlock: big.NewInt(15_050_000),
|
GrayGlacierBlock: big.NewInt(15_050_000),
|
||||||
TerminalTotalDifficulty: MainnetTerminalTotalDifficulty, // 58_750_000_000_000_000_000_000
|
TerminalTotalDifficulty: MainnetTerminalTotalDifficulty, // 58_750_000_000_000_000_000_000
|
||||||
TerminalTotalDifficultyPassed: true,
|
TerminalTotalDifficultyPassed: true,
|
||||||
|
ShanghaiTime: newUint64(1681338455),
|
||||||
Ethash: new(EthashConfig),
|
Ethash: new(EthashConfig),
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -23,7 +23,7 @@ import (
|
|||||||
const (
|
const (
|
||||||
VersionMajor = 1 // Major version component of the current release
|
VersionMajor = 1 // Major version component of the current release
|
||||||
VersionMinor = 11 // Minor version component of the current release
|
VersionMinor = 11 // Minor version component of the current release
|
||||||
VersionPatch = 3 // Patch version component of the current release
|
VersionPatch = 5 // Patch version component of the current release
|
||||||
VersionMeta = "stable" // Version metadata to append to the version string
|
VersionMeta = "stable" // Version metadata to append to the version string
|
||||||
)
|
)
|
||||||
|
|
||||||
|
@ -29,6 +29,7 @@ import (
|
|||||||
"sync"
|
"sync"
|
||||||
|
|
||||||
"github.com/ethereum/go-ethereum/rlp/internal/rlpstruct"
|
"github.com/ethereum/go-ethereum/rlp/internal/rlpstruct"
|
||||||
|
"github.com/holiman/uint256"
|
||||||
)
|
)
|
||||||
|
|
||||||
//lint:ignore ST1012 EOL is not an error.
|
//lint:ignore ST1012 EOL is not an error.
|
||||||
@ -52,6 +53,7 @@ var (
|
|||||||
errUintOverflow = errors.New("rlp: uint overflow")
|
errUintOverflow = errors.New("rlp: uint overflow")
|
||||||
errNoPointer = errors.New("rlp: interface given to Decode must be a pointer")
|
errNoPointer = errors.New("rlp: interface given to Decode must be a pointer")
|
||||||
errDecodeIntoNil = errors.New("rlp: pointer given to Decode must not be nil")
|
errDecodeIntoNil = errors.New("rlp: pointer given to Decode must not be nil")
|
||||||
|
errUint256Large = errors.New("rlp: value too large for uint256")
|
||||||
|
|
||||||
streamPool = sync.Pool{
|
streamPool = sync.Pool{
|
||||||
New: func() interface{} { return new(Stream) },
|
New: func() interface{} { return new(Stream) },
|
||||||
@ -148,6 +150,7 @@ func addErrorContext(err error, ctx string) error {
|
|||||||
var (
|
var (
|
||||||
decoderInterface = reflect.TypeOf(new(Decoder)).Elem()
|
decoderInterface = reflect.TypeOf(new(Decoder)).Elem()
|
||||||
bigInt = reflect.TypeOf(big.Int{})
|
bigInt = reflect.TypeOf(big.Int{})
|
||||||
|
u256Int = reflect.TypeOf(uint256.Int{})
|
||||||
)
|
)
|
||||||
|
|
||||||
func makeDecoder(typ reflect.Type, tags rlpstruct.Tags) (dec decoder, err error) {
|
func makeDecoder(typ reflect.Type, tags rlpstruct.Tags) (dec decoder, err error) {
|
||||||
@ -159,6 +162,10 @@ func makeDecoder(typ reflect.Type, tags rlpstruct.Tags) (dec decoder, err error)
|
|||||||
return decodeBigInt, nil
|
return decodeBigInt, nil
|
||||||
case typ.AssignableTo(bigInt):
|
case typ.AssignableTo(bigInt):
|
||||||
return decodeBigIntNoPtr, nil
|
return decodeBigIntNoPtr, nil
|
||||||
|
case typ == reflect.PtrTo(u256Int):
|
||||||
|
return decodeU256, nil
|
||||||
|
case typ == u256Int:
|
||||||
|
return decodeU256NoPtr, nil
|
||||||
case kind == reflect.Ptr:
|
case kind == reflect.Ptr:
|
||||||
return makePtrDecoder(typ, tags)
|
return makePtrDecoder(typ, tags)
|
||||||
case reflect.PtrTo(typ).Implements(decoderInterface):
|
case reflect.PtrTo(typ).Implements(decoderInterface):
|
||||||
@ -235,6 +242,24 @@ func decodeBigInt(s *Stream, val reflect.Value) error {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func decodeU256NoPtr(s *Stream, val reflect.Value) error {
|
||||||
|
return decodeU256(s, val.Addr())
|
||||||
|
}
|
||||||
|
|
||||||
|
func decodeU256(s *Stream, val reflect.Value) error {
|
||||||
|
i := val.Interface().(*uint256.Int)
|
||||||
|
if i == nil {
|
||||||
|
i = new(uint256.Int)
|
||||||
|
val.Set(reflect.ValueOf(i))
|
||||||
|
}
|
||||||
|
|
||||||
|
err := s.ReadUint256(i)
|
||||||
|
if err != nil {
|
||||||
|
return wrapStreamError(err, val.Type())
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
func makeListDecoder(typ reflect.Type, tag rlpstruct.Tags) (decoder, error) {
|
func makeListDecoder(typ reflect.Type, tag rlpstruct.Tags) (decoder, error) {
|
||||||
etype := typ.Elem()
|
etype := typ.Elem()
|
||||||
if etype.Kind() == reflect.Uint8 && !reflect.PtrTo(etype).Implements(decoderInterface) {
|
if etype.Kind() == reflect.Uint8 && !reflect.PtrTo(etype).Implements(decoderInterface) {
|
||||||
@ -863,6 +888,45 @@ func (s *Stream) decodeBigInt(dst *big.Int) error {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// ReadUint256 decodes the next value as a uint256.
|
||||||
|
func (s *Stream) ReadUint256(dst *uint256.Int) error {
|
||||||
|
var buffer []byte
|
||||||
|
kind, size, err := s.Kind()
|
||||||
|
switch {
|
||||||
|
case err != nil:
|
||||||
|
return err
|
||||||
|
case kind == List:
|
||||||
|
return ErrExpectedString
|
||||||
|
case kind == Byte:
|
||||||
|
buffer = s.uintbuf[:1]
|
||||||
|
buffer[0] = s.byteval
|
||||||
|
s.kind = -1 // re-arm Kind
|
||||||
|
case size == 0:
|
||||||
|
// Avoid zero-length read.
|
||||||
|
s.kind = -1
|
||||||
|
case size <= uint64(len(s.uintbuf)):
|
||||||
|
// All possible uint256 values fit into s.uintbuf.
|
||||||
|
buffer = s.uintbuf[:size]
|
||||||
|
if err := s.readFull(buffer); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
// Reject inputs where single byte encoding should have been used.
|
||||||
|
if size == 1 && buffer[0] < 128 {
|
||||||
|
return ErrCanonSize
|
||||||
|
}
|
||||||
|
default:
|
||||||
|
return errUint256Large
|
||||||
|
}
|
||||||
|
|
||||||
|
// Reject leading zero bytes.
|
||||||
|
if len(buffer) > 0 && buffer[0] == 0 {
|
||||||
|
return ErrCanonInt
|
||||||
|
}
|
||||||
|
// Set the integer bytes.
|
||||||
|
dst.SetBytes(buffer)
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
// Decode decodes a value and stores the result in the value pointed
|
// Decode decodes a value and stores the result in the value pointed
|
||||||
// to by val. Please see the documentation for the Decode function
|
// to by val. Please see the documentation for the Decode function
|
||||||
// to learn about the decoding rules.
|
// to learn about the decoding rules.
|
||||||
|
@ -28,6 +28,7 @@ import (
|
|||||||
"testing"
|
"testing"
|
||||||
|
|
||||||
"github.com/ethereum/go-ethereum/common/math"
|
"github.com/ethereum/go-ethereum/common/math"
|
||||||
|
"github.com/holiman/uint256"
|
||||||
)
|
)
|
||||||
|
|
||||||
func TestStreamKind(t *testing.T) {
|
func TestStreamKind(t *testing.T) {
|
||||||
@ -468,6 +469,10 @@ var (
|
|||||||
veryVeryBigInt = new(big.Int).Exp(veryBigInt, big.NewInt(8), nil)
|
veryVeryBigInt = new(big.Int).Exp(veryBigInt, big.NewInt(8), nil)
|
||||||
)
|
)
|
||||||
|
|
||||||
|
var (
|
||||||
|
veryBigInt256, _ = uint256.FromBig(veryBigInt)
|
||||||
|
)
|
||||||
|
|
||||||
var decodeTests = []decodeTest{
|
var decodeTests = []decodeTest{
|
||||||
// booleans
|
// booleans
|
||||||
{input: "01", ptr: new(bool), value: true},
|
{input: "01", ptr: new(bool), value: true},
|
||||||
@ -541,11 +546,27 @@ var decodeTests = []decodeTest{
|
|||||||
{input: "89FFFFFFFFFFFFFFFFFF", ptr: new(*big.Int), value: veryBigInt},
|
{input: "89FFFFFFFFFFFFFFFFFF", ptr: new(*big.Int), value: veryBigInt},
|
||||||
{input: "B848FFFFFFFFFFFFFFFFF800000000000000001BFFFFFFFFFFFFFFFFC8000000000000000045FFFFFFFFFFFFFFFFC800000000000000001BFFFFFFFFFFFFFFFFF8000000000000000001", ptr: new(*big.Int), value: veryVeryBigInt},
|
{input: "B848FFFFFFFFFFFFFFFFF800000000000000001BFFFFFFFFFFFFFFFFC8000000000000000045FFFFFFFFFFFFFFFFC800000000000000001BFFFFFFFFFFFFFFFFF8000000000000000001", ptr: new(*big.Int), value: veryVeryBigInt},
|
||||||
{input: "10", ptr: new(big.Int), value: *big.NewInt(16)}, // non-pointer also works
|
{input: "10", ptr: new(big.Int), value: *big.NewInt(16)}, // non-pointer also works
|
||||||
|
|
||||||
|
// big int errors
|
||||||
{input: "C0", ptr: new(*big.Int), error: "rlp: expected input string or byte for *big.Int"},
|
{input: "C0", ptr: new(*big.Int), error: "rlp: expected input string or byte for *big.Int"},
|
||||||
{input: "00", ptr: new(*big.Int), error: "rlp: non-canonical integer (leading zero bytes) for *big.Int"},
|
{input: "00", ptr: new(*big.Int), error: "rlp: non-canonical integer (leading zero bytes) for *big.Int"},
|
||||||
{input: "820001", ptr: new(*big.Int), error: "rlp: non-canonical integer (leading zero bytes) for *big.Int"},
|
{input: "820001", ptr: new(*big.Int), error: "rlp: non-canonical integer (leading zero bytes) for *big.Int"},
|
||||||
{input: "8105", ptr: new(*big.Int), error: "rlp: non-canonical size information for *big.Int"},
|
{input: "8105", ptr: new(*big.Int), error: "rlp: non-canonical size information for *big.Int"},
|
||||||
|
|
||||||
|
// uint256
|
||||||
|
{input: "80", ptr: new(*uint256.Int), value: uint256.NewInt(0)},
|
||||||
|
{input: "01", ptr: new(*uint256.Int), value: uint256.NewInt(1)},
|
||||||
|
{input: "88FFFFFFFFFFFFFFFF", ptr: new(*uint256.Int), value: uint256.NewInt(math.MaxUint64)},
|
||||||
|
{input: "89FFFFFFFFFFFFFFFFFF", ptr: new(*uint256.Int), value: veryBigInt256},
|
||||||
|
{input: "10", ptr: new(uint256.Int), value: *uint256.NewInt(16)}, // non-pointer also works
|
||||||
|
|
||||||
|
// uint256 errors
|
||||||
|
{input: "C0", ptr: new(*uint256.Int), error: "rlp: expected input string or byte for *uint256.Int"},
|
||||||
|
{input: "00", ptr: new(*uint256.Int), error: "rlp: non-canonical integer (leading zero bytes) for *uint256.Int"},
|
||||||
|
{input: "820001", ptr: new(*uint256.Int), error: "rlp: non-canonical integer (leading zero bytes) for *uint256.Int"},
|
||||||
|
{input: "8105", ptr: new(*uint256.Int), error: "rlp: non-canonical size information for *uint256.Int"},
|
||||||
|
{input: "A1FFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFF00", ptr: new(*uint256.Int), error: "rlp: value too large for uint256"},
|
||||||
|
|
||||||
// structs
|
// structs
|
||||||
{
|
{
|
||||||
input: "C50583343434",
|
input: "C50583343434",
|
||||||
@ -1223,6 +1244,27 @@ func BenchmarkDecodeBigInts(b *testing.B) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func BenchmarkDecodeU256Ints(b *testing.B) {
|
||||||
|
ints := make([]*uint256.Int, 200)
|
||||||
|
for i := range ints {
|
||||||
|
ints[i], _ = uint256.FromBig(math.BigPow(2, int64(i)))
|
||||||
|
}
|
||||||
|
enc, err := EncodeToBytes(ints)
|
||||||
|
if err != nil {
|
||||||
|
b.Fatal(err)
|
||||||
|
}
|
||||||
|
b.SetBytes(int64(len(enc)))
|
||||||
|
b.ReportAllocs()
|
||||||
|
b.ResetTimer()
|
||||||
|
|
||||||
|
var out []*uint256.Int
|
||||||
|
for i := 0; i < b.N; i++ {
|
||||||
|
if err := DecodeBytes(enc, &out); err != nil {
|
||||||
|
b.Fatal(err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
func encodeTestSlice(n uint) []byte {
|
func encodeTestSlice(n uint) []byte {
|
||||||
s := make([]uint, n)
|
s := make([]uint, n)
|
||||||
for i := uint(0); i < n; i++ {
|
for i := uint(0); i < n; i++ {
|
||||||
|
@ -17,10 +17,13 @@
|
|||||||
package rlp
|
package rlp
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"encoding/binary"
|
||||||
"io"
|
"io"
|
||||||
"math/big"
|
"math/big"
|
||||||
"reflect"
|
"reflect"
|
||||||
"sync"
|
"sync"
|
||||||
|
|
||||||
|
"github.com/holiman/uint256"
|
||||||
)
|
)
|
||||||
|
|
||||||
type encBuffer struct {
|
type encBuffer struct {
|
||||||
@ -169,6 +172,23 @@ func (w *encBuffer) writeBigInt(i *big.Int) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// writeUint256 writes z as an integer.
|
||||||
|
func (w *encBuffer) writeUint256(z *uint256.Int) {
|
||||||
|
bitlen := z.BitLen()
|
||||||
|
if bitlen <= 64 {
|
||||||
|
w.writeUint64(z.Uint64())
|
||||||
|
return
|
||||||
|
}
|
||||||
|
nBytes := byte((bitlen + 7) / 8)
|
||||||
|
var b [33]byte
|
||||||
|
binary.BigEndian.PutUint64(b[1:9], z[3])
|
||||||
|
binary.BigEndian.PutUint64(b[9:17], z[2])
|
||||||
|
binary.BigEndian.PutUint64(b[17:25], z[1])
|
||||||
|
binary.BigEndian.PutUint64(b[25:33], z[0])
|
||||||
|
b[32-nBytes] = 0x80 + nBytes
|
||||||
|
w.str = append(w.str, b[32-nBytes:]...)
|
||||||
|
}
|
||||||
|
|
||||||
// list adds a new list header to the header stack. It returns the index of the header.
|
// list adds a new list header to the header stack. It returns the index of the header.
|
||||||
// Call listEnd with this index after encoding the content of the list.
|
// Call listEnd with this index after encoding the content of the list.
|
||||||
func (buf *encBuffer) list() int {
|
func (buf *encBuffer) list() int {
|
||||||
@ -376,6 +396,11 @@ func (w EncoderBuffer) WriteBigInt(i *big.Int) {
|
|||||||
w.buf.writeBigInt(i)
|
w.buf.writeBigInt(i)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// WriteUint256 encodes uint256.Int as an RLP string.
|
||||||
|
func (w EncoderBuffer) WriteUint256(i *uint256.Int) {
|
||||||
|
w.buf.writeUint256(i)
|
||||||
|
}
|
||||||
|
|
||||||
// WriteBytes encodes b as an RLP string.
|
// WriteBytes encodes b as an RLP string.
|
||||||
func (w EncoderBuffer) WriteBytes(b []byte) {
|
func (w EncoderBuffer) WriteBytes(b []byte) {
|
||||||
w.buf.writeBytes(b)
|
w.buf.writeBytes(b)
|
||||||
|
@ -24,6 +24,7 @@ import (
|
|||||||
"reflect"
|
"reflect"
|
||||||
|
|
||||||
"github.com/ethereum/go-ethereum/rlp/internal/rlpstruct"
|
"github.com/ethereum/go-ethereum/rlp/internal/rlpstruct"
|
||||||
|
"github.com/holiman/uint256"
|
||||||
)
|
)
|
||||||
|
|
||||||
var (
|
var (
|
||||||
@ -144,6 +145,10 @@ func makeWriter(typ reflect.Type, ts rlpstruct.Tags) (writer, error) {
|
|||||||
return writeBigIntPtr, nil
|
return writeBigIntPtr, nil
|
||||||
case typ.AssignableTo(bigInt):
|
case typ.AssignableTo(bigInt):
|
||||||
return writeBigIntNoPtr, nil
|
return writeBigIntNoPtr, nil
|
||||||
|
case typ == reflect.PtrTo(u256Int):
|
||||||
|
return writeU256IntPtr, nil
|
||||||
|
case typ == u256Int:
|
||||||
|
return writeU256IntNoPtr, nil
|
||||||
case kind == reflect.Ptr:
|
case kind == reflect.Ptr:
|
||||||
return makePtrWriter(typ, ts)
|
return makePtrWriter(typ, ts)
|
||||||
case reflect.PtrTo(typ).Implements(encoderInterface):
|
case reflect.PtrTo(typ).Implements(encoderInterface):
|
||||||
@ -206,6 +211,22 @@ func writeBigIntNoPtr(val reflect.Value, w *encBuffer) error {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func writeU256IntPtr(val reflect.Value, w *encBuffer) error {
|
||||||
|
ptr := val.Interface().(*uint256.Int)
|
||||||
|
if ptr == nil {
|
||||||
|
w.str = append(w.str, 0x80)
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
w.writeUint256(ptr)
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func writeU256IntNoPtr(val reflect.Value, w *encBuffer) error {
|
||||||
|
i := val.Interface().(uint256.Int)
|
||||||
|
w.writeUint256(&i)
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
func writeBytes(val reflect.Value, w *encBuffer) error {
|
func writeBytes(val reflect.Value, w *encBuffer) error {
|
||||||
w.writeBytes(val.Bytes())
|
w.writeBytes(val.Bytes())
|
||||||
return nil
|
return nil
|
||||||
|
@ -27,6 +27,7 @@ import (
|
|||||||
"testing"
|
"testing"
|
||||||
|
|
||||||
"github.com/ethereum/go-ethereum/common/math"
|
"github.com/ethereum/go-ethereum/common/math"
|
||||||
|
"github.com/holiman/uint256"
|
||||||
)
|
)
|
||||||
|
|
||||||
type testEncoder struct {
|
type testEncoder struct {
|
||||||
@ -147,6 +148,30 @@ var encTests = []encTest{
|
|||||||
{val: big.NewInt(-1), error: "rlp: cannot encode negative big.Int"},
|
{val: big.NewInt(-1), error: "rlp: cannot encode negative big.Int"},
|
||||||
{val: *big.NewInt(-1), error: "rlp: cannot encode negative big.Int"},
|
{val: *big.NewInt(-1), error: "rlp: cannot encode negative big.Int"},
|
||||||
|
|
||||||
|
// uint256
|
||||||
|
{val: uint256.NewInt(0), output: "80"},
|
||||||
|
{val: uint256.NewInt(1), output: "01"},
|
||||||
|
{val: uint256.NewInt(127), output: "7F"},
|
||||||
|
{val: uint256.NewInt(128), output: "8180"},
|
||||||
|
{val: uint256.NewInt(256), output: "820100"},
|
||||||
|
{val: uint256.NewInt(1024), output: "820400"},
|
||||||
|
{val: uint256.NewInt(0xFFFFFF), output: "83FFFFFF"},
|
||||||
|
{val: uint256.NewInt(0xFFFFFFFF), output: "84FFFFFFFF"},
|
||||||
|
{val: uint256.NewInt(0xFFFFFFFFFF), output: "85FFFFFFFFFF"},
|
||||||
|
{val: uint256.NewInt(0xFFFFFFFFFFFF), output: "86FFFFFFFFFFFF"},
|
||||||
|
{val: uint256.NewInt(0xFFFFFFFFFFFFFF), output: "87FFFFFFFFFFFFFF"},
|
||||||
|
{
|
||||||
|
val: new(uint256.Int).SetBytes(unhex("102030405060708090A0B0C0D0E0F2")),
|
||||||
|
output: "8F102030405060708090A0B0C0D0E0F2",
|
||||||
|
},
|
||||||
|
{
|
||||||
|
val: new(uint256.Int).SetBytes(unhex("0100020003000400050006000700080009000A000B000C000D000E01")),
|
||||||
|
output: "9C0100020003000400050006000700080009000A000B000C000D000E01",
|
||||||
|
},
|
||||||
|
// non-pointer uint256.Int
|
||||||
|
{val: *uint256.NewInt(0), output: "80"},
|
||||||
|
{val: *uint256.NewInt(0xFFFFFF), output: "83FFFFFF"},
|
||||||
|
|
||||||
// byte arrays
|
// byte arrays
|
||||||
{val: [0]byte{}, output: "80"},
|
{val: [0]byte{}, output: "80"},
|
||||||
{val: [1]byte{0}, output: "00"},
|
{val: [1]byte{0}, output: "00"},
|
||||||
@ -256,6 +281,12 @@ var encTests = []encTest{
|
|||||||
output: "F90200CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376",
|
output: "F90200CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376CF84617364668471776572847A786376",
|
||||||
},
|
},
|
||||||
|
|
||||||
|
// Non-byte arrays are encoded as lists.
|
||||||
|
// Note that it is important to test [4]uint64 specifically,
|
||||||
|
// because that's the underlying type of uint256.Int.
|
||||||
|
{val: [4]uint32{1, 2, 3, 4}, output: "C401020304"},
|
||||||
|
{val: [4]uint64{1, 2, 3, 4}, output: "C401020304"},
|
||||||
|
|
||||||
// RawValue
|
// RawValue
|
||||||
{val: RawValue(unhex("01")), output: "01"},
|
{val: RawValue(unhex("01")), output: "01"},
|
||||||
{val: RawValue(unhex("82FFFF")), output: "82FFFF"},
|
{val: RawValue(unhex("82FFFF")), output: "82FFFF"},
|
||||||
@ -301,6 +332,7 @@ var encTests = []encTest{
|
|||||||
{val: (*[]byte)(nil), output: "80"},
|
{val: (*[]byte)(nil), output: "80"},
|
||||||
{val: (*[10]byte)(nil), output: "80"},
|
{val: (*[10]byte)(nil), output: "80"},
|
||||||
{val: (*big.Int)(nil), output: "80"},
|
{val: (*big.Int)(nil), output: "80"},
|
||||||
|
{val: (*uint256.Int)(nil), output: "80"},
|
||||||
{val: (*[]string)(nil), output: "C0"},
|
{val: (*[]string)(nil), output: "C0"},
|
||||||
{val: (*[10]string)(nil), output: "C0"},
|
{val: (*[10]string)(nil), output: "C0"},
|
||||||
{val: (*[]interface{})(nil), output: "C0"},
|
{val: (*[]interface{})(nil), output: "C0"},
|
||||||
@ -509,6 +541,23 @@ func BenchmarkEncodeBigInts(b *testing.B) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func BenchmarkEncodeU256Ints(b *testing.B) {
|
||||||
|
ints := make([]*uint256.Int, 200)
|
||||||
|
for i := range ints {
|
||||||
|
ints[i], _ = uint256.FromBig(math.BigPow(2, int64(i)))
|
||||||
|
}
|
||||||
|
out := bytes.NewBuffer(make([]byte, 0, 4096))
|
||||||
|
b.ResetTimer()
|
||||||
|
b.ReportAllocs()
|
||||||
|
|
||||||
|
for i := 0; i < b.N; i++ {
|
||||||
|
out.Reset()
|
||||||
|
if err := Encode(out, ints); err != nil {
|
||||||
|
b.Fatal(err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
func BenchmarkEncodeConcurrentInterface(b *testing.B) {
|
func BenchmarkEncodeConcurrentInterface(b *testing.B) {
|
||||||
type struct1 struct {
|
type struct1 struct {
|
||||||
A string
|
A string
|
||||||
|
@ -283,7 +283,7 @@ func (op byteArrayOp) genDecode(ctx *genContext) (string, string) {
|
|||||||
return resultV, b.String()
|
return resultV, b.String()
|
||||||
}
|
}
|
||||||
|
|
||||||
// bigIntNoPtrOp handles non-pointer big.Int.
|
// bigIntOp handles big.Int.
|
||||||
// This exists because big.Int has it's own decoder operation on rlp.Stream,
|
// This exists because big.Int has it's own decoder operation on rlp.Stream,
|
||||||
// but the decode method returns *big.Int, so it needs to be dereferenced.
|
// but the decode method returns *big.Int, so it needs to be dereferenced.
|
||||||
type bigIntOp struct {
|
type bigIntOp struct {
|
||||||
@ -330,6 +330,49 @@ func (op bigIntOp) genDecode(ctx *genContext) (string, string) {
|
|||||||
return result, b.String()
|
return result, b.String()
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// uint256Op handles "github.com/holiman/uint256".Int
|
||||||
|
type uint256Op struct {
|
||||||
|
pointer bool
|
||||||
|
}
|
||||||
|
|
||||||
|
func (op uint256Op) genWrite(ctx *genContext, v string) string {
|
||||||
|
var b bytes.Buffer
|
||||||
|
|
||||||
|
dst := v
|
||||||
|
if !op.pointer {
|
||||||
|
dst = "&" + v
|
||||||
|
}
|
||||||
|
fmt.Fprintf(&b, "w.WriteUint256(%s)\n", dst)
|
||||||
|
|
||||||
|
// Wrap with nil check.
|
||||||
|
if op.pointer {
|
||||||
|
code := b.String()
|
||||||
|
b.Reset()
|
||||||
|
fmt.Fprintf(&b, "if %s == nil {\n", v)
|
||||||
|
fmt.Fprintf(&b, " w.Write(rlp.EmptyString)")
|
||||||
|
fmt.Fprintf(&b, "} else {\n")
|
||||||
|
fmt.Fprint(&b, code)
|
||||||
|
fmt.Fprintf(&b, "}\n")
|
||||||
|
}
|
||||||
|
|
||||||
|
return b.String()
|
||||||
|
}
|
||||||
|
|
||||||
|
func (op uint256Op) genDecode(ctx *genContext) (string, string) {
|
||||||
|
ctx.addImport("github.com/holiman/uint256")
|
||||||
|
|
||||||
|
var b bytes.Buffer
|
||||||
|
resultV := ctx.temp()
|
||||||
|
fmt.Fprintf(&b, "var %s uint256.Int\n", resultV)
|
||||||
|
fmt.Fprintf(&b, "if err := dec.ReadUint256(&%s); err != nil { return err }\n", resultV)
|
||||||
|
|
||||||
|
result := resultV
|
||||||
|
if op.pointer {
|
||||||
|
result = "&" + resultV
|
||||||
|
}
|
||||||
|
return result, b.String()
|
||||||
|
}
|
||||||
|
|
||||||
// encoderDecoderOp handles rlp.Encoder and rlp.Decoder.
|
// encoderDecoderOp handles rlp.Encoder and rlp.Decoder.
|
||||||
// In order to be used with this, the type must implement both interfaces.
|
// In order to be used with this, the type must implement both interfaces.
|
||||||
// This restriction may be lifted in the future by creating separate ops for
|
// This restriction may be lifted in the future by creating separate ops for
|
||||||
@ -635,6 +678,9 @@ func (bctx *buildContext) makeOp(name *types.Named, typ types.Type, tags rlpstru
|
|||||||
if isBigInt(typ) {
|
if isBigInt(typ) {
|
||||||
return bigIntOp{}, nil
|
return bigIntOp{}, nil
|
||||||
}
|
}
|
||||||
|
if isUint256(typ) {
|
||||||
|
return uint256Op{}, nil
|
||||||
|
}
|
||||||
if typ == bctx.rawValueType {
|
if typ == bctx.rawValueType {
|
||||||
return bctx.makeRawValueOp(), nil
|
return bctx.makeRawValueOp(), nil
|
||||||
}
|
}
|
||||||
@ -647,6 +693,9 @@ func (bctx *buildContext) makeOp(name *types.Named, typ types.Type, tags rlpstru
|
|||||||
if isBigInt(typ.Elem()) {
|
if isBigInt(typ.Elem()) {
|
||||||
return bigIntOp{pointer: true}, nil
|
return bigIntOp{pointer: true}, nil
|
||||||
}
|
}
|
||||||
|
if isUint256(typ.Elem()) {
|
||||||
|
return uint256Op{pointer: true}, nil
|
||||||
|
}
|
||||||
// Encoder/Decoder interfaces.
|
// Encoder/Decoder interfaces.
|
||||||
if bctx.isEncoder(typ) {
|
if bctx.isEncoder(typ) {
|
||||||
if bctx.isDecoder(typ) {
|
if bctx.isDecoder(typ) {
|
||||||
|
@ -47,7 +47,7 @@ func init() {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
var tests = []string{"uints", "nil", "rawvalue", "optional", "bigint"}
|
var tests = []string{"uints", "nil", "rawvalue", "optional", "bigint", "uint256"}
|
||||||
|
|
||||||
func TestOutput(t *testing.T) {
|
func TestOutput(t *testing.T) {
|
||||||
for _, test := range tests {
|
for _, test := range tests {
|
||||||
|
10
rlp/rlpgen/testdata/uint256.in.txt
vendored
Normal file
10
rlp/rlpgen/testdata/uint256.in.txt
vendored
Normal file
@ -0,0 +1,10 @@
|
|||||||
|
// -*- mode: go -*-
|
||||||
|
|
||||||
|
package test
|
||||||
|
|
||||||
|
import "github.com/holiman/uint256"
|
||||||
|
|
||||||
|
type Test struct {
|
||||||
|
Int *uint256.Int
|
||||||
|
IntNoPtr uint256.Int
|
||||||
|
}
|
44
rlp/rlpgen/testdata/uint256.out.txt
vendored
Normal file
44
rlp/rlpgen/testdata/uint256.out.txt
vendored
Normal file
@ -0,0 +1,44 @@
|
|||||||
|
package test
|
||||||
|
|
||||||
|
import "github.com/ethereum/go-ethereum/rlp"
|
||||||
|
import "github.com/holiman/uint256"
|
||||||
|
import "io"
|
||||||
|
|
||||||
|
func (obj *Test) EncodeRLP(_w io.Writer) error {
|
||||||
|
w := rlp.NewEncoderBuffer(_w)
|
||||||
|
_tmp0 := w.List()
|
||||||
|
if obj.Int == nil {
|
||||||
|
w.Write(rlp.EmptyString)
|
||||||
|
} else {
|
||||||
|
w.WriteUint256(obj.Int)
|
||||||
|
}
|
||||||
|
w.WriteUint256(&obj.IntNoPtr)
|
||||||
|
w.ListEnd(_tmp0)
|
||||||
|
return w.Flush()
|
||||||
|
}
|
||||||
|
|
||||||
|
func (obj *Test) DecodeRLP(dec *rlp.Stream) error {
|
||||||
|
var _tmp0 Test
|
||||||
|
{
|
||||||
|
if _, err := dec.List(); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
// Int:
|
||||||
|
var _tmp1 uint256.Int
|
||||||
|
if err := dec.ReadUint256(&_tmp1); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
_tmp0.Int = &_tmp1
|
||||||
|
// IntNoPtr:
|
||||||
|
var _tmp2 uint256.Int
|
||||||
|
if err := dec.ReadUint256(&_tmp2); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
_tmp0.IntNoPtr = _tmp2
|
||||||
|
if err := dec.ListEnd(); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
*obj = _tmp0
|
||||||
|
return nil
|
||||||
|
}
|
@ -97,6 +97,16 @@ func isBigInt(typ types.Type) bool {
|
|||||||
return name.Pkg().Path() == "math/big" && name.Name() == "Int"
|
return name.Pkg().Path() == "math/big" && name.Name() == "Int"
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// isUint256 checks whether 'typ' is "github.com/holiman/uint256".Int.
|
||||||
|
func isUint256(typ types.Type) bool {
|
||||||
|
named, ok := typ.(*types.Named)
|
||||||
|
if !ok {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
name := named.Obj()
|
||||||
|
return name.Pkg().Path() == "github.com/holiman/uint256" && name.Name() == "Int"
|
||||||
|
}
|
||||||
|
|
||||||
// isByte checks whether the underlying type of 'typ' is uint8.
|
// isByte checks whether the underlying type of 'typ' is uint8.
|
||||||
func isByte(typ types.Type) bool {
|
func isByte(typ types.Type) bool {
|
||||||
basic, ok := resolveUnderlying(typ).(*types.Basic)
|
basic, ok := resolveUnderlying(typ).(*types.Basic)
|
||||||
|
@ -38,7 +38,7 @@ const (
|
|||||||
wsPingInterval = 30 * time.Second
|
wsPingInterval = 30 * time.Second
|
||||||
wsPingWriteTimeout = 5 * time.Second
|
wsPingWriteTimeout = 5 * time.Second
|
||||||
wsPongTimeout = 30 * time.Second
|
wsPongTimeout = 30 * time.Second
|
||||||
wsMessageSizeLimit = 15 * 1024 * 1024
|
wsMessageSizeLimit = 32 * 1024 * 1024
|
||||||
)
|
)
|
||||||
|
|
||||||
var wsBufferPool = new(sync.Pool)
|
var wsBufferPool = new(sync.Pool)
|
||||||
|
@ -742,112 +742,30 @@ func isPrimitiveTypeValid(primitiveType string) bool {
|
|||||||
primitiveType == "bool" ||
|
primitiveType == "bool" ||
|
||||||
primitiveType == "bool[]" ||
|
primitiveType == "bool[]" ||
|
||||||
primitiveType == "string" ||
|
primitiveType == "string" ||
|
||||||
primitiveType == "string[]" {
|
primitiveType == "string[]" ||
|
||||||
return true
|
primitiveType == "bytes" ||
|
||||||
}
|
|
||||||
if primitiveType == "bytes" ||
|
|
||||||
primitiveType == "bytes[]" ||
|
primitiveType == "bytes[]" ||
|
||||||
primitiveType == "bytes1" ||
|
primitiveType == "int" ||
|
||||||
primitiveType == "bytes1[]" ||
|
|
||||||
primitiveType == "bytes2" ||
|
|
||||||
primitiveType == "bytes2[]" ||
|
|
||||||
primitiveType == "bytes3" ||
|
|
||||||
primitiveType == "bytes3[]" ||
|
|
||||||
primitiveType == "bytes4" ||
|
|
||||||
primitiveType == "bytes4[]" ||
|
|
||||||
primitiveType == "bytes5" ||
|
|
||||||
primitiveType == "bytes5[]" ||
|
|
||||||
primitiveType == "bytes6" ||
|
|
||||||
primitiveType == "bytes6[]" ||
|
|
||||||
primitiveType == "bytes7" ||
|
|
||||||
primitiveType == "bytes7[]" ||
|
|
||||||
primitiveType == "bytes8" ||
|
|
||||||
primitiveType == "bytes8[]" ||
|
|
||||||
primitiveType == "bytes9" ||
|
|
||||||
primitiveType == "bytes9[]" ||
|
|
||||||
primitiveType == "bytes10" ||
|
|
||||||
primitiveType == "bytes10[]" ||
|
|
||||||
primitiveType == "bytes11" ||
|
|
||||||
primitiveType == "bytes11[]" ||
|
|
||||||
primitiveType == "bytes12" ||
|
|
||||||
primitiveType == "bytes12[]" ||
|
|
||||||
primitiveType == "bytes13" ||
|
|
||||||
primitiveType == "bytes13[]" ||
|
|
||||||
primitiveType == "bytes14" ||
|
|
||||||
primitiveType == "bytes14[]" ||
|
|
||||||
primitiveType == "bytes15" ||
|
|
||||||
primitiveType == "bytes15[]" ||
|
|
||||||
primitiveType == "bytes16" ||
|
|
||||||
primitiveType == "bytes16[]" ||
|
|
||||||
primitiveType == "bytes17" ||
|
|
||||||
primitiveType == "bytes17[]" ||
|
|
||||||
primitiveType == "bytes18" ||
|
|
||||||
primitiveType == "bytes18[]" ||
|
|
||||||
primitiveType == "bytes19" ||
|
|
||||||
primitiveType == "bytes19[]" ||
|
|
||||||
primitiveType == "bytes20" ||
|
|
||||||
primitiveType == "bytes20[]" ||
|
|
||||||
primitiveType == "bytes21" ||
|
|
||||||
primitiveType == "bytes21[]" ||
|
|
||||||
primitiveType == "bytes22" ||
|
|
||||||
primitiveType == "bytes22[]" ||
|
|
||||||
primitiveType == "bytes23" ||
|
|
||||||
primitiveType == "bytes23[]" ||
|
|
||||||
primitiveType == "bytes24" ||
|
|
||||||
primitiveType == "bytes24[]" ||
|
|
||||||
primitiveType == "bytes25" ||
|
|
||||||
primitiveType == "bytes25[]" ||
|
|
||||||
primitiveType == "bytes26" ||
|
|
||||||
primitiveType == "bytes26[]" ||
|
|
||||||
primitiveType == "bytes27" ||
|
|
||||||
primitiveType == "bytes27[]" ||
|
|
||||||
primitiveType == "bytes28" ||
|
|
||||||
primitiveType == "bytes28[]" ||
|
|
||||||
primitiveType == "bytes29" ||
|
|
||||||
primitiveType == "bytes29[]" ||
|
|
||||||
primitiveType == "bytes30" ||
|
|
||||||
primitiveType == "bytes30[]" ||
|
|
||||||
primitiveType == "bytes31" ||
|
|
||||||
primitiveType == "bytes31[]" ||
|
|
||||||
primitiveType == "bytes32" ||
|
|
||||||
primitiveType == "bytes32[]" {
|
|
||||||
return true
|
|
||||||
}
|
|
||||||
if primitiveType == "int" ||
|
|
||||||
primitiveType == "int[]" ||
|
primitiveType == "int[]" ||
|
||||||
primitiveType == "int8" ||
|
primitiveType == "uint" ||
|
||||||
primitiveType == "int8[]" ||
|
primitiveType == "uint[]" {
|
||||||
primitiveType == "int16" ||
|
|
||||||
primitiveType == "int16[]" ||
|
|
||||||
primitiveType == "int32" ||
|
|
||||||
primitiveType == "int32[]" ||
|
|
||||||
primitiveType == "int64" ||
|
|
||||||
primitiveType == "int64[]" ||
|
|
||||||
primitiveType == "int96" ||
|
|
||||||
primitiveType == "int96[]" ||
|
|
||||||
primitiveType == "int128" ||
|
|
||||||
primitiveType == "int128[]" ||
|
|
||||||
primitiveType == "int256" ||
|
|
||||||
primitiveType == "int256[]" {
|
|
||||||
return true
|
return true
|
||||||
}
|
}
|
||||||
if primitiveType == "uint" ||
|
// For 'bytesN', 'bytesN[]', we allow N from 1 to 32
|
||||||
primitiveType == "uint[]" ||
|
for n := 1; n <= 32; n++ {
|
||||||
primitiveType == "uint8" ||
|
// e.g. 'bytes28' or 'bytes28[]'
|
||||||
primitiveType == "uint8[]" ||
|
if primitiveType == fmt.Sprintf("bytes%d", n) || primitiveType == fmt.Sprintf("bytes%d[]", n) {
|
||||||
primitiveType == "uint16" ||
|
return true
|
||||||
primitiveType == "uint16[]" ||
|
}
|
||||||
primitiveType == "uint32" ||
|
}
|
||||||
primitiveType == "uint32[]" ||
|
// For 'intN','intN[]' and 'uintN','uintN[]' we allow N in increments of 8, from 8 up to 256
|
||||||
primitiveType == "uint64" ||
|
for n := 8; n <= 256; n += 8 {
|
||||||
primitiveType == "uint64[]" ||
|
if primitiveType == fmt.Sprintf("int%d", n) || primitiveType == fmt.Sprintf("int%d[]", n) {
|
||||||
primitiveType == "uint96" ||
|
return true
|
||||||
primitiveType == "uint96[]" ||
|
}
|
||||||
primitiveType == "uint128" ||
|
if primitiveType == fmt.Sprintf("uint%d", n) || primitiveType == fmt.Sprintf("uint%d[]", n) {
|
||||||
primitiveType == "uint128[]" ||
|
return true
|
||||||
primitiveType == "uint256" ||
|
}
|
||||||
primitiveType == "uint256[]" {
|
|
||||||
return true
|
|
||||||
}
|
}
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
|
@ -1,4 +1,4 @@
|
|||||||
// Copyright 2020 The go-ethereum Authors
|
// Copyright 2023 The go-ethereum Authors
|
||||||
// This file is part of the go-ethereum library.
|
// This file is part of the go-ethereum library.
|
||||||
//
|
//
|
||||||
// The go-ethereum library is free software: you can redistribute it and/or modify
|
// The go-ethereum library is free software: you can redistribute it and/or modify
|
||||||
@ -14,27 +14,27 @@
|
|||||||
// You should have received a copy of the GNU Lesser General Public License
|
// You should have received a copy of the GNU Lesser General Public License
|
||||||
// along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
|
// along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
|
||||||
|
|
||||||
package main
|
package apitypes
|
||||||
|
|
||||||
import (
|
import "testing"
|
||||||
"fmt"
|
|
||||||
"os"
|
|
||||||
|
|
||||||
"github.com/ethereum/go-ethereum/tests/fuzzers/modexp"
|
func TestIsPrimitive(t *testing.T) {
|
||||||
)
|
// Expected positives
|
||||||
|
for i, tc := range []string{
|
||||||
func main() {
|
"int24", "int24[]", "uint88", "uint88[]", "uint", "uint[]", "int256", "int256[]",
|
||||||
if len(os.Args) != 2 {
|
"uint96", "uint96[]", "int96", "int96[]", "bytes17[]", "bytes17",
|
||||||
fmt.Fprintf(os.Stderr, "Usage: debug <file>\n")
|
} {
|
||||||
fmt.Fprintf(os.Stderr, "Example\n")
|
if !isPrimitiveTypeValid(tc) {
|
||||||
fmt.Fprintf(os.Stderr, " $ debug ../crashers/4bbef6857c733a87ecf6fd8b9e7238f65eb9862a\n")
|
t.Errorf("test %d: expected '%v' to be a valid primitive", i, tc)
|
||||||
os.Exit(1)
|
}
|
||||||
}
|
}
|
||||||
crasher := os.Args[1]
|
// Expected negatives
|
||||||
data, err := os.ReadFile(crasher)
|
for i, tc := range []string{
|
||||||
if err != nil {
|
"int257", "int257[]", "uint88 ", "uint88 []", "uint257", "uint-1[]",
|
||||||
fmt.Fprintf(os.Stderr, "error loading crasher %v: %v", crasher, err)
|
"uint0", "uint0[]", "int95", "int95[]", "uint1", "uint1[]", "bytes33[]", "bytess",
|
||||||
os.Exit(1)
|
} {
|
||||||
|
if isPrimitiveTypeValid(tc) {
|
||||||
|
t.Errorf("test %d: expected '%v' to not be a valid primitive", i, tc)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
modexp.Fuzz(data)
|
|
||||||
}
|
}
|
@ -1,90 +0,0 @@
|
|||||||
// Copyright 2022 The go-ethereum Authors
|
|
||||||
// This file is part of the go-ethereum library.
|
|
||||||
//
|
|
||||||
// The go-ethereum library is free software: you can redistribute it and/or modify
|
|
||||||
// it under the terms of the GNU Lesser General Public License as published by
|
|
||||||
// the Free Software Foundation, either version 3 of the License, or
|
|
||||||
// (at your option) any later version.
|
|
||||||
//
|
|
||||||
// The go-ethereum library is distributed in the hope that it will be useful,
|
|
||||||
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
||||||
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
||||||
// GNU Lesser General Public License for more details.
|
|
||||||
//
|
|
||||||
// You should have received a copy of the GNU Lesser General Public License
|
|
||||||
// along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
|
|
||||||
|
|
||||||
package modexp
|
|
||||||
|
|
||||||
import (
|
|
||||||
"fmt"
|
|
||||||
"math/big"
|
|
||||||
|
|
||||||
"github.com/ethereum/go-ethereum/common"
|
|
||||||
"github.com/ethereum/go-ethereum/core/vm"
|
|
||||||
big2 "github.com/holiman/big"
|
|
||||||
)
|
|
||||||
|
|
||||||
// Fuzz is the fuzzing entry-point.
|
|
||||||
// The function must return
|
|
||||||
//
|
|
||||||
// - 1 if the fuzzer should increase priority of the
|
|
||||||
// given input during subsequent fuzzing (for example, the input is lexically
|
|
||||||
// correct and was parsed successfully);
|
|
||||||
// - -1 if the input must not be added to corpus even if gives new coverage; and
|
|
||||||
// - 0 otherwise
|
|
||||||
//
|
|
||||||
// other values are reserved for future use.
|
|
||||||
func Fuzz(input []byte) int {
|
|
||||||
if len(input) <= 96 {
|
|
||||||
return -1
|
|
||||||
}
|
|
||||||
// Abort on too expensive inputs
|
|
||||||
precomp := vm.PrecompiledContractsBerlin[common.BytesToAddress([]byte{5})]
|
|
||||||
if gas := precomp.RequiredGas(input); gas > 40_000_000 {
|
|
||||||
return 0
|
|
||||||
}
|
|
||||||
var (
|
|
||||||
baseLen = new(big.Int).SetBytes(getData(input, 0, 32)).Uint64()
|
|
||||||
expLen = new(big.Int).SetBytes(getData(input, 32, 32)).Uint64()
|
|
||||||
modLen = new(big.Int).SetBytes(getData(input, 64, 32)).Uint64()
|
|
||||||
)
|
|
||||||
// Handle a special case when both the base and mod length is zero
|
|
||||||
if baseLen == 0 && modLen == 0 {
|
|
||||||
return -1
|
|
||||||
}
|
|
||||||
input = input[96:]
|
|
||||||
// Retrieve the operands and execute the exponentiation
|
|
||||||
var (
|
|
||||||
base = new(big.Int).SetBytes(getData(input, 0, baseLen))
|
|
||||||
exp = new(big.Int).SetBytes(getData(input, baseLen, expLen))
|
|
||||||
mod = new(big.Int).SetBytes(getData(input, baseLen+expLen, modLen))
|
|
||||||
base2 = new(big2.Int).SetBytes(getData(input, 0, baseLen))
|
|
||||||
exp2 = new(big2.Int).SetBytes(getData(input, baseLen, expLen))
|
|
||||||
mod2 = new(big2.Int).SetBytes(getData(input, baseLen+expLen, modLen))
|
|
||||||
)
|
|
||||||
if mod.BitLen() == 0 {
|
|
||||||
// Modulo 0 is undefined, return zero
|
|
||||||
return -1
|
|
||||||
}
|
|
||||||
var a = new(big2.Int).Exp(base2, exp2, mod2).String()
|
|
||||||
var b = new(big.Int).Exp(base, exp, mod).String()
|
|
||||||
if a != b {
|
|
||||||
panic(fmt.Sprintf("Inequality %#x ^ %#x mod %#x \n have %s\n want %s", base, exp, mod, a, b))
|
|
||||||
}
|
|
||||||
return 1
|
|
||||||
}
|
|
||||||
|
|
||||||
// getData returns a slice from the data based on the start and size and pads
|
|
||||||
// up to size with zero's. This function is overflow safe.
|
|
||||||
func getData(data []byte, start uint64, size uint64) []byte {
|
|
||||||
length := uint64(len(data))
|
|
||||||
if start > length {
|
|
||||||
start = length
|
|
||||||
}
|
|
||||||
end := start + size
|
|
||||||
if end > length {
|
|
||||||
end = length
|
|
||||||
}
|
|
||||||
return common.RightPadBytes(data[start:end], int(size))
|
|
||||||
}
|
|
@ -90,6 +90,19 @@ var Forks = map[string]*params.ChainConfig{
|
|||||||
PetersburgBlock: big.NewInt(0),
|
PetersburgBlock: big.NewInt(0),
|
||||||
IstanbulBlock: big.NewInt(0),
|
IstanbulBlock: big.NewInt(0),
|
||||||
},
|
},
|
||||||
|
"MuirGlacier": {
|
||||||
|
ChainID: big.NewInt(1),
|
||||||
|
HomesteadBlock: big.NewInt(0),
|
||||||
|
EIP150Block: big.NewInt(0),
|
||||||
|
EIP155Block: big.NewInt(0),
|
||||||
|
EIP158Block: big.NewInt(0),
|
||||||
|
DAOForkBlock: big.NewInt(0),
|
||||||
|
ByzantiumBlock: big.NewInt(0),
|
||||||
|
ConstantinopleBlock: big.NewInt(0),
|
||||||
|
PetersburgBlock: big.NewInt(0),
|
||||||
|
IstanbulBlock: big.NewInt(0),
|
||||||
|
MuirGlacierBlock: big.NewInt(0),
|
||||||
|
},
|
||||||
"FrontierToHomesteadAt5": {
|
"FrontierToHomesteadAt5": {
|
||||||
ChainID: big.NewInt(1),
|
ChainID: big.NewInt(1),
|
||||||
HomesteadBlock: big.NewInt(5),
|
HomesteadBlock: big.NewInt(5),
|
||||||
|
@ -228,7 +228,7 @@ func runBenchmark(b *testing.B, t *StateTest) {
|
|||||||
evm := vm.NewEVM(context, txContext, statedb, config, vmconfig)
|
evm := vm.NewEVM(context, txContext, statedb, config, vmconfig)
|
||||||
|
|
||||||
// Create "contract" for sender to cache code analysis.
|
// Create "contract" for sender to cache code analysis.
|
||||||
sender := vm.NewContract(vm.AccountRef(msg.From()), vm.AccountRef(msg.From()),
|
sender := vm.NewContract(vm.AccountRef(msg.From), vm.AccountRef(msg.From),
|
||||||
nil, 0)
|
nil, 0)
|
||||||
|
|
||||||
var (
|
var (
|
||||||
@ -239,12 +239,12 @@ func runBenchmark(b *testing.B, t *StateTest) {
|
|||||||
b.ResetTimer()
|
b.ResetTimer()
|
||||||
for n := 0; n < b.N; n++ {
|
for n := 0; n < b.N; n++ {
|
||||||
snapshot := statedb.Snapshot()
|
snapshot := statedb.Snapshot()
|
||||||
statedb.Prepare(rules, msg.From(), context.Coinbase, msg.To(), vm.ActivePrecompiles(rules), msg.AccessList())
|
statedb.Prepare(rules, msg.From, context.Coinbase, msg.To, vm.ActivePrecompiles(rules), msg.AccessList)
|
||||||
b.StartTimer()
|
b.StartTimer()
|
||||||
start := time.Now()
|
start := time.Now()
|
||||||
|
|
||||||
// Execute the message.
|
// Execute the message.
|
||||||
_, leftOverGas, err := evm.Call(sender, *msg.To(), msg.Data(), msg.Gas(), msg.Value())
|
_, leftOverGas, err := evm.Call(sender, *msg.To, msg.Data, msg.GasLimit, msg.Value)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
b.Error(err)
|
b.Error(err)
|
||||||
return
|
return
|
||||||
@ -253,7 +253,7 @@ func runBenchmark(b *testing.B, t *StateTest) {
|
|||||||
b.StopTimer()
|
b.StopTimer()
|
||||||
elapsed += uint64(time.Since(start))
|
elapsed += uint64(time.Since(start))
|
||||||
refund += statedb.GetRefund()
|
refund += statedb.GetRefund()
|
||||||
gasUsed += msg.Gas() - leftOverGas
|
gasUsed += msg.GasLimit - leftOverGas
|
||||||
|
|
||||||
statedb.RevertToSnapshot(snapshot)
|
statedb.RevertToSnapshot(snapshot)
|
||||||
}
|
}
|
||||||
|
@ -329,7 +329,7 @@ func (t *StateTest) genesis(config *params.ChainConfig) *core.Genesis {
|
|||||||
return genesis
|
return genesis
|
||||||
}
|
}
|
||||||
|
|
||||||
func (tx *stTransaction) toMessage(ps stPostState, baseFee *big.Int) (core.Message, error) {
|
func (tx *stTransaction) toMessage(ps stPostState, baseFee *big.Int) (*core.Message, error) {
|
||||||
// Derive sender from private key if present.
|
// Derive sender from private key if present.
|
||||||
var from common.Address
|
var from common.Address
|
||||||
if len(tx.PrivateKey) > 0 {
|
if len(tx.PrivateKey) > 0 {
|
||||||
@ -397,8 +397,18 @@ func (tx *stTransaction) toMessage(ps stPostState, baseFee *big.Int) (core.Messa
|
|||||||
return nil, fmt.Errorf("no gas price provided")
|
return nil, fmt.Errorf("no gas price provided")
|
||||||
}
|
}
|
||||||
|
|
||||||
msg := types.NewMessage(from, to, tx.Nonce, value, gasLimit, gasPrice,
|
msg := &core.Message{
|
||||||
tx.MaxFeePerGas, tx.MaxPriorityFeePerGas, data, accessList, false)
|
From: from,
|
||||||
|
To: to,
|
||||||
|
Nonce: tx.Nonce,
|
||||||
|
Value: value,
|
||||||
|
GasLimit: gasLimit,
|
||||||
|
GasPrice: gasPrice,
|
||||||
|
GasFeeCap: tx.MaxFeePerGas,
|
||||||
|
GasTipCap: tx.MaxPriorityFeePerGas,
|
||||||
|
Data: data,
|
||||||
|
AccessList: accessList,
|
||||||
|
}
|
||||||
return msg, nil
|
return msg, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -33,29 +33,20 @@ type leaf struct {
|
|||||||
// insertion order.
|
// insertion order.
|
||||||
type committer struct {
|
type committer struct {
|
||||||
nodes *NodeSet
|
nodes *NodeSet
|
||||||
tracer *tracer
|
|
||||||
collectLeaf bool
|
collectLeaf bool
|
||||||
}
|
}
|
||||||
|
|
||||||
// newCommitter creates a new committer or picks one from the pool.
|
// newCommitter creates a new committer or picks one from the pool.
|
||||||
func newCommitter(owner common.Hash, tracer *tracer, collectLeaf bool) *committer {
|
func newCommitter(nodeset *NodeSet, collectLeaf bool) *committer {
|
||||||
return &committer{
|
return &committer{
|
||||||
nodes: NewNodeSet(owner),
|
nodes: nodeset,
|
||||||
tracer: tracer,
|
|
||||||
collectLeaf: collectLeaf,
|
collectLeaf: collectLeaf,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Commit collapses a node down into a hash node and returns it along with
|
// Commit collapses a node down into a hash node.
|
||||||
// the modified nodeset.
|
func (c *committer) Commit(n node) hashNode {
|
||||||
func (c *committer) Commit(n node) (hashNode, *NodeSet) {
|
return c.commit(nil, n).(hashNode)
|
||||||
h := c.commit(nil, n)
|
|
||||||
// Some nodes can be deleted from trie which can't be captured
|
|
||||||
// by committer itself. Iterate all deleted nodes tracked by
|
|
||||||
// tracer and marked them as deleted only if they are present
|
|
||||||
// in database previously.
|
|
||||||
c.tracer.markDeletions(c.nodes)
|
|
||||||
return h.(hashNode), c.nodes
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// commit collapses a node down into a hash node and returns it.
|
// commit collapses a node down into a hash node and returns it.
|
||||||
@ -74,9 +65,7 @@ func (c *committer) commit(path []byte, n node) node {
|
|||||||
// If the child is fullNode, recursively commit,
|
// If the child is fullNode, recursively commit,
|
||||||
// otherwise it can only be hashNode or valueNode.
|
// otherwise it can only be hashNode or valueNode.
|
||||||
if _, ok := cn.Val.(*fullNode); ok {
|
if _, ok := cn.Val.(*fullNode); ok {
|
||||||
childV := c.commit(append(path, cn.Key...), cn.Val)
|
collapsed.Val = c.commit(append(path, cn.Key...), cn.Val)
|
||||||
|
|
||||||
collapsed.Val = childV
|
|
||||||
}
|
}
|
||||||
// The key needs to be copied, since we're adding it to the
|
// The key needs to be copied, since we're adding it to the
|
||||||
// modified nodeset.
|
// modified nodeset.
|
||||||
@ -85,12 +74,6 @@ func (c *committer) commit(path []byte, n node) node {
|
|||||||
if hn, ok := hashedNode.(hashNode); ok {
|
if hn, ok := hashedNode.(hashNode); ok {
|
||||||
return hn
|
return hn
|
||||||
}
|
}
|
||||||
// The short node now is embedded in its parent. Mark the node as
|
|
||||||
// deleted if it's present in database previously. It's equivalent
|
|
||||||
// as deletion from database's perspective.
|
|
||||||
if prev := c.tracer.getPrev(path); len(prev) != 0 {
|
|
||||||
c.nodes.markDeleted(path, prev)
|
|
||||||
}
|
|
||||||
return collapsed
|
return collapsed
|
||||||
case *fullNode:
|
case *fullNode:
|
||||||
hashedKids := c.commitChildren(path, cn)
|
hashedKids := c.commitChildren(path, cn)
|
||||||
@ -101,12 +84,6 @@ func (c *committer) commit(path []byte, n node) node {
|
|||||||
if hn, ok := hashedNode.(hashNode); ok {
|
if hn, ok := hashedNode.(hashNode); ok {
|
||||||
return hn
|
return hn
|
||||||
}
|
}
|
||||||
// The full node now is embedded in its parent. Mark the node as
|
|
||||||
// deleted if it's present in database previously. It's equivalent
|
|
||||||
// as deletion from database's perspective.
|
|
||||||
if prev := c.tracer.getPrev(path); len(prev) != 0 {
|
|
||||||
c.nodes.markDeleted(path, prev)
|
|
||||||
}
|
|
||||||
return collapsed
|
return collapsed
|
||||||
case hashNode:
|
case hashNode:
|
||||||
return cn
|
return cn
|
||||||
@ -134,8 +111,7 @@ func (c *committer) commitChildren(path []byte, n *fullNode) [17]node {
|
|||||||
// Commit the child recursively and store the "hashed" value.
|
// Commit the child recursively and store the "hashed" value.
|
||||||
// Note the returned node can be some embedded nodes, so it's
|
// Note the returned node can be some embedded nodes, so it's
|
||||||
// possible the type is not hashNode.
|
// possible the type is not hashNode.
|
||||||
hashed := c.commit(append(path, byte(i)), child)
|
children[i] = c.commit(append(path, byte(i)), child)
|
||||||
children[i] = hashed
|
|
||||||
}
|
}
|
||||||
// For the 17th child, it's possible the type is valuenode.
|
// For the 17th child, it's possible the type is valuenode.
|
||||||
if n.Children[16] != nil {
|
if n.Children[16] != nil {
|
||||||
@ -155,6 +131,12 @@ func (c *committer) store(path []byte, n node) node {
|
|||||||
// usually is leaf node). But small value (less than 32bytes) is not
|
// usually is leaf node). But small value (less than 32bytes) is not
|
||||||
// our target (leaves in account trie only).
|
// our target (leaves in account trie only).
|
||||||
if hash == nil {
|
if hash == nil {
|
||||||
|
// The node is embedded in its parent, in other words, this node
|
||||||
|
// will not be stored in the database independently, mark it as
|
||||||
|
// deleted only if the node was existent in database before.
|
||||||
|
if _, ok := c.nodes.accessList[string(path)]; ok {
|
||||||
|
c.nodes.markDeleted(path)
|
||||||
|
}
|
||||||
return n
|
return n
|
||||||
}
|
}
|
||||||
// We have the hash already, estimate the RLP encoding-size of the node.
|
// We have the hash already, estimate the RLP encoding-size of the node.
|
||||||
@ -169,7 +151,7 @@ func (c *committer) store(path []byte, n node) node {
|
|||||||
}
|
}
|
||||||
)
|
)
|
||||||
// Collect the dirty node to nodeset for return.
|
// Collect the dirty node to nodeset for return.
|
||||||
c.nodes.markUpdated(path, mnode, c.tracer.getPrev(path))
|
c.nodes.markUpdated(path, mnode)
|
||||||
|
|
||||||
// Collect the corresponding leaf node if it's required. We don't check
|
// Collect the corresponding leaf node if it's required. We don't check
|
||||||
// full node since it's impossible to store value in fullNode. The key
|
// full node since it's impossible to store value in fullNode. The key
|
||||||
|
@ -801,13 +801,12 @@ func (db *Database) Update(nodes *MergedNodeSet) error {
|
|||||||
}
|
}
|
||||||
for _, owner := range order {
|
for _, owner := range order {
|
||||||
subset := nodes.sets[owner]
|
subset := nodes.sets[owner]
|
||||||
for _, path := range subset.updates.order {
|
subset.forEachWithOrder(func(path string, n *memoryNode) {
|
||||||
n, ok := subset.updates.nodes[path]
|
if n.isDeleted() {
|
||||||
if !ok {
|
return // ignore deletion
|
||||||
return fmt.Errorf("missing node %x %v", owner, path)
|
|
||||||
}
|
}
|
||||||
db.insert(n.hash, int(n.size), n.node)
|
db.insert(n.hash, int(n.size), n.node)
|
||||||
}
|
})
|
||||||
}
|
}
|
||||||
// Link up the account trie and storage trie if the node points
|
// Link up the account trie and storage trie if the node points
|
||||||
// to an account trie leaf.
|
// to an account trie leaf.
|
||||||
|
117
trie/nodeset.go
117
trie/nodeset.go
@ -19,6 +19,7 @@ package trie
|
|||||||
import (
|
import (
|
||||||
"fmt"
|
"fmt"
|
||||||
"reflect"
|
"reflect"
|
||||||
|
"sort"
|
||||||
"strings"
|
"strings"
|
||||||
|
|
||||||
"github.com/ethereum/go-ethereum/common"
|
"github.com/ethereum/go-ethereum/common"
|
||||||
@ -40,8 +41,8 @@ var memoryNodeSize = int(reflect.TypeOf(memoryNode{}).Size())
|
|||||||
|
|
||||||
// memorySize returns the total memory size used by this node.
|
// memorySize returns the total memory size used by this node.
|
||||||
// nolint:unused
|
// nolint:unused
|
||||||
func (n *memoryNode) memorySize(key int) int {
|
func (n *memoryNode) memorySize(pathlen int) int {
|
||||||
return int(n.size) + memoryNodeSize + key
|
return int(n.size) + memoryNodeSize + pathlen
|
||||||
}
|
}
|
||||||
|
|
||||||
// rlp returns the raw rlp encoded blob of the cached trie node, either directly
|
// rlp returns the raw rlp encoded blob of the cached trie node, either directly
|
||||||
@ -64,7 +65,13 @@ func (n *memoryNode) obj() node {
|
|||||||
return expandNode(n.hash[:], n.node)
|
return expandNode(n.hash[:], n.node)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// isDeleted returns the indicator if the node is marked as deleted.
|
||||||
|
func (n *memoryNode) isDeleted() bool {
|
||||||
|
return n.hash == (common.Hash{})
|
||||||
|
}
|
||||||
|
|
||||||
// nodeWithPrev wraps the memoryNode with the previous node value.
|
// nodeWithPrev wraps the memoryNode with the previous node value.
|
||||||
|
// nolint: unused
|
||||||
type nodeWithPrev struct {
|
type nodeWithPrev struct {
|
||||||
*memoryNode
|
*memoryNode
|
||||||
prev []byte // RLP-encoded previous value, nil means it's non-existent
|
prev []byte // RLP-encoded previous value, nil means it's non-existent
|
||||||
@ -79,64 +86,60 @@ func (n *nodeWithPrev) unwrap() *memoryNode {
|
|||||||
// memorySize returns the total memory size used by this node. It overloads
|
// memorySize returns the total memory size used by this node. It overloads
|
||||||
// the function in memoryNode by counting the size of previous value as well.
|
// the function in memoryNode by counting the size of previous value as well.
|
||||||
// nolint: unused
|
// nolint: unused
|
||||||
func (n *nodeWithPrev) memorySize(key int) int {
|
func (n *nodeWithPrev) memorySize(pathlen int) int {
|
||||||
return n.memoryNode.memorySize(key) + len(n.prev)
|
return n.memoryNode.memorySize(pathlen) + len(n.prev)
|
||||||
}
|
|
||||||
|
|
||||||
// nodesWithOrder represents a collection of dirty nodes which includes
|
|
||||||
// newly-inserted and updated nodes. The modification order of all nodes
|
|
||||||
// is represented by order list.
|
|
||||||
type nodesWithOrder struct {
|
|
||||||
order []string // the path list of dirty nodes, sort by insertion order
|
|
||||||
nodes map[string]*nodeWithPrev // the map of dirty nodes, keyed by node path
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// NodeSet contains all dirty nodes collected during the commit operation.
|
// NodeSet contains all dirty nodes collected during the commit operation.
|
||||||
// Each node is keyed by path. It's not thread-safe to use.
|
// Each node is keyed by path. It's not thread-safe to use.
|
||||||
type NodeSet struct {
|
type NodeSet struct {
|
||||||
owner common.Hash // the identifier of the trie
|
owner common.Hash // the identifier of the trie
|
||||||
updates *nodesWithOrder // the set of updated nodes(newly inserted, updated)
|
nodes map[string]*memoryNode // the set of dirty nodes(inserted, updated, deleted)
|
||||||
deletes map[string][]byte // the map of deleted nodes, keyed by node
|
leaves []*leaf // the list of dirty leaves
|
||||||
leaves []*leaf // the list of dirty leaves
|
updates int // the count of updated and inserted nodes
|
||||||
|
deletes int // the count of deleted nodes
|
||||||
|
|
||||||
|
// The list of accessed nodes, which records the original node value.
|
||||||
|
// The origin value is expected to be nil for newly inserted node
|
||||||
|
// and is expected to be non-nil for other types(updated, deleted).
|
||||||
|
accessList map[string][]byte
|
||||||
}
|
}
|
||||||
|
|
||||||
// NewNodeSet initializes an empty node set to be used for tracking dirty nodes
|
// NewNodeSet initializes an empty node set to be used for tracking dirty nodes
|
||||||
// from a specific account or storage trie. The owner is zero for the account
|
// from a specific account or storage trie. The owner is zero for the account
|
||||||
// trie and the owning account address hash for storage tries.
|
// trie and the owning account address hash for storage tries.
|
||||||
func NewNodeSet(owner common.Hash) *NodeSet {
|
func NewNodeSet(owner common.Hash, accessList map[string][]byte) *NodeSet {
|
||||||
return &NodeSet{
|
return &NodeSet{
|
||||||
owner: owner,
|
owner: owner,
|
||||||
updates: &nodesWithOrder{
|
nodes: make(map[string]*memoryNode),
|
||||||
nodes: make(map[string]*nodeWithPrev),
|
accessList: accessList,
|
||||||
},
|
|
||||||
deletes: make(map[string][]byte),
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
// forEachWithOrder iterates the dirty nodes with the order from bottom to top,
|
||||||
// NewNodeSetWithDeletion initializes the nodeset with provided deletion set.
|
// right to left, nodes with the longest path will be iterated first.
|
||||||
func NewNodeSetWithDeletion(owner common.Hash, paths [][]byte, prev [][]byte) *NodeSet {
|
func (set *NodeSet) forEachWithOrder(callback func(path string, n *memoryNode)) {
|
||||||
set := NewNodeSet(owner)
|
var paths sort.StringSlice
|
||||||
for i, path := range paths {
|
for path := range set.nodes {
|
||||||
set.markDeleted(path, prev[i])
|
paths = append(paths, path)
|
||||||
}
|
}
|
||||||
return set
|
// Bottom-up, longest path first
|
||||||
}
|
sort.Sort(sort.Reverse(paths))
|
||||||
*/
|
for _, path := range paths {
|
||||||
|
callback(path, set.nodes[path])
|
||||||
// markUpdated marks the node as dirty(newly-inserted or updated) with provided
|
|
||||||
// node path, node object along with its previous value.
|
|
||||||
func (set *NodeSet) markUpdated(path []byte, node *memoryNode, prev []byte) {
|
|
||||||
set.updates.order = append(set.updates.order, string(path))
|
|
||||||
set.updates.nodes[string(path)] = &nodeWithPrev{
|
|
||||||
memoryNode: node,
|
|
||||||
prev: prev,
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// markDeleted marks the node as deleted with provided path and previous value.
|
// markUpdated marks the node as dirty(newly-inserted or updated).
|
||||||
func (set *NodeSet) markDeleted(path []byte, prev []byte) {
|
func (set *NodeSet) markUpdated(path []byte, node *memoryNode) {
|
||||||
set.deletes[string(path)] = prev
|
set.nodes[string(path)] = node
|
||||||
|
set.updates += 1
|
||||||
|
}
|
||||||
|
|
||||||
|
// markDeleted marks the node as deleted.
|
||||||
|
func (set *NodeSet) markDeleted(path []byte) {
|
||||||
|
set.nodes[string(path)] = &memoryNode{}
|
||||||
|
set.deletes += 1
|
||||||
}
|
}
|
||||||
|
|
||||||
// addLeaf collects the provided leaf node into set.
|
// addLeaf collects the provided leaf node into set.
|
||||||
@ -144,16 +147,16 @@ func (set *NodeSet) addLeaf(node *leaf) {
|
|||||||
set.leaves = append(set.leaves, node)
|
set.leaves = append(set.leaves, node)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Size returns the number of updated and deleted nodes contained in the set.
|
// Size returns the number of dirty nodes in set.
|
||||||
func (set *NodeSet) Size() (int, int) {
|
func (set *NodeSet) Size() (int, int) {
|
||||||
return len(set.updates.order), len(set.deletes)
|
return set.updates, set.deletes
|
||||||
}
|
}
|
||||||
|
|
||||||
// Hashes returns the hashes of all updated nodes. TODO(rjl493456442) how can
|
// Hashes returns the hashes of all updated nodes. TODO(rjl493456442) how can
|
||||||
// we get rid of it?
|
// we get rid of it?
|
||||||
func (set *NodeSet) Hashes() []common.Hash {
|
func (set *NodeSet) Hashes() []common.Hash {
|
||||||
var ret []common.Hash
|
var ret []common.Hash
|
||||||
for _, node := range set.updates.nodes {
|
for _, node := range set.nodes {
|
||||||
ret = append(ret, node.hash)
|
ret = append(ret, node.hash)
|
||||||
}
|
}
|
||||||
return ret
|
return ret
|
||||||
@ -163,19 +166,23 @@ func (set *NodeSet) Hashes() []common.Hash {
|
|||||||
func (set *NodeSet) Summary() string {
|
func (set *NodeSet) Summary() string {
|
||||||
var out = new(strings.Builder)
|
var out = new(strings.Builder)
|
||||||
fmt.Fprintf(out, "nodeset owner: %v\n", set.owner)
|
fmt.Fprintf(out, "nodeset owner: %v\n", set.owner)
|
||||||
if set.updates != nil {
|
if set.nodes != nil {
|
||||||
for _, key := range set.updates.order {
|
for path, n := range set.nodes {
|
||||||
updated := set.updates.nodes[key]
|
// Deletion
|
||||||
if updated.prev != nil {
|
if n.isDeleted() {
|
||||||
fmt.Fprintf(out, " [*]: %x -> %v prev: %x\n", key, updated.hash, updated.prev)
|
fmt.Fprintf(out, " [-]: %x prev: %x\n", path, set.accessList[path])
|
||||||
} else {
|
continue
|
||||||
fmt.Fprintf(out, " [+]: %x -> %v\n", key, updated.hash)
|
|
||||||
}
|
}
|
||||||
|
// Insertion
|
||||||
|
origin, ok := set.accessList[path]
|
||||||
|
if !ok {
|
||||||
|
fmt.Fprintf(out, " [+]: %x -> %v\n", path, n.hash)
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
// Update
|
||||||
|
fmt.Fprintf(out, " [*]: %x -> %v prev: %x\n", path, n.hash, origin)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
for k, n := range set.deletes {
|
|
||||||
fmt.Fprintf(out, " [-]: %x -> %x\n", k, n)
|
|
||||||
}
|
|
||||||
for _, n := range set.leaves {
|
for _, n := range set.leaves {
|
||||||
fmt.Fprintf(out, "[leaf]: %v\n", n)
|
fmt.Fprintf(out, "[leaf]: %v\n", n)
|
||||||
}
|
}
|
||||||
|
@ -563,7 +563,7 @@ func VerifyRangeProof(rootHash common.Hash, firstKey []byte, lastKey []byte, key
|
|||||||
}
|
}
|
||||||
// Rebuild the trie with the leaf stream, the shape of trie
|
// Rebuild the trie with the leaf stream, the shape of trie
|
||||||
// should be same with the original one.
|
// should be same with the original one.
|
||||||
tr := &Trie{root: root, reader: newEmptyReader()}
|
tr := &Trie{root: root, reader: newEmptyReader(), tracer: newTracer()}
|
||||||
if empty {
|
if empty {
|
||||||
tr.root = nil
|
tr.root = nil
|
||||||
}
|
}
|
||||||
|
@ -20,6 +20,7 @@ import (
|
|||||||
"bytes"
|
"bytes"
|
||||||
crand "crypto/rand"
|
crand "crypto/rand"
|
||||||
"encoding/binary"
|
"encoding/binary"
|
||||||
|
"fmt"
|
||||||
mrand "math/rand"
|
mrand "math/rand"
|
||||||
"sort"
|
"sort"
|
||||||
"testing"
|
"testing"
|
||||||
@ -30,6 +31,24 @@ import (
|
|||||||
"github.com/ethereum/go-ethereum/ethdb/memorydb"
|
"github.com/ethereum/go-ethereum/ethdb/memorydb"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
// Prng is a pseudo random number generator seeded by strong randomness.
|
||||||
|
// The randomness is printed on startup in order to make failures reproducible.
|
||||||
|
var prng = initRnd()
|
||||||
|
|
||||||
|
func initRnd() *mrand.Rand {
|
||||||
|
var seed [8]byte
|
||||||
|
crand.Read(seed[:])
|
||||||
|
rnd := mrand.New(mrand.NewSource(int64(binary.LittleEndian.Uint64(seed[:]))))
|
||||||
|
fmt.Printf("Seed: %x\n", seed)
|
||||||
|
return rnd
|
||||||
|
}
|
||||||
|
|
||||||
|
func randBytes(n int) []byte {
|
||||||
|
r := make([]byte, n)
|
||||||
|
prng.Read(r)
|
||||||
|
return r
|
||||||
|
}
|
||||||
|
|
||||||
// makeProvers creates Merkle trie provers based on different implementations to
|
// makeProvers creates Merkle trie provers based on different implementations to
|
||||||
// test all variations.
|
// test all variations.
|
||||||
func makeProvers(trie *Trie) []func(key []byte) *memorydb.Database {
|
func makeProvers(trie *Trie) []func(key []byte) *memorydb.Database {
|
||||||
@ -1041,12 +1060,6 @@ func randomTrie(n int) (*Trie, map[string]*kv) {
|
|||||||
return trie, vals
|
return trie, vals
|
||||||
}
|
}
|
||||||
|
|
||||||
func randBytes(n int) []byte {
|
|
||||||
r := make([]byte, n)
|
|
||||||
crand.Read(r)
|
|
||||||
return r
|
|
||||||
}
|
|
||||||
|
|
||||||
func nonRandomTrie(n int) (*Trie, map[string]*kv) {
|
func nonRandomTrie(n int) (*Trie, map[string]*kv) {
|
||||||
trie := NewEmpty(NewDatabase(rawdb.NewMemoryDatabase()))
|
trie := NewEmpty(NewDatabase(rawdb.NewMemoryDatabase()))
|
||||||
vals := make(map[string]*kv)
|
vals := make(map[string]*kv)
|
||||||
|
@ -144,7 +144,9 @@ func (st *StackTrie) unmarshalBinary(r io.Reader) error {
|
|||||||
Val []byte
|
Val []byte
|
||||||
Key []byte
|
Key []byte
|
||||||
}
|
}
|
||||||
gob.NewDecoder(r).Decode(&dec)
|
if err := gob.NewDecoder(r).Decode(&dec); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
st.owner = dec.Owner
|
st.owner = dec.Owner
|
||||||
st.nodeType = dec.NodeType
|
st.nodeType = dec.NodeType
|
||||||
st.val = dec.Val
|
st.val = dec.Val
|
||||||
@ -158,7 +160,9 @@ func (st *StackTrie) unmarshalBinary(r io.Reader) error {
|
|||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
var child StackTrie
|
var child StackTrie
|
||||||
child.unmarshalBinary(r)
|
if err := child.unmarshalBinary(r); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
st.children[i] = &child
|
st.children[i] = &child
|
||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
|
@ -434,6 +434,7 @@ func TestDuplicateAvoidanceSync(t *testing.T) {
|
|||||||
// Tests that at any point in time during a sync, only complete sub-tries are in
|
// Tests that at any point in time during a sync, only complete sub-tries are in
|
||||||
// the database.
|
// the database.
|
||||||
func TestIncompleteSync(t *testing.T) {
|
func TestIncompleteSync(t *testing.T) {
|
||||||
|
t.Parallel()
|
||||||
// Create a random trie to copy
|
// Create a random trie to copy
|
||||||
srcDb, srcTrie, _ := makeTestTrie()
|
srcDb, srcTrie, _ := makeTestTrie()
|
||||||
|
|
||||||
|
125
trie/tracer.go
Normal file
125
trie/tracer.go
Normal file
@ -0,0 +1,125 @@
|
|||||||
|
// Copyright 2022 The go-ethereum Authors
|
||||||
|
// This file is part of the go-ethereum library.
|
||||||
|
//
|
||||||
|
// The go-ethereum library is free software: you can redistribute it and/or modify
|
||||||
|
// it under the terms of the GNU Lesser General Public License as published by
|
||||||
|
// the Free Software Foundation, either version 3 of the License, or
|
||||||
|
// (at your option) any later version.
|
||||||
|
//
|
||||||
|
// The go-ethereum library is distributed in the hope that it will be useful,
|
||||||
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||||
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||||
|
// GNU Lesser General Public License for more details.
|
||||||
|
//
|
||||||
|
// You should have received a copy of the GNU Lesser General Public License
|
||||||
|
// along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
|
||||||
|
|
||||||
|
package trie
|
||||||
|
|
||||||
|
import "github.com/ethereum/go-ethereum/common"
|
||||||
|
|
||||||
|
// tracer tracks the changes of trie nodes. During the trie operations,
|
||||||
|
// some nodes can be deleted from the trie, while these deleted nodes
|
||||||
|
// won't be captured by trie.Hasher or trie.Committer. Thus, these deleted
|
||||||
|
// nodes won't be removed from the disk at all. Tracer is an auxiliary tool
|
||||||
|
// used to track all insert and delete operations of trie and capture all
|
||||||
|
// deleted nodes eventually.
|
||||||
|
//
|
||||||
|
// The changed nodes can be mainly divided into two categories: the leaf
|
||||||
|
// node and intermediate node. The former is inserted/deleted by callers
|
||||||
|
// while the latter is inserted/deleted in order to follow the rule of trie.
|
||||||
|
// This tool can track all of them no matter the node is embedded in its
|
||||||
|
// parent or not, but valueNode is never tracked.
|
||||||
|
//
|
||||||
|
// Besides, it's also used for recording the original value of the nodes
|
||||||
|
// when they are resolved from the disk. The pre-value of the nodes will
|
||||||
|
// be used to construct trie history in the future.
|
||||||
|
//
|
||||||
|
// Note tracer is not thread-safe, callers should be responsible for handling
|
||||||
|
// the concurrency issues by themselves.
|
||||||
|
type tracer struct {
|
||||||
|
inserts map[string]struct{}
|
||||||
|
deletes map[string]struct{}
|
||||||
|
accessList map[string][]byte
|
||||||
|
}
|
||||||
|
|
||||||
|
// newTracer initializes the tracer for capturing trie changes.
|
||||||
|
func newTracer() *tracer {
|
||||||
|
return &tracer{
|
||||||
|
inserts: make(map[string]struct{}),
|
||||||
|
deletes: make(map[string]struct{}),
|
||||||
|
accessList: make(map[string][]byte),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// onRead tracks the newly loaded trie node and caches the rlp-encoded
|
||||||
|
// blob internally. Don't change the value outside of function since
|
||||||
|
// it's not deep-copied.
|
||||||
|
func (t *tracer) onRead(path []byte, val []byte) {
|
||||||
|
t.accessList[string(path)] = val
|
||||||
|
}
|
||||||
|
|
||||||
|
// onInsert tracks the newly inserted trie node. If it's already
|
||||||
|
// in the deletion set (resurrected node), then just wipe it from
|
||||||
|
// the deletion set as it's "untouched".
|
||||||
|
func (t *tracer) onInsert(path []byte) {
|
||||||
|
if _, present := t.deletes[string(path)]; present {
|
||||||
|
delete(t.deletes, string(path))
|
||||||
|
return
|
||||||
|
}
|
||||||
|
t.inserts[string(path)] = struct{}{}
|
||||||
|
}
|
||||||
|
|
||||||
|
// onDelete tracks the newly deleted trie node. If it's already
|
||||||
|
// in the addition set, then just wipe it from the addition set
|
||||||
|
// as it's untouched.
|
||||||
|
func (t *tracer) onDelete(path []byte) {
|
||||||
|
if _, present := t.inserts[string(path)]; present {
|
||||||
|
delete(t.inserts, string(path))
|
||||||
|
return
|
||||||
|
}
|
||||||
|
t.deletes[string(path)] = struct{}{}
|
||||||
|
}
|
||||||
|
|
||||||
|
// reset clears the content tracked by tracer.
|
||||||
|
func (t *tracer) reset() {
|
||||||
|
t.inserts = make(map[string]struct{})
|
||||||
|
t.deletes = make(map[string]struct{})
|
||||||
|
t.accessList = make(map[string][]byte)
|
||||||
|
}
|
||||||
|
|
||||||
|
// copy returns a deep copied tracer instance.
|
||||||
|
func (t *tracer) copy() *tracer {
|
||||||
|
var (
|
||||||
|
inserts = make(map[string]struct{})
|
||||||
|
deletes = make(map[string]struct{})
|
||||||
|
accessList = make(map[string][]byte)
|
||||||
|
)
|
||||||
|
for path := range t.inserts {
|
||||||
|
inserts[path] = struct{}{}
|
||||||
|
}
|
||||||
|
for path := range t.deletes {
|
||||||
|
deletes[path] = struct{}{}
|
||||||
|
}
|
||||||
|
for path, blob := range t.accessList {
|
||||||
|
accessList[path] = common.CopyBytes(blob)
|
||||||
|
}
|
||||||
|
return &tracer{
|
||||||
|
inserts: inserts,
|
||||||
|
deletes: deletes,
|
||||||
|
accessList: accessList,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// markDeletions puts all tracked deletions into the provided nodeset.
|
||||||
|
func (t *tracer) markDeletions(set *NodeSet) {
|
||||||
|
for path := range t.deletes {
|
||||||
|
// It's possible a few deleted nodes were embedded
|
||||||
|
// in their parent before, the deletions can be no
|
||||||
|
// effect by deleting nothing, filter them out.
|
||||||
|
if _, ok := set.accessList[path]; !ok {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
set.markDeleted([]byte(path))
|
||||||
|
}
|
||||||
|
}
|
368
trie/tracer_test.go
Normal file
368
trie/tracer_test.go
Normal file
@ -0,0 +1,368 @@
|
|||||||
|
// Copyright 2022 The go-ethereum Authors
|
||||||
|
// This file is part of the go-ethereum library.
|
||||||
|
//
|
||||||
|
// The go-ethereum library is free software: you can redistribute it and/or modify
|
||||||
|
// it under the terms of the GNU Lesser General Public License as published by
|
||||||
|
// the Free Software Foundation, either version 3 of the License, or
|
||||||
|
// (at your option) any later version.
|
||||||
|
//
|
||||||
|
// The go-ethereum library is distributed in the hope that it will be useful,
|
||||||
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||||
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||||
|
// GNU Lesser General Public License for more details.
|
||||||
|
//
|
||||||
|
// You should have received a copy of the GNU Lesser General Public License
|
||||||
|
// along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
|
||||||
|
|
||||||
|
package trie
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bytes"
|
||||||
|
"testing"
|
||||||
|
|
||||||
|
"github.com/ethereum/go-ethereum/common"
|
||||||
|
"github.com/ethereum/go-ethereum/core/rawdb"
|
||||||
|
)
|
||||||
|
|
||||||
|
var (
|
||||||
|
tiny = []struct{ k, v string }{
|
||||||
|
{"k1", "v1"},
|
||||||
|
{"k2", "v2"},
|
||||||
|
{"k3", "v3"},
|
||||||
|
}
|
||||||
|
nonAligned = []struct{ k, v string }{
|
||||||
|
{"do", "verb"},
|
||||||
|
{"ether", "wookiedoo"},
|
||||||
|
{"horse", "stallion"},
|
||||||
|
{"shaman", "horse"},
|
||||||
|
{"doge", "coin"},
|
||||||
|
{"dog", "puppy"},
|
||||||
|
{"somethingveryoddindeedthis is", "myothernodedata"},
|
||||||
|
}
|
||||||
|
standard = []struct{ k, v string }{
|
||||||
|
{string(randBytes(32)), "verb"},
|
||||||
|
{string(randBytes(32)), "wookiedoo"},
|
||||||
|
{string(randBytes(32)), "stallion"},
|
||||||
|
{string(randBytes(32)), "horse"},
|
||||||
|
{string(randBytes(32)), "coin"},
|
||||||
|
{string(randBytes(32)), "puppy"},
|
||||||
|
{string(randBytes(32)), "myothernodedata"},
|
||||||
|
}
|
||||||
|
)
|
||||||
|
|
||||||
|
func TestTrieTracer(t *testing.T) {
|
||||||
|
testTrieTracer(t, tiny)
|
||||||
|
testTrieTracer(t, nonAligned)
|
||||||
|
testTrieTracer(t, standard)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Tests if the trie diffs are tracked correctly. Tracer should capture
|
||||||
|
// all non-leaf dirty nodes, no matter the node is embedded or not.
|
||||||
|
func testTrieTracer(t *testing.T, vals []struct{ k, v string }) {
|
||||||
|
db := NewDatabase(rawdb.NewMemoryDatabase())
|
||||||
|
trie := NewEmpty(db)
|
||||||
|
|
||||||
|
// Determine all new nodes are tracked
|
||||||
|
for _, val := range vals {
|
||||||
|
trie.Update([]byte(val.k), []byte(val.v))
|
||||||
|
}
|
||||||
|
insertSet := copySet(trie.tracer.inserts) // copy before commit
|
||||||
|
deleteSet := copySet(trie.tracer.deletes) // copy before commit
|
||||||
|
root, nodes := trie.Commit(false)
|
||||||
|
db.Update(NewWithNodeSet(nodes))
|
||||||
|
|
||||||
|
seen := setKeys(iterNodes(db, root))
|
||||||
|
if !compareSet(insertSet, seen) {
|
||||||
|
t.Fatal("Unexpected insertion set")
|
||||||
|
}
|
||||||
|
if !compareSet(deleteSet, nil) {
|
||||||
|
t.Fatal("Unexpected deletion set")
|
||||||
|
}
|
||||||
|
|
||||||
|
// Determine all deletions are tracked
|
||||||
|
trie, _ = New(TrieID(root), db)
|
||||||
|
for _, val := range vals {
|
||||||
|
trie.Delete([]byte(val.k))
|
||||||
|
}
|
||||||
|
insertSet, deleteSet = copySet(trie.tracer.inserts), copySet(trie.tracer.deletes)
|
||||||
|
if !compareSet(insertSet, nil) {
|
||||||
|
t.Fatal("Unexpected insertion set")
|
||||||
|
}
|
||||||
|
if !compareSet(deleteSet, seen) {
|
||||||
|
t.Fatal("Unexpected deletion set")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Test that after inserting a new batch of nodes and deleting them immediately,
|
||||||
|
// the trie tracer should be cleared normally as no operation happened.
|
||||||
|
func TestTrieTracerNoop(t *testing.T) {
|
||||||
|
testTrieTracerNoop(t, tiny)
|
||||||
|
testTrieTracerNoop(t, nonAligned)
|
||||||
|
testTrieTracerNoop(t, standard)
|
||||||
|
}
|
||||||
|
|
||||||
|
func testTrieTracerNoop(t *testing.T, vals []struct{ k, v string }) {
|
||||||
|
trie := NewEmpty(NewDatabase(rawdb.NewMemoryDatabase()))
|
||||||
|
for _, val := range vals {
|
||||||
|
trie.Update([]byte(val.k), []byte(val.v))
|
||||||
|
}
|
||||||
|
for _, val := range vals {
|
||||||
|
trie.Delete([]byte(val.k))
|
||||||
|
}
|
||||||
|
if len(trie.tracer.inserts) != 0 {
|
||||||
|
t.Fatal("Unexpected insertion set")
|
||||||
|
}
|
||||||
|
if len(trie.tracer.deletes) != 0 {
|
||||||
|
t.Fatal("Unexpected deletion set")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Tests if the accessList is correctly tracked.
|
||||||
|
func TestAccessList(t *testing.T) {
|
||||||
|
testAccessList(t, tiny)
|
||||||
|
testAccessList(t, nonAligned)
|
||||||
|
testAccessList(t, standard)
|
||||||
|
}
|
||||||
|
|
||||||
|
func testAccessList(t *testing.T, vals []struct{ k, v string }) {
|
||||||
|
var (
|
||||||
|
db = NewDatabase(rawdb.NewMemoryDatabase())
|
||||||
|
trie = NewEmpty(db)
|
||||||
|
orig = trie.Copy()
|
||||||
|
)
|
||||||
|
// Create trie from scratch
|
||||||
|
for _, val := range vals {
|
||||||
|
trie.Update([]byte(val.k), []byte(val.v))
|
||||||
|
}
|
||||||
|
root, nodes := trie.Commit(false)
|
||||||
|
db.Update(NewWithNodeSet(nodes))
|
||||||
|
|
||||||
|
trie, _ = New(TrieID(root), db)
|
||||||
|
if err := verifyAccessList(orig, trie, nodes); err != nil {
|
||||||
|
t.Fatalf("Invalid accessList %v", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Update trie
|
||||||
|
trie, _ = New(TrieID(root), db)
|
||||||
|
orig = trie.Copy()
|
||||||
|
for _, val := range vals {
|
||||||
|
trie.Update([]byte(val.k), randBytes(32))
|
||||||
|
}
|
||||||
|
root, nodes = trie.Commit(false)
|
||||||
|
db.Update(NewWithNodeSet(nodes))
|
||||||
|
|
||||||
|
trie, _ = New(TrieID(root), db)
|
||||||
|
if err := verifyAccessList(orig, trie, nodes); err != nil {
|
||||||
|
t.Fatalf("Invalid accessList %v", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Add more new nodes
|
||||||
|
trie, _ = New(TrieID(root), db)
|
||||||
|
orig = trie.Copy()
|
||||||
|
var keys []string
|
||||||
|
for i := 0; i < 30; i++ {
|
||||||
|
key := randBytes(32)
|
||||||
|
keys = append(keys, string(key))
|
||||||
|
trie.Update(key, randBytes(32))
|
||||||
|
}
|
||||||
|
root, nodes = trie.Commit(false)
|
||||||
|
db.Update(NewWithNodeSet(nodes))
|
||||||
|
|
||||||
|
trie, _ = New(TrieID(root), db)
|
||||||
|
if err := verifyAccessList(orig, trie, nodes); err != nil {
|
||||||
|
t.Fatalf("Invalid accessList %v", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Partial deletions
|
||||||
|
trie, _ = New(TrieID(root), db)
|
||||||
|
orig = trie.Copy()
|
||||||
|
for _, key := range keys {
|
||||||
|
trie.Update([]byte(key), nil)
|
||||||
|
}
|
||||||
|
root, nodes = trie.Commit(false)
|
||||||
|
db.Update(NewWithNodeSet(nodes))
|
||||||
|
|
||||||
|
trie, _ = New(TrieID(root), db)
|
||||||
|
if err := verifyAccessList(orig, trie, nodes); err != nil {
|
||||||
|
t.Fatalf("Invalid accessList %v", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Delete all
|
||||||
|
trie, _ = New(TrieID(root), db)
|
||||||
|
orig = trie.Copy()
|
||||||
|
for _, val := range vals {
|
||||||
|
trie.Update([]byte(val.k), nil)
|
||||||
|
}
|
||||||
|
root, nodes = trie.Commit(false)
|
||||||
|
db.Update(NewWithNodeSet(nodes))
|
||||||
|
|
||||||
|
trie, _ = New(TrieID(root), db)
|
||||||
|
if err := verifyAccessList(orig, trie, nodes); err != nil {
|
||||||
|
t.Fatalf("Invalid accessList %v", err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Tests origin values won't be tracked in Iterator or Prover
|
||||||
|
func TestAccessListLeak(t *testing.T) {
|
||||||
|
var (
|
||||||
|
db = NewDatabase(rawdb.NewMemoryDatabase())
|
||||||
|
trie = NewEmpty(db)
|
||||||
|
)
|
||||||
|
// Create trie from scratch
|
||||||
|
for _, val := range standard {
|
||||||
|
trie.Update([]byte(val.k), []byte(val.v))
|
||||||
|
}
|
||||||
|
root, nodes := trie.Commit(false)
|
||||||
|
db.Update(NewWithNodeSet(nodes))
|
||||||
|
|
||||||
|
var cases = []struct {
|
||||||
|
op func(tr *Trie)
|
||||||
|
}{
|
||||||
|
{
|
||||||
|
func(tr *Trie) {
|
||||||
|
it := tr.NodeIterator(nil)
|
||||||
|
for it.Next(true) {
|
||||||
|
}
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
func(tr *Trie) {
|
||||||
|
it := NewIterator(tr.NodeIterator(nil))
|
||||||
|
for it.Next() {
|
||||||
|
}
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
func(tr *Trie) {
|
||||||
|
for _, val := range standard {
|
||||||
|
tr.Prove([]byte(val.k), 0, rawdb.NewMemoryDatabase())
|
||||||
|
}
|
||||||
|
},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
for _, c := range cases {
|
||||||
|
trie, _ = New(TrieID(root), db)
|
||||||
|
n1 := len(trie.tracer.accessList)
|
||||||
|
c.op(trie)
|
||||||
|
n2 := len(trie.tracer.accessList)
|
||||||
|
|
||||||
|
if n1 != n2 {
|
||||||
|
t.Fatalf("AccessList is leaked, prev %d after %d", n1, n2)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Tests whether the original tree node is correctly deleted after being embedded
|
||||||
|
// in its parent due to the smaller size of the original tree node.
|
||||||
|
func TestTinyTree(t *testing.T) {
|
||||||
|
var (
|
||||||
|
db = NewDatabase(rawdb.NewMemoryDatabase())
|
||||||
|
trie = NewEmpty(db)
|
||||||
|
)
|
||||||
|
for _, val := range tiny {
|
||||||
|
trie.Update([]byte(val.k), randBytes(32))
|
||||||
|
}
|
||||||
|
root, set := trie.Commit(false)
|
||||||
|
db.Update(NewWithNodeSet(set))
|
||||||
|
|
||||||
|
trie, _ = New(TrieID(root), db)
|
||||||
|
orig := trie.Copy()
|
||||||
|
for _, val := range tiny {
|
||||||
|
trie.Update([]byte(val.k), []byte(val.v))
|
||||||
|
}
|
||||||
|
root, set = trie.Commit(false)
|
||||||
|
db.Update(NewWithNodeSet(set))
|
||||||
|
|
||||||
|
trie, _ = New(TrieID(root), db)
|
||||||
|
if err := verifyAccessList(orig, trie, set); err != nil {
|
||||||
|
t.Fatalf("Invalid accessList %v", err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func compareSet(setA, setB map[string]struct{}) bool {
|
||||||
|
if len(setA) != len(setB) {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
for key := range setA {
|
||||||
|
if _, ok := setB[key]; !ok {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
|
||||||
|
func forNodes(tr *Trie) map[string][]byte {
|
||||||
|
var (
|
||||||
|
it = tr.NodeIterator(nil)
|
||||||
|
nodes = make(map[string][]byte)
|
||||||
|
)
|
||||||
|
for it.Next(true) {
|
||||||
|
if it.Leaf() {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
nodes[string(it.Path())] = common.CopyBytes(it.NodeBlob())
|
||||||
|
}
|
||||||
|
return nodes
|
||||||
|
}
|
||||||
|
|
||||||
|
func iterNodes(db *Database, root common.Hash) map[string][]byte {
|
||||||
|
tr, _ := New(TrieID(root), db)
|
||||||
|
return forNodes(tr)
|
||||||
|
}
|
||||||
|
|
||||||
|
func forHashedNodes(tr *Trie) map[string][]byte {
|
||||||
|
var (
|
||||||
|
it = tr.NodeIterator(nil)
|
||||||
|
nodes = make(map[string][]byte)
|
||||||
|
)
|
||||||
|
for it.Next(true) {
|
||||||
|
if it.Hash() == (common.Hash{}) {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
nodes[string(it.Path())] = common.CopyBytes(it.NodeBlob())
|
||||||
|
}
|
||||||
|
return nodes
|
||||||
|
}
|
||||||
|
|
||||||
|
func diffTries(trieA, trieB *Trie) (map[string][]byte, map[string][]byte, map[string][]byte) {
|
||||||
|
var (
|
||||||
|
nodesA = forHashedNodes(trieA)
|
||||||
|
nodesB = forHashedNodes(trieB)
|
||||||
|
inA = make(map[string][]byte) // hashed nodes in trie a but not b
|
||||||
|
inB = make(map[string][]byte) // hashed nodes in trie b but not a
|
||||||
|
both = make(map[string][]byte) // hashed nodes in both tries but different value
|
||||||
|
)
|
||||||
|
for path, blobA := range nodesA {
|
||||||
|
if blobB, ok := nodesB[path]; ok {
|
||||||
|
if bytes.Equal(blobA, blobB) {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
both[path] = blobA
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
inA[path] = blobA
|
||||||
|
}
|
||||||
|
for path, blobB := range nodesB {
|
||||||
|
if _, ok := nodesA[path]; ok {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
inB[path] = blobB
|
||||||
|
}
|
||||||
|
return inA, inB, both
|
||||||
|
}
|
||||||
|
|
||||||
|
func setKeys(set map[string][]byte) map[string]struct{} {
|
||||||
|
keys := make(map[string]struct{})
|
||||||
|
for k := range set {
|
||||||
|
keys[k] = struct{}{}
|
||||||
|
}
|
||||||
|
return keys
|
||||||
|
}
|
||||||
|
|
||||||
|
func copySet(set map[string]struct{}) map[string]struct{} {
|
||||||
|
copied := make(map[string]struct{})
|
||||||
|
for k := range set {
|
||||||
|
copied[k] = struct{}{}
|
||||||
|
}
|
||||||
|
return copied
|
||||||
|
}
|
28
trie/trie.go
28
trie/trie.go
@ -81,7 +81,7 @@ func New(id *ID, db NodeReader) (*Trie, error) {
|
|||||||
trie := &Trie{
|
trie := &Trie{
|
||||||
owner: id.Owner,
|
owner: id.Owner,
|
||||||
reader: reader,
|
reader: reader,
|
||||||
//tracer: newTracer(),
|
tracer: newTracer(),
|
||||||
}
|
}
|
||||||
if id.Root != (common.Hash{}) && id.Root != types.EmptyRootHash {
|
if id.Root != (common.Hash{}) && id.Root != types.EmptyRootHash {
|
||||||
rootnode, err := trie.resolveAndTrack(id.Root[:], nil)
|
rootnode, err := trie.resolveAndTrack(id.Root[:], nil)
|
||||||
@ -547,7 +547,7 @@ func (t *Trie) resolveAndTrack(n hashNode, prefix []byte) (node, error) {
|
|||||||
// Hash returns the root hash of the trie. It does not write to the
|
// Hash returns the root hash of the trie. It does not write to the
|
||||||
// database and can be used even if the trie doesn't have one.
|
// database and can be used even if the trie doesn't have one.
|
||||||
func (t *Trie) Hash() common.Hash {
|
func (t *Trie) Hash() common.Hash {
|
||||||
hash, cached, _ := t.hashRoot()
|
hash, cached := t.hashRoot()
|
||||||
t.root = cached
|
t.root = cached
|
||||||
return common.BytesToHash(hash.(hashNode))
|
return common.BytesToHash(hash.(hashNode))
|
||||||
}
|
}
|
||||||
@ -561,14 +561,14 @@ func (t *Trie) Hash() common.Hash {
|
|||||||
func (t *Trie) Commit(collectLeaf bool) (common.Hash, *NodeSet) {
|
func (t *Trie) Commit(collectLeaf bool) (common.Hash, *NodeSet) {
|
||||||
defer t.tracer.reset()
|
defer t.tracer.reset()
|
||||||
|
|
||||||
|
nodes := NewNodeSet(t.owner, t.tracer.accessList)
|
||||||
|
t.tracer.markDeletions(nodes)
|
||||||
|
|
||||||
// Trie is empty and can be classified into two types of situations:
|
// Trie is empty and can be classified into two types of situations:
|
||||||
// - The trie was empty and no update happens
|
// - The trie was empty and no update happens
|
||||||
// - The trie was non-empty and all nodes are dropped
|
// - The trie was non-empty and all nodes are dropped
|
||||||
if t.root == nil {
|
if t.root == nil {
|
||||||
// Wrap tracked deletions as the return
|
return types.EmptyRootHash, nodes
|
||||||
set := NewNodeSet(t.owner)
|
|
||||||
t.tracer.markDeletions(set)
|
|
||||||
return types.EmptyRootHash, set
|
|
||||||
}
|
}
|
||||||
// Derive the hash for all dirty nodes first. We hold the assumption
|
// Derive the hash for all dirty nodes first. We hold the assumption
|
||||||
// in the following procedure that all nodes are hashed.
|
// in the following procedure that all nodes are hashed.
|
||||||
@ -582,23 +582,23 @@ func (t *Trie) Commit(collectLeaf bool) (common.Hash, *NodeSet) {
|
|||||||
t.root = hashedNode
|
t.root = hashedNode
|
||||||
return rootHash, nil
|
return rootHash, nil
|
||||||
}
|
}
|
||||||
h := newCommitter(t.owner, t.tracer, collectLeaf)
|
t.root = newCommitter(nodes, collectLeaf).Commit(t.root)
|
||||||
newRoot, nodes := h.Commit(t.root)
|
|
||||||
t.root = newRoot
|
|
||||||
return rootHash, nodes
|
return rootHash, nodes
|
||||||
}
|
}
|
||||||
|
|
||||||
// hashRoot calculates the root hash of the given trie
|
// hashRoot calculates the root hash of the given trie
|
||||||
func (t *Trie) hashRoot() (node, node, error) {
|
func (t *Trie) hashRoot() (node, node) {
|
||||||
if t.root == nil {
|
if t.root == nil {
|
||||||
return hashNode(types.EmptyRootHash.Bytes()), nil, nil
|
return hashNode(types.EmptyRootHash.Bytes()), nil
|
||||||
}
|
}
|
||||||
// If the number of changes is below 100, we let one thread handle it
|
// If the number of changes is below 100, we let one thread handle it
|
||||||
h := newHasher(t.unhashed >= 100)
|
h := newHasher(t.unhashed >= 100)
|
||||||
defer returnHasherToPool(h)
|
defer func() {
|
||||||
|
returnHasherToPool(h)
|
||||||
|
t.unhashed = 0
|
||||||
|
}()
|
||||||
hashed, cached := h.hash(t.root, true)
|
hashed, cached := h.hash(t.root, true)
|
||||||
t.unhashed = 0
|
return hashed, cached
|
||||||
return hashed, cached, nil
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// Reset drops the referenced root node and cleans all internal state.
|
// Reset drops the referenced root node and cleans all internal state.
|
||||||
|
@ -18,7 +18,6 @@ package trie
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"bytes"
|
"bytes"
|
||||||
crand "crypto/rand"
|
|
||||||
"encoding/binary"
|
"encoding/binary"
|
||||||
"errors"
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
@ -403,6 +402,51 @@ func (randTest) Generate(r *rand.Rand, size int) reflect.Value {
|
|||||||
return reflect.ValueOf(steps)
|
return reflect.ValueOf(steps)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func verifyAccessList(old *Trie, new *Trie, set *NodeSet) error {
|
||||||
|
deletes, inserts, updates := diffTries(old, new)
|
||||||
|
|
||||||
|
// Check insertion set
|
||||||
|
for path := range inserts {
|
||||||
|
n, ok := set.nodes[path]
|
||||||
|
if !ok || n.isDeleted() {
|
||||||
|
return errors.New("expect new node")
|
||||||
|
}
|
||||||
|
_, ok = set.accessList[path]
|
||||||
|
if ok {
|
||||||
|
return errors.New("unexpected origin value")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
// Check deletion set
|
||||||
|
for path, blob := range deletes {
|
||||||
|
n, ok := set.nodes[path]
|
||||||
|
if !ok || !n.isDeleted() {
|
||||||
|
return errors.New("expect deleted node")
|
||||||
|
}
|
||||||
|
v, ok := set.accessList[path]
|
||||||
|
if !ok {
|
||||||
|
return errors.New("expect origin value")
|
||||||
|
}
|
||||||
|
if !bytes.Equal(v, blob) {
|
||||||
|
return errors.New("invalid origin value")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
// Check update set
|
||||||
|
for path, blob := range updates {
|
||||||
|
n, ok := set.nodes[path]
|
||||||
|
if !ok || n.isDeleted() {
|
||||||
|
return errors.New("expect updated node")
|
||||||
|
}
|
||||||
|
v, ok := set.accessList[path]
|
||||||
|
if !ok {
|
||||||
|
return errors.New("expect origin value")
|
||||||
|
}
|
||||||
|
if !bytes.Equal(v, blob) {
|
||||||
|
return errors.New("invalid origin value")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
func runRandTest(rt randTest) bool {
|
func runRandTest(rt randTest) bool {
|
||||||
var (
|
var (
|
||||||
triedb = NewDatabase(rawdb.NewMemoryDatabase())
|
triedb = NewDatabase(rawdb.NewMemoryDatabase())
|
||||||
@ -410,8 +454,6 @@ func runRandTest(rt randTest) bool {
|
|||||||
values = make(map[string]string) // tracks content of the trie
|
values = make(map[string]string) // tracks content of the trie
|
||||||
origTrie = NewEmpty(triedb)
|
origTrie = NewEmpty(triedb)
|
||||||
)
|
)
|
||||||
tr.tracer = newTracer()
|
|
||||||
|
|
||||||
for i, step := range rt {
|
for i, step := range rt {
|
||||||
// fmt.Printf("{op: %d, key: common.Hex2Bytes(\"%x\"), value: common.Hex2Bytes(\"%x\")}, // step %d\n",
|
// fmt.Printf("{op: %d, key: common.Hex2Bytes(\"%x\"), value: common.Hex2Bytes(\"%x\")}, // step %d\n",
|
||||||
// step.op, step.key, step.value, i)
|
// step.op, step.key, step.value, i)
|
||||||
@ -447,24 +489,6 @@ func runRandTest(rt randTest) bool {
|
|||||||
tr.Hash()
|
tr.Hash()
|
||||||
case opCommit:
|
case opCommit:
|
||||||
root, nodes := tr.Commit(true)
|
root, nodes := tr.Commit(true)
|
||||||
// Validity the returned nodeset
|
|
||||||
if nodes != nil {
|
|
||||||
for path, node := range nodes.updates.nodes {
|
|
||||||
blob, _, _ := origTrie.TryGetNode(hexToCompact([]byte(path)))
|
|
||||||
got := node.prev
|
|
||||||
if !bytes.Equal(blob, got) {
|
|
||||||
rt[i].err = fmt.Errorf("prevalue mismatch for 0x%x, got 0x%x want 0x%x", path, got, blob)
|
|
||||||
panic(rt[i].err)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
for path, prev := range nodes.deletes {
|
|
||||||
blob, _, _ := origTrie.TryGetNode(hexToCompact([]byte(path)))
|
|
||||||
if !bytes.Equal(blob, prev) {
|
|
||||||
rt[i].err = fmt.Errorf("prevalue mismatch for 0x%x, got 0x%x want 0x%x", path, prev, blob)
|
|
||||||
return false
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
if nodes != nil {
|
if nodes != nil {
|
||||||
triedb.Update(NewWithNodeSet(nodes))
|
triedb.Update(NewWithNodeSet(nodes))
|
||||||
}
|
}
|
||||||
@ -473,13 +497,13 @@ func runRandTest(rt randTest) bool {
|
|||||||
rt[i].err = err
|
rt[i].err = err
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
|
if nodes != nil {
|
||||||
|
if err := verifyAccessList(origTrie, newtr, nodes); err != nil {
|
||||||
|
rt[i].err = err
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
}
|
||||||
tr = newtr
|
tr = newtr
|
||||||
|
|
||||||
// Enable node tracing. Resolve the root node again explicitly
|
|
||||||
// since it's not captured at the beginning.
|
|
||||||
tr.tracer = newTracer()
|
|
||||||
tr.resolveAndTrack(root.Bytes(), nil)
|
|
||||||
|
|
||||||
origTrie = tr.Copy()
|
origTrie = tr.Copy()
|
||||||
case opItercheckhash:
|
case opItercheckhash:
|
||||||
checktr := NewEmpty(triedb)
|
checktr := NewEmpty(triedb)
|
||||||
@ -492,8 +516,6 @@ func runRandTest(rt randTest) bool {
|
|||||||
}
|
}
|
||||||
case opNodeDiff:
|
case opNodeDiff:
|
||||||
var (
|
var (
|
||||||
inserted = tr.tracer.insertList()
|
|
||||||
deleted = tr.tracer.deleteList()
|
|
||||||
origIter = origTrie.NodeIterator(nil)
|
origIter = origTrie.NodeIterator(nil)
|
||||||
curIter = tr.NodeIterator(nil)
|
curIter = tr.NodeIterator(nil)
|
||||||
origSeen = make(map[string]struct{})
|
origSeen = make(map[string]struct{})
|
||||||
@ -527,19 +549,19 @@ func runRandTest(rt randTest) bool {
|
|||||||
deleteExp[path] = struct{}{}
|
deleteExp[path] = struct{}{}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if len(insertExp) != len(inserted) {
|
if len(insertExp) != len(tr.tracer.inserts) {
|
||||||
rt[i].err = fmt.Errorf("insert set mismatch")
|
rt[i].err = fmt.Errorf("insert set mismatch")
|
||||||
}
|
}
|
||||||
if len(deleteExp) != len(deleted) {
|
if len(deleteExp) != len(tr.tracer.deletes) {
|
||||||
rt[i].err = fmt.Errorf("delete set mismatch")
|
rt[i].err = fmt.Errorf("delete set mismatch")
|
||||||
}
|
}
|
||||||
for _, insert := range inserted {
|
for insert := range tr.tracer.inserts {
|
||||||
if _, present := insertExp[string(insert)]; !present {
|
if _, present := insertExp[insert]; !present {
|
||||||
rt[i].err = fmt.Errorf("missing inserted node")
|
rt[i].err = fmt.Errorf("missing inserted node")
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
for _, del := range deleted {
|
for del := range tr.tracer.deletes {
|
||||||
if _, present := deleteExp[string(del)]; !present {
|
if _, present := deleteExp[del]; !present {
|
||||||
rt[i].err = fmt.Errorf("missing deleted node")
|
rt[i].err = fmt.Errorf("missing deleted node")
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -1123,13 +1145,14 @@ func deleteString(trie *Trie, k string) {
|
|||||||
|
|
||||||
func TestDecodeNode(t *testing.T) {
|
func TestDecodeNode(t *testing.T) {
|
||||||
t.Parallel()
|
t.Parallel()
|
||||||
|
|
||||||
var (
|
var (
|
||||||
hash = make([]byte, 20)
|
hash = make([]byte, 20)
|
||||||
elems = make([]byte, 20)
|
elems = make([]byte, 20)
|
||||||
)
|
)
|
||||||
for i := 0; i < 5000000; i++ {
|
for i := 0; i < 5000000; i++ {
|
||||||
crand.Read(hash)
|
prng.Read(hash)
|
||||||
crand.Read(elems)
|
prng.Read(elems)
|
||||||
decodeNode(hash, elems)
|
decodeNode(hash, elems)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -1,305 +0,0 @@
|
|||||||
// Copyright 2022 The go-ethereum Authors
|
|
||||||
// This file is part of the go-ethereum library.
|
|
||||||
//
|
|
||||||
// The go-ethereum library is free software: you can redistribute it and/or modify
|
|
||||||
// it under the terms of the GNU Lesser General Public License as published by
|
|
||||||
// the Free Software Foundation, either version 3 of the License, or
|
|
||||||
// (at your option) any later version.
|
|
||||||
//
|
|
||||||
// The go-ethereum library is distributed in the hope that it will be useful,
|
|
||||||
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
||||||
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
||||||
// GNU Lesser General Public License for more details.
|
|
||||||
//
|
|
||||||
// You should have received a copy of the GNU Lesser General Public License
|
|
||||||
// along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
|
|
||||||
|
|
||||||
package trie
|
|
||||||
|
|
||||||
import (
|
|
||||||
"bytes"
|
|
||||||
"testing"
|
|
||||||
|
|
||||||
"github.com/ethereum/go-ethereum/common"
|
|
||||||
"github.com/ethereum/go-ethereum/core/rawdb"
|
|
||||||
"github.com/ethereum/go-ethereum/core/types"
|
|
||||||
)
|
|
||||||
|
|
||||||
// Tests if the trie diffs are tracked correctly.
|
|
||||||
func TestTrieTracer(t *testing.T) {
|
|
||||||
db := NewDatabase(rawdb.NewMemoryDatabase())
|
|
||||||
trie := NewEmpty(db)
|
|
||||||
trie.tracer = newTracer()
|
|
||||||
|
|
||||||
// Insert a batch of entries, all the nodes should be marked as inserted
|
|
||||||
vals := []struct{ k, v string }{
|
|
||||||
{"do", "verb"},
|
|
||||||
{"ether", "wookiedoo"},
|
|
||||||
{"horse", "stallion"},
|
|
||||||
{"shaman", "horse"},
|
|
||||||
{"doge", "coin"},
|
|
||||||
{"dog", "puppy"},
|
|
||||||
{"somethingveryoddindeedthis is", "myothernodedata"},
|
|
||||||
}
|
|
||||||
for _, val := range vals {
|
|
||||||
trie.Update([]byte(val.k), []byte(val.v))
|
|
||||||
}
|
|
||||||
trie.Hash()
|
|
||||||
|
|
||||||
seen := make(map[string]struct{})
|
|
||||||
it := trie.NodeIterator(nil)
|
|
||||||
for it.Next(true) {
|
|
||||||
if it.Leaf() {
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
seen[string(it.Path())] = struct{}{}
|
|
||||||
}
|
|
||||||
inserted := trie.tracer.insertList()
|
|
||||||
if len(inserted) != len(seen) {
|
|
||||||
t.Fatalf("Unexpected inserted node tracked want %d got %d", len(seen), len(inserted))
|
|
||||||
}
|
|
||||||
for _, k := range inserted {
|
|
||||||
_, ok := seen[string(k)]
|
|
||||||
if !ok {
|
|
||||||
t.Fatalf("Unexpected inserted node")
|
|
||||||
}
|
|
||||||
}
|
|
||||||
deleted := trie.tracer.deleteList()
|
|
||||||
if len(deleted) != 0 {
|
|
||||||
t.Fatalf("Unexpected deleted node tracked %d", len(deleted))
|
|
||||||
}
|
|
||||||
|
|
||||||
// Commit the changes and re-create with new root
|
|
||||||
root, nodes := trie.Commit(false)
|
|
||||||
if err := db.Update(NewWithNodeSet(nodes)); err != nil {
|
|
||||||
t.Fatal(err)
|
|
||||||
}
|
|
||||||
trie, _ = New(TrieID(root), db)
|
|
||||||
trie.tracer = newTracer()
|
|
||||||
|
|
||||||
// Delete all the elements, check deletion set
|
|
||||||
for _, val := range vals {
|
|
||||||
trie.Delete([]byte(val.k))
|
|
||||||
}
|
|
||||||
trie.Hash()
|
|
||||||
|
|
||||||
inserted = trie.tracer.insertList()
|
|
||||||
if len(inserted) != 0 {
|
|
||||||
t.Fatalf("Unexpected inserted node tracked %d", len(inserted))
|
|
||||||
}
|
|
||||||
deleted = trie.tracer.deleteList()
|
|
||||||
if len(deleted) != len(seen) {
|
|
||||||
t.Fatalf("Unexpected deleted node tracked want %d got %d", len(seen), len(deleted))
|
|
||||||
}
|
|
||||||
for _, k := range deleted {
|
|
||||||
_, ok := seen[string(k)]
|
|
||||||
if !ok {
|
|
||||||
t.Fatalf("Unexpected inserted node")
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestTrieTracerNoop(t *testing.T) {
|
|
||||||
trie := NewEmpty(NewDatabase(rawdb.NewMemoryDatabase()))
|
|
||||||
trie.tracer = newTracer()
|
|
||||||
|
|
||||||
// Insert a batch of entries, all the nodes should be marked as inserted
|
|
||||||
vals := []struct{ k, v string }{
|
|
||||||
{"do", "verb"},
|
|
||||||
{"ether", "wookiedoo"},
|
|
||||||
{"horse", "stallion"},
|
|
||||||
{"shaman", "horse"},
|
|
||||||
{"doge", "coin"},
|
|
||||||
{"dog", "puppy"},
|
|
||||||
{"somethingveryoddindeedthis is", "myothernodedata"},
|
|
||||||
}
|
|
||||||
for _, val := range vals {
|
|
||||||
trie.Update([]byte(val.k), []byte(val.v))
|
|
||||||
}
|
|
||||||
for _, val := range vals {
|
|
||||||
trie.Delete([]byte(val.k))
|
|
||||||
}
|
|
||||||
if len(trie.tracer.insertList()) != 0 {
|
|
||||||
t.Fatalf("Unexpected inserted node tracked %d", len(trie.tracer.insertList()))
|
|
||||||
}
|
|
||||||
if len(trie.tracer.deleteList()) != 0 {
|
|
||||||
t.Fatalf("Unexpected deleted node tracked %d", len(trie.tracer.deleteList()))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestTrieTracePrevValue(t *testing.T) {
|
|
||||||
db := NewDatabase(rawdb.NewMemoryDatabase())
|
|
||||||
trie := NewEmpty(db)
|
|
||||||
trie.tracer = newTracer()
|
|
||||||
|
|
||||||
paths, blobs := trie.tracer.prevList()
|
|
||||||
if len(paths) != 0 || len(blobs) != 0 {
|
|
||||||
t.Fatalf("Nothing should be tracked")
|
|
||||||
}
|
|
||||||
// Insert a batch of entries, all the nodes should be marked as inserted
|
|
||||||
vals := []struct{ k, v string }{
|
|
||||||
{"do", "verb"},
|
|
||||||
{"ether", "wookiedoo"},
|
|
||||||
{"horse", "stallion"},
|
|
||||||
{"shaman", "horse"},
|
|
||||||
{"doge", "coin"},
|
|
||||||
{"dog", "puppy"},
|
|
||||||
{"somethingveryoddindeedthis is", "myothernodedata"},
|
|
||||||
}
|
|
||||||
for _, val := range vals {
|
|
||||||
trie.Update([]byte(val.k), []byte(val.v))
|
|
||||||
}
|
|
||||||
paths, blobs = trie.tracer.prevList()
|
|
||||||
if len(paths) != 0 || len(blobs) != 0 {
|
|
||||||
t.Fatalf("Nothing should be tracked")
|
|
||||||
}
|
|
||||||
|
|
||||||
// Commit the changes and re-create with new root
|
|
||||||
root, nodes := trie.Commit(false)
|
|
||||||
if err := db.Update(NewWithNodeSet(nodes)); err != nil {
|
|
||||||
t.Fatal(err)
|
|
||||||
}
|
|
||||||
trie, _ = New(TrieID(root), db)
|
|
||||||
trie.tracer = newTracer()
|
|
||||||
trie.resolveAndTrack(root.Bytes(), nil)
|
|
||||||
|
|
||||||
// Load all nodes in trie
|
|
||||||
for _, val := range vals {
|
|
||||||
trie.TryGet([]byte(val.k))
|
|
||||||
}
|
|
||||||
|
|
||||||
// Ensure all nodes are tracked by tracer with correct prev-values
|
|
||||||
iter := trie.NodeIterator(nil)
|
|
||||||
seen := make(map[string][]byte)
|
|
||||||
for iter.Next(true) {
|
|
||||||
// Embedded nodes are ignored since they are not present in
|
|
||||||
// database.
|
|
||||||
if iter.Hash() == (common.Hash{}) {
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
seen[string(iter.Path())] = common.CopyBytes(iter.NodeBlob())
|
|
||||||
}
|
|
||||||
|
|
||||||
paths, blobs = trie.tracer.prevList()
|
|
||||||
if len(paths) != len(seen) || len(blobs) != len(seen) {
|
|
||||||
t.Fatalf("Unexpected tracked values")
|
|
||||||
}
|
|
||||||
for i, path := range paths {
|
|
||||||
blob := blobs[i]
|
|
||||||
prev, ok := seen[string(path)]
|
|
||||||
if !ok {
|
|
||||||
t.Fatalf("Missing node %v", path)
|
|
||||||
}
|
|
||||||
if !bytes.Equal(blob, prev) {
|
|
||||||
t.Fatalf("Unexpected value path: %v, want: %v, got: %v", path, prev, blob)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// Re-open the trie and iterate the trie, ensure nothing will be tracked.
|
|
||||||
// Iterator will not link any loaded nodes to trie.
|
|
||||||
trie, _ = New(TrieID(root), db)
|
|
||||||
trie.tracer = newTracer()
|
|
||||||
|
|
||||||
iter = trie.NodeIterator(nil)
|
|
||||||
for iter.Next(true) {
|
|
||||||
}
|
|
||||||
paths, blobs = trie.tracer.prevList()
|
|
||||||
if len(paths) != 0 || len(blobs) != 0 {
|
|
||||||
t.Fatalf("Nothing should be tracked")
|
|
||||||
}
|
|
||||||
|
|
||||||
// Re-open the trie and generate proof for entries, ensure nothing will
|
|
||||||
// be tracked. Prover will not link any loaded nodes to trie.
|
|
||||||
trie, _ = New(TrieID(root), db)
|
|
||||||
trie.tracer = newTracer()
|
|
||||||
for _, val := range vals {
|
|
||||||
trie.Prove([]byte(val.k), 0, rawdb.NewMemoryDatabase())
|
|
||||||
}
|
|
||||||
paths, blobs = trie.tracer.prevList()
|
|
||||||
if len(paths) != 0 || len(blobs) != 0 {
|
|
||||||
t.Fatalf("Nothing should be tracked")
|
|
||||||
}
|
|
||||||
|
|
||||||
// Delete entries from trie, ensure all previous values are correct.
|
|
||||||
trie, _ = New(TrieID(root), db)
|
|
||||||
trie.tracer = newTracer()
|
|
||||||
trie.resolveAndTrack(root.Bytes(), nil)
|
|
||||||
|
|
||||||
for _, val := range vals {
|
|
||||||
trie.TryDelete([]byte(val.k))
|
|
||||||
}
|
|
||||||
paths, blobs = trie.tracer.prevList()
|
|
||||||
if len(paths) != len(seen) || len(blobs) != len(seen) {
|
|
||||||
t.Fatalf("Unexpected tracked values")
|
|
||||||
}
|
|
||||||
for i, path := range paths {
|
|
||||||
blob := blobs[i]
|
|
||||||
prev, ok := seen[string(path)]
|
|
||||||
if !ok {
|
|
||||||
t.Fatalf("Missing node %v", path)
|
|
||||||
}
|
|
||||||
if !bytes.Equal(blob, prev) {
|
|
||||||
t.Fatalf("Unexpected value path: %v, want: %v, got: %v", path, prev, blob)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestDeleteAll(t *testing.T) {
|
|
||||||
db := NewDatabase(rawdb.NewMemoryDatabase())
|
|
||||||
trie := NewEmpty(db)
|
|
||||||
trie.tracer = newTracer()
|
|
||||||
|
|
||||||
// Insert a batch of entries, all the nodes should be marked as inserted
|
|
||||||
vals := []struct{ k, v string }{
|
|
||||||
{"do", "verb"},
|
|
||||||
{"ether", "wookiedoo"},
|
|
||||||
{"horse", "stallion"},
|
|
||||||
{"shaman", "horse"},
|
|
||||||
{"doge", "coin"},
|
|
||||||
{"dog", "puppy"},
|
|
||||||
{"somethingveryoddindeedthis is", "myothernodedata"},
|
|
||||||
}
|
|
||||||
for _, val := range vals {
|
|
||||||
trie.Update([]byte(val.k), []byte(val.v))
|
|
||||||
}
|
|
||||||
root, set := trie.Commit(false)
|
|
||||||
if err := db.Update(NewWithNodeSet(set)); err != nil {
|
|
||||||
t.Fatal(err)
|
|
||||||
}
|
|
||||||
// Delete entries from trie, ensure all values are detected
|
|
||||||
trie, _ = New(TrieID(root), db)
|
|
||||||
trie.tracer = newTracer()
|
|
||||||
trie.resolveAndTrack(root.Bytes(), nil)
|
|
||||||
|
|
||||||
// Iterate all existent nodes
|
|
||||||
var (
|
|
||||||
it = trie.NodeIterator(nil)
|
|
||||||
nodes = make(map[string][]byte)
|
|
||||||
)
|
|
||||||
for it.Next(true) {
|
|
||||||
if it.Hash() != (common.Hash{}) {
|
|
||||||
nodes[string(it.Path())] = common.CopyBytes(it.NodeBlob())
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// Perform deletion to purge the entire trie
|
|
||||||
for _, val := range vals {
|
|
||||||
trie.Delete([]byte(val.k))
|
|
||||||
}
|
|
||||||
root, set = trie.Commit(false)
|
|
||||||
if root != types.EmptyRootHash {
|
|
||||||
t.Fatalf("Invalid trie root %v", root)
|
|
||||||
}
|
|
||||||
for path, blob := range set.deletes {
|
|
||||||
prev, ok := nodes[path]
|
|
||||||
if !ok {
|
|
||||||
t.Fatalf("Extra node deleted %v", []byte(path))
|
|
||||||
}
|
|
||||||
if !bytes.Equal(prev, blob) {
|
|
||||||
t.Fatalf("Unexpected previous value %v", []byte(path))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
if len(set.deletes) != len(nodes) {
|
|
||||||
t.Fatalf("Unexpected deletion set")
|
|
||||||
}
|
|
||||||
}
|
|
199
trie/utils.go
199
trie/utils.go
@ -1,199 +0,0 @@
|
|||||||
// Copyright 2022 The go-ethereum Authors
|
|
||||||
// This file is part of the go-ethereum library.
|
|
||||||
//
|
|
||||||
// The go-ethereum library is free software: you can redistribute it and/or modify
|
|
||||||
// it under the terms of the GNU Lesser General Public License as published by
|
|
||||||
// the Free Software Foundation, either version 3 of the License, or
|
|
||||||
// (at your option) any later version.
|
|
||||||
//
|
|
||||||
// The go-ethereum library is distributed in the hope that it will be useful,
|
|
||||||
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
||||||
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
||||||
// GNU Lesser General Public License for more details.
|
|
||||||
//
|
|
||||||
// You should have received a copy of the GNU Lesser General Public License
|
|
||||||
// along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
|
|
||||||
|
|
||||||
package trie
|
|
||||||
|
|
||||||
// tracer tracks the changes of trie nodes. During the trie operations,
|
|
||||||
// some nodes can be deleted from the trie, while these deleted nodes
|
|
||||||
// won't be captured by trie.Hasher or trie.Committer. Thus, these deleted
|
|
||||||
// nodes won't be removed from the disk at all. Tracer is an auxiliary tool
|
|
||||||
// used to track all insert and delete operations of trie and capture all
|
|
||||||
// deleted nodes eventually.
|
|
||||||
//
|
|
||||||
// The changed nodes can be mainly divided into two categories: the leaf
|
|
||||||
// node and intermediate node. The former is inserted/deleted by callers
|
|
||||||
// while the latter is inserted/deleted in order to follow the rule of trie.
|
|
||||||
// This tool can track all of them no matter the node is embedded in its
|
|
||||||
// parent or not, but valueNode is never tracked.
|
|
||||||
//
|
|
||||||
// Besides, it's also used for recording the original value of the nodes
|
|
||||||
// when they are resolved from the disk. The pre-value of the nodes will
|
|
||||||
// be used to construct reverse-diffs in the future.
|
|
||||||
//
|
|
||||||
// Note tracer is not thread-safe, callers should be responsible for handling
|
|
||||||
// the concurrency issues by themselves.
|
|
||||||
type tracer struct {
|
|
||||||
insert map[string]struct{}
|
|
||||||
delete map[string]struct{}
|
|
||||||
origin map[string][]byte
|
|
||||||
}
|
|
||||||
|
|
||||||
// newTracer initializes the tracer for capturing trie changes.
|
|
||||||
func newTracer() *tracer {
|
|
||||||
return &tracer{
|
|
||||||
insert: make(map[string]struct{}),
|
|
||||||
delete: make(map[string]struct{}),
|
|
||||||
origin: make(map[string][]byte),
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// onRead tracks the newly loaded trie node and caches the rlp-encoded blob internally.
|
|
||||||
// Don't change the value outside of function since it's not deep-copied.
|
|
||||||
func (t *tracer) onRead(path []byte, val []byte) {
|
|
||||||
// Tracer isn't used right now, remove this check later.
|
|
||||||
if t == nil {
|
|
||||||
return
|
|
||||||
}
|
|
||||||
t.origin[string(path)] = val
|
|
||||||
}
|
|
||||||
|
|
||||||
// onInsert tracks the newly inserted trie node. If it's already in the deletion set
|
|
||||||
// (resurrected node), then just wipe it from the deletion set as the "untouched".
|
|
||||||
func (t *tracer) onInsert(path []byte) {
|
|
||||||
// Tracer isn't used right now, remove this check later.
|
|
||||||
if t == nil {
|
|
||||||
return
|
|
||||||
}
|
|
||||||
if _, present := t.delete[string(path)]; present {
|
|
||||||
delete(t.delete, string(path))
|
|
||||||
return
|
|
||||||
}
|
|
||||||
t.insert[string(path)] = struct{}{}
|
|
||||||
}
|
|
||||||
|
|
||||||
// onDelete tracks the newly deleted trie node. If it's already
|
|
||||||
// in the addition set, then just wipe it from the addition set
|
|
||||||
// as it's untouched.
|
|
||||||
func (t *tracer) onDelete(path []byte) {
|
|
||||||
// Tracer isn't used right now, remove this check later.
|
|
||||||
if t == nil {
|
|
||||||
return
|
|
||||||
}
|
|
||||||
if _, present := t.insert[string(path)]; present {
|
|
||||||
delete(t.insert, string(path))
|
|
||||||
return
|
|
||||||
}
|
|
||||||
t.delete[string(path)] = struct{}{}
|
|
||||||
}
|
|
||||||
|
|
||||||
// insertList returns the tracked inserted trie nodes in list format.
|
|
||||||
func (t *tracer) insertList() [][]byte {
|
|
||||||
// Tracer isn't used right now, remove this check later.
|
|
||||||
if t == nil {
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
var ret [][]byte
|
|
||||||
for path := range t.insert {
|
|
||||||
ret = append(ret, []byte(path))
|
|
||||||
}
|
|
||||||
return ret
|
|
||||||
}
|
|
||||||
|
|
||||||
// deleteList returns the tracked deleted trie nodes in list format.
|
|
||||||
func (t *tracer) deleteList() [][]byte {
|
|
||||||
// Tracer isn't used right now, remove this check later.
|
|
||||||
if t == nil {
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
var ret [][]byte
|
|
||||||
for path := range t.delete {
|
|
||||||
ret = append(ret, []byte(path))
|
|
||||||
}
|
|
||||||
return ret
|
|
||||||
}
|
|
||||||
|
|
||||||
// prevList returns the tracked node blobs in list format.
|
|
||||||
func (t *tracer) prevList() ([][]byte, [][]byte) {
|
|
||||||
// Tracer isn't used right now, remove this check later.
|
|
||||||
if t == nil {
|
|
||||||
return nil, nil
|
|
||||||
}
|
|
||||||
var (
|
|
||||||
paths [][]byte
|
|
||||||
blobs [][]byte
|
|
||||||
)
|
|
||||||
for path, blob := range t.origin {
|
|
||||||
paths = append(paths, []byte(path))
|
|
||||||
blobs = append(blobs, blob)
|
|
||||||
}
|
|
||||||
return paths, blobs
|
|
||||||
}
|
|
||||||
|
|
||||||
// getPrev returns the cached original value of the specified node.
|
|
||||||
func (t *tracer) getPrev(path []byte) []byte {
|
|
||||||
// Tracer isn't used right now, remove this check later.
|
|
||||||
if t == nil {
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
return t.origin[string(path)]
|
|
||||||
}
|
|
||||||
|
|
||||||
// reset clears the content tracked by tracer.
|
|
||||||
func (t *tracer) reset() {
|
|
||||||
// Tracer isn't used right now, remove this check later.
|
|
||||||
if t == nil {
|
|
||||||
return
|
|
||||||
}
|
|
||||||
t.insert = make(map[string]struct{})
|
|
||||||
t.delete = make(map[string]struct{})
|
|
||||||
t.origin = make(map[string][]byte)
|
|
||||||
}
|
|
||||||
|
|
||||||
// copy returns a deep copied tracer instance.
|
|
||||||
func (t *tracer) copy() *tracer {
|
|
||||||
// Tracer isn't used right now, remove this check later.
|
|
||||||
if t == nil {
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
var (
|
|
||||||
insert = make(map[string]struct{})
|
|
||||||
delete = make(map[string]struct{})
|
|
||||||
origin = make(map[string][]byte)
|
|
||||||
)
|
|
||||||
for key := range t.insert {
|
|
||||||
insert[key] = struct{}{}
|
|
||||||
}
|
|
||||||
for key := range t.delete {
|
|
||||||
delete[key] = struct{}{}
|
|
||||||
}
|
|
||||||
for key, val := range t.origin {
|
|
||||||
origin[key] = val
|
|
||||||
}
|
|
||||||
return &tracer{
|
|
||||||
insert: insert,
|
|
||||||
delete: delete,
|
|
||||||
origin: origin,
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// markDeletions puts all tracked deletions into the provided nodeset.
|
|
||||||
func (t *tracer) markDeletions(set *NodeSet) {
|
|
||||||
// Tracer isn't used right now, remove this check later.
|
|
||||||
if t == nil {
|
|
||||||
return
|
|
||||||
}
|
|
||||||
for _, path := range t.deleteList() {
|
|
||||||
// There are a few possibilities for this scenario(the node is deleted
|
|
||||||
// but not present in database previously), for example the node was
|
|
||||||
// embedded in the parent and now deleted from the trie. In this case
|
|
||||||
// it's noop from database's perspective.
|
|
||||||
val := t.getPrev(path)
|
|
||||||
if len(val) == 0 {
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
set.markDeleted(path, val)
|
|
||||||
}
|
|
||||||
}
|
|
Loading…
Reference in New Issue
Block a user