forked from cerc-io/plugeth
2219 lines
75 KiB
Go
2219 lines
75 KiB
Go
// Copyright 2015 The go-ethereum Authors
|
|
// This file is part of go-ethereum.
|
|
//
|
|
// go-ethereum is free software: you can redistribute it and/or modify
|
|
// it under the terms of the GNU General Public License as published by
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
// (at your option) any later version.
|
|
//
|
|
// go-ethereum is distributed in the hope that it will be useful,
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
// GNU General Public License for more details.
|
|
//
|
|
// You should have received a copy of the GNU General Public License
|
|
// along with go-ethereum. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
// Package utils contains internal helper functions for go-ethereum commands.
|
|
package utils
|
|
|
|
import (
|
|
"context"
|
|
"crypto/ecdsa"
|
|
"encoding/hex"
|
|
"errors"
|
|
"fmt"
|
|
"math"
|
|
"math/big"
|
|
"net"
|
|
"net/http"
|
|
"os"
|
|
"path/filepath"
|
|
godebug "runtime/debug"
|
|
"strconv"
|
|
"strings"
|
|
"time"
|
|
|
|
"github.com/ethereum/go-ethereum/accounts"
|
|
"github.com/ethereum/go-ethereum/accounts/keystore"
|
|
"github.com/ethereum/go-ethereum/common"
|
|
"github.com/ethereum/go-ethereum/common/fdlimit"
|
|
"github.com/ethereum/go-ethereum/core"
|
|
"github.com/ethereum/go-ethereum/core/rawdb"
|
|
"github.com/ethereum/go-ethereum/core/txpool/legacypool"
|
|
"github.com/ethereum/go-ethereum/core/vm"
|
|
"github.com/ethereum/go-ethereum/crypto"
|
|
"github.com/ethereum/go-ethereum/crypto/kzg4844"
|
|
"github.com/ethereum/go-ethereum/eth"
|
|
"github.com/ethereum/go-ethereum/eth/catalyst"
|
|
"github.com/ethereum/go-ethereum/eth/downloader"
|
|
"github.com/ethereum/go-ethereum/eth/ethconfig"
|
|
"github.com/ethereum/go-ethereum/eth/filters"
|
|
"github.com/ethereum/go-ethereum/eth/gasprice"
|
|
"github.com/ethereum/go-ethereum/eth/tracers"
|
|
"github.com/ethereum/go-ethereum/ethdb"
|
|
"github.com/ethereum/go-ethereum/ethdb/remotedb"
|
|
"github.com/ethereum/go-ethereum/ethstats"
|
|
"github.com/ethereum/go-ethereum/graphql"
|
|
"github.com/ethereum/go-ethereum/internal/ethapi"
|
|
"github.com/ethereum/go-ethereum/internal/flags"
|
|
"github.com/ethereum/go-ethereum/log"
|
|
"github.com/ethereum/go-ethereum/metrics"
|
|
"github.com/ethereum/go-ethereum/metrics/exp"
|
|
"github.com/ethereum/go-ethereum/metrics/influxdb"
|
|
"github.com/ethereum/go-ethereum/miner"
|
|
"github.com/ethereum/go-ethereum/node"
|
|
"github.com/ethereum/go-ethereum/p2p"
|
|
"github.com/ethereum/go-ethereum/p2p/enode"
|
|
"github.com/ethereum/go-ethereum/p2p/nat"
|
|
"github.com/ethereum/go-ethereum/p2p/netutil"
|
|
"github.com/ethereum/go-ethereum/params"
|
|
"github.com/ethereum/go-ethereum/rpc"
|
|
"github.com/ethereum/go-ethereum/triedb"
|
|
"github.com/ethereum/go-ethereum/triedb/hashdb"
|
|
"github.com/ethereum/go-ethereum/triedb/pathdb"
|
|
pcsclite "github.com/gballet/go-libpcsclite"
|
|
gopsutil "github.com/shirou/gopsutil/mem"
|
|
"github.com/urfave/cli/v2"
|
|
)
|
|
|
|
// These are all the command line flags we support.
|
|
// If you add to this list, please remember to include the
|
|
// flag in the appropriate command definition.
|
|
//
|
|
// The flags are defined here so their names and help texts
|
|
// are the same for all commands.
|
|
|
|
var (
|
|
// General settings
|
|
//begin PluGeth code injection
|
|
PluginsDirFlag = &flags.DirectoryFlag{
|
|
Name: "pluginsdir",
|
|
Usage: "Directory for plugins",
|
|
Value: flags.DirectoryString(filepath.Join("<datadir>", "plugins")),
|
|
Category: flags.EthCategory,
|
|
}
|
|
//end PluGeth code injection
|
|
DataDirFlag = &flags.DirectoryFlag{
|
|
Name: "datadir",
|
|
Usage: "Data directory for the databases and keystore",
|
|
Value: flags.DirectoryString(node.DefaultDataDir()),
|
|
Category: flags.EthCategory,
|
|
}
|
|
RemoteDBFlag = &cli.StringFlag{
|
|
Name: "remotedb",
|
|
Usage: "URL for remote database",
|
|
Category: flags.LoggingCategory,
|
|
}
|
|
DBEngineFlag = &cli.StringFlag{
|
|
Name: "db.engine",
|
|
Usage: "Backing database implementation to use ('pebble' or 'leveldb')",
|
|
Value: node.DefaultConfig.DBEngine,
|
|
Category: flags.EthCategory,
|
|
}
|
|
AncientFlag = &flags.DirectoryFlag{
|
|
Name: "datadir.ancient",
|
|
Usage: "Root directory for ancient data (default = inside chaindata)",
|
|
Category: flags.EthCategory,
|
|
}
|
|
MinFreeDiskSpaceFlag = &flags.DirectoryFlag{
|
|
Name: "datadir.minfreedisk",
|
|
Usage: "Minimum free disk space in MB, once reached triggers auto shut down (default = --cache.gc converted to MB, 0 = disabled)",
|
|
Category: flags.EthCategory,
|
|
}
|
|
KeyStoreDirFlag = &flags.DirectoryFlag{
|
|
Name: "keystore",
|
|
Usage: "Directory for the keystore (default = inside the datadir)",
|
|
Category: flags.AccountCategory,
|
|
}
|
|
USBFlag = &cli.BoolFlag{
|
|
Name: "usb",
|
|
Usage: "Enable monitoring and management of USB hardware wallets",
|
|
Category: flags.AccountCategory,
|
|
}
|
|
SmartCardDaemonPathFlag = &cli.StringFlag{
|
|
Name: "pcscdpath",
|
|
Usage: "Path to the smartcard daemon (pcscd) socket file",
|
|
Value: pcsclite.PCSCDSockName,
|
|
Category: flags.AccountCategory,
|
|
}
|
|
NetworkIdFlag = &cli.Uint64Flag{
|
|
Name: "networkid",
|
|
Usage: "Explicitly set network id (integer)(For testnets: use --goerli, --sepolia, --holesky instead)",
|
|
Value: ethconfig.Defaults.NetworkId,
|
|
Category: flags.EthCategory,
|
|
}
|
|
MainnetFlag = &cli.BoolFlag{
|
|
Name: "mainnet",
|
|
Usage: "Ethereum mainnet",
|
|
Category: flags.EthCategory,
|
|
}
|
|
GoerliFlag = &cli.BoolFlag{
|
|
Name: "goerli",
|
|
Usage: "Görli network: pre-configured proof-of-authority test network",
|
|
Category: flags.EthCategory,
|
|
}
|
|
SepoliaFlag = &cli.BoolFlag{
|
|
Name: "sepolia",
|
|
Usage: "Sepolia network: pre-configured proof-of-work test network",
|
|
Category: flags.EthCategory,
|
|
}
|
|
HoleskyFlag = &cli.BoolFlag{
|
|
Name: "holesky",
|
|
Usage: "Holesky network: pre-configured proof-of-stake test network",
|
|
Category: flags.EthCategory,
|
|
}
|
|
// Dev mode
|
|
DeveloperFlag = &cli.BoolFlag{
|
|
Name: "dev",
|
|
Usage: "Ephemeral proof-of-authority network with a pre-funded developer account, mining enabled",
|
|
Category: flags.DevCategory,
|
|
}
|
|
DeveloperPeriodFlag = &cli.Uint64Flag{
|
|
Name: "dev.period",
|
|
Usage: "Block period to use in developer mode (0 = mine only if transaction pending)",
|
|
Category: flags.DevCategory,
|
|
}
|
|
DeveloperGasLimitFlag = &cli.Uint64Flag{
|
|
Name: "dev.gaslimit",
|
|
Usage: "Initial block gas limit",
|
|
Value: 11500000,
|
|
Category: flags.DevCategory,
|
|
}
|
|
|
|
IdentityFlag = &cli.StringFlag{
|
|
Name: "identity",
|
|
Usage: "Custom node name",
|
|
Category: flags.NetworkingCategory,
|
|
}
|
|
DocRootFlag = &flags.DirectoryFlag{
|
|
Name: "docroot",
|
|
Usage: "Document Root for HTTPClient file scheme",
|
|
Value: flags.DirectoryString(flags.HomeDir()),
|
|
Category: flags.APICategory,
|
|
}
|
|
ExitWhenSyncedFlag = &cli.BoolFlag{
|
|
Name: "exitwhensynced",
|
|
Usage: "Exits after block synchronisation completes",
|
|
Category: flags.EthCategory,
|
|
}
|
|
|
|
// Dump command options.
|
|
IterativeOutputFlag = &cli.BoolFlag{
|
|
Name: "iterative",
|
|
Usage: "Print streaming JSON iteratively, delimited by newlines",
|
|
Value: true,
|
|
}
|
|
ExcludeStorageFlag = &cli.BoolFlag{
|
|
Name: "nostorage",
|
|
Usage: "Exclude storage entries (save db lookups)",
|
|
}
|
|
IncludeIncompletesFlag = &cli.BoolFlag{
|
|
Name: "incompletes",
|
|
Usage: "Include accounts for which we don't have the address (missing preimage)",
|
|
}
|
|
ExcludeCodeFlag = &cli.BoolFlag{
|
|
Name: "nocode",
|
|
Usage: "Exclude contract code (save db lookups)",
|
|
}
|
|
StartKeyFlag = &cli.StringFlag{
|
|
Name: "start",
|
|
Usage: "Start position. Either a hash or address",
|
|
Value: "0x0000000000000000000000000000000000000000000000000000000000000000",
|
|
}
|
|
DumpLimitFlag = &cli.Uint64Flag{
|
|
Name: "limit",
|
|
Usage: "Max number of elements (0 = no limit)",
|
|
Value: 0,
|
|
}
|
|
|
|
defaultSyncMode = ethconfig.Defaults.SyncMode
|
|
SnapshotFlag = &cli.BoolFlag{
|
|
Name: "snapshot",
|
|
Usage: `Enables snapshot-database mode (default = enable)`,
|
|
Value: true,
|
|
Category: flags.EthCategory,
|
|
}
|
|
LightKDFFlag = &cli.BoolFlag{
|
|
Name: "lightkdf",
|
|
Usage: "Reduce key-derivation RAM & CPU usage at some expense of KDF strength",
|
|
Category: flags.AccountCategory,
|
|
}
|
|
EthRequiredBlocksFlag = &cli.StringFlag{
|
|
Name: "eth.requiredblocks",
|
|
Usage: "Comma separated block number-to-hash mappings to require for peering (<number>=<hash>)",
|
|
Category: flags.EthCategory,
|
|
}
|
|
BloomFilterSizeFlag = &cli.Uint64Flag{
|
|
Name: "bloomfilter.size",
|
|
Usage: "Megabytes of memory allocated to bloom-filter for pruning",
|
|
Value: 2048,
|
|
Category: flags.EthCategory,
|
|
}
|
|
OverrideCancun = &cli.Uint64Flag{
|
|
Name: "override.cancun",
|
|
Usage: "Manually specify the Cancun fork timestamp, overriding the bundled setting",
|
|
Category: flags.EthCategory,
|
|
}
|
|
OverrideVerkle = &cli.Uint64Flag{
|
|
Name: "override.verkle",
|
|
Usage: "Manually specify the Verkle fork timestamp, overriding the bundled setting",
|
|
Category: flags.EthCategory,
|
|
}
|
|
SyncModeFlag = &flags.TextMarshalerFlag{
|
|
Name: "syncmode",
|
|
Usage: `Blockchain sync mode ("snap" or "full")`,
|
|
Value: &defaultSyncMode,
|
|
Category: flags.StateCategory,
|
|
}
|
|
GCModeFlag = &cli.StringFlag{
|
|
Name: "gcmode",
|
|
Usage: `Blockchain garbage collection mode, only relevant in state.scheme=hash ("full", "archive")`,
|
|
Value: "full",
|
|
Category: flags.StateCategory,
|
|
}
|
|
StateSchemeFlag = &cli.StringFlag{
|
|
Name: "state.scheme",
|
|
Usage: "Scheme to use for storing ethereum state ('hash' or 'path')",
|
|
Category: flags.StateCategory,
|
|
}
|
|
StateHistoryFlag = &cli.Uint64Flag{
|
|
Name: "history.state",
|
|
Usage: "Number of recent blocks to retain state history for (default = 90,000 blocks, 0 = entire chain)",
|
|
Value: ethconfig.Defaults.StateHistory,
|
|
Category: flags.StateCategory,
|
|
}
|
|
TransactionHistoryFlag = &cli.Uint64Flag{
|
|
Name: "history.transactions",
|
|
Usage: "Number of recent blocks to maintain transactions index for (default = about one year, 0 = entire chain)",
|
|
Value: ethconfig.Defaults.TransactionHistory,
|
|
Category: flags.StateCategory,
|
|
}
|
|
// Transaction pool settings
|
|
TxPoolLocalsFlag = &cli.StringFlag{
|
|
Name: "txpool.locals",
|
|
Usage: "Comma separated accounts to treat as locals (no flush, priority inclusion)",
|
|
Category: flags.TxPoolCategory,
|
|
}
|
|
TxPoolNoLocalsFlag = &cli.BoolFlag{
|
|
Name: "txpool.nolocals",
|
|
Usage: "Disables price exemptions for locally submitted transactions",
|
|
Category: flags.TxPoolCategory,
|
|
}
|
|
TxPoolJournalFlag = &cli.StringFlag{
|
|
Name: "txpool.journal",
|
|
Usage: "Disk journal for local transaction to survive node restarts",
|
|
Value: ethconfig.Defaults.TxPool.Journal,
|
|
Category: flags.TxPoolCategory,
|
|
}
|
|
TxPoolRejournalFlag = &cli.DurationFlag{
|
|
Name: "txpool.rejournal",
|
|
Usage: "Time interval to regenerate the local transaction journal",
|
|
Value: ethconfig.Defaults.TxPool.Rejournal,
|
|
Category: flags.TxPoolCategory,
|
|
}
|
|
TxPoolPriceLimitFlag = &cli.Uint64Flag{
|
|
Name: "txpool.pricelimit",
|
|
Usage: "Minimum gas price tip to enforce for acceptance into the pool",
|
|
Value: ethconfig.Defaults.TxPool.PriceLimit,
|
|
Category: flags.TxPoolCategory,
|
|
}
|
|
TxPoolPriceBumpFlag = &cli.Uint64Flag{
|
|
Name: "txpool.pricebump",
|
|
Usage: "Price bump percentage to replace an already existing transaction",
|
|
Value: ethconfig.Defaults.TxPool.PriceBump,
|
|
Category: flags.TxPoolCategory,
|
|
}
|
|
TxPoolAccountSlotsFlag = &cli.Uint64Flag{
|
|
Name: "txpool.accountslots",
|
|
Usage: "Minimum number of executable transaction slots guaranteed per account",
|
|
Value: ethconfig.Defaults.TxPool.AccountSlots,
|
|
Category: flags.TxPoolCategory,
|
|
}
|
|
TxPoolGlobalSlotsFlag = &cli.Uint64Flag{
|
|
Name: "txpool.globalslots",
|
|
Usage: "Maximum number of executable transaction slots for all accounts",
|
|
Value: ethconfig.Defaults.TxPool.GlobalSlots,
|
|
Category: flags.TxPoolCategory,
|
|
}
|
|
TxPoolAccountQueueFlag = &cli.Uint64Flag{
|
|
Name: "txpool.accountqueue",
|
|
Usage: "Maximum number of non-executable transaction slots permitted per account",
|
|
Value: ethconfig.Defaults.TxPool.AccountQueue,
|
|
Category: flags.TxPoolCategory,
|
|
}
|
|
TxPoolGlobalQueueFlag = &cli.Uint64Flag{
|
|
Name: "txpool.globalqueue",
|
|
Usage: "Maximum number of non-executable transaction slots for all accounts",
|
|
Value: ethconfig.Defaults.TxPool.GlobalQueue,
|
|
Category: flags.TxPoolCategory,
|
|
}
|
|
TxPoolLifetimeFlag = &cli.DurationFlag{
|
|
Name: "txpool.lifetime",
|
|
Usage: "Maximum amount of time non-executable transaction are queued",
|
|
Value: ethconfig.Defaults.TxPool.Lifetime,
|
|
Category: flags.TxPoolCategory,
|
|
}
|
|
// Blob transaction pool settings
|
|
BlobPoolDataDirFlag = &cli.StringFlag{
|
|
Name: "blobpool.datadir",
|
|
Usage: "Data directory to store blob transactions in",
|
|
Value: ethconfig.Defaults.BlobPool.Datadir,
|
|
Category: flags.BlobPoolCategory,
|
|
}
|
|
BlobPoolDataCapFlag = &cli.Uint64Flag{
|
|
Name: "blobpool.datacap",
|
|
Usage: "Disk space to allocate for pending blob transactions (soft limit)",
|
|
Value: ethconfig.Defaults.BlobPool.Datacap,
|
|
Category: flags.BlobPoolCategory,
|
|
}
|
|
BlobPoolPriceBumpFlag = &cli.Uint64Flag{
|
|
Name: "blobpool.pricebump",
|
|
Usage: "Price bump percentage to replace an already existing blob transaction",
|
|
Value: ethconfig.Defaults.BlobPool.PriceBump,
|
|
Category: flags.BlobPoolCategory,
|
|
}
|
|
// Performance tuning settings
|
|
CacheFlag = &cli.IntFlag{
|
|
Name: "cache",
|
|
Usage: "Megabytes of memory allocated to internal caching (default = 4096 mainnet full node, 128 light mode)",
|
|
Value: 1024,
|
|
Category: flags.PerfCategory,
|
|
}
|
|
CacheDatabaseFlag = &cli.IntFlag{
|
|
Name: "cache.database",
|
|
Usage: "Percentage of cache memory allowance to use for database io",
|
|
Value: 50,
|
|
Category: flags.PerfCategory,
|
|
}
|
|
CacheTrieFlag = &cli.IntFlag{
|
|
Name: "cache.trie",
|
|
Usage: "Percentage of cache memory allowance to use for trie caching (default = 15% full mode, 30% archive mode)",
|
|
Value: 15,
|
|
Category: flags.PerfCategory,
|
|
}
|
|
CacheGCFlag = &cli.IntFlag{
|
|
Name: "cache.gc",
|
|
Usage: "Percentage of cache memory allowance to use for trie pruning (default = 25% full mode, 0% archive mode)",
|
|
Value: 25,
|
|
Category: flags.PerfCategory,
|
|
}
|
|
CacheSnapshotFlag = &cli.IntFlag{
|
|
Name: "cache.snapshot",
|
|
Usage: "Percentage of cache memory allowance to use for snapshot caching (default = 10% full mode, 20% archive mode)",
|
|
Value: 10,
|
|
Category: flags.PerfCategory,
|
|
}
|
|
CacheNoPrefetchFlag = &cli.BoolFlag{
|
|
Name: "cache.noprefetch",
|
|
Usage: "Disable heuristic state prefetch during block import (less CPU and disk IO, more time waiting for data)",
|
|
Category: flags.PerfCategory,
|
|
}
|
|
CachePreimagesFlag = &cli.BoolFlag{
|
|
Name: "cache.preimages",
|
|
Usage: "Enable recording the SHA3/keccak preimages of trie keys",
|
|
Category: flags.PerfCategory,
|
|
}
|
|
CacheLogSizeFlag = &cli.IntFlag{
|
|
Name: "cache.blocklogs",
|
|
Usage: "Size (in number of blocks) of the log cache for filtering",
|
|
Category: flags.PerfCategory,
|
|
Value: ethconfig.Defaults.FilterLogCacheSize,
|
|
}
|
|
FDLimitFlag = &cli.IntFlag{
|
|
Name: "fdlimit",
|
|
Usage: "Raise the open file descriptor resource limit (default = system fd limit)",
|
|
Category: flags.PerfCategory,
|
|
}
|
|
CryptoKZGFlag = &cli.StringFlag{
|
|
Name: "crypto.kzg",
|
|
Usage: "KZG library implementation to use; gokzg (recommended) or ckzg",
|
|
Value: "gokzg",
|
|
Category: flags.PerfCategory,
|
|
}
|
|
|
|
// Miner settings
|
|
MiningEnabledFlag = &cli.BoolFlag{
|
|
Name: "mine",
|
|
Usage: "Enable mining",
|
|
Category: flags.MinerCategory,
|
|
}
|
|
MinerGasLimitFlag = &cli.Uint64Flag{
|
|
Name: "miner.gaslimit",
|
|
Usage: "Target gas ceiling for mined blocks",
|
|
Value: ethconfig.Defaults.Miner.GasCeil,
|
|
Category: flags.MinerCategory,
|
|
}
|
|
MinerGasPriceFlag = &flags.BigFlag{
|
|
Name: "miner.gasprice",
|
|
Usage: "Minimum gas price for mining a transaction",
|
|
Value: ethconfig.Defaults.Miner.GasPrice,
|
|
Category: flags.MinerCategory,
|
|
}
|
|
MinerEtherbaseFlag = &cli.StringFlag{
|
|
Name: "miner.etherbase",
|
|
Usage: "0x prefixed public address for block mining rewards",
|
|
Category: flags.MinerCategory,
|
|
}
|
|
MinerExtraDataFlag = &cli.StringFlag{
|
|
Name: "miner.extradata",
|
|
Usage: "Block extra data set by the miner (default = client version)",
|
|
Category: flags.MinerCategory,
|
|
}
|
|
MinerRecommitIntervalFlag = &cli.DurationFlag{
|
|
Name: "miner.recommit",
|
|
Usage: "Time interval to recreate the block being mined",
|
|
Value: ethconfig.Defaults.Miner.Recommit,
|
|
Category: flags.MinerCategory,
|
|
}
|
|
MinerNewPayloadTimeout = &cli.DurationFlag{
|
|
Name: "miner.newpayload-timeout",
|
|
Usage: "Specify the maximum time allowance for creating a new payload",
|
|
Value: ethconfig.Defaults.Miner.NewPayloadTimeout,
|
|
Category: flags.MinerCategory,
|
|
}
|
|
|
|
// Account settings
|
|
UnlockedAccountFlag = &cli.StringFlag{
|
|
Name: "unlock",
|
|
Usage: "Comma separated list of accounts to unlock",
|
|
Value: "",
|
|
Category: flags.AccountCategory,
|
|
}
|
|
PasswordFileFlag = &cli.PathFlag{
|
|
Name: "password",
|
|
Usage: "Password file to use for non-interactive password input",
|
|
TakesFile: true,
|
|
Category: flags.AccountCategory,
|
|
}
|
|
ExternalSignerFlag = &cli.StringFlag{
|
|
Name: "signer",
|
|
Usage: "External signer (url or path to ipc file)",
|
|
Value: "",
|
|
Category: flags.AccountCategory,
|
|
}
|
|
InsecureUnlockAllowedFlag = &cli.BoolFlag{
|
|
Name: "allow-insecure-unlock",
|
|
Usage: "Allow insecure account unlocking when account-related RPCs are exposed by http",
|
|
Category: flags.AccountCategory,
|
|
}
|
|
|
|
// EVM settings
|
|
VMEnableDebugFlag = &cli.BoolFlag{
|
|
Name: "vmdebug",
|
|
Usage: "Record information useful for VM and contract debugging",
|
|
Category: flags.VMCategory,
|
|
}
|
|
|
|
// API options.
|
|
RPCGlobalGasCapFlag = &cli.Uint64Flag{
|
|
Name: "rpc.gascap",
|
|
Usage: "Sets a cap on gas that can be used in eth_call/estimateGas (0=infinite)",
|
|
Value: ethconfig.Defaults.RPCGasCap,
|
|
Category: flags.APICategory,
|
|
}
|
|
RPCGlobalEVMTimeoutFlag = &cli.DurationFlag{
|
|
Name: "rpc.evmtimeout",
|
|
Usage: "Sets a timeout used for eth_call (0=infinite)",
|
|
Value: ethconfig.Defaults.RPCEVMTimeout,
|
|
Category: flags.APICategory,
|
|
}
|
|
RPCGlobalTxFeeCapFlag = &cli.Float64Flag{
|
|
Name: "rpc.txfeecap",
|
|
Usage: "Sets a cap on transaction fee (in ether) that can be sent via the RPC APIs (0 = no cap)",
|
|
Value: ethconfig.Defaults.RPCTxFeeCap,
|
|
Category: flags.APICategory,
|
|
}
|
|
// Authenticated RPC HTTP settings
|
|
AuthListenFlag = &cli.StringFlag{
|
|
Name: "authrpc.addr",
|
|
Usage: "Listening address for authenticated APIs",
|
|
Value: node.DefaultConfig.AuthAddr,
|
|
Category: flags.APICategory,
|
|
}
|
|
AuthPortFlag = &cli.IntFlag{
|
|
Name: "authrpc.port",
|
|
Usage: "Listening port for authenticated APIs",
|
|
Value: node.DefaultConfig.AuthPort,
|
|
Category: flags.APICategory,
|
|
}
|
|
AuthVirtualHostsFlag = &cli.StringFlag{
|
|
Name: "authrpc.vhosts",
|
|
Usage: "Comma separated list of virtual hostnames from which to accept requests (server enforced). Accepts '*' wildcard.",
|
|
Value: strings.Join(node.DefaultConfig.AuthVirtualHosts, ","),
|
|
Category: flags.APICategory,
|
|
}
|
|
JWTSecretFlag = &flags.DirectoryFlag{
|
|
Name: "authrpc.jwtsecret",
|
|
Usage: "Path to a JWT secret to use for authenticated RPC endpoints",
|
|
Category: flags.APICategory,
|
|
}
|
|
|
|
// Logging and debug settings
|
|
EthStatsURLFlag = &cli.StringFlag{
|
|
Name: "ethstats",
|
|
Usage: "Reporting URL of a ethstats service (nodename:secret@host:port)",
|
|
Category: flags.MetricsCategory,
|
|
}
|
|
NoCompactionFlag = &cli.BoolFlag{
|
|
Name: "nocompaction",
|
|
Usage: "Disables db compaction after import",
|
|
Category: flags.LoggingCategory,
|
|
}
|
|
|
|
// MISC settings
|
|
SyncTargetFlag = &cli.StringFlag{
|
|
Name: "synctarget",
|
|
Usage: `Hash of the block to full sync to (dev testing feature)`,
|
|
TakesFile: true,
|
|
Category: flags.MiscCategory,
|
|
}
|
|
|
|
// RPC settings
|
|
IPCDisabledFlag = &cli.BoolFlag{
|
|
Name: "ipcdisable",
|
|
Usage: "Disable the IPC-RPC server",
|
|
Category: flags.APICategory,
|
|
}
|
|
IPCPathFlag = &flags.DirectoryFlag{
|
|
Name: "ipcpath",
|
|
Usage: "Filename for IPC socket/pipe within the datadir (explicit paths escape it)",
|
|
Category: flags.APICategory,
|
|
}
|
|
HTTPEnabledFlag = &cli.BoolFlag{
|
|
Name: "http",
|
|
Usage: "Enable the HTTP-RPC server",
|
|
Category: flags.APICategory,
|
|
}
|
|
HTTPListenAddrFlag = &cli.StringFlag{
|
|
Name: "http.addr",
|
|
Usage: "HTTP-RPC server listening interface",
|
|
Value: node.DefaultHTTPHost,
|
|
Category: flags.APICategory,
|
|
}
|
|
HTTPPortFlag = &cli.IntFlag{
|
|
Name: "http.port",
|
|
Usage: "HTTP-RPC server listening port",
|
|
Value: node.DefaultHTTPPort,
|
|
Category: flags.APICategory,
|
|
}
|
|
HTTPCORSDomainFlag = &cli.StringFlag{
|
|
Name: "http.corsdomain",
|
|
Usage: "Comma separated list of domains from which to accept cross origin requests (browser enforced)",
|
|
Value: "",
|
|
Category: flags.APICategory,
|
|
}
|
|
HTTPVirtualHostsFlag = &cli.StringFlag{
|
|
Name: "http.vhosts",
|
|
Usage: "Comma separated list of virtual hostnames from which to accept requests (server enforced). Accepts '*' wildcard.",
|
|
Value: strings.Join(node.DefaultConfig.HTTPVirtualHosts, ","),
|
|
Category: flags.APICategory,
|
|
}
|
|
HTTPApiFlag = &cli.StringFlag{
|
|
Name: "http.api",
|
|
Usage: "API's offered over the HTTP-RPC interface",
|
|
Value: "",
|
|
Category: flags.APICategory,
|
|
}
|
|
HTTPPathPrefixFlag = &cli.StringFlag{
|
|
Name: "http.rpcprefix",
|
|
Usage: "HTTP path path prefix on which JSON-RPC is served. Use '/' to serve on all paths.",
|
|
Value: "",
|
|
Category: flags.APICategory,
|
|
}
|
|
GraphQLEnabledFlag = &cli.BoolFlag{
|
|
Name: "graphql",
|
|
Usage: "Enable GraphQL on the HTTP-RPC server. Note that GraphQL can only be started if an HTTP server is started as well.",
|
|
Category: flags.APICategory,
|
|
}
|
|
GraphQLCORSDomainFlag = &cli.StringFlag{
|
|
Name: "graphql.corsdomain",
|
|
Usage: "Comma separated list of domains from which to accept cross origin requests (browser enforced)",
|
|
Value: "",
|
|
Category: flags.APICategory,
|
|
}
|
|
GraphQLVirtualHostsFlag = &cli.StringFlag{
|
|
Name: "graphql.vhosts",
|
|
Usage: "Comma separated list of virtual hostnames from which to accept requests (server enforced). Accepts '*' wildcard.",
|
|
Value: strings.Join(node.DefaultConfig.GraphQLVirtualHosts, ","),
|
|
Category: flags.APICategory,
|
|
}
|
|
WSEnabledFlag = &cli.BoolFlag{
|
|
Name: "ws",
|
|
Usage: "Enable the WS-RPC server",
|
|
Category: flags.APICategory,
|
|
}
|
|
WSListenAddrFlag = &cli.StringFlag{
|
|
Name: "ws.addr",
|
|
Usage: "WS-RPC server listening interface",
|
|
Value: node.DefaultWSHost,
|
|
Category: flags.APICategory,
|
|
}
|
|
WSPortFlag = &cli.IntFlag{
|
|
Name: "ws.port",
|
|
Usage: "WS-RPC server listening port",
|
|
Value: node.DefaultWSPort,
|
|
Category: flags.APICategory,
|
|
}
|
|
WSApiFlag = &cli.StringFlag{
|
|
Name: "ws.api",
|
|
Usage: "API's offered over the WS-RPC interface",
|
|
Value: "",
|
|
Category: flags.APICategory,
|
|
}
|
|
WSAllowedOriginsFlag = &cli.StringFlag{
|
|
Name: "ws.origins",
|
|
Usage: "Origins from which to accept websockets requests",
|
|
Value: "",
|
|
Category: flags.APICategory,
|
|
}
|
|
WSPathPrefixFlag = &cli.StringFlag{
|
|
Name: "ws.rpcprefix",
|
|
Usage: "HTTP path prefix on which JSON-RPC is served. Use '/' to serve on all paths.",
|
|
Value: "",
|
|
Category: flags.APICategory,
|
|
}
|
|
ExecFlag = &cli.StringFlag{
|
|
Name: "exec",
|
|
Usage: "Execute JavaScript statement",
|
|
Category: flags.APICategory,
|
|
}
|
|
PreloadJSFlag = &cli.StringFlag{
|
|
Name: "preload",
|
|
Usage: "Comma separated list of JavaScript files to preload into the console",
|
|
Category: flags.APICategory,
|
|
}
|
|
AllowUnprotectedTxs = &cli.BoolFlag{
|
|
Name: "rpc.allow-unprotected-txs",
|
|
Usage: "Allow for unprotected (non EIP155 signed) transactions to be submitted via RPC",
|
|
Category: flags.APICategory,
|
|
}
|
|
BatchRequestLimit = &cli.IntFlag{
|
|
Name: "rpc.batch-request-limit",
|
|
Usage: "Maximum number of requests in a batch",
|
|
Value: node.DefaultConfig.BatchRequestLimit,
|
|
Category: flags.APICategory,
|
|
}
|
|
BatchResponseMaxSize = &cli.IntFlag{
|
|
Name: "rpc.batch-response-max-size",
|
|
Usage: "Maximum number of bytes returned from a batched call",
|
|
Value: node.DefaultConfig.BatchResponseMaxSize,
|
|
Category: flags.APICategory,
|
|
}
|
|
EnablePersonal = &cli.BoolFlag{
|
|
Name: "rpc.enabledeprecatedpersonal",
|
|
Usage: "Enables the (deprecated) personal namespace",
|
|
Category: flags.APICategory,
|
|
}
|
|
|
|
// Network Settings
|
|
MaxPeersFlag = &cli.IntFlag{
|
|
Name: "maxpeers",
|
|
Usage: "Maximum number of network peers (network disabled if set to 0)",
|
|
Value: node.DefaultConfig.P2P.MaxPeers,
|
|
Category: flags.NetworkingCategory,
|
|
}
|
|
MaxPendingPeersFlag = &cli.IntFlag{
|
|
Name: "maxpendpeers",
|
|
Usage: "Maximum number of pending connection attempts (defaults used if set to 0)",
|
|
Value: node.DefaultConfig.P2P.MaxPendingPeers,
|
|
Category: flags.NetworkingCategory,
|
|
}
|
|
ListenPortFlag = &cli.IntFlag{
|
|
Name: "port",
|
|
Usage: "Network listening port",
|
|
Value: 30303,
|
|
Category: flags.NetworkingCategory,
|
|
}
|
|
BootnodesFlag = &cli.StringFlag{
|
|
Name: "bootnodes",
|
|
Usage: "Comma separated enode URLs for P2P discovery bootstrap",
|
|
Value: "",
|
|
Category: flags.NetworkingCategory,
|
|
}
|
|
NodeKeyFileFlag = &cli.StringFlag{
|
|
Name: "nodekey",
|
|
Usage: "P2P node key file",
|
|
Category: flags.NetworkingCategory,
|
|
}
|
|
NodeKeyHexFlag = &cli.StringFlag{
|
|
Name: "nodekeyhex",
|
|
Usage: "P2P node key as hex (for testing)",
|
|
Category: flags.NetworkingCategory,
|
|
}
|
|
NATFlag = &cli.StringFlag{
|
|
Name: "nat",
|
|
Usage: "NAT port mapping mechanism (any|none|upnp|pmp|pmp:<IP>|extip:<IP>)",
|
|
Value: "any",
|
|
Category: flags.NetworkingCategory,
|
|
}
|
|
NoDiscoverFlag = &cli.BoolFlag{
|
|
Name: "nodiscover",
|
|
Usage: "Disables the peer discovery mechanism (manual peer addition)",
|
|
Category: flags.NetworkingCategory,
|
|
}
|
|
DiscoveryV4Flag = &cli.BoolFlag{
|
|
Name: "discovery.v4",
|
|
Aliases: []string{"discv4"},
|
|
Usage: "Enables the V4 discovery mechanism",
|
|
Category: flags.NetworkingCategory,
|
|
Value: true,
|
|
}
|
|
DiscoveryV5Flag = &cli.BoolFlag{
|
|
Name: "discovery.v5",
|
|
Aliases: []string{"discv5"},
|
|
Usage: "Enables the experimental RLPx V5 (Topic Discovery) mechanism",
|
|
Category: flags.NetworkingCategory,
|
|
}
|
|
NetrestrictFlag = &cli.StringFlag{
|
|
Name: "netrestrict",
|
|
Usage: "Restricts network communication to the given IP networks (CIDR masks)",
|
|
Category: flags.NetworkingCategory,
|
|
}
|
|
DNSDiscoveryFlag = &cli.StringFlag{
|
|
Name: "discovery.dns",
|
|
Usage: "Sets DNS discovery entry points (use \"\" to disable DNS)",
|
|
Category: flags.NetworkingCategory,
|
|
}
|
|
DiscoveryPortFlag = &cli.IntFlag{
|
|
Name: "discovery.port",
|
|
Usage: "Use a custom UDP port for P2P discovery",
|
|
Value: 30303,
|
|
Category: flags.NetworkingCategory,
|
|
}
|
|
|
|
// Console
|
|
JSpathFlag = &flags.DirectoryFlag{
|
|
Name: "jspath",
|
|
Usage: "JavaScript root path for `loadScript`",
|
|
Value: flags.DirectoryString("."),
|
|
Category: flags.APICategory,
|
|
}
|
|
HttpHeaderFlag = &cli.StringSliceFlag{
|
|
Name: "header",
|
|
Aliases: []string{"H"},
|
|
Usage: "Pass custom headers to the RPC server when using --" + RemoteDBFlag.Name + " or the geth attach console. This flag can be given multiple times.",
|
|
Category: flags.APICategory,
|
|
}
|
|
|
|
// Gas price oracle settings
|
|
GpoBlocksFlag = &cli.IntFlag{
|
|
Name: "gpo.blocks",
|
|
Usage: "Number of recent blocks to check for gas prices",
|
|
Value: ethconfig.Defaults.GPO.Blocks,
|
|
Category: flags.GasPriceCategory,
|
|
}
|
|
GpoPercentileFlag = &cli.IntFlag{
|
|
Name: "gpo.percentile",
|
|
Usage: "Suggested gas price is the given percentile of a set of recent transaction gas prices",
|
|
Value: ethconfig.Defaults.GPO.Percentile,
|
|
Category: flags.GasPriceCategory,
|
|
}
|
|
GpoMaxGasPriceFlag = &cli.Int64Flag{
|
|
Name: "gpo.maxprice",
|
|
Usage: "Maximum transaction priority fee (or gasprice before London fork) to be recommended by gpo",
|
|
Value: ethconfig.Defaults.GPO.MaxPrice.Int64(),
|
|
Category: flags.GasPriceCategory,
|
|
}
|
|
GpoIgnoreGasPriceFlag = &cli.Int64Flag{
|
|
Name: "gpo.ignoreprice",
|
|
Usage: "Gas price below which gpo will ignore transactions",
|
|
Value: ethconfig.Defaults.GPO.IgnorePrice.Int64(),
|
|
Category: flags.GasPriceCategory,
|
|
}
|
|
|
|
// Metrics flags
|
|
MetricsEnabledFlag = &cli.BoolFlag{
|
|
Name: "metrics",
|
|
Usage: "Enable metrics collection and reporting",
|
|
Category: flags.MetricsCategory,
|
|
}
|
|
MetricsEnabledExpensiveFlag = &cli.BoolFlag{
|
|
Name: "metrics.expensive",
|
|
Usage: "Enable expensive metrics collection and reporting",
|
|
Category: flags.MetricsCategory,
|
|
}
|
|
|
|
// MetricsHTTPFlag defines the endpoint for a stand-alone metrics HTTP endpoint.
|
|
// Since the pprof service enables sensitive/vulnerable behavior, this allows a user
|
|
// to enable a public-OK metrics endpoint without having to worry about ALSO exposing
|
|
// other profiling behavior or information.
|
|
MetricsHTTPFlag = &cli.StringFlag{
|
|
Name: "metrics.addr",
|
|
Usage: `Enable stand-alone metrics HTTP server listening interface.`,
|
|
Category: flags.MetricsCategory,
|
|
}
|
|
MetricsPortFlag = &cli.IntFlag{
|
|
Name: "metrics.port",
|
|
Usage: `Metrics HTTP server listening port.
|
|
Please note that --` + MetricsHTTPFlag.Name + ` must be set to start the server.`,
|
|
Value: metrics.DefaultConfig.Port,
|
|
Category: flags.MetricsCategory,
|
|
}
|
|
MetricsEnableInfluxDBFlag = &cli.BoolFlag{
|
|
Name: "metrics.influxdb",
|
|
Usage: "Enable metrics export/push to an external InfluxDB database",
|
|
Category: flags.MetricsCategory,
|
|
}
|
|
MetricsInfluxDBEndpointFlag = &cli.StringFlag{
|
|
Name: "metrics.influxdb.endpoint",
|
|
Usage: "InfluxDB API endpoint to report metrics to",
|
|
Value: metrics.DefaultConfig.InfluxDBEndpoint,
|
|
Category: flags.MetricsCategory,
|
|
}
|
|
MetricsInfluxDBDatabaseFlag = &cli.StringFlag{
|
|
Name: "metrics.influxdb.database",
|
|
Usage: "InfluxDB database name to push reported metrics to",
|
|
Value: metrics.DefaultConfig.InfluxDBDatabase,
|
|
Category: flags.MetricsCategory,
|
|
}
|
|
MetricsInfluxDBUsernameFlag = &cli.StringFlag{
|
|
Name: "metrics.influxdb.username",
|
|
Usage: "Username to authorize access to the database",
|
|
Value: metrics.DefaultConfig.InfluxDBUsername,
|
|
Category: flags.MetricsCategory,
|
|
}
|
|
MetricsInfluxDBPasswordFlag = &cli.StringFlag{
|
|
Name: "metrics.influxdb.password",
|
|
Usage: "Password to authorize access to the database",
|
|
Value: metrics.DefaultConfig.InfluxDBPassword,
|
|
Category: flags.MetricsCategory,
|
|
}
|
|
// Tags are part of every measurement sent to InfluxDB. Queries on tags are faster in InfluxDB.
|
|
// For example `host` tag could be used so that we can group all nodes and average a measurement
|
|
// across all of them, but also so that we can select a specific node and inspect its measurements.
|
|
// https://docs.influxdata.com/influxdb/v1.4/concepts/key_concepts/#tag-key
|
|
MetricsInfluxDBTagsFlag = &cli.StringFlag{
|
|
Name: "metrics.influxdb.tags",
|
|
Usage: "Comma-separated InfluxDB tags (key/values) attached to all measurements",
|
|
Value: metrics.DefaultConfig.InfluxDBTags,
|
|
Category: flags.MetricsCategory,
|
|
}
|
|
|
|
MetricsEnableInfluxDBV2Flag = &cli.BoolFlag{
|
|
Name: "metrics.influxdbv2",
|
|
Usage: "Enable metrics export/push to an external InfluxDB v2 database",
|
|
Category: flags.MetricsCategory,
|
|
}
|
|
|
|
MetricsInfluxDBTokenFlag = &cli.StringFlag{
|
|
Name: "metrics.influxdb.token",
|
|
Usage: "Token to authorize access to the database (v2 only)",
|
|
Value: metrics.DefaultConfig.InfluxDBToken,
|
|
Category: flags.MetricsCategory,
|
|
}
|
|
|
|
MetricsInfluxDBBucketFlag = &cli.StringFlag{
|
|
Name: "metrics.influxdb.bucket",
|
|
Usage: "InfluxDB bucket name to push reported metrics to (v2 only)",
|
|
Value: metrics.DefaultConfig.InfluxDBBucket,
|
|
Category: flags.MetricsCategory,
|
|
}
|
|
|
|
MetricsInfluxDBOrganizationFlag = &cli.StringFlag{
|
|
Name: "metrics.influxdb.organization",
|
|
Usage: "InfluxDB organization name (v2 only)",
|
|
Value: metrics.DefaultConfig.InfluxDBOrganization,
|
|
Category: flags.MetricsCategory,
|
|
}
|
|
)
|
|
|
|
var (
|
|
// TestnetFlags is the flag group of all built-in supported testnets.
|
|
TestnetFlags = []cli.Flag{
|
|
GoerliFlag,
|
|
SepoliaFlag,
|
|
HoleskyFlag,
|
|
}
|
|
// NetworkFlags is the flag group of all built-in supported networks.
|
|
NetworkFlags = append([]cli.Flag{MainnetFlag}, TestnetFlags...)
|
|
|
|
// DatabaseFlags is the flag group of all database flags.
|
|
DatabaseFlags = []cli.Flag{
|
|
DataDirFlag,
|
|
AncientFlag,
|
|
RemoteDBFlag,
|
|
DBEngineFlag,
|
|
StateSchemeFlag,
|
|
HttpHeaderFlag,
|
|
}
|
|
)
|
|
|
|
// MakeDataDir retrieves the currently requested data directory, terminating
|
|
// if none (or the empty string) is specified. If the node is starting a testnet,
|
|
// then a subdirectory of the specified datadir will be used.
|
|
func MakeDataDir(ctx *cli.Context) string {
|
|
if path := ctx.String(DataDirFlag.Name); path != "" {
|
|
if ctx.Bool(GoerliFlag.Name) {
|
|
return filepath.Join(path, "goerli")
|
|
}
|
|
if ctx.Bool(SepoliaFlag.Name) {
|
|
return filepath.Join(path, "sepolia")
|
|
}
|
|
if ctx.Bool(HoleskyFlag.Name) {
|
|
return filepath.Join(path, "holesky")
|
|
}
|
|
return path
|
|
}
|
|
Fatalf("Cannot determine default data directory, please set manually (--datadir)")
|
|
return ""
|
|
}
|
|
|
|
// setNodeKey creates a node key from set command line flags, either loading it
|
|
// from a file or as a specified hex value. If neither flags were provided, this
|
|
// method returns nil and an ephemeral key is to be generated.
|
|
func setNodeKey(ctx *cli.Context, cfg *p2p.Config) {
|
|
var (
|
|
hex = ctx.String(NodeKeyHexFlag.Name)
|
|
file = ctx.String(NodeKeyFileFlag.Name)
|
|
key *ecdsa.PrivateKey
|
|
err error
|
|
)
|
|
switch {
|
|
case file != "" && hex != "":
|
|
Fatalf("Options %q and %q are mutually exclusive", NodeKeyFileFlag.Name, NodeKeyHexFlag.Name)
|
|
case file != "":
|
|
if key, err = crypto.LoadECDSA(file); err != nil {
|
|
Fatalf("Option %q: %v", NodeKeyFileFlag.Name, err)
|
|
}
|
|
cfg.PrivateKey = key
|
|
case hex != "":
|
|
if key, err = crypto.HexToECDSA(hex); err != nil {
|
|
Fatalf("Option %q: %v", NodeKeyHexFlag.Name, err)
|
|
}
|
|
cfg.PrivateKey = key
|
|
}
|
|
}
|
|
|
|
// setNodeUserIdent creates the user identifier from CLI flags.
|
|
func setNodeUserIdent(ctx *cli.Context, cfg *node.Config) {
|
|
if identity := ctx.String(IdentityFlag.Name); len(identity) > 0 {
|
|
cfg.UserIdent = identity
|
|
}
|
|
}
|
|
|
|
// setBootstrapNodes creates a list of bootstrap nodes from the command line
|
|
// flags, reverting to pre-configured ones if none have been specified.
|
|
// Priority order for bootnodes configuration:
|
|
//
|
|
// 1. --bootnodes flag
|
|
// 2. Config file
|
|
// 3. Network preset flags (e.g. --goerli)
|
|
// 4. default to mainnet nodes
|
|
func setBootstrapNodes(ctx *cli.Context, cfg *p2p.Config) {
|
|
urls := params.MainnetBootnodes
|
|
// begin PluGeth injection
|
|
if pluginUrls := pluginSetBootstrapNodes(); pluginUrls != nil {
|
|
urls = pluginUrls
|
|
}
|
|
// end PluGeth injection
|
|
if ctx.IsSet(BootnodesFlag.Name) {
|
|
urls = SplitAndTrim(ctx.String(BootnodesFlag.Name))
|
|
} else {
|
|
if cfg.BootstrapNodes != nil {
|
|
return // Already set by config file, don't apply defaults.
|
|
}
|
|
switch {
|
|
case ctx.Bool(HoleskyFlag.Name):
|
|
urls = params.HoleskyBootnodes
|
|
case ctx.Bool(SepoliaFlag.Name):
|
|
urls = params.SepoliaBootnodes
|
|
case ctx.Bool(GoerliFlag.Name):
|
|
urls = params.GoerliBootnodes
|
|
}
|
|
}
|
|
cfg.BootstrapNodes = mustParseBootnodes(urls)
|
|
}
|
|
|
|
func mustParseBootnodes(urls []string) []*enode.Node {
|
|
nodes := make([]*enode.Node, 0, len(urls))
|
|
for _, url := range urls {
|
|
if url != "" {
|
|
node, err := enode.Parse(enode.ValidSchemes, url)
|
|
if err != nil {
|
|
log.Crit("Bootstrap URL invalid", "enode", url, "err", err)
|
|
return nil
|
|
}
|
|
nodes = append(nodes, node)
|
|
}
|
|
}
|
|
return nodes
|
|
}
|
|
|
|
// setBootstrapNodesV5 creates a list of bootstrap nodes from the command line
|
|
// flags, reverting to pre-configured ones if none have been specified.
|
|
func setBootstrapNodesV5(ctx *cli.Context, cfg *p2p.Config) {
|
|
urls := params.V5Bootnodes
|
|
switch {
|
|
case ctx.IsSet(BootnodesFlag.Name):
|
|
urls = SplitAndTrim(ctx.String(BootnodesFlag.Name))
|
|
case cfg.BootstrapNodesV5 != nil:
|
|
return // already set, don't apply defaults.
|
|
}
|
|
|
|
cfg.BootstrapNodesV5 = make([]*enode.Node, 0, len(urls))
|
|
for _, url := range urls {
|
|
if url != "" {
|
|
node, err := enode.Parse(enode.ValidSchemes, url)
|
|
if err != nil {
|
|
log.Error("Bootstrap URL invalid", "enode", url, "err", err)
|
|
continue
|
|
}
|
|
cfg.BootstrapNodesV5 = append(cfg.BootstrapNodesV5, node)
|
|
}
|
|
}
|
|
}
|
|
|
|
// setListenAddress creates TCP/UDP listening address strings from set command
|
|
// line flags
|
|
func setListenAddress(ctx *cli.Context, cfg *p2p.Config) {
|
|
if ctx.IsSet(ListenPortFlag.Name) {
|
|
cfg.ListenAddr = fmt.Sprintf(":%d", ctx.Int(ListenPortFlag.Name))
|
|
}
|
|
if ctx.IsSet(DiscoveryPortFlag.Name) {
|
|
cfg.DiscAddr = fmt.Sprintf(":%d", ctx.Int(DiscoveryPortFlag.Name))
|
|
}
|
|
}
|
|
|
|
// setNAT creates a port mapper from command line flags.
|
|
func setNAT(ctx *cli.Context, cfg *p2p.Config) {
|
|
if ctx.IsSet(NATFlag.Name) {
|
|
natif, err := nat.Parse(ctx.String(NATFlag.Name))
|
|
if err != nil {
|
|
Fatalf("Option %s: %v", NATFlag.Name, err)
|
|
}
|
|
cfg.NAT = natif
|
|
}
|
|
}
|
|
|
|
// SplitAndTrim splits input separated by a comma
|
|
// and trims excessive white space from the substrings.
|
|
func SplitAndTrim(input string) (ret []string) {
|
|
l := strings.Split(input, ",")
|
|
for _, r := range l {
|
|
if r = strings.TrimSpace(r); r != "" {
|
|
ret = append(ret, r)
|
|
}
|
|
}
|
|
return ret
|
|
}
|
|
|
|
// setHTTP creates the HTTP RPC listener interface string from the set
|
|
// command line flags, returning empty if the HTTP endpoint is disabled.
|
|
func setHTTP(ctx *cli.Context, cfg *node.Config) {
|
|
if ctx.Bool(HTTPEnabledFlag.Name) {
|
|
if cfg.HTTPHost == "" {
|
|
cfg.HTTPHost = "127.0.0.1"
|
|
}
|
|
if ctx.IsSet(HTTPListenAddrFlag.Name) {
|
|
cfg.HTTPHost = ctx.String(HTTPListenAddrFlag.Name)
|
|
}
|
|
}
|
|
|
|
if ctx.IsSet(HTTPPortFlag.Name) {
|
|
cfg.HTTPPort = ctx.Int(HTTPPortFlag.Name)
|
|
}
|
|
|
|
if ctx.IsSet(AuthListenFlag.Name) {
|
|
cfg.AuthAddr = ctx.String(AuthListenFlag.Name)
|
|
}
|
|
|
|
if ctx.IsSet(AuthPortFlag.Name) {
|
|
cfg.AuthPort = ctx.Int(AuthPortFlag.Name)
|
|
}
|
|
|
|
if ctx.IsSet(AuthVirtualHostsFlag.Name) {
|
|
cfg.AuthVirtualHosts = SplitAndTrim(ctx.String(AuthVirtualHostsFlag.Name))
|
|
}
|
|
|
|
if ctx.IsSet(HTTPCORSDomainFlag.Name) {
|
|
cfg.HTTPCors = SplitAndTrim(ctx.String(HTTPCORSDomainFlag.Name))
|
|
}
|
|
|
|
if ctx.IsSet(HTTPApiFlag.Name) {
|
|
cfg.HTTPModules = SplitAndTrim(ctx.String(HTTPApiFlag.Name))
|
|
}
|
|
|
|
if ctx.IsSet(HTTPVirtualHostsFlag.Name) {
|
|
cfg.HTTPVirtualHosts = SplitAndTrim(ctx.String(HTTPVirtualHostsFlag.Name))
|
|
}
|
|
|
|
if ctx.IsSet(HTTPPathPrefixFlag.Name) {
|
|
cfg.HTTPPathPrefix = ctx.String(HTTPPathPrefixFlag.Name)
|
|
}
|
|
if ctx.IsSet(AllowUnprotectedTxs.Name) {
|
|
cfg.AllowUnprotectedTxs = ctx.Bool(AllowUnprotectedTxs.Name)
|
|
}
|
|
|
|
if ctx.IsSet(BatchRequestLimit.Name) {
|
|
cfg.BatchRequestLimit = ctx.Int(BatchRequestLimit.Name)
|
|
}
|
|
|
|
if ctx.IsSet(BatchResponseMaxSize.Name) {
|
|
cfg.BatchResponseMaxSize = ctx.Int(BatchResponseMaxSize.Name)
|
|
}
|
|
}
|
|
|
|
// setGraphQL creates the GraphQL listener interface string from the set
|
|
// command line flags, returning empty if the GraphQL endpoint is disabled.
|
|
func setGraphQL(ctx *cli.Context, cfg *node.Config) {
|
|
if ctx.IsSet(GraphQLCORSDomainFlag.Name) {
|
|
cfg.GraphQLCors = SplitAndTrim(ctx.String(GraphQLCORSDomainFlag.Name))
|
|
}
|
|
if ctx.IsSet(GraphQLVirtualHostsFlag.Name) {
|
|
cfg.GraphQLVirtualHosts = SplitAndTrim(ctx.String(GraphQLVirtualHostsFlag.Name))
|
|
}
|
|
}
|
|
|
|
// setWS creates the WebSocket RPC listener interface string from the set
|
|
// command line flags, returning empty if the HTTP endpoint is disabled.
|
|
func setWS(ctx *cli.Context, cfg *node.Config) {
|
|
if ctx.Bool(WSEnabledFlag.Name) {
|
|
if cfg.WSHost == "" {
|
|
cfg.WSHost = "127.0.0.1"
|
|
}
|
|
if ctx.IsSet(WSListenAddrFlag.Name) {
|
|
cfg.WSHost = ctx.String(WSListenAddrFlag.Name)
|
|
}
|
|
}
|
|
if ctx.IsSet(WSPortFlag.Name) {
|
|
cfg.WSPort = ctx.Int(WSPortFlag.Name)
|
|
}
|
|
|
|
if ctx.IsSet(WSAllowedOriginsFlag.Name) {
|
|
cfg.WSOrigins = SplitAndTrim(ctx.String(WSAllowedOriginsFlag.Name))
|
|
}
|
|
|
|
if ctx.IsSet(WSApiFlag.Name) {
|
|
cfg.WSModules = SplitAndTrim(ctx.String(WSApiFlag.Name))
|
|
}
|
|
|
|
if ctx.IsSet(WSPathPrefixFlag.Name) {
|
|
cfg.WSPathPrefix = ctx.String(WSPathPrefixFlag.Name)
|
|
}
|
|
}
|
|
|
|
// setIPC creates an IPC path configuration from the set command line flags,
|
|
// returning an empty string if IPC was explicitly disabled, or the set path.
|
|
func setIPC(ctx *cli.Context, cfg *node.Config) {
|
|
CheckExclusive(ctx, IPCDisabledFlag, IPCPathFlag)
|
|
switch {
|
|
case ctx.Bool(IPCDisabledFlag.Name):
|
|
cfg.IPCPath = ""
|
|
case ctx.IsSet(IPCPathFlag.Name):
|
|
cfg.IPCPath = ctx.String(IPCPathFlag.Name)
|
|
}
|
|
}
|
|
|
|
// setLes shows the deprecation warnings for LES flags.
|
|
func setLes(ctx *cli.Context, cfg *ethconfig.Config) {
|
|
if ctx.IsSet(LightServeFlag.Name) {
|
|
log.Warn("The light server has been deprecated, please remove this flag", "flag", LightServeFlag.Name)
|
|
}
|
|
if ctx.IsSet(LightIngressFlag.Name) {
|
|
log.Warn("The light server has been deprecated, please remove this flag", "flag", LightIngressFlag.Name)
|
|
}
|
|
if ctx.IsSet(LightEgressFlag.Name) {
|
|
log.Warn("The light server has been deprecated, please remove this flag", "flag", LightEgressFlag.Name)
|
|
}
|
|
if ctx.IsSet(LightMaxPeersFlag.Name) {
|
|
log.Warn("The light server has been deprecated, please remove this flag", "flag", LightMaxPeersFlag.Name)
|
|
}
|
|
if ctx.IsSet(LightNoPruneFlag.Name) {
|
|
log.Warn("The light server has been deprecated, please remove this flag", "flag", LightNoPruneFlag.Name)
|
|
}
|
|
if ctx.IsSet(LightNoSyncServeFlag.Name) {
|
|
log.Warn("The light server has been deprecated, please remove this flag", "flag", LightNoSyncServeFlag.Name)
|
|
}
|
|
}
|
|
|
|
// MakeDatabaseHandles raises out the number of allowed file handles per process
|
|
// for Geth and returns half of the allowance to assign to the database.
|
|
func MakeDatabaseHandles(max int) int {
|
|
limit, err := fdlimit.Maximum()
|
|
if err != nil {
|
|
Fatalf("Failed to retrieve file descriptor allowance: %v", err)
|
|
}
|
|
switch {
|
|
case max == 0:
|
|
// User didn't specify a meaningful value, use system limits
|
|
case max < 128:
|
|
// User specified something unhealthy, just use system defaults
|
|
log.Error("File descriptor limit invalid (<128)", "had", max, "updated", limit)
|
|
case max > limit:
|
|
// User requested more than the OS allows, notify that we can't allocate it
|
|
log.Warn("Requested file descriptors denied by OS", "req", max, "limit", limit)
|
|
default:
|
|
// User limit is meaningful and within allowed range, use that
|
|
limit = max
|
|
}
|
|
raised, err := fdlimit.Raise(uint64(limit))
|
|
if err != nil {
|
|
Fatalf("Failed to raise file descriptor allowance: %v", err)
|
|
}
|
|
return int(raised / 2) // Leave half for networking and other stuff
|
|
}
|
|
|
|
// MakeAddress converts an account specified directly as a hex encoded string or
|
|
// a key index in the key store to an internal account representation.
|
|
func MakeAddress(ks *keystore.KeyStore, account string) (accounts.Account, error) {
|
|
// If the specified account is a valid address, return it
|
|
if common.IsHexAddress(account) {
|
|
return accounts.Account{Address: common.HexToAddress(account)}, nil
|
|
}
|
|
// Otherwise try to interpret the account as a keystore index
|
|
index, err := strconv.Atoi(account)
|
|
if err != nil || index < 0 {
|
|
return accounts.Account{}, fmt.Errorf("invalid account address or index %q", account)
|
|
}
|
|
log.Warn("-------------------------------------------------------------------")
|
|
log.Warn("Referring to accounts by order in the keystore folder is dangerous!")
|
|
log.Warn("This functionality is deprecated and will be removed in the future!")
|
|
log.Warn("Please use explicit addresses! (can search via `geth account list`)")
|
|
log.Warn("-------------------------------------------------------------------")
|
|
|
|
accs := ks.Accounts()
|
|
if len(accs) <= index {
|
|
return accounts.Account{}, fmt.Errorf("index %d higher than number of accounts %d", index, len(accs))
|
|
}
|
|
return accs[index], nil
|
|
}
|
|
|
|
// setEtherbase retrieves the etherbase from the directly specified command line flags.
|
|
func setEtherbase(ctx *cli.Context, cfg *ethconfig.Config) {
|
|
if !ctx.IsSet(MinerEtherbaseFlag.Name) {
|
|
return
|
|
}
|
|
addr := ctx.String(MinerEtherbaseFlag.Name)
|
|
if strings.HasPrefix(addr, "0x") || strings.HasPrefix(addr, "0X") {
|
|
addr = addr[2:]
|
|
}
|
|
b, err := hex.DecodeString(addr)
|
|
if err != nil || len(b) != common.AddressLength {
|
|
Fatalf("-%s: invalid etherbase address %q", MinerEtherbaseFlag.Name, addr)
|
|
return
|
|
}
|
|
cfg.Miner.Etherbase = common.BytesToAddress(b)
|
|
}
|
|
|
|
// MakePasswordList reads password lines from the file specified by the global --password flag.
|
|
func MakePasswordList(ctx *cli.Context) []string {
|
|
path := ctx.Path(PasswordFileFlag.Name)
|
|
if path == "" {
|
|
return nil
|
|
}
|
|
text, err := os.ReadFile(path)
|
|
if err != nil {
|
|
Fatalf("Failed to read password file: %v", err)
|
|
}
|
|
lines := strings.Split(string(text), "\n")
|
|
// Sanitise DOS line endings.
|
|
for i := range lines {
|
|
lines[i] = strings.TrimRight(lines[i], "\r")
|
|
}
|
|
return lines
|
|
}
|
|
|
|
func SetP2PConfig(ctx *cli.Context, cfg *p2p.Config) {
|
|
setNodeKey(ctx, cfg)
|
|
setNAT(ctx, cfg)
|
|
setListenAddress(ctx, cfg)
|
|
setBootstrapNodes(ctx, cfg)
|
|
setBootstrapNodesV5(ctx, cfg)
|
|
|
|
if ctx.IsSet(MaxPeersFlag.Name) {
|
|
cfg.MaxPeers = ctx.Int(MaxPeersFlag.Name)
|
|
}
|
|
ethPeers := cfg.MaxPeers
|
|
log.Info("Maximum peer count", "ETH", ethPeers, "total", cfg.MaxPeers)
|
|
|
|
if ctx.IsSet(MaxPendingPeersFlag.Name) {
|
|
cfg.MaxPendingPeers = ctx.Int(MaxPendingPeersFlag.Name)
|
|
}
|
|
if ctx.IsSet(NoDiscoverFlag.Name) {
|
|
cfg.NoDiscovery = true
|
|
}
|
|
|
|
CheckExclusive(ctx, DiscoveryV4Flag, NoDiscoverFlag)
|
|
CheckExclusive(ctx, DiscoveryV5Flag, NoDiscoverFlag)
|
|
cfg.DiscoveryV4 = ctx.Bool(DiscoveryV4Flag.Name)
|
|
cfg.DiscoveryV5 = ctx.Bool(DiscoveryV5Flag.Name)
|
|
|
|
if netrestrict := ctx.String(NetrestrictFlag.Name); netrestrict != "" {
|
|
list, err := netutil.ParseNetlist(netrestrict)
|
|
if err != nil {
|
|
Fatalf("Option %q: %v", NetrestrictFlag.Name, err)
|
|
}
|
|
cfg.NetRestrict = list
|
|
}
|
|
|
|
if ctx.Bool(DeveloperFlag.Name) {
|
|
// --dev mode can't use p2p networking.
|
|
cfg.MaxPeers = 0
|
|
cfg.ListenAddr = ""
|
|
cfg.NoDial = true
|
|
cfg.NoDiscovery = true
|
|
cfg.DiscoveryV5 = false
|
|
}
|
|
}
|
|
|
|
// SetNodeConfig applies node-related command line flags to the config.
|
|
func SetNodeConfig(ctx *cli.Context, cfg *node.Config) {
|
|
SetP2PConfig(ctx, &cfg.P2P)
|
|
setIPC(ctx, cfg)
|
|
setHTTP(ctx, cfg)
|
|
setGraphQL(ctx, cfg)
|
|
setWS(ctx, cfg)
|
|
setNodeUserIdent(ctx, cfg)
|
|
SetDataDir(ctx, cfg)
|
|
setSmartCard(ctx, cfg)
|
|
|
|
if ctx.IsSet(JWTSecretFlag.Name) {
|
|
cfg.JWTSecret = ctx.String(JWTSecretFlag.Name)
|
|
}
|
|
|
|
if ctx.IsSet(EnablePersonal.Name) {
|
|
cfg.EnablePersonal = true
|
|
}
|
|
|
|
if ctx.IsSet(ExternalSignerFlag.Name) {
|
|
cfg.ExternalSigner = ctx.String(ExternalSignerFlag.Name)
|
|
}
|
|
|
|
if ctx.IsSet(KeyStoreDirFlag.Name) {
|
|
cfg.KeyStoreDir = ctx.String(KeyStoreDirFlag.Name)
|
|
}
|
|
if ctx.IsSet(DeveloperFlag.Name) {
|
|
cfg.UseLightweightKDF = true
|
|
}
|
|
if ctx.IsSet(LightKDFFlag.Name) {
|
|
cfg.UseLightweightKDF = ctx.Bool(LightKDFFlag.Name)
|
|
}
|
|
if ctx.IsSet(NoUSBFlag.Name) || cfg.NoUSB {
|
|
log.Warn("Option nousb is deprecated and USB is deactivated by default. Use --usb to enable")
|
|
}
|
|
if ctx.IsSet(USBFlag.Name) {
|
|
cfg.USB = ctx.Bool(USBFlag.Name)
|
|
}
|
|
if ctx.IsSet(InsecureUnlockAllowedFlag.Name) {
|
|
cfg.InsecureUnlockAllowed = ctx.Bool(InsecureUnlockAllowedFlag.Name)
|
|
}
|
|
if ctx.IsSet(DBEngineFlag.Name) {
|
|
dbEngine := ctx.String(DBEngineFlag.Name)
|
|
if dbEngine != "leveldb" && dbEngine != "pebble" {
|
|
Fatalf("Invalid choice for db.engine '%s', allowed 'leveldb' or 'pebble'", dbEngine)
|
|
}
|
|
log.Info(fmt.Sprintf("Using %s as db engine", dbEngine))
|
|
cfg.DBEngine = dbEngine
|
|
}
|
|
// deprecation notice for log debug flags (TODO: find a more appropriate place to put these?)
|
|
if ctx.IsSet(LogBacktraceAtFlag.Name) {
|
|
log.Warn("log.backtrace flag is deprecated")
|
|
}
|
|
if ctx.IsSet(LogDebugFlag.Name) {
|
|
log.Warn("log.debug flag is deprecated")
|
|
}
|
|
}
|
|
|
|
func setSmartCard(ctx *cli.Context, cfg *node.Config) {
|
|
// Skip enabling smartcards if no path is set
|
|
path := ctx.String(SmartCardDaemonPathFlag.Name)
|
|
if path == "" {
|
|
return
|
|
}
|
|
// Sanity check that the smartcard path is valid
|
|
fi, err := os.Stat(path)
|
|
if err != nil {
|
|
log.Info("Smartcard socket not found, disabling", "err", err)
|
|
return
|
|
}
|
|
if fi.Mode()&os.ModeType != os.ModeSocket {
|
|
log.Error("Invalid smartcard daemon path", "path", path, "type", fi.Mode().String())
|
|
return
|
|
}
|
|
// Smartcard daemon path exists and is a socket, enable it
|
|
cfg.SmartCardDaemonPath = path
|
|
}
|
|
|
|
func SetDataDir(ctx *cli.Context, cfg *node.Config) {
|
|
// begin PluGeth injection
|
|
pluginPath := pluginDefaultDataDir(node.DefaultDataDir())
|
|
switch {
|
|
case pluginPath != "" && ctx.String(DataDirFlag.Name) == node.DefaultDataDir():
|
|
cfg.DataDir = pluginPath
|
|
// end PluGeth injection
|
|
case ctx.IsSet(DataDirFlag.Name):
|
|
cfg.DataDir = ctx.String(DataDirFlag.Name)
|
|
case ctx.Bool(DeveloperFlag.Name):
|
|
cfg.DataDir = "" // unless explicitly requested, use memory databases
|
|
case ctx.Bool(GoerliFlag.Name) && cfg.DataDir == node.DefaultDataDir():
|
|
cfg.DataDir = filepath.Join(node.DefaultDataDir(), "goerli")
|
|
case ctx.Bool(SepoliaFlag.Name) && cfg.DataDir == node.DefaultDataDir():
|
|
cfg.DataDir = filepath.Join(node.DefaultDataDir(), "sepolia")
|
|
case ctx.Bool(HoleskyFlag.Name) && cfg.DataDir == node.DefaultDataDir():
|
|
cfg.DataDir = filepath.Join(node.DefaultDataDir(), "holesky")
|
|
}
|
|
}
|
|
|
|
func setGPO(ctx *cli.Context, cfg *gasprice.Config) {
|
|
if ctx.IsSet(GpoBlocksFlag.Name) {
|
|
cfg.Blocks = ctx.Int(GpoBlocksFlag.Name)
|
|
}
|
|
if ctx.IsSet(GpoPercentileFlag.Name) {
|
|
cfg.Percentile = ctx.Int(GpoPercentileFlag.Name)
|
|
}
|
|
if ctx.IsSet(GpoMaxGasPriceFlag.Name) {
|
|
cfg.MaxPrice = big.NewInt(ctx.Int64(GpoMaxGasPriceFlag.Name))
|
|
}
|
|
if ctx.IsSet(GpoIgnoreGasPriceFlag.Name) {
|
|
cfg.IgnorePrice = big.NewInt(ctx.Int64(GpoIgnoreGasPriceFlag.Name))
|
|
}
|
|
}
|
|
|
|
func setTxPool(ctx *cli.Context, cfg *legacypool.Config) {
|
|
if ctx.IsSet(TxPoolLocalsFlag.Name) {
|
|
locals := strings.Split(ctx.String(TxPoolLocalsFlag.Name), ",")
|
|
for _, account := range locals {
|
|
if trimmed := strings.TrimSpace(account); !common.IsHexAddress(trimmed) {
|
|
Fatalf("Invalid account in --txpool.locals: %s", trimmed)
|
|
} else {
|
|
cfg.Locals = append(cfg.Locals, common.HexToAddress(account))
|
|
}
|
|
}
|
|
}
|
|
if ctx.IsSet(TxPoolNoLocalsFlag.Name) {
|
|
cfg.NoLocals = ctx.Bool(TxPoolNoLocalsFlag.Name)
|
|
}
|
|
if ctx.IsSet(TxPoolJournalFlag.Name) {
|
|
cfg.Journal = ctx.String(TxPoolJournalFlag.Name)
|
|
}
|
|
if ctx.IsSet(TxPoolRejournalFlag.Name) {
|
|
cfg.Rejournal = ctx.Duration(TxPoolRejournalFlag.Name)
|
|
}
|
|
if ctx.IsSet(TxPoolPriceLimitFlag.Name) {
|
|
cfg.PriceLimit = ctx.Uint64(TxPoolPriceLimitFlag.Name)
|
|
}
|
|
if ctx.IsSet(TxPoolPriceBumpFlag.Name) {
|
|
cfg.PriceBump = ctx.Uint64(TxPoolPriceBumpFlag.Name)
|
|
}
|
|
if ctx.IsSet(TxPoolAccountSlotsFlag.Name) {
|
|
cfg.AccountSlots = ctx.Uint64(TxPoolAccountSlotsFlag.Name)
|
|
}
|
|
if ctx.IsSet(TxPoolGlobalSlotsFlag.Name) {
|
|
cfg.GlobalSlots = ctx.Uint64(TxPoolGlobalSlotsFlag.Name)
|
|
}
|
|
if ctx.IsSet(TxPoolAccountQueueFlag.Name) {
|
|
cfg.AccountQueue = ctx.Uint64(TxPoolAccountQueueFlag.Name)
|
|
}
|
|
if ctx.IsSet(TxPoolGlobalQueueFlag.Name) {
|
|
cfg.GlobalQueue = ctx.Uint64(TxPoolGlobalQueueFlag.Name)
|
|
}
|
|
if ctx.IsSet(TxPoolLifetimeFlag.Name) {
|
|
cfg.Lifetime = ctx.Duration(TxPoolLifetimeFlag.Name)
|
|
}
|
|
}
|
|
|
|
func setMiner(ctx *cli.Context, cfg *miner.Config) {
|
|
if ctx.IsSet(MinerExtraDataFlag.Name) {
|
|
cfg.ExtraData = []byte(ctx.String(MinerExtraDataFlag.Name))
|
|
}
|
|
if ctx.IsSet(MinerGasLimitFlag.Name) {
|
|
cfg.GasCeil = ctx.Uint64(MinerGasLimitFlag.Name)
|
|
}
|
|
if ctx.IsSet(MinerGasPriceFlag.Name) {
|
|
cfg.GasPrice = flags.GlobalBig(ctx, MinerGasPriceFlag.Name)
|
|
}
|
|
if ctx.IsSet(MinerRecommitIntervalFlag.Name) {
|
|
cfg.Recommit = ctx.Duration(MinerRecommitIntervalFlag.Name)
|
|
}
|
|
if ctx.IsSet(MinerNewPayloadTimeout.Name) {
|
|
cfg.NewPayloadTimeout = ctx.Duration(MinerNewPayloadTimeout.Name)
|
|
}
|
|
}
|
|
|
|
func setRequiredBlocks(ctx *cli.Context, cfg *ethconfig.Config) {
|
|
requiredBlocks := ctx.String(EthRequiredBlocksFlag.Name)
|
|
if requiredBlocks == "" {
|
|
if ctx.IsSet(LegacyWhitelistFlag.Name) {
|
|
log.Warn("The flag --whitelist is deprecated and will be removed, please use --eth.requiredblocks")
|
|
requiredBlocks = ctx.String(LegacyWhitelistFlag.Name)
|
|
} else {
|
|
return
|
|
}
|
|
}
|
|
cfg.RequiredBlocks = make(map[uint64]common.Hash)
|
|
for _, entry := range strings.Split(requiredBlocks, ",") {
|
|
parts := strings.Split(entry, "=")
|
|
if len(parts) != 2 {
|
|
Fatalf("Invalid required block entry: %s", entry)
|
|
}
|
|
number, err := strconv.ParseUint(parts[0], 0, 64)
|
|
if err != nil {
|
|
Fatalf("Invalid required block number %s: %v", parts[0], err)
|
|
}
|
|
var hash common.Hash
|
|
if err = hash.UnmarshalText([]byte(parts[1])); err != nil {
|
|
Fatalf("Invalid required block hash %s: %v", parts[1], err)
|
|
}
|
|
cfg.RequiredBlocks[number] = hash
|
|
}
|
|
}
|
|
|
|
// CheckExclusive verifies that only a single instance of the provided flags was
|
|
// set by the user. Each flag might optionally be followed by a string type to
|
|
// specialize it further.
|
|
func CheckExclusive(ctx *cli.Context, args ...interface{}) {
|
|
set := make([]string, 0, 1)
|
|
for i := 0; i < len(args); i++ {
|
|
// Make sure the next argument is a flag and skip if not set
|
|
flag, ok := args[i].(cli.Flag)
|
|
if !ok {
|
|
panic(fmt.Sprintf("invalid argument, not cli.Flag type: %T", args[i]))
|
|
}
|
|
// Check if next arg extends current and expand its name if so
|
|
name := flag.Names()[0]
|
|
|
|
if i+1 < len(args) {
|
|
switch option := args[i+1].(type) {
|
|
case string:
|
|
// Extended flag check, make sure value set doesn't conflict with passed in option
|
|
if ctx.String(flag.Names()[0]) == option {
|
|
name += "=" + option
|
|
set = append(set, "--"+name)
|
|
}
|
|
// shift arguments and continue
|
|
i++
|
|
continue
|
|
|
|
case cli.Flag:
|
|
default:
|
|
panic(fmt.Sprintf("invalid argument, not cli.Flag or string extension: %T", args[i+1]))
|
|
}
|
|
}
|
|
// Mark the flag if it's set
|
|
if ctx.IsSet(flag.Names()[0]) {
|
|
set = append(set, "--"+name)
|
|
}
|
|
}
|
|
if len(set) > 1 {
|
|
Fatalf("Flags %v can't be used at the same time", strings.Join(set, ", "))
|
|
}
|
|
}
|
|
|
|
// SetEthConfig applies eth-related command line flags to the config.
|
|
func SetEthConfig(ctx *cli.Context, stack *node.Node, cfg *ethconfig.Config) {
|
|
// Avoid conflicting network flags
|
|
CheckExclusive(ctx, MainnetFlag, DeveloperFlag, GoerliFlag, SepoliaFlag, HoleskyFlag)
|
|
CheckExclusive(ctx, DeveloperFlag, ExternalSignerFlag) // Can't use both ephemeral unlocked and external signer
|
|
|
|
// Set configurations from CLI flags
|
|
setEtherbase(ctx, cfg)
|
|
setGPO(ctx, &cfg.GPO)
|
|
setTxPool(ctx, &cfg.TxPool)
|
|
setMiner(ctx, &cfg.Miner)
|
|
setRequiredBlocks(ctx, cfg)
|
|
setLes(ctx, cfg)
|
|
|
|
// beginPluGethInjection
|
|
if pluginNetworkId := pluginNetworkId(); pluginNetworkId != nil {
|
|
cfg.NetworkId = *pluginNetworkId
|
|
}
|
|
if cfg.EthDiscoveryURLs == nil {
|
|
var lightMode bool
|
|
if cfg.SyncMode == downloader.LightSync {
|
|
lightMode = true
|
|
}
|
|
cfg.EthDiscoveryURLs = pluginETHDiscoveryURLs(lightMode)
|
|
cfg.SnapDiscoveryURLs = pluginSnapDiscoveryURLs()
|
|
}
|
|
//end PluGeth injection
|
|
|
|
// Cap the cache allowance and tune the garbage collector
|
|
mem, err := gopsutil.VirtualMemory()
|
|
if err == nil {
|
|
if 32<<(^uintptr(0)>>63) == 32 && mem.Total > 2*1024*1024*1024 {
|
|
log.Warn("Lowering memory allowance on 32bit arch", "available", mem.Total/1024/1024, "addressable", 2*1024)
|
|
mem.Total = 2 * 1024 * 1024 * 1024
|
|
}
|
|
allowance := int(mem.Total / 1024 / 1024 / 3)
|
|
if cache := ctx.Int(CacheFlag.Name); cache > allowance {
|
|
log.Warn("Sanitizing cache to Go's GC limits", "provided", cache, "updated", allowance)
|
|
ctx.Set(CacheFlag.Name, strconv.Itoa(allowance))
|
|
}
|
|
}
|
|
// Ensure Go's GC ignores the database cache for trigger percentage
|
|
cache := ctx.Int(CacheFlag.Name)
|
|
gogc := math.Max(20, math.Min(100, 100/(float64(cache)/1024)))
|
|
|
|
log.Debug("Sanitizing Go's GC trigger", "percent", int(gogc))
|
|
godebug.SetGCPercent(int(gogc))
|
|
|
|
if ctx.IsSet(SyncTargetFlag.Name) {
|
|
cfg.SyncMode = downloader.FullSync // dev sync target forces full sync
|
|
} else if ctx.IsSet(SyncModeFlag.Name) {
|
|
cfg.SyncMode = *flags.GlobalTextMarshaler(ctx, SyncModeFlag.Name).(*downloader.SyncMode)
|
|
}
|
|
if ctx.IsSet(NetworkIdFlag.Name) {
|
|
cfg.NetworkId = ctx.Uint64(NetworkIdFlag.Name)
|
|
}
|
|
if ctx.IsSet(CacheFlag.Name) || ctx.IsSet(CacheDatabaseFlag.Name) {
|
|
cfg.DatabaseCache = ctx.Int(CacheFlag.Name) * ctx.Int(CacheDatabaseFlag.Name) / 100
|
|
}
|
|
cfg.DatabaseHandles = MakeDatabaseHandles(ctx.Int(FDLimitFlag.Name))
|
|
if ctx.IsSet(AncientFlag.Name) {
|
|
cfg.DatabaseFreezer = ctx.String(AncientFlag.Name)
|
|
}
|
|
|
|
if gcmode := ctx.String(GCModeFlag.Name); gcmode != "full" && gcmode != "archive" {
|
|
Fatalf("--%s must be either 'full' or 'archive'", GCModeFlag.Name)
|
|
}
|
|
if ctx.IsSet(GCModeFlag.Name) {
|
|
cfg.NoPruning = ctx.String(GCModeFlag.Name) == "archive"
|
|
}
|
|
if ctx.IsSet(CacheNoPrefetchFlag.Name) {
|
|
cfg.NoPrefetch = ctx.Bool(CacheNoPrefetchFlag.Name)
|
|
}
|
|
// Read the value from the flag no matter if it's set or not.
|
|
cfg.Preimages = ctx.Bool(CachePreimagesFlag.Name)
|
|
if cfg.NoPruning && !cfg.Preimages {
|
|
cfg.Preimages = true
|
|
log.Info("Enabling recording of key preimages since archive mode is used")
|
|
}
|
|
if ctx.IsSet(StateHistoryFlag.Name) {
|
|
cfg.StateHistory = ctx.Uint64(StateHistoryFlag.Name)
|
|
}
|
|
if ctx.IsSet(StateSchemeFlag.Name) {
|
|
cfg.StateScheme = ctx.String(StateSchemeFlag.Name)
|
|
}
|
|
// Parse transaction history flag, if user is still using legacy config
|
|
// file with 'TxLookupLimit' configured, copy the value to 'TransactionHistory'.
|
|
if cfg.TransactionHistory == ethconfig.Defaults.TransactionHistory && cfg.TxLookupLimit != ethconfig.Defaults.TxLookupLimit {
|
|
log.Warn("The config option 'TxLookupLimit' is deprecated and will be removed, please use 'TransactionHistory'")
|
|
cfg.TransactionHistory = cfg.TxLookupLimit
|
|
}
|
|
if ctx.IsSet(TransactionHistoryFlag.Name) {
|
|
cfg.TransactionHistory = ctx.Uint64(TransactionHistoryFlag.Name)
|
|
} else if ctx.IsSet(TxLookupLimitFlag.Name) {
|
|
log.Warn("The flag --txlookuplimit is deprecated and will be removed, please use --history.transactions")
|
|
cfg.TransactionHistory = ctx.Uint64(TxLookupLimitFlag.Name)
|
|
}
|
|
if ctx.String(GCModeFlag.Name) == "archive" && cfg.TransactionHistory != 0 {
|
|
cfg.TransactionHistory = 0
|
|
log.Warn("Disabled transaction unindexing for archive node")
|
|
}
|
|
if ctx.IsSet(CacheFlag.Name) || ctx.IsSet(CacheTrieFlag.Name) {
|
|
cfg.TrieCleanCache = ctx.Int(CacheFlag.Name) * ctx.Int(CacheTrieFlag.Name) / 100
|
|
}
|
|
if ctx.IsSet(CacheFlag.Name) || ctx.IsSet(CacheGCFlag.Name) {
|
|
cfg.TrieDirtyCache = ctx.Int(CacheFlag.Name) * ctx.Int(CacheGCFlag.Name) / 100
|
|
}
|
|
if ctx.IsSet(CacheFlag.Name) || ctx.IsSet(CacheSnapshotFlag.Name) {
|
|
cfg.SnapshotCache = ctx.Int(CacheFlag.Name) * ctx.Int(CacheSnapshotFlag.Name) / 100
|
|
}
|
|
if ctx.IsSet(CacheLogSizeFlag.Name) {
|
|
cfg.FilterLogCacheSize = ctx.Int(CacheLogSizeFlag.Name)
|
|
}
|
|
if !ctx.Bool(SnapshotFlag.Name) || cfg.SnapshotCache == 0 {
|
|
// If snap-sync is requested, this flag is also required
|
|
if cfg.SyncMode == downloader.SnapSync {
|
|
if !ctx.Bool(SnapshotFlag.Name) {
|
|
log.Warn("Snap sync requested, enabling --snapshot")
|
|
}
|
|
if cfg.SnapshotCache == 0 {
|
|
log.Warn("Snap sync requested, resetting --cache.snapshot")
|
|
cfg.SnapshotCache = ctx.Int(CacheFlag.Name) * CacheSnapshotFlag.Value / 100
|
|
}
|
|
} else {
|
|
cfg.TrieCleanCache += cfg.SnapshotCache
|
|
cfg.SnapshotCache = 0 // Disabled
|
|
}
|
|
}
|
|
if ctx.IsSet(DocRootFlag.Name) {
|
|
cfg.DocRoot = ctx.String(DocRootFlag.Name)
|
|
}
|
|
if ctx.IsSet(VMEnableDebugFlag.Name) {
|
|
// TODO(fjl): force-enable this in --dev mode
|
|
cfg.EnablePreimageRecording = ctx.Bool(VMEnableDebugFlag.Name)
|
|
}
|
|
|
|
if ctx.IsSet(RPCGlobalGasCapFlag.Name) {
|
|
cfg.RPCGasCap = ctx.Uint64(RPCGlobalGasCapFlag.Name)
|
|
}
|
|
if cfg.RPCGasCap != 0 {
|
|
log.Info("Set global gas cap", "cap", cfg.RPCGasCap)
|
|
} else {
|
|
log.Info("Global gas cap disabled")
|
|
}
|
|
if ctx.IsSet(RPCGlobalEVMTimeoutFlag.Name) {
|
|
cfg.RPCEVMTimeout = ctx.Duration(RPCGlobalEVMTimeoutFlag.Name)
|
|
}
|
|
if ctx.IsSet(RPCGlobalTxFeeCapFlag.Name) {
|
|
cfg.RPCTxFeeCap = ctx.Float64(RPCGlobalTxFeeCapFlag.Name)
|
|
}
|
|
if ctx.IsSet(NoDiscoverFlag.Name) {
|
|
cfg.EthDiscoveryURLs, cfg.SnapDiscoveryURLs = []string{}, []string{}
|
|
} else if ctx.IsSet(DNSDiscoveryFlag.Name) {
|
|
urls := ctx.String(DNSDiscoveryFlag.Name)
|
|
if urls == "" {
|
|
cfg.EthDiscoveryURLs = []string{}
|
|
} else {
|
|
cfg.EthDiscoveryURLs = SplitAndTrim(urls)
|
|
}
|
|
}
|
|
// Override any default configs for hard coded networks.
|
|
switch {
|
|
case ctx.Bool(MainnetFlag.Name):
|
|
if !ctx.IsSet(NetworkIdFlag.Name) {
|
|
cfg.NetworkId = 1
|
|
}
|
|
cfg.Genesis = core.DefaultGenesisBlock()
|
|
SetDNSDiscoveryDefaults(cfg, params.MainnetGenesisHash)
|
|
case ctx.Bool(HoleskyFlag.Name):
|
|
if !ctx.IsSet(NetworkIdFlag.Name) {
|
|
cfg.NetworkId = 17000
|
|
}
|
|
cfg.Genesis = core.DefaultHoleskyGenesisBlock()
|
|
SetDNSDiscoveryDefaults(cfg, params.HoleskyGenesisHash)
|
|
case ctx.Bool(SepoliaFlag.Name):
|
|
if !ctx.IsSet(NetworkIdFlag.Name) {
|
|
cfg.NetworkId = 11155111
|
|
}
|
|
cfg.Genesis = core.DefaultSepoliaGenesisBlock()
|
|
SetDNSDiscoveryDefaults(cfg, params.SepoliaGenesisHash)
|
|
case ctx.Bool(GoerliFlag.Name):
|
|
if !ctx.IsSet(NetworkIdFlag.Name) {
|
|
cfg.NetworkId = 5
|
|
}
|
|
cfg.Genesis = core.DefaultGoerliGenesisBlock()
|
|
SetDNSDiscoveryDefaults(cfg, params.GoerliGenesisHash)
|
|
case ctx.Bool(DeveloperFlag.Name):
|
|
if !ctx.IsSet(NetworkIdFlag.Name) {
|
|
cfg.NetworkId = 1337
|
|
}
|
|
cfg.SyncMode = downloader.FullSync
|
|
// Create new developer account or reuse existing one
|
|
var (
|
|
developer accounts.Account
|
|
passphrase string
|
|
err error
|
|
)
|
|
if list := MakePasswordList(ctx); len(list) > 0 {
|
|
// Just take the first value. Although the function returns a possible multiple values and
|
|
// some usages iterate through them as attempts, that doesn't make sense in this setting,
|
|
// when we're definitely concerned with only one account.
|
|
passphrase = list[0]
|
|
}
|
|
|
|
// Unlock the developer account by local keystore.
|
|
var ks *keystore.KeyStore
|
|
if keystores := stack.AccountManager().Backends(keystore.KeyStoreType); len(keystores) > 0 {
|
|
ks = keystores[0].(*keystore.KeyStore)
|
|
}
|
|
if ks == nil {
|
|
Fatalf("Keystore is not available")
|
|
}
|
|
|
|
// Figure out the dev account address.
|
|
// setEtherbase has been called above, configuring the miner address from command line flags.
|
|
if cfg.Miner.Etherbase != (common.Address{}) {
|
|
developer = accounts.Account{Address: cfg.Miner.Etherbase}
|
|
} else if accs := ks.Accounts(); len(accs) > 0 {
|
|
developer = ks.Accounts()[0]
|
|
} else {
|
|
developer, err = ks.NewAccount(passphrase)
|
|
if err != nil {
|
|
Fatalf("Failed to create developer account: %v", err)
|
|
}
|
|
}
|
|
// Make sure the address is configured as fee recipient, otherwise
|
|
// the miner will fail to start.
|
|
cfg.Miner.Etherbase = developer.Address
|
|
|
|
if err := ks.Unlock(developer, passphrase); err != nil {
|
|
Fatalf("Failed to unlock developer account: %v", err)
|
|
}
|
|
log.Info("Using developer account", "address", developer.Address)
|
|
|
|
// Create a new developer genesis block or reuse existing one
|
|
cfg.Genesis = core.DeveloperGenesisBlock(ctx.Uint64(DeveloperGasLimitFlag.Name), &developer.Address)
|
|
if ctx.IsSet(DataDirFlag.Name) {
|
|
chaindb := tryMakeReadOnlyDatabase(ctx, stack)
|
|
if rawdb.ReadCanonicalHash(chaindb, 0) != (common.Hash{}) {
|
|
cfg.Genesis = nil // fallback to db content
|
|
|
|
//validate genesis has PoS enabled in block 0
|
|
genesis, err := core.ReadGenesis(chaindb)
|
|
if err != nil {
|
|
Fatalf("Could not read genesis from database: %v", err)
|
|
}
|
|
if !genesis.Config.TerminalTotalDifficultyPassed {
|
|
Fatalf("Bad developer-mode genesis configuration: terminalTotalDifficultyPassed must be true in developer mode")
|
|
}
|
|
if genesis.Config.TerminalTotalDifficulty == nil {
|
|
Fatalf("Bad developer-mode genesis configuration: terminalTotalDifficulty must be specified.")
|
|
}
|
|
if genesis.Difficulty.Cmp(genesis.Config.TerminalTotalDifficulty) != 1 {
|
|
Fatalf("Bad developer-mode genesis configuration: genesis block difficulty must be > terminalTotalDifficulty")
|
|
}
|
|
}
|
|
chaindb.Close()
|
|
}
|
|
if !ctx.IsSet(MinerGasPriceFlag.Name) {
|
|
cfg.Miner.GasPrice = big.NewInt(1)
|
|
}
|
|
default:
|
|
if cfg.NetworkId == 1 {
|
|
SetDNSDiscoveryDefaults(cfg, params.MainnetGenesisHash)
|
|
}
|
|
}
|
|
|
|
//begin plugeth injection
|
|
if genesis := pluginGenesisBlock(); genesis != nil {
|
|
chaindb := MakeChainDatabase(ctx, stack, false)
|
|
cfg.Genesis = genesis
|
|
rawdb.WriteChainConfig(chaindb, genesis.ToBlock().Hash(), genesis.Config)
|
|
chaindb.Close()
|
|
}
|
|
//end plugeth injection
|
|
|
|
// Set any dangling config values
|
|
if ctx.String(CryptoKZGFlag.Name) != "gokzg" && ctx.String(CryptoKZGFlag.Name) != "ckzg" {
|
|
Fatalf("--%s flag must be 'gokzg' or 'ckzg'", CryptoKZGFlag.Name)
|
|
}
|
|
log.Info("Initializing the KZG library", "backend", ctx.String(CryptoKZGFlag.Name))
|
|
if err := kzg4844.UseCKZG(ctx.String(CryptoKZGFlag.Name) == "ckzg"); err != nil {
|
|
Fatalf("Failed to set KZG library implementation to %s: %v", ctx.String(CryptoKZGFlag.Name), err)
|
|
}
|
|
}
|
|
|
|
// SetDNSDiscoveryDefaults configures DNS discovery with the given URL if
|
|
// no URLs are set.
|
|
func SetDNSDiscoveryDefaults(cfg *ethconfig.Config, genesis common.Hash) {
|
|
if cfg.EthDiscoveryURLs != nil {
|
|
return // already set through flags/config
|
|
}
|
|
protocol := "all"
|
|
if url := params.KnownDNSNetwork(genesis, protocol); url != "" {
|
|
cfg.EthDiscoveryURLs = []string{url}
|
|
cfg.SnapDiscoveryURLs = cfg.EthDiscoveryURLs
|
|
}
|
|
}
|
|
|
|
// RegisterEthService adds an Ethereum client to the stack.
|
|
// The second return value is the full node instance.
|
|
func RegisterEthService(stack *node.Node, cfg *ethconfig.Config) (ethapi.Backend, *eth.Ethereum) {
|
|
backend, err := eth.New(stack, cfg)
|
|
if err != nil {
|
|
Fatalf("Failed to register the Ethereum service: %v", err)
|
|
}
|
|
stack.RegisterAPIs(tracers.APIs(backend.APIBackend))
|
|
return backend.APIBackend, backend
|
|
}
|
|
|
|
// RegisterEthStatsService configures the Ethereum Stats daemon and adds it to the node.
|
|
func RegisterEthStatsService(stack *node.Node, backend ethapi.Backend, url string) {
|
|
if err := ethstats.New(stack, backend, backend.Engine(), url); err != nil {
|
|
Fatalf("Failed to register the Ethereum Stats service: %v", err)
|
|
}
|
|
}
|
|
|
|
// RegisterGraphQLService adds the GraphQL API to the node.
|
|
func RegisterGraphQLService(stack *node.Node, backend ethapi.Backend, filterSystem *filters.FilterSystem, cfg *node.Config) {
|
|
err := graphql.New(stack, backend, filterSystem, cfg.GraphQLCors, cfg.GraphQLVirtualHosts)
|
|
if err != nil {
|
|
Fatalf("Failed to register the GraphQL service: %v", err)
|
|
}
|
|
}
|
|
|
|
// RegisterFilterAPI adds the eth log filtering RPC API to the node.
|
|
func RegisterFilterAPI(stack *node.Node, backend ethapi.Backend, ethcfg *ethconfig.Config) *filters.FilterSystem {
|
|
filterSystem := filters.NewFilterSystem(backend, filters.Config{
|
|
LogCacheSize: ethcfg.FilterLogCacheSize,
|
|
})
|
|
stack.RegisterAPIs([]rpc.API{{
|
|
Namespace: "eth",
|
|
Service: filters.NewFilterAPI(filterSystem, false),
|
|
}})
|
|
return filterSystem
|
|
}
|
|
|
|
// RegisterFullSyncTester adds the full-sync tester service into node.
|
|
func RegisterFullSyncTester(stack *node.Node, eth *eth.Ethereum, target common.Hash) {
|
|
catalyst.RegisterFullSyncTester(stack, eth, target)
|
|
log.Info("Registered full-sync tester", "hash", target)
|
|
}
|
|
|
|
func SetupMetrics(ctx *cli.Context) {
|
|
if metrics.Enabled {
|
|
log.Info("Enabling metrics collection")
|
|
|
|
var (
|
|
enableExport = ctx.Bool(MetricsEnableInfluxDBFlag.Name)
|
|
enableExportV2 = ctx.Bool(MetricsEnableInfluxDBV2Flag.Name)
|
|
)
|
|
|
|
if enableExport || enableExportV2 {
|
|
CheckExclusive(ctx, MetricsEnableInfluxDBFlag, MetricsEnableInfluxDBV2Flag)
|
|
|
|
v1FlagIsSet := ctx.IsSet(MetricsInfluxDBUsernameFlag.Name) ||
|
|
ctx.IsSet(MetricsInfluxDBPasswordFlag.Name)
|
|
|
|
v2FlagIsSet := ctx.IsSet(MetricsInfluxDBTokenFlag.Name) ||
|
|
ctx.IsSet(MetricsInfluxDBOrganizationFlag.Name) ||
|
|
ctx.IsSet(MetricsInfluxDBBucketFlag.Name)
|
|
|
|
if enableExport && v2FlagIsSet {
|
|
Fatalf("Flags --influxdb.metrics.organization, --influxdb.metrics.token, --influxdb.metrics.bucket are only available for influxdb-v2")
|
|
} else if enableExportV2 && v1FlagIsSet {
|
|
Fatalf("Flags --influxdb.metrics.username, --influxdb.metrics.password are only available for influxdb-v1")
|
|
}
|
|
}
|
|
|
|
var (
|
|
endpoint = ctx.String(MetricsInfluxDBEndpointFlag.Name)
|
|
database = ctx.String(MetricsInfluxDBDatabaseFlag.Name)
|
|
username = ctx.String(MetricsInfluxDBUsernameFlag.Name)
|
|
password = ctx.String(MetricsInfluxDBPasswordFlag.Name)
|
|
|
|
token = ctx.String(MetricsInfluxDBTokenFlag.Name)
|
|
bucket = ctx.String(MetricsInfluxDBBucketFlag.Name)
|
|
organization = ctx.String(MetricsInfluxDBOrganizationFlag.Name)
|
|
)
|
|
|
|
if enableExport {
|
|
tagsMap := SplitTagsFlag(ctx.String(MetricsInfluxDBTagsFlag.Name))
|
|
|
|
log.Info("Enabling metrics export to InfluxDB")
|
|
|
|
go influxdb.InfluxDBWithTags(metrics.DefaultRegistry, 10*time.Second, endpoint, database, username, password, "geth.", tagsMap)
|
|
} else if enableExportV2 {
|
|
tagsMap := SplitTagsFlag(ctx.String(MetricsInfluxDBTagsFlag.Name))
|
|
|
|
log.Info("Enabling metrics export to InfluxDB (v2)")
|
|
|
|
go influxdb.InfluxDBV2WithTags(metrics.DefaultRegistry, 10*time.Second, endpoint, token, bucket, organization, "geth.", tagsMap)
|
|
}
|
|
|
|
if ctx.IsSet(MetricsHTTPFlag.Name) {
|
|
address := net.JoinHostPort(ctx.String(MetricsHTTPFlag.Name), fmt.Sprintf("%d", ctx.Int(MetricsPortFlag.Name)))
|
|
log.Info("Enabling stand-alone metrics HTTP endpoint", "address", address)
|
|
exp.Setup(address)
|
|
} else if ctx.IsSet(MetricsPortFlag.Name) {
|
|
log.Warn(fmt.Sprintf("--%s specified without --%s, metrics server will not start.", MetricsPortFlag.Name, MetricsHTTPFlag.Name))
|
|
}
|
|
}
|
|
}
|
|
|
|
func SplitTagsFlag(tagsFlag string) map[string]string {
|
|
tags := strings.Split(tagsFlag, ",")
|
|
tagsMap := map[string]string{}
|
|
|
|
for _, t := range tags {
|
|
if t != "" {
|
|
kv := strings.Split(t, "=")
|
|
|
|
if len(kv) == 2 {
|
|
tagsMap[kv[0]] = kv[1]
|
|
}
|
|
}
|
|
}
|
|
|
|
return tagsMap
|
|
}
|
|
|
|
// MakeChainDatabase opens a database using the flags passed to the client and will hard crash if it fails.
|
|
func MakeChainDatabase(ctx *cli.Context, stack *node.Node, readonly bool) ethdb.Database {
|
|
var (
|
|
cache = ctx.Int(CacheFlag.Name) * ctx.Int(CacheDatabaseFlag.Name) / 100
|
|
handles = MakeDatabaseHandles(ctx.Int(FDLimitFlag.Name))
|
|
err error
|
|
chainDb ethdb.Database
|
|
)
|
|
switch {
|
|
case ctx.IsSet(RemoteDBFlag.Name):
|
|
log.Info("Using remote db", "url", ctx.String(RemoteDBFlag.Name), "headers", len(ctx.StringSlice(HttpHeaderFlag.Name)))
|
|
client, err := DialRPCWithHeaders(ctx.String(RemoteDBFlag.Name), ctx.StringSlice(HttpHeaderFlag.Name))
|
|
if err != nil {
|
|
break
|
|
}
|
|
chainDb = remotedb.New(client)
|
|
case ctx.String(SyncModeFlag.Name) == "light":
|
|
chainDb, err = stack.OpenDatabase("lightchaindata", cache, handles, "", readonly)
|
|
default:
|
|
chainDb, err = stack.OpenDatabaseWithFreezer("chaindata", cache, handles, ctx.String(AncientFlag.Name), "", readonly)
|
|
}
|
|
if err != nil {
|
|
Fatalf("Could not open database: %v", err)
|
|
}
|
|
return chainDb
|
|
}
|
|
|
|
// tryMakeReadOnlyDatabase try to open the chain database in read-only mode,
|
|
// or fallback to write mode if the database is not initialized.
|
|
func tryMakeReadOnlyDatabase(ctx *cli.Context, stack *node.Node) ethdb.Database {
|
|
// If the database doesn't exist we need to open it in write-mode to allow
|
|
// the engine to create files.
|
|
readonly := true
|
|
if rawdb.PreexistingDatabase(stack.ResolvePath("chaindata")) == "" {
|
|
readonly = false
|
|
}
|
|
return MakeChainDatabase(ctx, stack, readonly)
|
|
}
|
|
|
|
func IsNetworkPreset(ctx *cli.Context) bool {
|
|
for _, flag := range NetworkFlags {
|
|
bFlag, _ := flag.(*cli.BoolFlag)
|
|
if ctx.IsSet(bFlag.Name) {
|
|
return true
|
|
}
|
|
}
|
|
return false
|
|
}
|
|
|
|
func DialRPCWithHeaders(endpoint string, headers []string) (*rpc.Client, error) {
|
|
if endpoint == "" {
|
|
return nil, errors.New("endpoint must be specified")
|
|
}
|
|
if strings.HasPrefix(endpoint, "rpc:") || strings.HasPrefix(endpoint, "ipc:") {
|
|
// Backwards compatibility with geth < 1.5 which required
|
|
// these prefixes.
|
|
endpoint = endpoint[4:]
|
|
}
|
|
var opts []rpc.ClientOption
|
|
if len(headers) > 0 {
|
|
customHeaders := make(http.Header)
|
|
for _, h := range headers {
|
|
kv := strings.Split(h, ":")
|
|
if len(kv) != 2 {
|
|
return nil, fmt.Errorf("invalid http header directive: %q", h)
|
|
}
|
|
customHeaders.Add(kv[0], kv[1])
|
|
}
|
|
opts = append(opts, rpc.WithHeaders(customHeaders))
|
|
}
|
|
return rpc.DialOptions(context.Background(), endpoint, opts...)
|
|
}
|
|
|
|
func MakeGenesis(ctx *cli.Context) *core.Genesis {
|
|
var genesis *core.Genesis
|
|
switch {
|
|
case ctx.Bool(MainnetFlag.Name):
|
|
genesis = core.DefaultGenesisBlock()
|
|
case ctx.Bool(HoleskyFlag.Name):
|
|
genesis = core.DefaultHoleskyGenesisBlock()
|
|
case ctx.Bool(SepoliaFlag.Name):
|
|
genesis = core.DefaultSepoliaGenesisBlock()
|
|
case ctx.Bool(GoerliFlag.Name):
|
|
genesis = core.DefaultGoerliGenesisBlock()
|
|
case ctx.Bool(DeveloperFlag.Name):
|
|
Fatalf("Developer chains are ephemeral")
|
|
}
|
|
//begin plugeth injection
|
|
if genesis == nil {
|
|
genesis = pluginGenesisBlock()
|
|
}
|
|
//end plugeth injection
|
|
return genesis
|
|
}
|
|
|
|
// MakeChain creates a chain manager from set command line flags.
|
|
func MakeChain(ctx *cli.Context, stack *node.Node, readonly bool) (*core.BlockChain, ethdb.Database) {
|
|
var (
|
|
gspec = MakeGenesis(ctx)
|
|
chainDb = MakeChainDatabase(ctx, stack, readonly)
|
|
)
|
|
config, err := core.LoadChainConfig(chainDb, gspec)
|
|
if err != nil {
|
|
Fatalf("%v", err)
|
|
}
|
|
engine, err := ethconfig.CreateConsensusEngine(config, chainDb)
|
|
if err != nil {
|
|
Fatalf("%v", err)
|
|
}
|
|
if gcmode := ctx.String(GCModeFlag.Name); gcmode != "full" && gcmode != "archive" {
|
|
Fatalf("--%s must be either 'full' or 'archive'", GCModeFlag.Name)
|
|
}
|
|
scheme, err := rawdb.ParseStateScheme(ctx.String(StateSchemeFlag.Name), chainDb)
|
|
if err != nil {
|
|
Fatalf("%v", err)
|
|
}
|
|
cache := &core.CacheConfig{
|
|
TrieCleanLimit: ethconfig.Defaults.TrieCleanCache,
|
|
TrieCleanNoPrefetch: ctx.Bool(CacheNoPrefetchFlag.Name),
|
|
TrieDirtyLimit: ethconfig.Defaults.TrieDirtyCache,
|
|
TrieDirtyDisabled: ctx.String(GCModeFlag.Name) == "archive",
|
|
TrieTimeLimit: ethconfig.Defaults.TrieTimeout,
|
|
SnapshotLimit: ethconfig.Defaults.SnapshotCache,
|
|
Preimages: ctx.Bool(CachePreimagesFlag.Name),
|
|
StateScheme: scheme,
|
|
StateHistory: ctx.Uint64(StateHistoryFlag.Name),
|
|
}
|
|
if cache.TrieDirtyDisabled && !cache.Preimages {
|
|
cache.Preimages = true
|
|
log.Info("Enabling recording of key preimages since archive mode is used")
|
|
}
|
|
if !ctx.Bool(SnapshotFlag.Name) {
|
|
cache.SnapshotLimit = 0 // Disabled
|
|
}
|
|
// If we're in readonly, do not bother generating snapshot data.
|
|
if readonly {
|
|
cache.SnapshotNoBuild = true
|
|
}
|
|
|
|
if ctx.IsSet(CacheFlag.Name) || ctx.IsSet(CacheTrieFlag.Name) {
|
|
cache.TrieCleanLimit = ctx.Int(CacheFlag.Name) * ctx.Int(CacheTrieFlag.Name) / 100
|
|
}
|
|
if ctx.IsSet(CacheFlag.Name) || ctx.IsSet(CacheGCFlag.Name) {
|
|
cache.TrieDirtyLimit = ctx.Int(CacheFlag.Name) * ctx.Int(CacheGCFlag.Name) / 100
|
|
}
|
|
vmcfg := vm.Config{EnablePreimageRecording: ctx.Bool(VMEnableDebugFlag.Name)}
|
|
|
|
// Disable transaction indexing/unindexing by default.
|
|
chain, err := core.NewBlockChain(chainDb, cache, gspec, nil, engine, vmcfg, nil, nil)
|
|
if err != nil {
|
|
Fatalf("Can't create BlockChain: %v", err)
|
|
}
|
|
return chain, chainDb
|
|
}
|
|
|
|
// MakeConsolePreloads retrieves the absolute paths for the console JavaScript
|
|
// scripts to preload before starting.
|
|
func MakeConsolePreloads(ctx *cli.Context) []string {
|
|
// Skip preloading if there's nothing to preload
|
|
if ctx.String(PreloadJSFlag.Name) == "" {
|
|
return nil
|
|
}
|
|
// Otherwise resolve absolute paths and return them
|
|
var preloads []string
|
|
|
|
for _, file := range strings.Split(ctx.String(PreloadJSFlag.Name), ",") {
|
|
preloads = append(preloads, strings.TrimSpace(file))
|
|
}
|
|
return preloads
|
|
}
|
|
|
|
// MakeTrieDatabase constructs a trie database based on the configured scheme.
|
|
func MakeTrieDatabase(ctx *cli.Context, disk ethdb.Database, preimage bool, readOnly bool, isVerkle bool) *triedb.Database {
|
|
config := &triedb.Config{
|
|
Preimages: preimage,
|
|
IsVerkle: isVerkle,
|
|
}
|
|
scheme, err := rawdb.ParseStateScheme(ctx.String(StateSchemeFlag.Name), disk)
|
|
if err != nil {
|
|
Fatalf("%v", err)
|
|
}
|
|
if scheme == rawdb.HashScheme {
|
|
// Read-only mode is not implemented in hash mode,
|
|
// ignore the parameter silently. TODO(rjl493456442)
|
|
// please config it if read mode is implemented.
|
|
config.HashDB = hashdb.Defaults
|
|
return triedb.NewDatabase(disk, config)
|
|
}
|
|
if readOnly {
|
|
config.PathDB = pathdb.ReadOnly
|
|
} else {
|
|
config.PathDB = pathdb.Defaults
|
|
}
|
|
return triedb.NewDatabase(disk, config)
|
|
}
|