1888 lines
50 KiB
Go
1888 lines
50 KiB
Go
// Code generated by github.com/filecoin-project/lotus/node/config/cfgdocgen. DO NOT EDIT.
|
|
|
|
package config
|
|
|
|
type DocField struct {
|
|
Name string
|
|
Type string
|
|
Comment string
|
|
}
|
|
|
|
var Doc = map[string][]DocField{
|
|
"API": {
|
|
{
|
|
Name: "ListenAddress",
|
|
Type: "string",
|
|
|
|
Comment: `Binding address for the Lotus API`,
|
|
},
|
|
{
|
|
Name: "RemoteListenAddress",
|
|
Type: "string",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "Timeout",
|
|
Type: "Duration",
|
|
|
|
Comment: ``,
|
|
},
|
|
},
|
|
"ApisConfig": {
|
|
{
|
|
Name: "ChainApiInfo",
|
|
Type: "[]string",
|
|
|
|
Comment: `ChainApiInfo is the API endpoint for the Lotus daemon.`,
|
|
},
|
|
{
|
|
Name: "StorageRPCSecret",
|
|
Type: "string",
|
|
|
|
Comment: `RPC Secret for the storage subsystem.
|
|
If integrating with lotus-miner this must match the value from
|
|
cat ~/.lotusminer/keystore/MF2XI2BNNJ3XILLQOJUXMYLUMU | jq -r .PrivateKey`,
|
|
},
|
|
},
|
|
"Backup": {
|
|
{
|
|
Name: "DisableMetadataLog",
|
|
Type: "bool",
|
|
|
|
Comment: `When set to true disables metadata log (.lotus/kvlog). This can save disk
|
|
space by reducing metadata redundancy.
|
|
|
|
Note that in case of metadata corruption it might be much harder to recover
|
|
your node if metadata log is disabled`,
|
|
},
|
|
},
|
|
"BatchFeeConfig": {
|
|
{
|
|
Name: "Base",
|
|
Type: "types.FIL",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "PerSector",
|
|
Type: "types.FIL",
|
|
|
|
Comment: ``,
|
|
},
|
|
},
|
|
"Chainstore": {
|
|
{
|
|
Name: "EnableSplitstore",
|
|
Type: "bool",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "Splitstore",
|
|
Type: "Splitstore",
|
|
|
|
Comment: ``,
|
|
},
|
|
},
|
|
"Client": {
|
|
{
|
|
Name: "UseIpfs",
|
|
Type: "bool",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "IpfsOnlineMode",
|
|
Type: "bool",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "IpfsMAddr",
|
|
Type: "string",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "IpfsUseForRetrieval",
|
|
Type: "bool",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "SimultaneousTransfersForStorage",
|
|
Type: "uint64",
|
|
|
|
Comment: `The maximum number of simultaneous data transfers between the client
|
|
and storage providers for storage deals`,
|
|
},
|
|
{
|
|
Name: "SimultaneousTransfersForRetrieval",
|
|
Type: "uint64",
|
|
|
|
Comment: `The maximum number of simultaneous data transfers between the client
|
|
and storage providers for retrieval deals`,
|
|
},
|
|
{
|
|
Name: "OffChainRetrieval",
|
|
Type: "bool",
|
|
|
|
Comment: `Require that retrievals perform no on-chain operations. Paid retrievals
|
|
without existing payment channels with available funds will fail instead
|
|
of automatically performing on-chain operations.`,
|
|
},
|
|
},
|
|
"Common": {
|
|
{
|
|
Name: "API",
|
|
Type: "API",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "Backup",
|
|
Type: "Backup",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "Logging",
|
|
Type: "Logging",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "Libp2p",
|
|
Type: "Libp2p",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "Pubsub",
|
|
Type: "Pubsub",
|
|
|
|
Comment: ``,
|
|
},
|
|
},
|
|
"CurioAddresses": {
|
|
{
|
|
Name: "PreCommitControl",
|
|
Type: "[]string",
|
|
|
|
Comment: `Addresses to send PreCommit messages from`,
|
|
},
|
|
{
|
|
Name: "CommitControl",
|
|
Type: "[]string",
|
|
|
|
Comment: `Addresses to send Commit messages from`,
|
|
},
|
|
{
|
|
Name: "TerminateControl",
|
|
Type: "[]string",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "DisableOwnerFallback",
|
|
Type: "bool",
|
|
|
|
Comment: `DisableOwnerFallback disables usage of the owner address for messages
|
|
sent automatically`,
|
|
},
|
|
{
|
|
Name: "DisableWorkerFallback",
|
|
Type: "bool",
|
|
|
|
Comment: `DisableWorkerFallback disables usage of the worker address for messages
|
|
sent automatically, if control addresses are configured.
|
|
A control address that doesn't have enough funds will still be chosen
|
|
over the worker address if this flag is set.`,
|
|
},
|
|
{
|
|
Name: "MinerAddresses",
|
|
Type: "[]string",
|
|
|
|
Comment: `MinerAddresses are the addresses of the miner actors to use for sending messages`,
|
|
},
|
|
},
|
|
"CurioConfig": {
|
|
{
|
|
Name: "Subsystems",
|
|
Type: "CurioSubsystemsConfig",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "Fees",
|
|
Type: "CurioFees",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "Addresses",
|
|
Type: "[]CurioAddresses",
|
|
|
|
Comment: `Addresses of wallets per MinerAddress (one of the fields).`,
|
|
},
|
|
{
|
|
Name: "Proving",
|
|
Type: "ProvingConfig",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "Journal",
|
|
Type: "JournalConfig",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "Apis",
|
|
Type: "ApisConfig",
|
|
|
|
Comment: ``,
|
|
},
|
|
},
|
|
"CurioFees": {
|
|
{
|
|
Name: "DefaultMaxFee",
|
|
Type: "types.FIL",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "MaxPreCommitGasFee",
|
|
Type: "types.FIL",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "MaxCommitGasFee",
|
|
Type: "types.FIL",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "MaxPreCommitBatchGasFee",
|
|
Type: "BatchFeeConfig",
|
|
|
|
Comment: `maxBatchFee = maxBase + maxPerSector * nSectors`,
|
|
},
|
|
{
|
|
Name: "MaxCommitBatchGasFee",
|
|
Type: "BatchFeeConfig",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "MaxTerminateGasFee",
|
|
Type: "types.FIL",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "MaxWindowPoStGasFee",
|
|
Type: "types.FIL",
|
|
|
|
Comment: `WindowPoSt is a high-value operation, so the default fee should be high.`,
|
|
},
|
|
{
|
|
Name: "MaxPublishDealsFee",
|
|
Type: "types.FIL",
|
|
|
|
Comment: ``,
|
|
},
|
|
},
|
|
"CurioSubsystemsConfig": {
|
|
{
|
|
Name: "EnableWindowPost",
|
|
Type: "bool",
|
|
|
|
Comment: `EnableWindowPost enables window post to be executed on this lotus-provider instance. Each machine in the cluster
|
|
with WindowPoSt enabled will also participate in the window post scheduler. It is possible to have multiple
|
|
machines with WindowPoSt enabled which will provide redundancy, and in case of multiple partitions per deadline,
|
|
will allow for parallel processing of partitions.
|
|
|
|
It is possible to have instances handling both WindowPoSt and WinningPoSt, which can provide redundancy without
|
|
the need for additional machines. In setups like this it is generally recommended to run
|
|
partitionsPerDeadline+1 machines.`,
|
|
},
|
|
{
|
|
Name: "WindowPostMaxTasks",
|
|
Type: "int",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "EnableWinningPost",
|
|
Type: "bool",
|
|
|
|
Comment: `EnableWinningPost enables winning post to be executed on this lotus-provider instance.
|
|
Each machine in the cluster with WinningPoSt enabled will also participate in the winning post scheduler.
|
|
It is possible to mix machines with WindowPoSt and WinningPoSt enabled, for details see the EnableWindowPost
|
|
documentation.`,
|
|
},
|
|
{
|
|
Name: "WinningPostMaxTasks",
|
|
Type: "int",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "EnableParkPiece",
|
|
Type: "bool",
|
|
|
|
Comment: `EnableParkPiece enables the "piece parking" task to run on this node. This task is responsible for fetching
|
|
pieces from the network and storing them in the storage subsystem until sectors are sealed. This task is
|
|
only applicable when integrating with boost, and should be enabled on nodes which will hold deal data
|
|
from boost until sectors containing the related pieces have the TreeD/TreeR constructed.
|
|
Note that future Curio implementations will have a separate task type for fetching pieces from the internet.`,
|
|
},
|
|
{
|
|
Name: "ParkPieceMaxTasks",
|
|
Type: "int",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "EnableSealSDR",
|
|
Type: "bool",
|
|
|
|
Comment: `EnableSealSDR enables SDR tasks to run. SDR is the long sequential computation
|
|
creating 11 layer files in sector cache directory.
|
|
|
|
SDR is the first task in the sealing pipeline. It's inputs are just the hash of the
|
|
unsealed data (CommD), sector number, miner id, and the seal proof type.
|
|
It's outputs are the 11 layer files in the sector cache directory.
|
|
|
|
In lotus-miner this was run as part of PreCommit1.`,
|
|
},
|
|
{
|
|
Name: "SealSDRMaxTasks",
|
|
Type: "int",
|
|
|
|
Comment: `The maximum amount of SDR tasks that can run simultaneously. Note that the maximum number of tasks will
|
|
also be bounded by resources available on the machine.`,
|
|
},
|
|
{
|
|
Name: "EnableSealSDRTrees",
|
|
Type: "bool",
|
|
|
|
Comment: `EnableSealSDRTrees enables the SDR pipeline tree-building task to run.
|
|
This task handles encoding of unsealed data into last sdr layer and building
|
|
of TreeR, TreeC and TreeD.
|
|
|
|
This task runs after SDR
|
|
TreeD is first computed with optional input of unsealed data
|
|
TreeR is computed from replica, which is first computed as field
|
|
addition of the last SDR layer and the bottom layer of TreeD (which is the unsealed data)
|
|
TreeC is computed from the 11 SDR layers
|
|
The 3 trees will later be used to compute the PoRep proof.
|
|
|
|
In case of SyntheticPoRep challenges for PoRep will be pre-generated at this step, and trees and layers
|
|
will be dropped. SyntheticPoRep works by pre-generating a very large set of challenges (~30GiB on disk)
|
|
then using a small subset of them for the actual PoRep computation. This allows for significant scratch space
|
|
saving between PreCommit and PoRep generation at the expense of more computation (generating challenges in this step)
|
|
|
|
In lotus-miner this was run as part of PreCommit2 (TreeD was run in PreCommit1).
|
|
Note that nodes with SDRTrees enabled will also answer to Finalize tasks,
|
|
which just remove unneeded tree data after PoRep is computed.`,
|
|
},
|
|
{
|
|
Name: "SealSDRTreesMaxTasks",
|
|
Type: "int",
|
|
|
|
Comment: `The maximum amount of SealSDRTrees tasks that can run simultaneously. Note that the maximum number of tasks will
|
|
also be bounded by resources available on the machine.`,
|
|
},
|
|
{
|
|
Name: "FinalizeMaxTasks",
|
|
Type: "int",
|
|
|
|
Comment: `FinalizeMaxTasks is the maximum amount of finalize tasks that can run simultaneously.
|
|
The finalize task is enabled on all machines which also handle SDRTrees tasks. Finalize ALWAYS runs on whichever
|
|
machine holds sector cache files, as it removes unneeded tree data after PoRep is computed.
|
|
Finalize will run in parallel with the SubmitCommitMsg task.`,
|
|
},
|
|
{
|
|
Name: "EnableSendPrecommitMsg",
|
|
Type: "bool",
|
|
|
|
Comment: `EnableSendPrecommitMsg enables the sending of precommit messages to the chain
|
|
from this lotus-provider instance.
|
|
This runs after SDRTrees and uses the output CommD / CommR (roots of TreeD / TreeR) for the message`,
|
|
},
|
|
{
|
|
Name: "EnablePoRepProof",
|
|
Type: "bool",
|
|
|
|
Comment: `EnablePoRepProof enables the computation of the porep proof
|
|
|
|
This task runs after interactive-porep seed becomes available, which happens 150 epochs (75min) after the
|
|
precommit message lands on chain. This task should run on a machine with a GPU. Vanilla PoRep proofs are
|
|
requested from the machine which holds sector cache files which most likely is the machine which ran the SDRTrees
|
|
task.
|
|
|
|
In lotus-miner this was Commit1 / Commit2`,
|
|
},
|
|
{
|
|
Name: "PoRepProofMaxTasks",
|
|
Type: "int",
|
|
|
|
Comment: `The maximum amount of PoRepProof tasks that can run simultaneously. Note that the maximum number of tasks will
|
|
also be bounded by resources available on the machine.`,
|
|
},
|
|
{
|
|
Name: "EnableSendCommitMsg",
|
|
Type: "bool",
|
|
|
|
Comment: `EnableSendCommitMsg enables the sending of commit messages to the chain
|
|
from this lotus-provider instance.`,
|
|
},
|
|
{
|
|
Name: "EnableMoveStorage",
|
|
Type: "bool",
|
|
|
|
Comment: `EnableMoveStorage enables the move-into-long-term-storage task to run on this lotus-provider instance.
|
|
This tasks should only be enabled on nodes with long-term storage.
|
|
|
|
The MoveStorage task is the last task in the sealing pipeline. It moves the sealed sector data from the
|
|
SDRTrees machine into long-term storage. This task runs after the Finalize task.`,
|
|
},
|
|
{
|
|
Name: "MoveStorageMaxTasks",
|
|
Type: "int",
|
|
|
|
Comment: `The maximum amount of MoveStorage tasks that can run simultaneously. Note that the maximum number of tasks will
|
|
also be bounded by resources available on the machine. It is recommended that this value is set to a number which
|
|
uses all available network (or disk) bandwidth on the machine without causing bottlenecks.`,
|
|
},
|
|
{
|
|
Name: "EnableWebGui",
|
|
Type: "bool",
|
|
|
|
Comment: `EnableWebGui enables the web GUI on this lotus-provider instance. The UI has minimal local overhead, but it should
|
|
only need to be run on a single machine in the cluster.`,
|
|
},
|
|
{
|
|
Name: "GuiAddress",
|
|
Type: "string",
|
|
|
|
Comment: `The address that should listen for Web GUI requests.`,
|
|
},
|
|
},
|
|
"DAGStoreConfig": {
|
|
{
|
|
Name: "RootDir",
|
|
Type: "string",
|
|
|
|
Comment: `Path to the dagstore root directory. This directory contains three
|
|
subdirectories, which can be symlinked to alternative locations if
|
|
need be:
|
|
- ./transients: caches unsealed deals that have been fetched from the
|
|
storage subsystem for serving retrievals.
|
|
- ./indices: stores shard indices.
|
|
- ./datastore: holds the KV store tracking the state of every shard
|
|
known to the DAG store.
|
|
Default value: <LOTUS_MARKETS_PATH>/dagstore (split deployment) or
|
|
<LOTUS_MINER_PATH>/dagstore (monolith deployment)`,
|
|
},
|
|
{
|
|
Name: "MaxConcurrentIndex",
|
|
Type: "int",
|
|
|
|
Comment: `The maximum amount of indexing jobs that can run simultaneously.
|
|
0 means unlimited.
|
|
Default value: 5.`,
|
|
},
|
|
{
|
|
Name: "MaxConcurrentReadyFetches",
|
|
Type: "int",
|
|
|
|
Comment: `The maximum amount of unsealed deals that can be fetched simultaneously
|
|
from the storage subsystem. 0 means unlimited.
|
|
Default value: 0 (unlimited).`,
|
|
},
|
|
{
|
|
Name: "MaxConcurrentUnseals",
|
|
Type: "int",
|
|
|
|
Comment: `The maximum amount of unseals that can be processed simultaneously
|
|
from the storage subsystem. 0 means unlimited.
|
|
Default value: 0 (unlimited).`,
|
|
},
|
|
{
|
|
Name: "MaxConcurrencyStorageCalls",
|
|
Type: "int",
|
|
|
|
Comment: `The maximum number of simultaneous inflight API calls to the storage
|
|
subsystem.
|
|
Default value: 100.`,
|
|
},
|
|
{
|
|
Name: "GCInterval",
|
|
Type: "Duration",
|
|
|
|
Comment: `The time between calls to periodic dagstore GC, in time.Duration string
|
|
representation, e.g. 1m, 5m, 1h.
|
|
Default value: 1 minute.`,
|
|
},
|
|
},
|
|
"DealmakingConfig": {
|
|
{
|
|
Name: "ConsiderOnlineStorageDeals",
|
|
Type: "bool",
|
|
|
|
Comment: `When enabled, the miner can accept online deals`,
|
|
},
|
|
{
|
|
Name: "ConsiderOfflineStorageDeals",
|
|
Type: "bool",
|
|
|
|
Comment: `When enabled, the miner can accept offline deals`,
|
|
},
|
|
{
|
|
Name: "ConsiderOnlineRetrievalDeals",
|
|
Type: "bool",
|
|
|
|
Comment: `When enabled, the miner can accept retrieval deals`,
|
|
},
|
|
{
|
|
Name: "ConsiderOfflineRetrievalDeals",
|
|
Type: "bool",
|
|
|
|
Comment: `When enabled, the miner can accept offline retrieval deals`,
|
|
},
|
|
{
|
|
Name: "ConsiderVerifiedStorageDeals",
|
|
Type: "bool",
|
|
|
|
Comment: `When enabled, the miner can accept verified deals`,
|
|
},
|
|
{
|
|
Name: "ConsiderUnverifiedStorageDeals",
|
|
Type: "bool",
|
|
|
|
Comment: `When enabled, the miner can accept unverified deals`,
|
|
},
|
|
{
|
|
Name: "PieceCidBlocklist",
|
|
Type: "[]cid.Cid",
|
|
|
|
Comment: `A list of Data CIDs to reject when making deals`,
|
|
},
|
|
{
|
|
Name: "ExpectedSealDuration",
|
|
Type: "Duration",
|
|
|
|
Comment: `Maximum expected amount of time getting the deal into a sealed sector will take
|
|
This includes the time the deal will need to get transferred and published
|
|
before being assigned to a sector`,
|
|
},
|
|
{
|
|
Name: "MaxDealStartDelay",
|
|
Type: "Duration",
|
|
|
|
Comment: `Maximum amount of time proposed deal StartEpoch can be in future`,
|
|
},
|
|
{
|
|
Name: "PublishMsgPeriod",
|
|
Type: "Duration",
|
|
|
|
Comment: `When a deal is ready to publish, the amount of time to wait for more
|
|
deals to be ready to publish before publishing them all as a batch`,
|
|
},
|
|
{
|
|
Name: "MaxDealsPerPublishMsg",
|
|
Type: "uint64",
|
|
|
|
Comment: `The maximum number of deals to include in a single PublishStorageDeals
|
|
message`,
|
|
},
|
|
{
|
|
Name: "MaxProviderCollateralMultiplier",
|
|
Type: "uint64",
|
|
|
|
Comment: `The maximum collateral that the provider will put up against a deal,
|
|
as a multiplier of the minimum collateral bound`,
|
|
},
|
|
{
|
|
Name: "MaxStagingDealsBytes",
|
|
Type: "int64",
|
|
|
|
Comment: `The maximum allowed disk usage size in bytes of staging deals not yet
|
|
passed to the sealing node by the markets service. 0 is unlimited.`,
|
|
},
|
|
{
|
|
Name: "SimultaneousTransfersForStorage",
|
|
Type: "uint64",
|
|
|
|
Comment: `The maximum number of parallel online data transfers for storage deals`,
|
|
},
|
|
{
|
|
Name: "SimultaneousTransfersForStoragePerClient",
|
|
Type: "uint64",
|
|
|
|
Comment: `The maximum number of simultaneous data transfers from any single client
|
|
for storage deals.
|
|
Unset by default (0), and values higher than SimultaneousTransfersForStorage
|
|
will have no effect; i.e. the total number of simultaneous data transfers
|
|
across all storage clients is bound by SimultaneousTransfersForStorage
|
|
regardless of this number.`,
|
|
},
|
|
{
|
|
Name: "SimultaneousTransfersForRetrieval",
|
|
Type: "uint64",
|
|
|
|
Comment: `The maximum number of parallel online data transfers for retrieval deals`,
|
|
},
|
|
{
|
|
Name: "StartEpochSealingBuffer",
|
|
Type: "uint64",
|
|
|
|
Comment: `Minimum start epoch buffer to give time for sealing of sector with deal.`,
|
|
},
|
|
{
|
|
Name: "Filter",
|
|
Type: "string",
|
|
|
|
Comment: `A command used for fine-grained evaluation of storage deals
|
|
see https://lotus.filecoin.io/storage-providers/advanced-configurations/market/#using-filters-for-fine-grained-storage-and-retrieval-deal-acceptance for more details`,
|
|
},
|
|
{
|
|
Name: "RetrievalFilter",
|
|
Type: "string",
|
|
|
|
Comment: `A command used for fine-grained evaluation of retrieval deals
|
|
see https://lotus.filecoin.io/storage-providers/advanced-configurations/market/#using-filters-for-fine-grained-storage-and-retrieval-deal-acceptance for more details`,
|
|
},
|
|
{
|
|
Name: "RetrievalPricing",
|
|
Type: "*RetrievalPricing",
|
|
|
|
Comment: ``,
|
|
},
|
|
},
|
|
"EventsConfig": {
|
|
{
|
|
Name: "DisableRealTimeFilterAPI",
|
|
Type: "bool",
|
|
|
|
Comment: `DisableRealTimeFilterAPI will disable the RealTimeFilterAPI that can create and query filters for actor events as they are emitted.
|
|
The API is enabled when Fevm.EnableEthRPC or EnableActorEventsAPI is true, but can be disabled selectively with this flag.`,
|
|
},
|
|
{
|
|
Name: "DisableHistoricFilterAPI",
|
|
Type: "bool",
|
|
|
|
Comment: `DisableHistoricFilterAPI will disable the HistoricFilterAPI that can create and query filters for actor events
|
|
that occurred in the past. HistoricFilterAPI maintains a queryable index of events.
|
|
The API is enabled when Fevm.EnableEthRPC or EnableActorEventsAPI is true, but can be disabled selectively with this flag.`,
|
|
},
|
|
{
|
|
Name: "EnableActorEventsAPI",
|
|
Type: "bool",
|
|
|
|
Comment: `EnableActorEventsAPI enables the Actor events API that enables clients to consume events
|
|
emitted by (smart contracts + built-in Actors).
|
|
This will also enable the RealTimeFilterAPI and HistoricFilterAPI by default, but they can be
|
|
disabled by setting their respective Disable* options.`,
|
|
},
|
|
{
|
|
Name: "FilterTTL",
|
|
Type: "Duration",
|
|
|
|
Comment: `FilterTTL specifies the time to live for actor event filters. Filters that haven't been accessed longer than
|
|
this time become eligible for automatic deletion.`,
|
|
},
|
|
{
|
|
Name: "MaxFilters",
|
|
Type: "int",
|
|
|
|
Comment: `MaxFilters specifies the maximum number of filters that may exist at any one time.`,
|
|
},
|
|
{
|
|
Name: "MaxFilterResults",
|
|
Type: "int",
|
|
|
|
Comment: `MaxFilterResults specifies the maximum number of results that can be accumulated by an actor event filter.`,
|
|
},
|
|
{
|
|
Name: "MaxFilterHeightRange",
|
|
Type: "uint64",
|
|
|
|
Comment: `MaxFilterHeightRange specifies the maximum range of heights that can be used in a filter (to avoid querying
|
|
the entire chain)`,
|
|
},
|
|
{
|
|
Name: "DatabasePath",
|
|
Type: "string",
|
|
|
|
Comment: `DatabasePath is the full path to a sqlite database that will be used to index actor events to
|
|
support the historic filter APIs. If the database does not exist it will be created. The directory containing
|
|
the database must already exist and be writeable. If a relative path is provided here, sqlite treats it as
|
|
relative to the CWD (current working directory).`,
|
|
},
|
|
},
|
|
"FaultReporterConfig": {
|
|
{
|
|
Name: "EnableConsensusFaultReporter",
|
|
Type: "bool",
|
|
|
|
Comment: `EnableConsensusFaultReporter controls whether the node will monitor and
|
|
report consensus faults. When enabled, the node will watch for malicious
|
|
behaviors like double-mining and parent grinding, and submit reports to the
|
|
network. This can earn reporter rewards, but is not guaranteed. Nodes should
|
|
enable fault reporting with care, as it may increase resource usage, and may
|
|
generate gas fees without earning rewards.`,
|
|
},
|
|
{
|
|
Name: "ConsensusFaultReporterDataDir",
|
|
Type: "string",
|
|
|
|
Comment: `ConsensusFaultReporterDataDir is the path where fault reporter state will be
|
|
persisted. This directory should have adequate space and permissions for the
|
|
node process.`,
|
|
},
|
|
{
|
|
Name: "ConsensusFaultReporterAddress",
|
|
Type: "string",
|
|
|
|
Comment: `ConsensusFaultReporterAddress is the wallet address used for submitting
|
|
ReportConsensusFault messages. It will pay for gas fees, and receive any
|
|
rewards. This address should have adequate funds to cover gas fees.`,
|
|
},
|
|
},
|
|
"FeeConfig": {
|
|
{
|
|
Name: "DefaultMaxFee",
|
|
Type: "types.FIL",
|
|
|
|
Comment: ``,
|
|
},
|
|
},
|
|
"FevmConfig": {
|
|
{
|
|
Name: "EnableEthRPC",
|
|
Type: "bool",
|
|
|
|
Comment: `EnableEthRPC enables eth_ rpc, and enables storing a mapping of eth transaction hashes to filecoin message Cids.
|
|
This will also enable the RealTimeFilterAPI and HistoricFilterAPI by default, but they can be disabled by config options above.`,
|
|
},
|
|
{
|
|
Name: "EthTxHashMappingLifetimeDays",
|
|
Type: "int",
|
|
|
|
Comment: `EthTxHashMappingLifetimeDays the transaction hash lookup database will delete mappings that have been stored for more than x days
|
|
Set to 0 to keep all mappings`,
|
|
},
|
|
{
|
|
Name: "Events",
|
|
Type: "DeprecatedEvents",
|
|
|
|
Comment: ``,
|
|
},
|
|
},
|
|
"FullNode": {
|
|
{
|
|
Name: "Client",
|
|
Type: "Client",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "Wallet",
|
|
Type: "Wallet",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "Fees",
|
|
Type: "FeeConfig",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "Chainstore",
|
|
Type: "Chainstore",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "Fevm",
|
|
Type: "FevmConfig",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "Events",
|
|
Type: "EventsConfig",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "Index",
|
|
Type: "IndexConfig",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "FaultReporter",
|
|
Type: "FaultReporterConfig",
|
|
|
|
Comment: ``,
|
|
},
|
|
},
|
|
"HarmonyDB": {
|
|
{
|
|
Name: "Hosts",
|
|
Type: "[]string",
|
|
|
|
Comment: `HOSTS is a list of hostnames to nodes running YugabyteDB
|
|
in a cluster. Only 1 is required`,
|
|
},
|
|
{
|
|
Name: "Username",
|
|
Type: "string",
|
|
|
|
Comment: `The Yugabyte server's username with full credentials to operate on Lotus' Database. Blank for default.`,
|
|
},
|
|
{
|
|
Name: "Password",
|
|
Type: "string",
|
|
|
|
Comment: `The password for the related username. Blank for default.`,
|
|
},
|
|
{
|
|
Name: "Database",
|
|
Type: "string",
|
|
|
|
Comment: `The database (logical partition) within Yugabyte. Blank for default.`,
|
|
},
|
|
{
|
|
Name: "Port",
|
|
Type: "string",
|
|
|
|
Comment: `The port to find Yugabyte. Blank for default.`,
|
|
},
|
|
},
|
|
"IndexConfig": {
|
|
{
|
|
Name: "EnableMsgIndex",
|
|
Type: "bool",
|
|
|
|
Comment: `EXPERIMENTAL FEATURE. USE WITH CAUTION
|
|
EnableMsgIndex enables indexing of messages on chain.`,
|
|
},
|
|
},
|
|
"IndexProviderConfig": {
|
|
{
|
|
Name: "Enable",
|
|
Type: "bool",
|
|
|
|
Comment: `Enable set whether to enable indexing announcement to the network and expose endpoints that
|
|
allow indexer nodes to process announcements. Enabled by default.`,
|
|
},
|
|
{
|
|
Name: "EntriesCacheCapacity",
|
|
Type: "int",
|
|
|
|
Comment: `EntriesCacheCapacity sets the maximum capacity to use for caching the indexing advertisement
|
|
entries. Defaults to 1024 if not specified. The cache is evicted using LRU policy. The
|
|
maximum storage used by the cache is a factor of EntriesCacheCapacity, EntriesChunkSize and
|
|
the length of multihashes being advertised. For example, advertising 128-bit long multihashes
|
|
with the default EntriesCacheCapacity, and EntriesChunkSize means the cache size can grow to
|
|
256MiB when full.`,
|
|
},
|
|
{
|
|
Name: "EntriesChunkSize",
|
|
Type: "int",
|
|
|
|
Comment: `EntriesChunkSize sets the maximum number of multihashes to include in a single entries chunk.
|
|
Defaults to 16384 if not specified. Note that chunks are chained together for indexing
|
|
advertisements that include more multihashes than the configured EntriesChunkSize.`,
|
|
},
|
|
{
|
|
Name: "TopicName",
|
|
Type: "string",
|
|
|
|
Comment: `TopicName sets the topic name on which the changes to the advertised content are announced.
|
|
If not explicitly specified, the topic name is automatically inferred from the network name
|
|
in following format: '/indexer/ingest/<network-name>'
|
|
Defaults to empty, which implies the topic name is inferred from network name.`,
|
|
},
|
|
{
|
|
Name: "PurgeCacheOnStart",
|
|
Type: "bool",
|
|
|
|
Comment: `PurgeCacheOnStart sets whether to clear any cached entries chunks when the provider engine
|
|
starts. By default, the cache is rehydrated from previously cached entries stored in
|
|
datastore if any is present.`,
|
|
},
|
|
},
|
|
"JournalConfig": {
|
|
{
|
|
Name: "DisabledEvents",
|
|
Type: "string",
|
|
|
|
Comment: `Events of the form: "system1:event1,system1:event2[,...]"`,
|
|
},
|
|
},
|
|
"Libp2p": {
|
|
{
|
|
Name: "ListenAddresses",
|
|
Type: "[]string",
|
|
|
|
Comment: `Binding address for the libp2p host - 0 means random port.
|
|
Format: multiaddress; see https://multiformats.io/multiaddr/`,
|
|
},
|
|
{
|
|
Name: "AnnounceAddresses",
|
|
Type: "[]string",
|
|
|
|
Comment: `Addresses to explicitally announce to other peers. If not specified,
|
|
all interface addresses are announced
|
|
Format: multiaddress`,
|
|
},
|
|
{
|
|
Name: "NoAnnounceAddresses",
|
|
Type: "[]string",
|
|
|
|
Comment: `Addresses to not announce
|
|
Format: multiaddress`,
|
|
},
|
|
{
|
|
Name: "BootstrapPeers",
|
|
Type: "[]string",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "ProtectedPeers",
|
|
Type: "[]string",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "DisableNatPortMap",
|
|
Type: "bool",
|
|
|
|
Comment: `When not disabled (default), lotus asks NAT devices (e.g., routers), to
|
|
open up an external port and forward it to the port lotus is running on.
|
|
When this works (i.e., when your router supports NAT port forwarding),
|
|
it makes the local lotus node accessible from the public internet`,
|
|
},
|
|
{
|
|
Name: "ConnMgrLow",
|
|
Type: "uint",
|
|
|
|
Comment: `ConnMgrLow is the number of connections that the basic connection manager
|
|
will trim down to.`,
|
|
},
|
|
{
|
|
Name: "ConnMgrHigh",
|
|
Type: "uint",
|
|
|
|
Comment: `ConnMgrHigh is the number of connections that, when exceeded, will trigger
|
|
a connection GC operation. Note: protected/recently formed connections don't
|
|
count towards this limit.`,
|
|
},
|
|
{
|
|
Name: "ConnMgrGrace",
|
|
Type: "Duration",
|
|
|
|
Comment: `ConnMgrGrace is a time duration that new connections are immune from being
|
|
closed by the connection manager.`,
|
|
},
|
|
},
|
|
"Logging": {
|
|
{
|
|
Name: "SubsystemLevels",
|
|
Type: "map[string]string",
|
|
|
|
Comment: `SubsystemLevels specify per-subsystem log levels`,
|
|
},
|
|
},
|
|
"MinerAddressConfig": {
|
|
{
|
|
Name: "PreCommitControl",
|
|
Type: "[]string",
|
|
|
|
Comment: `Addresses to send PreCommit messages from`,
|
|
},
|
|
{
|
|
Name: "CommitControl",
|
|
Type: "[]string",
|
|
|
|
Comment: `Addresses to send Commit messages from`,
|
|
},
|
|
{
|
|
Name: "TerminateControl",
|
|
Type: "[]string",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "DealPublishControl",
|
|
Type: "[]string",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "DisableOwnerFallback",
|
|
Type: "bool",
|
|
|
|
Comment: `DisableOwnerFallback disables usage of the owner address for messages
|
|
sent automatically`,
|
|
},
|
|
{
|
|
Name: "DisableWorkerFallback",
|
|
Type: "bool",
|
|
|
|
Comment: `DisableWorkerFallback disables usage of the worker address for messages
|
|
sent automatically, if control addresses are configured.
|
|
A control address that doesn't have enough funds will still be chosen
|
|
over the worker address if this flag is set.`,
|
|
},
|
|
},
|
|
"MinerFeeConfig": {
|
|
{
|
|
Name: "MaxPreCommitGasFee",
|
|
Type: "types.FIL",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "MaxCommitGasFee",
|
|
Type: "types.FIL",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "MaxPreCommitBatchGasFee",
|
|
Type: "BatchFeeConfig",
|
|
|
|
Comment: `maxBatchFee = maxBase + maxPerSector * nSectors`,
|
|
},
|
|
{
|
|
Name: "MaxCommitBatchGasFee",
|
|
Type: "BatchFeeConfig",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "MaxTerminateGasFee",
|
|
Type: "types.FIL",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "MaxWindowPoStGasFee",
|
|
Type: "types.FIL",
|
|
|
|
Comment: `WindowPoSt is a high-value operation, so the default fee should be high.`,
|
|
},
|
|
{
|
|
Name: "MaxPublishDealsFee",
|
|
Type: "types.FIL",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "MaxMarketBalanceAddFee",
|
|
Type: "types.FIL",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "MaximizeWindowPoStFeeCap",
|
|
Type: "bool",
|
|
|
|
Comment: ``,
|
|
},
|
|
},
|
|
"MinerSubsystemConfig": {
|
|
{
|
|
Name: "EnableMining",
|
|
Type: "bool",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "EnableSealing",
|
|
Type: "bool",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "EnableSectorStorage",
|
|
Type: "bool",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "EnableMarkets",
|
|
Type: "bool",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "EnableSectorIndexDB",
|
|
Type: "bool",
|
|
|
|
Comment: `When enabled, the sector index will reside in an external database
|
|
as opposed to the local KV store in the miner process
|
|
This is useful to allow workers to bypass the lotus miner to access sector information`,
|
|
},
|
|
{
|
|
Name: "SealerApiInfo",
|
|
Type: "string",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "SectorIndexApiInfo",
|
|
Type: "string",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "DisableWindowPoSt",
|
|
Type: "bool",
|
|
|
|
Comment: `When window post is enabled, the miner will automatically submit window post proofs
|
|
for all sectors that are eligible for window post
|
|
IF WINDOW POST IS DISABLED, THE MINER WILL NOT SUBMIT WINDOW POST PROOFS
|
|
THIS WILL RESULT IN FAULTS AND PENALTIES IF NO OTHER MECHANISM IS RUNNING
|
|
TO SUBMIT WINDOW POST PROOFS.
|
|
Note: This option is entirely disabling the window post scheduler,
|
|
not just the builtin PoSt computation like Proving.DisableBuiltinWindowPoSt.
|
|
This option will stop lotus-miner from performing any actions related
|
|
to window post, including scheduling, submitting proofs, and recovering
|
|
sectors.`,
|
|
},
|
|
{
|
|
Name: "DisableWinningPoSt",
|
|
Type: "bool",
|
|
|
|
Comment: `When winning post is disabled, the miner process will NOT attempt to mine
|
|
blocks. This should only be set when there's an external process mining
|
|
blocks on behalf of the miner.
|
|
When disabled and no external block producers are configured, all potential
|
|
block rewards will be missed!`,
|
|
},
|
|
},
|
|
"ProvingConfig": {
|
|
{
|
|
Name: "ParallelCheckLimit",
|
|
Type: "int",
|
|
|
|
Comment: `Maximum number of sector checks to run in parallel. (0 = unlimited)
|
|
|
|
WARNING: Setting this value too high may make the node crash by running out of stack
|
|
WARNING: Setting this value too low may make sector challenge reading much slower, resulting in failed PoSt due
|
|
to late submission.
|
|
|
|
After changing this option, confirm that the new value works in your setup by invoking
|
|
'lotus-miner proving compute window-post 0'`,
|
|
},
|
|
{
|
|
Name: "SingleCheckTimeout",
|
|
Type: "Duration",
|
|
|
|
Comment: `Maximum amount of time a proving pre-check can take for a sector. If the check times out the sector will be skipped
|
|
|
|
WARNING: Setting this value too low risks in sectors being skipped even though they are accessible, just reading the
|
|
test challenge took longer than this timeout
|
|
WARNING: Setting this value too high risks missing PoSt deadline in case IO operations related to this sector are
|
|
blocked (e.g. in case of disconnected NFS mount)`,
|
|
},
|
|
{
|
|
Name: "PartitionCheckTimeout",
|
|
Type: "Duration",
|
|
|
|
Comment: `Maximum amount of time a proving pre-check can take for an entire partition. If the check times out, sectors in
|
|
the partition which didn't get checked on time will be skipped
|
|
|
|
WARNING: Setting this value too low risks in sectors being skipped even though they are accessible, just reading the
|
|
test challenge took longer than this timeout
|
|
WARNING: Setting this value too high risks missing PoSt deadline in case IO operations related to this partition are
|
|
blocked or slow`,
|
|
},
|
|
{
|
|
Name: "DisableBuiltinWindowPoSt",
|
|
Type: "bool",
|
|
|
|
Comment: `Disable Window PoSt computation on the lotus-miner process even if no window PoSt workers are present.
|
|
|
|
WARNING: If no windowPoSt workers are connected, window PoSt WILL FAIL resulting in faulty sectors which will need
|
|
to be recovered. Before enabling this option, make sure your PoSt workers work correctly.
|
|
|
|
After changing this option, confirm that the new value works in your setup by invoking
|
|
'lotus-miner proving compute window-post 0'`,
|
|
},
|
|
{
|
|
Name: "DisableBuiltinWinningPoSt",
|
|
Type: "bool",
|
|
|
|
Comment: `Disable Winning PoSt computation on the lotus-miner process even if no winning PoSt workers are present.
|
|
|
|
WARNING: If no WinningPoSt workers are connected, Winning PoSt WILL FAIL resulting in lost block rewards.
|
|
Before enabling this option, make sure your PoSt workers work correctly.`,
|
|
},
|
|
{
|
|
Name: "DisableWDPoStPreChecks",
|
|
Type: "bool",
|
|
|
|
Comment: `Disable WindowPoSt provable sector readability checks.
|
|
|
|
In normal operation, when preparing to compute WindowPoSt, lotus-miner will perform a round of reading challenges
|
|
from all sectors to confirm that those sectors can be proven. Challenges read in this process are discarded, as
|
|
we're only interested in checking that sector data can be read.
|
|
|
|
When using builtin proof computation (no PoSt workers, and DisableBuiltinWindowPoSt is set to false), this process
|
|
can save a lot of time and compute resources in the case that some sectors are not readable - this is caused by
|
|
the builtin logic not skipping snark computation when some sectors need to be skipped.
|
|
|
|
When using PoSt workers, this process is mostly redundant, with PoSt workers challenges will be read once, and
|
|
if challenges for some sectors aren't readable, those sectors will just get skipped.
|
|
|
|
Disabling sector pre-checks will slightly reduce IO load when proving sectors, possibly resulting in shorter
|
|
time to produce window PoSt. In setups with good IO capabilities the effect of this option on proving time should
|
|
be negligible.
|
|
|
|
NOTE: It likely is a bad idea to disable sector pre-checks in setups with no PoSt workers.
|
|
|
|
NOTE: Even when this option is enabled, recovering sectors will be checked before recovery declaration message is
|
|
sent to the chain
|
|
|
|
After changing this option, confirm that the new value works in your setup by invoking
|
|
'lotus-miner proving compute window-post 0'`,
|
|
},
|
|
{
|
|
Name: "MaxPartitionsPerPoStMessage",
|
|
Type: "int",
|
|
|
|
Comment: `Maximum number of partitions to prove in a single SubmitWindowPoSt messace. 0 = network limit (3 in nv21)
|
|
|
|
A single partition may contain up to 2349 32GiB sectors, or 2300 64GiB sectors.
|
|
//
|
|
Note that setting this value lower may result in less efficient gas use - more messages will be sent,
|
|
to prove each deadline, resulting in more total gas use (but each message will have lower gas limit)
|
|
|
|
Setting this value above the network limit has no effect`,
|
|
},
|
|
{
|
|
Name: "MaxPartitionsPerRecoveryMessage",
|
|
Type: "int",
|
|
|
|
Comment: `In some cases when submitting DeclareFaultsRecovered messages,
|
|
there may be too many recoveries to fit in a BlockGasLimit.
|
|
In those cases it may be necessary to set this value to something low (eg 1);
|
|
Note that setting this value lower may result in less efficient gas use - more messages will be sent than needed,
|
|
resulting in more total gas use (but each message will have lower gas limit)`,
|
|
},
|
|
{
|
|
Name: "SingleRecoveringPartitionPerPostMessage",
|
|
Type: "bool",
|
|
|
|
Comment: `Enable single partition per PoSt Message for partitions containing recovery sectors
|
|
|
|
In cases when submitting PoSt messages which contain recovering sectors, the default network limit may still be
|
|
too high to fit in the block gas limit. In those cases, it becomes useful to only house the single partition
|
|
with recovering sectors in the post message
|
|
|
|
Note that setting this value lower may result in less efficient gas use - more messages will be sent,
|
|
to prove each deadline, resulting in more total gas use (but each message will have lower gas limit)`,
|
|
},
|
|
},
|
|
"Pubsub": {
|
|
{
|
|
Name: "Bootstrapper",
|
|
Type: "bool",
|
|
|
|
Comment: `Run the node in bootstrap-node mode`,
|
|
},
|
|
{
|
|
Name: "DirectPeers",
|
|
Type: "[]string",
|
|
|
|
Comment: `DirectPeers specifies peers with direct peering agreements. These peers are
|
|
connected outside of the mesh, with all (valid) message unconditionally
|
|
forwarded to them. The router will maintain open connections to these peers.
|
|
Note that the peering agreement should be reciprocal with direct peers
|
|
symmetrically configured at both ends.
|
|
Type: Array of multiaddress peerinfo strings, must include peerid (/p2p/12D3K...`,
|
|
},
|
|
{
|
|
Name: "IPColocationWhitelist",
|
|
Type: "[]string",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "RemoteTracer",
|
|
Type: "string",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "JsonTracer",
|
|
Type: "string",
|
|
|
|
Comment: `Path to file that will be used to output tracer content in JSON format.
|
|
If present tracer will save data to defined file.
|
|
Format: file path`,
|
|
},
|
|
{
|
|
Name: "ElasticSearchTracer",
|
|
Type: "string",
|
|
|
|
Comment: `Connection string for elasticsearch instance.
|
|
If present tracer will save data to elasticsearch.
|
|
Format: https://<username>:<password>@<elasticsearch_url>:<port>/`,
|
|
},
|
|
{
|
|
Name: "ElasticSearchIndex",
|
|
Type: "string",
|
|
|
|
Comment: `Name of elasticsearch index that will be used to save tracer data.
|
|
This property is used only if ElasticSearchTracer propery is set.`,
|
|
},
|
|
{
|
|
Name: "TracerSourceAuth",
|
|
Type: "string",
|
|
|
|
Comment: `Auth token that will be passed with logs to elasticsearch - used for weighted peers score.`,
|
|
},
|
|
},
|
|
"RetrievalPricing": {
|
|
{
|
|
Name: "Strategy",
|
|
Type: "string",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "Default",
|
|
Type: "*RetrievalPricingDefault",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "External",
|
|
Type: "*RetrievalPricingExternal",
|
|
|
|
Comment: ``,
|
|
},
|
|
},
|
|
"RetrievalPricingDefault": {
|
|
{
|
|
Name: "VerifiedDealsFreeTransfer",
|
|
Type: "bool",
|
|
|
|
Comment: `VerifiedDealsFreeTransfer configures zero fees for data transfer for a retrieval deal
|
|
of a payloadCid that belongs to a verified storage deal.
|
|
This parameter is ONLY applicable if the retrieval pricing policy strategy has been configured to "default".
|
|
default value is true`,
|
|
},
|
|
},
|
|
"RetrievalPricingExternal": {
|
|
{
|
|
Name: "Path",
|
|
Type: "string",
|
|
|
|
Comment: `Path of the external script that will be run to price a retrieval deal.
|
|
This parameter is ONLY applicable if the retrieval pricing policy strategy has been configured to "external".`,
|
|
},
|
|
},
|
|
"SealerConfig": {
|
|
{
|
|
Name: "ParallelFetchLimit",
|
|
Type: "int",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "AllowSectorDownload",
|
|
Type: "bool",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "AllowAddPiece",
|
|
Type: "bool",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "AllowPreCommit1",
|
|
Type: "bool",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "AllowPreCommit2",
|
|
Type: "bool",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "AllowCommit",
|
|
Type: "bool",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "AllowUnseal",
|
|
Type: "bool",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "AllowReplicaUpdate",
|
|
Type: "bool",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "AllowProveReplicaUpdate2",
|
|
Type: "bool",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "AllowRegenSectorKey",
|
|
Type: "bool",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "LocalWorkerName",
|
|
Type: "string",
|
|
|
|
Comment: `LocalWorkerName specifies a custom name for the builtin worker.
|
|
If set to an empty string (default) os hostname will be used`,
|
|
},
|
|
{
|
|
Name: "Assigner",
|
|
Type: "string",
|
|
|
|
Comment: `Assigner specifies the worker assigner to use when scheduling tasks.
|
|
"utilization" (default) - assign tasks to workers with lowest utilization.
|
|
"spread" - assign tasks to as many distinct workers as possible.`,
|
|
},
|
|
{
|
|
Name: "DisallowRemoteFinalize",
|
|
Type: "bool",
|
|
|
|
Comment: `DisallowRemoteFinalize when set to true will force all Finalize tasks to
|
|
run on workers with local access to both long-term storage and the sealing
|
|
path containing the sector.
|
|
--
|
|
WARNING: Only set this if all workers have access to long-term storage
|
|
paths. If this flag is enabled, and there are workers without long-term
|
|
storage access, sectors will not be moved from them, and Finalize tasks
|
|
will appear to be stuck.
|
|
--
|
|
If you see stuck Finalize tasks after enabling this setting, check
|
|
'lotus-miner sealing sched-diag' and 'lotus-miner storage find [sector num]'`,
|
|
},
|
|
{
|
|
Name: "ResourceFiltering",
|
|
Type: "ResourceFilteringStrategy",
|
|
|
|
Comment: `ResourceFiltering instructs the system which resource filtering strategy
|
|
to use when evaluating tasks against this worker. An empty value defaults
|
|
to "hardware".`,
|
|
},
|
|
},
|
|
"SealingConfig": {
|
|
{
|
|
Name: "MaxWaitDealsSectors",
|
|
Type: "uint64",
|
|
|
|
Comment: `Upper bound on how many sectors can be waiting for more deals to be packed in it before it begins sealing at any given time.
|
|
If the miner is accepting multiple deals in parallel, up to MaxWaitDealsSectors of new sectors will be created.
|
|
If more than MaxWaitDealsSectors deals are accepted in parallel, only MaxWaitDealsSectors deals will be processed in parallel
|
|
Note that setting this number too high in relation to deal ingestion rate may result in poor sector packing efficiency
|
|
0 = no limit`,
|
|
},
|
|
{
|
|
Name: "MaxSealingSectors",
|
|
Type: "uint64",
|
|
|
|
Comment: `Upper bound on how many sectors can be sealing+upgrading at the same time when creating new CC sectors (0 = unlimited)`,
|
|
},
|
|
{
|
|
Name: "MaxSealingSectorsForDeals",
|
|
Type: "uint64",
|
|
|
|
Comment: `Upper bound on how many sectors can be sealing+upgrading at the same time when creating new sectors with deals (0 = unlimited)`,
|
|
},
|
|
{
|
|
Name: "PreferNewSectorsForDeals",
|
|
Type: "bool",
|
|
|
|
Comment: `Prefer creating new sectors even if there are sectors Available for upgrading.
|
|
This setting combined with MaxUpgradingSectors set to a value higher than MaxSealingSectorsForDeals makes it
|
|
possible to use fast sector upgrades to handle high volumes of storage deals, while still using the simple sealing
|
|
flow when the volume of storage deals is lower.`,
|
|
},
|
|
{
|
|
Name: "MaxUpgradingSectors",
|
|
Type: "uint64",
|
|
|
|
Comment: `Upper bound on how many sectors can be sealing+upgrading at the same time when upgrading CC sectors with deals (0 = MaxSealingSectorsForDeals)`,
|
|
},
|
|
{
|
|
Name: "MinUpgradeSectorExpiration",
|
|
Type: "uint64",
|
|
|
|
Comment: `When set to a non-zero value, minimum number of epochs until sector expiration required for sectors to be considered
|
|
for upgrades (0 = DealMinDuration = 180 days = 518400 epochs)
|
|
|
|
Note that if all deals waiting in the input queue have lifetimes longer than this value, upgrade sectors will be
|
|
required to have expiration of at least the soonest-ending deal`,
|
|
},
|
|
{
|
|
Name: "MinTargetUpgradeSectorExpiration",
|
|
Type: "uint64",
|
|
|
|
Comment: `DEPRECATED: Target expiration is no longer used`,
|
|
},
|
|
{
|
|
Name: "CommittedCapacitySectorLifetime",
|
|
Type: "Duration",
|
|
|
|
Comment: `CommittedCapacitySectorLifetime is the duration a Committed Capacity (CC) sector will
|
|
live before it must be extended or converted into sector containing deals before it is
|
|
terminated. Value must be between 180-1278 days (1278 in nv21, 540 before nv21).`,
|
|
},
|
|
{
|
|
Name: "WaitDealsDelay",
|
|
Type: "Duration",
|
|
|
|
Comment: `Period of time that a newly created sector will wait for more deals to be packed in to before it starts to seal.
|
|
Sectors which are fully filled will start sealing immediately`,
|
|
},
|
|
{
|
|
Name: "AlwaysKeepUnsealedCopy",
|
|
Type: "bool",
|
|
|
|
Comment: `Whether to keep unsealed copies of deal data regardless of whether the client requested that. This lets the miner
|
|
avoid the relatively high cost of unsealing the data later, at the cost of more storage space`,
|
|
},
|
|
{
|
|
Name: "FinalizeEarly",
|
|
Type: "bool",
|
|
|
|
Comment: `Run sector finalization before submitting sector proof to the chain`,
|
|
},
|
|
{
|
|
Name: "MakeNewSectorForDeals",
|
|
Type: "bool",
|
|
|
|
Comment: `Whether new sectors are created to pack incoming deals
|
|
When this is set to false no new sectors will be created for sealing incoming deals
|
|
This is useful for forcing all deals to be assigned as snap deals to sectors marked for upgrade`,
|
|
},
|
|
{
|
|
Name: "MakeCCSectorsAvailable",
|
|
Type: "bool",
|
|
|
|
Comment: `After sealing CC sectors, make them available for upgrading with deals`,
|
|
},
|
|
{
|
|
Name: "CollateralFromMinerBalance",
|
|
Type: "bool",
|
|
|
|
Comment: `Whether to use available miner balance for sector collateral instead of sending it with each message`,
|
|
},
|
|
{
|
|
Name: "AvailableBalanceBuffer",
|
|
Type: "types.FIL",
|
|
|
|
Comment: `Minimum available balance to keep in the miner actor before sending it with messages`,
|
|
},
|
|
{
|
|
Name: "DisableCollateralFallback",
|
|
Type: "bool",
|
|
|
|
Comment: `Don't send collateral with messages even if there is no available balance in the miner actor`,
|
|
},
|
|
{
|
|
Name: "MaxPreCommitBatch",
|
|
Type: "int",
|
|
|
|
Comment: `maximum precommit batch size - batches will be sent immediately above this size`,
|
|
},
|
|
{
|
|
Name: "PreCommitBatchWait",
|
|
Type: "Duration",
|
|
|
|
Comment: `how long to wait before submitting a batch after crossing the minimum batch size`,
|
|
},
|
|
{
|
|
Name: "PreCommitBatchSlack",
|
|
Type: "Duration",
|
|
|
|
Comment: `time buffer for forceful batch submission before sectors/deal in batch would start expiring`,
|
|
},
|
|
{
|
|
Name: "AggregateCommits",
|
|
Type: "bool",
|
|
|
|
Comment: `enable / disable commit aggregation (takes effect after nv13)`,
|
|
},
|
|
{
|
|
Name: "MinCommitBatch",
|
|
Type: "int",
|
|
|
|
Comment: `minimum batched commit size - batches above this size will eventually be sent on a timeout`,
|
|
},
|
|
{
|
|
Name: "MaxCommitBatch",
|
|
Type: "int",
|
|
|
|
Comment: `maximum batched commit size - batches will be sent immediately above this size`,
|
|
},
|
|
{
|
|
Name: "CommitBatchWait",
|
|
Type: "Duration",
|
|
|
|
Comment: `how long to wait before submitting a batch after crossing the minimum batch size`,
|
|
},
|
|
{
|
|
Name: "CommitBatchSlack",
|
|
Type: "Duration",
|
|
|
|
Comment: `time buffer for forceful batch submission before sectors/deals in batch would start expiring`,
|
|
},
|
|
{
|
|
Name: "BatchPreCommitAboveBaseFee",
|
|
Type: "types.FIL",
|
|
|
|
Comment: `network BaseFee below which to stop doing precommit batching, instead
|
|
sending precommit messages to the chain individually. When the basefee is
|
|
below this threshold, precommit messages will get sent out immediately.`,
|
|
},
|
|
{
|
|
Name: "AggregateAboveBaseFee",
|
|
Type: "types.FIL",
|
|
|
|
Comment: `network BaseFee below which to stop doing commit aggregation, instead
|
|
submitting proofs to the chain individually`,
|
|
},
|
|
{
|
|
Name: "MaxSectorProveCommitsSubmittedPerEpoch",
|
|
Type: "uint64",
|
|
|
|
Comment: `When submitting several sector prove commit messages simultaneously, this option allows you to
|
|
stagger the number of prove commits submitted per epoch
|
|
This is done because gas estimates for ProveCommits are non deterministic and increasing as a large
|
|
number of sectors get committed within the same epoch resulting in occasionally failed msgs.
|
|
Submitting a smaller number of prove commits per epoch would reduce the possibility of failed msgs`,
|
|
},
|
|
{
|
|
Name: "TerminateBatchMax",
|
|
Type: "uint64",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "TerminateBatchMin",
|
|
Type: "uint64",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "TerminateBatchWait",
|
|
Type: "Duration",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "UseSyntheticPoRep",
|
|
Type: "bool",
|
|
|
|
Comment: `UseSyntheticPoRep, when set to true, will reduce the amount of cache data held on disk after the completion of PreCommit 2 to 11GiB.`,
|
|
},
|
|
{
|
|
Name: "RequireActivationSuccess",
|
|
Type: "bool",
|
|
|
|
Comment: `Whether to abort if any sector activation in a batch fails (newly sealed sectors, only with ProveCommitSectors3).`,
|
|
},
|
|
{
|
|
Name: "RequireActivationSuccessUpdate",
|
|
Type: "bool",
|
|
|
|
Comment: `Whether to abort if any piece activation notification returns a non-zero exit code (newly sealed sectors, only with ProveCommitSectors3).`,
|
|
},
|
|
{
|
|
Name: "RequireNotificationSuccess",
|
|
Type: "bool",
|
|
|
|
Comment: `Whether to abort if any sector activation in a batch fails (updating sectors, only with ProveReplicaUpdates3).`,
|
|
},
|
|
{
|
|
Name: "RequireNotificationSuccessUpdate",
|
|
Type: "bool",
|
|
|
|
Comment: `Whether to abort if any piece activation notification returns a non-zero exit code (updating sectors, only with ProveReplicaUpdates3).`,
|
|
},
|
|
},
|
|
"Splitstore": {
|
|
{
|
|
Name: "ColdStoreType",
|
|
Type: "string",
|
|
|
|
Comment: `ColdStoreType specifies the type of the coldstore.
|
|
It can be "discard" (default) for discarding cold blocks, "messages" to store only messages or "universal" to store all chain state..`,
|
|
},
|
|
{
|
|
Name: "HotStoreType",
|
|
Type: "string",
|
|
|
|
Comment: `HotStoreType specifies the type of the hotstore.
|
|
Only currently supported value is "badger".`,
|
|
},
|
|
{
|
|
Name: "MarkSetType",
|
|
Type: "string",
|
|
|
|
Comment: `MarkSetType specifies the type of the markset.
|
|
It can be "map" for in memory marking or "badger" (default) for on-disk marking.`,
|
|
},
|
|
{
|
|
Name: "HotStoreMessageRetention",
|
|
Type: "uint64",
|
|
|
|
Comment: `HotStoreMessageRetention specifies the retention policy for messages, in finalities beyond
|
|
the compaction boundary; default is 0.`,
|
|
},
|
|
{
|
|
Name: "HotStoreFullGCFrequency",
|
|
Type: "uint64",
|
|
|
|
Comment: `HotStoreFullGCFrequency specifies how often to perform a full (moving) GC on the hotstore.
|
|
A value of 0 disables, while a value 1 will do full GC in every compaction.
|
|
Default is 20 (about once a week).`,
|
|
},
|
|
{
|
|
Name: "HotStoreMaxSpaceTarget",
|
|
Type: "uint64",
|
|
|
|
Comment: `HotStoreMaxSpaceTarget sets a target max disk size for the hotstore. Splitstore GC
|
|
will run moving GC if disk utilization gets within a threshold (150 GB) of the target.
|
|
Splitstore GC will NOT run moving GC if the total size of the move would get
|
|
within 50 GB of the target, and instead will run a more aggressive online GC.
|
|
If both HotStoreFullGCFrequency and HotStoreMaxSpaceTarget are set then splitstore
|
|
GC will trigger moving GC if either configuration condition is met.
|
|
A reasonable minimum is 2x fully GCed hotstore size + 50 G buffer.
|
|
At this minimum size moving GC happens every time, any smaller and moving GC won't
|
|
be able to run. In spring 2023 this minimum is ~550 GB.`,
|
|
},
|
|
{
|
|
Name: "HotStoreMaxSpaceThreshold",
|
|
Type: "uint64",
|
|
|
|
Comment: `When HotStoreMaxSpaceTarget is set Moving GC will be triggered when total moving size
|
|
exceeds HotstoreMaxSpaceTarget - HotstoreMaxSpaceThreshold`,
|
|
},
|
|
{
|
|
Name: "HotstoreMaxSpaceSafetyBuffer",
|
|
Type: "uint64",
|
|
|
|
Comment: `Safety buffer to prevent moving GC from overflowing disk when HotStoreMaxSpaceTarget
|
|
is set. Moving GC will not occur when total moving size exceeds
|
|
HotstoreMaxSpaceTarget - HotstoreMaxSpaceSafetyBuffer`,
|
|
},
|
|
},
|
|
"StorageMiner": {
|
|
{
|
|
Name: "Subsystems",
|
|
Type: "MinerSubsystemConfig",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "Dealmaking",
|
|
Type: "DealmakingConfig",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "IndexProvider",
|
|
Type: "IndexProviderConfig",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "Proving",
|
|
Type: "ProvingConfig",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "Sealing",
|
|
Type: "SealingConfig",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "Storage",
|
|
Type: "SealerConfig",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "Fees",
|
|
Type: "MinerFeeConfig",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "Addresses",
|
|
Type: "MinerAddressConfig",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "DAGStore",
|
|
Type: "DAGStoreConfig",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "HarmonyDB",
|
|
Type: "HarmonyDB",
|
|
|
|
Comment: ``,
|
|
},
|
|
},
|
|
"Wallet": {
|
|
{
|
|
Name: "RemoteBackend",
|
|
Type: "string",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "EnableLedger",
|
|
Type: "bool",
|
|
|
|
Comment: ``,
|
|
},
|
|
{
|
|
Name: "DisableLocal",
|
|
Type: "bool",
|
|
|
|
Comment: ``,
|
|
},
|
|
},
|
|
}
|