2021-07-23 11:55:50 +00:00
|
|
|
package config
|
|
|
|
|
|
|
|
import (
|
|
|
|
"github.com/filecoin-project/lotus/chain/types"
|
|
|
|
)
|
|
|
|
|
|
|
|
// // NOTE: ONLY PUT STRUCT DEFINITIONS IN THIS FILE
|
2021-07-23 14:02:45 +00:00
|
|
|
// //
|
|
|
|
// // After making edits here, run 'make cfgdoc-gen' (or 'make gen')
|
2021-07-23 11:55:50 +00:00
|
|
|
|
|
|
|
// Common is common config between full node and miner
|
|
|
|
type Common struct {
|
2022-03-10 10:58:31 +00:00
|
|
|
API API
|
|
|
|
Backup Backup
|
|
|
|
Logging Logging
|
|
|
|
Libp2p Libp2p
|
|
|
|
Pubsub Pubsub
|
2021-07-23 11:55:50 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// FullNode is a full node config
|
|
|
|
type FullNode struct {
|
|
|
|
Common
|
2023-08-01 15:28:47 +00:00
|
|
|
Wallet Wallet
|
|
|
|
Fees FeeConfig
|
|
|
|
Chainstore Chainstore
|
|
|
|
Fevm FevmConfig
|
chore: Merge nv22 into master (#11699)
* [WIP] feat: Add nv22 skeleton
Addition of Network Version 22 skeleton
* update FFI
* feat: drand: refactor round verification
* feat: sealing: Support nv22 DDO features in the sealing pipeline (#11226)
* Initial work supporting DDO pieces in lotus-miner
* sealing: Update pipeline input to operate on UniversalPiece
* sealing: Update pipeline checks/sealing states to operate on UniversalPiece
* sealing: Make pipeline build with UniversalPiece
* move PieceDealInfo out of api
* make gen
* make sealing pipeline unit tests pass
* fix itest ensemble build
* don't panic in SectorsStatus with deals
* stop linter from complaining about checkPieces
* fix sector import tests
* mod tidy
* sealing: Add logic for (pre)committing DDO sectors
* sealing: state-types with method defs
* DDO non-snap pipeline works(?), DDO Itests
* DDO support in snapdeals pipeline
* make gen
* update actor bundles
* update the gst market fix
* fix: chain: use PreCommitSectorsBatch2 when setting up genesis
* some bug fixes
* integration working changes
* update actor bundles
* Make TestOnboardRawPieceSnap pass
* Appease the linter
* Make deadlines test pass with v12 actors
* Update go-state-types, abstract market DealState
* make gen
* mod tidy, lint fixes
* Fix some more tests
* Bump version in master
Bump version in master
* Make gen
Make gen
* fix sender
* fix: lotus-provider: Fix winning PoSt
* fix: sql Scan cannot write to an object
* Actually show miner-addrs in info-log
Actually show miner-addrs in lotus-provider info-log
* [WIP] feat: Add nv22 skeleton
Addition of Network Version 22 skeleton
* update FFI
* ddo is now nv22
* make gen
* temp actor bundle with ddo
* use working go-state-types
* gst with v13 market migration
* update bundle, builtin.MethodsMiner.ProveCommitSectors2 -> 3
* actually working v13 migration, v13 migration itest
* Address review
* sealing: Correct DDO snap pledge math
* itests: Mixed ddo itest
* pipeline: Fix sectorWeight
* sealing: convert market deals into PAMs in mixed sectors
* sealing: make market to ddo conversion work
* fix lint
* update gst
* Update actors and GST to lastest integ branch
* commit batcher: Update ProveCommitSectors3Params builder logic
* make gen
* use builtin-actors master
* ddo: address review
* itests: Add commd assertions to ddo tests
* make gen
* gst with fixed types
* config knobs for RequireActivationSuccess
* storage: Drop obsolete flaky tasts
---------
Co-authored-by: Jennifer Wang <jiayingw703@gmail.com>
Co-authored-by: Aayush <arajasek94@gmail.com>
Co-authored-by: Shrenuj Bansal <shrenuj.bansal@protocol.ai>
Co-authored-by: Phi <orjan.roren@gmail.com>
Co-authored-by: Andrew Jackson (Ajax) <snadrus@gmail.com>
Co-authored-by: TippyFlits <james.bluett@protocol.ai>
* feat: implement FIP-0063
* chore: deps: update to go-multiaddr v0.12.2 (#11602)
* feat: fvm: update the FVM/FFI to v4.1 (#11608) (#11612)
This:
1. Adds nv22 support.
2. Updates the message tracing format.
Co-authored-by: Steven Allen <steven@stebalien.com>
* AggregateProofType nil when doing batch updates
Use latest nv22 go-state-types version with matching update
* Update to v13.0.0-rc.2 bundle
* chore: Upgrade heights and codename
Update upgrade heights
Co-Authored-By: Steven Allen <steven@stebalien.com>
* Update epoch after nv22 DRAND switch
Update epoch after nv22 DRAND switch
* Update Mango codename to Phoneix
Make the codename for the Drand-change inline with Dragon style.
* Add UpgradePhoenixHeight to API params
* set UpgradePhoenixHeight to be one hour after Dragon
* Make gen
Make gen and UpgradePhoenixHeight in butterfly and local devnet to be in line with Calibration and Mainnet
* Update epoch heights (#11637)
Update epoch heights
* new: add forest bootstrap nodes (#11636)
Signed-off-by: samuelarogbonlo <sbayo971@gmail.com>
* Merge pull request #11491 from filecoin-project/fix/remove-decommissioned-pl-bootstrap-nodes
Remove PL operated bootstrap nodes from mainnet.pi
* feat: api: new verified registry methods to get all allocations and claims (#11631)
* new verireg methods
* update changelog and add itest
* update itest and cli
* update new method's support till v9
* remove gateway APIs
* fix cli internal var names
* chore:: backport #11609 to the feat/nv22 branch (#11644)
* feat: api: improve the correctness of Eth's trace_block (#11609)
* Improve the correctness of Eth's trace_block
- Improve encoding/decoding of parameters and return values:
- Encode "native" parameters and return values with Solidity ABI.
- Correctly decode parameters to "create" calls.
- Use the correct (ish) output for "create" calls.
- Handle all forms of "create".
- Make robust with respect to reverts:
- Use the actor ID/address from the trace instead of looking it up in
the state-tree (may not exist in the state-tree due to a revert).
- Gracefully handle failed actor/contract creation.
- Improve performance:
- We avoid looking anything up in the state-tree when translating the
trace, which should significantly improve performance.
- Improve code readability:
- Remove all "backtracking" logic.
- Use an "environment" struct to store temporary state instead of
attaching it to the trace.
- Fix random bugs:
- Fix an allocation bug in the "address" logic (need to set the
capacity before modifying the slice).
- Improved error checking/handling.
- Use correct types for `trace_block` action/results (create, call, etc.).
- And use the correct types for Result/Action structs instead of reusing the same "Call" action every time.
- Improve error messages.
* Make gen
Make gen
---------
Co-authored-by: Steven Allen <steven@stebalien.com>
* fix: add UpgradePhoenixHeight to StateGetNetworkParams (#11648)
* chore: deps: update to go-state-types v13.0.0-rc.1
* do NOT update the cache when running the real migration
* Merge pull request #11632 from hanabi1224/hm/drand-test
feat: drand quicknet: allow scheduling drand quicknet upgrade before nv22 on 2k devnet
* chore: deps: update to go-state-types v13.0.0-rc.2
chore: deps: update to go-state-types v13.0.0-rc.2
* feat: set migration config UpgradeEpoch for v13 actors upgrade
* Built-in actor events first draft
* itest for DDO non-market verified data w/ builtin actor events
* Tests for builtin actor events API
* Clean up DDO+Events tests, add lots of explainer comments
* Minor tweaks to events types
* Avoid duplicate messages when looking for receipts
* Rename internal events modules for clarity
* Adjust actor event API after review
* s/ActorEvents/Events/g in global config
* Manage event sending rate for SubscribeActorEvents
* Terminate SubscribeActorEvents chan when at max height
* Document future API changes
* More clarity in actor event API docs
* More post-review changes, lots of tests for SubscribeActorEvents
Use BlockDelay as the window for receiving events on the SubscribeActorEvents
channel. We expect the user to have received the initial batch of historical
events (if any) in one block's time. For real-time events we expect them to
not fall behind by roughly one block's time.
* Remove duplicate code from actor event type marshalling tests
Reduce verbosity and remove duplicate test logic from actor event types
JSON marshalling tests.
* Rename actor events test to follow go convention
Add missing `s` to `actor_events` test file to follow golang convention
used across the repo.
* Run actor events table tests in deterministic order
Refactor `map` usage for actor event table tests to ensure deterministic
test execution order, making debugging potential issues easier. If
non-determinism is a target, leverage Go's built-in parallel testing
capabilities.
* Reduce scope for filter removal failure when getting actor events
Use a fresh context to remove the temporary filter installed solely to
get the actor events. This should reduce chances of failure in a case
where the original context may be expired/cancelled.
Refactor removal into a `defer` statement for a more readable, concise
return statement.
* Use fixed RNG seed for actor event tests
Improve determinism in actor event tests by using a fixed RNG seed. This
makes up a more reproducible test suit.
* Use provided libraries to assert eventual conditions
Use the functionalities already provided by `testify` to assert eventual
conditions, and remove the use of `time.Sleep`.
Remove duplicate code in utility functions that are already defined.
Refactor assertion helper functions to use consistent terminology:
"require" implies fatal error, whereas "assert" implies error where the
test may proceed executing.
* Update changelog for actor events APIs
* Fix concerns and docs identified by review
* Update actor bundle to v13.0.0-rc3
Update actor bundle to v13.0.0-rc3
* Prep Lotus v1.26.0-rc1
- For sanity reverting the mainnet upgrade epoch to 99999999, and then only set it when cutting the final release
-Update Calibnet CIDs to v13.0.0-rc3
- Add GetActorEvents, SubscribeActorEvents, GetAllClaims and GetAllAllocations methods to the changelog
Co-Authored-By: Jiaying Wang <42981373+jennijuju@users.noreply.github.com>
* Update CHANGELOG.md
Co-authored-by: Masih H. Derkani <m@derkani.org>
* Make gen
Make gen
* fix: beacon: validate drand change at nv16 correctly
* bump to v1.26.0-rc2
* test: cleanup ddo verified itest, extract steps to functions
also add allocation-removed event case
* test: extract verified DDO test to separate file, add more checks
* test: add additional actor events checks
* Add verification for "deal-activated" actor event
* docs(drand): document the meaning of "IsChained" (#11692)
* Resolve conflicts
I encountered multiple issues when trying to run make gen. And these changes fixed a couple of them:
- go mod tidy
- Remove RaftState/RaftLeader
- Revert `if ts.Height() > claim.TermMax+claim.TermStart || !cctx.IsSet("expired")` to the what is in the release/v1.26.0: `if tsHeight > val.TermMax || !expired`
* fixup imports, make jen
* Update version
Update version in master to v1.27.0-dev
* Update node/impl/full/dummy.go
Co-authored-by: Łukasz Magiera <magik6k@users.noreply.github.com>
* Adjust ListClaimsCmd
Adjust ListClaimsCmd according to review
---------
Signed-off-by: samuelarogbonlo <sbayo971@gmail.com>
Co-authored-by: TippyFlits <james.bluett@protocol.ai>
Co-authored-by: Aayush <arajasek94@gmail.com>
Co-authored-by: Łukasz Magiera <magik6k@users.noreply.github.com>
Co-authored-by: Jennifer Wang <jiayingw703@gmail.com>
Co-authored-by: Shrenuj Bansal <shrenuj.bansal@protocol.ai>
Co-authored-by: Andrew Jackson (Ajax) <snadrus@gmail.com>
Co-authored-by: Steven Allen <steven@stebalien.com>
Co-authored-by: Rod Vagg <rod@vagg.org>
Co-authored-by: Samuel Arogbonlo <47984109+samuelarogbonlo@users.noreply.github.com>
Co-authored-by: LexLuthr <88259624+LexLuthr@users.noreply.github.com>
Co-authored-by: tom123222 <160735201+tom123222@users.noreply.github.com>
Co-authored-by: Aarsh Shah <aarshkshah1992@gmail.com>
Co-authored-by: Masih H. Derkani <m@derkani.org>
Co-authored-by: Jiaying Wang <42981373+jennijuju@users.noreply.github.com>
2024-03-12 09:33:58 +00:00
|
|
|
Events EventsConfig
|
2023-08-01 15:28:47 +00:00
|
|
|
Index IndexConfig
|
|
|
|
FaultReporter FaultReporterConfig
|
2021-07-23 11:55:50 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// // Common
|
|
|
|
|
|
|
|
type Backup struct {
|
2021-07-23 13:40:30 +00:00
|
|
|
// When set to true disables metadata log (.lotus/kvlog). This can save disk
|
|
|
|
// space by reducing metadata redundancy.
|
|
|
|
//
|
|
|
|
// Note that in case of metadata corruption it might be much harder to recover
|
|
|
|
// your node if metadata log is disabled
|
2021-07-23 11:55:50 +00:00
|
|
|
DisableMetadataLog bool
|
|
|
|
}
|
|
|
|
|
2022-03-10 10:58:31 +00:00
|
|
|
// Logging is the logging system config
|
|
|
|
type Logging struct {
|
|
|
|
// SubsystemLevels specify per-subsystem log levels
|
|
|
|
SubsystemLevels map[string]string
|
|
|
|
}
|
|
|
|
|
2021-07-23 11:55:50 +00:00
|
|
|
// StorageMiner is a miner config
|
|
|
|
type StorageMiner struct {
|
|
|
|
Common
|
|
|
|
|
2024-06-21 18:48:22 +00:00
|
|
|
Subsystems MinerSubsystemConfig
|
|
|
|
Dealmaking DealmakingConfig
|
|
|
|
Proving ProvingConfig
|
|
|
|
Sealing SealingConfig
|
|
|
|
Storage SealerConfig
|
|
|
|
Fees MinerFeeConfig
|
|
|
|
Addresses MinerAddressConfig
|
|
|
|
HarmonyDB HarmonyDB
|
integrate DAG store and CARv2 in deal-making (#6671)
This commit removes badger from the deal-making processes, and
moves to a new architecture with the dagstore as the cental
component on the miner-side, and CARv2s on the client-side.
Every deal that has been handed off to the sealing subsystem becomes
a shard in the dagstore. Shards are mounted via the LotusMount, which
teaches the dagstore how to load the related piece when serving
retrievals.
When the miner starts the Lotus for the first time with this patch,
we will perform a one-time migration of all active deals into the
dagstore. This is a lightweight process, and it consists simply
of registering the shards in the dagstore.
Shards are backed by the unsealed copy of the piece. This is currently
a CARv1. However, the dagstore keeps CARv2 indices for all pieces, so
when it's time to acquire a shard to serve a retrieval, the unsealed
CARv1 is joined with its index (safeguarded by the dagstore), to form
a read-only blockstore, thus taking the place of the monolithic
badger.
Data transfers have been adjusted to interface directly with CARv2 files.
On inbound transfers (client retrievals, miner storage deals), we stream
the received data into a CARv2 ReadWrite blockstore. On outbound transfers
(client storage deals, miner retrievals), we serve the data off a CARv2
ReadOnly blockstore.
Client-side imports are managed by the refactored *imports.Manager
component (when not using IPFS integration). Just like it before, we use
the go-filestore library to avoid duplicating the data from the original
file in the resulting UnixFS DAG (concretely the leaves). However, the
target of those imports are what we call "ref-CARv2s": CARv2 files placed
under the `$LOTUS_PATH/imports` directory, containing the intermediate
nodes in full, and the leaves as positional references to the original file
on disk.
Client-side retrievals are placed into CARv2 files in the location:
`$LOTUS_PATH/retrievals`.
A new set of `Dagstore*` JSON-RPC operations and `lotus-miner dagstore`
subcommands have been introduced on the miner-side to inspect and manage
the dagstore.
Despite moving to a CARv2-backed system, the IPFS integration has been
respected, and it continues to be possible to make storage deals with data
held in an IPFS node, and to perform retrievals directly into an IPFS node.
NOTE: because the "staging" and "client" Badger blockstores are no longer
used, existing imports on the client will be rendered useless. On startup,
Lotus will enumerate all imports and print WARN statements on the log for
each import that needs to be reimported. These log lines contain these
messages:
- import lacks carv2 path; import will not work; please reimport
- import has missing/broken carv2; please reimport
At the end, we will print a "sanity check completed" message indicating
the count of imports found, and how many were deemed broken.
Co-authored-by: Aarsh Shah <aarshkshah1992@gmail.com>
Co-authored-by: Dirk McCormick <dirkmdev@gmail.com>
Co-authored-by: Raúl Kripalani <raul@protocol.ai>
Co-authored-by: Dirk McCormick <dirkmdev@gmail.com>
2021-08-16 22:34:32 +00:00
|
|
|
}
|
|
|
|
|
2024-03-15 21:38:13 +00:00
|
|
|
type CurioConfig struct {
|
|
|
|
Subsystems CurioSubsystemsConfig
|
2023-09-20 03:48:39 +00:00
|
|
|
|
2024-03-15 21:38:13 +00:00
|
|
|
Fees CurioFees
|
2024-01-20 00:48:29 +00:00
|
|
|
|
|
|
|
// Addresses of wallets per MinerAddress (one of the fields).
|
2024-03-15 21:38:13 +00:00
|
|
|
Addresses []CurioAddresses
|
2024-04-02 16:21:21 +00:00
|
|
|
Proving CurioProvingConfig
|
2024-04-12 02:36:38 +00:00
|
|
|
Ingest CurioIngestConfig
|
2023-10-30 16:19:07 +00:00
|
|
|
Journal JournalConfig
|
|
|
|
Apis ApisConfig
|
2024-06-21 18:48:22 +00:00
|
|
|
Alerting CurioAlerting
|
2023-10-16 15:28:58 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
type ApisConfig struct {
|
2023-11-09 00:01:44 +00:00
|
|
|
// ChainApiInfo is the API endpoint for the Lotus daemon.
|
|
|
|
ChainApiInfo []string
|
2023-10-31 13:45:38 +00:00
|
|
|
|
|
|
|
// RPC Secret for the storage subsystem.
|
|
|
|
// If integrating with lotus-miner this must match the value from
|
|
|
|
// cat ~/.lotusminer/keystore/MF2XI2BNNJ3XILLQOJUXMYLUMU | jq -r .PrivateKey
|
|
|
|
StorageRPCSecret string
|
2023-10-16 15:28:58 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
type JournalConfig struct {
|
|
|
|
//Events of the form: "system1:event1,system1:event2[,...]"
|
|
|
|
DisabledEvents string
|
2023-09-20 03:48:39 +00:00
|
|
|
}
|
|
|
|
|
2024-03-15 21:38:13 +00:00
|
|
|
type CurioSubsystemsConfig struct {
|
2024-04-04 12:06:21 +00:00
|
|
|
// EnableWindowPost enables window post to be executed on this curio instance. Each machine in the cluster
|
2024-02-21 15:51:29 +00:00
|
|
|
// with WindowPoSt enabled will also participate in the window post scheduler. It is possible to have multiple
|
|
|
|
// machines with WindowPoSt enabled which will provide redundancy, and in case of multiple partitions per deadline,
|
|
|
|
// will allow for parallel processing of partitions.
|
|
|
|
//
|
|
|
|
// It is possible to have instances handling both WindowPoSt and WinningPoSt, which can provide redundancy without
|
|
|
|
// the need for additional machines. In setups like this it is generally recommended to run
|
|
|
|
// partitionsPerDeadline+1 machines.
|
|
|
|
EnableWindowPost bool
|
|
|
|
WindowPostMaxTasks int
|
|
|
|
|
2024-04-04 12:06:21 +00:00
|
|
|
// EnableWinningPost enables winning post to be executed on this curio instance.
|
2024-02-21 15:51:29 +00:00
|
|
|
// Each machine in the cluster with WinningPoSt enabled will also participate in the winning post scheduler.
|
|
|
|
// It is possible to mix machines with WindowPoSt and WinningPoSt enabled, for details see the EnableWindowPost
|
|
|
|
// documentation.
|
2023-11-10 19:17:05 +00:00
|
|
|
EnableWinningPost bool
|
|
|
|
WinningPostMaxTasks int
|
2023-12-12 05:16:57 +00:00
|
|
|
|
2024-02-28 19:57:12 +00:00
|
|
|
// EnableParkPiece enables the "piece parking" task to run on this node. This task is responsible for fetching
|
|
|
|
// pieces from the network and storing them in the storage subsystem until sectors are sealed. This task is
|
|
|
|
// only applicable when integrating with boost, and should be enabled on nodes which will hold deal data
|
|
|
|
// from boost until sectors containing the related pieces have the TreeD/TreeR constructed.
|
|
|
|
// Note that future Curio implementations will have a separate task type for fetching pieces from the internet.
|
|
|
|
EnableParkPiece bool
|
|
|
|
ParkPieceMaxTasks int
|
|
|
|
|
2023-12-19 11:16:38 +00:00
|
|
|
// EnableSealSDR enables SDR tasks to run. SDR is the long sequential computation
|
2024-02-21 15:51:29 +00:00
|
|
|
// creating 11 layer files in sector cache directory.
|
|
|
|
//
|
|
|
|
// SDR is the first task in the sealing pipeline. It's inputs are just the hash of the
|
|
|
|
// unsealed data (CommD), sector number, miner id, and the seal proof type.
|
|
|
|
// It's outputs are the 11 layer files in the sector cache directory.
|
|
|
|
//
|
2023-12-20 13:45:19 +00:00
|
|
|
// In lotus-miner this was run as part of PreCommit1.
|
2024-02-21 15:51:29 +00:00
|
|
|
EnableSealSDR bool
|
|
|
|
|
|
|
|
// The maximum amount of SDR tasks that can run simultaneously. Note that the maximum number of tasks will
|
|
|
|
// also be bounded by resources available on the machine.
|
2023-12-19 11:16:38 +00:00
|
|
|
SealSDRMaxTasks int
|
2023-12-20 13:45:19 +00:00
|
|
|
|
|
|
|
// EnableSealSDRTrees enables the SDR pipeline tree-building task to run.
|
|
|
|
// This task handles encoding of unsealed data into last sdr layer and building
|
|
|
|
// of TreeR, TreeC and TreeD.
|
2024-02-21 15:51:29 +00:00
|
|
|
//
|
|
|
|
// This task runs after SDR
|
|
|
|
// TreeD is first computed with optional input of unsealed data
|
|
|
|
// TreeR is computed from replica, which is first computed as field
|
|
|
|
// addition of the last SDR layer and the bottom layer of TreeD (which is the unsealed data)
|
|
|
|
// TreeC is computed from the 11 SDR layers
|
|
|
|
// The 3 trees will later be used to compute the PoRep proof.
|
|
|
|
//
|
|
|
|
// In case of SyntheticPoRep challenges for PoRep will be pre-generated at this step, and trees and layers
|
|
|
|
// will be dropped. SyntheticPoRep works by pre-generating a very large set of challenges (~30GiB on disk)
|
|
|
|
// then using a small subset of them for the actual PoRep computation. This allows for significant scratch space
|
|
|
|
// saving between PreCommit and PoRep generation at the expense of more computation (generating challenges in this step)
|
|
|
|
//
|
2023-12-20 13:45:19 +00:00
|
|
|
// In lotus-miner this was run as part of PreCommit2 (TreeD was run in PreCommit1).
|
2024-01-30 19:05:47 +00:00
|
|
|
// Note that nodes with SDRTrees enabled will also answer to Finalize tasks,
|
|
|
|
// which just remove unneeded tree data after PoRep is computed.
|
2024-02-21 15:51:29 +00:00
|
|
|
EnableSealSDRTrees bool
|
|
|
|
|
|
|
|
// The maximum amount of SealSDRTrees tasks that can run simultaneously. Note that the maximum number of tasks will
|
|
|
|
// also be bounded by resources available on the machine.
|
2023-12-20 13:45:19 +00:00
|
|
|
SealSDRTreesMaxTasks int
|
2024-02-21 15:51:29 +00:00
|
|
|
|
|
|
|
// FinalizeMaxTasks is the maximum amount of finalize tasks that can run simultaneously.
|
|
|
|
// The finalize task is enabled on all machines which also handle SDRTrees tasks. Finalize ALWAYS runs on whichever
|
|
|
|
// machine holds sector cache files, as it removes unneeded tree data after PoRep is computed.
|
|
|
|
// Finalize will run in parallel with the SubmitCommitMsg task.
|
|
|
|
FinalizeMaxTasks int
|
2024-01-05 15:10:34 +00:00
|
|
|
|
|
|
|
// EnableSendPrecommitMsg enables the sending of precommit messages to the chain
|
2024-04-04 12:06:21 +00:00
|
|
|
// from this curio instance.
|
2024-02-21 15:51:29 +00:00
|
|
|
// This runs after SDRTrees and uses the output CommD / CommR (roots of TreeD / TreeR) for the message
|
2024-01-05 15:10:34 +00:00
|
|
|
EnableSendPrecommitMsg bool
|
2024-01-11 15:47:07 +00:00
|
|
|
|
2024-02-21 15:51:29 +00:00
|
|
|
// EnablePoRepProof enables the computation of the porep proof
|
|
|
|
//
|
|
|
|
// This task runs after interactive-porep seed becomes available, which happens 150 epochs (75min) after the
|
|
|
|
// precommit message lands on chain. This task should run on a machine with a GPU. Vanilla PoRep proofs are
|
|
|
|
// requested from the machine which holds sector cache files which most likely is the machine which ran the SDRTrees
|
|
|
|
// task.
|
|
|
|
//
|
2024-01-11 15:47:07 +00:00
|
|
|
// In lotus-miner this was Commit1 / Commit2
|
2024-02-21 15:51:29 +00:00
|
|
|
EnablePoRepProof bool
|
|
|
|
|
|
|
|
// The maximum amount of PoRepProof tasks that can run simultaneously. Note that the maximum number of tasks will
|
|
|
|
// also be bounded by resources available on the machine.
|
2024-01-11 15:47:07 +00:00
|
|
|
PoRepProofMaxTasks int
|
2024-01-11 18:24:28 +00:00
|
|
|
|
|
|
|
// EnableSendCommitMsg enables the sending of commit messages to the chain
|
2024-04-04 12:06:21 +00:00
|
|
|
// from this curio instance.
|
2024-01-11 18:24:28 +00:00
|
|
|
EnableSendCommitMsg bool
|
2024-01-12 10:03:37 +00:00
|
|
|
|
2024-04-04 12:06:21 +00:00
|
|
|
// EnableMoveStorage enables the move-into-long-term-storage task to run on this curio instance.
|
2024-02-21 15:51:29 +00:00
|
|
|
// This tasks should only be enabled on nodes with long-term storage.
|
|
|
|
//
|
|
|
|
// The MoveStorage task is the last task in the sealing pipeline. It moves the sealed sector data from the
|
|
|
|
// SDRTrees machine into long-term storage. This task runs after the Finalize task.
|
|
|
|
EnableMoveStorage bool
|
|
|
|
|
|
|
|
// The maximum amount of MoveStorage tasks that can run simultaneously. Note that the maximum number of tasks will
|
|
|
|
// also be bounded by resources available on the machine. It is recommended that this value is set to a number which
|
|
|
|
// uses all available network (or disk) bandwidth on the machine without causing bottlenecks.
|
2024-01-30 19:05:47 +00:00
|
|
|
MoveStorageMaxTasks int
|
|
|
|
|
2024-04-03 16:31:43 +00:00
|
|
|
// BoostAdapters is a list of tuples of miner address and port/ip to listen for market (e.g. boost) requests.
|
|
|
|
// This interface is compatible with the lotus-miner RPC, implementing a subset needed for storage market operations.
|
2024-06-21 18:48:22 +00:00
|
|
|
// Strings should be in the format "actor:ip:port". IP cannot be 0.0.0.0. We recommend using a private IP.
|
|
|
|
// Example: "f0123:127.0.0.1:32100". Multiple addresses can be specified.
|
2024-04-03 16:31:43 +00:00
|
|
|
//
|
|
|
|
// When a market node like boost gives Curio's market RPC a deal to placing into a sector, Curio will first store the
|
|
|
|
// deal data in a temporary location "Piece Park" before assigning it to a sector. This requires that at least one
|
|
|
|
// node in the cluster has the EnableParkPiece option enabled and has sufficient scratch space to store the deal data.
|
|
|
|
// This is different from lotus-miner which stored the deal data into an "unsealed" sector as soon as the deal was
|
|
|
|
// received. Deal data in PiecePark is accessed when the sector TreeD and TreeR are computed, but isn't needed for
|
|
|
|
// the initial SDR layers computation. Pieces in PiecePark are removed after all sectors referencing the piece are
|
|
|
|
// sealed.
|
|
|
|
//
|
|
|
|
// To get API info for boost configuration run 'curio market rpc-info'
|
|
|
|
//
|
|
|
|
// NOTE: All deal data will flow through this service, so it should be placed on a machine running boost or on
|
|
|
|
// a machine which handles ParkPiece tasks.
|
|
|
|
BoostAdapters []string
|
|
|
|
|
2024-04-04 12:06:21 +00:00
|
|
|
// EnableWebGui enables the web GUI on this curio instance. The UI has minimal local overhead, but it should
|
2024-02-21 15:51:29 +00:00
|
|
|
// only need to be run on a single machine in the cluster.
|
2024-01-12 10:03:37 +00:00
|
|
|
EnableWebGui bool
|
2024-02-21 15:51:29 +00:00
|
|
|
|
2024-01-12 10:03:37 +00:00
|
|
|
// The address that should listen for Web GUI requests.
|
|
|
|
GuiAddress string
|
integrate DAG store and CARv2 in deal-making (#6671)
This commit removes badger from the deal-making processes, and
moves to a new architecture with the dagstore as the cental
component on the miner-side, and CARv2s on the client-side.
Every deal that has been handed off to the sealing subsystem becomes
a shard in the dagstore. Shards are mounted via the LotusMount, which
teaches the dagstore how to load the related piece when serving
retrievals.
When the miner starts the Lotus for the first time with this patch,
we will perform a one-time migration of all active deals into the
dagstore. This is a lightweight process, and it consists simply
of registering the shards in the dagstore.
Shards are backed by the unsealed copy of the piece. This is currently
a CARv1. However, the dagstore keeps CARv2 indices for all pieces, so
when it's time to acquire a shard to serve a retrieval, the unsealed
CARv1 is joined with its index (safeguarded by the dagstore), to form
a read-only blockstore, thus taking the place of the monolithic
badger.
Data transfers have been adjusted to interface directly with CARv2 files.
On inbound transfers (client retrievals, miner storage deals), we stream
the received data into a CARv2 ReadWrite blockstore. On outbound transfers
(client storage deals, miner retrievals), we serve the data off a CARv2
ReadOnly blockstore.
Client-side imports are managed by the refactored *imports.Manager
component (when not using IPFS integration). Just like it before, we use
the go-filestore library to avoid duplicating the data from the original
file in the resulting UnixFS DAG (concretely the leaves). However, the
target of those imports are what we call "ref-CARv2s": CARv2 files placed
under the `$LOTUS_PATH/imports` directory, containing the intermediate
nodes in full, and the leaves as positional references to the original file
on disk.
Client-side retrievals are placed into CARv2 files in the location:
`$LOTUS_PATH/retrievals`.
A new set of `Dagstore*` JSON-RPC operations and `lotus-miner dagstore`
subcommands have been introduced on the miner-side to inspect and manage
the dagstore.
Despite moving to a CARv2-backed system, the IPFS integration has been
respected, and it continues to be possible to make storage deals with data
held in an IPFS node, and to perform retrievals directly into an IPFS node.
NOTE: because the "staging" and "client" Badger blockstores are no longer
used, existing imports on the client will be rendered useless. On startup,
Lotus will enumerate all imports and print WARN statements on the log for
each import that needs to be reimported. These log lines contain these
messages:
- import lacks carv2 path; import will not work; please reimport
- import has missing/broken carv2; please reimport
At the end, we will print a "sanity check completed" message indicating
the count of imports found, and how many were deemed broken.
Co-authored-by: Aarsh Shah <aarshkshah1992@gmail.com>
Co-authored-by: Dirk McCormick <dirkmdev@gmail.com>
Co-authored-by: Raúl Kripalani <raul@protocol.ai>
Co-authored-by: Dirk McCormick <dirkmdev@gmail.com>
2021-08-16 22:34:32 +00:00
|
|
|
}
|
|
|
|
|
2021-07-23 11:55:50 +00:00
|
|
|
type MinerSubsystemConfig struct {
|
|
|
|
EnableMining bool
|
|
|
|
EnableSealing bool
|
|
|
|
EnableSectorStorage bool
|
|
|
|
|
2023-08-16 14:57:08 +00:00
|
|
|
// When enabled, the sector index will reside in an external database
|
|
|
|
// as opposed to the local KV store in the miner process
|
|
|
|
// This is useful to allow workers to bypass the lotus miner to access sector information
|
2023-08-09 00:59:21 +00:00
|
|
|
EnableSectorIndexDB bool
|
2021-07-23 11:55:50 +00:00
|
|
|
|
|
|
|
SealerApiInfo string // if EnableSealing == false
|
|
|
|
SectorIndexApiInfo string // if EnableSectorStorage == false
|
2023-11-09 17:11:10 +00:00
|
|
|
|
|
|
|
// When window post is enabled, the miner will automatically submit window post proofs
|
|
|
|
// for all sectors that are eligible for window post
|
|
|
|
// IF WINDOW POST IS DISABLED, THE MINER WILL NOT SUBMIT WINDOW POST PROOFS
|
|
|
|
// THIS WILL RESULT IN FAULTS AND PENALTIES IF NO OTHER MECHANISM IS RUNNING
|
2023-11-09 17:22:08 +00:00
|
|
|
// TO SUBMIT WINDOW POST PROOFS.
|
|
|
|
// Note: This option is entirely disabling the window post scheduler,
|
|
|
|
// not just the builtin PoSt computation like Proving.DisableBuiltinWindowPoSt.
|
|
|
|
// This option will stop lotus-miner from performing any actions related
|
|
|
|
// to window post, including scheduling, submitting proofs, and recovering
|
|
|
|
// sectors.
|
2023-11-09 17:11:10 +00:00
|
|
|
DisableWindowPoSt bool
|
2023-11-11 11:19:58 +00:00
|
|
|
|
|
|
|
// When winning post is disabled, the miner process will NOT attempt to mine
|
|
|
|
// blocks. This should only be set when there's an external process mining
|
|
|
|
// blocks on behalf of the miner.
|
|
|
|
// When disabled and no external block producers are configured, all potential
|
|
|
|
// block rewards will be missed!
|
|
|
|
DisableWinningPoSt bool
|
2021-07-23 11:55:50 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
type DealmakingConfig struct {
|
2021-09-30 12:35:23 +00:00
|
|
|
// Minimum start epoch buffer to give time for sealing of sector with deal.
|
|
|
|
StartEpochSealingBuffer uint64
|
2021-07-23 11:55:50 +00:00
|
|
|
}
|
|
|
|
|
2022-03-29 01:19:11 +00:00
|
|
|
type ProvingConfig struct {
|
|
|
|
// Maximum number of sector checks to run in parallel. (0 = unlimited)
|
2022-07-01 19:24:54 +00:00
|
|
|
//
|
|
|
|
// WARNING: Setting this value too high may make the node crash by running out of stack
|
|
|
|
// WARNING: Setting this value too low may make sector challenge reading much slower, resulting in failed PoSt due
|
|
|
|
// to late submission.
|
2022-07-01 20:20:05 +00:00
|
|
|
//
|
|
|
|
// After changing this option, confirm that the new value works in your setup by invoking
|
|
|
|
// 'lotus-miner proving compute window-post 0'
|
2022-03-29 01:19:11 +00:00
|
|
|
ParallelCheckLimit int
|
|
|
|
|
2022-11-17 17:25:30 +00:00
|
|
|
// Maximum amount of time a proving pre-check can take for a sector. If the check times out the sector will be skipped
|
|
|
|
//
|
|
|
|
// WARNING: Setting this value too low risks in sectors being skipped even though they are accessible, just reading the
|
|
|
|
// test challenge took longer than this timeout
|
|
|
|
// WARNING: Setting this value too high risks missing PoSt deadline in case IO operations related to this sector are
|
|
|
|
// blocked (e.g. in case of disconnected NFS mount)
|
|
|
|
SingleCheckTimeout Duration
|
|
|
|
|
|
|
|
// Maximum amount of time a proving pre-check can take for an entire partition. If the check times out, sectors in
|
|
|
|
// the partition which didn't get checked on time will be skipped
|
|
|
|
//
|
|
|
|
// WARNING: Setting this value too low risks in sectors being skipped even though they are accessible, just reading the
|
|
|
|
// test challenge took longer than this timeout
|
|
|
|
// WARNING: Setting this value too high risks missing PoSt deadline in case IO operations related to this partition are
|
|
|
|
// blocked or slow
|
|
|
|
PartitionCheckTimeout Duration
|
|
|
|
|
2022-07-01 19:24:54 +00:00
|
|
|
// Disable Window PoSt computation on the lotus-miner process even if no window PoSt workers are present.
|
|
|
|
//
|
|
|
|
// WARNING: If no windowPoSt workers are connected, window PoSt WILL FAIL resulting in faulty sectors which will need
|
|
|
|
// to be recovered. Before enabling this option, make sure your PoSt workers work correctly.
|
2022-07-01 20:20:05 +00:00
|
|
|
//
|
|
|
|
// After changing this option, confirm that the new value works in your setup by invoking
|
|
|
|
// 'lotus-miner proving compute window-post 0'
|
2022-07-01 19:24:54 +00:00
|
|
|
DisableBuiltinWindowPoSt bool
|
|
|
|
|
|
|
|
// Disable Winning PoSt computation on the lotus-miner process even if no winning PoSt workers are present.
|
|
|
|
//
|
|
|
|
// WARNING: If no WinningPoSt workers are connected, Winning PoSt WILL FAIL resulting in lost block rewards.
|
|
|
|
// Before enabling this option, make sure your PoSt workers work correctly.
|
|
|
|
DisableBuiltinWinningPoSt bool
|
2022-07-01 20:20:05 +00:00
|
|
|
|
|
|
|
// Disable WindowPoSt provable sector readability checks.
|
|
|
|
//
|
|
|
|
// In normal operation, when preparing to compute WindowPoSt, lotus-miner will perform a round of reading challenges
|
|
|
|
// from all sectors to confirm that those sectors can be proven. Challenges read in this process are discarded, as
|
2022-07-04 15:17:00 +00:00
|
|
|
// we're only interested in checking that sector data can be read.
|
2022-07-01 20:20:05 +00:00
|
|
|
//
|
|
|
|
// When using builtin proof computation (no PoSt workers, and DisableBuiltinWindowPoSt is set to false), this process
|
|
|
|
// can save a lot of time and compute resources in the case that some sectors are not readable - this is caused by
|
|
|
|
// the builtin logic not skipping snark computation when some sectors need to be skipped.
|
|
|
|
//
|
|
|
|
// When using PoSt workers, this process is mostly redundant, with PoSt workers challenges will be read once, and
|
|
|
|
// if challenges for some sectors aren't readable, those sectors will just get skipped.
|
|
|
|
//
|
2022-07-04 15:17:00 +00:00
|
|
|
// Disabling sector pre-checks will slightly reduce IO load when proving sectors, possibly resulting in shorter
|
2022-07-01 20:20:05 +00:00
|
|
|
// time to produce window PoSt. In setups with good IO capabilities the effect of this option on proving time should
|
|
|
|
// be negligible.
|
|
|
|
//
|
|
|
|
// NOTE: It likely is a bad idea to disable sector pre-checks in setups with no PoSt workers.
|
|
|
|
//
|
|
|
|
// NOTE: Even when this option is enabled, recovering sectors will be checked before recovery declaration message is
|
|
|
|
// sent to the chain
|
|
|
|
//
|
|
|
|
// After changing this option, confirm that the new value works in your setup by invoking
|
|
|
|
// 'lotus-miner proving compute window-post 0'
|
|
|
|
DisableWDPoStPreChecks bool
|
2022-07-07 10:33:40 +00:00
|
|
|
|
2023-10-10 20:34:45 +00:00
|
|
|
// Maximum number of partitions to prove in a single SubmitWindowPoSt messace. 0 = network limit (3 in nv21)
|
2022-07-07 10:33:40 +00:00
|
|
|
//
|
|
|
|
// A single partition may contain up to 2349 32GiB sectors, or 2300 64GiB sectors.
|
2023-10-10 20:34:45 +00:00
|
|
|
// //
|
2022-10-04 18:33:18 +00:00
|
|
|
// Note that setting this value lower may result in less efficient gas use - more messages will be sent,
|
2022-07-07 10:33:40 +00:00
|
|
|
// to prove each deadline, resulting in more total gas use (but each message will have lower gas limit)
|
|
|
|
//
|
|
|
|
// Setting this value above the network limit has no effect
|
2022-07-07 14:52:22 +00:00
|
|
|
MaxPartitionsPerPoStMessage int
|
|
|
|
|
|
|
|
// Maximum number of partitions to declare in a single DeclareFaultsRecovered message. 0 = no limit.
|
|
|
|
|
|
|
|
// In some cases when submitting DeclareFaultsRecovered messages,
|
|
|
|
// there may be too many recoveries to fit in a BlockGasLimit.
|
|
|
|
// In those cases it may be necessary to set this value to something low (eg 1);
|
|
|
|
// Note that setting this value lower may result in less efficient gas use - more messages will be sent than needed,
|
|
|
|
// resulting in more total gas use (but each message will have lower gas limit)
|
|
|
|
MaxPartitionsPerRecoveryMessage int
|
2022-10-04 18:33:18 +00:00
|
|
|
|
2022-10-04 19:21:55 +00:00
|
|
|
// Enable single partition per PoSt Message for partitions containing recovery sectors
|
2022-10-04 18:33:18 +00:00
|
|
|
//
|
|
|
|
// In cases when submitting PoSt messages which contain recovering sectors, the default network limit may still be
|
|
|
|
// too high to fit in the block gas limit. In those cases, it becomes useful to only house the single partition
|
2022-10-04 19:21:55 +00:00
|
|
|
// with recovering sectors in the post message
|
2022-10-04 18:33:18 +00:00
|
|
|
//
|
|
|
|
// Note that setting this value lower may result in less efficient gas use - more messages will be sent,
|
|
|
|
// to prove each deadline, resulting in more total gas use (but each message will have lower gas limit)
|
|
|
|
SingleRecoveringPartitionPerPostMessage bool
|
2022-03-29 01:19:11 +00:00
|
|
|
}
|
|
|
|
|
2021-07-23 11:55:50 +00:00
|
|
|
type SealingConfig struct {
|
2021-07-23 13:40:30 +00:00
|
|
|
// Upper bound on how many sectors can be waiting for more deals to be packed in it before it begins sealing at any given time.
|
|
|
|
// If the miner is accepting multiple deals in parallel, up to MaxWaitDealsSectors of new sectors will be created.
|
|
|
|
// If more than MaxWaitDealsSectors deals are accepted in parallel, only MaxWaitDealsSectors deals will be processed in parallel
|
|
|
|
// Note that setting this number too high in relation to deal ingestion rate may result in poor sector packing efficiency
|
2021-07-23 11:55:50 +00:00
|
|
|
// 0 = no limit
|
|
|
|
MaxWaitDealsSectors uint64
|
|
|
|
|
2022-03-17 20:12:42 +00:00
|
|
|
// Upper bound on how many sectors can be sealing+upgrading at the same time when creating new CC sectors (0 = unlimited)
|
2021-07-23 11:55:50 +00:00
|
|
|
MaxSealingSectors uint64
|
|
|
|
|
2022-03-17 20:12:42 +00:00
|
|
|
// Upper bound on how many sectors can be sealing+upgrading at the same time when creating new sectors with deals (0 = unlimited)
|
2021-07-23 11:55:50 +00:00
|
|
|
MaxSealingSectorsForDeals uint64
|
|
|
|
|
2022-03-17 20:27:10 +00:00
|
|
|
// Prefer creating new sectors even if there are sectors Available for upgrading.
|
2022-03-17 20:12:42 +00:00
|
|
|
// This setting combined with MaxUpgradingSectors set to a value higher than MaxSealingSectorsForDeals makes it
|
|
|
|
// possible to use fast sector upgrades to handle high volumes of storage deals, while still using the simple sealing
|
|
|
|
// flow when the volume of storage deals is lower.
|
|
|
|
PreferNewSectorsForDeals bool
|
|
|
|
|
|
|
|
// Upper bound on how many sectors can be sealing+upgrading at the same time when upgrading CC sectors with deals (0 = MaxSealingSectorsForDeals)
|
|
|
|
MaxUpgradingSectors uint64
|
|
|
|
|
2022-09-14 10:45:22 +00:00
|
|
|
// When set to a non-zero value, minimum number of epochs until sector expiration required for sectors to be considered
|
|
|
|
// for upgrades (0 = DealMinDuration = 180 days = 518400 epochs)
|
|
|
|
//
|
|
|
|
// Note that if all deals waiting in the input queue have lifetimes longer than this value, upgrade sectors will be
|
|
|
|
// required to have expiration of at least the soonest-ending deal
|
|
|
|
MinUpgradeSectorExpiration uint64
|
|
|
|
|
2023-05-23 19:34:27 +00:00
|
|
|
// DEPRECATED: Target expiration is no longer used
|
2022-09-14 10:45:22 +00:00
|
|
|
MinTargetUpgradeSectorExpiration uint64
|
|
|
|
|
2021-07-24 00:36:29 +00:00
|
|
|
// CommittedCapacitySectorLifetime is the duration a Committed Capacity (CC) sector will
|
|
|
|
// live before it must be extended or converted into sector containing deals before it is
|
2023-10-11 15:30:30 +00:00
|
|
|
// terminated. Value must be between 180-1278 days (1278 in nv21, 540 before nv21).
|
2021-07-24 00:36:29 +00:00
|
|
|
CommittedCapacitySectorLifetime Duration
|
|
|
|
|
2021-07-23 13:40:30 +00:00
|
|
|
// Period of time that a newly created sector will wait for more deals to be packed in to before it starts to seal.
|
|
|
|
// Sectors which are fully filled will start sealing immediately
|
2021-07-23 11:55:50 +00:00
|
|
|
WaitDealsDelay Duration
|
|
|
|
|
2021-07-23 13:40:30 +00:00
|
|
|
// Whether to keep unsealed copies of deal data regardless of whether the client requested that. This lets the miner
|
|
|
|
// avoid the relatively high cost of unsealing the data later, at the cost of more storage space
|
2021-07-23 11:55:50 +00:00
|
|
|
AlwaysKeepUnsealedCopy bool
|
|
|
|
|
|
|
|
// Run sector finalization before submitting sector proof to the chain
|
|
|
|
FinalizeEarly bool
|
|
|
|
|
2022-03-26 19:50:21 +00:00
|
|
|
// Whether new sectors are created to pack incoming deals
|
|
|
|
// When this is set to false no new sectors will be created for sealing incoming deals
|
|
|
|
// This is useful for forcing all deals to be assigned as snap deals to sectors marked for upgrade
|
|
|
|
MakeNewSectorForDeals bool
|
|
|
|
|
2022-03-16 18:29:47 +00:00
|
|
|
// After sealing CC sectors, make them available for upgrading with deals
|
|
|
|
MakeCCSectorsAvailable bool
|
|
|
|
|
2021-07-23 11:55:50 +00:00
|
|
|
// Whether to use available miner balance for sector collateral instead of sending it with each message
|
|
|
|
CollateralFromMinerBalance bool
|
|
|
|
// Minimum available balance to keep in the miner actor before sending it with messages
|
|
|
|
AvailableBalanceBuffer types.FIL
|
|
|
|
// Don't send collateral with messages even if there is no available balance in the miner actor
|
|
|
|
DisableCollateralFallback bool
|
|
|
|
|
|
|
|
// maximum precommit batch size - batches will be sent immediately above this size
|
|
|
|
MaxPreCommitBatch int
|
|
|
|
// how long to wait before submitting a batch after crossing the minimum batch size
|
|
|
|
PreCommitBatchWait Duration
|
|
|
|
// time buffer for forceful batch submission before sectors/deal in batch would start expiring
|
|
|
|
PreCommitBatchSlack Duration
|
|
|
|
|
|
|
|
// enable / disable commit aggregation (takes effect after nv13)
|
|
|
|
AggregateCommits bool
|
2022-05-23 16:49:01 +00:00
|
|
|
// minimum batched commit size - batches above this size will eventually be sent on a timeout
|
2021-07-23 11:55:50 +00:00
|
|
|
MinCommitBatch int
|
2022-05-23 16:49:01 +00:00
|
|
|
// maximum batched commit size - batches will be sent immediately above this size
|
2021-07-23 11:55:50 +00:00
|
|
|
MaxCommitBatch int
|
|
|
|
// how long to wait before submitting a batch after crossing the minimum batch size
|
|
|
|
CommitBatchWait Duration
|
|
|
|
// time buffer for forceful batch submission before sectors/deals in batch would start expiring
|
|
|
|
CommitBatchSlack Duration
|
|
|
|
|
2021-09-30 14:53:12 +00:00
|
|
|
// network BaseFee below which to stop doing precommit batching, instead
|
2023-08-08 11:04:21 +00:00
|
|
|
// sending precommit messages to the chain individually. When the basefee is
|
|
|
|
// below this threshold, precommit messages will get sent out immediately.
|
2021-09-30 14:53:12 +00:00
|
|
|
BatchPreCommitAboveBaseFee types.FIL
|
|
|
|
|
2021-07-23 11:55:50 +00:00
|
|
|
// network BaseFee below which to stop doing commit aggregation, instead
|
|
|
|
// submitting proofs to the chain individually
|
|
|
|
AggregateAboveBaseFee types.FIL
|
|
|
|
|
2023-04-01 23:30:32 +00:00
|
|
|
// When submitting several sector prove commit messages simultaneously, this option allows you to
|
|
|
|
// stagger the number of prove commits submitted per epoch
|
|
|
|
// This is done because gas estimates for ProveCommits are non deterministic and increasing as a large
|
|
|
|
// number of sectors get committed within the same epoch resulting in occasionally failed msgs.
|
|
|
|
// Submitting a smaller number of prove commits per epoch would reduce the possibility of failed msgs
|
|
|
|
MaxSectorProveCommitsSubmittedPerEpoch uint64
|
|
|
|
|
2021-07-23 11:55:50 +00:00
|
|
|
TerminateBatchMax uint64
|
|
|
|
TerminateBatchMin uint64
|
|
|
|
TerminateBatchWait Duration
|
|
|
|
|
|
|
|
// Keep this many sectors in sealing pipeline, start CC if needed
|
|
|
|
// todo TargetSealingSectors uint64
|
|
|
|
|
|
|
|
// todo TargetSectors - stop auto-pleding new sectors after this many sectors are sealed, default CC upgrade for deals sectors if above
|
2023-05-29 23:11:52 +00:00
|
|
|
|
2023-09-26 06:22:25 +00:00
|
|
|
// UseSyntheticPoRep, when set to true, will reduce the amount of cache data held on disk after the completion of PreCommit 2 to 11GiB.
|
2023-07-12 11:51:50 +00:00
|
|
|
UseSyntheticPoRep bool
|
chore: Merge nv22 into master (#11699)
* [WIP] feat: Add nv22 skeleton
Addition of Network Version 22 skeleton
* update FFI
* feat: drand: refactor round verification
* feat: sealing: Support nv22 DDO features in the sealing pipeline (#11226)
* Initial work supporting DDO pieces in lotus-miner
* sealing: Update pipeline input to operate on UniversalPiece
* sealing: Update pipeline checks/sealing states to operate on UniversalPiece
* sealing: Make pipeline build with UniversalPiece
* move PieceDealInfo out of api
* make gen
* make sealing pipeline unit tests pass
* fix itest ensemble build
* don't panic in SectorsStatus with deals
* stop linter from complaining about checkPieces
* fix sector import tests
* mod tidy
* sealing: Add logic for (pre)committing DDO sectors
* sealing: state-types with method defs
* DDO non-snap pipeline works(?), DDO Itests
* DDO support in snapdeals pipeline
* make gen
* update actor bundles
* update the gst market fix
* fix: chain: use PreCommitSectorsBatch2 when setting up genesis
* some bug fixes
* integration working changes
* update actor bundles
* Make TestOnboardRawPieceSnap pass
* Appease the linter
* Make deadlines test pass with v12 actors
* Update go-state-types, abstract market DealState
* make gen
* mod tidy, lint fixes
* Fix some more tests
* Bump version in master
Bump version in master
* Make gen
Make gen
* fix sender
* fix: lotus-provider: Fix winning PoSt
* fix: sql Scan cannot write to an object
* Actually show miner-addrs in info-log
Actually show miner-addrs in lotus-provider info-log
* [WIP] feat: Add nv22 skeleton
Addition of Network Version 22 skeleton
* update FFI
* ddo is now nv22
* make gen
* temp actor bundle with ddo
* use working go-state-types
* gst with v13 market migration
* update bundle, builtin.MethodsMiner.ProveCommitSectors2 -> 3
* actually working v13 migration, v13 migration itest
* Address review
* sealing: Correct DDO snap pledge math
* itests: Mixed ddo itest
* pipeline: Fix sectorWeight
* sealing: convert market deals into PAMs in mixed sectors
* sealing: make market to ddo conversion work
* fix lint
* update gst
* Update actors and GST to lastest integ branch
* commit batcher: Update ProveCommitSectors3Params builder logic
* make gen
* use builtin-actors master
* ddo: address review
* itests: Add commd assertions to ddo tests
* make gen
* gst with fixed types
* config knobs for RequireActivationSuccess
* storage: Drop obsolete flaky tasts
---------
Co-authored-by: Jennifer Wang <jiayingw703@gmail.com>
Co-authored-by: Aayush <arajasek94@gmail.com>
Co-authored-by: Shrenuj Bansal <shrenuj.bansal@protocol.ai>
Co-authored-by: Phi <orjan.roren@gmail.com>
Co-authored-by: Andrew Jackson (Ajax) <snadrus@gmail.com>
Co-authored-by: TippyFlits <james.bluett@protocol.ai>
* feat: implement FIP-0063
* chore: deps: update to go-multiaddr v0.12.2 (#11602)
* feat: fvm: update the FVM/FFI to v4.1 (#11608) (#11612)
This:
1. Adds nv22 support.
2. Updates the message tracing format.
Co-authored-by: Steven Allen <steven@stebalien.com>
* AggregateProofType nil when doing batch updates
Use latest nv22 go-state-types version with matching update
* Update to v13.0.0-rc.2 bundle
* chore: Upgrade heights and codename
Update upgrade heights
Co-Authored-By: Steven Allen <steven@stebalien.com>
* Update epoch after nv22 DRAND switch
Update epoch after nv22 DRAND switch
* Update Mango codename to Phoneix
Make the codename for the Drand-change inline with Dragon style.
* Add UpgradePhoenixHeight to API params
* set UpgradePhoenixHeight to be one hour after Dragon
* Make gen
Make gen and UpgradePhoenixHeight in butterfly and local devnet to be in line with Calibration and Mainnet
* Update epoch heights (#11637)
Update epoch heights
* new: add forest bootstrap nodes (#11636)
Signed-off-by: samuelarogbonlo <sbayo971@gmail.com>
* Merge pull request #11491 from filecoin-project/fix/remove-decommissioned-pl-bootstrap-nodes
Remove PL operated bootstrap nodes from mainnet.pi
* feat: api: new verified registry methods to get all allocations and claims (#11631)
* new verireg methods
* update changelog and add itest
* update itest and cli
* update new method's support till v9
* remove gateway APIs
* fix cli internal var names
* chore:: backport #11609 to the feat/nv22 branch (#11644)
* feat: api: improve the correctness of Eth's trace_block (#11609)
* Improve the correctness of Eth's trace_block
- Improve encoding/decoding of parameters and return values:
- Encode "native" parameters and return values with Solidity ABI.
- Correctly decode parameters to "create" calls.
- Use the correct (ish) output for "create" calls.
- Handle all forms of "create".
- Make robust with respect to reverts:
- Use the actor ID/address from the trace instead of looking it up in
the state-tree (may not exist in the state-tree due to a revert).
- Gracefully handle failed actor/contract creation.
- Improve performance:
- We avoid looking anything up in the state-tree when translating the
trace, which should significantly improve performance.
- Improve code readability:
- Remove all "backtracking" logic.
- Use an "environment" struct to store temporary state instead of
attaching it to the trace.
- Fix random bugs:
- Fix an allocation bug in the "address" logic (need to set the
capacity before modifying the slice).
- Improved error checking/handling.
- Use correct types for `trace_block` action/results (create, call, etc.).
- And use the correct types for Result/Action structs instead of reusing the same "Call" action every time.
- Improve error messages.
* Make gen
Make gen
---------
Co-authored-by: Steven Allen <steven@stebalien.com>
* fix: add UpgradePhoenixHeight to StateGetNetworkParams (#11648)
* chore: deps: update to go-state-types v13.0.0-rc.1
* do NOT update the cache when running the real migration
* Merge pull request #11632 from hanabi1224/hm/drand-test
feat: drand quicknet: allow scheduling drand quicknet upgrade before nv22 on 2k devnet
* chore: deps: update to go-state-types v13.0.0-rc.2
chore: deps: update to go-state-types v13.0.0-rc.2
* feat: set migration config UpgradeEpoch for v13 actors upgrade
* Built-in actor events first draft
* itest for DDO non-market verified data w/ builtin actor events
* Tests for builtin actor events API
* Clean up DDO+Events tests, add lots of explainer comments
* Minor tweaks to events types
* Avoid duplicate messages when looking for receipts
* Rename internal events modules for clarity
* Adjust actor event API after review
* s/ActorEvents/Events/g in global config
* Manage event sending rate for SubscribeActorEvents
* Terminate SubscribeActorEvents chan when at max height
* Document future API changes
* More clarity in actor event API docs
* More post-review changes, lots of tests for SubscribeActorEvents
Use BlockDelay as the window for receiving events on the SubscribeActorEvents
channel. We expect the user to have received the initial batch of historical
events (if any) in one block's time. For real-time events we expect them to
not fall behind by roughly one block's time.
* Remove duplicate code from actor event type marshalling tests
Reduce verbosity and remove duplicate test logic from actor event types
JSON marshalling tests.
* Rename actor events test to follow go convention
Add missing `s` to `actor_events` test file to follow golang convention
used across the repo.
* Run actor events table tests in deterministic order
Refactor `map` usage for actor event table tests to ensure deterministic
test execution order, making debugging potential issues easier. If
non-determinism is a target, leverage Go's built-in parallel testing
capabilities.
* Reduce scope for filter removal failure when getting actor events
Use a fresh context to remove the temporary filter installed solely to
get the actor events. This should reduce chances of failure in a case
where the original context may be expired/cancelled.
Refactor removal into a `defer` statement for a more readable, concise
return statement.
* Use fixed RNG seed for actor event tests
Improve determinism in actor event tests by using a fixed RNG seed. This
makes up a more reproducible test suit.
* Use provided libraries to assert eventual conditions
Use the functionalities already provided by `testify` to assert eventual
conditions, and remove the use of `time.Sleep`.
Remove duplicate code in utility functions that are already defined.
Refactor assertion helper functions to use consistent terminology:
"require" implies fatal error, whereas "assert" implies error where the
test may proceed executing.
* Update changelog for actor events APIs
* Fix concerns and docs identified by review
* Update actor bundle to v13.0.0-rc3
Update actor bundle to v13.0.0-rc3
* Prep Lotus v1.26.0-rc1
- For sanity reverting the mainnet upgrade epoch to 99999999, and then only set it when cutting the final release
-Update Calibnet CIDs to v13.0.0-rc3
- Add GetActorEvents, SubscribeActorEvents, GetAllClaims and GetAllAllocations methods to the changelog
Co-Authored-By: Jiaying Wang <42981373+jennijuju@users.noreply.github.com>
* Update CHANGELOG.md
Co-authored-by: Masih H. Derkani <m@derkani.org>
* Make gen
Make gen
* fix: beacon: validate drand change at nv16 correctly
* bump to v1.26.0-rc2
* test: cleanup ddo verified itest, extract steps to functions
also add allocation-removed event case
* test: extract verified DDO test to separate file, add more checks
* test: add additional actor events checks
* Add verification for "deal-activated" actor event
* docs(drand): document the meaning of "IsChained" (#11692)
* Resolve conflicts
I encountered multiple issues when trying to run make gen. And these changes fixed a couple of them:
- go mod tidy
- Remove RaftState/RaftLeader
- Revert `if ts.Height() > claim.TermMax+claim.TermStart || !cctx.IsSet("expired")` to the what is in the release/v1.26.0: `if tsHeight > val.TermMax || !expired`
* fixup imports, make jen
* Update version
Update version in master to v1.27.0-dev
* Update node/impl/full/dummy.go
Co-authored-by: Łukasz Magiera <magik6k@users.noreply.github.com>
* Adjust ListClaimsCmd
Adjust ListClaimsCmd according to review
---------
Signed-off-by: samuelarogbonlo <sbayo971@gmail.com>
Co-authored-by: TippyFlits <james.bluett@protocol.ai>
Co-authored-by: Aayush <arajasek94@gmail.com>
Co-authored-by: Łukasz Magiera <magik6k@users.noreply.github.com>
Co-authored-by: Jennifer Wang <jiayingw703@gmail.com>
Co-authored-by: Shrenuj Bansal <shrenuj.bansal@protocol.ai>
Co-authored-by: Andrew Jackson (Ajax) <snadrus@gmail.com>
Co-authored-by: Steven Allen <steven@stebalien.com>
Co-authored-by: Rod Vagg <rod@vagg.org>
Co-authored-by: Samuel Arogbonlo <47984109+samuelarogbonlo@users.noreply.github.com>
Co-authored-by: LexLuthr <88259624+LexLuthr@users.noreply.github.com>
Co-authored-by: tom123222 <160735201+tom123222@users.noreply.github.com>
Co-authored-by: Aarsh Shah <aarshkshah1992@gmail.com>
Co-authored-by: Masih H. Derkani <m@derkani.org>
Co-authored-by: Jiaying Wang <42981373+jennijuju@users.noreply.github.com>
2024-03-12 09:33:58 +00:00
|
|
|
|
|
|
|
// Whether to abort if any sector activation in a batch fails (newly sealed sectors, only with ProveCommitSectors3).
|
|
|
|
RequireActivationSuccess bool
|
|
|
|
// Whether to abort if any piece activation notification returns a non-zero exit code (newly sealed sectors, only with ProveCommitSectors3).
|
|
|
|
RequireActivationSuccessUpdate bool
|
|
|
|
// Whether to abort if any sector activation in a batch fails (updating sectors, only with ProveReplicaUpdates3).
|
|
|
|
RequireNotificationSuccess bool
|
|
|
|
// Whether to abort if any piece activation notification returns a non-zero exit code (updating sectors, only with ProveReplicaUpdates3).
|
|
|
|
RequireNotificationSuccessUpdate bool
|
2021-07-23 11:55:50 +00:00
|
|
|
}
|
|
|
|
|
2022-03-29 01:19:11 +00:00
|
|
|
type SealerConfig struct {
|
|
|
|
ParallelFetchLimit int
|
|
|
|
|
2022-09-06 09:06:30 +00:00
|
|
|
AllowSectorDownload bool
|
2022-03-29 01:19:11 +00:00
|
|
|
AllowAddPiece bool
|
|
|
|
AllowPreCommit1 bool
|
|
|
|
AllowPreCommit2 bool
|
|
|
|
AllowCommit bool
|
|
|
|
AllowUnseal bool
|
|
|
|
AllowReplicaUpdate bool
|
|
|
|
AllowProveReplicaUpdate2 bool
|
|
|
|
AllowRegenSectorKey bool
|
|
|
|
|
2022-08-03 10:54:32 +00:00
|
|
|
// LocalWorkerName specifies a custom name for the builtin worker.
|
|
|
|
// If set to an empty string (default) os hostname will be used
|
|
|
|
LocalWorkerName string
|
|
|
|
|
2022-05-23 14:58:43 +00:00
|
|
|
// Assigner specifies the worker assigner to use when scheduling tasks.
|
|
|
|
// "utilization" (default) - assign tasks to workers with lowest utilization.
|
|
|
|
// "spread" - assign tasks to as many distinct workers as possible.
|
|
|
|
Assigner string
|
|
|
|
|
2022-05-23 21:53:25 +00:00
|
|
|
// DisallowRemoteFinalize when set to true will force all Finalize tasks to
|
|
|
|
// run on workers with local access to both long-term storage and the sealing
|
|
|
|
// path containing the sector.
|
2022-05-23 23:33:56 +00:00
|
|
|
// --
|
2022-05-23 21:53:25 +00:00
|
|
|
// WARNING: Only set this if all workers have access to long-term storage
|
|
|
|
// paths. If this flag is enabled, and there are workers without long-term
|
|
|
|
// storage access, sectors will not be moved from them, and Finalize tasks
|
|
|
|
// will appear to be stuck.
|
2022-05-23 23:33:56 +00:00
|
|
|
// --
|
2022-05-23 21:53:25 +00:00
|
|
|
// If you see stuck Finalize tasks after enabling this setting, check
|
|
|
|
// 'lotus-miner sealing sched-diag' and 'lotus-miner storage find [sector num]'
|
|
|
|
DisallowRemoteFinalize bool
|
|
|
|
|
2022-03-29 01:19:11 +00:00
|
|
|
// ResourceFiltering instructs the system which resource filtering strategy
|
|
|
|
// to use when evaluating tasks against this worker. An empty value defaults
|
|
|
|
// to "hardware".
|
2022-10-31 17:15:09 +00:00
|
|
|
ResourceFiltering ResourceFilteringStrategy
|
2022-03-29 01:19:11 +00:00
|
|
|
}
|
|
|
|
|
2021-07-23 11:55:50 +00:00
|
|
|
type BatchFeeConfig struct {
|
|
|
|
Base types.FIL
|
|
|
|
PerSector types.FIL
|
|
|
|
}
|
|
|
|
|
|
|
|
type MinerFeeConfig struct {
|
|
|
|
MaxPreCommitGasFee types.FIL
|
|
|
|
MaxCommitGasFee types.FIL
|
|
|
|
|
|
|
|
// maxBatchFee = maxBase + maxPerSector * nSectors
|
|
|
|
MaxPreCommitBatchGasFee BatchFeeConfig
|
|
|
|
MaxCommitBatchGasFee BatchFeeConfig
|
|
|
|
|
2021-07-23 13:40:30 +00:00
|
|
|
MaxTerminateGasFee types.FIL
|
|
|
|
// WindowPoSt is a high-value operation, so the default fee should be high.
|
2021-07-23 11:55:50 +00:00
|
|
|
MaxWindowPoStGasFee types.FIL
|
|
|
|
MaxPublishDealsFee types.FIL
|
|
|
|
MaxMarketBalanceAddFee types.FIL
|
2022-11-29 10:25:41 +00:00
|
|
|
|
|
|
|
MaximizeWindowPoStFeeCap bool
|
2021-07-23 11:55:50 +00:00
|
|
|
}
|
|
|
|
|
2024-03-15 21:38:13 +00:00
|
|
|
type CurioFees struct {
|
2023-09-20 03:48:39 +00:00
|
|
|
DefaultMaxFee types.FIL
|
|
|
|
MaxPreCommitGasFee types.FIL
|
|
|
|
MaxCommitGasFee types.FIL
|
|
|
|
|
|
|
|
// maxBatchFee = maxBase + maxPerSector * nSectors
|
|
|
|
MaxPreCommitBatchGasFee BatchFeeConfig
|
|
|
|
MaxCommitBatchGasFee BatchFeeConfig
|
|
|
|
|
|
|
|
MaxTerminateGasFee types.FIL
|
|
|
|
// WindowPoSt is a high-value operation, so the default fee should be high.
|
|
|
|
MaxWindowPoStGasFee types.FIL
|
|
|
|
MaxPublishDealsFee types.FIL
|
|
|
|
}
|
2021-07-23 11:55:50 +00:00
|
|
|
type MinerAddressConfig struct {
|
2021-07-23 13:40:30 +00:00
|
|
|
// Addresses to send PreCommit messages from
|
|
|
|
PreCommitControl []string
|
|
|
|
// Addresses to send Commit messages from
|
2021-07-23 11:55:50 +00:00
|
|
|
CommitControl []string
|
|
|
|
TerminateControl []string
|
|
|
|
DealPublishControl []string
|
|
|
|
|
|
|
|
// DisableOwnerFallback disables usage of the owner address for messages
|
|
|
|
// sent automatically
|
|
|
|
DisableOwnerFallback bool
|
|
|
|
// DisableWorkerFallback disables usage of the worker address for messages
|
|
|
|
// sent automatically, if control addresses are configured.
|
|
|
|
// A control address that doesn't have enough funds will still be chosen
|
|
|
|
// over the worker address if this flag is set.
|
|
|
|
DisableWorkerFallback bool
|
|
|
|
}
|
|
|
|
|
2024-03-15 21:38:13 +00:00
|
|
|
type CurioAddresses struct {
|
2023-09-20 03:48:39 +00:00
|
|
|
// Addresses to send PreCommit messages from
|
|
|
|
PreCommitControl []string
|
|
|
|
// Addresses to send Commit messages from
|
|
|
|
CommitControl []string
|
|
|
|
TerminateControl []string
|
|
|
|
|
|
|
|
// DisableOwnerFallback disables usage of the owner address for messages
|
2021-07-23 11:55:50 +00:00
|
|
|
// sent automatically
|
|
|
|
DisableOwnerFallback bool
|
|
|
|
// DisableWorkerFallback disables usage of the worker address for messages
|
|
|
|
// sent automatically, if control addresses are configured.
|
|
|
|
// A control address that doesn't have enough funds will still be chosen
|
|
|
|
// over the worker address if this flag is set.
|
|
|
|
DisableWorkerFallback bool
|
2023-10-25 00:26:13 +00:00
|
|
|
|
|
|
|
// MinerAddresses are the addresses of the miner actors to use for sending messages
|
|
|
|
MinerAddresses []string
|
2021-07-23 11:55:50 +00:00
|
|
|
}
|
|
|
|
|
2024-04-02 16:21:21 +00:00
|
|
|
type CurioProvingConfig struct {
|
|
|
|
// Maximum number of sector checks to run in parallel. (0 = unlimited)
|
|
|
|
//
|
|
|
|
// WARNING: Setting this value too high may make the node crash by running out of stack
|
|
|
|
// WARNING: Setting this value too low may make sector challenge reading much slower, resulting in failed PoSt due
|
|
|
|
// to late submission.
|
|
|
|
//
|
|
|
|
// After changing this option, confirm that the new value works in your setup by invoking
|
|
|
|
// 'lotus-miner proving compute window-post 0'
|
|
|
|
ParallelCheckLimit int
|
|
|
|
|
|
|
|
// Maximum amount of time a proving pre-check can take for a sector. If the check times out the sector will be skipped
|
|
|
|
//
|
|
|
|
// WARNING: Setting this value too low risks in sectors being skipped even though they are accessible, just reading the
|
|
|
|
// test challenge took longer than this timeout
|
|
|
|
// WARNING: Setting this value too high risks missing PoSt deadline in case IO operations related to this sector are
|
|
|
|
// blocked (e.g. in case of disconnected NFS mount)
|
|
|
|
SingleCheckTimeout Duration
|
|
|
|
|
|
|
|
// Maximum amount of time a proving pre-check can take for an entire partition. If the check times out, sectors in
|
|
|
|
// the partition which didn't get checked on time will be skipped
|
|
|
|
//
|
|
|
|
// WARNING: Setting this value too low risks in sectors being skipped even though they are accessible, just reading the
|
|
|
|
// test challenge took longer than this timeout
|
|
|
|
// WARNING: Setting this value too high risks missing PoSt deadline in case IO operations related to this partition are
|
|
|
|
// blocked or slow
|
|
|
|
PartitionCheckTimeout Duration
|
|
|
|
|
|
|
|
// Disable WindowPoSt provable sector readability checks.
|
|
|
|
//
|
|
|
|
// In normal operation, when preparing to compute WindowPoSt, lotus-miner will perform a round of reading challenges
|
|
|
|
// from all sectors to confirm that those sectors can be proven. Challenges read in this process are discarded, as
|
|
|
|
// we're only interested in checking that sector data can be read.
|
|
|
|
//
|
|
|
|
// When using builtin proof computation (no PoSt workers, and DisableBuiltinWindowPoSt is set to false), this process
|
|
|
|
// can save a lot of time and compute resources in the case that some sectors are not readable - this is caused by
|
|
|
|
// the builtin logic not skipping snark computation when some sectors need to be skipped.
|
|
|
|
//
|
|
|
|
// When using PoSt workers, this process is mostly redundant, with PoSt workers challenges will be read once, and
|
|
|
|
// if challenges for some sectors aren't readable, those sectors will just get skipped.
|
|
|
|
//
|
|
|
|
// Disabling sector pre-checks will slightly reduce IO load when proving sectors, possibly resulting in shorter
|
|
|
|
// time to produce window PoSt. In setups with good IO capabilities the effect of this option on proving time should
|
|
|
|
// be negligible.
|
|
|
|
//
|
|
|
|
// NOTE: It likely is a bad idea to disable sector pre-checks in setups with no PoSt workers.
|
|
|
|
//
|
|
|
|
// NOTE: Even when this option is enabled, recovering sectors will be checked before recovery declaration message is
|
|
|
|
// sent to the chain
|
|
|
|
//
|
|
|
|
// After changing this option, confirm that the new value works in your setup by invoking
|
|
|
|
// 'lotus-miner proving compute window-post 0'
|
|
|
|
DisableWDPoStPreChecks bool
|
|
|
|
|
|
|
|
// Maximum number of partitions to prove in a single SubmitWindowPoSt messace. 0 = network limit (3 in nv21)
|
|
|
|
//
|
|
|
|
// A single partition may contain up to 2349 32GiB sectors, or 2300 64GiB sectors.
|
|
|
|
// //
|
|
|
|
// Note that setting this value lower may result in less efficient gas use - more messages will be sent,
|
|
|
|
// to prove each deadline, resulting in more total gas use (but each message will have lower gas limit)
|
|
|
|
//
|
|
|
|
// Setting this value above the network limit has no effect
|
|
|
|
MaxPartitionsPerPoStMessage int
|
|
|
|
|
|
|
|
// Maximum number of partitions to declare in a single DeclareFaultsRecovered message. 0 = no limit.
|
|
|
|
|
|
|
|
// In some cases when submitting DeclareFaultsRecovered messages,
|
|
|
|
// there may be too many recoveries to fit in a BlockGasLimit.
|
|
|
|
// In those cases it may be necessary to set this value to something low (eg 1);
|
|
|
|
// Note that setting this value lower may result in less efficient gas use - more messages will be sent than needed,
|
|
|
|
// resulting in more total gas use (but each message will have lower gas limit)
|
|
|
|
MaxPartitionsPerRecoveryMessage int
|
|
|
|
|
|
|
|
// Enable single partition per PoSt Message for partitions containing recovery sectors
|
|
|
|
//
|
|
|
|
// In cases when submitting PoSt messages which contain recovering sectors, the default network limit may still be
|
|
|
|
// too high to fit in the block gas limit. In those cases, it becomes useful to only house the single partition
|
|
|
|
// with recovering sectors in the post message
|
|
|
|
//
|
|
|
|
// Note that setting this value lower may result in less efficient gas use - more messages will be sent,
|
|
|
|
// to prove each deadline, resulting in more total gas use (but each message will have lower gas limit)
|
|
|
|
SingleRecoveringPartitionPerPostMessage bool
|
|
|
|
}
|
|
|
|
|
2024-04-12 02:36:38 +00:00
|
|
|
type CurioIngestConfig struct {
|
|
|
|
// Maximum number of sectors that can be queued waiting for SDR to start processing.
|
|
|
|
// 0 = unlimited
|
|
|
|
// Note: This mechanism will delay taking deal data from markets, providing backpressure to the market subsystem.
|
|
|
|
// The SDR queue includes deals which are in the process of entering the sealing pipeline - size of this queue
|
|
|
|
// will also impact the maximum number of ParkPiece tasks which can run concurrently.
|
|
|
|
//
|
|
|
|
// SDR queue is the first queue in the sealing pipeline, meaning that it should be used as the primary backpressure mechanism.
|
|
|
|
MaxQueueSDR int
|
|
|
|
|
|
|
|
// Maximum number of sectors that can be queued waiting for SDRTrees to start processing.
|
|
|
|
// 0 = unlimited
|
|
|
|
// Note: This mechanism will delay taking deal data from markets, providing backpressure to the market subsystem.
|
|
|
|
// In case of the trees tasks it is possible that this queue grows more than this limit, the backpressure is only
|
|
|
|
// applied to sectors entering the pipeline.
|
|
|
|
MaxQueueTrees int
|
|
|
|
|
|
|
|
// Maximum number of sectors that can be queued waiting for PoRep to start processing.
|
|
|
|
// 0 = unlimited
|
|
|
|
// Note: This mechanism will delay taking deal data from markets, providing backpressure to the market subsystem.
|
|
|
|
// Like with the trees tasks, it is possible that this queue grows more than this limit, the backpressure is only
|
|
|
|
// applied to sectors entering the pipeline.
|
|
|
|
MaxQueuePoRep int
|
|
|
|
}
|
|
|
|
|
2021-07-23 11:55:50 +00:00
|
|
|
// API contains configs for API endpoint
|
|
|
|
type API struct {
|
2021-07-23 12:55:19 +00:00
|
|
|
// Binding address for the Lotus API
|
2021-07-23 11:55:50 +00:00
|
|
|
ListenAddress string
|
|
|
|
RemoteListenAddress string
|
|
|
|
Timeout Duration
|
|
|
|
}
|
|
|
|
|
|
|
|
// Libp2p contains configs for libp2p
|
|
|
|
type Libp2p struct {
|
2021-07-23 13:40:30 +00:00
|
|
|
// Binding address for the libp2p host - 0 means random port.
|
|
|
|
// Format: multiaddress; see https://multiformats.io/multiaddr/
|
|
|
|
ListenAddresses []string
|
|
|
|
// Addresses to explicitally announce to other peers. If not specified,
|
|
|
|
// all interface addresses are announced
|
|
|
|
// Format: multiaddress
|
|
|
|
AnnounceAddresses []string
|
|
|
|
// Addresses to not announce
|
|
|
|
// Format: multiaddress
|
2021-07-23 11:55:50 +00:00
|
|
|
NoAnnounceAddresses []string
|
|
|
|
BootstrapPeers []string
|
|
|
|
ProtectedPeers []string
|
|
|
|
|
2021-08-27 17:14:29 +00:00
|
|
|
// When not disabled (default), lotus asks NAT devices (e.g., routers), to
|
|
|
|
// open up an external port and forward it to the port lotus is running on.
|
|
|
|
// When this works (i.e., when your router supports NAT port forwarding),
|
|
|
|
// it makes the local lotus node accessible from the public internet
|
|
|
|
DisableNatPortMap bool
|
|
|
|
|
2021-08-27 17:16:50 +00:00
|
|
|
// ConnMgrLow is the number of connections that the basic connection manager
|
|
|
|
// will trim down to.
|
|
|
|
ConnMgrLow uint
|
|
|
|
// ConnMgrHigh is the number of connections that, when exceeded, will trigger
|
|
|
|
// a connection GC operation. Note: protected/recently formed connections don't
|
|
|
|
// count towards this limit.
|
|
|
|
ConnMgrHigh uint
|
|
|
|
// ConnMgrGrace is a time duration that new connections are immune from being
|
|
|
|
// closed by the connection manager.
|
2021-07-23 11:55:50 +00:00
|
|
|
ConnMgrGrace Duration
|
|
|
|
}
|
|
|
|
|
|
|
|
type Pubsub struct {
|
2021-07-23 13:40:30 +00:00
|
|
|
// Run the node in bootstrap-node mode
|
|
|
|
Bootstrapper bool
|
|
|
|
// DirectPeers specifies peers with direct peering agreements. These peers are
|
|
|
|
// connected outside of the mesh, with all (valid) message unconditionally
|
|
|
|
// forwarded to them. The router will maintain open connections to these peers.
|
|
|
|
// Note that the peering agreement should be reciprocal with direct peers
|
|
|
|
// symmetrically configured at both ends.
|
|
|
|
// Type: Array of multiaddress peerinfo strings, must include peerid (/p2p/12D3K...
|
2021-07-23 11:55:50 +00:00
|
|
|
DirectPeers []string
|
|
|
|
IPColocationWhitelist []string
|
|
|
|
RemoteTracer string
|
2021-09-29 10:51:43 +00:00
|
|
|
// Path to file that will be used to output tracer content in JSON format.
|
|
|
|
// If present tracer will save data to defined file.
|
|
|
|
// Format: file path
|
|
|
|
JsonTracer string
|
|
|
|
// Connection string for elasticsearch instance.
|
|
|
|
// If present tracer will save data to elasticsearch.
|
|
|
|
// Format: https://<username>:<password>@<elasticsearch_url>:<port>/
|
|
|
|
ElasticSearchTracer string
|
|
|
|
// Name of elasticsearch index that will be used to save tracer data.
|
|
|
|
// This property is used only if ElasticSearchTracer propery is set.
|
|
|
|
ElasticSearchIndex string
|
|
|
|
// Auth token that will be passed with logs to elasticsearch - used for weighted peers score.
|
|
|
|
TracerSourceAuth string
|
2021-07-23 11:55:50 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
type Chainstore struct {
|
|
|
|
EnableSplitstore bool
|
|
|
|
Splitstore Splitstore
|
|
|
|
}
|
|
|
|
|
|
|
|
type Splitstore struct {
|
2021-07-24 05:59:15 +00:00
|
|
|
// ColdStoreType specifies the type of the coldstore.
|
2023-09-22 13:24:55 +00:00
|
|
|
// It can be "discard" (default) for discarding cold blocks, "messages" to store only messages or "universal" to store all chain state..
|
2021-07-23 11:55:50 +00:00
|
|
|
ColdStoreType string
|
2021-07-24 05:59:15 +00:00
|
|
|
// HotStoreType specifies the type of the hotstore.
|
|
|
|
// Only currently supported value is "badger".
|
|
|
|
HotStoreType string
|
|
|
|
// MarkSetType specifies the type of the markset.
|
2022-02-06 09:21:16 +00:00
|
|
|
// It can be "map" for in memory marking or "badger" (default) for on-disk marking.
|
2021-07-24 05:59:15 +00:00
|
|
|
MarkSetType string
|
|
|
|
|
|
|
|
// HotStoreMessageRetention specifies the retention policy for messages, in finalities beyond
|
|
|
|
// the compaction boundary; default is 0.
|
2021-07-23 11:55:50 +00:00
|
|
|
HotStoreMessageRetention uint64
|
2021-07-27 06:53:22 +00:00
|
|
|
// HotStoreFullGCFrequency specifies how often to perform a full (moving) GC on the hotstore.
|
|
|
|
// A value of 0 disables, while a value 1 will do full GC in every compaction.
|
2021-07-24 05:59:15 +00:00
|
|
|
// Default is 20 (about once a week).
|
2021-07-27 06:53:22 +00:00
|
|
|
HotStoreFullGCFrequency uint64
|
2023-03-07 14:38:27 +00:00
|
|
|
// HotStoreMaxSpaceTarget sets a target max disk size for the hotstore. Splitstore GC
|
|
|
|
// will run moving GC if disk utilization gets within a threshold (150 GB) of the target.
|
|
|
|
// Splitstore GC will NOT run moving GC if the total size of the move would get
|
|
|
|
// within 50 GB of the target, and instead will run a more aggressive online GC.
|
|
|
|
// If both HotStoreFullGCFrequency and HotStoreMaxSpaceTarget are set then splitstore
|
|
|
|
// GC will trigger moving GC if either configuration condition is met.
|
2023-03-09 01:11:39 +00:00
|
|
|
// A reasonable minimum is 2x fully GCed hotstore size + 50 G buffer.
|
|
|
|
// At this minimum size moving GC happens every time, any smaller and moving GC won't
|
|
|
|
// be able to run. In spring 2023 this minimum is ~550 GB.
|
2023-03-07 14:38:27 +00:00
|
|
|
HotStoreMaxSpaceTarget uint64
|
2023-03-09 15:40:14 +00:00
|
|
|
|
|
|
|
// When HotStoreMaxSpaceTarget is set Moving GC will be triggered when total moving size
|
|
|
|
// exceeds HotstoreMaxSpaceTarget - HotstoreMaxSpaceThreshold
|
|
|
|
HotStoreMaxSpaceThreshold uint64
|
|
|
|
|
|
|
|
// Safety buffer to prevent moving GC from overflowing disk when HotStoreMaxSpaceTarget
|
|
|
|
// is set. Moving GC will not occur when total moving size exceeds
|
|
|
|
// HotstoreMaxSpaceTarget - HotstoreMaxSpaceSafetyBuffer
|
|
|
|
HotstoreMaxSpaceSafetyBuffer uint64
|
2021-07-23 11:55:50 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// // Full Node
|
|
|
|
type Wallet struct {
|
|
|
|
RemoteBackend string
|
|
|
|
EnableLedger bool
|
|
|
|
DisableLocal bool
|
|
|
|
}
|
|
|
|
|
|
|
|
type FeeConfig struct {
|
|
|
|
DefaultMaxFee types.FIL
|
|
|
|
}
|
2022-09-13 17:05:48 +00:00
|
|
|
|
2023-01-19 22:44:58 +00:00
|
|
|
type FevmConfig struct {
|
|
|
|
// EnableEthRPC enables eth_ rpc, and enables storing a mapping of eth transaction hashes to filecoin message Cids.
|
|
|
|
// This will also enable the RealTimeFilterAPI and HistoricFilterAPI by default, but they can be disabled by config options above.
|
|
|
|
EnableEthRPC bool
|
|
|
|
|
|
|
|
// EthTxHashMappingLifetimeDays the transaction hash lookup database will delete mappings that have been stored for more than x days
|
|
|
|
// Set to 0 to keep all mappings
|
|
|
|
EthTxHashMappingLifetimeDays int
|
|
|
|
|
2024-03-08 07:43:39 +00:00
|
|
|
Events DeprecatedEvents `toml:"Events,omitempty"`
|
2023-01-19 22:44:58 +00:00
|
|
|
}
|
|
|
|
|
2024-03-08 07:43:39 +00:00
|
|
|
type DeprecatedEvents struct {
|
|
|
|
// DisableRealTimeFilterAPI is DEPRECATED and will be removed in a future release. Use Events.DisableRealTimeFilterAPI instead.
|
|
|
|
DisableRealTimeFilterAPI bool `moved:"Events.DisableRealTimeFilterAPI" toml:"DisableRealTimeFilterAPI,omitempty"`
|
|
|
|
|
|
|
|
// DisableHistoricFilterAPI is DEPRECATED and will be removed in a future release. Use Events.DisableHistoricFilterAPI instead.
|
|
|
|
DisableHistoricFilterAPI bool `moved:"Events.DisableHistoricFilterAPI" toml:"DisableHistoricFilterAPI,omitempty"`
|
|
|
|
|
|
|
|
// FilterTTL is DEPRECATED and will be removed in a future release. Use Events.FilterTTL instead.
|
|
|
|
FilterTTL Duration `moved:"Events.FilterTTL" toml:"FilterTTL,omitzero"`
|
|
|
|
|
|
|
|
// MaxFilters is DEPRECATED and will be removed in a future release. Use Events.MaxFilters instead.
|
|
|
|
MaxFilters int `moved:"Events.MaxFilters" toml:"MaxFilters,omitzero"`
|
|
|
|
|
|
|
|
// MaxFilterResults is DEPRECATED and will be removed in a future release. Use Events.MaxFilterResults instead.
|
|
|
|
MaxFilterResults int `moved:"Events.MaxFilterResults" toml:"MaxFilterResults,omitzero"`
|
|
|
|
|
|
|
|
// MaxFilterHeightRange is DEPRECATED and will be removed in a future release. Use Events.MaxFilterHeightRange instead.
|
|
|
|
MaxFilterHeightRange uint64 `moved:"Events.MaxFilterHeightRange" toml:"MaxFilterHeightRange,omitzero"`
|
|
|
|
|
|
|
|
// DatabasePath is DEPRECATED and will be removed in a future release. Use Events.DatabasePath instead.
|
|
|
|
DatabasePath string `moved:"Events.DatabasePath" toml:"DatabasePath,omitempty"`
|
|
|
|
}
|
|
|
|
|
|
|
|
type EventsConfig struct {
|
2023-01-19 22:44:58 +00:00
|
|
|
// DisableRealTimeFilterAPI will disable the RealTimeFilterAPI that can create and query filters for actor events as they are emitted.
|
2024-03-08 07:43:39 +00:00
|
|
|
// The API is enabled when Fevm.EnableEthRPC or EnableActorEventsAPI is true, but can be disabled selectively with this flag.
|
2023-01-19 22:44:58 +00:00
|
|
|
DisableRealTimeFilterAPI bool
|
2023-01-13 19:11:13 +00:00
|
|
|
|
2023-01-19 22:44:58 +00:00
|
|
|
// DisableHistoricFilterAPI will disable the HistoricFilterAPI that can create and query filters for actor events
|
|
|
|
// that occurred in the past. HistoricFilterAPI maintains a queryable index of events.
|
2024-03-08 07:43:39 +00:00
|
|
|
// The API is enabled when Fevm.EnableEthRPC or EnableActorEventsAPI is true, but can be disabled selectively with this flag.
|
2023-01-19 22:44:58 +00:00
|
|
|
DisableHistoricFilterAPI bool
|
2023-01-13 19:11:13 +00:00
|
|
|
|
2024-03-08 07:43:39 +00:00
|
|
|
// EnableActorEventsAPI enables the Actor events API that enables clients to consume events
|
|
|
|
// emitted by (smart contracts + built-in Actors).
|
|
|
|
// This will also enable the RealTimeFilterAPI and HistoricFilterAPI by default, but they can be
|
|
|
|
// disabled by setting their respective Disable* options.
|
|
|
|
EnableActorEventsAPI bool
|
|
|
|
|
2023-01-13 19:11:13 +00:00
|
|
|
// FilterTTL specifies the time to live for actor event filters. Filters that haven't been accessed longer than
|
|
|
|
// this time become eligible for automatic deletion.
|
|
|
|
FilterTTL Duration
|
|
|
|
|
|
|
|
// MaxFilters specifies the maximum number of filters that may exist at any one time.
|
|
|
|
MaxFilters int
|
|
|
|
|
|
|
|
// MaxFilterResults specifies the maximum number of results that can be accumulated by an actor event filter.
|
|
|
|
MaxFilterResults int
|
|
|
|
|
|
|
|
// MaxFilterHeightRange specifies the maximum range of heights that can be used in a filter (to avoid querying
|
|
|
|
// the entire chain)
|
|
|
|
MaxFilterHeightRange uint64
|
|
|
|
|
2023-01-19 22:44:58 +00:00
|
|
|
// DatabasePath is the full path to a sqlite database that will be used to index actor events to
|
2023-01-13 19:11:13 +00:00
|
|
|
// support the historic filter APIs. If the database does not exist it will be created. The directory containing
|
|
|
|
// the database must already exist and be writeable. If a relative path is provided here, sqlite treats it as
|
|
|
|
// relative to the CWD (current working directory).
|
2023-01-19 22:44:58 +00:00
|
|
|
DatabasePath string
|
2023-01-13 19:11:13 +00:00
|
|
|
|
|
|
|
// Others, not implemented yet:
|
|
|
|
// Set a limit on the number of active websocket subscriptions (may be zero)
|
|
|
|
// Set a timeout for subscription clients
|
|
|
|
// Set upper bound on index size
|
|
|
|
}
|
2023-03-13 10:01:04 +00:00
|
|
|
|
|
|
|
type IndexConfig struct {
|
2023-06-09 16:46:27 +00:00
|
|
|
// EXPERIMENTAL FEATURE. USE WITH CAUTION
|
2023-03-13 10:01:04 +00:00
|
|
|
// EnableMsgIndex enables indexing of messages on chain.
|
|
|
|
EnableMsgIndex bool
|
|
|
|
}
|
2023-08-01 15:28:47 +00:00
|
|
|
|
2023-07-14 23:05:49 +00:00
|
|
|
type HarmonyDB struct {
|
|
|
|
// HOSTS is a list of hostnames to nodes running YugabyteDB
|
|
|
|
// in a cluster. Only 1 is required
|
|
|
|
Hosts []string
|
|
|
|
|
|
|
|
// The Yugabyte server's username with full credentials to operate on Lotus' Database. Blank for default.
|
|
|
|
Username string
|
|
|
|
|
|
|
|
// The password for the related username. Blank for default.
|
|
|
|
Password string
|
|
|
|
|
|
|
|
// The database (logical partition) within Yugabyte. Blank for default.
|
|
|
|
Database string
|
|
|
|
|
|
|
|
// The port to find Yugabyte. Blank for default.
|
|
|
|
Port string
|
|
|
|
}
|
chore: Merge nv22 into master (#11699)
* [WIP] feat: Add nv22 skeleton
Addition of Network Version 22 skeleton
* update FFI
* feat: drand: refactor round verification
* feat: sealing: Support nv22 DDO features in the sealing pipeline (#11226)
* Initial work supporting DDO pieces in lotus-miner
* sealing: Update pipeline input to operate on UniversalPiece
* sealing: Update pipeline checks/sealing states to operate on UniversalPiece
* sealing: Make pipeline build with UniversalPiece
* move PieceDealInfo out of api
* make gen
* make sealing pipeline unit tests pass
* fix itest ensemble build
* don't panic in SectorsStatus with deals
* stop linter from complaining about checkPieces
* fix sector import tests
* mod tidy
* sealing: Add logic for (pre)committing DDO sectors
* sealing: state-types with method defs
* DDO non-snap pipeline works(?), DDO Itests
* DDO support in snapdeals pipeline
* make gen
* update actor bundles
* update the gst market fix
* fix: chain: use PreCommitSectorsBatch2 when setting up genesis
* some bug fixes
* integration working changes
* update actor bundles
* Make TestOnboardRawPieceSnap pass
* Appease the linter
* Make deadlines test pass with v12 actors
* Update go-state-types, abstract market DealState
* make gen
* mod tidy, lint fixes
* Fix some more tests
* Bump version in master
Bump version in master
* Make gen
Make gen
* fix sender
* fix: lotus-provider: Fix winning PoSt
* fix: sql Scan cannot write to an object
* Actually show miner-addrs in info-log
Actually show miner-addrs in lotus-provider info-log
* [WIP] feat: Add nv22 skeleton
Addition of Network Version 22 skeleton
* update FFI
* ddo is now nv22
* make gen
* temp actor bundle with ddo
* use working go-state-types
* gst with v13 market migration
* update bundle, builtin.MethodsMiner.ProveCommitSectors2 -> 3
* actually working v13 migration, v13 migration itest
* Address review
* sealing: Correct DDO snap pledge math
* itests: Mixed ddo itest
* pipeline: Fix sectorWeight
* sealing: convert market deals into PAMs in mixed sectors
* sealing: make market to ddo conversion work
* fix lint
* update gst
* Update actors and GST to lastest integ branch
* commit batcher: Update ProveCommitSectors3Params builder logic
* make gen
* use builtin-actors master
* ddo: address review
* itests: Add commd assertions to ddo tests
* make gen
* gst with fixed types
* config knobs for RequireActivationSuccess
* storage: Drop obsolete flaky tasts
---------
Co-authored-by: Jennifer Wang <jiayingw703@gmail.com>
Co-authored-by: Aayush <arajasek94@gmail.com>
Co-authored-by: Shrenuj Bansal <shrenuj.bansal@protocol.ai>
Co-authored-by: Phi <orjan.roren@gmail.com>
Co-authored-by: Andrew Jackson (Ajax) <snadrus@gmail.com>
Co-authored-by: TippyFlits <james.bluett@protocol.ai>
* feat: implement FIP-0063
* chore: deps: update to go-multiaddr v0.12.2 (#11602)
* feat: fvm: update the FVM/FFI to v4.1 (#11608) (#11612)
This:
1. Adds nv22 support.
2. Updates the message tracing format.
Co-authored-by: Steven Allen <steven@stebalien.com>
* AggregateProofType nil when doing batch updates
Use latest nv22 go-state-types version with matching update
* Update to v13.0.0-rc.2 bundle
* chore: Upgrade heights and codename
Update upgrade heights
Co-Authored-By: Steven Allen <steven@stebalien.com>
* Update epoch after nv22 DRAND switch
Update epoch after nv22 DRAND switch
* Update Mango codename to Phoneix
Make the codename for the Drand-change inline with Dragon style.
* Add UpgradePhoenixHeight to API params
* set UpgradePhoenixHeight to be one hour after Dragon
* Make gen
Make gen and UpgradePhoenixHeight in butterfly and local devnet to be in line with Calibration and Mainnet
* Update epoch heights (#11637)
Update epoch heights
* new: add forest bootstrap nodes (#11636)
Signed-off-by: samuelarogbonlo <sbayo971@gmail.com>
* Merge pull request #11491 from filecoin-project/fix/remove-decommissioned-pl-bootstrap-nodes
Remove PL operated bootstrap nodes from mainnet.pi
* feat: api: new verified registry methods to get all allocations and claims (#11631)
* new verireg methods
* update changelog and add itest
* update itest and cli
* update new method's support till v9
* remove gateway APIs
* fix cli internal var names
* chore:: backport #11609 to the feat/nv22 branch (#11644)
* feat: api: improve the correctness of Eth's trace_block (#11609)
* Improve the correctness of Eth's trace_block
- Improve encoding/decoding of parameters and return values:
- Encode "native" parameters and return values with Solidity ABI.
- Correctly decode parameters to "create" calls.
- Use the correct (ish) output for "create" calls.
- Handle all forms of "create".
- Make robust with respect to reverts:
- Use the actor ID/address from the trace instead of looking it up in
the state-tree (may not exist in the state-tree due to a revert).
- Gracefully handle failed actor/contract creation.
- Improve performance:
- We avoid looking anything up in the state-tree when translating the
trace, which should significantly improve performance.
- Improve code readability:
- Remove all "backtracking" logic.
- Use an "environment" struct to store temporary state instead of
attaching it to the trace.
- Fix random bugs:
- Fix an allocation bug in the "address" logic (need to set the
capacity before modifying the slice).
- Improved error checking/handling.
- Use correct types for `trace_block` action/results (create, call, etc.).
- And use the correct types for Result/Action structs instead of reusing the same "Call" action every time.
- Improve error messages.
* Make gen
Make gen
---------
Co-authored-by: Steven Allen <steven@stebalien.com>
* fix: add UpgradePhoenixHeight to StateGetNetworkParams (#11648)
* chore: deps: update to go-state-types v13.0.0-rc.1
* do NOT update the cache when running the real migration
* Merge pull request #11632 from hanabi1224/hm/drand-test
feat: drand quicknet: allow scheduling drand quicknet upgrade before nv22 on 2k devnet
* chore: deps: update to go-state-types v13.0.0-rc.2
chore: deps: update to go-state-types v13.0.0-rc.2
* feat: set migration config UpgradeEpoch for v13 actors upgrade
* Built-in actor events first draft
* itest for DDO non-market verified data w/ builtin actor events
* Tests for builtin actor events API
* Clean up DDO+Events tests, add lots of explainer comments
* Minor tweaks to events types
* Avoid duplicate messages when looking for receipts
* Rename internal events modules for clarity
* Adjust actor event API after review
* s/ActorEvents/Events/g in global config
* Manage event sending rate for SubscribeActorEvents
* Terminate SubscribeActorEvents chan when at max height
* Document future API changes
* More clarity in actor event API docs
* More post-review changes, lots of tests for SubscribeActorEvents
Use BlockDelay as the window for receiving events on the SubscribeActorEvents
channel. We expect the user to have received the initial batch of historical
events (if any) in one block's time. For real-time events we expect them to
not fall behind by roughly one block's time.
* Remove duplicate code from actor event type marshalling tests
Reduce verbosity and remove duplicate test logic from actor event types
JSON marshalling tests.
* Rename actor events test to follow go convention
Add missing `s` to `actor_events` test file to follow golang convention
used across the repo.
* Run actor events table tests in deterministic order
Refactor `map` usage for actor event table tests to ensure deterministic
test execution order, making debugging potential issues easier. If
non-determinism is a target, leverage Go's built-in parallel testing
capabilities.
* Reduce scope for filter removal failure when getting actor events
Use a fresh context to remove the temporary filter installed solely to
get the actor events. This should reduce chances of failure in a case
where the original context may be expired/cancelled.
Refactor removal into a `defer` statement for a more readable, concise
return statement.
* Use fixed RNG seed for actor event tests
Improve determinism in actor event tests by using a fixed RNG seed. This
makes up a more reproducible test suit.
* Use provided libraries to assert eventual conditions
Use the functionalities already provided by `testify` to assert eventual
conditions, and remove the use of `time.Sleep`.
Remove duplicate code in utility functions that are already defined.
Refactor assertion helper functions to use consistent terminology:
"require" implies fatal error, whereas "assert" implies error where the
test may proceed executing.
* Update changelog for actor events APIs
* Fix concerns and docs identified by review
* Update actor bundle to v13.0.0-rc3
Update actor bundle to v13.0.0-rc3
* Prep Lotus v1.26.0-rc1
- For sanity reverting the mainnet upgrade epoch to 99999999, and then only set it when cutting the final release
-Update Calibnet CIDs to v13.0.0-rc3
- Add GetActorEvents, SubscribeActorEvents, GetAllClaims and GetAllAllocations methods to the changelog
Co-Authored-By: Jiaying Wang <42981373+jennijuju@users.noreply.github.com>
* Update CHANGELOG.md
Co-authored-by: Masih H. Derkani <m@derkani.org>
* Make gen
Make gen
* fix: beacon: validate drand change at nv16 correctly
* bump to v1.26.0-rc2
* test: cleanup ddo verified itest, extract steps to functions
also add allocation-removed event case
* test: extract verified DDO test to separate file, add more checks
* test: add additional actor events checks
* Add verification for "deal-activated" actor event
* docs(drand): document the meaning of "IsChained" (#11692)
* Resolve conflicts
I encountered multiple issues when trying to run make gen. And these changes fixed a couple of them:
- go mod tidy
- Remove RaftState/RaftLeader
- Revert `if ts.Height() > claim.TermMax+claim.TermStart || !cctx.IsSet("expired")` to the what is in the release/v1.26.0: `if tsHeight > val.TermMax || !expired`
* fixup imports, make jen
* Update version
Update version in master to v1.27.0-dev
* Update node/impl/full/dummy.go
Co-authored-by: Łukasz Magiera <magik6k@users.noreply.github.com>
* Adjust ListClaimsCmd
Adjust ListClaimsCmd according to review
---------
Signed-off-by: samuelarogbonlo <sbayo971@gmail.com>
Co-authored-by: TippyFlits <james.bluett@protocol.ai>
Co-authored-by: Aayush <arajasek94@gmail.com>
Co-authored-by: Łukasz Magiera <magik6k@users.noreply.github.com>
Co-authored-by: Jennifer Wang <jiayingw703@gmail.com>
Co-authored-by: Shrenuj Bansal <shrenuj.bansal@protocol.ai>
Co-authored-by: Andrew Jackson (Ajax) <snadrus@gmail.com>
Co-authored-by: Steven Allen <steven@stebalien.com>
Co-authored-by: Rod Vagg <rod@vagg.org>
Co-authored-by: Samuel Arogbonlo <47984109+samuelarogbonlo@users.noreply.github.com>
Co-authored-by: LexLuthr <88259624+LexLuthr@users.noreply.github.com>
Co-authored-by: tom123222 <160735201+tom123222@users.noreply.github.com>
Co-authored-by: Aarsh Shah <aarshkshah1992@gmail.com>
Co-authored-by: Masih H. Derkani <m@derkani.org>
Co-authored-by: Jiaying Wang <42981373+jennijuju@users.noreply.github.com>
2024-03-12 09:33:58 +00:00
|
|
|
|
2023-08-01 15:28:47 +00:00
|
|
|
type FaultReporterConfig struct {
|
|
|
|
// EnableConsensusFaultReporter controls whether the node will monitor and
|
|
|
|
// report consensus faults. When enabled, the node will watch for malicious
|
|
|
|
// behaviors like double-mining and parent grinding, and submit reports to the
|
|
|
|
// network. This can earn reporter rewards, but is not guaranteed. Nodes should
|
|
|
|
// enable fault reporting with care, as it may increase resource usage, and may
|
|
|
|
// generate gas fees without earning rewards.
|
|
|
|
EnableConsensusFaultReporter bool
|
|
|
|
|
|
|
|
// ConsensusFaultReporterDataDir is the path where fault reporter state will be
|
|
|
|
// persisted. This directory should have adequate space and permissions for the
|
|
|
|
// node process.
|
|
|
|
ConsensusFaultReporterDataDir string
|
|
|
|
|
|
|
|
// ConsensusFaultReporterAddress is the wallet address used for submitting
|
|
|
|
// ReportConsensusFault messages. It will pay for gas fees, and receive any
|
|
|
|
// rewards. This address should have adequate funds to cover gas fees.
|
|
|
|
ConsensusFaultReporterAddress string
|
|
|
|
}
|
2024-06-21 18:48:22 +00:00
|
|
|
|
|
|
|
type CurioAlerting struct {
|
|
|
|
// PagerDutyEventURL is URL for PagerDuty.com Events API v2 URL. Events sent to this API URL are ultimately
|
|
|
|
// routed to a PagerDuty.com service and processed.
|
|
|
|
// The default is sufficient for integration with the stock commercial PagerDuty.com company's service.
|
|
|
|
PagerDutyEventURL string
|
|
|
|
|
|
|
|
// PageDutyIntegrationKey is the integration key for a PagerDuty.com service. You can find this unique service
|
|
|
|
// identifier in the integration page for the service.
|
|
|
|
PageDutyIntegrationKey string
|
|
|
|
|
|
|
|
// MinimumWalletBalance is the minimum balance all active wallets. If the balance is below this value, an
|
|
|
|
// alerts will be triggered for the wallet
|
|
|
|
MinimumWalletBalance types.FIL
|
|
|
|
}
|