Chore: Backports to the release/v1.26.0
branch (#11713)
* enable storing events (#11712) * fix: commit batch: Always go through commit batcher (#11704) * fix: commit batch: Always go through commit batcher * fix sealing fsm tests * sealing pipeline: Fix panic on padding pieces in WaitDeals (#11708) * sealing pipeline: Fix panic on padding pieces in WaitDeals * sealing pipeline: Catch panics * sealing pipeline: Output DDO pieces in SectorStatus (#11709) * sealing pipeline: Fix failing ProveCommit3 aggregate (#11710) * itests: Repro failing ProveCommit3 aggregate * commit batch: Correctly sort sectors in processBatchV2 * fix imports * ci: Bigger instance for sector_pledge test * itests: Use Must-Post mining in TestPledgeBatching --------- Co-authored-by: Aarsh Shah <aarshkshah1992@gmail.com> Co-authored-by: Łukasz Magiera <magik6k@users.noreply.github.com>
This commit is contained in:
parent
ea2d0790b4
commit
f929ae17d7
@ -945,6 +945,7 @@ workflows:
|
|||||||
- build
|
- build
|
||||||
suite: itest-sector_pledge
|
suite: itest-sector_pledge
|
||||||
target: "./itests/sector_pledge_test.go"
|
target: "./itests/sector_pledge_test.go"
|
||||||
|
resource_class: 2xlarge
|
||||||
get-params: true
|
get-params: true
|
||||||
|
|
||||||
- test:
|
- test:
|
||||||
|
@ -551,7 +551,7 @@ workflows:
|
|||||||
- build
|
- build
|
||||||
suite: itest-[[ $name ]]
|
suite: itest-[[ $name ]]
|
||||||
target: "./itests/[[ $file ]]"
|
target: "./itests/[[ $file ]]"
|
||||||
[[- if or (eq $name "worker") (eq $name "deals_concurrent") (eq $name "wdpost_worker_config")]]
|
[[- if or (eq $name "worker") (eq $name "deals_concurrent") (eq $name "wdpost_worker_config") (eq $name "sector_pledge")]]
|
||||||
resource_class: 2xlarge
|
resource_class: 2xlarge
|
||||||
[[- end]]
|
[[- end]]
|
||||||
[[- if or (eq $name "wdpost") (eq $name "sector_pledge")]]
|
[[- if or (eq $name "wdpost") (eq $name "sector_pledge")]]
|
||||||
|
@ -35,7 +35,13 @@ var DefaultEnsembleOpts = ensembleOpts{
|
|||||||
}
|
}
|
||||||
|
|
||||||
// MockProofs activates mock proofs for the entire ensemble.
|
// MockProofs activates mock proofs for the entire ensemble.
|
||||||
func MockProofs() EnsembleOpt {
|
func MockProofs(e ...bool) EnsembleOpt {
|
||||||
|
if len(e) > 0 && !e[0] {
|
||||||
|
return func(opts *ensembleOpts) error {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
return func(opts *ensembleOpts) error {
|
return func(opts *ensembleOpts) error {
|
||||||
opts.mockProofs = true
|
opts.mockProofs = true
|
||||||
// since we're using mock proofs, we don't need to download
|
// since we're using mock proofs, we don't need to download
|
||||||
|
@ -12,6 +12,7 @@ import (
|
|||||||
"github.com/stretchr/testify/require"
|
"github.com/stretchr/testify/require"
|
||||||
|
|
||||||
"github.com/filecoin-project/go-state-types/abi"
|
"github.com/filecoin-project/go-state-types/abi"
|
||||||
|
"github.com/filecoin-project/go-state-types/big"
|
||||||
miner5 "github.com/filecoin-project/specs-actors/v5/actors/builtin/miner"
|
miner5 "github.com/filecoin-project/specs-actors/v5/actors/builtin/miner"
|
||||||
|
|
||||||
"github.com/filecoin-project/lotus/api"
|
"github.com/filecoin-project/lotus/api"
|
||||||
@ -39,7 +40,7 @@ func TestPledgeSectors(t *testing.T) {
|
|||||||
defer cancel()
|
defer cancel()
|
||||||
|
|
||||||
_, miner, ens := kit.EnsembleMinimal(t, kit.MockProofs())
|
_, miner, ens := kit.EnsembleMinimal(t, kit.MockProofs())
|
||||||
ens.InterconnectAll().BeginMining(blockTime)
|
ens.InterconnectAll().BeginMiningMustPost(blockTime)
|
||||||
|
|
||||||
miner.PledgeSectors(ctx, nSectors, 0, nil)
|
miner.PledgeSectors(ctx, nSectors, 0, nil)
|
||||||
}
|
}
|
||||||
@ -65,12 +66,18 @@ func TestPledgeBatching(t *testing.T) {
|
|||||||
//stm: @SECTOR_PRE_COMMIT_FLUSH_001, @SECTOR_COMMIT_FLUSH_001
|
//stm: @SECTOR_PRE_COMMIT_FLUSH_001, @SECTOR_COMMIT_FLUSH_001
|
||||||
blockTime := 50 * time.Millisecond
|
blockTime := 50 * time.Millisecond
|
||||||
|
|
||||||
runTest := func(t *testing.T, nSectors int) {
|
runTest := func(t *testing.T, nSectors int, aggregate bool) {
|
||||||
ctx, cancel := context.WithCancel(context.Background())
|
ctx, cancel := context.WithCancel(context.Background())
|
||||||
defer cancel()
|
defer cancel()
|
||||||
|
|
||||||
client, miner, ens := kit.EnsembleMinimal(t, kit.MockProofs())
|
kit.QuietMiningLogs()
|
||||||
ens.InterconnectAll().BeginMining(blockTime)
|
|
||||||
|
client, miner, ens := kit.EnsembleMinimal(t, kit.MockProofs(!aggregate), kit.MutateSealingConfig(func(sc *config.SealingConfig) {
|
||||||
|
if aggregate {
|
||||||
|
sc.AggregateAboveBaseFee = types.FIL(big.Zero())
|
||||||
|
}
|
||||||
|
}))
|
||||||
|
ens.InterconnectAll().BeginMiningMustPost(blockTime)
|
||||||
|
|
||||||
client.WaitTillChain(ctx, kit.HeightAtLeast(10))
|
client.WaitTillChain(ctx, kit.HeightAtLeast(10))
|
||||||
|
|
||||||
@ -114,7 +121,10 @@ func TestPledgeBatching(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
t.Run("100", func(t *testing.T) {
|
t.Run("100", func(t *testing.T) {
|
||||||
runTest(t, 100)
|
runTest(t, 100, false)
|
||||||
|
})
|
||||||
|
t.Run("10-agg", func(t *testing.T) {
|
||||||
|
runTest(t, 10, true)
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -225,7 +225,7 @@ func ConfigFullNode(c interface{}) Option {
|
|||||||
// If the Eth JSON-RPC is enabled, enable storing events at the ChainStore.
|
// If the Eth JSON-RPC is enabled, enable storing events at the ChainStore.
|
||||||
// This is the case even if real-time and historic filtering are disabled,
|
// This is the case even if real-time and historic filtering are disabled,
|
||||||
// as it enables us to serve logs in eth_getTransactionReceipt.
|
// as it enables us to serve logs in eth_getTransactionReceipt.
|
||||||
If(cfg.Fevm.EnableEthRPC, Override(StoreEventsKey, modules.EnableStoringEvents)),
|
If(cfg.Fevm.EnableEthRPC || cfg.Events.EnableActorEventsAPI, Override(StoreEventsKey, modules.EnableStoringEvents)),
|
||||||
|
|
||||||
Override(new(dtypes.ClientImportMgr), modules.ClientImportMgr),
|
Override(new(dtypes.ClientImportMgr), modules.ClientImportMgr),
|
||||||
|
|
||||||
|
@ -209,7 +209,7 @@ func (b *CommitBatcher) maybeStartBatch(notif bool) ([]sealiface.CommitBatchRes,
|
|||||||
return nil, xerrors.Errorf("getting config: %w", err)
|
return nil, xerrors.Errorf("getting config: %w", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
if notif && total < cfg.MaxCommitBatch {
|
if notif && total < cfg.MaxCommitBatch && cfg.AggregateCommits {
|
||||||
return nil, nil
|
return nil, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -233,7 +233,7 @@ func (b *CommitBatcher) maybeStartBatch(notif bool) ([]sealiface.CommitBatchRes,
|
|||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
|
|
||||||
individual := (total < cfg.MinCommitBatch) || (total < miner.MinAggregatedSectors) || blackedOut()
|
individual := (total < cfg.MinCommitBatch) || (total < miner.MinAggregatedSectors) || blackedOut() || !cfg.AggregateCommits
|
||||||
|
|
||||||
if !individual && !cfg.AggregateAboveBaseFee.Equals(big.Zero()) {
|
if !individual && !cfg.AggregateAboveBaseFee.Equals(big.Zero()) {
|
||||||
if ts.MinTicketBlock().ParentBaseFee.LessThan(cfg.AggregateAboveBaseFee) {
|
if ts.MinTicketBlock().ParentBaseFee.LessThan(cfg.AggregateAboveBaseFee) {
|
||||||
@ -331,6 +331,9 @@ func (b *CommitBatcher) processBatchV2(cfg sealiface.Config, sectors []abi.Secto
|
|||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// sort sectors by number
|
||||||
|
sort.Slice(sectors, func(i, j int) bool { return sectors[i] < sectors[j] })
|
||||||
|
|
||||||
total := len(sectors)
|
total := len(sectors)
|
||||||
|
|
||||||
res := sealiface.CommitBatchRes{
|
res := sealiface.CommitBatchRes{
|
||||||
@ -371,10 +374,6 @@ func (b *CommitBatcher) processBatchV2(cfg sealiface.Config, sectors []abi.Secto
|
|||||||
return nil, nil
|
return nil, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
sort.Slice(infos, func(i, j int) bool {
|
|
||||||
return infos[i].Number < infos[j].Number
|
|
||||||
})
|
|
||||||
|
|
||||||
proofs := make([][]byte, 0, total)
|
proofs := make([][]byte, 0, total)
|
||||||
for _, info := range infos {
|
for _, info := range infos {
|
||||||
proofs = append(proofs, b.todo[info.Number].Proof)
|
proofs = append(proofs, b.todo[info.Number].Proof)
|
||||||
@ -444,13 +443,13 @@ func (b *CommitBatcher) processBatchV2(cfg sealiface.Config, sectors []abi.Secto
|
|||||||
enc := new(bytes.Buffer)
|
enc := new(bytes.Buffer)
|
||||||
if err := params.MarshalCBOR(enc); err != nil {
|
if err := params.MarshalCBOR(enc); err != nil {
|
||||||
res.Error = err.Error()
|
res.Error = err.Error()
|
||||||
return []sealiface.CommitBatchRes{res}, xerrors.Errorf("couldn't serialize ProveCommitSectors2Params: %w", err)
|
return []sealiface.CommitBatchRes{res}, xerrors.Errorf("couldn't serialize ProveCommitSectors3Params: %w", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
_, err = simulateMsgGas(b.mctx, b.api, from, b.maddr, builtin.MethodsMiner.ProveCommitSectors3, needFunds, maxFee, enc.Bytes())
|
_, err = simulateMsgGas(b.mctx, b.api, from, b.maddr, builtin.MethodsMiner.ProveCommitSectors3, needFunds, maxFee, enc.Bytes())
|
||||||
|
|
||||||
if err != nil && (!api.ErrorIsIn(err, []error{&api.ErrOutOfGas{}}) || len(sectors) < miner.MinAggregatedSectors*2) {
|
if err != nil && (!api.ErrorIsIn(err, []error{&api.ErrOutOfGas{}}) || len(sectors) < miner.MinAggregatedSectors*2) {
|
||||||
log.Errorf("simulating CommitBatch message failed: %s", err)
|
log.Errorf("simulating CommitBatch message failed (%x): %s", enc.Bytes(), err)
|
||||||
res.Error = err.Error()
|
res.Error = err.Error()
|
||||||
return []sealiface.CommitBatchRes{res}, xerrors.Errorf("simulating CommitBatch message failed: %w", err)
|
return []sealiface.CommitBatchRes{res}, xerrors.Errorf("simulating CommitBatch message failed: %w", err)
|
||||||
}
|
}
|
||||||
@ -474,7 +473,7 @@ func (b *CommitBatcher) processBatchV2(cfg sealiface.Config, sectors []abi.Secto
|
|||||||
|
|
||||||
res.Msg = &mcid
|
res.Msg = &mcid
|
||||||
|
|
||||||
log.Infow("Sent ProveCommitSectors2 message", "cid", mcid, "from", from, "todo", total, "sectors", len(infos))
|
log.Infow("Sent ProveCommitSectors3 message", "cid", mcid, "from", from, "todo", total, "sectors", len(infos))
|
||||||
|
|
||||||
return []sealiface.CommitBatchRes{res}, nil
|
return []sealiface.CommitBatchRes{res}, nil
|
||||||
}
|
}
|
||||||
@ -591,7 +590,7 @@ func (b *CommitBatcher) processBatchV1(cfg sealiface.Config, sectors []abi.Secto
|
|||||||
_, err = simulateMsgGas(b.mctx, b.api, from, b.maddr, builtin.MethodsMiner.ProveCommitAggregate, needFunds, maxFee, enc.Bytes())
|
_, err = simulateMsgGas(b.mctx, b.api, from, b.maddr, builtin.MethodsMiner.ProveCommitAggregate, needFunds, maxFee, enc.Bytes())
|
||||||
|
|
||||||
if err != nil && (!api.ErrorIsIn(err, []error{&api.ErrOutOfGas{}}) || len(sectors) < miner.MinAggregatedSectors*2) {
|
if err != nil && (!api.ErrorIsIn(err, []error{&api.ErrOutOfGas{}}) || len(sectors) < miner.MinAggregatedSectors*2) {
|
||||||
log.Errorf("simulating CommitBatch message failed: %s", err)
|
log.Errorf("simulating CommitBatch message failed (%x): %s", enc.Bytes(), err)
|
||||||
res.Error = err.Error()
|
res.Error = err.Error()
|
||||||
return []sealiface.CommitBatchRes{res}, xerrors.Errorf("simulating CommitBatch message failed: %w", err)
|
return []sealiface.CommitBatchRes{res}, xerrors.Errorf("simulating CommitBatch message failed: %w", err)
|
||||||
}
|
}
|
||||||
|
@ -11,6 +11,7 @@ import (
|
|||||||
"net/http"
|
"net/http"
|
||||||
"os"
|
"os"
|
||||||
"reflect"
|
"reflect"
|
||||||
|
"runtime"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"golang.org/x/xerrors"
|
"golang.org/x/xerrors"
|
||||||
@ -39,8 +40,27 @@ func (m *Sealing) Plan(events []statemachine.Event, user interface{}) (interface
|
|||||||
return nil, processed, nil
|
return nil, processed, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
return func(ctx statemachine.Context, si SectorInfo) error {
|
return func(ctx statemachine.Context, si SectorInfo) (err error) {
|
||||||
err := next(ctx, si)
|
// handle panics
|
||||||
|
defer func() {
|
||||||
|
if r := recover(); r != nil {
|
||||||
|
buf := make([]byte, 1<<16)
|
||||||
|
n := runtime.Stack(buf, false)
|
||||||
|
buf = buf[:n]
|
||||||
|
|
||||||
|
l := Log{
|
||||||
|
Timestamp: uint64(time.Now().Unix()),
|
||||||
|
Message: fmt.Sprintf("panic: %v\n%s", r, buf),
|
||||||
|
Kind: "panic",
|
||||||
|
}
|
||||||
|
si.logAppend(l)
|
||||||
|
|
||||||
|
err = fmt.Errorf("panic: %v\n%s", r, buf)
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
|
||||||
|
// execute the next state
|
||||||
|
err = next(ctx, si)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
log.Errorf("unhandled sector error (%d): %+v", si.SectorNumber, err)
|
log.Errorf("unhandled sector error (%d): %+v", si.SectorNumber, err)
|
||||||
return nil
|
return nil
|
||||||
@ -127,8 +147,8 @@ var fsmPlanners = map[SectorState]func(events []statemachine.Event, state *Secto
|
|||||||
),
|
),
|
||||||
Committing: planCommitting,
|
Committing: planCommitting,
|
||||||
CommitFinalize: planOne(
|
CommitFinalize: planOne(
|
||||||
on(SectorFinalized{}, SubmitCommit),
|
on(SectorFinalized{}, SubmitCommitAggregate),
|
||||||
on(SectorFinalizedAvailable{}, SubmitCommit),
|
on(SectorFinalizedAvailable{}, SubmitCommitAggregate),
|
||||||
on(SectorFinalizeFailed{}, CommitFinalizeFailed),
|
on(SectorFinalizeFailed{}, CommitFinalizeFailed),
|
||||||
),
|
),
|
||||||
SubmitCommit: planOne(
|
SubmitCommit: planOne(
|
||||||
@ -674,7 +694,7 @@ func planCommitting(events []statemachine.Event, state *SectorInfo) (uint64, err
|
|||||||
}
|
}
|
||||||
case SectorCommitted: // the normal case
|
case SectorCommitted: // the normal case
|
||||||
e.apply(state)
|
e.apply(state)
|
||||||
state.State = SubmitCommit
|
state.State = SubmitCommitAggregate
|
||||||
case SectorProofReady: // early finalize
|
case SectorProofReady: // early finalize
|
||||||
e.apply(state)
|
e.apply(state)
|
||||||
state.State = CommitFinalize
|
state.State = CommitFinalize
|
||||||
|
@ -70,10 +70,10 @@ func TestHappyPath(t *testing.T) {
|
|||||||
require.Equal(m.t, m.state.State, Committing)
|
require.Equal(m.t, m.state.State, Committing)
|
||||||
|
|
||||||
m.planSingle(SectorCommitted{})
|
m.planSingle(SectorCommitted{})
|
||||||
require.Equal(m.t, m.state.State, SubmitCommit)
|
require.Equal(m.t, m.state.State, SubmitCommitAggregate)
|
||||||
|
|
||||||
m.planSingle(SectorCommitSubmitted{})
|
m.planSingle(SectorCommitAggregateSent{})
|
||||||
require.Equal(m.t, m.state.State, CommitWait)
|
require.Equal(m.t, m.state.State, CommitAggregateWait)
|
||||||
|
|
||||||
m.planSingle(SectorProving{})
|
m.planSingle(SectorProving{})
|
||||||
require.Equal(m.t, m.state.State, FinalizeSector)
|
require.Equal(m.t, m.state.State, FinalizeSector)
|
||||||
@ -81,7 +81,7 @@ func TestHappyPath(t *testing.T) {
|
|||||||
m.planSingle(SectorFinalized{})
|
m.planSingle(SectorFinalized{})
|
||||||
require.Equal(m.t, m.state.State, Proving)
|
require.Equal(m.t, m.state.State, Proving)
|
||||||
|
|
||||||
expected := []SectorState{Packing, GetTicket, PreCommit1, PreCommit2, SubmitPreCommitBatch, PreCommitBatchWait, WaitSeed, Committing, SubmitCommit, CommitWait, FinalizeSector, Proving}
|
expected := []SectorState{Packing, GetTicket, PreCommit1, PreCommit2, SubmitPreCommitBatch, PreCommitBatchWait, WaitSeed, Committing, SubmitCommitAggregate, CommitAggregateWait, FinalizeSector, Proving}
|
||||||
for i, n := range notif {
|
for i, n := range notif {
|
||||||
if n.before.State != expected[i] {
|
if n.before.State != expected[i] {
|
||||||
t.Fatalf("expected before state: %s, got: %s", expected[i], n.before.State)
|
t.Fatalf("expected before state: %s, got: %s", expected[i], n.before.State)
|
||||||
@ -135,9 +135,6 @@ func TestHappyPathFinalizeEarly(t *testing.T) {
|
|||||||
require.Equal(m.t, m.state.State, CommitFinalize)
|
require.Equal(m.t, m.state.State, CommitFinalize)
|
||||||
|
|
||||||
m.planSingle(SectorFinalized{})
|
m.planSingle(SectorFinalized{})
|
||||||
require.Equal(m.t, m.state.State, SubmitCommit)
|
|
||||||
|
|
||||||
m.planSingle(SectorSubmitCommitAggregate{})
|
|
||||||
require.Equal(m.t, m.state.State, SubmitCommitAggregate)
|
require.Equal(m.t, m.state.State, SubmitCommitAggregate)
|
||||||
|
|
||||||
m.planSingle(SectorCommitAggregateSent{})
|
m.planSingle(SectorCommitAggregateSent{})
|
||||||
@ -149,7 +146,7 @@ func TestHappyPathFinalizeEarly(t *testing.T) {
|
|||||||
m.planSingle(SectorFinalized{})
|
m.planSingle(SectorFinalized{})
|
||||||
require.Equal(m.t, m.state.State, Proving)
|
require.Equal(m.t, m.state.State, Proving)
|
||||||
|
|
||||||
expected := []SectorState{Packing, GetTicket, PreCommit1, PreCommit2, SubmitPreCommitBatch, PreCommitBatchWait, WaitSeed, Committing, CommitFinalize, SubmitCommit, SubmitCommitAggregate, CommitAggregateWait, FinalizeSector, Proving}
|
expected := []SectorState{Packing, GetTicket, PreCommit1, PreCommit2, SubmitPreCommitBatch, PreCommitBatchWait, WaitSeed, Committing, CommitFinalize, SubmitCommitAggregate, CommitAggregateWait, FinalizeSector, Proving}
|
||||||
for i, n := range notif {
|
for i, n := range notif {
|
||||||
if n.before.State != expected[i] {
|
if n.before.State != expected[i] {
|
||||||
t.Fatalf("expected before state: %s, got: %s", expected[i], n.before.State)
|
t.Fatalf("expected before state: %s, got: %s", expected[i], n.before.State)
|
||||||
@ -188,9 +185,9 @@ func TestCommitFinalizeFailed(t *testing.T) {
|
|||||||
require.Equal(m.t, m.state.State, CommitFinalize)
|
require.Equal(m.t, m.state.State, CommitFinalize)
|
||||||
|
|
||||||
m.planSingle(SectorFinalized{})
|
m.planSingle(SectorFinalized{})
|
||||||
require.Equal(m.t, m.state.State, SubmitCommit)
|
require.Equal(m.t, m.state.State, SubmitCommitAggregate)
|
||||||
|
|
||||||
expected := []SectorState{Committing, CommitFinalize, CommitFinalizeFailed, CommitFinalize, SubmitCommit}
|
expected := []SectorState{Committing, CommitFinalize, CommitFinalizeFailed, CommitFinalize, SubmitCommitAggregate}
|
||||||
for i, n := range notif {
|
for i, n := range notif {
|
||||||
if n.before.State != expected[i] {
|
if n.before.State != expected[i] {
|
||||||
t.Fatalf("expected before state: %s, got: %s", expected[i], n.before.State)
|
t.Fatalf("expected before state: %s, got: %s", expected[i], n.before.State)
|
||||||
@ -242,10 +239,10 @@ func TestSeedRevert(t *testing.T) {
|
|||||||
// not changing the seed this time
|
// not changing the seed this time
|
||||||
_, _, err = m.s.plan([]statemachine.Event{{User: SectorSeedReady{SeedValue: nil, SeedEpoch: 5}}, {User: SectorCommitted{}}}, m.state)
|
_, _, err = m.s.plan([]statemachine.Event{{User: SectorSeedReady{SeedValue: nil, SeedEpoch: 5}}, {User: SectorCommitted{}}}, m.state)
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
require.Equal(m.t, m.state.State, SubmitCommit)
|
require.Equal(m.t, m.state.State, SubmitCommitAggregate)
|
||||||
|
|
||||||
m.planSingle(SectorCommitSubmitted{})
|
m.planSingle(SectorCommitAggregateSent{})
|
||||||
require.Equal(m.t, m.state.State, CommitWait)
|
require.Equal(m.t, m.state.State, CommitAggregateWait)
|
||||||
|
|
||||||
m.planSingle(SectorProving{})
|
m.planSingle(SectorProving{})
|
||||||
require.Equal(m.t, m.state.State, FinalizeSector)
|
require.Equal(m.t, m.state.State, FinalizeSector)
|
||||||
|
@ -34,12 +34,16 @@ func (m *Sealing) handleWaitDeals(ctx statemachine.Context, sector SectorInfo) e
|
|||||||
for _, piece := range sector.Pieces {
|
for _, piece := range sector.Pieces {
|
||||||
used += piece.Piece().Size.Unpadded()
|
used += piece.Piece().Size.Unpadded()
|
||||||
|
|
||||||
|
if !piece.HasDealInfo() {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
endEpoch, err := piece.EndEpoch()
|
endEpoch, err := piece.EndEpoch()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return xerrors.Errorf("piece.EndEpoch: %w", err)
|
return xerrors.Errorf("piece.EndEpoch: %w", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
if piece.HasDealInfo() && endEpoch > lastDealEnd {
|
if endEpoch > lastDealEnd {
|
||||||
lastDealEnd = endEpoch
|
lastDealEnd = endEpoch
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -953,20 +957,30 @@ func (m *Sealing) SectorsStatus(ctx context.Context, sid abi.SectorNumber, showO
|
|||||||
return api.SectorInfo{}, err
|
return api.SectorInfo{}, err
|
||||||
}
|
}
|
||||||
|
|
||||||
|
nv, err := m.Api.StateNetworkVersion(ctx, types.EmptyTSK)
|
||||||
|
if err != nil {
|
||||||
|
return api.SectorInfo{}, xerrors.Errorf("getting network version: %w", err)
|
||||||
|
}
|
||||||
|
|
||||||
deals := make([]abi.DealID, len(info.Pieces))
|
deals := make([]abi.DealID, len(info.Pieces))
|
||||||
pieces := make([]api.SectorPiece, len(info.Pieces))
|
pieces := make([]api.SectorPiece, len(info.Pieces))
|
||||||
for i, piece := range info.Pieces {
|
for i, piece := range info.Pieces {
|
||||||
// todo make this work with DDO deals in some reasonable way
|
|
||||||
|
|
||||||
pieces[i].Piece = piece.Piece()
|
pieces[i].Piece = piece.Piece()
|
||||||
if !piece.HasDealInfo() || piece.Impl().PublishCid == nil {
|
|
||||||
|
if !piece.HasDealInfo() {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
pdi := piece.Impl()
|
||||||
|
if pdi.Valid(nv) != nil {
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
|
|
||||||
pdi := piece.DealInfo().Impl() // copy
|
|
||||||
pieces[i].DealInfo = &pdi
|
pieces[i].DealInfo = &pdi
|
||||||
|
|
||||||
deals[i] = piece.DealInfo().Impl().DealID
|
if pdi.PublishCid != nil {
|
||||||
|
deals[i] = pdi.DealID
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
log := make([]api.SectorLog, len(info.Log))
|
log := make([]api.SectorLog, len(info.Log))
|
||||||
|
@ -94,7 +94,7 @@ const (
|
|||||||
CommitFinalizeFailed SectorState = "CommitFinalizeFailed"
|
CommitFinalizeFailed SectorState = "CommitFinalizeFailed"
|
||||||
|
|
||||||
// single commit
|
// single commit
|
||||||
SubmitCommit SectorState = "SubmitCommit" // send commit message to the chain
|
SubmitCommit SectorState = "SubmitCommit" // send commit message to the chain (deprecated)
|
||||||
CommitWait SectorState = "CommitWait" // wait for the commit message to land on chain
|
CommitWait SectorState = "CommitWait" // wait for the commit message to land on chain
|
||||||
|
|
||||||
SubmitCommitAggregate SectorState = "SubmitCommitAggregate"
|
SubmitCommitAggregate SectorState = "SubmitCommitAggregate"
|
||||||
|
@ -18,7 +18,6 @@ import (
|
|||||||
"github.com/filecoin-project/go-state-types/abi"
|
"github.com/filecoin-project/go-state-types/abi"
|
||||||
actorstypes "github.com/filecoin-project/go-state-types/actors"
|
actorstypes "github.com/filecoin-project/go-state-types/actors"
|
||||||
"github.com/filecoin-project/go-state-types/big"
|
"github.com/filecoin-project/go-state-types/big"
|
||||||
"github.com/filecoin-project/go-state-types/builtin"
|
|
||||||
miner2 "github.com/filecoin-project/go-state-types/builtin/v13/miner"
|
miner2 "github.com/filecoin-project/go-state-types/builtin/v13/miner"
|
||||||
verifreg13 "github.com/filecoin-project/go-state-types/builtin/v13/verifreg"
|
verifreg13 "github.com/filecoin-project/go-state-types/builtin/v13/verifreg"
|
||||||
"github.com/filecoin-project/go-state-types/builtin/v9/verifreg"
|
"github.com/filecoin-project/go-state-types/builtin/v9/verifreg"
|
||||||
@ -740,90 +739,11 @@ func (m *Sealing) handleCommitting(ctx statemachine.Context, sector SectorInfo)
|
|||||||
}
|
}
|
||||||
|
|
||||||
func (m *Sealing) handleSubmitCommit(ctx statemachine.Context, sector SectorInfo) error {
|
func (m *Sealing) handleSubmitCommit(ctx statemachine.Context, sector SectorInfo) error {
|
||||||
// TODO: Deprecate this path, always go through batcher, just respect the AggregateCommits config in there
|
// like precommit this is a deprecated state, but we keep it around for
|
||||||
|
// existing state machines
|
||||||
cfg, err := m.getConfig()
|
// todo: drop after nv21
|
||||||
if err != nil {
|
|
||||||
return xerrors.Errorf("getting config: %w", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
if cfg.AggregateCommits {
|
|
||||||
nv, err := m.Api.StateNetworkVersion(ctx.Context(), types.EmptyTSK)
|
|
||||||
if err != nil {
|
|
||||||
return xerrors.Errorf("getting network version: %w", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
if nv >= network.Version13 {
|
|
||||||
return ctx.Send(SectorSubmitCommitAggregate{})
|
return ctx.Send(SectorSubmitCommitAggregate{})
|
||||||
}
|
}
|
||||||
}
|
|
||||||
|
|
||||||
ts, err := m.Api.ChainHead(ctx.Context())
|
|
||||||
if err != nil {
|
|
||||||
log.Errorf("handleSubmitCommit: api error, not proceeding: %+v", err)
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
if err := m.checkCommit(ctx.Context(), sector, sector.Proof, ts.Key()); err != nil {
|
|
||||||
return ctx.Send(SectorCommitFailed{xerrors.Errorf("commit check error: %w", err)})
|
|
||||||
}
|
|
||||||
|
|
||||||
enc := new(bytes.Buffer)
|
|
||||||
params := &miner.ProveCommitSectorParams{
|
|
||||||
SectorNumber: sector.SectorNumber,
|
|
||||||
Proof: sector.Proof,
|
|
||||||
}
|
|
||||||
|
|
||||||
if err := params.MarshalCBOR(enc); err != nil {
|
|
||||||
return ctx.Send(SectorCommitFailed{xerrors.Errorf("could not serialize commit sector parameters: %w", err)})
|
|
||||||
}
|
|
||||||
|
|
||||||
mi, err := m.Api.StateMinerInfo(ctx.Context(), m.maddr, ts.Key())
|
|
||||||
if err != nil {
|
|
||||||
log.Errorf("handleCommitting: api error, not proceeding: %+v", err)
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
pci, err := m.Api.StateSectorPreCommitInfo(ctx.Context(), m.maddr, sector.SectorNumber, ts.Key())
|
|
||||||
if err != nil {
|
|
||||||
return xerrors.Errorf("getting precommit info: %w", err)
|
|
||||||
}
|
|
||||||
if pci == nil {
|
|
||||||
return ctx.Send(SectorCommitFailed{error: xerrors.Errorf("precommit info not found on chain")})
|
|
||||||
}
|
|
||||||
|
|
||||||
collateral, err := m.Api.StateMinerInitialPledgeCollateral(ctx.Context(), m.maddr, pci.Info, ts.Key())
|
|
||||||
if err != nil {
|
|
||||||
return xerrors.Errorf("getting initial pledge collateral: %w", err)
|
|
||||||
}
|
|
||||||
|
|
||||||
collateral = big.Sub(collateral, pci.PreCommitDeposit)
|
|
||||||
if collateral.LessThan(big.Zero()) {
|
|
||||||
collateral = big.Zero()
|
|
||||||
}
|
|
||||||
|
|
||||||
collateral, err = collateralSendAmount(ctx.Context(), m.Api, m.maddr, cfg, collateral)
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
|
|
||||||
goodFunds := big.Add(collateral, big.Int(m.feeCfg.MaxCommitGasFee))
|
|
||||||
|
|
||||||
from, _, err := m.addrSel.AddressFor(ctx.Context(), m.Api, mi, api.CommitAddr, goodFunds, collateral)
|
|
||||||
if err != nil {
|
|
||||||
return ctx.Send(SectorCommitFailed{xerrors.Errorf("no good address to send commit message from: %w", err)})
|
|
||||||
}
|
|
||||||
|
|
||||||
// TODO: check seed / ticket / deals are up to date
|
|
||||||
mcid, err := sendMsg(ctx.Context(), m.Api, from, m.maddr, builtin.MethodsMiner.ProveCommitSector, collateral, big.Int(m.feeCfg.MaxCommitGasFee), enc.Bytes())
|
|
||||||
if err != nil {
|
|
||||||
return ctx.Send(SectorCommitFailed{xerrors.Errorf("pushing message to mpool: %w", err)})
|
|
||||||
}
|
|
||||||
|
|
||||||
return ctx.Send(SectorCommitSubmitted{
|
|
||||||
Message: mcid,
|
|
||||||
})
|
|
||||||
}
|
|
||||||
|
|
||||||
// processPieces returns either:
|
// processPieces returns either:
|
||||||
// - a list of piece activation manifests
|
// - a list of piece activation manifests
|
||||||
|
@ -289,10 +289,18 @@ func (sp *SafeSectorPiece) handleDealInfo(params handleDealInfoParams) error {
|
|||||||
// SectorPiece Proxy
|
// SectorPiece Proxy
|
||||||
|
|
||||||
func (sp *SafeSectorPiece) Impl() piece.PieceDealInfo {
|
func (sp *SafeSectorPiece) Impl() piece.PieceDealInfo {
|
||||||
|
if !sp.HasDealInfo() {
|
||||||
|
return piece.PieceDealInfo{}
|
||||||
|
}
|
||||||
|
|
||||||
return sp.real.DealInfo.Impl()
|
return sp.real.DealInfo.Impl()
|
||||||
}
|
}
|
||||||
|
|
||||||
func (sp *SafeSectorPiece) String() string {
|
func (sp *SafeSectorPiece) String() string {
|
||||||
|
if !sp.HasDealInfo() {
|
||||||
|
return "<no deal info>"
|
||||||
|
}
|
||||||
|
|
||||||
return sp.real.DealInfo.String()
|
return sp.real.DealInfo.String()
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -305,21 +313,41 @@ func (sp *SafeSectorPiece) Valid(nv network.Version) error {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func (sp *SafeSectorPiece) StartEpoch() (abi.ChainEpoch, error) {
|
func (sp *SafeSectorPiece) StartEpoch() (abi.ChainEpoch, error) {
|
||||||
|
if !sp.HasDealInfo() {
|
||||||
|
return 0, xerrors.Errorf("no deal info")
|
||||||
|
}
|
||||||
|
|
||||||
return sp.real.DealInfo.StartEpoch()
|
return sp.real.DealInfo.StartEpoch()
|
||||||
}
|
}
|
||||||
|
|
||||||
func (sp *SafeSectorPiece) EndEpoch() (abi.ChainEpoch, error) {
|
func (sp *SafeSectorPiece) EndEpoch() (abi.ChainEpoch, error) {
|
||||||
|
if !sp.HasDealInfo() {
|
||||||
|
return 0, xerrors.Errorf("no deal info")
|
||||||
|
}
|
||||||
|
|
||||||
return sp.real.DealInfo.EndEpoch()
|
return sp.real.DealInfo.EndEpoch()
|
||||||
}
|
}
|
||||||
|
|
||||||
func (sp *SafeSectorPiece) PieceCID() cid.Cid {
|
func (sp *SafeSectorPiece) PieceCID() cid.Cid {
|
||||||
|
if !sp.HasDealInfo() {
|
||||||
|
return sp.real.Piece.PieceCID
|
||||||
|
}
|
||||||
|
|
||||||
return sp.real.DealInfo.PieceCID()
|
return sp.real.DealInfo.PieceCID()
|
||||||
}
|
}
|
||||||
|
|
||||||
func (sp *SafeSectorPiece) KeepUnsealedRequested() bool {
|
func (sp *SafeSectorPiece) KeepUnsealedRequested() bool {
|
||||||
|
if !sp.HasDealInfo() {
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
return sp.real.DealInfo.KeepUnsealedRequested()
|
return sp.real.DealInfo.KeepUnsealedRequested()
|
||||||
}
|
}
|
||||||
|
|
||||||
func (sp *SafeSectorPiece) GetAllocation(ctx context.Context, aapi piece.AllocationAPI, tsk types.TipSetKey) (*verifreg.Allocation, error) {
|
func (sp *SafeSectorPiece) GetAllocation(ctx context.Context, aapi piece.AllocationAPI, tsk types.TipSetKey) (*verifreg.Allocation, error) {
|
||||||
|
if !sp.HasDealInfo() {
|
||||||
|
return nil, xerrors.Errorf("no deal info")
|
||||||
|
}
|
||||||
|
|
||||||
return sp.real.DealInfo.GetAllocation(ctx, aapi, tsk)
|
return sp.real.DealInfo.GetAllocation(ctx, aapi, tsk)
|
||||||
}
|
}
|
||||||
|
Loading…
Reference in New Issue
Block a user