225 lines
7.1 KiB
Go
225 lines
7.1 KiB
Go
// stm: #integration
|
|
package itests
|
|
|
|
import (
|
|
"context"
|
|
"fmt"
|
|
"strings"
|
|
"testing"
|
|
"time"
|
|
|
|
"github.com/stretchr/testify/assert"
|
|
"github.com/stretchr/testify/require"
|
|
|
|
"github.com/filecoin-project/go-state-types/abi"
|
|
"github.com/filecoin-project/lotus/api"
|
|
"github.com/filecoin-project/lotus/build"
|
|
"github.com/filecoin-project/lotus/chain/types"
|
|
"github.com/filecoin-project/lotus/itests/kit"
|
|
"github.com/filecoin-project/lotus/node/config"
|
|
"github.com/filecoin-project/lotus/node/impl"
|
|
sealing "github.com/filecoin-project/lotus/storage/pipeline"
|
|
miner5 "github.com/filecoin-project/specs-actors/v5/actors/builtin/miner"
|
|
)
|
|
|
|
func TestPledgeSectors(t *testing.T) {
|
|
//stm: @CHAIN_SYNCER_LOAD_GENESIS_001, @CHAIN_SYNCER_FETCH_TIPSET_001,
|
|
//stm: @CHAIN_SYNCER_START_001, @CHAIN_SYNCER_SYNC_001, @BLOCKCHAIN_BEACON_VALIDATE_BLOCK_VALUES_01
|
|
//stm: @CHAIN_SYNCER_COLLECT_CHAIN_001, @CHAIN_SYNCER_COLLECT_HEADERS_001, @CHAIN_SYNCER_VALIDATE_TIPSET_001
|
|
//stm: @CHAIN_SYNCER_NEW_PEER_HEAD_001, @CHAIN_SYNCER_VALIDATE_MESSAGE_META_001, @CHAIN_SYNCER_STOP_001
|
|
|
|
//stm: @CHAIN_INCOMING_HANDLE_INCOMING_BLOCKS_001, @CHAIN_INCOMING_VALIDATE_BLOCK_PUBSUB_001, @CHAIN_INCOMING_VALIDATE_MESSAGE_PUBSUB_001
|
|
kit.QuietMiningLogs()
|
|
|
|
blockTime := 50 * time.Millisecond
|
|
|
|
runTest := func(t *testing.T, nSectors int) {
|
|
ctx, cancel := context.WithCancel(context.Background())
|
|
defer cancel()
|
|
|
|
_, miner, ens := kit.EnsembleMinimal(t, kit.MockProofs())
|
|
ens.InterconnectAll().BeginMining(blockTime)
|
|
|
|
miner.PledgeSectors(ctx, nSectors, 0, nil)
|
|
}
|
|
|
|
t.Run("1", func(t *testing.T) {
|
|
runTest(t, 1)
|
|
})
|
|
|
|
t.Run("100", func(t *testing.T) {
|
|
runTest(t, 100)
|
|
})
|
|
|
|
t.Run("1000", func(t *testing.T) {
|
|
if testing.Short() { // takes ~16s
|
|
t.Skip("skipping test in short mode")
|
|
}
|
|
|
|
runTest(t, 1000)
|
|
})
|
|
}
|
|
|
|
func TestPledgeBatching(t *testing.T) {
|
|
//stm: @SECTOR_PRE_COMMIT_FLUSH_001, @SECTOR_COMMIT_FLUSH_001
|
|
blockTime := 50 * time.Millisecond
|
|
|
|
runTest := func(t *testing.T, nSectors int) {
|
|
ctx, cancel := context.WithCancel(context.Background())
|
|
defer cancel()
|
|
|
|
client, miner, ens := kit.EnsembleMinimal(t, kit.MockProofs())
|
|
ens.InterconnectAll().BeginMining(blockTime)
|
|
|
|
client.WaitTillChain(ctx, kit.HeightAtLeast(10))
|
|
|
|
toCheck := miner.StartPledge(ctx, nSectors, 0, nil)
|
|
|
|
for len(toCheck) > 0 {
|
|
states := map[api.SectorState]int{}
|
|
|
|
for n := range toCheck {
|
|
st, err := miner.SectorsStatus(ctx, n, false)
|
|
require.NoError(t, err)
|
|
states[st.State]++
|
|
if st.State == api.SectorState(sealing.Proving) {
|
|
delete(toCheck, n)
|
|
}
|
|
if strings.Contains(string(st.State), "Fail") {
|
|
t.Fatal("sector in a failed state", st.State)
|
|
}
|
|
}
|
|
if states[api.SectorState(sealing.SubmitPreCommitBatch)] == nSectors ||
|
|
(states[api.SectorState(sealing.SubmitPreCommitBatch)] > 0 && states[api.SectorState(sealing.PreCommit1)] == 0 && states[api.SectorState(sealing.PreCommit2)] == 0) {
|
|
pcb, err := miner.SectorPreCommitFlush(ctx)
|
|
require.NoError(t, err)
|
|
if pcb != nil {
|
|
fmt.Printf("PRECOMMIT BATCH: %+v\n", pcb)
|
|
}
|
|
}
|
|
|
|
if states[api.SectorState(sealing.SubmitCommitAggregate)] == nSectors ||
|
|
(states[api.SectorState(sealing.SubmitCommitAggregate)] > 0 && states[api.SectorState(sealing.WaitSeed)] == 0 && states[api.SectorState(sealing.Committing)] == 0) {
|
|
cb, err := miner.SectorCommitFlush(ctx)
|
|
require.NoError(t, err)
|
|
if cb != nil {
|
|
fmt.Printf("COMMIT BATCH: %+v\n", cb)
|
|
}
|
|
}
|
|
|
|
build.Clock.Sleep(100 * time.Millisecond)
|
|
fmt.Printf("WaitSeal: %d %+v\n", len(toCheck), states)
|
|
}
|
|
}
|
|
|
|
t.Run("100", func(t *testing.T) {
|
|
runTest(t, 100)
|
|
})
|
|
}
|
|
|
|
func TestPledgeMaxBatching(t *testing.T) {
|
|
//stm: @CHAIN_SYNCER_LOAD_GENESIS_001, @CHAIN_SYNCER_FETCH_TIPSET_001,
|
|
//stm: @CHAIN_SYNCER_START_001, @CHAIN_SYNCER_SYNC_001, @BLOCKCHAIN_BEACON_VALIDATE_BLOCK_VALUES_01
|
|
//stm: @CHAIN_SYNCER_COLLECT_CHAIN_001, @CHAIN_SYNCER_COLLECT_HEADERS_001, @CHAIN_SYNCER_VALIDATE_TIPSET_001
|
|
//stm: @CHAIN_SYNCER_NEW_PEER_HEAD_001, @CHAIN_SYNCER_VALIDATE_MESSAGE_META_001, @CHAIN_SYNCER_STOP_001
|
|
|
|
//stm: @CHAIN_INCOMING_HANDLE_INCOMING_BLOCKS_001, @CHAIN_INCOMING_VALIDATE_BLOCK_PUBSUB_001, @CHAIN_INCOMING_VALIDATE_MESSAGE_PUBSUB_001
|
|
blockTime := 50 * time.Millisecond
|
|
|
|
runTest := func(t *testing.T) {
|
|
nSectors := miner5.MaxAggregatedSectors
|
|
ctx, cancel := context.WithCancel(context.Background())
|
|
defer cancel()
|
|
|
|
_, full, miner, ens := kit.EnsembleTwoOne(t, kit.MockProofs())
|
|
ens.InterconnectAll().BeginMining(blockTime)
|
|
m, ok := miner.StorageMiner.(*impl.StorageMinerAPI)
|
|
require.True(t, ok)
|
|
cfg, err := m.GetSealingConfigFunc()
|
|
require.NoError(t, err)
|
|
cfg.MinCommitBatch = miner5.MaxAggregatedSectors
|
|
require.NoError(t, m.SetSealingConfigFunc(cfg))
|
|
|
|
toCheck := miner.StartPledge(ctx, nSectors, 0, nil)
|
|
var lastSectorNo abi.SectorNumber
|
|
|
|
for len(toCheck) > 0 {
|
|
states := map[api.SectorState]int{}
|
|
|
|
for n := range toCheck {
|
|
lastSectorNo = n
|
|
st, err := miner.SectorsStatus(ctx, n, false)
|
|
require.NoError(t, err)
|
|
states[st.State]++
|
|
if st.State == api.SectorState(sealing.Proving) {
|
|
delete(toCheck, n)
|
|
}
|
|
if strings.Contains(string(st.State), "Fail") {
|
|
t.Fatal("sector in a failed state", st.State)
|
|
}
|
|
}
|
|
if states[api.SectorState(sealing.SubmitPreCommitBatch)] == nSectors ||
|
|
(states[api.SectorState(sealing.SubmitPreCommitBatch)] > 0 && states[api.SectorState(sealing.PreCommit1)] == 0 && states[api.SectorState(sealing.PreCommit2)] == 0) {
|
|
pcb, err := miner.SectorPreCommitFlush(ctx)
|
|
require.NoError(t, err)
|
|
if pcb != nil {
|
|
fmt.Printf("PRECOMMIT BATCH: %+v\n", pcb)
|
|
}
|
|
}
|
|
|
|
if states[api.SectorState(sealing.SubmitCommitAggregate)] == nSectors {
|
|
cb, err := miner.SectorCommitFlush(ctx)
|
|
require.NoError(t, err)
|
|
if cb != nil {
|
|
fmt.Printf("COMMIT BATCH: %+v\n", cb)
|
|
}
|
|
}
|
|
|
|
build.Clock.Sleep(100 * time.Millisecond)
|
|
fmt.Printf("WaitSeal: %d %+v\n", len(toCheck), states)
|
|
}
|
|
|
|
// Wait for flushed ProveCommitAggregate to land on chain
|
|
st, err := miner.SectorsStatus(ctx, lastSectorNo, false)
|
|
require.NoError(t, err)
|
|
for st.State == api.SectorState(sealing.CommitAggregateWait) {
|
|
build.Clock.Sleep(100 * time.Millisecond)
|
|
}
|
|
|
|
// Ensure that max aggregate message has propagated to the other node by checking current state
|
|
//stm: @CHAIN_STATE_MINER_SECTORS_001
|
|
sectorInfosAfter, err := full.StateMinerSectors(ctx, miner.ActorAddr, nil, types.EmptyTSK)
|
|
require.NoError(t, err)
|
|
assert.Equal(t, miner5.MaxAggregatedSectors+kit.DefaultPresealsPerBootstrapMiner, len(sectorInfosAfter))
|
|
}
|
|
|
|
t.Run("Force max prove commit aggregate size", runTest)
|
|
}
|
|
|
|
func TestPledgeSynth(t *testing.T) {
|
|
kit.QuietMiningLogs()
|
|
|
|
blockTime := 50 * time.Millisecond
|
|
|
|
runTest := func(t *testing.T, nSectors int) {
|
|
ctx, cancel := context.WithCancel(context.Background())
|
|
defer cancel()
|
|
|
|
_, miner, ens := kit.EnsembleMinimal(t, kit.MutateSealingConfig(func(sc *config.SealingConfig) {
|
|
sc.UseSyntheticPoRep = true
|
|
})) // no mock proofs
|
|
|
|
ens.InterconnectAll().BeginMiningMustPost(blockTime)
|
|
|
|
miner.PledgeSectors(ctx, nSectors, 0, nil)
|
|
}
|
|
|
|
t.Run("1", func(t *testing.T) {
|
|
runTest(t, 1)
|
|
})
|
|
|
|
t.Run("3", func(t *testing.T) {
|
|
runTest(t, 3)
|
|
})
|
|
}
|