deals tests: begin migration.
This commit is contained in:
parent
cf0150e057
commit
329970934a
1
go.mod
1
go.mod
@ -16,6 +16,7 @@ require (
|
||||
github.com/chzyer/readline v0.0.0-20180603132655-2972be24d48e
|
||||
github.com/cockroachdb/pebble v0.0.0-20201001221639-879f3bfeef07
|
||||
github.com/coreos/go-systemd/v22 v22.1.0
|
||||
github.com/davecgh/go-spew v1.1.1 // indirect
|
||||
github.com/detailyang/go-fallocate v0.0.0-20180908115635-432fa640bd2e
|
||||
github.com/dgraph-io/badger/v2 v2.2007.2
|
||||
github.com/docker/go-units v0.4.0
|
||||
|
508
itests/deals_test.go
Normal file
508
itests/deals_test.go
Normal file
@ -0,0 +1,508 @@
|
||||
package itests
|
||||
|
||||
import (
|
||||
"bytes"
|
||||
"context"
|
||||
"fmt"
|
||||
"testing"
|
||||
"time"
|
||||
|
||||
"github.com/davecgh/go-spew/spew"
|
||||
"github.com/filecoin-project/go-fil-markets/storagemarket"
|
||||
"github.com/filecoin-project/go-state-types/abi"
|
||||
"github.com/filecoin-project/lotus/api"
|
||||
"github.com/filecoin-project/lotus/chain/actors/builtin/market"
|
||||
"github.com/filecoin-project/lotus/chain/types"
|
||||
"github.com/filecoin-project/lotus/itests/kit"
|
||||
"github.com/filecoin-project/lotus/markets/storageadapter"
|
||||
"github.com/filecoin-project/lotus/node"
|
||||
market2 "github.com/filecoin-project/specs-actors/v2/actors/builtin/market"
|
||||
"github.com/stretchr/testify/require"
|
||||
)
|
||||
|
||||
//
|
||||
// func TestDealCycle(t *testing.T) {
|
||||
// kit.QuietMiningLogs()
|
||||
//
|
||||
// blockTime := 10 * time.Millisecond
|
||||
//
|
||||
// // For these tests where the block time is artificially short, just use
|
||||
// // a deal start epoch that is guaranteed to be far enough in the future
|
||||
// // so that the deal starts sealing in time
|
||||
// dealStartEpoch := abi.ChainEpoch(2 << 12)
|
||||
//
|
||||
// t.Run("TestFullDealCycle_Single", func(t *testing.T) {
|
||||
// runFullDealCycles(t, 1, kit.MockMinerBuilder, blockTime, false, false, dealStartEpoch)
|
||||
// })
|
||||
// t.Run("TestFullDealCycle_Two", func(t *testing.T) {
|
||||
// runFullDealCycles(t, 2, kit.MockMinerBuilder, blockTime, false, false, dealStartEpoch)
|
||||
// })
|
||||
// t.Run("WithExportedCAR", func(t *testing.T) {
|
||||
// runFullDealCycles(t, 1, kit.MockMinerBuilder, blockTime, true, false, dealStartEpoch)
|
||||
// })
|
||||
// t.Run("TestFastRetrievalDealCycle", func(t *testing.T) {
|
||||
// runFastRetrievalDealFlowT(t, kit.MockMinerBuilder, blockTime, dealStartEpoch)
|
||||
// })
|
||||
// t.Run("TestZeroPricePerByteRetrievalDealFlow", func(t *testing.T) {
|
||||
// runZeroPricePerByteRetrievalDealFlow(t, kit.MockMinerBuilder, blockTime, dealStartEpoch)
|
||||
// })
|
||||
// }
|
||||
//
|
||||
// func TestAPIDealFlowReal(t *testing.T) {
|
||||
// if testing.Short() {
|
||||
// t.Skip("skipping test in short mode")
|
||||
// }
|
||||
//
|
||||
// kit.QuietMiningLogs()
|
||||
//
|
||||
// // TODO: just set this globally?
|
||||
// oldDelay := policy.GetPreCommitChallengeDelay()
|
||||
// policy.SetPreCommitChallengeDelay(5)
|
||||
// t.Cleanup(func() {
|
||||
// policy.SetPreCommitChallengeDelay(oldDelay)
|
||||
// })
|
||||
//
|
||||
// t.Run("basic", func(t *testing.T) {
|
||||
// runFullDealCycles(t, 1, kit.FullNodeBuilder, time.Second, false, false, 0)
|
||||
// })
|
||||
//
|
||||
// t.Run("fast-retrieval", func(t *testing.T) {
|
||||
// runFullDealCycles(t, 1, kit.FullNodeBuilder, time.Second, false, true, 0)
|
||||
// })
|
||||
//
|
||||
// t.Run("retrieval-second", func(t *testing.T) {
|
||||
// runSecondDealRetrievalTest(t, kit.FullNodeBuilder, time.Second)
|
||||
// })
|
||||
// }
|
||||
|
||||
func TestPublishDealsBatching(t *testing.T) {
|
||||
var (
|
||||
ctx = context.Background()
|
||||
publishPeriod = 10 * time.Second
|
||||
maxDealsPerMsg = uint64(2) // Set max deals per publish deals message to 2
|
||||
startEpoch = abi.ChainEpoch(2 << 12)
|
||||
)
|
||||
|
||||
kit.QuietMiningLogs()
|
||||
|
||||
opts := node.Override(new(*storageadapter.DealPublisher),
|
||||
storageadapter.NewDealPublisher(nil, storageadapter.PublishMsgConfig{
|
||||
Period: publishPeriod,
|
||||
MaxDealsPerMsg: maxDealsPerMsg,
|
||||
}),
|
||||
)
|
||||
|
||||
client, miner, ens := kit.EnsembleMinimum(t, kit.MockProofs(), kit.ExtraNodeOpts(opts))
|
||||
ens.InterconnectAll().BeginMining(10 * time.Millisecond)
|
||||
|
||||
dh := kit.NewDealHarness(t, client, miner)
|
||||
|
||||
fmt.Println("***********************")
|
||||
spew.Dump(client.NetPeers(context.Background()))
|
||||
|
||||
// Starts a deal and waits until it's published
|
||||
runDealTillPublish := func(rseed int) {
|
||||
res, _, _, err := kit.CreateImportFile(ctx, client, rseed, 0)
|
||||
require.NoError(t, err)
|
||||
|
||||
upds, err := client.ClientGetDealUpdates(ctx)
|
||||
require.NoError(t, err)
|
||||
|
||||
dh.StartDeal(ctx, res.Root, false, startEpoch)
|
||||
|
||||
// TODO: this sleep is only necessary because deals don't immediately get logged in the dealstore, we should fix this
|
||||
time.Sleep(time.Second)
|
||||
|
||||
done := make(chan struct{})
|
||||
go func() {
|
||||
for upd := range upds {
|
||||
if upd.DataRef.Root == res.Root && upd.State == storagemarket.StorageDealAwaitingPreCommit {
|
||||
done <- struct{}{}
|
||||
}
|
||||
}
|
||||
}()
|
||||
<-done
|
||||
}
|
||||
|
||||
// Run three deals in parallel
|
||||
done := make(chan struct{}, maxDealsPerMsg+1)
|
||||
for rseed := 1; rseed <= 3; rseed++ {
|
||||
rseed := rseed
|
||||
go func() {
|
||||
runDealTillPublish(rseed)
|
||||
done <- struct{}{}
|
||||
}()
|
||||
}
|
||||
|
||||
// Wait for two of the deals to be published
|
||||
for i := 0; i < int(maxDealsPerMsg); i++ {
|
||||
<-done
|
||||
}
|
||||
|
||||
// Expect a single PublishStorageDeals message that includes the first two deals
|
||||
msgCids, err := client.StateListMessages(ctx, &api.MessageMatch{To: market.Address}, types.EmptyTSK, 1)
|
||||
require.NoError(t, err)
|
||||
count := 0
|
||||
for _, msgCid := range msgCids {
|
||||
msg, err := client.ChainGetMessage(ctx, msgCid)
|
||||
require.NoError(t, err)
|
||||
|
||||
if msg.Method == market.Methods.PublishStorageDeals {
|
||||
count++
|
||||
var pubDealsParams market2.PublishStorageDealsParams
|
||||
err = pubDealsParams.UnmarshalCBOR(bytes.NewReader(msg.Params))
|
||||
require.NoError(t, err)
|
||||
require.Len(t, pubDealsParams.Deals, int(maxDealsPerMsg))
|
||||
}
|
||||
}
|
||||
require.Equal(t, 1, count)
|
||||
|
||||
// The third deal should be published once the publish period expires.
|
||||
// Allow a little padding as it takes a moment for the state change to
|
||||
// be noticed by the client.
|
||||
padding := 10 * time.Second
|
||||
select {
|
||||
case <-time.After(publishPeriod + padding):
|
||||
require.Fail(t, "Expected 3rd deal to be published once publish period elapsed")
|
||||
case <-done: // Success
|
||||
}
|
||||
}
|
||||
|
||||
//
|
||||
// func TestDealMining(t *testing.T) {
|
||||
// // test making a deal with a fresh miner, and see if it starts to mine.
|
||||
// if testing.Short() {
|
||||
// t.Skip("skipping test in short mode")
|
||||
// }
|
||||
//
|
||||
// kit.QuietMiningLogs()
|
||||
//
|
||||
// b := kit.MockMinerBuilder
|
||||
// blocktime := 50 * time.Millisecond
|
||||
//
|
||||
// ctx := context.Background()
|
||||
// fulls, miners := b(t,
|
||||
// kit.OneFull,
|
||||
// []kit.StorageMiner{
|
||||
// {Full: 0, Preseal: kit.PresealGenesis},
|
||||
// {Full: 0, Preseal: 0}, // TODO: Add support for miners on non-first full node
|
||||
// })
|
||||
// client := fulls[0].FullNode.(*impl.FullNodeAPI)
|
||||
// genesisMiner := miners[0]
|
||||
// provider := miners[1]
|
||||
//
|
||||
// addrinfo, err := client.NetAddrsListen(ctx)
|
||||
// if err != nil {
|
||||
// t.Fatal(err)
|
||||
// }
|
||||
//
|
||||
// if err := provider.NetConnect(ctx, addrinfo); err != nil {
|
||||
// t.Fatal(err)
|
||||
// }
|
||||
//
|
||||
// if err := genesisMiner.NetConnect(ctx, addrinfo); err != nil {
|
||||
// t.Fatal(err)
|
||||
// }
|
||||
//
|
||||
// time.Sleep(time.Second)
|
||||
//
|
||||
// data := make([]byte, 600)
|
||||
// rand.New(rand.NewSource(5)).Read(data)
|
||||
//
|
||||
// r := bytes.NewReader(data)
|
||||
// fcid, err := client.ClientImportLocal(ctx, r)
|
||||
// if err != nil {
|
||||
// t.Fatal(err)
|
||||
// }
|
||||
//
|
||||
// fmt.Println("FILE CID: ", fcid)
|
||||
//
|
||||
// var mine int32 = 1
|
||||
// done := make(chan struct{})
|
||||
// minedTwo := make(chan struct{})
|
||||
//
|
||||
// m2addr, err := miners[1].ActorAddress(context.TODO())
|
||||
// if err != nil {
|
||||
// t.Fatal(err)
|
||||
// }
|
||||
//
|
||||
// go func() {
|
||||
// defer close(done)
|
||||
//
|
||||
// complChan := minedTwo
|
||||
// for atomic.LoadInt32(&mine) != 0 {
|
||||
// wait := make(chan int)
|
||||
// mdone := func(mined bool, _ abi.ChainEpoch, err error) {
|
||||
// n := 0
|
||||
// if mined {
|
||||
// n = 1
|
||||
// }
|
||||
// wait <- n
|
||||
// }
|
||||
//
|
||||
// if err := miners[0].MineOne(ctx, miner.MineReq{Done: mdone}); err != nil {
|
||||
// t.Error(err)
|
||||
// }
|
||||
//
|
||||
// if err := miners[1].MineOne(ctx, miner.MineReq{Done: mdone}); err != nil {
|
||||
// t.Error(err)
|
||||
// }
|
||||
//
|
||||
// expect := <-wait
|
||||
// expect += <-wait
|
||||
//
|
||||
// time.Sleep(blocktime)
|
||||
// if expect == 0 {
|
||||
// // null block
|
||||
// continue
|
||||
// }
|
||||
//
|
||||
// var nodeOneMined bool
|
||||
// for _, node := range miners {
|
||||
// mb, err := node.MiningBase(ctx)
|
||||
// if err != nil {
|
||||
// t.Error(err)
|
||||
// return
|
||||
// }
|
||||
//
|
||||
// for _, b := range mb.Blocks() {
|
||||
// if b.Miner == m2addr {
|
||||
// nodeOneMined = true
|
||||
// break
|
||||
// }
|
||||
// }
|
||||
//
|
||||
// }
|
||||
//
|
||||
// if nodeOneMined && complChan != nil {
|
||||
// close(complChan)
|
||||
// complChan = nil
|
||||
// }
|
||||
//
|
||||
// }
|
||||
// }()
|
||||
//
|
||||
// dh := kit.NewDealHarness(t, client, provider)
|
||||
//
|
||||
// deal := dh.StartDeal(ctx, fcid, false, 0)
|
||||
//
|
||||
// // TODO: this sleep is only necessary because deals don't immediately get logged in the dealstore, we should fix this
|
||||
// time.Sleep(time.Second)
|
||||
//
|
||||
// dh.WaitDealSealed(ctx, deal, false, false, nil)
|
||||
//
|
||||
// <-minedTwo
|
||||
//
|
||||
// atomic.StoreInt32(&mine, 0)
|
||||
// fmt.Println("shutting down mining")
|
||||
// <-done
|
||||
// }
|
||||
//
|
||||
// func TestOfflineDealFlow(t *testing.T) {
|
||||
// blocktime := 10 * time.Millisecond
|
||||
//
|
||||
// // For these tests where the block time is artificially short, just use
|
||||
// // a deal start epoch that is guaranteed to be far enough in the future
|
||||
// // so that the deal starts sealing in time
|
||||
// startEpoch := abi.ChainEpoch(2 << 12)
|
||||
//
|
||||
// runTest := func(t *testing.T, fastRet bool) {
|
||||
// ctx := context.Background()
|
||||
// fulls, miners := kit.MockMinerBuilder(t, kit.OneFull, kit.OneMiner)
|
||||
// client, miner := fulls[0].FullNode.(*impl.FullNodeAPI), miners[0]
|
||||
//
|
||||
// kit.ConnectAndStartMining(t, blocktime, miner, client)
|
||||
//
|
||||
// dh := kit.NewDealHarness(t, client, miner)
|
||||
//
|
||||
// // Create a random file and import on the client.
|
||||
// res, path, data, err := kit.CreateImportFile(ctx, client, 1, 0)
|
||||
// require.NoError(t, err)
|
||||
//
|
||||
// // Get the piece size and commP
|
||||
// fcid := res.Root
|
||||
// pieceInfo, err := client.ClientDealPieceCID(ctx, fcid)
|
||||
// require.NoError(t, err)
|
||||
// fmt.Println("FILE CID: ", fcid)
|
||||
//
|
||||
// // Create a storage deal with the miner
|
||||
// maddr, err := miner.ActorAddress(ctx)
|
||||
// require.NoError(t, err)
|
||||
//
|
||||
// addr, err := client.WalletDefaultAddress(ctx)
|
||||
// require.NoError(t, err)
|
||||
//
|
||||
// // Manual storage deal (offline deal)
|
||||
// dataRef := &storagemarket.DataRef{
|
||||
// TransferType: storagemarket.TTManual,
|
||||
// Root: fcid,
|
||||
// PieceCid: &pieceInfo.PieceCID,
|
||||
// PieceSize: pieceInfo.PieceSize.Unpadded(),
|
||||
// }
|
||||
//
|
||||
// proposalCid, err := client.ClientStartDeal(ctx, &api.StartDealParams{
|
||||
// Data: dataRef,
|
||||
// Wallet: addr,
|
||||
// Miner: maddr,
|
||||
// EpochPrice: types.NewInt(1000000),
|
||||
// DealStartEpoch: startEpoch,
|
||||
// MinBlocksDuration: uint64(build.MinDealDuration),
|
||||
// FastRetrieval: fastRet,
|
||||
// })
|
||||
// require.NoError(t, err)
|
||||
//
|
||||
// // Wait for the deal to reach StorageDealCheckForAcceptance on the client
|
||||
// cd, err := client.ClientGetDealInfo(ctx, *proposalCid)
|
||||
// require.NoError(t, err)
|
||||
// require.Eventually(t, func() bool {
|
||||
// cd, _ := client.ClientGetDealInfo(ctx, *proposalCid)
|
||||
// return cd.State == storagemarket.StorageDealCheckForAcceptance
|
||||
// }, 30*time.Second, 1*time.Second, "actual deal status is %s", storagemarket.DealStates[cd.State])
|
||||
//
|
||||
// // Create a CAR file from the raw file
|
||||
// carFileDir, err := ioutil.TempDir(os.TempDir(), "test-make-deal-car")
|
||||
// require.NoError(t, err)
|
||||
// carFilePath := filepath.Join(carFileDir, "out.car")
|
||||
// err = client.ClientGenCar(ctx, api.FileRef{Path: path}, carFilePath)
|
||||
// require.NoError(t, err)
|
||||
//
|
||||
// // Import the CAR file on the miner - this is the equivalent to
|
||||
// // transferring the file across the wire in a normal (non-offline) deal
|
||||
// err = miner.DealsImportData(ctx, *proposalCid, carFilePath)
|
||||
// require.NoError(t, err)
|
||||
//
|
||||
// // Wait for the deal to be published
|
||||
// dh.WaitDealPublished(ctx, proposalCid)
|
||||
//
|
||||
// t.Logf("deal published, retrieving")
|
||||
//
|
||||
// // Retrieve the deal
|
||||
// dh.PerformRetrieval(ctx, fcid, &pieceInfo.PieceCID, false, data)
|
||||
// }
|
||||
//
|
||||
// t.Run("NormalRetrieval", func(t *testing.T) {
|
||||
// runTest(t, false)
|
||||
// })
|
||||
// t.Run("FastRetrieval", func(t *testing.T) {
|
||||
// runTest(t, true)
|
||||
// })
|
||||
//
|
||||
// }
|
||||
//
|
||||
// func runFullDealCycles(t *testing.T, n int, b kit.APIBuilder, blocktime time.Duration, carExport, fastRet bool, startEpoch abi.ChainEpoch) {
|
||||
// full, _, ens := kit.EnsembleMinimum(t)
|
||||
// ens.BeginMining()
|
||||
// dh := kit.NewDealHarness(t, client, miner)
|
||||
//
|
||||
// baseseed := 6
|
||||
// for i := 0; i < n; i++ {
|
||||
// dh.MakeFullDeal(context.Background(), baseseed+i, carExport, fastRet, startEpoch)
|
||||
// }
|
||||
// }
|
||||
//
|
||||
// func runFastRetrievalDealFlowT(t *testing.T, b kit.APIBuilder, blocktime time.Duration, startEpoch abi.ChainEpoch) {
|
||||
// ctx := context.Background()
|
||||
//
|
||||
// var (
|
||||
// nb = kit.NewNodeBuilder(t)
|
||||
// full = nb.FullNode()
|
||||
// miner = nb.Miner(full)
|
||||
// )
|
||||
//
|
||||
// nb.Create()
|
||||
//
|
||||
// kit.ConnectAndStartMining(t, blocktime, miner, full)
|
||||
//
|
||||
// dh := kit.NewDealHarness(t, full, miner)
|
||||
// data := make([]byte, 1600)
|
||||
// rand.New(rand.NewSource(int64(8))).Read(data)
|
||||
//
|
||||
// r := bytes.NewReader(data)
|
||||
// fcid, err := full.FullNode.(*impl.FullNodeAPI).ClientImportLocal(ctx, r)
|
||||
// require.NoError(t, err)
|
||||
//
|
||||
// fmt.Println("FILE CID: ", fcid)
|
||||
//
|
||||
// deal := dh.StartDeal(ctx, fcid, true, startEpoch)
|
||||
// dh.WaitDealPublished(ctx, deal)
|
||||
//
|
||||
// fmt.Println("deal published, retrieving")
|
||||
//
|
||||
// // Retrieval
|
||||
// info, err := full.ClientGetDealInfo(ctx, *deal)
|
||||
// require.NoError(t, err)
|
||||
//
|
||||
// dh.PerformRetrieval(ctx, fcid, &info.PieceCID, false, data)
|
||||
// }
|
||||
//
|
||||
// func runSecondDealRetrievalTest(t *testing.T, b kit.APIBuilder, blocktime time.Duration) {
|
||||
// ctx := context.Background()
|
||||
//
|
||||
// fulls, miners := b(t, kit.OneFull, kit.OneMiner)
|
||||
// client, miner := fulls[0].FullNode.(*impl.FullNodeAPI), miners[0]
|
||||
//
|
||||
// kit.ConnectAndStartMining(t, blocktime, miner, client)
|
||||
//
|
||||
// dh := kit.NewDealHarness(t, client, miner)
|
||||
//
|
||||
// {
|
||||
// data1 := make([]byte, 800)
|
||||
// rand.New(rand.NewSource(int64(3))).Read(data1)
|
||||
// r := bytes.NewReader(data1)
|
||||
//
|
||||
// fcid1, err := client.ClientImportLocal(ctx, r)
|
||||
// if err != nil {
|
||||
// t.Fatal(err)
|
||||
// }
|
||||
//
|
||||
// data2 := make([]byte, 800)
|
||||
// rand.New(rand.NewSource(int64(9))).Read(data2)
|
||||
// r2 := bytes.NewReader(data2)
|
||||
//
|
||||
// fcid2, err := client.ClientImportLocal(ctx, r2)
|
||||
// if err != nil {
|
||||
// t.Fatal(err)
|
||||
// }
|
||||
//
|
||||
// deal1 := dh.StartDeal(ctx, fcid1, true, 0)
|
||||
//
|
||||
// // TODO: this sleep is only necessary because deals don't immediately get logged in the dealstore, we should fix this
|
||||
// time.Sleep(time.Second)
|
||||
// dh.WaitDealSealed(ctx, deal1, true, false, nil)
|
||||
//
|
||||
// deal2 := dh.StartDeal(ctx, fcid2, true, 0)
|
||||
//
|
||||
// time.Sleep(time.Second)
|
||||
// dh.WaitDealSealed(ctx, deal2, false, false, nil)
|
||||
//
|
||||
// // Retrieval
|
||||
// info, err := client.ClientGetDealInfo(ctx, *deal2)
|
||||
// require.NoError(t, err)
|
||||
//
|
||||
// rf, _ := miner.SectorsRefs(ctx)
|
||||
// fmt.Printf("refs: %+v\n", rf)
|
||||
//
|
||||
// dh.PerformRetrieval(ctx, fcid2, &info.PieceCID, false, data2)
|
||||
// }
|
||||
// }
|
||||
//
|
||||
// func runZeroPricePerByteRetrievalDealFlow(t *testing.T, b kit.APIBuilder, blocktime time.Duration, startEpoch abi.ChainEpoch) {
|
||||
// ctx := context.Background()
|
||||
//
|
||||
// fulls, miners := b(t, kit.OneFull, kit.OneMiner)
|
||||
// client, miner := fulls[0].FullNode.(*impl.FullNodeAPI), miners[0]
|
||||
//
|
||||
// kit.ConnectAndStartMining(t, blocktime, miner, client)
|
||||
//
|
||||
// dh := kit.NewDealHarness(t, client, miner)
|
||||
//
|
||||
// // Set price-per-byte to zero
|
||||
// ask, err := miner.MarketGetRetrievalAsk(ctx)
|
||||
// require.NoError(t, err)
|
||||
//
|
||||
// ask.PricePerByte = abi.NewTokenAmount(0)
|
||||
// err = miner.MarketSetRetrievalAsk(ctx, ask)
|
||||
// require.NoError(t, err)
|
||||
//
|
||||
// dh.MakeFullDeal(ctx, 6, false, false, startEpoch)
|
||||
// }
|
@ -1,527 +0,0 @@
|
||||
package itests
|
||||
|
||||
import (
|
||||
"bytes"
|
||||
"context"
|
||||
"fmt"
|
||||
"io/ioutil"
|
||||
"math/rand"
|
||||
"os"
|
||||
"path/filepath"
|
||||
"sync/atomic"
|
||||
"testing"
|
||||
"time"
|
||||
|
||||
"github.com/filecoin-project/go-fil-markets/storagemarket"
|
||||
"github.com/filecoin-project/go-state-types/abi"
|
||||
"github.com/filecoin-project/lotus/api"
|
||||
"github.com/filecoin-project/lotus/build"
|
||||
"github.com/filecoin-project/lotus/chain/actors/builtin/market"
|
||||
"github.com/filecoin-project/lotus/chain/actors/policy"
|
||||
"github.com/filecoin-project/lotus/chain/types"
|
||||
"github.com/filecoin-project/lotus/itests/kit"
|
||||
"github.com/filecoin-project/lotus/markets/storageadapter"
|
||||
"github.com/filecoin-project/lotus/miner"
|
||||
"github.com/filecoin-project/lotus/node"
|
||||
"github.com/filecoin-project/lotus/node/impl"
|
||||
"github.com/filecoin-project/lotus/node/impl/client"
|
||||
market2 "github.com/filecoin-project/specs-actors/v2/actors/builtin/market"
|
||||
"github.com/stretchr/testify/require"
|
||||
)
|
||||
|
||||
func TestDealCycle(t *testing.T) {
|
||||
kit.QuietMiningLogs()
|
||||
|
||||
blockTime := 10 * time.Millisecond
|
||||
|
||||
// For these tests where the block time is artificially short, just use
|
||||
// a deal start epoch that is guaranteed to be far enough in the future
|
||||
// so that the deal starts sealing in time
|
||||
dealStartEpoch := abi.ChainEpoch(2 << 12)
|
||||
|
||||
t.Run("TestFullDealCycle_Single", func(t *testing.T) {
|
||||
runFullDealCycles(t, 1, kit.MockMinerBuilder, blockTime, false, false, dealStartEpoch)
|
||||
})
|
||||
t.Run("TestFullDealCycle_Two", func(t *testing.T) {
|
||||
runFullDealCycles(t, 2, kit.MockMinerBuilder, blockTime, false, false, dealStartEpoch)
|
||||
})
|
||||
t.Run("WithExportedCAR", func(t *testing.T) {
|
||||
runFullDealCycles(t, 1, kit.MockMinerBuilder, blockTime, true, false, dealStartEpoch)
|
||||
})
|
||||
t.Run("TestFastRetrievalDealCycle", func(t *testing.T) {
|
||||
runFastRetrievalDealFlowT(t, kit.MockMinerBuilder, blockTime, dealStartEpoch)
|
||||
})
|
||||
t.Run("TestZeroPricePerByteRetrievalDealFlow", func(t *testing.T) {
|
||||
runZeroPricePerByteRetrievalDealFlow(t, kit.MockMinerBuilder, blockTime, dealStartEpoch)
|
||||
})
|
||||
}
|
||||
|
||||
func TestAPIDealFlowReal(t *testing.T) {
|
||||
if testing.Short() {
|
||||
t.Skip("skipping test in short mode")
|
||||
}
|
||||
|
||||
kit.QuietMiningLogs()
|
||||
|
||||
// TODO: just set this globally?
|
||||
oldDelay := policy.GetPreCommitChallengeDelay()
|
||||
policy.SetPreCommitChallengeDelay(5)
|
||||
t.Cleanup(func() {
|
||||
policy.SetPreCommitChallengeDelay(oldDelay)
|
||||
})
|
||||
|
||||
t.Run("basic", func(t *testing.T) {
|
||||
runFullDealCycles(t, 1, kit.FullNodeBuilder, time.Second, false, false, 0)
|
||||
})
|
||||
|
||||
t.Run("fast-retrieval", func(t *testing.T) {
|
||||
runFullDealCycles(t, 1, kit.FullNodeBuilder, time.Second, false, true, 0)
|
||||
})
|
||||
|
||||
t.Run("retrieval-second", func(t *testing.T) {
|
||||
runSecondDealRetrievalTest(t, kit.FullNodeBuilder, time.Second)
|
||||
})
|
||||
}
|
||||
|
||||
func TestPublishDealsBatching(t *testing.T) {
|
||||
ctx := context.Background()
|
||||
|
||||
kit.QuietMiningLogs()
|
||||
|
||||
b := kit.MockMinerBuilder
|
||||
blocktime := 10 * time.Millisecond
|
||||
startEpoch := abi.ChainEpoch(2 << 12)
|
||||
|
||||
publishPeriod := 10 * time.Second
|
||||
maxDealsPerMsg := uint64(2)
|
||||
|
||||
// Set max deals per publish deals message to 2
|
||||
minerDef := []kit.StorageMiner{{
|
||||
Full: 0,
|
||||
Opts: node.Override(
|
||||
new(*storageadapter.DealPublisher),
|
||||
storageadapter.NewDealPublisher(nil, storageadapter.PublishMsgConfig{
|
||||
Period: publishPeriod,
|
||||
MaxDealsPerMsg: maxDealsPerMsg,
|
||||
})),
|
||||
Preseal: kit.PresealGenesis,
|
||||
}}
|
||||
|
||||
// Create a connect client and miner node
|
||||
n, sn := b(t, kit.OneFull, minerDef)
|
||||
client := n[0].FullNode.(*impl.FullNodeAPI)
|
||||
miner := sn[0]
|
||||
|
||||
kit.ConnectAndStartMining(t, blocktime, miner, client)
|
||||
|
||||
dh := kit.NewDealHarness(t, client, miner)
|
||||
|
||||
// Starts a deal and waits until it's published
|
||||
runDealTillPublish := func(rseed int) {
|
||||
res, _, _, err := kit.CreateImportFile(ctx, client, rseed, 0)
|
||||
require.NoError(t, err)
|
||||
|
||||
upds, err := client.ClientGetDealUpdates(ctx)
|
||||
require.NoError(t, err)
|
||||
|
||||
dh.StartDeal(ctx, res.Root, false, startEpoch)
|
||||
|
||||
// TODO: this sleep is only necessary because deals don't immediately get logged in the dealstore, we should fix this
|
||||
time.Sleep(time.Second)
|
||||
|
||||
done := make(chan struct{})
|
||||
go func() {
|
||||
for upd := range upds {
|
||||
if upd.DataRef.Root == res.Root && upd.State == storagemarket.StorageDealAwaitingPreCommit {
|
||||
done <- struct{}{}
|
||||
}
|
||||
}
|
||||
}()
|
||||
<-done
|
||||
}
|
||||
|
||||
// Run three deals in parallel
|
||||
done := make(chan struct{}, maxDealsPerMsg+1)
|
||||
for rseed := 1; rseed <= 3; rseed++ {
|
||||
rseed := rseed
|
||||
go func() {
|
||||
runDealTillPublish(rseed)
|
||||
done <- struct{}{}
|
||||
}()
|
||||
}
|
||||
|
||||
// Wait for two of the deals to be published
|
||||
for i := 0; i < int(maxDealsPerMsg); i++ {
|
||||
<-done
|
||||
}
|
||||
|
||||
// Expect a single PublishStorageDeals message that includes the first two deals
|
||||
msgCids, err := client.StateListMessages(ctx, &api.MessageMatch{To: market.Address}, types.EmptyTSK, 1)
|
||||
require.NoError(t, err)
|
||||
count := 0
|
||||
for _, msgCid := range msgCids {
|
||||
msg, err := client.ChainGetMessage(ctx, msgCid)
|
||||
require.NoError(t, err)
|
||||
|
||||
if msg.Method == market.Methods.PublishStorageDeals {
|
||||
count++
|
||||
var pubDealsParams market2.PublishStorageDealsParams
|
||||
err = pubDealsParams.UnmarshalCBOR(bytes.NewReader(msg.Params))
|
||||
require.NoError(t, err)
|
||||
require.Len(t, pubDealsParams.Deals, int(maxDealsPerMsg))
|
||||
}
|
||||
}
|
||||
require.Equal(t, 1, count)
|
||||
|
||||
// The third deal should be published once the publish period expires.
|
||||
// Allow a little padding as it takes a moment for the state change to
|
||||
// be noticed by the client.
|
||||
padding := 10 * time.Second
|
||||
select {
|
||||
case <-time.After(publishPeriod + padding):
|
||||
require.Fail(t, "Expected 3rd deal to be published once publish period elapsed")
|
||||
case <-done: // Success
|
||||
}
|
||||
}
|
||||
|
||||
func TestDealMining(t *testing.T) {
|
||||
// test making a deal with a fresh miner, and see if it starts to mine.
|
||||
if testing.Short() {
|
||||
t.Skip("skipping test in short mode")
|
||||
}
|
||||
|
||||
kit.QuietMiningLogs()
|
||||
|
||||
b := kit.MockMinerBuilder
|
||||
blocktime := 50 * time.Millisecond
|
||||
|
||||
ctx := context.Background()
|
||||
fulls, miners := b(t,
|
||||
kit.OneFull,
|
||||
[]kit.StorageMiner{
|
||||
{Full: 0, Preseal: kit.PresealGenesis},
|
||||
{Full: 0, Preseal: 0}, // TODO: Add support for miners on non-first full node
|
||||
})
|
||||
client := fulls[0].FullNode.(*impl.FullNodeAPI)
|
||||
genesisMiner := miners[0]
|
||||
provider := miners[1]
|
||||
|
||||
addrinfo, err := client.NetAddrsListen(ctx)
|
||||
if err != nil {
|
||||
t.Fatal(err)
|
||||
}
|
||||
|
||||
if err := provider.NetConnect(ctx, addrinfo); err != nil {
|
||||
t.Fatal(err)
|
||||
}
|
||||
|
||||
if err := genesisMiner.NetConnect(ctx, addrinfo); err != nil {
|
||||
t.Fatal(err)
|
||||
}
|
||||
|
||||
time.Sleep(time.Second)
|
||||
|
||||
data := make([]byte, 600)
|
||||
rand.New(rand.NewSource(5)).Read(data)
|
||||
|
||||
r := bytes.NewReader(data)
|
||||
fcid, err := client.ClientImportLocal(ctx, r)
|
||||
if err != nil {
|
||||
t.Fatal(err)
|
||||
}
|
||||
|
||||
fmt.Println("FILE CID: ", fcid)
|
||||
|
||||
var mine int32 = 1
|
||||
done := make(chan struct{})
|
||||
minedTwo := make(chan struct{})
|
||||
|
||||
m2addr, err := miners[1].ActorAddress(context.TODO())
|
||||
if err != nil {
|
||||
t.Fatal(err)
|
||||
}
|
||||
|
||||
go func() {
|
||||
defer close(done)
|
||||
|
||||
complChan := minedTwo
|
||||
for atomic.LoadInt32(&mine) != 0 {
|
||||
wait := make(chan int)
|
||||
mdone := func(mined bool, _ abi.ChainEpoch, err error) {
|
||||
n := 0
|
||||
if mined {
|
||||
n = 1
|
||||
}
|
||||
wait <- n
|
||||
}
|
||||
|
||||
if err := miners[0].MineOne(ctx, miner.MineReq{Done: mdone}); err != nil {
|
||||
t.Error(err)
|
||||
}
|
||||
|
||||
if err := miners[1].MineOne(ctx, miner.MineReq{Done: mdone}); err != nil {
|
||||
t.Error(err)
|
||||
}
|
||||
|
||||
expect := <-wait
|
||||
expect += <-wait
|
||||
|
||||
time.Sleep(blocktime)
|
||||
if expect == 0 {
|
||||
// null block
|
||||
continue
|
||||
}
|
||||
|
||||
var nodeOneMined bool
|
||||
for _, node := range miners {
|
||||
mb, err := node.MiningBase(ctx)
|
||||
if err != nil {
|
||||
t.Error(err)
|
||||
return
|
||||
}
|
||||
|
||||
for _, b := range mb.Blocks() {
|
||||
if b.Miner == m2addr {
|
||||
nodeOneMined = true
|
||||
break
|
||||
}
|
||||
}
|
||||
|
||||
}
|
||||
|
||||
if nodeOneMined && complChan != nil {
|
||||
close(complChan)
|
||||
complChan = nil
|
||||
}
|
||||
|
||||
}
|
||||
}()
|
||||
|
||||
dh := kit.NewDealHarness(t, client, provider)
|
||||
|
||||
deal := dh.StartDeal(ctx, fcid, false, 0)
|
||||
|
||||
// TODO: this sleep is only necessary because deals don't immediately get logged in the dealstore, we should fix this
|
||||
time.Sleep(time.Second)
|
||||
|
||||
dh.WaitDealSealed(ctx, deal, false, false, nil)
|
||||
|
||||
<-minedTwo
|
||||
|
||||
atomic.StoreInt32(&mine, 0)
|
||||
fmt.Println("shutting down mining")
|
||||
<-done
|
||||
}
|
||||
|
||||
func TestOfflineDealFlow(t *testing.T) {
|
||||
blocktime := 10 * time.Millisecond
|
||||
|
||||
// For these tests where the block time is artificially short, just use
|
||||
// a deal start epoch that is guaranteed to be far enough in the future
|
||||
// so that the deal starts sealing in time
|
||||
startEpoch := abi.ChainEpoch(2 << 12)
|
||||
|
||||
runTest := func(t *testing.T, fastRet bool) {
|
||||
ctx := context.Background()
|
||||
fulls, miners := kit.MockMinerBuilder(t, kit.OneFull, kit.OneMiner)
|
||||
client, miner := fulls[0].FullNode.(*impl.FullNodeAPI), miners[0]
|
||||
|
||||
kit.ConnectAndStartMining(t, blocktime, miner, client)
|
||||
|
||||
dh := kit.NewDealHarness(t, client, miner)
|
||||
|
||||
// Create a random file and import on the client.
|
||||
res, path, data, err := kit.CreateImportFile(ctx, client, 1, 0)
|
||||
require.NoError(t, err)
|
||||
|
||||
// Get the piece size and commP
|
||||
fcid := res.Root
|
||||
pieceInfo, err := client.ClientDealPieceCID(ctx, fcid)
|
||||
require.NoError(t, err)
|
||||
fmt.Println("FILE CID: ", fcid)
|
||||
|
||||
// Create a storage deal with the miner
|
||||
maddr, err := miner.ActorAddress(ctx)
|
||||
require.NoError(t, err)
|
||||
|
||||
addr, err := client.WalletDefaultAddress(ctx)
|
||||
require.NoError(t, err)
|
||||
|
||||
// Manual storage deal (offline deal)
|
||||
dataRef := &storagemarket.DataRef{
|
||||
TransferType: storagemarket.TTManual,
|
||||
Root: fcid,
|
||||
PieceCid: &pieceInfo.PieceCID,
|
||||
PieceSize: pieceInfo.PieceSize.Unpadded(),
|
||||
}
|
||||
|
||||
proposalCid, err := client.ClientStartDeal(ctx, &api.StartDealParams{
|
||||
Data: dataRef,
|
||||
Wallet: addr,
|
||||
Miner: maddr,
|
||||
EpochPrice: types.NewInt(1000000),
|
||||
DealStartEpoch: startEpoch,
|
||||
MinBlocksDuration: uint64(build.MinDealDuration),
|
||||
FastRetrieval: fastRet,
|
||||
})
|
||||
require.NoError(t, err)
|
||||
|
||||
// Wait for the deal to reach StorageDealCheckForAcceptance on the client
|
||||
cd, err := client.ClientGetDealInfo(ctx, *proposalCid)
|
||||
require.NoError(t, err)
|
||||
require.Eventually(t, func() bool {
|
||||
cd, _ := client.ClientGetDealInfo(ctx, *proposalCid)
|
||||
return cd.State == storagemarket.StorageDealCheckForAcceptance
|
||||
}, 30*time.Second, 1*time.Second, "actual deal status is %s", storagemarket.DealStates[cd.State])
|
||||
|
||||
// Create a CAR file from the raw file
|
||||
carFileDir, err := ioutil.TempDir(os.TempDir(), "test-make-deal-car")
|
||||
require.NoError(t, err)
|
||||
carFilePath := filepath.Join(carFileDir, "out.car")
|
||||
err = client.ClientGenCar(ctx, api.FileRef{Path: path}, carFilePath)
|
||||
require.NoError(t, err)
|
||||
|
||||
// Import the CAR file on the miner - this is the equivalent to
|
||||
// transferring the file across the wire in a normal (non-offline) deal
|
||||
err = miner.DealsImportData(ctx, *proposalCid, carFilePath)
|
||||
require.NoError(t, err)
|
||||
|
||||
// Wait for the deal to be published
|
||||
dh.WaitDealPublished(ctx, proposalCid)
|
||||
|
||||
t.Logf("deal published, retrieving")
|
||||
|
||||
// Retrieve the deal
|
||||
dh.TestRetrieval(ctx, fcid, &pieceInfo.PieceCID, false, data)
|
||||
}
|
||||
|
||||
t.Run("NormalRetrieval", func(t *testing.T) {
|
||||
runTest(t, false)
|
||||
})
|
||||
t.Run("FastRetrieval", func(t *testing.T) {
|
||||
runTest(t, true)
|
||||
})
|
||||
|
||||
}
|
||||
|
||||
func runFullDealCycles(t *testing.T, n int, b kit.APIBuilder, blocktime time.Duration, carExport, fastRet bool, startEpoch abi.ChainEpoch) {
|
||||
|
||||
fulls, miners := b(t, kit.OneFull, kit.OneMiner)
|
||||
client, miner := fulls[0].FullNode.(*impl.FullNodeAPI), miners[0]
|
||||
|
||||
kit.ConnectAndStartMining(t, blocktime, miner, client)
|
||||
|
||||
dh := kit.NewDealHarness(t, client, miner)
|
||||
|
||||
baseseed := 6
|
||||
for i := 0; i < n; i++ {
|
||||
dh.MakeFullDeal(context.Background(), baseseed+i, carExport, fastRet, startEpoch)
|
||||
}
|
||||
}
|
||||
|
||||
func runFastRetrievalDealFlowT(t *testing.T, b kit.APIBuilder, blocktime time.Duration, startEpoch abi.ChainEpoch) {
|
||||
ctx := context.Background()
|
||||
|
||||
var (
|
||||
nb = kit.NewNodeBuilder(t)
|
||||
full = nb.FullNode()
|
||||
miner = nb.Miner(full)
|
||||
)
|
||||
|
||||
nb.Create()
|
||||
|
||||
kit.ConnectAndStartMining(t, blocktime, miner, full)
|
||||
|
||||
dh := kit.NewDealHarness(t, full, miner)
|
||||
data := make([]byte, 1600)
|
||||
rand.New(rand.NewSource(int64(8))).Read(data)
|
||||
|
||||
r := bytes.NewReader(data)
|
||||
fcid, err := full.FullNode.(*impl.FullNodeAPI).ClientImportLocal(ctx, r)
|
||||
require.NoError(t, err)
|
||||
|
||||
fmt.Println("FILE CID: ", fcid)
|
||||
|
||||
deal := dh.StartDeal(ctx, fcid, true, startEpoch)
|
||||
dh.WaitDealPublished(ctx, deal)
|
||||
|
||||
fmt.Println("deal published, retrieving")
|
||||
|
||||
// Retrieval
|
||||
info, err := full.ClientGetDealInfo(ctx, *deal)
|
||||
require.NoError(t, err)
|
||||
|
||||
dh.TestRetrieval(ctx, fcid, &info.PieceCID, false, data)
|
||||
}
|
||||
|
||||
func runSecondDealRetrievalTest(t *testing.T, b kit.APIBuilder, blocktime time.Duration) {
|
||||
ctx := context.Background()
|
||||
|
||||
fulls, miners := b(t, kit.OneFull, kit.OneMiner)
|
||||
client, miner := fulls[0].FullNode.(*impl.FullNodeAPI), miners[0]
|
||||
|
||||
kit.ConnectAndStartMining(t, blocktime, miner, client)
|
||||
|
||||
dh := kit.NewDealHarness(t, client, miner)
|
||||
|
||||
{
|
||||
data1 := make([]byte, 800)
|
||||
rand.New(rand.NewSource(int64(3))).Read(data1)
|
||||
r := bytes.NewReader(data1)
|
||||
|
||||
fcid1, err := client.ClientImportLocal(ctx, r)
|
||||
if err != nil {
|
||||
t.Fatal(err)
|
||||
}
|
||||
|
||||
data2 := make([]byte, 800)
|
||||
rand.New(rand.NewSource(int64(9))).Read(data2)
|
||||
r2 := bytes.NewReader(data2)
|
||||
|
||||
fcid2, err := client.ClientImportLocal(ctx, r2)
|
||||
if err != nil {
|
||||
t.Fatal(err)
|
||||
}
|
||||
|
||||
deal1 := dh.StartDeal(ctx, fcid1, true, 0)
|
||||
|
||||
// TODO: this sleep is only necessary because deals don't immediately get logged in the dealstore, we should fix this
|
||||
time.Sleep(time.Second)
|
||||
dh.WaitDealSealed(ctx, deal1, true, false, nil)
|
||||
|
||||
deal2 := dh.StartDeal(ctx, fcid2, true, 0)
|
||||
|
||||
time.Sleep(time.Second)
|
||||
dh.WaitDealSealed(ctx, deal2, false, false, nil)
|
||||
|
||||
// Retrieval
|
||||
info, err := client.ClientGetDealInfo(ctx, *deal2)
|
||||
require.NoError(t, err)
|
||||
|
||||
rf, _ := miner.SectorsRefs(ctx)
|
||||
fmt.Printf("refs: %+v\n", rf)
|
||||
|
||||
dh.TestRetrieval(ctx, fcid2, &info.PieceCID, false, data2)
|
||||
}
|
||||
}
|
||||
|
||||
func runZeroPricePerByteRetrievalDealFlow(t *testing.T, b kit.APIBuilder, blocktime time.Duration, startEpoch abi.ChainEpoch) {
|
||||
ctx := context.Background()
|
||||
|
||||
fulls, miners := b(t, kit.OneFull, kit.OneMiner)
|
||||
client, miner := fulls[0].FullNode.(*impl.FullNodeAPI), miners[0]
|
||||
|
||||
kit.ConnectAndStartMining(t, blocktime, miner, client)
|
||||
|
||||
dh := kit.NewDealHarness(t, client, miner)
|
||||
|
||||
// Set price-per-byte to zero
|
||||
ask, err := miner.MarketGetRetrievalAsk(ctx)
|
||||
require.NoError(t, err)
|
||||
|
||||
ask.PricePerByte = abi.NewTokenAmount(0)
|
||||
err = miner.MarketSetRetrievalAsk(ctx, ask)
|
||||
require.NoError(t, err)
|
||||
|
||||
dh.MakeFullDeal(ctx, 6, false, false, startEpoch)
|
||||
}
|
@ -114,12 +114,13 @@ func NewEnsemble(t *testing.T, opts ...BuilderOpt) *Ensemble {
|
||||
}
|
||||
|
||||
type NodeOpts struct {
|
||||
balance abi.TokenAmount
|
||||
lite bool
|
||||
sectors int
|
||||
mockProofs bool
|
||||
rpc bool
|
||||
ownerKey *wallet.Key
|
||||
balance abi.TokenAmount
|
||||
lite bool
|
||||
sectors int
|
||||
mockProofs bool
|
||||
rpc bool
|
||||
ownerKey *wallet.Key
|
||||
extraNodeOpts []node.Option
|
||||
}
|
||||
|
||||
var DefaultNodeOpts = NodeOpts{
|
||||
@ -182,6 +183,13 @@ func OwnerAddr(wk *wallet.Key) NodeOpt {
|
||||
}
|
||||
}
|
||||
|
||||
func ExtraNodeOpts(extra ...node.Option) NodeOpt {
|
||||
return func(opts *NodeOpts) error {
|
||||
opts.extraNodeOpts = extra
|
||||
return nil
|
||||
}
|
||||
}
|
||||
|
||||
// FullNode enrolls a new full node.
|
||||
func (n *Ensemble) FullNode(full *TestFullNode, opts ...NodeOpt) *Ensemble {
|
||||
options := DefaultNodeOpts
|
||||
@ -300,6 +308,10 @@ func (n *Ensemble) Start() *Ensemble {
|
||||
n.mn = mocknet.New(ctx)
|
||||
}
|
||||
|
||||
// ---------------------
|
||||
// FULL NODES
|
||||
// ---------------------
|
||||
|
||||
// Create all inactive full nodes.
|
||||
for i, full := range n.inactive.fullnodes {
|
||||
opts := []node.Option{
|
||||
@ -313,6 +325,9 @@ func (n *Ensemble) Start() *Ensemble {
|
||||
node.Override(new(dtypes.Bootstrapper), dtypes.Bootstrapper(true)),
|
||||
}
|
||||
|
||||
// append any node builder options.
|
||||
opts = append(opts, full.options.extraNodeOpts...)
|
||||
|
||||
// Either generate the genesis or inject it.
|
||||
if i == 0 && !n.bootstrapped {
|
||||
opts = append(opts, node.Override(new(modules.Genesis), testing2.MakeGenesisMem(&n.genesisBlock, *gtempl)))
|
||||
@ -322,7 +337,10 @@ func (n *Ensemble) Start() *Ensemble {
|
||||
|
||||
// Are we mocking proofs?
|
||||
if full.options.mockProofs {
|
||||
opts = append(opts, node.Override(new(ffiwrapper.Verifier), mock.MockVerifier))
|
||||
opts = append(opts,
|
||||
node.Override(new(ffiwrapper.Verifier), mock.MockVerifier),
|
||||
node.Override(new(ffiwrapper.Prover), mock.MockProver),
|
||||
)
|
||||
}
|
||||
|
||||
// Construct the full node.
|
||||
@ -356,6 +374,10 @@ func (n *Ensemble) Start() *Ensemble {
|
||||
err := n.mn.LinkAll()
|
||||
require.NoError(n.t, err)
|
||||
|
||||
// ---------------------
|
||||
// MINERS
|
||||
// ---------------------
|
||||
|
||||
// Create all inactive miners.
|
||||
for i, m := range n.inactive.miners {
|
||||
if n.bootstrapped {
|
||||
@ -469,23 +491,35 @@ func (n *Ensemble) Start() *Ensemble {
|
||||
node.Override(new(*lotusminer.Miner), lotusminer.NewTestMiner(mineBlock, m.ActorAddr)),
|
||||
}
|
||||
|
||||
// append any node builder options.
|
||||
opts = append(opts, m.options.extraNodeOpts...)
|
||||
|
||||
idAddr, err := address.IDFromAddress(m.ActorAddr)
|
||||
require.NoError(n.t, err)
|
||||
|
||||
if !n.bootstrapped && m.options.mockProofs {
|
||||
s := n.genesis.miners[i].Sectors
|
||||
sectors := make([]abi.SectorID, len(s))
|
||||
for i, sector := range s {
|
||||
sectors[i] = abi.SectorID{
|
||||
// preload preseals if the network still hasn't bootstrapped.
|
||||
var presealSectors []abi.SectorID
|
||||
if !n.bootstrapped {
|
||||
sectors := n.genesis.miners[i].Sectors
|
||||
for _, sector := range sectors {
|
||||
presealSectors = append(presealSectors, abi.SectorID{
|
||||
Miner: abi.ActorID(idAddr),
|
||||
Number: sector.SectorID,
|
||||
}
|
||||
})
|
||||
}
|
||||
}
|
||||
|
||||
if m.options.mockProofs {
|
||||
opts = append(opts,
|
||||
node.Override(new(sectorstorage.SectorManager), func() (sectorstorage.SectorManager, error) {
|
||||
return mock.NewMockSectorMgr(sectors), nil
|
||||
node.Override(new(*mock.SectorMgr), func() (*mock.SectorMgr, error) {
|
||||
return mock.NewMockSectorMgr(presealSectors), nil
|
||||
}),
|
||||
node.Override(new(sectorstorage.SectorManager), node.From(new(*mock.SectorMgr))),
|
||||
node.Override(new(sectorstorage.Unsealer), node.From(new(*mock.SectorMgr))),
|
||||
node.Override(new(sectorstorage.PieceProvider), node.From(new(*mock.SectorMgr))),
|
||||
|
||||
node.Override(new(ffiwrapper.Verifier), mock.MockVerifier),
|
||||
node.Override(new(ffiwrapper.Prover), mock.MockProver),
|
||||
node.Unset(new(*sectorstorage.Manager)),
|
||||
)
|
||||
}
|
||||
@ -532,9 +566,7 @@ func (n *Ensemble) Start() *Ensemble {
|
||||
require.NoError(n.t, err)
|
||||
|
||||
if !n.bootstrapped && len(n.active.miners) > 0 {
|
||||
// We have *just* bootstrapped, so
|
||||
// mine 2 blocks to setup some CE stuff
|
||||
// in some actors
|
||||
// We have *just* bootstrapped, so mine 2 blocks to setup some CE stuff in some actors
|
||||
var wait sync.Mutex
|
||||
wait.Lock()
|
||||
|
||||
@ -557,22 +589,32 @@ func (n *Ensemble) Start() *Ensemble {
|
||||
return n
|
||||
}
|
||||
|
||||
// InterconnectAll connects all full nodes one to another. We do not need to
|
||||
// take action with miners, because miners only stay connected to their full
|
||||
// nodes over JSON-RPC.
|
||||
// InterconnectAll connects all miners and full nodes to one another.
|
||||
func (n *Ensemble) InterconnectAll() *Ensemble {
|
||||
// connect full nodes to miners.
|
||||
for _, from := range n.active.fullnodes {
|
||||
for _, to := range n.active.miners {
|
||||
// []*TestMiner to []api.CommonAPI type coercion not possible
|
||||
// so cannot use variadic form.
|
||||
n.Connect(from, to)
|
||||
}
|
||||
}
|
||||
|
||||
// connect full nodes between each other, skipping ourselves.
|
||||
last := len(n.active.fullnodes) - 1
|
||||
for i, from := range n.active.fullnodes {
|
||||
if i == last {
|
||||
continue
|
||||
}
|
||||
n.Connect(from, n.active.fullnodes[i+1:]...)
|
||||
for _, to := range n.active.fullnodes[i+1:] {
|
||||
n.Connect(from, to)
|
||||
}
|
||||
}
|
||||
return n
|
||||
}
|
||||
|
||||
// Connect connects one full node to the provided full nodes.
|
||||
func (n *Ensemble) Connect(from *TestFullNode, to ...*TestFullNode) *Ensemble {
|
||||
func (n *Ensemble) Connect(from api.Common, to ...api.Common) *Ensemble {
|
||||
addr, err := from.NetAddrsListen(context.Background())
|
||||
require.NoError(n.t, err)
|
||||
|
||||
@ -584,7 +626,8 @@ func (n *Ensemble) Connect(from *TestFullNode, to ...*TestFullNode) *Ensemble {
|
||||
}
|
||||
|
||||
// BeginMining kicks off mining for the specified miners. If nil or 0-length,
|
||||
// it will kick off mining for all enrolled and active miners.
|
||||
// it will kick off mining for all enrolled and active miners. It also adds a
|
||||
// cleanup function to stop all mining operations on test teardown.
|
||||
func (n *Ensemble) BeginMining(blocktime time.Duration, miners ...*TestMiner) []*BlockMiner {
|
||||
ctx := context.Background()
|
||||
|
||||
@ -601,6 +644,8 @@ func (n *Ensemble) BeginMining(blocktime time.Duration, miners ...*TestMiner) []
|
||||
for _, m := range miners {
|
||||
bm := NewBlockMiner(n.t, m)
|
||||
bm.MineBlocks(ctx, blocktime)
|
||||
n.t.Cleanup(bm.Stop)
|
||||
|
||||
bms = append(bms, bm)
|
||||
}
|
||||
|
||||
|
Loading…
Reference in New Issue
Block a user