deals tests: begin migration.

This commit is contained in:
Raúl Kripalani 2021-06-10 13:25:36 +01:00
parent cf0150e057
commit 329970934a
4 changed files with 578 additions and 551 deletions

1
go.mod
View File

@ -16,6 +16,7 @@ require (
github.com/chzyer/readline v0.0.0-20180603132655-2972be24d48e github.com/chzyer/readline v0.0.0-20180603132655-2972be24d48e
github.com/cockroachdb/pebble v0.0.0-20201001221639-879f3bfeef07 github.com/cockroachdb/pebble v0.0.0-20201001221639-879f3bfeef07
github.com/coreos/go-systemd/v22 v22.1.0 github.com/coreos/go-systemd/v22 v22.1.0
github.com/davecgh/go-spew v1.1.1 // indirect
github.com/detailyang/go-fallocate v0.0.0-20180908115635-432fa640bd2e github.com/detailyang/go-fallocate v0.0.0-20180908115635-432fa640bd2e
github.com/dgraph-io/badger/v2 v2.2007.2 github.com/dgraph-io/badger/v2 v2.2007.2
github.com/docker/go-units v0.4.0 github.com/docker/go-units v0.4.0

508
itests/deals_test.go Normal file
View File

@ -0,0 +1,508 @@
package itests
import (
"bytes"
"context"
"fmt"
"testing"
"time"
"github.com/davecgh/go-spew/spew"
"github.com/filecoin-project/go-fil-markets/storagemarket"
"github.com/filecoin-project/go-state-types/abi"
"github.com/filecoin-project/lotus/api"
"github.com/filecoin-project/lotus/chain/actors/builtin/market"
"github.com/filecoin-project/lotus/chain/types"
"github.com/filecoin-project/lotus/itests/kit"
"github.com/filecoin-project/lotus/markets/storageadapter"
"github.com/filecoin-project/lotus/node"
market2 "github.com/filecoin-project/specs-actors/v2/actors/builtin/market"
"github.com/stretchr/testify/require"
)
//
// func TestDealCycle(t *testing.T) {
// kit.QuietMiningLogs()
//
// blockTime := 10 * time.Millisecond
//
// // For these tests where the block time is artificially short, just use
// // a deal start epoch that is guaranteed to be far enough in the future
// // so that the deal starts sealing in time
// dealStartEpoch := abi.ChainEpoch(2 << 12)
//
// t.Run("TestFullDealCycle_Single", func(t *testing.T) {
// runFullDealCycles(t, 1, kit.MockMinerBuilder, blockTime, false, false, dealStartEpoch)
// })
// t.Run("TestFullDealCycle_Two", func(t *testing.T) {
// runFullDealCycles(t, 2, kit.MockMinerBuilder, blockTime, false, false, dealStartEpoch)
// })
// t.Run("WithExportedCAR", func(t *testing.T) {
// runFullDealCycles(t, 1, kit.MockMinerBuilder, blockTime, true, false, dealStartEpoch)
// })
// t.Run("TestFastRetrievalDealCycle", func(t *testing.T) {
// runFastRetrievalDealFlowT(t, kit.MockMinerBuilder, blockTime, dealStartEpoch)
// })
// t.Run("TestZeroPricePerByteRetrievalDealFlow", func(t *testing.T) {
// runZeroPricePerByteRetrievalDealFlow(t, kit.MockMinerBuilder, blockTime, dealStartEpoch)
// })
// }
//
// func TestAPIDealFlowReal(t *testing.T) {
// if testing.Short() {
// t.Skip("skipping test in short mode")
// }
//
// kit.QuietMiningLogs()
//
// // TODO: just set this globally?
// oldDelay := policy.GetPreCommitChallengeDelay()
// policy.SetPreCommitChallengeDelay(5)
// t.Cleanup(func() {
// policy.SetPreCommitChallengeDelay(oldDelay)
// })
//
// t.Run("basic", func(t *testing.T) {
// runFullDealCycles(t, 1, kit.FullNodeBuilder, time.Second, false, false, 0)
// })
//
// t.Run("fast-retrieval", func(t *testing.T) {
// runFullDealCycles(t, 1, kit.FullNodeBuilder, time.Second, false, true, 0)
// })
//
// t.Run("retrieval-second", func(t *testing.T) {
// runSecondDealRetrievalTest(t, kit.FullNodeBuilder, time.Second)
// })
// }
func TestPublishDealsBatching(t *testing.T) {
var (
ctx = context.Background()
publishPeriod = 10 * time.Second
maxDealsPerMsg = uint64(2) // Set max deals per publish deals message to 2
startEpoch = abi.ChainEpoch(2 << 12)
)
kit.QuietMiningLogs()
opts := node.Override(new(*storageadapter.DealPublisher),
storageadapter.NewDealPublisher(nil, storageadapter.PublishMsgConfig{
Period: publishPeriod,
MaxDealsPerMsg: maxDealsPerMsg,
}),
)
client, miner, ens := kit.EnsembleMinimum(t, kit.MockProofs(), kit.ExtraNodeOpts(opts))
ens.InterconnectAll().BeginMining(10 * time.Millisecond)
dh := kit.NewDealHarness(t, client, miner)
fmt.Println("***********************")
spew.Dump(client.NetPeers(context.Background()))
// Starts a deal and waits until it's published
runDealTillPublish := func(rseed int) {
res, _, _, err := kit.CreateImportFile(ctx, client, rseed, 0)
require.NoError(t, err)
upds, err := client.ClientGetDealUpdates(ctx)
require.NoError(t, err)
dh.StartDeal(ctx, res.Root, false, startEpoch)
// TODO: this sleep is only necessary because deals don't immediately get logged in the dealstore, we should fix this
time.Sleep(time.Second)
done := make(chan struct{})
go func() {
for upd := range upds {
if upd.DataRef.Root == res.Root && upd.State == storagemarket.StorageDealAwaitingPreCommit {
done <- struct{}{}
}
}
}()
<-done
}
// Run three deals in parallel
done := make(chan struct{}, maxDealsPerMsg+1)
for rseed := 1; rseed <= 3; rseed++ {
rseed := rseed
go func() {
runDealTillPublish(rseed)
done <- struct{}{}
}()
}
// Wait for two of the deals to be published
for i := 0; i < int(maxDealsPerMsg); i++ {
<-done
}
// Expect a single PublishStorageDeals message that includes the first two deals
msgCids, err := client.StateListMessages(ctx, &api.MessageMatch{To: market.Address}, types.EmptyTSK, 1)
require.NoError(t, err)
count := 0
for _, msgCid := range msgCids {
msg, err := client.ChainGetMessage(ctx, msgCid)
require.NoError(t, err)
if msg.Method == market.Methods.PublishStorageDeals {
count++
var pubDealsParams market2.PublishStorageDealsParams
err = pubDealsParams.UnmarshalCBOR(bytes.NewReader(msg.Params))
require.NoError(t, err)
require.Len(t, pubDealsParams.Deals, int(maxDealsPerMsg))
}
}
require.Equal(t, 1, count)
// The third deal should be published once the publish period expires.
// Allow a little padding as it takes a moment for the state change to
// be noticed by the client.
padding := 10 * time.Second
select {
case <-time.After(publishPeriod + padding):
require.Fail(t, "Expected 3rd deal to be published once publish period elapsed")
case <-done: // Success
}
}
//
// func TestDealMining(t *testing.T) {
// // test making a deal with a fresh miner, and see if it starts to mine.
// if testing.Short() {
// t.Skip("skipping test in short mode")
// }
//
// kit.QuietMiningLogs()
//
// b := kit.MockMinerBuilder
// blocktime := 50 * time.Millisecond
//
// ctx := context.Background()
// fulls, miners := b(t,
// kit.OneFull,
// []kit.StorageMiner{
// {Full: 0, Preseal: kit.PresealGenesis},
// {Full: 0, Preseal: 0}, // TODO: Add support for miners on non-first full node
// })
// client := fulls[0].FullNode.(*impl.FullNodeAPI)
// genesisMiner := miners[0]
// provider := miners[1]
//
// addrinfo, err := client.NetAddrsListen(ctx)
// if err != nil {
// t.Fatal(err)
// }
//
// if err := provider.NetConnect(ctx, addrinfo); err != nil {
// t.Fatal(err)
// }
//
// if err := genesisMiner.NetConnect(ctx, addrinfo); err != nil {
// t.Fatal(err)
// }
//
// time.Sleep(time.Second)
//
// data := make([]byte, 600)
// rand.New(rand.NewSource(5)).Read(data)
//
// r := bytes.NewReader(data)
// fcid, err := client.ClientImportLocal(ctx, r)
// if err != nil {
// t.Fatal(err)
// }
//
// fmt.Println("FILE CID: ", fcid)
//
// var mine int32 = 1
// done := make(chan struct{})
// minedTwo := make(chan struct{})
//
// m2addr, err := miners[1].ActorAddress(context.TODO())
// if err != nil {
// t.Fatal(err)
// }
//
// go func() {
// defer close(done)
//
// complChan := minedTwo
// for atomic.LoadInt32(&mine) != 0 {
// wait := make(chan int)
// mdone := func(mined bool, _ abi.ChainEpoch, err error) {
// n := 0
// if mined {
// n = 1
// }
// wait <- n
// }
//
// if err := miners[0].MineOne(ctx, miner.MineReq{Done: mdone}); err != nil {
// t.Error(err)
// }
//
// if err := miners[1].MineOne(ctx, miner.MineReq{Done: mdone}); err != nil {
// t.Error(err)
// }
//
// expect := <-wait
// expect += <-wait
//
// time.Sleep(blocktime)
// if expect == 0 {
// // null block
// continue
// }
//
// var nodeOneMined bool
// for _, node := range miners {
// mb, err := node.MiningBase(ctx)
// if err != nil {
// t.Error(err)
// return
// }
//
// for _, b := range mb.Blocks() {
// if b.Miner == m2addr {
// nodeOneMined = true
// break
// }
// }
//
// }
//
// if nodeOneMined && complChan != nil {
// close(complChan)
// complChan = nil
// }
//
// }
// }()
//
// dh := kit.NewDealHarness(t, client, provider)
//
// deal := dh.StartDeal(ctx, fcid, false, 0)
//
// // TODO: this sleep is only necessary because deals don't immediately get logged in the dealstore, we should fix this
// time.Sleep(time.Second)
//
// dh.WaitDealSealed(ctx, deal, false, false, nil)
//
// <-minedTwo
//
// atomic.StoreInt32(&mine, 0)
// fmt.Println("shutting down mining")
// <-done
// }
//
// func TestOfflineDealFlow(t *testing.T) {
// blocktime := 10 * time.Millisecond
//
// // For these tests where the block time is artificially short, just use
// // a deal start epoch that is guaranteed to be far enough in the future
// // so that the deal starts sealing in time
// startEpoch := abi.ChainEpoch(2 << 12)
//
// runTest := func(t *testing.T, fastRet bool) {
// ctx := context.Background()
// fulls, miners := kit.MockMinerBuilder(t, kit.OneFull, kit.OneMiner)
// client, miner := fulls[0].FullNode.(*impl.FullNodeAPI), miners[0]
//
// kit.ConnectAndStartMining(t, blocktime, miner, client)
//
// dh := kit.NewDealHarness(t, client, miner)
//
// // Create a random file and import on the client.
// res, path, data, err := kit.CreateImportFile(ctx, client, 1, 0)
// require.NoError(t, err)
//
// // Get the piece size and commP
// fcid := res.Root
// pieceInfo, err := client.ClientDealPieceCID(ctx, fcid)
// require.NoError(t, err)
// fmt.Println("FILE CID: ", fcid)
//
// // Create a storage deal with the miner
// maddr, err := miner.ActorAddress(ctx)
// require.NoError(t, err)
//
// addr, err := client.WalletDefaultAddress(ctx)
// require.NoError(t, err)
//
// // Manual storage deal (offline deal)
// dataRef := &storagemarket.DataRef{
// TransferType: storagemarket.TTManual,
// Root: fcid,
// PieceCid: &pieceInfo.PieceCID,
// PieceSize: pieceInfo.PieceSize.Unpadded(),
// }
//
// proposalCid, err := client.ClientStartDeal(ctx, &api.StartDealParams{
// Data: dataRef,
// Wallet: addr,
// Miner: maddr,
// EpochPrice: types.NewInt(1000000),
// DealStartEpoch: startEpoch,
// MinBlocksDuration: uint64(build.MinDealDuration),
// FastRetrieval: fastRet,
// })
// require.NoError(t, err)
//
// // Wait for the deal to reach StorageDealCheckForAcceptance on the client
// cd, err := client.ClientGetDealInfo(ctx, *proposalCid)
// require.NoError(t, err)
// require.Eventually(t, func() bool {
// cd, _ := client.ClientGetDealInfo(ctx, *proposalCid)
// return cd.State == storagemarket.StorageDealCheckForAcceptance
// }, 30*time.Second, 1*time.Second, "actual deal status is %s", storagemarket.DealStates[cd.State])
//
// // Create a CAR file from the raw file
// carFileDir, err := ioutil.TempDir(os.TempDir(), "test-make-deal-car")
// require.NoError(t, err)
// carFilePath := filepath.Join(carFileDir, "out.car")
// err = client.ClientGenCar(ctx, api.FileRef{Path: path}, carFilePath)
// require.NoError(t, err)
//
// // Import the CAR file on the miner - this is the equivalent to
// // transferring the file across the wire in a normal (non-offline) deal
// err = miner.DealsImportData(ctx, *proposalCid, carFilePath)
// require.NoError(t, err)
//
// // Wait for the deal to be published
// dh.WaitDealPublished(ctx, proposalCid)
//
// t.Logf("deal published, retrieving")
//
// // Retrieve the deal
// dh.PerformRetrieval(ctx, fcid, &pieceInfo.PieceCID, false, data)
// }
//
// t.Run("NormalRetrieval", func(t *testing.T) {
// runTest(t, false)
// })
// t.Run("FastRetrieval", func(t *testing.T) {
// runTest(t, true)
// })
//
// }
//
// func runFullDealCycles(t *testing.T, n int, b kit.APIBuilder, blocktime time.Duration, carExport, fastRet bool, startEpoch abi.ChainEpoch) {
// full, _, ens := kit.EnsembleMinimum(t)
// ens.BeginMining()
// dh := kit.NewDealHarness(t, client, miner)
//
// baseseed := 6
// for i := 0; i < n; i++ {
// dh.MakeFullDeal(context.Background(), baseseed+i, carExport, fastRet, startEpoch)
// }
// }
//
// func runFastRetrievalDealFlowT(t *testing.T, b kit.APIBuilder, blocktime time.Duration, startEpoch abi.ChainEpoch) {
// ctx := context.Background()
//
// var (
// nb = kit.NewNodeBuilder(t)
// full = nb.FullNode()
// miner = nb.Miner(full)
// )
//
// nb.Create()
//
// kit.ConnectAndStartMining(t, blocktime, miner, full)
//
// dh := kit.NewDealHarness(t, full, miner)
// data := make([]byte, 1600)
// rand.New(rand.NewSource(int64(8))).Read(data)
//
// r := bytes.NewReader(data)
// fcid, err := full.FullNode.(*impl.FullNodeAPI).ClientImportLocal(ctx, r)
// require.NoError(t, err)
//
// fmt.Println("FILE CID: ", fcid)
//
// deal := dh.StartDeal(ctx, fcid, true, startEpoch)
// dh.WaitDealPublished(ctx, deal)
//
// fmt.Println("deal published, retrieving")
//
// // Retrieval
// info, err := full.ClientGetDealInfo(ctx, *deal)
// require.NoError(t, err)
//
// dh.PerformRetrieval(ctx, fcid, &info.PieceCID, false, data)
// }
//
// func runSecondDealRetrievalTest(t *testing.T, b kit.APIBuilder, blocktime time.Duration) {
// ctx := context.Background()
//
// fulls, miners := b(t, kit.OneFull, kit.OneMiner)
// client, miner := fulls[0].FullNode.(*impl.FullNodeAPI), miners[0]
//
// kit.ConnectAndStartMining(t, blocktime, miner, client)
//
// dh := kit.NewDealHarness(t, client, miner)
//
// {
// data1 := make([]byte, 800)
// rand.New(rand.NewSource(int64(3))).Read(data1)
// r := bytes.NewReader(data1)
//
// fcid1, err := client.ClientImportLocal(ctx, r)
// if err != nil {
// t.Fatal(err)
// }
//
// data2 := make([]byte, 800)
// rand.New(rand.NewSource(int64(9))).Read(data2)
// r2 := bytes.NewReader(data2)
//
// fcid2, err := client.ClientImportLocal(ctx, r2)
// if err != nil {
// t.Fatal(err)
// }
//
// deal1 := dh.StartDeal(ctx, fcid1, true, 0)
//
// // TODO: this sleep is only necessary because deals don't immediately get logged in the dealstore, we should fix this
// time.Sleep(time.Second)
// dh.WaitDealSealed(ctx, deal1, true, false, nil)
//
// deal2 := dh.StartDeal(ctx, fcid2, true, 0)
//
// time.Sleep(time.Second)
// dh.WaitDealSealed(ctx, deal2, false, false, nil)
//
// // Retrieval
// info, err := client.ClientGetDealInfo(ctx, *deal2)
// require.NoError(t, err)
//
// rf, _ := miner.SectorsRefs(ctx)
// fmt.Printf("refs: %+v\n", rf)
//
// dh.PerformRetrieval(ctx, fcid2, &info.PieceCID, false, data2)
// }
// }
//
// func runZeroPricePerByteRetrievalDealFlow(t *testing.T, b kit.APIBuilder, blocktime time.Duration, startEpoch abi.ChainEpoch) {
// ctx := context.Background()
//
// fulls, miners := b(t, kit.OneFull, kit.OneMiner)
// client, miner := fulls[0].FullNode.(*impl.FullNodeAPI), miners[0]
//
// kit.ConnectAndStartMining(t, blocktime, miner, client)
//
// dh := kit.NewDealHarness(t, client, miner)
//
// // Set price-per-byte to zero
// ask, err := miner.MarketGetRetrievalAsk(ctx)
// require.NoError(t, err)
//
// ask.PricePerByte = abi.NewTokenAmount(0)
// err = miner.MarketSetRetrievalAsk(ctx, ask)
// require.NoError(t, err)
//
// dh.MakeFullDeal(ctx, 6, false, false, startEpoch)
// }

View File

@ -1,527 +0,0 @@
package itests
import (
"bytes"
"context"
"fmt"
"io/ioutil"
"math/rand"
"os"
"path/filepath"
"sync/atomic"
"testing"
"time"
"github.com/filecoin-project/go-fil-markets/storagemarket"
"github.com/filecoin-project/go-state-types/abi"
"github.com/filecoin-project/lotus/api"
"github.com/filecoin-project/lotus/build"
"github.com/filecoin-project/lotus/chain/actors/builtin/market"
"github.com/filecoin-project/lotus/chain/actors/policy"
"github.com/filecoin-project/lotus/chain/types"
"github.com/filecoin-project/lotus/itests/kit"
"github.com/filecoin-project/lotus/markets/storageadapter"
"github.com/filecoin-project/lotus/miner"
"github.com/filecoin-project/lotus/node"
"github.com/filecoin-project/lotus/node/impl"
"github.com/filecoin-project/lotus/node/impl/client"
market2 "github.com/filecoin-project/specs-actors/v2/actors/builtin/market"
"github.com/stretchr/testify/require"
)
func TestDealCycle(t *testing.T) {
kit.QuietMiningLogs()
blockTime := 10 * time.Millisecond
// For these tests where the block time is artificially short, just use
// a deal start epoch that is guaranteed to be far enough in the future
// so that the deal starts sealing in time
dealStartEpoch := abi.ChainEpoch(2 << 12)
t.Run("TestFullDealCycle_Single", func(t *testing.T) {
runFullDealCycles(t, 1, kit.MockMinerBuilder, blockTime, false, false, dealStartEpoch)
})
t.Run("TestFullDealCycle_Two", func(t *testing.T) {
runFullDealCycles(t, 2, kit.MockMinerBuilder, blockTime, false, false, dealStartEpoch)
})
t.Run("WithExportedCAR", func(t *testing.T) {
runFullDealCycles(t, 1, kit.MockMinerBuilder, blockTime, true, false, dealStartEpoch)
})
t.Run("TestFastRetrievalDealCycle", func(t *testing.T) {
runFastRetrievalDealFlowT(t, kit.MockMinerBuilder, blockTime, dealStartEpoch)
})
t.Run("TestZeroPricePerByteRetrievalDealFlow", func(t *testing.T) {
runZeroPricePerByteRetrievalDealFlow(t, kit.MockMinerBuilder, blockTime, dealStartEpoch)
})
}
func TestAPIDealFlowReal(t *testing.T) {
if testing.Short() {
t.Skip("skipping test in short mode")
}
kit.QuietMiningLogs()
// TODO: just set this globally?
oldDelay := policy.GetPreCommitChallengeDelay()
policy.SetPreCommitChallengeDelay(5)
t.Cleanup(func() {
policy.SetPreCommitChallengeDelay(oldDelay)
})
t.Run("basic", func(t *testing.T) {
runFullDealCycles(t, 1, kit.FullNodeBuilder, time.Second, false, false, 0)
})
t.Run("fast-retrieval", func(t *testing.T) {
runFullDealCycles(t, 1, kit.FullNodeBuilder, time.Second, false, true, 0)
})
t.Run("retrieval-second", func(t *testing.T) {
runSecondDealRetrievalTest(t, kit.FullNodeBuilder, time.Second)
})
}
func TestPublishDealsBatching(t *testing.T) {
ctx := context.Background()
kit.QuietMiningLogs()
b := kit.MockMinerBuilder
blocktime := 10 * time.Millisecond
startEpoch := abi.ChainEpoch(2 << 12)
publishPeriod := 10 * time.Second
maxDealsPerMsg := uint64(2)
// Set max deals per publish deals message to 2
minerDef := []kit.StorageMiner{{
Full: 0,
Opts: node.Override(
new(*storageadapter.DealPublisher),
storageadapter.NewDealPublisher(nil, storageadapter.PublishMsgConfig{
Period: publishPeriod,
MaxDealsPerMsg: maxDealsPerMsg,
})),
Preseal: kit.PresealGenesis,
}}
// Create a connect client and miner node
n, sn := b(t, kit.OneFull, minerDef)
client := n[0].FullNode.(*impl.FullNodeAPI)
miner := sn[0]
kit.ConnectAndStartMining(t, blocktime, miner, client)
dh := kit.NewDealHarness(t, client, miner)
// Starts a deal and waits until it's published
runDealTillPublish := func(rseed int) {
res, _, _, err := kit.CreateImportFile(ctx, client, rseed, 0)
require.NoError(t, err)
upds, err := client.ClientGetDealUpdates(ctx)
require.NoError(t, err)
dh.StartDeal(ctx, res.Root, false, startEpoch)
// TODO: this sleep is only necessary because deals don't immediately get logged in the dealstore, we should fix this
time.Sleep(time.Second)
done := make(chan struct{})
go func() {
for upd := range upds {
if upd.DataRef.Root == res.Root && upd.State == storagemarket.StorageDealAwaitingPreCommit {
done <- struct{}{}
}
}
}()
<-done
}
// Run three deals in parallel
done := make(chan struct{}, maxDealsPerMsg+1)
for rseed := 1; rseed <= 3; rseed++ {
rseed := rseed
go func() {
runDealTillPublish(rseed)
done <- struct{}{}
}()
}
// Wait for two of the deals to be published
for i := 0; i < int(maxDealsPerMsg); i++ {
<-done
}
// Expect a single PublishStorageDeals message that includes the first two deals
msgCids, err := client.StateListMessages(ctx, &api.MessageMatch{To: market.Address}, types.EmptyTSK, 1)
require.NoError(t, err)
count := 0
for _, msgCid := range msgCids {
msg, err := client.ChainGetMessage(ctx, msgCid)
require.NoError(t, err)
if msg.Method == market.Methods.PublishStorageDeals {
count++
var pubDealsParams market2.PublishStorageDealsParams
err = pubDealsParams.UnmarshalCBOR(bytes.NewReader(msg.Params))
require.NoError(t, err)
require.Len(t, pubDealsParams.Deals, int(maxDealsPerMsg))
}
}
require.Equal(t, 1, count)
// The third deal should be published once the publish period expires.
// Allow a little padding as it takes a moment for the state change to
// be noticed by the client.
padding := 10 * time.Second
select {
case <-time.After(publishPeriod + padding):
require.Fail(t, "Expected 3rd deal to be published once publish period elapsed")
case <-done: // Success
}
}
func TestDealMining(t *testing.T) {
// test making a deal with a fresh miner, and see if it starts to mine.
if testing.Short() {
t.Skip("skipping test in short mode")
}
kit.QuietMiningLogs()
b := kit.MockMinerBuilder
blocktime := 50 * time.Millisecond
ctx := context.Background()
fulls, miners := b(t,
kit.OneFull,
[]kit.StorageMiner{
{Full: 0, Preseal: kit.PresealGenesis},
{Full: 0, Preseal: 0}, // TODO: Add support for miners on non-first full node
})
client := fulls[0].FullNode.(*impl.FullNodeAPI)
genesisMiner := miners[0]
provider := miners[1]
addrinfo, err := client.NetAddrsListen(ctx)
if err != nil {
t.Fatal(err)
}
if err := provider.NetConnect(ctx, addrinfo); err != nil {
t.Fatal(err)
}
if err := genesisMiner.NetConnect(ctx, addrinfo); err != nil {
t.Fatal(err)
}
time.Sleep(time.Second)
data := make([]byte, 600)
rand.New(rand.NewSource(5)).Read(data)
r := bytes.NewReader(data)
fcid, err := client.ClientImportLocal(ctx, r)
if err != nil {
t.Fatal(err)
}
fmt.Println("FILE CID: ", fcid)
var mine int32 = 1
done := make(chan struct{})
minedTwo := make(chan struct{})
m2addr, err := miners[1].ActorAddress(context.TODO())
if err != nil {
t.Fatal(err)
}
go func() {
defer close(done)
complChan := minedTwo
for atomic.LoadInt32(&mine) != 0 {
wait := make(chan int)
mdone := func(mined bool, _ abi.ChainEpoch, err error) {
n := 0
if mined {
n = 1
}
wait <- n
}
if err := miners[0].MineOne(ctx, miner.MineReq{Done: mdone}); err != nil {
t.Error(err)
}
if err := miners[1].MineOne(ctx, miner.MineReq{Done: mdone}); err != nil {
t.Error(err)
}
expect := <-wait
expect += <-wait
time.Sleep(blocktime)
if expect == 0 {
// null block
continue
}
var nodeOneMined bool
for _, node := range miners {
mb, err := node.MiningBase(ctx)
if err != nil {
t.Error(err)
return
}
for _, b := range mb.Blocks() {
if b.Miner == m2addr {
nodeOneMined = true
break
}
}
}
if nodeOneMined && complChan != nil {
close(complChan)
complChan = nil
}
}
}()
dh := kit.NewDealHarness(t, client, provider)
deal := dh.StartDeal(ctx, fcid, false, 0)
// TODO: this sleep is only necessary because deals don't immediately get logged in the dealstore, we should fix this
time.Sleep(time.Second)
dh.WaitDealSealed(ctx, deal, false, false, nil)
<-minedTwo
atomic.StoreInt32(&mine, 0)
fmt.Println("shutting down mining")
<-done
}
func TestOfflineDealFlow(t *testing.T) {
blocktime := 10 * time.Millisecond
// For these tests where the block time is artificially short, just use
// a deal start epoch that is guaranteed to be far enough in the future
// so that the deal starts sealing in time
startEpoch := abi.ChainEpoch(2 << 12)
runTest := func(t *testing.T, fastRet bool) {
ctx := context.Background()
fulls, miners := kit.MockMinerBuilder(t, kit.OneFull, kit.OneMiner)
client, miner := fulls[0].FullNode.(*impl.FullNodeAPI), miners[0]
kit.ConnectAndStartMining(t, blocktime, miner, client)
dh := kit.NewDealHarness(t, client, miner)
// Create a random file and import on the client.
res, path, data, err := kit.CreateImportFile(ctx, client, 1, 0)
require.NoError(t, err)
// Get the piece size and commP
fcid := res.Root
pieceInfo, err := client.ClientDealPieceCID(ctx, fcid)
require.NoError(t, err)
fmt.Println("FILE CID: ", fcid)
// Create a storage deal with the miner
maddr, err := miner.ActorAddress(ctx)
require.NoError(t, err)
addr, err := client.WalletDefaultAddress(ctx)
require.NoError(t, err)
// Manual storage deal (offline deal)
dataRef := &storagemarket.DataRef{
TransferType: storagemarket.TTManual,
Root: fcid,
PieceCid: &pieceInfo.PieceCID,
PieceSize: pieceInfo.PieceSize.Unpadded(),
}
proposalCid, err := client.ClientStartDeal(ctx, &api.StartDealParams{
Data: dataRef,
Wallet: addr,
Miner: maddr,
EpochPrice: types.NewInt(1000000),
DealStartEpoch: startEpoch,
MinBlocksDuration: uint64(build.MinDealDuration),
FastRetrieval: fastRet,
})
require.NoError(t, err)
// Wait for the deal to reach StorageDealCheckForAcceptance on the client
cd, err := client.ClientGetDealInfo(ctx, *proposalCid)
require.NoError(t, err)
require.Eventually(t, func() bool {
cd, _ := client.ClientGetDealInfo(ctx, *proposalCid)
return cd.State == storagemarket.StorageDealCheckForAcceptance
}, 30*time.Second, 1*time.Second, "actual deal status is %s", storagemarket.DealStates[cd.State])
// Create a CAR file from the raw file
carFileDir, err := ioutil.TempDir(os.TempDir(), "test-make-deal-car")
require.NoError(t, err)
carFilePath := filepath.Join(carFileDir, "out.car")
err = client.ClientGenCar(ctx, api.FileRef{Path: path}, carFilePath)
require.NoError(t, err)
// Import the CAR file on the miner - this is the equivalent to
// transferring the file across the wire in a normal (non-offline) deal
err = miner.DealsImportData(ctx, *proposalCid, carFilePath)
require.NoError(t, err)
// Wait for the deal to be published
dh.WaitDealPublished(ctx, proposalCid)
t.Logf("deal published, retrieving")
// Retrieve the deal
dh.TestRetrieval(ctx, fcid, &pieceInfo.PieceCID, false, data)
}
t.Run("NormalRetrieval", func(t *testing.T) {
runTest(t, false)
})
t.Run("FastRetrieval", func(t *testing.T) {
runTest(t, true)
})
}
func runFullDealCycles(t *testing.T, n int, b kit.APIBuilder, blocktime time.Duration, carExport, fastRet bool, startEpoch abi.ChainEpoch) {
fulls, miners := b(t, kit.OneFull, kit.OneMiner)
client, miner := fulls[0].FullNode.(*impl.FullNodeAPI), miners[0]
kit.ConnectAndStartMining(t, blocktime, miner, client)
dh := kit.NewDealHarness(t, client, miner)
baseseed := 6
for i := 0; i < n; i++ {
dh.MakeFullDeal(context.Background(), baseseed+i, carExport, fastRet, startEpoch)
}
}
func runFastRetrievalDealFlowT(t *testing.T, b kit.APIBuilder, blocktime time.Duration, startEpoch abi.ChainEpoch) {
ctx := context.Background()
var (
nb = kit.NewNodeBuilder(t)
full = nb.FullNode()
miner = nb.Miner(full)
)
nb.Create()
kit.ConnectAndStartMining(t, blocktime, miner, full)
dh := kit.NewDealHarness(t, full, miner)
data := make([]byte, 1600)
rand.New(rand.NewSource(int64(8))).Read(data)
r := bytes.NewReader(data)
fcid, err := full.FullNode.(*impl.FullNodeAPI).ClientImportLocal(ctx, r)
require.NoError(t, err)
fmt.Println("FILE CID: ", fcid)
deal := dh.StartDeal(ctx, fcid, true, startEpoch)
dh.WaitDealPublished(ctx, deal)
fmt.Println("deal published, retrieving")
// Retrieval
info, err := full.ClientGetDealInfo(ctx, *deal)
require.NoError(t, err)
dh.TestRetrieval(ctx, fcid, &info.PieceCID, false, data)
}
func runSecondDealRetrievalTest(t *testing.T, b kit.APIBuilder, blocktime time.Duration) {
ctx := context.Background()
fulls, miners := b(t, kit.OneFull, kit.OneMiner)
client, miner := fulls[0].FullNode.(*impl.FullNodeAPI), miners[0]
kit.ConnectAndStartMining(t, blocktime, miner, client)
dh := kit.NewDealHarness(t, client, miner)
{
data1 := make([]byte, 800)
rand.New(rand.NewSource(int64(3))).Read(data1)
r := bytes.NewReader(data1)
fcid1, err := client.ClientImportLocal(ctx, r)
if err != nil {
t.Fatal(err)
}
data2 := make([]byte, 800)
rand.New(rand.NewSource(int64(9))).Read(data2)
r2 := bytes.NewReader(data2)
fcid2, err := client.ClientImportLocal(ctx, r2)
if err != nil {
t.Fatal(err)
}
deal1 := dh.StartDeal(ctx, fcid1, true, 0)
// TODO: this sleep is only necessary because deals don't immediately get logged in the dealstore, we should fix this
time.Sleep(time.Second)
dh.WaitDealSealed(ctx, deal1, true, false, nil)
deal2 := dh.StartDeal(ctx, fcid2, true, 0)
time.Sleep(time.Second)
dh.WaitDealSealed(ctx, deal2, false, false, nil)
// Retrieval
info, err := client.ClientGetDealInfo(ctx, *deal2)
require.NoError(t, err)
rf, _ := miner.SectorsRefs(ctx)
fmt.Printf("refs: %+v\n", rf)
dh.TestRetrieval(ctx, fcid2, &info.PieceCID, false, data2)
}
}
func runZeroPricePerByteRetrievalDealFlow(t *testing.T, b kit.APIBuilder, blocktime time.Duration, startEpoch abi.ChainEpoch) {
ctx := context.Background()
fulls, miners := b(t, kit.OneFull, kit.OneMiner)
client, miner := fulls[0].FullNode.(*impl.FullNodeAPI), miners[0]
kit.ConnectAndStartMining(t, blocktime, miner, client)
dh := kit.NewDealHarness(t, client, miner)
// Set price-per-byte to zero
ask, err := miner.MarketGetRetrievalAsk(ctx)
require.NoError(t, err)
ask.PricePerByte = abi.NewTokenAmount(0)
err = miner.MarketSetRetrievalAsk(ctx, ask)
require.NoError(t, err)
dh.MakeFullDeal(ctx, 6, false, false, startEpoch)
}

View File

@ -114,12 +114,13 @@ func NewEnsemble(t *testing.T, opts ...BuilderOpt) *Ensemble {
} }
type NodeOpts struct { type NodeOpts struct {
balance abi.TokenAmount balance abi.TokenAmount
lite bool lite bool
sectors int sectors int
mockProofs bool mockProofs bool
rpc bool rpc bool
ownerKey *wallet.Key ownerKey *wallet.Key
extraNodeOpts []node.Option
} }
var DefaultNodeOpts = NodeOpts{ var DefaultNodeOpts = NodeOpts{
@ -182,6 +183,13 @@ func OwnerAddr(wk *wallet.Key) NodeOpt {
} }
} }
func ExtraNodeOpts(extra ...node.Option) NodeOpt {
return func(opts *NodeOpts) error {
opts.extraNodeOpts = extra
return nil
}
}
// FullNode enrolls a new full node. // FullNode enrolls a new full node.
func (n *Ensemble) FullNode(full *TestFullNode, opts ...NodeOpt) *Ensemble { func (n *Ensemble) FullNode(full *TestFullNode, opts ...NodeOpt) *Ensemble {
options := DefaultNodeOpts options := DefaultNodeOpts
@ -300,6 +308,10 @@ func (n *Ensemble) Start() *Ensemble {
n.mn = mocknet.New(ctx) n.mn = mocknet.New(ctx)
} }
// ---------------------
// FULL NODES
// ---------------------
// Create all inactive full nodes. // Create all inactive full nodes.
for i, full := range n.inactive.fullnodes { for i, full := range n.inactive.fullnodes {
opts := []node.Option{ opts := []node.Option{
@ -313,6 +325,9 @@ func (n *Ensemble) Start() *Ensemble {
node.Override(new(dtypes.Bootstrapper), dtypes.Bootstrapper(true)), node.Override(new(dtypes.Bootstrapper), dtypes.Bootstrapper(true)),
} }
// append any node builder options.
opts = append(opts, full.options.extraNodeOpts...)
// Either generate the genesis or inject it. // Either generate the genesis or inject it.
if i == 0 && !n.bootstrapped { if i == 0 && !n.bootstrapped {
opts = append(opts, node.Override(new(modules.Genesis), testing2.MakeGenesisMem(&n.genesisBlock, *gtempl))) opts = append(opts, node.Override(new(modules.Genesis), testing2.MakeGenesisMem(&n.genesisBlock, *gtempl)))
@ -322,7 +337,10 @@ func (n *Ensemble) Start() *Ensemble {
// Are we mocking proofs? // Are we mocking proofs?
if full.options.mockProofs { if full.options.mockProofs {
opts = append(opts, node.Override(new(ffiwrapper.Verifier), mock.MockVerifier)) opts = append(opts,
node.Override(new(ffiwrapper.Verifier), mock.MockVerifier),
node.Override(new(ffiwrapper.Prover), mock.MockProver),
)
} }
// Construct the full node. // Construct the full node.
@ -356,6 +374,10 @@ func (n *Ensemble) Start() *Ensemble {
err := n.mn.LinkAll() err := n.mn.LinkAll()
require.NoError(n.t, err) require.NoError(n.t, err)
// ---------------------
// MINERS
// ---------------------
// Create all inactive miners. // Create all inactive miners.
for i, m := range n.inactive.miners { for i, m := range n.inactive.miners {
if n.bootstrapped { if n.bootstrapped {
@ -469,23 +491,35 @@ func (n *Ensemble) Start() *Ensemble {
node.Override(new(*lotusminer.Miner), lotusminer.NewTestMiner(mineBlock, m.ActorAddr)), node.Override(new(*lotusminer.Miner), lotusminer.NewTestMiner(mineBlock, m.ActorAddr)),
} }
// append any node builder options.
opts = append(opts, m.options.extraNodeOpts...)
idAddr, err := address.IDFromAddress(m.ActorAddr) idAddr, err := address.IDFromAddress(m.ActorAddr)
require.NoError(n.t, err) require.NoError(n.t, err)
if !n.bootstrapped && m.options.mockProofs { // preload preseals if the network still hasn't bootstrapped.
s := n.genesis.miners[i].Sectors var presealSectors []abi.SectorID
sectors := make([]abi.SectorID, len(s)) if !n.bootstrapped {
for i, sector := range s { sectors := n.genesis.miners[i].Sectors
sectors[i] = abi.SectorID{ for _, sector := range sectors {
presealSectors = append(presealSectors, abi.SectorID{
Miner: abi.ActorID(idAddr), Miner: abi.ActorID(idAddr),
Number: sector.SectorID, Number: sector.SectorID,
} })
} }
}
if m.options.mockProofs {
opts = append(opts, opts = append(opts,
node.Override(new(sectorstorage.SectorManager), func() (sectorstorage.SectorManager, error) { node.Override(new(*mock.SectorMgr), func() (*mock.SectorMgr, error) {
return mock.NewMockSectorMgr(sectors), nil return mock.NewMockSectorMgr(presealSectors), nil
}), }),
node.Override(new(sectorstorage.SectorManager), node.From(new(*mock.SectorMgr))),
node.Override(new(sectorstorage.Unsealer), node.From(new(*mock.SectorMgr))),
node.Override(new(sectorstorage.PieceProvider), node.From(new(*mock.SectorMgr))),
node.Override(new(ffiwrapper.Verifier), mock.MockVerifier), node.Override(new(ffiwrapper.Verifier), mock.MockVerifier),
node.Override(new(ffiwrapper.Prover), mock.MockProver),
node.Unset(new(*sectorstorage.Manager)), node.Unset(new(*sectorstorage.Manager)),
) )
} }
@ -532,9 +566,7 @@ func (n *Ensemble) Start() *Ensemble {
require.NoError(n.t, err) require.NoError(n.t, err)
if !n.bootstrapped && len(n.active.miners) > 0 { if !n.bootstrapped && len(n.active.miners) > 0 {
// We have *just* bootstrapped, so // We have *just* bootstrapped, so mine 2 blocks to setup some CE stuff in some actors
// mine 2 blocks to setup some CE stuff
// in some actors
var wait sync.Mutex var wait sync.Mutex
wait.Lock() wait.Lock()
@ -557,22 +589,32 @@ func (n *Ensemble) Start() *Ensemble {
return n return n
} }
// InterconnectAll connects all full nodes one to another. We do not need to // InterconnectAll connects all miners and full nodes to one another.
// take action with miners, because miners only stay connected to their full
// nodes over JSON-RPC.
func (n *Ensemble) InterconnectAll() *Ensemble { func (n *Ensemble) InterconnectAll() *Ensemble {
// connect full nodes to miners.
for _, from := range n.active.fullnodes {
for _, to := range n.active.miners {
// []*TestMiner to []api.CommonAPI type coercion not possible
// so cannot use variadic form.
n.Connect(from, to)
}
}
// connect full nodes between each other, skipping ourselves.
last := len(n.active.fullnodes) - 1 last := len(n.active.fullnodes) - 1
for i, from := range n.active.fullnodes { for i, from := range n.active.fullnodes {
if i == last { if i == last {
continue continue
} }
n.Connect(from, n.active.fullnodes[i+1:]...) for _, to := range n.active.fullnodes[i+1:] {
n.Connect(from, to)
}
} }
return n return n
} }
// Connect connects one full node to the provided full nodes. // Connect connects one full node to the provided full nodes.
func (n *Ensemble) Connect(from *TestFullNode, to ...*TestFullNode) *Ensemble { func (n *Ensemble) Connect(from api.Common, to ...api.Common) *Ensemble {
addr, err := from.NetAddrsListen(context.Background()) addr, err := from.NetAddrsListen(context.Background())
require.NoError(n.t, err) require.NoError(n.t, err)
@ -584,7 +626,8 @@ func (n *Ensemble) Connect(from *TestFullNode, to ...*TestFullNode) *Ensemble {
} }
// BeginMining kicks off mining for the specified miners. If nil or 0-length, // BeginMining kicks off mining for the specified miners. If nil or 0-length,
// it will kick off mining for all enrolled and active miners. // it will kick off mining for all enrolled and active miners. It also adds a
// cleanup function to stop all mining operations on test teardown.
func (n *Ensemble) BeginMining(blocktime time.Duration, miners ...*TestMiner) []*BlockMiner { func (n *Ensemble) BeginMining(blocktime time.Duration, miners ...*TestMiner) []*BlockMiner {
ctx := context.Background() ctx := context.Background()
@ -601,6 +644,8 @@ func (n *Ensemble) BeginMining(blocktime time.Duration, miners ...*TestMiner) []
for _, m := range miners { for _, m := range miners {
bm := NewBlockMiner(n.t, m) bm := NewBlockMiner(n.t, m)
bm.MineBlocks(ctx, blocktime) bm.MineBlocks(ctx, blocktime)
n.t.Cleanup(bm.Stop)
bms = append(bms, bm) bms = append(bms, bm)
} }