2021-06-10 12:25:36 +00:00
|
|
|
package itests
|
|
|
|
|
|
|
|
import (
|
|
|
|
"bytes"
|
|
|
|
"context"
|
2021-06-10 17:25:02 +00:00
|
|
|
"fmt"
|
2021-06-10 14:54:16 +00:00
|
|
|
"path/filepath"
|
2021-06-10 12:25:36 +00:00
|
|
|
"testing"
|
|
|
|
"time"
|
|
|
|
|
|
|
|
"github.com/filecoin-project/go-fil-markets/storagemarket"
|
|
|
|
"github.com/filecoin-project/go-state-types/abi"
|
|
|
|
"github.com/filecoin-project/lotus/api"
|
2021-06-10 14:54:16 +00:00
|
|
|
"github.com/filecoin-project/lotus/build"
|
2021-06-10 12:25:36 +00:00
|
|
|
"github.com/filecoin-project/lotus/chain/actors/builtin/market"
|
2021-06-22 11:08:36 +00:00
|
|
|
"github.com/filecoin-project/lotus/chain/actors/policy"
|
2021-06-10 12:25:36 +00:00
|
|
|
"github.com/filecoin-project/lotus/chain/types"
|
2021-06-11 04:05:20 +00:00
|
|
|
"github.com/filecoin-project/lotus/extern/sector-storage/storiface"
|
2021-06-18 18:45:29 +00:00
|
|
|
"github.com/filecoin-project/lotus/itests/kit"
|
2021-06-10 12:25:36 +00:00
|
|
|
"github.com/filecoin-project/lotus/markets/storageadapter"
|
|
|
|
"github.com/filecoin-project/lotus/node"
|
|
|
|
market2 "github.com/filecoin-project/specs-actors/v2/actors/builtin/market"
|
|
|
|
"github.com/stretchr/testify/require"
|
2021-06-10 17:25:02 +00:00
|
|
|
"golang.org/x/sync/errgroup"
|
2021-06-10 12:25:36 +00:00
|
|
|
)
|
|
|
|
|
2021-06-10 17:25:02 +00:00
|
|
|
func TestDealCyclesConcurrent(t *testing.T) {
|
2021-06-11 17:26:25 +00:00
|
|
|
if testing.Short() {
|
|
|
|
t.Skip("skipping test in short mode")
|
|
|
|
}
|
|
|
|
|
2021-06-18 18:45:29 +00:00
|
|
|
kit.QuietMiningLogs()
|
2021-06-10 17:25:02 +00:00
|
|
|
|
2021-06-22 11:08:36 +00:00
|
|
|
oldDelay := policy.GetPreCommitChallengeDelay()
|
|
|
|
policy.SetPreCommitChallengeDelay(5)
|
|
|
|
t.Cleanup(func() {
|
|
|
|
policy.SetPreCommitChallengeDelay(oldDelay)
|
|
|
|
})
|
|
|
|
|
2021-06-22 10:05:59 +00:00
|
|
|
blockTime := 10 * time.Millisecond
|
2021-06-10 17:25:02 +00:00
|
|
|
|
|
|
|
// For these tests where the block time is artificially short, just use
|
|
|
|
// a deal start epoch that is guaranteed to be far enough in the future
|
|
|
|
// so that the deal starts sealing in time
|
2021-06-11 17:26:25 +00:00
|
|
|
startEpoch := abi.ChainEpoch(2 << 12)
|
2021-06-10 17:25:02 +00:00
|
|
|
|
|
|
|
runTest := func(t *testing.T, n int, fastRetrieval bool, carExport bool) {
|
2021-06-22 10:05:59 +00:00
|
|
|
api.RunningNodeType = api.NodeMiner // TODO(anteva): fix me
|
2021-06-18 16:40:33 +00:00
|
|
|
|
2021-06-22 09:28:23 +00:00
|
|
|
////TODO: add miner
|
|
|
|
//client, main, _, _ := kit2.EnsembleWithMarket(t, kit2.MockProofs(), kit2.ThroughRPC())
|
2021-06-18 16:40:33 +00:00
|
|
|
|
2021-06-22 09:28:23 +00:00
|
|
|
//dh := kit.NewDealHarness(t, client, main)
|
2021-06-21 17:21:10 +00:00
|
|
|
client, miner, ens := kit.EnsembleMinimal(t, kit.MockProofs())
|
2021-06-10 17:25:02 +00:00
|
|
|
ens.InterconnectAll().BeginMining(blockTime)
|
2021-06-18 18:45:29 +00:00
|
|
|
dh := kit.NewDealHarness(t, client, miner)
|
2021-06-11 17:26:25 +00:00
|
|
|
|
|
|
|
runConcurrentDeals(t, dh, fullDealCyclesOpts{
|
|
|
|
n: n,
|
|
|
|
fastRetrieval: fastRetrieval,
|
|
|
|
carExport: carExport,
|
|
|
|
startEpoch: startEpoch,
|
|
|
|
})
|
2021-06-10 17:25:02 +00:00
|
|
|
}
|
|
|
|
|
2021-06-21 20:35:39 +00:00
|
|
|
// TODO: add 2, 4, 8, more when this graphsync issue is fixed: https://github.com/ipfs/go-graphsync/issues/175#
|
2021-06-18 16:40:33 +00:00
|
|
|
cycles := []int{1}
|
2021-06-10 17:25:02 +00:00
|
|
|
for _, n := range cycles {
|
2021-06-13 23:10:37 +00:00
|
|
|
n := n
|
2021-06-10 17:25:02 +00:00
|
|
|
ns := fmt.Sprintf("%d", n)
|
2021-06-18 16:40:33 +00:00
|
|
|
//t.Run(ns+"-fastretrieval-CAR", func(t *testing.T) { runTest(t, n, true, true) })
|
|
|
|
//t.Run(ns+"-fastretrieval-NoCAR", func(t *testing.T) { runTest(t, n, true, false) })
|
|
|
|
//t.Run(ns+"-stdretrieval-CAR", func(t *testing.T) { runTest(t, n, true, false) })
|
2021-06-10 17:25:02 +00:00
|
|
|
t.Run(ns+"-stdretrieval-NoCAR", func(t *testing.T) { runTest(t, n, false, false) })
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2021-06-11 17:26:25 +00:00
|
|
|
type fullDealCyclesOpts struct {
|
|
|
|
n int
|
|
|
|
fastRetrieval bool
|
|
|
|
carExport bool
|
|
|
|
startEpoch abi.ChainEpoch
|
|
|
|
}
|
|
|
|
|
2021-06-18 18:45:29 +00:00
|
|
|
func runConcurrentDeals(t *testing.T, dh *kit.DealHarness, opts fullDealCyclesOpts) {
|
2021-06-11 17:26:25 +00:00
|
|
|
errgrp, _ := errgroup.WithContext(context.Background())
|
|
|
|
for i := 0; i < opts.n; i++ {
|
|
|
|
i := i
|
|
|
|
errgrp.Go(func() (err error) {
|
|
|
|
defer func() {
|
|
|
|
// This is necessary because golang can't deal with test
|
|
|
|
// failures being reported from children goroutines ¯\_(ツ)_/¯
|
|
|
|
if r := recover(); r != nil {
|
|
|
|
err = fmt.Errorf("deal failed: %s", r)
|
|
|
|
}
|
|
|
|
}()
|
2021-06-21 17:19:26 +00:00
|
|
|
deal, res, inPath := dh.MakeOnlineDeal(context.Background(), kit.MakeFullDealParams{
|
|
|
|
Rseed: 5 + i,
|
|
|
|
FastRet: opts.fastRetrieval,
|
|
|
|
StartEpoch: opts.startEpoch,
|
|
|
|
})
|
2021-06-11 17:26:25 +00:00
|
|
|
outPath := dh.PerformRetrieval(context.Background(), deal, res.Root, opts.carExport)
|
2021-06-18 18:45:29 +00:00
|
|
|
kit.AssertFilesEqual(t, inPath, outPath)
|
2021-06-11 17:26:25 +00:00
|
|
|
return nil
|
|
|
|
})
|
|
|
|
}
|
|
|
|
require.NoError(t, errgrp.Wait())
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestDealsWithSealingAndRPC(t *testing.T) {
|
|
|
|
if testing.Short() {
|
|
|
|
t.Skip("skipping test in short mode")
|
|
|
|
}
|
|
|
|
|
2021-06-18 18:45:29 +00:00
|
|
|
kit.QuietMiningLogs()
|
2021-06-11 17:26:25 +00:00
|
|
|
|
2021-06-22 11:08:36 +00:00
|
|
|
oldDelay := policy.GetPreCommitChallengeDelay()
|
|
|
|
policy.SetPreCommitChallengeDelay(5)
|
|
|
|
t.Cleanup(func() {
|
|
|
|
policy.SetPreCommitChallengeDelay(oldDelay)
|
|
|
|
})
|
|
|
|
|
2021-06-11 17:26:25 +00:00
|
|
|
var blockTime = 1 * time.Second
|
|
|
|
|
2021-06-22 10:05:59 +00:00
|
|
|
client, miner, ens := kit.EnsembleMinimal(t, kit.ThroughRPC(), kit.WithAllSubsystems()) // no mock proofs.
|
2021-06-11 17:26:25 +00:00
|
|
|
ens.InterconnectAll().BeginMining(blockTime)
|
2021-06-18 18:45:29 +00:00
|
|
|
dh := kit.NewDealHarness(t, client, miner)
|
2021-06-11 17:26:25 +00:00
|
|
|
|
|
|
|
t.Run("stdretrieval", func(t *testing.T) {
|
|
|
|
runConcurrentDeals(t, dh, fullDealCyclesOpts{n: 1})
|
|
|
|
})
|
|
|
|
|
|
|
|
t.Run("fastretrieval", func(t *testing.T) {
|
|
|
|
runConcurrentDeals(t, dh, fullDealCyclesOpts{n: 1, fastRetrieval: true})
|
|
|
|
})
|
|
|
|
|
|
|
|
t.Run("fastretrieval-twodeals-sequential", func(t *testing.T) {
|
|
|
|
runConcurrentDeals(t, dh, fullDealCyclesOpts{n: 1, fastRetrieval: true})
|
|
|
|
runConcurrentDeals(t, dh, fullDealCyclesOpts{n: 1, fastRetrieval: true})
|
|
|
|
})
|
|
|
|
}
|
2021-06-10 12:25:36 +00:00
|
|
|
|
2021-06-21 17:19:26 +00:00
|
|
|
func TestQuotePriceForUnsealedRetrieval(t *testing.T) {
|
|
|
|
var (
|
|
|
|
ctx = context.Background()
|
2021-06-22 12:33:44 +00:00
|
|
|
blocktime = time.Second
|
2021-06-21 17:19:26 +00:00
|
|
|
)
|
2021-06-11 04:05:20 +00:00
|
|
|
|
2021-06-21 17:19:26 +00:00
|
|
|
kit.QuietMiningLogs()
|
|
|
|
|
2021-06-21 20:20:45 +00:00
|
|
|
client, miner, ens := kit.EnsembleMinimal(t)
|
2021-06-21 17:19:26 +00:00
|
|
|
ens.InterconnectAll().BeginMining(blocktime)
|
|
|
|
|
|
|
|
var (
|
|
|
|
ppb = int64(1)
|
|
|
|
unsealPrice = int64(77)
|
|
|
|
)
|
2021-06-11 04:05:20 +00:00
|
|
|
|
|
|
|
// Set unsealed price to non-zero
|
|
|
|
ask, err := miner.MarketGetRetrievalAsk(ctx)
|
|
|
|
require.NoError(t, err)
|
|
|
|
ask.PricePerByte = abi.NewTokenAmount(ppb)
|
|
|
|
ask.UnsealPrice = abi.NewTokenAmount(unsealPrice)
|
|
|
|
err = miner.MarketSetRetrievalAsk(ctx, ask)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
dh := kit.NewDealHarness(t, client, miner)
|
|
|
|
|
2021-06-21 17:19:26 +00:00
|
|
|
deal1, res1, _ := dh.MakeOnlineDeal(ctx, kit.MakeFullDealParams{Rseed: 6})
|
2021-06-11 04:05:20 +00:00
|
|
|
|
|
|
|
// one more storage deal for the same data
|
2021-06-21 17:19:26 +00:00
|
|
|
_, res2, _ := dh.MakeOnlineDeal(ctx, kit.MakeFullDealParams{Rseed: 6})
|
|
|
|
require.Equal(t, res1.Root, res2.Root)
|
|
|
|
|
|
|
|
// Retrieval
|
|
|
|
dealInfo, err := client.ClientGetDealInfo(ctx, *deal1)
|
|
|
|
require.NoError(t, err)
|
2021-06-11 04:05:20 +00:00
|
|
|
|
|
|
|
// fetch quote -> zero for unsealed price since unsealed file already exists.
|
2021-06-21 17:19:26 +00:00
|
|
|
offers, err := client.ClientFindData(ctx, res1.Root, &dealInfo.PieceCID)
|
2021-06-11 04:05:20 +00:00
|
|
|
require.NoError(t, err)
|
|
|
|
require.Len(t, offers, 2)
|
|
|
|
require.Equal(t, offers[0], offers[1])
|
|
|
|
require.Equal(t, uint64(0), offers[0].UnsealPrice.Uint64())
|
2021-06-21 17:19:26 +00:00
|
|
|
require.Equal(t, dealInfo.Size*uint64(ppb), offers[0].MinPrice.Uint64())
|
2021-06-11 04:05:20 +00:00
|
|
|
|
|
|
|
// remove ONLY one unsealed file
|
|
|
|
ss, err := miner.StorageList(context.Background())
|
|
|
|
require.NoError(t, err)
|
|
|
|
_, err = miner.SectorsList(ctx)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
iLoop:
|
|
|
|
for storeID, sd := range ss {
|
|
|
|
for _, sector := range sd {
|
2021-06-21 17:19:26 +00:00
|
|
|
err := miner.StorageDropSector(ctx, storeID, sector.SectorID, storiface.FTUnsealed)
|
|
|
|
require.NoError(t, err)
|
|
|
|
break iLoop // remove ONLY one
|
2021-06-11 04:05:20 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// get retrieval quote -> zero for unsealed price as unsealed file exists.
|
2021-06-21 17:19:26 +00:00
|
|
|
offers, err = client.ClientFindData(ctx, res1.Root, &dealInfo.PieceCID)
|
2021-06-11 04:05:20 +00:00
|
|
|
require.NoError(t, err)
|
|
|
|
require.Len(t, offers, 2)
|
|
|
|
require.Equal(t, offers[0], offers[1])
|
|
|
|
require.Equal(t, uint64(0), offers[0].UnsealPrice.Uint64())
|
2021-06-21 17:19:26 +00:00
|
|
|
require.Equal(t, dealInfo.Size*uint64(ppb), offers[0].MinPrice.Uint64())
|
2021-06-11 04:05:20 +00:00
|
|
|
|
|
|
|
// remove the other unsealed file as well
|
|
|
|
ss, err = miner.StorageList(context.Background())
|
|
|
|
require.NoError(t, err)
|
|
|
|
_, err = miner.SectorsList(ctx)
|
|
|
|
require.NoError(t, err)
|
|
|
|
for storeID, sd := range ss {
|
|
|
|
for _, sector := range sd {
|
|
|
|
require.NoError(t, miner.StorageDropSector(ctx, storeID, sector.SectorID, storiface.FTUnsealed))
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// fetch quote -> non-zero for unseal price as we no more unsealed files.
|
2021-06-21 17:19:26 +00:00
|
|
|
offers, err = client.ClientFindData(ctx, res1.Root, &dealInfo.PieceCID)
|
2021-06-11 04:05:20 +00:00
|
|
|
require.NoError(t, err)
|
|
|
|
require.Len(t, offers, 2)
|
|
|
|
require.Equal(t, offers[0], offers[1])
|
|
|
|
require.Equal(t, uint64(unsealPrice), offers[0].UnsealPrice.Uint64())
|
2021-06-21 17:19:26 +00:00
|
|
|
total := (dealInfo.Size * uint64(ppb)) + uint64(unsealPrice)
|
2021-06-11 04:05:20 +00:00
|
|
|
require.Equal(t, total, offers[0].MinPrice.Uint64())
|
2021-06-21 17:19:26 +00:00
|
|
|
|
2021-05-18 20:32:10 +00:00
|
|
|
}
|
|
|
|
|
2021-06-10 12:25:36 +00:00
|
|
|
func TestPublishDealsBatching(t *testing.T) {
|
|
|
|
var (
|
|
|
|
ctx = context.Background()
|
|
|
|
publishPeriod = 10 * time.Second
|
|
|
|
maxDealsPerMsg = uint64(2) // Set max deals per publish deals message to 2
|
|
|
|
startEpoch = abi.ChainEpoch(2 << 12)
|
|
|
|
)
|
|
|
|
|
2021-06-22 11:08:36 +00:00
|
|
|
oldDelay := policy.GetPreCommitChallengeDelay()
|
|
|
|
policy.SetPreCommitChallengeDelay(5)
|
|
|
|
t.Cleanup(func() {
|
|
|
|
policy.SetPreCommitChallengeDelay(oldDelay)
|
|
|
|
})
|
|
|
|
|
2021-06-18 18:45:29 +00:00
|
|
|
kit.QuietMiningLogs()
|
2021-06-10 12:25:36 +00:00
|
|
|
|
|
|
|
opts := node.Override(new(*storageadapter.DealPublisher),
|
|
|
|
storageadapter.NewDealPublisher(nil, storageadapter.PublishMsgConfig{
|
|
|
|
Period: publishPeriod,
|
|
|
|
MaxDealsPerMsg: maxDealsPerMsg,
|
|
|
|
}),
|
|
|
|
)
|
|
|
|
|
2021-06-18 18:45:29 +00:00
|
|
|
client, miner, ens := kit.EnsembleMinimal(t, kit.MockProofs(), kit.ConstructorOpts(opts))
|
2021-06-10 12:25:36 +00:00
|
|
|
ens.InterconnectAll().BeginMining(10 * time.Millisecond)
|
|
|
|
|
2021-06-18 18:45:29 +00:00
|
|
|
dh := kit.NewDealHarness(t, client, miner)
|
2021-06-10 12:25:36 +00:00
|
|
|
|
|
|
|
// Starts a deal and waits until it's published
|
|
|
|
runDealTillPublish := func(rseed int) {
|
2021-06-10 17:25:02 +00:00
|
|
|
res, _ := client.CreateImportFile(ctx, rseed, 0)
|
2021-06-10 12:25:36 +00:00
|
|
|
|
|
|
|
upds, err := client.ClientGetDealUpdates(ctx)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
dh.StartDeal(ctx, res.Root, false, startEpoch)
|
|
|
|
|
|
|
|
// TODO: this sleep is only necessary because deals don't immediately get logged in the dealstore, we should fix this
|
|
|
|
time.Sleep(time.Second)
|
|
|
|
|
|
|
|
done := make(chan struct{})
|
|
|
|
go func() {
|
|
|
|
for upd := range upds {
|
|
|
|
if upd.DataRef.Root == res.Root && upd.State == storagemarket.StorageDealAwaitingPreCommit {
|
|
|
|
done <- struct{}{}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}()
|
|
|
|
<-done
|
|
|
|
}
|
|
|
|
|
|
|
|
// Run three deals in parallel
|
|
|
|
done := make(chan struct{}, maxDealsPerMsg+1)
|
|
|
|
for rseed := 1; rseed <= 3; rseed++ {
|
|
|
|
rseed := rseed
|
|
|
|
go func() {
|
|
|
|
runDealTillPublish(rseed)
|
|
|
|
done <- struct{}{}
|
|
|
|
}()
|
|
|
|
}
|
|
|
|
|
|
|
|
// Wait for two of the deals to be published
|
|
|
|
for i := 0; i < int(maxDealsPerMsg); i++ {
|
|
|
|
<-done
|
|
|
|
}
|
|
|
|
|
|
|
|
// Expect a single PublishStorageDeals message that includes the first two deals
|
|
|
|
msgCids, err := client.StateListMessages(ctx, &api.MessageMatch{To: market.Address}, types.EmptyTSK, 1)
|
|
|
|
require.NoError(t, err)
|
|
|
|
count := 0
|
|
|
|
for _, msgCid := range msgCids {
|
|
|
|
msg, err := client.ChainGetMessage(ctx, msgCid)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
if msg.Method == market.Methods.PublishStorageDeals {
|
|
|
|
count++
|
|
|
|
var pubDealsParams market2.PublishStorageDealsParams
|
|
|
|
err = pubDealsParams.UnmarshalCBOR(bytes.NewReader(msg.Params))
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.Len(t, pubDealsParams.Deals, int(maxDealsPerMsg))
|
|
|
|
}
|
|
|
|
}
|
|
|
|
require.Equal(t, 1, count)
|
|
|
|
|
|
|
|
// The third deal should be published once the publish period expires.
|
|
|
|
// Allow a little padding as it takes a moment for the state change to
|
|
|
|
// be noticed by the client.
|
|
|
|
padding := 10 * time.Second
|
|
|
|
select {
|
|
|
|
case <-time.After(publishPeriod + padding):
|
|
|
|
require.Fail(t, "Expected 3rd deal to be published once publish period elapsed")
|
|
|
|
case <-done: // Success
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2021-06-10 13:04:39 +00:00
|
|
|
func TestFirstDealEnablesMining(t *testing.T) {
|
|
|
|
// test making a deal with a fresh miner, and see if it starts to mine.
|
|
|
|
if testing.Short() {
|
|
|
|
t.Skip("skipping test in short mode")
|
|
|
|
}
|
|
|
|
|
2021-06-18 18:45:29 +00:00
|
|
|
kit.QuietMiningLogs()
|
2021-06-10 13:04:39 +00:00
|
|
|
|
|
|
|
var (
|
2021-06-18 18:45:29 +00:00
|
|
|
client kit.TestFullNode
|
|
|
|
genMiner kit.TestMiner // bootstrap
|
|
|
|
provider kit.TestMiner // no sectors, will need to create one
|
2021-06-10 13:04:39 +00:00
|
|
|
)
|
|
|
|
|
2021-06-18 18:45:29 +00:00
|
|
|
ens := kit.NewEnsemble(t, kit.MockProofs())
|
2021-06-11 17:26:25 +00:00
|
|
|
ens.FullNode(&client)
|
|
|
|
ens.Miner(&genMiner, &client)
|
2021-06-18 18:45:29 +00:00
|
|
|
ens.Miner(&provider, &client, kit.PresealSectors(0))
|
2021-06-10 13:04:39 +00:00
|
|
|
ens.Start().InterconnectAll().BeginMining(50 * time.Millisecond)
|
|
|
|
|
|
|
|
ctx := context.Background()
|
|
|
|
|
2021-06-18 18:45:29 +00:00
|
|
|
dh := kit.NewDealHarness(t, &client, &provider)
|
2021-06-10 13:04:39 +00:00
|
|
|
|
2021-06-10 17:25:02 +00:00
|
|
|
ref, _ := client.CreateImportFile(ctx, 5, 0)
|
2021-06-10 13:04:39 +00:00
|
|
|
|
|
|
|
t.Log("FILE CID:", ref.Root)
|
|
|
|
|
|
|
|
ctx, cancel := context.WithCancel(ctx)
|
|
|
|
defer cancel()
|
|
|
|
|
|
|
|
// start a goroutine to monitor head changes from the client
|
|
|
|
// once the provider has mined a block, thanks to the power acquired from the deal,
|
|
|
|
// we pass the test.
|
|
|
|
providerMined := make(chan struct{})
|
2021-06-10 17:25:02 +00:00
|
|
|
|
2021-06-10 13:04:39 +00:00
|
|
|
go func() {
|
2021-06-18 18:45:29 +00:00
|
|
|
_ = client.WaitTillChain(ctx, kit.BlockMinedBy(provider.ActorAddr))
|
2021-06-17 20:58:29 +00:00
|
|
|
close(providerMined)
|
2021-06-10 13:04:39 +00:00
|
|
|
}()
|
|
|
|
|
|
|
|
// now perform the deal.
|
|
|
|
deal := dh.StartDeal(ctx, ref.Root, false, 0)
|
|
|
|
|
|
|
|
// TODO: this sleep is only necessary because deals don't immediately get logged in the dealstore, we should fix this
|
|
|
|
time.Sleep(time.Second)
|
|
|
|
|
|
|
|
dh.WaitDealSealed(ctx, deal, false, false, nil)
|
|
|
|
|
|
|
|
<-providerMined
|
|
|
|
}
|
|
|
|
|
2021-06-10 14:54:16 +00:00
|
|
|
func TestOfflineDealFlow(t *testing.T) {
|
|
|
|
blocktime := 10 * time.Millisecond
|
|
|
|
|
|
|
|
// For these tests where the block time is artificially short, just use
|
|
|
|
// a deal start epoch that is guaranteed to be far enough in the future
|
|
|
|
// so that the deal starts sealing in time
|
|
|
|
startEpoch := abi.ChainEpoch(2 << 12)
|
|
|
|
|
2021-06-22 11:08:36 +00:00
|
|
|
oldDelay := policy.GetPreCommitChallengeDelay()
|
|
|
|
policy.SetPreCommitChallengeDelay(5)
|
|
|
|
t.Cleanup(func() {
|
|
|
|
policy.SetPreCommitChallengeDelay(oldDelay)
|
|
|
|
})
|
|
|
|
|
2021-06-10 14:54:16 +00:00
|
|
|
runTest := func(t *testing.T, fastRet bool) {
|
|
|
|
ctx := context.Background()
|
2021-06-18 18:45:29 +00:00
|
|
|
client, miner, ens := kit.EnsembleMinimal(t, kit.MockProofs())
|
2021-06-10 14:54:16 +00:00
|
|
|
ens.InterconnectAll().BeginMining(blocktime)
|
|
|
|
|
2021-06-18 18:45:29 +00:00
|
|
|
dh := kit.NewDealHarness(t, client, miner)
|
2021-06-10 14:54:16 +00:00
|
|
|
|
|
|
|
// Create a random file and import on the client.
|
2021-06-10 17:25:02 +00:00
|
|
|
res, inFile := client.CreateImportFile(ctx, 1, 0)
|
2021-06-10 14:54:16 +00:00
|
|
|
|
|
|
|
// Get the piece size and commP
|
2021-06-10 17:25:02 +00:00
|
|
|
rootCid := res.Root
|
|
|
|
pieceInfo, err := client.ClientDealPieceCID(ctx, rootCid)
|
2021-06-10 14:54:16 +00:00
|
|
|
require.NoError(t, err)
|
2021-06-10 17:25:02 +00:00
|
|
|
t.Log("FILE CID:", rootCid)
|
2021-06-10 14:54:16 +00:00
|
|
|
|
|
|
|
// Create a storage deal with the miner
|
|
|
|
maddr, err := miner.ActorAddress(ctx)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
addr, err := client.WalletDefaultAddress(ctx)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
// Manual storage deal (offline deal)
|
|
|
|
dataRef := &storagemarket.DataRef{
|
|
|
|
TransferType: storagemarket.TTManual,
|
2021-06-10 17:25:02 +00:00
|
|
|
Root: rootCid,
|
2021-06-10 14:54:16 +00:00
|
|
|
PieceCid: &pieceInfo.PieceCID,
|
|
|
|
PieceSize: pieceInfo.PieceSize.Unpadded(),
|
|
|
|
}
|
|
|
|
|
|
|
|
proposalCid, err := client.ClientStartDeal(ctx, &api.StartDealParams{
|
|
|
|
Data: dataRef,
|
|
|
|
Wallet: addr,
|
|
|
|
Miner: maddr,
|
|
|
|
EpochPrice: types.NewInt(1000000),
|
|
|
|
DealStartEpoch: startEpoch,
|
|
|
|
MinBlocksDuration: uint64(build.MinDealDuration),
|
|
|
|
FastRetrieval: fastRet,
|
|
|
|
})
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
// Wait for the deal to reach StorageDealCheckForAcceptance on the client
|
|
|
|
cd, err := client.ClientGetDealInfo(ctx, *proposalCid)
|
|
|
|
require.NoError(t, err)
|
|
|
|
require.Eventually(t, func() bool {
|
|
|
|
cd, _ := client.ClientGetDealInfo(ctx, *proposalCid)
|
|
|
|
return cd.State == storagemarket.StorageDealCheckForAcceptance
|
|
|
|
}, 30*time.Second, 1*time.Second, "actual deal status is %s", storagemarket.DealStates[cd.State])
|
|
|
|
|
|
|
|
// Create a CAR file from the raw file
|
2021-06-10 17:25:02 +00:00
|
|
|
carFileDir := t.TempDir()
|
2021-06-10 14:54:16 +00:00
|
|
|
carFilePath := filepath.Join(carFileDir, "out.car")
|
2021-06-10 17:25:02 +00:00
|
|
|
err = client.ClientGenCar(ctx, api.FileRef{Path: inFile}, carFilePath)
|
2021-06-10 14:54:16 +00:00
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
// Import the CAR file on the miner - this is the equivalent to
|
|
|
|
// transferring the file across the wire in a normal (non-offline) deal
|
|
|
|
err = miner.DealsImportData(ctx, *proposalCid, carFilePath)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
// Wait for the deal to be published
|
|
|
|
dh.WaitDealPublished(ctx, proposalCid)
|
|
|
|
|
|
|
|
t.Logf("deal published, retrieving")
|
|
|
|
|
|
|
|
// Retrieve the deal
|
2021-06-10 17:25:02 +00:00
|
|
|
outFile := dh.PerformRetrieval(ctx, proposalCid, rootCid, false)
|
2021-06-10 14:54:16 +00:00
|
|
|
|
2021-06-18 18:45:29 +00:00
|
|
|
kit.AssertFilesEqual(t, inFile, outFile)
|
2021-05-20 15:12:42 +00:00
|
|
|
|
2021-06-10 17:25:02 +00:00
|
|
|
}
|
2021-06-10 14:54:16 +00:00
|
|
|
|
2021-06-11 17:26:25 +00:00
|
|
|
t.Run("stdretrieval", func(t *testing.T) { runTest(t, false) })
|
|
|
|
t.Run("fastretrieval", func(t *testing.T) { runTest(t, true) })
|
2021-06-10 14:54:16 +00:00
|
|
|
}
|
|
|
|
|
2021-06-13 22:43:22 +00:00
|
|
|
func TestZeroPricePerByteRetrieval(t *testing.T) {
|
|
|
|
if testing.Short() {
|
|
|
|
t.Skip("skipping test in short mode")
|
|
|
|
}
|
|
|
|
|
2021-06-22 11:08:36 +00:00
|
|
|
oldDelay := policy.GetPreCommitChallengeDelay()
|
|
|
|
policy.SetPreCommitChallengeDelay(5)
|
|
|
|
t.Cleanup(func() {
|
|
|
|
policy.SetPreCommitChallengeDelay(oldDelay)
|
|
|
|
})
|
|
|
|
|
2021-06-18 18:45:29 +00:00
|
|
|
kit.QuietMiningLogs()
|
2021-06-13 22:43:22 +00:00
|
|
|
|
|
|
|
var (
|
|
|
|
blockTime = 10 * time.Millisecond
|
|
|
|
startEpoch = abi.ChainEpoch(2 << 12)
|
|
|
|
)
|
|
|
|
|
2021-06-18 18:45:29 +00:00
|
|
|
client, miner, ens := kit.EnsembleMinimal(t, kit.MockProofs())
|
2021-06-13 22:43:22 +00:00
|
|
|
ens.InterconnectAll().BeginMining(blockTime)
|
|
|
|
|
|
|
|
ctx := context.Background()
|
|
|
|
|
|
|
|
ask, err := miner.MarketGetRetrievalAsk(ctx)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
|
|
|
ask.PricePerByte = abi.NewTokenAmount(0)
|
|
|
|
err = miner.MarketSetRetrievalAsk(ctx, ask)
|
|
|
|
require.NoError(t, err)
|
|
|
|
|
2021-06-18 18:45:29 +00:00
|
|
|
dh := kit.NewDealHarness(t, client, miner)
|
2021-06-13 22:43:22 +00:00
|
|
|
runConcurrentDeals(t, dh, fullDealCyclesOpts{
|
|
|
|
n: 1,
|
|
|
|
startEpoch: startEpoch,
|
|
|
|
})
|
|
|
|
}
|