package kit import ( "context" "errors" "fmt" "os" "strings" "testing" "time" "github.com/ipfs/boxo/files" dag "github.com/ipfs/boxo/ipld/merkledag" dstest "github.com/ipfs/boxo/ipld/merkledag/test" unixfile "github.com/ipfs/boxo/ipld/unixfs/file" "github.com/ipfs/go-cid" ipldcbor "github.com/ipfs/go-ipld-cbor" ipld "github.com/ipfs/go-ipld-format" "github.com/ipld/go-car" _ "github.com/ipld/go-ipld-prime/codec/dagcbor" "github.com/stretchr/testify/require" "golang.org/x/sync/errgroup" "github.com/filecoin-project/go-address" "github.com/filecoin-project/go-fil-markets/retrievalmarket" "github.com/filecoin-project/go-fil-markets/shared_testutil" "github.com/filecoin-project/go-fil-markets/storagemarket" "github.com/filecoin-project/go-state-types/abi" "github.com/filecoin-project/lotus/api" "github.com/filecoin-project/lotus/build" "github.com/filecoin-project/lotus/chain/types" sealing "github.com/filecoin-project/lotus/storage/pipeline" ) type DealHarness struct { t *testing.T client *TestFullNode main *TestMiner market *TestMiner } type MakeFullDealParams struct { Rseed int FastRet bool StartEpoch abi.ChainEpoch UseCARFileForStorageDeal bool // SuspendUntilCryptoeconStable suspends deal-making, until cryptoecon // parameters are stabilised. This affects projected collateral, and tests // will fail in network version 13 and higher if deals are started too soon // after network birth. // // The reason is that the formula for collateral calculation takes // circulating supply into account: // // [portion of power this deal will be] * [~1% of tokens]. // // In the first epochs after genesis, the total circulating supply is // changing dramatically in percentual terms. Therefore, if the deal is // proposed too soon, by the time it gets published on chain, the quoted // provider collateral will no longer be valid. // // The observation is that deals fail with: // // GasEstimateMessageGas error: estimating gas used: message execution // failed: exit 16, reason: Provider collateral out of bounds. (RetCode=16) // // Enabling this will suspend deal-making until the network has reached a // height of 300. SuspendUntilCryptoeconStable bool } // NewDealHarness creates a test harness that contains testing utilities for deals. func NewDealHarness(t *testing.T, client *TestFullNode, main *TestMiner, market *TestMiner) *DealHarness { return &DealHarness{ t: t, client: client, main: main, market: market, } } // MakeOnlineDeal makes an online deal, generating a random file with the // supplied seed, and setting the specified fast retrieval flag and start epoch // on the storage deal. It returns when the deal is sealed. // // TODO: convert input parameters to struct, and add size as an input param. func (dh *DealHarness) MakeOnlineDeal(ctx context.Context, params MakeFullDealParams) (deal *cid.Cid, res *api.ImportRes, path string) { deal, res, path = dh.StartRandomDeal(ctx, params) fmt.Printf("WAIT DEAL SEALEDS START\n") dh.WaitDealSealed(ctx, deal, false, false, nil) fmt.Printf("WAIT DEAL SEALEDS END\n") return deal, res, path } func (dh *DealHarness) StartRandomDeal(ctx context.Context, params MakeFullDealParams) (deal *cid.Cid, res *api.ImportRes, path string) { if params.UseCARFileForStorageDeal { res, _, path = dh.client.ClientImportCARFile(ctx, params.Rseed, 200) } else { res, path = dh.client.CreateImportFile(ctx, params.Rseed, 0) } dh.t.Logf("FILE CID: %s", res.Root) if params.SuspendUntilCryptoeconStable { dh.t.Logf("deal-making suspending until cryptecon parameters have stabilised") ts := dh.client.WaitTillChain(ctx, HeightAtLeast(300)) dh.t.Logf("deal-making continuing; current height is %d", ts.Height()) } dp := dh.DefaultStartDealParams() dp.Data.Root = res.Root dp.DealStartEpoch = params.StartEpoch dp.FastRetrieval = params.FastRet deal = dh.StartDeal(ctx, dp) return deal, res, path } func (dh *DealHarness) DefaultStartDealParams() api.StartDealParams { dp := api.StartDealParams{ Data: &storagemarket.DataRef{TransferType: storagemarket.TTGraphsync}, EpochPrice: types.NewInt(1000000), MinBlocksDuration: uint64(build.MinDealDuration), } var err error dp.Miner, err = dh.main.ActorAddress(context.Background()) require.NoError(dh.t, err) dp.Wallet, err = dh.client.WalletDefaultAddress(context.Background()) require.NoError(dh.t, err) return dp } // StartDeal starts a storage deal between the client and the miner. func (dh *DealHarness) StartDeal(ctx context.Context, dealParams api.StartDealParams) *cid.Cid { dealProposalCid, err := dh.client.ClientStartDeal(ctx, &dealParams) require.NoError(dh.t, err) return dealProposalCid } // WaitDealSealed waits until the deal is sealed. func (dh *DealHarness) WaitDealSealed(ctx context.Context, deal *cid.Cid, noseal, noSealStart bool, cb func()) { loop: for { di, err := dh.client.ClientGetDealInfo(ctx, *deal) require.NoError(dh.t, err) switch di.State { case storagemarket.StorageDealAwaitingPreCommit, storagemarket.StorageDealSealing: if noseal { return } if !noSealStart { dh.StartSealingWaiting(ctx) } case storagemarket.StorageDealProposalRejected: dh.t.Fatal("deal rejected") case storagemarket.StorageDealFailing: dh.t.Fatal("deal failed") case storagemarket.StorageDealError: dh.t.Fatal("deal errored", di.Message) case storagemarket.StorageDealActive: dh.t.Log("COMPLETE", di) break loop } mds, err := dh.market.MarketListIncompleteDeals(ctx) require.NoError(dh.t, err) var minerState storagemarket.StorageDealStatus for _, md := range mds { if md.DealID == di.DealID { minerState = md.State break } } dh.t.Logf("Deal %d state: client:%s provider:%s\n", di.DealID, storagemarket.DealStates[di.State], storagemarket.DealStates[minerState]) time.Sleep(time.Second / 2) if cb != nil { cb() } } fmt.Printf("WAIT DEAL SEALED LOOP BROKEN\n") } // WaitDealSealedQuiet waits until the deal is sealed, without logging anything. func (dh *DealHarness) WaitDealSealedQuiet(ctx context.Context, deal *cid.Cid, noseal, noSealStart bool, cb func()) { loop: for { di, err := dh.client.ClientGetDealInfo(ctx, *deal) require.NoError(dh.t, err) switch di.State { case storagemarket.StorageDealAwaitingPreCommit, storagemarket.StorageDealSealing: if noseal { return } if !noSealStart { dh.StartSealingWaiting(ctx) } case storagemarket.StorageDealProposalRejected: dh.t.Fatal("deal rejected") case storagemarket.StorageDealFailing: dh.t.Fatal("deal failed") case storagemarket.StorageDealError: dh.t.Fatal("deal errored", di.Message) case storagemarket.StorageDealActive: break loop } _, err = dh.market.MarketListIncompleteDeals(ctx) require.NoError(dh.t, err) time.Sleep(time.Second / 2) if cb != nil { cb() } } } func (dh *DealHarness) ExpectDealFailure(ctx context.Context, deal *cid.Cid, errs string) error { for { di, err := dh.client.ClientGetDealInfo(ctx, *deal) require.NoError(dh.t, err) switch di.State { case storagemarket.StorageDealAwaitingPreCommit, storagemarket.StorageDealSealing: return fmt.Errorf("deal is sealing, and we expected an error: %s", errs) case storagemarket.StorageDealProposalRejected: if strings.Contains(di.Message, errs) { return nil } return fmt.Errorf("unexpected error: %s ; expected: %s", di.Message, errs) case storagemarket.StorageDealFailing: if strings.Contains(di.Message, errs) { return nil } return fmt.Errorf("unexpected error: %s ; expected: %s", di.Message, errs) case storagemarket.StorageDealError: if strings.Contains(di.Message, errs) { return nil } return fmt.Errorf("unexpected error: %s ; expected: %s", di.Message, errs) case storagemarket.StorageDealActive: return errors.New("expected to get an error, but didn't get one") } mds, err := dh.market.MarketListIncompleteDeals(ctx) require.NoError(dh.t, err) var minerState storagemarket.StorageDealStatus for _, md := range mds { if md.DealID == di.DealID { minerState = md.State break } } dh.t.Logf("Deal %d state: client:%s provider:%s\n", di.DealID, storagemarket.DealStates[di.State], storagemarket.DealStates[minerState]) time.Sleep(time.Second / 2) } } // WaitDealPublished waits until the deal is published. func (dh *DealHarness) WaitDealPublished(ctx context.Context, deal *cid.Cid) { subCtx, cancel := context.WithCancel(ctx) defer cancel() updates, err := dh.market.MarketGetDealUpdates(subCtx) require.NoError(dh.t, err) for { select { case <-ctx.Done(): dh.t.Fatal("context timeout") case di := <-updates: if deal.Equals(di.ProposalCid) { switch di.State { case storagemarket.StorageDealProposalRejected: dh.t.Fatal("deal rejected") case storagemarket.StorageDealFailing: dh.t.Fatal("deal failed") case storagemarket.StorageDealError: dh.t.Fatal("deal errored", di.Message) case storagemarket.StorageDealFinalizing, storagemarket.StorageDealAwaitingPreCommit, storagemarket.StorageDealSealing, storagemarket.StorageDealActive: dh.t.Log("COMPLETE", di) return } dh.t.Log("Deal state: ", storagemarket.DealStates[di.State]) } } } } func (dh *DealHarness) StartSealingWaiting(ctx context.Context) { snums, err := dh.main.SectorsListNonGenesis(ctx) require.NoError(dh.t, err) for _, snum := range snums { si, err := dh.main.SectorsStatus(ctx, snum, false) require.NoError(dh.t, err) dh.t.Logf("Sector state <%d>-[%d]:, %s", snum, si.SealProof, si.State) if si.State == api.SectorState(sealing.WaitDeals) { require.NoError(dh.t, dh.main.SectorStartSealing(ctx, snum)) } dh.main.FlushSealingBatches(ctx) } } func (dh *DealHarness) PerformRetrieval(ctx context.Context, deal *cid.Cid, root cid.Cid, carExport bool, offers ...api.QueryOffer) (path string) { return dh.PerformRetrievalWithOrder(ctx, deal, root, carExport, func(offer api.QueryOffer, a address.Address) api.RetrievalOrder { return offer.Order(a) }, offers...) } func (dh *DealHarness) PerformRetrievalWithOrder(ctx context.Context, deal *cid.Cid, root cid.Cid, carExport bool, makeOrder func(api.QueryOffer, address.Address) api.RetrievalOrder, offers ...api.QueryOffer) (path string) { var offer api.QueryOffer if len(offers) == 0 { // perform retrieval. info, err := dh.client.ClientGetDealInfo(ctx, *deal) require.NoError(dh.t, err) offers, err := dh.client.ClientFindData(ctx, root, &info.PieceCID) require.NoError(dh.t, err) require.NotEmpty(dh.t, offers, "no offers") offer = offers[0] } else { offer = offers[0] } carFile := dh.t.TempDir() + string(os.PathSeparator) + "ret-car-" + root.String() caddr, err := dh.client.WalletDefaultAddress(ctx) require.NoError(dh.t, err) updatesCtx, cancel := context.WithCancel(ctx) updates, err := dh.client.ClientGetRetrievalUpdates(updatesCtx) require.NoError(dh.t, err) order := makeOrder(offer, caddr) retrievalRes, err := dh.client.ClientRetrieve(ctx, order) require.NoError(dh.t, err) consumeEvents: for { var evt api.RetrievalInfo select { case <-updatesCtx.Done(): dh.t.Fatal("Retrieval Timed Out") case evt = <-updates: if evt.ID != retrievalRes.DealID { continue } } switch evt.Status { case retrievalmarket.DealStatusCompleted: break consumeEvents case retrievalmarket.DealStatusRejected: dh.t.Fatalf("Retrieval Proposal Rejected: %s", evt.Message) case retrievalmarket.DealStatusDealNotFound, retrievalmarket.DealStatusErrored: dh.t.Fatalf("Retrieval Error: %s", evt.Message) } } cancel() if order.RemoteStore != nil { // if we're retrieving into a remote store, skip export return "" } require.NoError(dh.t, dh.client.ClientExport(ctx, api.ExportRef{ Root: root, DealID: retrievalRes.DealID, }, api.FileRef{ Path: carFile, IsCAR: carExport, })) ret := carFile return ret } func (dh *DealHarness) ExtractFileFromCAR(ctx context.Context, file *os.File) string { bserv := dstest.Bserv() ch, err := car.LoadCar(ctx, bserv.Blockstore(), file) require.NoError(dh.t, err) blk, err := bserv.GetBlock(ctx, ch.Roots[0]) require.NoError(dh.t, err) reg := ipld.Registry{} reg.Register(cid.DagProtobuf, dag.DecodeProtobufBlock) reg.Register(cid.DagCBOR, ipldcbor.DecodeBlock) reg.Register(cid.Raw, dag.DecodeRawBlock) nd, err := reg.Decode(blk) require.NoError(dh.t, err) dserv := dag.NewDAGService(bserv) fil, err := unixfile.NewUnixfsFile(ctx, dserv, nd) require.NoError(dh.t, err) tmpfile := dh.t.TempDir() + string(os.PathSeparator) + "file-in-car" + nd.Cid().String() err = files.WriteTo(fil, tmpfile) require.NoError(dh.t, err) return tmpfile } type RunConcurrentDealsOpts struct { N int FastRetrieval bool CarExport bool StartEpoch abi.ChainEpoch UseCARFileForStorageDeal bool IndexProvider *shared_testutil.MockIndexProvider } func (dh *DealHarness) RunConcurrentDeals(opts RunConcurrentDealsOpts) { ctx := context.Background() errgrp, _ := errgroup.WithContext(context.Background()) for i := 0; i < opts.N; i++ { i := i errgrp.Go(func() (err error) { defer dh.t.Logf("finished concurrent deal %d/%d", i, opts.N) defer func() { // This is necessary because golang can't deal with test // failures being reported from children goroutines ¯\_(ツ)_/¯ if r := recover(); r != nil { err = fmt.Errorf("deal failed: %s", r) } }() dh.t.Logf("making storage deal %d/%d", i, opts.N) deal, res, inPath := dh.MakeOnlineDeal(context.Background(), MakeFullDealParams{ Rseed: 5 + i, FastRet: opts.FastRetrieval, StartEpoch: opts.StartEpoch, UseCARFileForStorageDeal: opts.UseCARFileForStorageDeal, }) // Check that the storage provider announced the deal to indexers if opts.IndexProvider != nil { notifs := opts.IndexProvider.GetNotifs() _, ok := notifs[string(deal.Bytes())] require.True(dh.t, ok) } dh.t.Logf("retrieving deal %d/%d", i, opts.N) outPath := dh.PerformRetrieval(context.Background(), deal, res.Root, opts.CarExport) if opts.CarExport { f, err := os.Open(outPath) require.NoError(dh.t, err) actualFile := dh.ExtractFileFromCAR(ctx, f) require.NoError(dh.t, f.Close()) AssertFilesEqual(dh.t, inPath, actualFile) } else { AssertFilesEqual(dh.t, inPath, outPath) } return nil }) } require.NoError(dh.t, errgrp.Wait()) }