// stm: #integration package itests import ( "context" "strings" "testing" "time" logging "github.com/ipfs/go-log/v2" "github.com/libp2p/go-libp2p/core/peer" "github.com/stretchr/testify/require" "golang.org/x/xerrors" "github.com/filecoin-project/go-address" "github.com/filecoin-project/go-state-types/big" "github.com/filecoin-project/go-state-types/exitcode" lapi "github.com/filecoin-project/lotus/api" "github.com/filecoin-project/lotus/build" "github.com/filecoin-project/lotus/chain/types" "github.com/filecoin-project/lotus/itests/kit" ) func TestAPI(t *testing.T) { //stm: @CHAIN_SYNCER_LOAD_GENESIS_001, @CHAIN_SYNCER_FETCH_TIPSET_001, //stm: @CHAIN_SYNCER_START_001, @CHAIN_SYNCER_SYNC_001, @BLOCKCHAIN_BEACON_VALIDATE_BLOCK_VALUES_01 //stm: @CHAIN_SYNCER_COLLECT_CHAIN_001, @CHAIN_SYNCER_COLLECT_HEADERS_001, @CHAIN_SYNCER_VALIDATE_TIPSET_001 //stm: @CHAIN_SYNCER_NEW_PEER_HEAD_001, @CHAIN_SYNCER_VALIDATE_MESSAGE_META_001, @CHAIN_SYNCER_STOP_001 //stm: @CHAIN_STATE_MINER_INFO_001 t.Run("direct", func(t *testing.T) { runAPITest(t) }) t.Run("rpc", func(t *testing.T) { runAPITest(t, kit.ThroughRPC()) }) } type apiSuite struct { opts []interface{} } // runAPITest is the entry point to API test suite func runAPITest(t *testing.T, opts ...interface{}) { ts := apiSuite{opts: opts} t.Run("version", ts.testVersion) t.Run("id", ts.testID) t.Run("testConnectTwo", ts.testConnectTwo) t.Run("testMining", ts.testMining) t.Run("testMiningReal", ts.testMiningReal) t.Run("testSlowNotify", ts.testSlowNotify) t.Run("testSearchMsg", ts.testSearchMsg) t.Run("testOutOfGasError", ts.testOutOfGasError) t.Run("testLookupNotFoundError", ts.testLookupNotFoundError) t.Run("testNonGenesisMiner", ts.testNonGenesisMiner) } func (ts *apiSuite) testVersion(t *testing.T) { lapi.RunningNodeType = lapi.NodeFull t.Cleanup(func() { lapi.RunningNodeType = lapi.NodeUnknown }) full, _, _ := kit.EnsembleMinimal(t, ts.opts...) v, err := full.Version(context.Background()) require.NoError(t, err) versions := strings.Split(v.Version, "+") require.NotZero(t, len(versions), "empty version") require.Equal(t, versions[0], build.BuildVersion) } func (ts *apiSuite) testID(t *testing.T) { ctx := context.Background() full, _, _ := kit.EnsembleMinimal(t, ts.opts...) id, err := full.ID(ctx) if err != nil { t.Fatal(err) } require.Regexp(t, "^12", id.Pretty()) } func (ts *apiSuite) testConnectTwo(t *testing.T) { ctx := context.Background() one, two, _, ens := kit.EnsembleTwoOne(t, ts.opts...) p, err := one.NetPeers(ctx) require.NoError(t, err) require.Empty(t, p, "node one has peers") p, err = two.NetPeers(ctx) require.NoError(t, err) require.Empty(t, p, "node two has peers") ens.InterconnectAll() peers, err := one.NetPeers(ctx) require.NoError(t, err) countPeerIDs := func(peers []peer.AddrInfo) int { peerIDs := make(map[peer.ID]struct{}) for _, p := range peers { peerIDs[p.ID] = struct{}{} } return len(peerIDs) } require.Equal(t, countPeerIDs(peers), 2, "node one doesn't have 2 peers") peers, err = two.NetPeers(ctx) require.NoError(t, err) require.Equal(t, countPeerIDs(peers), 2, "node one doesn't have 2 peers") } func (ts *apiSuite) testSearchMsg(t *testing.T) { ctx := context.Background() full, _, ens := kit.EnsembleMinimal(t, ts.opts...) senderAddr, err := full.WalletDefaultAddress(ctx) require.NoError(t, err) msg := &types.Message{ From: senderAddr, To: senderAddr, Value: big.Zero(), } ens.BeginMining(100 * time.Millisecond) sm, err := full.MpoolPushMessage(ctx, msg, nil) require.NoError(t, err) //stm: @CHAIN_STATE_WAIT_MSG_001 res, err := full.StateWaitMsg(ctx, sm.Cid(), 1, lapi.LookbackNoLimit, true) require.NoError(t, err) require.Equal(t, exitcode.Ok, res.Receipt.ExitCode, "message not successful") //stm: @CHAIN_STATE_SEARCH_MSG_001 searchRes, err := full.StateSearchMsg(ctx, types.EmptyTSK, sm.Cid(), lapi.LookbackNoLimit, true) require.NoError(t, err) require.NotNil(t, searchRes) require.Equalf(t, res.TipSet, searchRes.TipSet, "search ts: %s, different from wait ts: %s", searchRes.TipSet, res.TipSet) } func (ts *apiSuite) testOutOfGasError(t *testing.T) { ctx := context.Background() full, _, _ := kit.EnsembleMinimal(t, ts.opts...) senderAddr, err := full.WalletDefaultAddress(ctx) require.NoError(t, err) // the gas estimator API executes the message with gasLimit = BlockGasLimit // Lowering it to 2 will cause it to run out of gas, testing the failure case we want originalLimit := build.BlockGasLimit build.BlockGasLimit = 2 defer func() { build.BlockGasLimit = originalLimit }() msg := &types.Message{ From: senderAddr, To: senderAddr, Value: big.Zero(), } _, err = full.GasEstimateMessageGas(ctx, msg, nil, types.EmptyTSK) require.Error(t, err, "should have failed") require.True(t, xerrors.Is(err, &lapi.ErrOutOfGas{})) } func (ts *apiSuite) testLookupNotFoundError(t *testing.T) { ctx := context.Background() full, _, _ := kit.EnsembleMinimal(t, ts.opts...) addr, err := full.WalletNew(ctx, types.KTSecp256k1) require.NoError(t, err) _, err = full.StateLookupID(ctx, addr, types.EmptyTSK) require.Error(t, err) require.True(t, xerrors.Is(err, &lapi.ErrActorNotFound{})) } func (ts *apiSuite) testMining(t *testing.T) { ctx := context.Background() full, miner, _ := kit.EnsembleMinimal(t, ts.opts...) newHeads, err := full.ChainNotify(ctx) require.NoError(t, err) initHead := (<-newHeads)[0] baseHeight := initHead.Val.Height() h1, err := full.ChainHead(ctx) require.NoError(t, err) require.Equal(t, int64(h1.Height()), int64(baseHeight)) bm := kit.NewBlockMiner(t, miner) bm.MineUntilBlock(ctx, full, nil) require.NoError(t, err) <-newHeads h2, err := full.ChainHead(ctx) require.NoError(t, err) require.Greater(t, int64(h2.Height()), int64(h1.Height())) bm.MineUntilBlock(ctx, full, nil) require.NoError(t, err) <-newHeads h3, err := full.ChainHead(ctx) require.NoError(t, err) require.Greater(t, int64(h3.Height()), int64(h2.Height())) } func (ts *apiSuite) testMiningReal(t *testing.T) { build.InsecurePoStValidation = false defer func() { build.InsecurePoStValidation = true }() ts.testMining(t) } func (ts *apiSuite) testSlowNotify(t *testing.T) { _ = logging.SetLogLevel("rpc", "ERROR") ctx, cancel := context.WithCancel(context.Background()) defer cancel() full, miner, _ := kit.EnsembleMinimal(t, ts.opts...) // Subscribe a bunch of times to make sure we fill up any RPC buffers. var newHeadsChans []<-chan []*lapi.HeadChange for i := 0; i < 100; i++ { newHeads, err := full.ChainNotify(ctx) require.NoError(t, err) newHeadsChans = append(newHeadsChans, newHeads) } initHead := (<-newHeadsChans[0])[0] baseHeight := initHead.Val.Height() bm := kit.NewBlockMiner(t, miner) bm.MineBlocks(ctx, time.Microsecond) full.WaitTillChain(ctx, kit.HeightAtLeast(baseHeight+100)) // Make sure they were all closed, draining any buffered events first. for _, ch := range newHeadsChans { var ok bool for ok { select { case _, ok = <-ch: default: t.Fatal("expected new heads channel to be closed") } } } // Make sure we can resubscribe and everything still works. newHeads, err := full.ChainNotify(ctx) require.NoError(t, err) for i := 0; i < 10; i++ { _, ok := <-newHeads require.True(t, ok, "notify channel closed") } } func (ts *apiSuite) testNonGenesisMiner(t *testing.T) { ctx := context.Background() full, genesisMiner, ens := kit.EnsembleMinimal(t, append(ts.opts, kit.MockProofs())...) ens.InterconnectAll().BeginMining(4 * time.Millisecond) time.Sleep(1 * time.Second) gaa, err := genesisMiner.ActorAddress(ctx) require.NoError(t, err) _, err = full.StateMinerInfo(ctx, gaa, types.EmptyTSK) require.NoError(t, err) var newMiner kit.TestMiner ens.Miner(&newMiner, full, kit.OwnerAddr(full.DefaultKey), kit.SectorSize(2<<10), kit.WithAllSubsystems(), ).Start().InterconnectAll() ta, err := newMiner.ActorAddress(ctx) require.NoError(t, err) tid, err := address.IDFromAddress(ta) require.NoError(t, err) require.Equal(t, uint64(1001), tid) }