package vm import ( "bytes" "context" "fmt" "io" "math" "os" "sort" "sync" "sync/atomic" "time" "github.com/ipfs/go-cid" cbor "github.com/ipfs/go-ipld-cbor" cbg "github.com/whyrusleeping/cbor-gen" "golang.org/x/xerrors" ffi "github.com/filecoin-project/filecoin-ffi" ffi_cgo "github.com/filecoin-project/filecoin-ffi/cgo" "github.com/filecoin-project/go-address" "github.com/filecoin-project/go-state-types/abi" actorstypes "github.com/filecoin-project/go-state-types/actors" "github.com/filecoin-project/go-state-types/exitcode" "github.com/filecoin-project/go-state-types/manifest" "github.com/filecoin-project/go-state-types/network" "github.com/filecoin-project/lotus/blockstore" "github.com/filecoin-project/lotus/build" "github.com/filecoin-project/lotus/chain/actors" "github.com/filecoin-project/lotus/chain/actors/adt" "github.com/filecoin-project/lotus/chain/actors/aerrors" "github.com/filecoin-project/lotus/chain/actors/builtin/miner" "github.com/filecoin-project/lotus/chain/actors/policy" "github.com/filecoin-project/lotus/chain/state" "github.com/filecoin-project/lotus/chain/types" "github.com/filecoin-project/lotus/lib/sigs" "github.com/filecoin-project/lotus/node/bundle" ) var _ Interface = (*FVM)(nil) var _ ffi_cgo.Externs = (*FvmExtern)(nil) type FvmExtern struct { Rand blockstore.Blockstore epoch abi.ChainEpoch lbState LookbackStateGetter tsGet TipSetGetter base cid.Cid } type FvmGasCharge struct { Name string TotalGas int64 ComputeGas int64 StorageGas int64 } // This may eventually become identical to ExecutionTrace, but we can make incremental progress towards that type FvmExecutionTrace struct { Msg *types.Message MsgRct *types.MessageReceipt Error string GasCharges []FvmGasCharge `cborgen:"maxlen=1000000000"` Subcalls []FvmExecutionTrace `cborgen:"maxlen=1000000000"` } func (t *FvmExecutionTrace) ToExecutionTrace() types.ExecutionTrace { if t == nil { return types.ExecutionTrace{} } ret := types.ExecutionTrace{ Msg: t.Msg, MsgRct: t.MsgRct, Error: t.Error, Subcalls: nil, // Should be nil when there are no subcalls for backwards compatibility } if len(t.GasCharges) > 0 { ret.GasCharges = make([]*types.GasTrace, len(t.GasCharges)) for i, v := range t.GasCharges { ret.GasCharges[i] = &types.GasTrace{ Name: v.Name, TotalGas: v.TotalGas, ComputeGas: v.ComputeGas, StorageGas: v.StorageGas, } } } if len(t.Subcalls) > 0 { ret.Subcalls = make([]types.ExecutionTrace, len(t.Subcalls)) for i, v := range t.Subcalls { ret.Subcalls[i] = v.ToExecutionTrace() } } return ret } func (x *FvmExtern) TipsetCid(ctx context.Context, epoch abi.ChainEpoch) (cid.Cid, error) { tsk, err := x.tsGet(ctx, epoch) if err != nil { return cid.Undef, err } return tsk.Cid() } // VerifyConsensusFault is similar to the one in syscalls.go used by the Lotus VM, except it never errors // Errors are logged and "no fault" is returned, which is functionally what go-actors does anyway func (x *FvmExtern) VerifyConsensusFault(ctx context.Context, a, b, extra []byte) (*ffi_cgo.ConsensusFault, int64) { totalGas := int64(0) ret := &ffi_cgo.ConsensusFault{ Type: ffi_cgo.ConsensusFaultNone, } // Note that block syntax is not validated. Any validly signed block will be accepted pursuant to the below conditions. // Whether or not it could ever have been accepted in a chain is not checked/does not matter here. // for that reason when checking block parent relationships, rather than instantiating a Tipset to do so // (which runs a syntactic check), we do it directly on the CIDs. // (0) cheap preliminary checks // can blocks be decoded properly? var blockA, blockB types.BlockHeader if decodeErr := blockA.UnmarshalCBOR(bytes.NewReader(a)); decodeErr != nil { log.Info("invalid consensus fault: cannot decode first block header: %w", decodeErr) return ret, totalGas } if decodeErr := blockB.UnmarshalCBOR(bytes.NewReader(b)); decodeErr != nil { log.Info("invalid consensus fault: cannot decode second block header: %w", decodeErr) return ret, totalGas } // are blocks the same? if blockA.Cid().Equals(blockB.Cid()) { log.Info("invalid consensus fault: submitted blocks are the same") return ret, totalGas } // (1) check conditions necessary to any consensus fault // were blocks mined by same miner? if blockA.Miner != blockB.Miner { log.Info("invalid consensus fault: blocks not mined by the same miner") return ret, totalGas } // block a must be earlier or equal to block b, epoch wise (ie at least as early in the chain). if blockB.Height < blockA.Height { log.Info("invalid consensus fault: first block must not be of higher height than second") return ret, totalGas } ret.Epoch = blockB.Height faultType := ffi_cgo.ConsensusFaultNone // (2) check for the consensus faults themselves // (a) double-fork mining fault if blockA.Height == blockB.Height { faultType = ffi_cgo.ConsensusFaultDoubleForkMining } // (b) time-offset mining fault // strictly speaking no need to compare heights based on double fork mining check above, // but at same height this would be a different fault. if types.CidArrsEqual(blockA.Parents, blockB.Parents) && blockA.Height != blockB.Height { faultType = ffi_cgo.ConsensusFaultTimeOffsetMining } // (c) parent-grinding fault // Here extra is the "witness", a third block that shows the connection between A and B as // A's sibling and B's parent. // Specifically, since A is of lower height, it must be that B was mined omitting A from its tipset // // B // | // [A, C] var blockC types.BlockHeader if len(extra) > 0 { if decodeErr := blockC.UnmarshalCBOR(bytes.NewReader(extra)); decodeErr != nil { log.Info("invalid consensus fault: cannot decode extra: %w", decodeErr) return ret, totalGas } if types.CidArrsEqual(blockA.Parents, blockC.Parents) && blockA.Height == blockC.Height && types.CidArrsContains(blockB.Parents, blockC.Cid()) && !types.CidArrsContains(blockB.Parents, blockA.Cid()) { faultType = ffi_cgo.ConsensusFaultParentGrinding } } // (3) return if no consensus fault by now if faultType == ffi_cgo.ConsensusFaultNone { log.Info("invalid consensus fault: no fault detected") return ret, totalGas } // else // (4) expensive final checks // check blocks are properly signed by their respective miner // note we do not need to check extra's: it is a parent to block b // which itself is signed, so it was willingly included by the miner gasA, sigErr := x.verifyBlockSig(ctx, &blockA) totalGas += gasA if sigErr != nil { log.Info("invalid consensus fault: cannot verify first block sig: %w", sigErr) return ret, totalGas } gas2, sigErr := x.verifyBlockSig(ctx, &blockB) totalGas += gas2 if sigErr != nil { log.Info("invalid consensus fault: cannot verify second block sig: %w", sigErr) return ret, totalGas } ret.Type = faultType ret.Target = blockA.Miner return ret, totalGas } func (x *FvmExtern) verifyBlockSig(ctx context.Context, blk *types.BlockHeader) (int64, error) { waddr, gasUsed, err := x.workerKeyAtLookback(ctx, blk.Miner, blk.Height) if err != nil { return gasUsed, err } return gasUsed, sigs.CheckBlockSignature(ctx, blk, waddr) } func (x *FvmExtern) workerKeyAtLookback(ctx context.Context, minerId address.Address, height abi.ChainEpoch) (address.Address, int64, error) { if height < x.epoch-policy.ChainFinality { return address.Undef, 0, xerrors.Errorf("cannot get worker key (currEpoch %d, height %d)", x.epoch, height) } gasUsed := int64(0) gasAdder := func(gc GasCharge) { // technically not overflow safe, but that's fine gasUsed += gc.Total() } cstWithoutGas := cbor.NewCborStore(x.Blockstore) cbb := &gasChargingBlocks{gasAdder, PricelistByEpoch(x.epoch), x.Blockstore} cstWithGas := cbor.NewCborStore(cbb) lbState, err := x.lbState(ctx, height) if err != nil { return address.Undef, gasUsed, err } // get appropriate miner actor act, err := lbState.GetActor(minerId) if err != nil { return address.Undef, gasUsed, err } // use that to get the miner state mas, err := miner.Load(adt.WrapStore(ctx, cstWithGas), act) if err != nil { return address.Undef, gasUsed, err } info, err := mas.Info() if err != nil { return address.Undef, gasUsed, err } stateTree, err := state.LoadStateTree(cstWithoutGas, x.base) if err != nil { return address.Undef, gasUsed, err } raddr, err := ResolveToDeterministicAddr(stateTree, cstWithGas, info.Worker) if err != nil { return address.Undef, gasUsed, err } return raddr, gasUsed, nil } type FVM struct { fvm *ffi.FVM nv network.Version } func defaultFVMOpts(ctx context.Context, opts *VMOpts) (*ffi.FVMOpts, error) { state, err := state.LoadStateTree(cbor.NewCborStore(opts.Bstore), opts.StateBase) if err != nil { return nil, xerrors.Errorf("loading state tree: %w", err) } circToReport, err := opts.CircSupplyCalc(ctx, opts.Epoch, state) if err != nil { return nil, xerrors.Errorf("calculating circ supply: %w", err) } return &ffi.FVMOpts{ FVMVersion: 0, Externs: &FvmExtern{ Rand: opts.Rand, Blockstore: opts.Bstore, lbState: opts.LookbackState, tsGet: opts.TipSetGetter, base: opts.StateBase, epoch: opts.Epoch, }, Epoch: opts.Epoch, Timestamp: opts.Timestamp, ChainID: build.Eip155ChainId, BaseFee: opts.BaseFee, BaseCircSupply: circToReport, NetworkVersion: opts.NetworkVersion, StateBase: opts.StateBase, Tracing: opts.Tracing || EnableDetailedTracing, Debug: build.ActorDebugging, }, nil } func NewFVM(ctx context.Context, opts *VMOpts) (*FVM, error) { fvmOpts, err := defaultFVMOpts(ctx, opts) if err != nil { return nil, xerrors.Errorf("creating fvm opts: %w", err) } fvm, err := ffi.CreateFVM(fvmOpts) if err != nil { return nil, xerrors.Errorf("failed to create FVM: %w", err) } return &FVM{ fvm: fvm, nv: opts.NetworkVersion, }, nil } func NewDebugFVM(ctx context.Context, opts *VMOpts) (*FVM, error) { baseBstore := opts.Bstore overlayBstore := blockstore.NewMemorySync() cborStore := cbor.NewCborStore(overlayBstore) vmBstore := blockstore.NewTieredBstore(overlayBstore, baseBstore) opts.Bstore = vmBstore fvmOpts, err := defaultFVMOpts(ctx, opts) if err != nil { return nil, xerrors.Errorf("creating fvm opts: %w", err) } fvmOpts.Debug = true putMapping := func(ar map[cid.Cid]cid.Cid) (cid.Cid, error) { var mapping xMapping mapping.redirects = make([]xRedirect, 0, len(ar)) for from, to := range ar { mapping.redirects = append(mapping.redirects, xRedirect{from: from, to: to}) } sort.Slice(mapping.redirects, func(i, j int) bool { return bytes.Compare(mapping.redirects[i].from.Bytes(), mapping.redirects[j].from.Bytes()) < 0 }) // Passing this as a pointer of structs has proven to be an enormous PiTA; hence this code. mappingCid, err := cborStore.Put(context.TODO(), &mapping) if err != nil { return cid.Undef, err } return mappingCid, nil } createMapping := func(debugBundlePath string) error { mfCid, err := bundle.LoadBundleFromFile(ctx, overlayBstore, debugBundlePath) if err != nil { return xerrors.Errorf("loading debug bundle: %w", err) } mf, err := actors.LoadManifest(ctx, mfCid, adt.WrapStore(ctx, cborStore)) if err != nil { return xerrors.Errorf("loading debug manifest: %w", err) } av, err := actorstypes.VersionForNetwork(opts.NetworkVersion) if err != nil { return xerrors.Errorf("getting actors version: %w", err) } // create actor redirect mapping actorRedirect := make(map[cid.Cid]cid.Cid) for _, key := range manifest.GetBuiltinActorsKeys(av) { from, ok := actors.GetActorCodeID(av, key) if !ok { log.Warnf("actor missing in the from manifest %s", key) continue } to, ok := mf.Get(key) if !ok { log.Warnf("actor missing in the to manifest %s", key) continue } actorRedirect[from] = to } if len(actorRedirect) > 0 { mappingCid, err := putMapping(actorRedirect) if err != nil { return xerrors.Errorf("error writing redirect mapping: %w", err) } fvmOpts.ActorRedirect = mappingCid } return nil } av, err := actorstypes.VersionForNetwork(opts.NetworkVersion) if err != nil { return nil, xerrors.Errorf("error determining actors version for network version %d: %w", opts.NetworkVersion, err) } debugBundlePath := os.Getenv(fmt.Sprintf("LOTUS_FVM_DEBUG_BUNDLE_V%d", av)) if debugBundlePath != "" { if err := createMapping(debugBundlePath); err != nil { log.Errorf("failed to create v%d debug mapping", av) } } fvm, err := ffi.CreateFVM(fvmOpts) if err != nil { return nil, err } return &FVM{ fvm: fvm, nv: opts.NetworkVersion, }, nil } func (vm *FVM) ApplyMessage(ctx context.Context, cmsg types.ChainMsg) (*ApplyRet, error) { start := build.Clock.Now() defer atomic.AddUint64(&StatApplied, 1) vmMsg := cmsg.VMMessage() msgBytes, err := vmMsg.Serialize() if err != nil { return nil, xerrors.Errorf("serializing msg: %w", err) } ret, err := vm.fvm.ApplyMessage(msgBytes, uint(cmsg.ChainLength())) if err != nil { return nil, xerrors.Errorf("applying msg: %w", err) } duration := time.Since(start) var receipt types.MessageReceipt if vm.nv >= network.Version18 { receipt = types.NewMessageReceiptV1(exitcode.ExitCode(ret.ExitCode), ret.Return, ret.GasUsed, ret.EventsRoot) } else { receipt = types.NewMessageReceiptV0(exitcode.ExitCode(ret.ExitCode), ret.Return, ret.GasUsed) } var aerr aerrors.ActorError if ret.ExitCode != 0 { amsg := ret.FailureInfo if amsg == "" { amsg = "unknown error" } aerr = aerrors.New(exitcode.ExitCode(ret.ExitCode), amsg) } var et types.ExecutionTrace if len(ret.ExecTraceBytes) != 0 { var fvmEt FvmExecutionTrace if err = fvmEt.UnmarshalCBOR(bytes.NewReader(ret.ExecTraceBytes)); err != nil { return nil, xerrors.Errorf("failed to unmarshal exectrace: %w", err) } et = fvmEt.ToExecutionTrace() } // Set the top-level exectrace info from the message and receipt for backwards compatibility et.Msg = vmMsg et.MsgRct = &receipt et.Duration = duration if aerr != nil { et.Error = aerr.Error() } return &ApplyRet{ MessageReceipt: receipt, GasCosts: &GasOutputs{ BaseFeeBurn: ret.BaseFeeBurn, OverEstimationBurn: ret.OverEstimationBurn, MinerPenalty: ret.MinerPenalty, MinerTip: ret.MinerTip, Refund: ret.Refund, GasRefund: ret.GasRefund, GasBurned: ret.GasBurned, }, ActorErr: aerr, ExecutionTrace: et, Duration: duration, }, nil } func (vm *FVM) ApplyImplicitMessage(ctx context.Context, cmsg *types.Message) (*ApplyRet, error) { start := build.Clock.Now() defer atomic.AddUint64(&StatApplied, 1) cmsg.GasLimit = math.MaxInt64 / 2 vmMsg := cmsg.VMMessage() msgBytes, err := vmMsg.Serialize() if err != nil { return nil, xerrors.Errorf("serializing msg: %w", err) } ret, err := vm.fvm.ApplyImplicitMessage(msgBytes) if err != nil { return nil, xerrors.Errorf("applying msg: %w", err) } duration := time.Since(start) var receipt types.MessageReceipt if vm.nv >= network.Version18 { receipt = types.NewMessageReceiptV1(exitcode.ExitCode(ret.ExitCode), ret.Return, ret.GasUsed, ret.EventsRoot) } else { receipt = types.NewMessageReceiptV0(exitcode.ExitCode(ret.ExitCode), ret.Return, ret.GasUsed) } var aerr aerrors.ActorError if ret.ExitCode != 0 { amsg := ret.FailureInfo if amsg == "" { amsg = "unknown error" } aerr = aerrors.New(exitcode.ExitCode(ret.ExitCode), amsg) } var et types.ExecutionTrace if len(ret.ExecTraceBytes) != 0 { var fvmEt FvmExecutionTrace if err = fvmEt.UnmarshalCBOR(bytes.NewReader(ret.ExecTraceBytes)); err != nil { return nil, xerrors.Errorf("failed to unmarshal exectrace: %w", err) } et = fvmEt.ToExecutionTrace() } else { et.Msg = vmMsg et.MsgRct = &receipt et.Duration = duration if aerr != nil { et.Error = aerr.Error() } } applyRet := &ApplyRet{ MessageReceipt: receipt, ActorErr: aerr, ExecutionTrace: et, Duration: duration, } if ret.ExitCode != 0 { return applyRet, fmt.Errorf("implicit message failed with exit code: %d and error: %w", ret.ExitCode, applyRet.ActorErr) } return applyRet, nil } func (vm *FVM) Flush(ctx context.Context) (cid.Cid, error) { return vm.fvm.Flush() } type dualExecutionFVM struct { main *FVM debug *FVM } var _ Interface = (*dualExecutionFVM)(nil) func NewDualExecutionFVM(ctx context.Context, opts *VMOpts) (Interface, error) { main, err := NewFVM(ctx, opts) if err != nil { return nil, err } debug, err := NewDebugFVM(ctx, opts) if err != nil { return nil, err } return &dualExecutionFVM{ main: main, debug: debug, }, nil } func (vm *dualExecutionFVM) ApplyMessage(ctx context.Context, cmsg types.ChainMsg) (ret *ApplyRet, err error) { var wg sync.WaitGroup wg.Add(2) go func() { defer wg.Done() ret, err = vm.main.ApplyMessage(ctx, cmsg) }() go func() { defer wg.Done() if _, err := vm.debug.ApplyMessage(ctx, cmsg); err != nil { log.Errorf("debug execution failed: %w", err) } }() wg.Wait() return ret, err } func (vm *dualExecutionFVM) ApplyImplicitMessage(ctx context.Context, msg *types.Message) (ret *ApplyRet, err error) { var wg sync.WaitGroup wg.Add(2) go func() { defer wg.Done() ret, err = vm.main.ApplyImplicitMessage(ctx, msg) }() go func() { defer wg.Done() if _, err := vm.debug.ApplyImplicitMessage(ctx, msg); err != nil { log.Errorf("debug execution failed: %s", err) } }() wg.Wait() return ret, err } func (vm *dualExecutionFVM) Flush(ctx context.Context) (cid.Cid, error) { return vm.main.Flush(ctx) } // Passing this as a pointer of structs has proven to be an enormous PiTA; hence this code. type xRedirect struct{ from, to cid.Cid } type xMapping struct{ redirects []xRedirect } func (m *xMapping) MarshalCBOR(w io.Writer) error { scratch := make([]byte, 9) if err := cbg.WriteMajorTypeHeaderBuf(scratch, w, cbg.MajArray, uint64(len(m.redirects))); err != nil { return err } for _, v := range m.redirects { if err := v.MarshalCBOR(w); err != nil { return err } } return nil } func (r *xRedirect) MarshalCBOR(w io.Writer) error { scratch := make([]byte, 9) if err := cbg.WriteMajorTypeHeaderBuf(scratch, w, cbg.MajArray, uint64(2)); err != nil { return err } if err := cbg.WriteCidBuf(scratch, w, r.from); err != nil { return xerrors.Errorf("failed to write cid field from: %w", err) } if err := cbg.WriteCidBuf(scratch, w, r.to); err != nil { return xerrors.Errorf("failed to write cid field from: %w", err) } return nil }