2a6beb6a39
This PR removes the newly added txpool.Transaction wrapper type, and instead adds a way of keeping the blob sidecar within types.Transaction. It's better this way because most code in go-ethereum does not care about blob transactions, and probably never will. This will start mattering especially on the client side of RPC, where all APIs are based on types.Transaction. Users need to be able to use the same signing flows they already have. However, since blobs are only allowed in some places but not others, we will now need to add checks to avoid creating invalid blocks. I'm still trying to figure out the best place to do some of these. The way I have it currently is as follows: - In block validation (import), txs are verified not to have a blob sidecar. - In miner, we strip off the sidecar when committing the transaction into the block. - In TxPool validation, txs must have a sidecar to be added into the blobpool. - Note there is a special case here: when transactions are re-added because of a chain reorg, we cannot use the transactions gathered from the old chain blocks as-is, because they will be missing their blobs. This was previously handled by storing the blobs into the 'blobpool limbo'. The code has now changed to store the full transaction in the limbo instead, but it might be confusing for code readers why we're not simply adding the types.Transaction we already have. Code changes summary: - txpool.Transaction removed and all uses replaced by types.Transaction again - blobpool now stores types.Transaction instead of defining its own blobTx format for storage - the blobpool limbo now stores types.Transaction instead of storing only the blobs - checks to validate the presence/absence of the blob sidecar added in certain critical places
566 lines
19 KiB
Go
566 lines
19 KiB
Go
// Copyright 2021 The go-ethereum Authors
|
|
// This file is part of the go-ethereum library.
|
|
//
|
|
// The go-ethereum library is free software: you can redistribute it and/or modify
|
|
// it under the terms of the GNU Lesser General Public License as published by
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
// (at your option) any later version.
|
|
//
|
|
// The go-ethereum library is distributed in the hope that it will be useful,
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
// GNU Lesser General Public License for more details.
|
|
//
|
|
// You should have received a copy of the GNU Lesser General Public License
|
|
// along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
package les
|
|
|
|
import (
|
|
"encoding/binary"
|
|
"encoding/json"
|
|
|
|
"github.com/ethereum/go-ethereum/common"
|
|
"github.com/ethereum/go-ethereum/core"
|
|
"github.com/ethereum/go-ethereum/core/state"
|
|
"github.com/ethereum/go-ethereum/core/txpool"
|
|
"github.com/ethereum/go-ethereum/core/types"
|
|
"github.com/ethereum/go-ethereum/light"
|
|
"github.com/ethereum/go-ethereum/log"
|
|
"github.com/ethereum/go-ethereum/metrics"
|
|
"github.com/ethereum/go-ethereum/rlp"
|
|
"github.com/ethereum/go-ethereum/trie"
|
|
)
|
|
|
|
// serverBackend defines the backend functions needed for serving LES requests
|
|
type serverBackend interface {
|
|
ArchiveMode() bool
|
|
AddTxsSync() bool
|
|
BlockChain() *core.BlockChain
|
|
TxPool() *txpool.TxPool
|
|
GetHelperTrie(typ uint, index uint64) *trie.Trie
|
|
}
|
|
|
|
// Decoder is implemented by the messages passed to the handler functions
|
|
type Decoder interface {
|
|
Decode(val interface{}) error
|
|
}
|
|
|
|
// RequestType is a static struct that describes an LES request type and references
|
|
// its handler function.
|
|
type RequestType struct {
|
|
Name string
|
|
MaxCount uint64
|
|
InPacketsMeter, InTrafficMeter, OutPacketsMeter, OutTrafficMeter metrics.Meter
|
|
ServingTimeMeter metrics.Timer
|
|
Handle func(msg Decoder) (serve serveRequestFn, reqID, amount uint64, err error)
|
|
}
|
|
|
|
// serveRequestFn is returned by the request handler functions after decoding the request.
|
|
// This function does the actual request serving using the supplied backend. waitOrStop is
|
|
// called between serving individual request items and may block if the serving process
|
|
// needs to be throttled. If it returns false then the process is terminated.
|
|
// The reply is not sent by this function yet. The flow control feedback value is supplied
|
|
// by the protocol handler when calling the send function of the returned reply struct.
|
|
type serveRequestFn func(backend serverBackend, peer *clientPeer, waitOrStop func() bool) *reply
|
|
|
|
// Les3 contains the request types supported by les/2 and les/3
|
|
var Les3 = map[uint64]RequestType{
|
|
GetBlockHeadersMsg: {
|
|
Name: "block header request",
|
|
MaxCount: MaxHeaderFetch,
|
|
InPacketsMeter: miscInHeaderPacketsMeter,
|
|
InTrafficMeter: miscInHeaderTrafficMeter,
|
|
OutPacketsMeter: miscOutHeaderPacketsMeter,
|
|
OutTrafficMeter: miscOutHeaderTrafficMeter,
|
|
ServingTimeMeter: miscServingTimeHeaderTimer,
|
|
Handle: handleGetBlockHeaders,
|
|
},
|
|
GetBlockBodiesMsg: {
|
|
Name: "block bodies request",
|
|
MaxCount: MaxBodyFetch,
|
|
InPacketsMeter: miscInBodyPacketsMeter,
|
|
InTrafficMeter: miscInBodyTrafficMeter,
|
|
OutPacketsMeter: miscOutBodyPacketsMeter,
|
|
OutTrafficMeter: miscOutBodyTrafficMeter,
|
|
ServingTimeMeter: miscServingTimeBodyTimer,
|
|
Handle: handleGetBlockBodies,
|
|
},
|
|
GetCodeMsg: {
|
|
Name: "code request",
|
|
MaxCount: MaxCodeFetch,
|
|
InPacketsMeter: miscInCodePacketsMeter,
|
|
InTrafficMeter: miscInCodeTrafficMeter,
|
|
OutPacketsMeter: miscOutCodePacketsMeter,
|
|
OutTrafficMeter: miscOutCodeTrafficMeter,
|
|
ServingTimeMeter: miscServingTimeCodeTimer,
|
|
Handle: handleGetCode,
|
|
},
|
|
GetReceiptsMsg: {
|
|
Name: "receipts request",
|
|
MaxCount: MaxReceiptFetch,
|
|
InPacketsMeter: miscInReceiptPacketsMeter,
|
|
InTrafficMeter: miscInReceiptTrafficMeter,
|
|
OutPacketsMeter: miscOutReceiptPacketsMeter,
|
|
OutTrafficMeter: miscOutReceiptTrafficMeter,
|
|
ServingTimeMeter: miscServingTimeReceiptTimer,
|
|
Handle: handleGetReceipts,
|
|
},
|
|
GetProofsV2Msg: {
|
|
Name: "les/2 proofs request",
|
|
MaxCount: MaxProofsFetch,
|
|
InPacketsMeter: miscInTrieProofPacketsMeter,
|
|
InTrafficMeter: miscInTrieProofTrafficMeter,
|
|
OutPacketsMeter: miscOutTrieProofPacketsMeter,
|
|
OutTrafficMeter: miscOutTrieProofTrafficMeter,
|
|
ServingTimeMeter: miscServingTimeTrieProofTimer,
|
|
Handle: handleGetProofs,
|
|
},
|
|
GetHelperTrieProofsMsg: {
|
|
Name: "helper trie proof request",
|
|
MaxCount: MaxHelperTrieProofsFetch,
|
|
InPacketsMeter: miscInHelperTriePacketsMeter,
|
|
InTrafficMeter: miscInHelperTrieTrafficMeter,
|
|
OutPacketsMeter: miscOutHelperTriePacketsMeter,
|
|
OutTrafficMeter: miscOutHelperTrieTrafficMeter,
|
|
ServingTimeMeter: miscServingTimeHelperTrieTimer,
|
|
Handle: handleGetHelperTrieProofs,
|
|
},
|
|
SendTxV2Msg: {
|
|
Name: "new transactions",
|
|
MaxCount: MaxTxSend,
|
|
InPacketsMeter: miscInTxsPacketsMeter,
|
|
InTrafficMeter: miscInTxsTrafficMeter,
|
|
OutPacketsMeter: miscOutTxsPacketsMeter,
|
|
OutTrafficMeter: miscOutTxsTrafficMeter,
|
|
ServingTimeMeter: miscServingTimeTxTimer,
|
|
Handle: handleSendTx,
|
|
},
|
|
GetTxStatusMsg: {
|
|
Name: "transaction status query request",
|
|
MaxCount: MaxTxStatus,
|
|
InPacketsMeter: miscInTxStatusPacketsMeter,
|
|
InTrafficMeter: miscInTxStatusTrafficMeter,
|
|
OutPacketsMeter: miscOutTxStatusPacketsMeter,
|
|
OutTrafficMeter: miscOutTxStatusTrafficMeter,
|
|
ServingTimeMeter: miscServingTimeTxStatusTimer,
|
|
Handle: handleGetTxStatus,
|
|
},
|
|
}
|
|
|
|
// handleGetBlockHeaders handles a block header request
|
|
func handleGetBlockHeaders(msg Decoder) (serveRequestFn, uint64, uint64, error) {
|
|
var r GetBlockHeadersPacket
|
|
if err := msg.Decode(&r); err != nil {
|
|
return nil, 0, 0, err
|
|
}
|
|
return func(backend serverBackend, p *clientPeer, waitOrStop func() bool) *reply {
|
|
// Gather headers until the fetch or network limits is reached
|
|
var (
|
|
bc = backend.BlockChain()
|
|
hashMode = r.Query.Origin.Hash != (common.Hash{})
|
|
first = true
|
|
maxNonCanonical = uint64(100)
|
|
bytes common.StorageSize
|
|
headers []*types.Header
|
|
unknown bool
|
|
)
|
|
for !unknown && len(headers) < int(r.Query.Amount) && bytes < softResponseLimit {
|
|
if !first && !waitOrStop() {
|
|
return nil
|
|
}
|
|
// Retrieve the next header satisfying the r
|
|
var origin *types.Header
|
|
if hashMode {
|
|
if first {
|
|
origin = bc.GetHeaderByHash(r.Query.Origin.Hash)
|
|
if origin != nil {
|
|
r.Query.Origin.Number = origin.Number.Uint64()
|
|
}
|
|
} else {
|
|
origin = bc.GetHeader(r.Query.Origin.Hash, r.Query.Origin.Number)
|
|
}
|
|
} else {
|
|
origin = bc.GetHeaderByNumber(r.Query.Origin.Number)
|
|
}
|
|
if origin == nil {
|
|
break
|
|
}
|
|
headers = append(headers, origin)
|
|
bytes += estHeaderRlpSize
|
|
|
|
// Advance to the next header of the r
|
|
switch {
|
|
case hashMode && r.Query.Reverse:
|
|
// Hash based traversal towards the genesis block
|
|
ancestor := r.Query.Skip + 1
|
|
if ancestor == 0 {
|
|
unknown = true
|
|
} else {
|
|
r.Query.Origin.Hash, r.Query.Origin.Number = bc.GetAncestor(r.Query.Origin.Hash, r.Query.Origin.Number, ancestor, &maxNonCanonical)
|
|
unknown = r.Query.Origin.Hash == common.Hash{}
|
|
}
|
|
case hashMode && !r.Query.Reverse:
|
|
// Hash based traversal towards the leaf block
|
|
var (
|
|
current = origin.Number.Uint64()
|
|
next = current + r.Query.Skip + 1
|
|
)
|
|
if next <= current {
|
|
infos, _ := json.Marshal(p.Peer.Info())
|
|
p.Log().Warn("GetBlockHeaders skip overflow attack", "current", current, "skip", r.Query.Skip, "next", next, "attacker", string(infos))
|
|
unknown = true
|
|
} else {
|
|
if header := bc.GetHeaderByNumber(next); header != nil {
|
|
nextHash := header.Hash()
|
|
expOldHash, _ := bc.GetAncestor(nextHash, next, r.Query.Skip+1, &maxNonCanonical)
|
|
if expOldHash == r.Query.Origin.Hash {
|
|
r.Query.Origin.Hash, r.Query.Origin.Number = nextHash, next
|
|
} else {
|
|
unknown = true
|
|
}
|
|
} else {
|
|
unknown = true
|
|
}
|
|
}
|
|
case r.Query.Reverse:
|
|
// Number based traversal towards the genesis block
|
|
if r.Query.Origin.Number >= r.Query.Skip+1 {
|
|
r.Query.Origin.Number -= r.Query.Skip + 1
|
|
} else {
|
|
unknown = true
|
|
}
|
|
|
|
case !r.Query.Reverse:
|
|
// Number based traversal towards the leaf block
|
|
r.Query.Origin.Number += r.Query.Skip + 1
|
|
}
|
|
first = false
|
|
}
|
|
return p.replyBlockHeaders(r.ReqID, headers)
|
|
}, r.ReqID, r.Query.Amount, nil
|
|
}
|
|
|
|
// handleGetBlockBodies handles a block body request
|
|
func handleGetBlockBodies(msg Decoder) (serveRequestFn, uint64, uint64, error) {
|
|
var r GetBlockBodiesPacket
|
|
if err := msg.Decode(&r); err != nil {
|
|
return nil, 0, 0, err
|
|
}
|
|
return func(backend serverBackend, p *clientPeer, waitOrStop func() bool) *reply {
|
|
var (
|
|
bytes int
|
|
bodies []rlp.RawValue
|
|
)
|
|
bc := backend.BlockChain()
|
|
for i, hash := range r.Hashes {
|
|
if i != 0 && !waitOrStop() {
|
|
return nil
|
|
}
|
|
if bytes >= softResponseLimit {
|
|
break
|
|
}
|
|
body := bc.GetBodyRLP(hash)
|
|
if body == nil {
|
|
p.bumpInvalid()
|
|
continue
|
|
}
|
|
bodies = append(bodies, body)
|
|
bytes += len(body)
|
|
}
|
|
return p.replyBlockBodiesRLP(r.ReqID, bodies)
|
|
}, r.ReqID, uint64(len(r.Hashes)), nil
|
|
}
|
|
|
|
// handleGetCode handles a contract code request
|
|
func handleGetCode(msg Decoder) (serveRequestFn, uint64, uint64, error) {
|
|
var r GetCodePacket
|
|
if err := msg.Decode(&r); err != nil {
|
|
return nil, 0, 0, err
|
|
}
|
|
return func(backend serverBackend, p *clientPeer, waitOrStop func() bool) *reply {
|
|
var (
|
|
bytes int
|
|
data [][]byte
|
|
)
|
|
bc := backend.BlockChain()
|
|
for i, request := range r.Reqs {
|
|
if i != 0 && !waitOrStop() {
|
|
return nil
|
|
}
|
|
// Look up the root hash belonging to the request
|
|
header := bc.GetHeaderByHash(request.BHash)
|
|
if header == nil {
|
|
p.Log().Warn("Failed to retrieve associate header for code", "hash", request.BHash)
|
|
p.bumpInvalid()
|
|
continue
|
|
}
|
|
// Refuse to search stale state data in the database since looking for
|
|
// a non-exist key is kind of expensive.
|
|
local := bc.CurrentHeader().Number.Uint64()
|
|
if !backend.ArchiveMode() && header.Number.Uint64()+core.TriesInMemory <= local {
|
|
p.Log().Debug("Reject stale code request", "number", header.Number.Uint64(), "head", local)
|
|
p.bumpInvalid()
|
|
continue
|
|
}
|
|
address := common.BytesToAddress(request.AccountAddress)
|
|
account, err := getAccount(bc.TrieDB(), header.Root, address)
|
|
if err != nil {
|
|
p.Log().Warn("Failed to retrieve account for code", "block", header.Number, "hash", header.Hash(), "account", address, "err", err)
|
|
p.bumpInvalid()
|
|
continue
|
|
}
|
|
code, err := bc.StateCache().ContractCode(address, common.BytesToHash(account.CodeHash))
|
|
if err != nil {
|
|
p.Log().Warn("Failed to retrieve account code", "block", header.Number, "hash", header.Hash(), "account", address, "codehash", common.BytesToHash(account.CodeHash), "err", err)
|
|
continue
|
|
}
|
|
// Accumulate the code and abort if enough data was retrieved
|
|
data = append(data, code)
|
|
if bytes += len(code); bytes >= softResponseLimit {
|
|
break
|
|
}
|
|
}
|
|
return p.replyCode(r.ReqID, data)
|
|
}, r.ReqID, uint64(len(r.Reqs)), nil
|
|
}
|
|
|
|
// handleGetReceipts handles a block receipts request
|
|
func handleGetReceipts(msg Decoder) (serveRequestFn, uint64, uint64, error) {
|
|
var r GetReceiptsPacket
|
|
if err := msg.Decode(&r); err != nil {
|
|
return nil, 0, 0, err
|
|
}
|
|
return func(backend serverBackend, p *clientPeer, waitOrStop func() bool) *reply {
|
|
var (
|
|
bytes int
|
|
receipts []rlp.RawValue
|
|
)
|
|
bc := backend.BlockChain()
|
|
for i, hash := range r.Hashes {
|
|
if i != 0 && !waitOrStop() {
|
|
return nil
|
|
}
|
|
if bytes >= softResponseLimit {
|
|
break
|
|
}
|
|
// Retrieve the requested block's receipts, skipping if unknown to us
|
|
results := bc.GetReceiptsByHash(hash)
|
|
if results == nil {
|
|
if header := bc.GetHeaderByHash(hash); header == nil || header.ReceiptHash != types.EmptyReceiptsHash {
|
|
p.bumpInvalid()
|
|
continue
|
|
}
|
|
}
|
|
// If known, encode and queue for response packet
|
|
if encoded, err := rlp.EncodeToBytes(results); err != nil {
|
|
log.Error("Failed to encode receipt", "err", err)
|
|
} else {
|
|
receipts = append(receipts, encoded)
|
|
bytes += len(encoded)
|
|
}
|
|
}
|
|
return p.replyReceiptsRLP(r.ReqID, receipts)
|
|
}, r.ReqID, uint64(len(r.Hashes)), nil
|
|
}
|
|
|
|
// handleGetProofs handles a proof request
|
|
func handleGetProofs(msg Decoder) (serveRequestFn, uint64, uint64, error) {
|
|
var r GetProofsPacket
|
|
if err := msg.Decode(&r); err != nil {
|
|
return nil, 0, 0, err
|
|
}
|
|
return func(backend serverBackend, p *clientPeer, waitOrStop func() bool) *reply {
|
|
var (
|
|
lastBHash common.Hash
|
|
root common.Hash
|
|
header *types.Header
|
|
err error
|
|
)
|
|
bc := backend.BlockChain()
|
|
nodes := light.NewNodeSet()
|
|
|
|
for i, request := range r.Reqs {
|
|
if i != 0 && !waitOrStop() {
|
|
return nil
|
|
}
|
|
// Look up the root hash belonging to the request
|
|
if request.BHash != lastBHash {
|
|
root, lastBHash = common.Hash{}, request.BHash
|
|
|
|
if header = bc.GetHeaderByHash(request.BHash); header == nil {
|
|
p.Log().Warn("Failed to retrieve header for proof", "hash", request.BHash)
|
|
p.bumpInvalid()
|
|
continue
|
|
}
|
|
// Refuse to search stale state data in the database since looking for
|
|
// a non-exist key is kind of expensive.
|
|
local := bc.CurrentHeader().Number.Uint64()
|
|
if !backend.ArchiveMode() && header.Number.Uint64()+core.TriesInMemory <= local {
|
|
p.Log().Debug("Reject stale trie request", "number", header.Number.Uint64(), "head", local)
|
|
p.bumpInvalid()
|
|
continue
|
|
}
|
|
root = header.Root
|
|
}
|
|
// If a header lookup failed (non existent), ignore subsequent requests for the same header
|
|
if root == (common.Hash{}) {
|
|
p.bumpInvalid()
|
|
continue
|
|
}
|
|
// Open the account or storage trie for the request
|
|
statedb := bc.StateCache()
|
|
|
|
var trie state.Trie
|
|
switch len(request.AccountAddress) {
|
|
case 0:
|
|
// No account key specified, open an account trie
|
|
trie, err = statedb.OpenTrie(root)
|
|
if trie == nil || err != nil {
|
|
p.Log().Warn("Failed to open storage trie for proof", "block", header.Number, "hash", header.Hash(), "root", root, "err", err)
|
|
continue
|
|
}
|
|
default:
|
|
// Account key specified, open a storage trie
|
|
address := common.BytesToAddress(request.AccountAddress)
|
|
account, err := getAccount(bc.TrieDB(), root, address)
|
|
if err != nil {
|
|
p.Log().Warn("Failed to retrieve account for proof", "block", header.Number, "hash", header.Hash(), "account", address, "err", err)
|
|
p.bumpInvalid()
|
|
continue
|
|
}
|
|
trie, err = statedb.OpenStorageTrie(root, address, account.Root)
|
|
if trie == nil || err != nil {
|
|
p.Log().Warn("Failed to open storage trie for proof", "block", header.Number, "hash", header.Hash(), "account", address, "root", account.Root, "err", err)
|
|
continue
|
|
}
|
|
}
|
|
// Prove the user's request from the account or storage trie
|
|
if err := trie.Prove(request.Key, nodes); err != nil {
|
|
p.Log().Warn("Failed to prove state request", "block", header.Number, "hash", header.Hash(), "err", err)
|
|
continue
|
|
}
|
|
if nodes.DataSize() >= softResponseLimit {
|
|
break
|
|
}
|
|
}
|
|
return p.replyProofsV2(r.ReqID, nodes.NodeList())
|
|
}, r.ReqID, uint64(len(r.Reqs)), nil
|
|
}
|
|
|
|
// handleGetHelperTrieProofs handles a helper trie proof request
|
|
func handleGetHelperTrieProofs(msg Decoder) (serveRequestFn, uint64, uint64, error) {
|
|
var r GetHelperTrieProofsPacket
|
|
if err := msg.Decode(&r); err != nil {
|
|
return nil, 0, 0, err
|
|
}
|
|
return func(backend serverBackend, p *clientPeer, waitOrStop func() bool) *reply {
|
|
var (
|
|
lastIdx uint64
|
|
lastType uint
|
|
auxTrie *trie.Trie
|
|
auxBytes int
|
|
auxData [][]byte
|
|
)
|
|
bc := backend.BlockChain()
|
|
nodes := light.NewNodeSet()
|
|
for i, request := range r.Reqs {
|
|
if i != 0 && !waitOrStop() {
|
|
return nil
|
|
}
|
|
if auxTrie == nil || request.Type != lastType || request.TrieIdx != lastIdx {
|
|
lastType, lastIdx = request.Type, request.TrieIdx
|
|
auxTrie = backend.GetHelperTrie(request.Type, request.TrieIdx)
|
|
}
|
|
if auxTrie == nil {
|
|
return nil
|
|
}
|
|
// TODO(rjl493456442) short circuit if the proving is failed.
|
|
// The original client side code has a dirty hack to retrieve
|
|
// the headers with no valid proof. Keep the compatibility for
|
|
// legacy les protocol and drop this hack when the les2/3 are
|
|
// not supported.
|
|
err := auxTrie.Prove(request.Key, nodes)
|
|
if p.version >= lpv4 && err != nil {
|
|
return nil
|
|
}
|
|
if request.Type == htCanonical && request.AuxReq == htAuxHeader && len(request.Key) == 8 {
|
|
header := bc.GetHeaderByNumber(binary.BigEndian.Uint64(request.Key))
|
|
data, err := rlp.EncodeToBytes(header)
|
|
if err != nil {
|
|
log.Error("Failed to encode header", "err", err)
|
|
return nil
|
|
}
|
|
auxData = append(auxData, data)
|
|
auxBytes += len(data)
|
|
}
|
|
if nodes.DataSize()+auxBytes >= softResponseLimit {
|
|
break
|
|
}
|
|
}
|
|
return p.replyHelperTrieProofs(r.ReqID, HelperTrieResps{Proofs: nodes.NodeList(), AuxData: auxData})
|
|
}, r.ReqID, uint64(len(r.Reqs)), nil
|
|
}
|
|
|
|
// handleSendTx handles a transaction propagation request
|
|
func handleSendTx(msg Decoder) (serveRequestFn, uint64, uint64, error) {
|
|
var r SendTxPacket
|
|
if err := msg.Decode(&r); err != nil {
|
|
return nil, 0, 0, err
|
|
}
|
|
amount := uint64(len(r.Txs))
|
|
return func(backend serverBackend, p *clientPeer, waitOrStop func() bool) *reply {
|
|
stats := make([]light.TxStatus, len(r.Txs))
|
|
for i, tx := range r.Txs {
|
|
if i != 0 && !waitOrStop() {
|
|
return nil
|
|
}
|
|
hash := tx.Hash()
|
|
stats[i] = txStatus(backend, hash)
|
|
if stats[i].Status == txpool.TxStatusUnknown {
|
|
if errs := backend.TxPool().Add([]*types.Transaction{tx}, false, backend.AddTxsSync()); errs[0] != nil {
|
|
stats[i].Error = errs[0].Error()
|
|
continue
|
|
}
|
|
stats[i] = txStatus(backend, hash)
|
|
}
|
|
}
|
|
return p.replyTxStatus(r.ReqID, stats)
|
|
}, r.ReqID, amount, nil
|
|
}
|
|
|
|
// handleGetTxStatus handles a transaction status query
|
|
func handleGetTxStatus(msg Decoder) (serveRequestFn, uint64, uint64, error) {
|
|
var r GetTxStatusPacket
|
|
if err := msg.Decode(&r); err != nil {
|
|
return nil, 0, 0, err
|
|
}
|
|
return func(backend serverBackend, p *clientPeer, waitOrStop func() bool) *reply {
|
|
stats := make([]light.TxStatus, len(r.Hashes))
|
|
for i, hash := range r.Hashes {
|
|
if i != 0 && !waitOrStop() {
|
|
return nil
|
|
}
|
|
stats[i] = txStatus(backend, hash)
|
|
}
|
|
return p.replyTxStatus(r.ReqID, stats)
|
|
}, r.ReqID, uint64(len(r.Hashes)), nil
|
|
}
|
|
|
|
// txStatus returns the status of a specified transaction.
|
|
func txStatus(b serverBackend, hash common.Hash) light.TxStatus {
|
|
var stat light.TxStatus
|
|
// Looking the transaction in txpool first.
|
|
stat.Status = b.TxPool().Status(hash)
|
|
|
|
// If the transaction is unknown to the pool, try looking it up locally.
|
|
if stat.Status == txpool.TxStatusUnknown {
|
|
lookup := b.BlockChain().GetTransactionLookup(hash)
|
|
if lookup != nil {
|
|
stat.Status = txpool.TxStatusIncluded
|
|
stat.Lookup = lookup
|
|
}
|
|
}
|
|
return stat
|
|
}
|