forked from cerc-io/plugeth
rpc: refactor subscriptions and filters
This commit is contained in:
parent
3b39d4d1c1
commit
47ff813012
@ -114,6 +114,28 @@ func (h *Header) UnmarshalJSON(data []byte) error {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (h *Header) MarshalJSON() ([]byte, error) {
|
||||||
|
fields := map[string]interface{}{
|
||||||
|
"hash": h.Hash(),
|
||||||
|
"parentHash": h.ParentHash,
|
||||||
|
"number": fmt.Sprintf("%#x", h.Number),
|
||||||
|
"nonce": h.Nonce,
|
||||||
|
"receiptRoot": h.ReceiptHash,
|
||||||
|
"logsBloom": h.Bloom,
|
||||||
|
"sha3Uncles": h.UncleHash,
|
||||||
|
"stateRoot": h.Root,
|
||||||
|
"miner": h.Coinbase,
|
||||||
|
"difficulty": fmt.Sprintf("%#x", h.Difficulty),
|
||||||
|
"extraData": fmt.Sprintf("0x%x", h.Extra),
|
||||||
|
"gasLimit": fmt.Sprintf("%#x", h.GasLimit),
|
||||||
|
"gasUsed": fmt.Sprintf("%#x", h.GasUsed),
|
||||||
|
"timestamp": fmt.Sprintf("%#x", h.Time),
|
||||||
|
"transactionsRoot": h.TxHash,
|
||||||
|
}
|
||||||
|
|
||||||
|
return json.Marshal(fields)
|
||||||
|
}
|
||||||
|
|
||||||
func rlpHash(x interface{}) (h common.Hash) {
|
func rlpHash(x interface{}) (h common.Hash) {
|
||||||
hw := sha3.NewKeccak256()
|
hw := sha3.NewKeccak256()
|
||||||
rlp.Encode(hw, x)
|
rlp.Encode(hw, x)
|
||||||
|
@ -19,10 +19,9 @@ package downloader
|
|||||||
import (
|
import (
|
||||||
"sync"
|
"sync"
|
||||||
|
|
||||||
"golang.org/x/net/context"
|
|
||||||
|
|
||||||
"github.com/ethereum/go-ethereum/event"
|
"github.com/ethereum/go-ethereum/event"
|
||||||
"github.com/ethereum/go-ethereum/rpc"
|
"github.com/ethereum/go-ethereum/rpc"
|
||||||
|
"golang.org/x/net/context"
|
||||||
)
|
)
|
||||||
|
|
||||||
// PublicDownloaderAPI provides an API which gives information about the current synchronisation status.
|
// PublicDownloaderAPI provides an API which gives information about the current synchronisation status.
|
||||||
@ -30,25 +29,48 @@ import (
|
|||||||
type PublicDownloaderAPI struct {
|
type PublicDownloaderAPI struct {
|
||||||
d *Downloader
|
d *Downloader
|
||||||
mux *event.TypeMux
|
mux *event.TypeMux
|
||||||
muSyncSubscriptions sync.Mutex
|
installSyncSubscription chan chan interface{}
|
||||||
syncSubscriptions map[string]rpc.Subscription
|
uninstallSyncSubscription chan *uninstallSyncSubscriptionRequest
|
||||||
}
|
}
|
||||||
|
|
||||||
// NewPublicDownloaderAPI create a new PublicDownloaderAPI.
|
// NewPublicDownloaderAPI create a new PublicDownloaderAPI. The API has an internal event loop that
|
||||||
|
// listens for events from the downloader through the global event mux. In case it receives one of
|
||||||
|
// these events it broadcasts it to all syncing subscriptions that are installed through the
|
||||||
|
// installSyncSubscription channel.
|
||||||
func NewPublicDownloaderAPI(d *Downloader, m *event.TypeMux) *PublicDownloaderAPI {
|
func NewPublicDownloaderAPI(d *Downloader, m *event.TypeMux) *PublicDownloaderAPI {
|
||||||
api := &PublicDownloaderAPI{d: d, mux: m, syncSubscriptions: make(map[string]rpc.Subscription)}
|
api := &PublicDownloaderAPI{
|
||||||
|
d: d,
|
||||||
|
mux: m,
|
||||||
|
installSyncSubscription: make(chan chan interface{}),
|
||||||
|
uninstallSyncSubscription: make(chan *uninstallSyncSubscriptionRequest),
|
||||||
|
}
|
||||||
|
|
||||||
go api.run()
|
go api.eventLoop()
|
||||||
|
|
||||||
return api
|
return api
|
||||||
}
|
}
|
||||||
|
|
||||||
func (api *PublicDownloaderAPI) run() {
|
// eventLoop runs an loop until the event mux closes. It will install and uninstall new
|
||||||
sub := api.mux.Subscribe(StartEvent{}, DoneEvent{}, FailedEvent{})
|
// sync subscriptions and broadcasts sync status updates to the installed sync subscriptions.
|
||||||
|
func (api *PublicDownloaderAPI) eventLoop() {
|
||||||
|
var (
|
||||||
|
sub = api.mux.Subscribe(StartEvent{}, DoneEvent{}, FailedEvent{})
|
||||||
|
syncSubscriptions = make(map[chan interface{}]struct{})
|
||||||
|
)
|
||||||
|
|
||||||
|
for {
|
||||||
|
select {
|
||||||
|
case i := <-api.installSyncSubscription:
|
||||||
|
syncSubscriptions[i] = struct{}{}
|
||||||
|
case u := <-api.uninstallSyncSubscription:
|
||||||
|
delete(syncSubscriptions, u.c)
|
||||||
|
close(u.uninstalled)
|
||||||
|
case event := <-sub.Chan():
|
||||||
|
if event == nil {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
for event := range sub.Chan() {
|
|
||||||
var notification interface{}
|
var notification interface{}
|
||||||
|
|
||||||
switch event.Data.(type) {
|
switch event.Data.(type) {
|
||||||
case StartEvent:
|
case StartEvent:
|
||||||
result := &SyncingResult{Syncing: true}
|
result := &SyncingResult{Syncing: true}
|
||||||
@ -57,15 +79,42 @@ func (api *PublicDownloaderAPI) run() {
|
|||||||
case DoneEvent, FailedEvent:
|
case DoneEvent, FailedEvent:
|
||||||
notification = false
|
notification = false
|
||||||
}
|
}
|
||||||
|
// broadcast
|
||||||
|
for c := range syncSubscriptions {
|
||||||
|
c <- notification
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
api.muSyncSubscriptions.Lock()
|
// Syncing provides information when this nodes starts synchronising with the Ethereum network and when it's finished.
|
||||||
for id, sub := range api.syncSubscriptions {
|
func (api *PublicDownloaderAPI) Syncing(ctx context.Context) (*rpc.Subscription, error) {
|
||||||
if sub.Notify(notification) == rpc.ErrNotificationNotFound {
|
notifier, supported := rpc.NotifierFromContext(ctx)
|
||||||
delete(api.syncSubscriptions, id)
|
if !supported {
|
||||||
|
return &rpc.Subscription{}, rpc.ErrNotificationsUnsupported
|
||||||
|
}
|
||||||
|
|
||||||
|
rpcSub := notifier.CreateSubscription()
|
||||||
|
|
||||||
|
go func() {
|
||||||
|
statuses := make(chan interface{})
|
||||||
|
sub := api.SubscribeSyncStatus(statuses)
|
||||||
|
|
||||||
|
for {
|
||||||
|
select {
|
||||||
|
case status := <-statuses:
|
||||||
|
notifier.Notify(rpcSub.ID, status)
|
||||||
|
case <-rpcSub.Err():
|
||||||
|
sub.Unsubscribe()
|
||||||
|
return
|
||||||
|
case <-notifier.Closed():
|
||||||
|
sub.Unsubscribe()
|
||||||
|
return
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
api.muSyncSubscriptions.Unlock()
|
}()
|
||||||
}
|
|
||||||
|
return rpcSub, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// Progress gives progress indications when the node is synchronising with the Ethereum network.
|
// Progress gives progress indications when the node is synchronising with the Ethereum network.
|
||||||
@ -83,26 +132,42 @@ type SyncingResult struct {
|
|||||||
Status Progress `json:"status"`
|
Status Progress `json:"status"`
|
||||||
}
|
}
|
||||||
|
|
||||||
// Syncing provides information when this nodes starts synchronising with the Ethereum network and when it's finished.
|
// uninstallSyncSubscriptionRequest uninstalles a syncing subscription in the API event loop.
|
||||||
func (api *PublicDownloaderAPI) Syncing(ctx context.Context) (rpc.Subscription, error) {
|
type uninstallSyncSubscriptionRequest struct {
|
||||||
notifier, supported := rpc.NotifierFromContext(ctx)
|
c chan interface{}
|
||||||
if !supported {
|
uninstalled chan interface{}
|
||||||
return nil, rpc.ErrNotificationsUnsupported
|
|
||||||
}
|
}
|
||||||
|
|
||||||
subscription, err := notifier.NewSubscription(func(id string) {
|
// SyncStatusSubscription represents a syncing subscription.
|
||||||
api.muSyncSubscriptions.Lock()
|
type SyncStatusSubscription struct {
|
||||||
delete(api.syncSubscriptions, id)
|
api *PublicDownloaderAPI // register subscription in event loop of this api instance
|
||||||
api.muSyncSubscriptions.Unlock()
|
c chan interface{} // channel where events are broadcasted to
|
||||||
|
unsubOnce sync.Once // make sure unsubscribe logic is executed once
|
||||||
|
}
|
||||||
|
|
||||||
|
// Unsubscribe uninstalls the subscription from the DownloadAPI event loop.
|
||||||
|
// The status channel that was passed to subscribeSyncStatus isn't used anymore
|
||||||
|
// after this method returns.
|
||||||
|
func (s *SyncStatusSubscription) Unsubscribe() {
|
||||||
|
s.unsubOnce.Do(func() {
|
||||||
|
req := uninstallSyncSubscriptionRequest{s.c, make(chan interface{})}
|
||||||
|
s.api.uninstallSyncSubscription <- &req
|
||||||
|
|
||||||
|
for {
|
||||||
|
select {
|
||||||
|
case <-s.c:
|
||||||
|
// drop new status events until uninstall confirmation
|
||||||
|
continue
|
||||||
|
case <-req.uninstalled:
|
||||||
|
return
|
||||||
|
}
|
||||||
|
}
|
||||||
})
|
})
|
||||||
|
|
||||||
if err != nil {
|
|
||||||
return nil, err
|
|
||||||
}
|
}
|
||||||
|
|
||||||
api.muSyncSubscriptions.Lock()
|
// SubscribeSyncStatus creates a subscription that will broadcast new synchronisation updates.
|
||||||
api.syncSubscriptions[subscription.ID()] = subscription
|
// The given channel must receive interface values, the result can either
|
||||||
api.muSyncSubscriptions.Unlock()
|
func (api *PublicDownloaderAPI) SubscribeSyncStatus(status chan interface{}) *SyncStatusSubscription {
|
||||||
|
api.installSyncSubscription <- status
|
||||||
return subscription, nil
|
return &SyncStatusSubscription{api: api, c: status}
|
||||||
}
|
}
|
||||||
|
@ -17,292 +17,414 @@
|
|||||||
package filters
|
package filters
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"crypto/rand"
|
|
||||||
"encoding/hex"
|
"encoding/hex"
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
"errors"
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
|
"math/big"
|
||||||
"sync"
|
"sync"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
|
"golang.org/x/net/context"
|
||||||
|
|
||||||
"github.com/ethereum/go-ethereum/common"
|
"github.com/ethereum/go-ethereum/common"
|
||||||
"github.com/ethereum/go-ethereum/core/types"
|
"github.com/ethereum/go-ethereum/core/types"
|
||||||
"github.com/ethereum/go-ethereum/core/vm"
|
|
||||||
"github.com/ethereum/go-ethereum/ethdb"
|
"github.com/ethereum/go-ethereum/ethdb"
|
||||||
"github.com/ethereum/go-ethereum/event"
|
"github.com/ethereum/go-ethereum/event"
|
||||||
"github.com/ethereum/go-ethereum/rpc"
|
"github.com/ethereum/go-ethereum/rpc"
|
||||||
|
|
||||||
"golang.org/x/net/context"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
var (
|
var (
|
||||||
filterTickerTime = 5 * time.Minute
|
deadline = 5 * time.Minute // consider a filter inactive if it has not been polled for within deadline
|
||||||
)
|
)
|
||||||
|
|
||||||
// byte will be inferred
|
// filter is a helper struct that holds meta information over the filter type
|
||||||
const (
|
// and associated subscription in the event system.
|
||||||
unknownFilterTy = iota
|
type filter struct {
|
||||||
blockFilterTy
|
typ Type
|
||||||
transactionFilterTy
|
deadline *time.Timer // filter is inactiv when deadline triggers
|
||||||
logFilterTy
|
hashes []common.Hash
|
||||||
)
|
crit FilterCriteria
|
||||||
|
logs []Log
|
||||||
|
s *Subscription // associated subscription in event system
|
||||||
|
}
|
||||||
|
|
||||||
// PublicFilterAPI offers support to create and manage filters. This will allow external clients to retrieve various
|
// PublicFilterAPI offers support to create and manage filters. This will allow external clients to retrieve various
|
||||||
// information related to the Ethereum protocol such als blocks, transactions and logs.
|
// information related to the Ethereum protocol such als blocks, transactions and logs.
|
||||||
type PublicFilterAPI struct {
|
type PublicFilterAPI struct {
|
||||||
mux *event.TypeMux
|
mux *event.TypeMux
|
||||||
|
|
||||||
quit chan struct{}
|
quit chan struct{}
|
||||||
chainDb ethdb.Database
|
chainDb ethdb.Database
|
||||||
|
events *EventSystem
|
||||||
filterManager *FilterSystem
|
filtersMu sync.Mutex
|
||||||
|
filters map[rpc.ID]*filter
|
||||||
filterMapMu sync.RWMutex
|
|
||||||
filterMapping map[string]int // maps between filter internal filter identifiers and external filter identifiers
|
|
||||||
|
|
||||||
logMu sync.RWMutex
|
|
||||||
logQueue map[int]*logQueue
|
|
||||||
|
|
||||||
blockMu sync.RWMutex
|
|
||||||
blockQueue map[int]*hashQueue
|
|
||||||
|
|
||||||
transactionMu sync.RWMutex
|
|
||||||
transactionQueue map[int]*hashQueue
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// NewPublicFilterAPI returns a new PublicFilterAPI instance.
|
// NewPublicFilterAPI returns a new PublicFilterAPI instance.
|
||||||
func NewPublicFilterAPI(chainDb ethdb.Database, mux *event.TypeMux) *PublicFilterAPI {
|
func NewPublicFilterAPI(chainDb ethdb.Database, mux *event.TypeMux) *PublicFilterAPI {
|
||||||
svc := &PublicFilterAPI{
|
api := &PublicFilterAPI{
|
||||||
mux: mux,
|
mux: mux,
|
||||||
chainDb: chainDb,
|
chainDb: chainDb,
|
||||||
filterManager: NewFilterSystem(mux),
|
events: NewEventSystem(mux),
|
||||||
filterMapping: make(map[string]int),
|
filters: make(map[rpc.ID]*filter),
|
||||||
logQueue: make(map[int]*logQueue),
|
|
||||||
blockQueue: make(map[int]*hashQueue),
|
|
||||||
transactionQueue: make(map[int]*hashQueue),
|
|
||||||
}
|
|
||||||
go svc.start()
|
|
||||||
return svc
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// Stop quits the work loop.
|
go api.timeoutLoop()
|
||||||
func (s *PublicFilterAPI) Stop() {
|
|
||||||
close(s.quit)
|
return api
|
||||||
}
|
}
|
||||||
|
|
||||||
// start the work loop, wait and process events.
|
// timeoutLoop runs every 5 minutes and deletes filters that have not been recently used.
|
||||||
func (s *PublicFilterAPI) start() {
|
// Tt is started when the api is created.
|
||||||
timer := time.NewTicker(2 * time.Second)
|
func (api *PublicFilterAPI) timeoutLoop() {
|
||||||
defer timer.Stop()
|
ticker := time.NewTicker(5 * time.Minute)
|
||||||
done:
|
for {
|
||||||
|
<-ticker.C
|
||||||
|
api.filtersMu.Lock()
|
||||||
|
for id, f := range api.filters {
|
||||||
|
select {
|
||||||
|
case <-f.deadline.C:
|
||||||
|
f.s.Unsubscribe()
|
||||||
|
delete(api.filters, id)
|
||||||
|
default:
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
}
|
||||||
|
api.filtersMu.Unlock()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// NewPendingTransactionFilter creates a filter that fetches pending transaction hashes
|
||||||
|
// as transactions enter the pending state.
|
||||||
|
//
|
||||||
|
// It is part of the filter package because this filter can be used throug the
|
||||||
|
// `eth_getFilterChanges` polling method that is also used for log filters.
|
||||||
|
//
|
||||||
|
// https://github.com/ethereum/wiki/wiki/JSON-RPC#eth_newpendingtransactionfilter
|
||||||
|
func (api *PublicFilterAPI) NewPendingTransactionFilter() rpc.ID {
|
||||||
|
var (
|
||||||
|
pendingTxs = make(chan common.Hash)
|
||||||
|
pendingTxSub = api.events.SubscribePendingTxEvents(pendingTxs)
|
||||||
|
)
|
||||||
|
|
||||||
|
api.filtersMu.Lock()
|
||||||
|
api.filters[pendingTxSub.ID] = &filter{typ: PendingTransactionsSubscription, deadline: time.NewTimer(deadline), hashes: make([]common.Hash, 0), s: pendingTxSub}
|
||||||
|
api.filtersMu.Unlock()
|
||||||
|
|
||||||
|
go func() {
|
||||||
for {
|
for {
|
||||||
select {
|
select {
|
||||||
case <-timer.C:
|
case ph := <-pendingTxs:
|
||||||
s.filterManager.Lock() // lock order like filterLoop()
|
api.filtersMu.Lock()
|
||||||
s.logMu.Lock()
|
if f, found := api.filters[pendingTxSub.ID]; found {
|
||||||
for id, filter := range s.logQueue {
|
f.hashes = append(f.hashes, ph)
|
||||||
if time.Since(filter.timeout) > filterTickerTime {
|
}
|
||||||
s.filterManager.Remove(id)
|
api.filtersMu.Unlock()
|
||||||
delete(s.logQueue, id)
|
case <-pendingTxSub.Err():
|
||||||
|
api.filtersMu.Lock()
|
||||||
|
delete(api.filters, pendingTxSub.ID)
|
||||||
|
api.filtersMu.Unlock()
|
||||||
|
return
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
s.logMu.Unlock()
|
}()
|
||||||
|
|
||||||
s.blockMu.Lock()
|
return pendingTxSub.ID
|
||||||
for id, filter := range s.blockQueue {
|
|
||||||
if time.Since(filter.timeout) > filterTickerTime {
|
|
||||||
s.filterManager.Remove(id)
|
|
||||||
delete(s.blockQueue, id)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
s.blockMu.Unlock()
|
|
||||||
|
|
||||||
s.transactionMu.Lock()
|
|
||||||
for id, filter := range s.transactionQueue {
|
|
||||||
if time.Since(filter.timeout) > filterTickerTime {
|
|
||||||
s.filterManager.Remove(id)
|
|
||||||
delete(s.transactionQueue, id)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
s.transactionMu.Unlock()
|
|
||||||
s.filterManager.Unlock()
|
|
||||||
case <-s.quit:
|
|
||||||
break done
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// NewPendingTransactions creates a subscription that is triggered each time a transaction
|
||||||
|
// enters the transaction pool and was signed from one of the transactions this nodes manages.
|
||||||
|
func (api *PublicFilterAPI) NewPendingTransactions(ctx context.Context) (*rpc.Subscription, error) {
|
||||||
|
notifier, supported := rpc.NotifierFromContext(ctx)
|
||||||
|
if !supported {
|
||||||
|
return &rpc.Subscription{}, rpc.ErrNotificationsUnsupported
|
||||||
}
|
}
|
||||||
|
|
||||||
// NewBlockFilter create a new filter that returns blocks that are included into the canonical chain.
|
rpcSub := notifier.CreateSubscription()
|
||||||
func (s *PublicFilterAPI) NewBlockFilter() (string, error) {
|
|
||||||
// protect filterManager.Add() and setting of filter fields
|
|
||||||
s.filterManager.Lock()
|
|
||||||
defer s.filterManager.Unlock()
|
|
||||||
|
|
||||||
externalId, err := newFilterId()
|
go func() {
|
||||||
if err != nil {
|
txHashes := make(chan common.Hash)
|
||||||
return "", err
|
pendingTxSub := api.events.SubscribePendingTxEvents(txHashes)
|
||||||
}
|
|
||||||
|
|
||||||
filter := New(s.chainDb)
|
for {
|
||||||
id, err := s.filterManager.Add(filter, ChainFilter)
|
select {
|
||||||
if err != nil {
|
case h := <-txHashes:
|
||||||
return "", err
|
notifier.Notify(rpcSub.ID, h)
|
||||||
}
|
case <-rpcSub.Err():
|
||||||
|
pendingTxSub.Unsubscribe()
|
||||||
s.blockMu.Lock()
|
return
|
||||||
s.blockQueue[id] = &hashQueue{timeout: time.Now()}
|
case <-notifier.Closed():
|
||||||
s.blockMu.Unlock()
|
pendingTxSub.Unsubscribe()
|
||||||
|
return
|
||||||
filter.BlockCallback = func(block *types.Block, logs vm.Logs) {
|
|
||||||
s.blockMu.Lock()
|
|
||||||
defer s.blockMu.Unlock()
|
|
||||||
|
|
||||||
if queue := s.blockQueue[id]; queue != nil {
|
|
||||||
queue.add(block.Hash())
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
}()
|
||||||
|
|
||||||
s.filterMapMu.Lock()
|
return rpcSub, nil
|
||||||
s.filterMapping[externalId] = id
|
|
||||||
s.filterMapMu.Unlock()
|
|
||||||
|
|
||||||
return externalId, nil
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// NewPendingTransactionFilter creates a filter that returns new pending transactions.
|
// NewBlockFilter creates a filter that fetches blocks that are imported into the chain.
|
||||||
func (s *PublicFilterAPI) NewPendingTransactionFilter() (string, error) {
|
// It is part of the filter package since polling goes with eth_getFilterChanges.
|
||||||
// protect filterManager.Add() and setting of filter fields
|
//
|
||||||
s.filterManager.Lock()
|
// https://github.com/ethereum/wiki/wiki/JSON-RPC#eth_newblockfilter
|
||||||
defer s.filterManager.Unlock()
|
func (api *PublicFilterAPI) NewBlockFilter() rpc.ID {
|
||||||
|
var (
|
||||||
|
headers = make(chan *types.Header)
|
||||||
|
headerSub = api.events.SubscribeNewHeads(headers)
|
||||||
|
)
|
||||||
|
|
||||||
externalId, err := newFilterId()
|
api.filtersMu.Lock()
|
||||||
if err != nil {
|
api.filters[headerSub.ID] = &filter{typ: BlocksSubscription, deadline: time.NewTimer(deadline), hashes: make([]common.Hash, 0), s: headerSub}
|
||||||
return "", err
|
api.filtersMu.Unlock()
|
||||||
|
|
||||||
|
go func() {
|
||||||
|
for {
|
||||||
|
select {
|
||||||
|
case h := <-headers:
|
||||||
|
api.filtersMu.Lock()
|
||||||
|
if f, found := api.filters[headerSub.ID]; found {
|
||||||
|
f.hashes = append(f.hashes, h.Hash())
|
||||||
}
|
}
|
||||||
|
api.filtersMu.Unlock()
|
||||||
filter := New(s.chainDb)
|
case <-headerSub.Err():
|
||||||
id, err := s.filterManager.Add(filter, PendingTxFilter)
|
api.filtersMu.Lock()
|
||||||
if err != nil {
|
delete(api.filters, headerSub.ID)
|
||||||
return "", err
|
api.filtersMu.Unlock()
|
||||||
}
|
return
|
||||||
|
|
||||||
s.transactionMu.Lock()
|
|
||||||
s.transactionQueue[id] = &hashQueue{timeout: time.Now()}
|
|
||||||
s.transactionMu.Unlock()
|
|
||||||
|
|
||||||
filter.TransactionCallback = func(tx *types.Transaction) {
|
|
||||||
s.transactionMu.Lock()
|
|
||||||
defer s.transactionMu.Unlock()
|
|
||||||
|
|
||||||
if queue := s.transactionQueue[id]; queue != nil {
|
|
||||||
queue.add(tx.Hash())
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
}()
|
||||||
|
|
||||||
s.filterMapMu.Lock()
|
return headerSub.ID
|
||||||
s.filterMapping[externalId] = id
|
|
||||||
s.filterMapMu.Unlock()
|
|
||||||
|
|
||||||
return externalId, nil
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// newLogFilter creates a new log filter.
|
// NewHeads send a notification each time a new (header) block is appended to the chain.
|
||||||
func (s *PublicFilterAPI) newLogFilter(earliest, latest int64, addresses []common.Address, topics [][]common.Hash, callback func(log *vm.Log, removed bool)) (int, error) {
|
func (api *PublicFilterAPI) NewHeads(ctx context.Context) (*rpc.Subscription, error) {
|
||||||
// protect filterManager.Add() and setting of filter fields
|
notifier, supported := rpc.NotifierFromContext(ctx)
|
||||||
s.filterManager.Lock()
|
if !supported {
|
||||||
defer s.filterManager.Unlock()
|
return &rpc.Subscription{}, rpc.ErrNotificationsUnsupported
|
||||||
|
|
||||||
filter := New(s.chainDb)
|
|
||||||
id, err := s.filterManager.Add(filter, LogFilter)
|
|
||||||
if err != nil {
|
|
||||||
return 0, err
|
|
||||||
}
|
}
|
||||||
|
|
||||||
s.logMu.Lock()
|
rpcSub := notifier.CreateSubscription()
|
||||||
s.logQueue[id] = &logQueue{timeout: time.Now()}
|
|
||||||
s.logMu.Unlock()
|
|
||||||
|
|
||||||
filter.SetBeginBlock(earliest)
|
go func() {
|
||||||
filter.SetEndBlock(latest)
|
headers := make(chan *types.Header)
|
||||||
filter.SetAddresses(addresses)
|
headersSub := api.events.SubscribeNewHeads(headers)
|
||||||
filter.SetTopics(topics)
|
|
||||||
filter.LogCallback = func(log *vm.Log, removed bool) {
|
|
||||||
if callback != nil {
|
|
||||||
callback(log, removed)
|
|
||||||
} else {
|
|
||||||
s.logMu.Lock()
|
|
||||||
defer s.logMu.Unlock()
|
|
||||||
if queue := s.logQueue[id]; queue != nil {
|
|
||||||
queue.add(vmlog{log, removed})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
return id, nil
|
for {
|
||||||
|
select {
|
||||||
|
case h := <-headers:
|
||||||
|
notifier.Notify(rpcSub.ID, h)
|
||||||
|
case <-rpcSub.Err():
|
||||||
|
headersSub.Unsubscribe()
|
||||||
|
return
|
||||||
|
case <-notifier.Closed():
|
||||||
|
headersSub.Unsubscribe()
|
||||||
|
return
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
|
||||||
|
return rpcSub, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// Logs creates a subscription that fires for all new log that match the given filter criteria.
|
// Logs creates a subscription that fires for all new log that match the given filter criteria.
|
||||||
func (s *PublicFilterAPI) Logs(ctx context.Context, args NewFilterArgs) (rpc.Subscription, error) {
|
func (api *PublicFilterAPI) Logs(ctx context.Context, crit FilterCriteria) (*rpc.Subscription, error) {
|
||||||
notifier, supported := rpc.NotifierFromContext(ctx)
|
notifier, supported := rpc.NotifierFromContext(ctx)
|
||||||
if !supported {
|
if !supported {
|
||||||
return nil, rpc.ErrNotificationsUnsupported
|
return &rpc.Subscription{}, rpc.ErrNotificationsUnsupported
|
||||||
}
|
}
|
||||||
|
|
||||||
var (
|
rpcSub := notifier.CreateSubscription()
|
||||||
externalId string
|
|
||||||
subscription rpc.Subscription
|
|
||||||
err error
|
|
||||||
)
|
|
||||||
|
|
||||||
if externalId, err = newFilterId(); err != nil {
|
go func() {
|
||||||
return nil, err
|
matchedLogs := make(chan []Log)
|
||||||
|
logsSub := api.events.SubscribeLogs(crit, matchedLogs)
|
||||||
|
|
||||||
|
for {
|
||||||
|
select {
|
||||||
|
case logs := <-matchedLogs:
|
||||||
|
for _, log := range logs {
|
||||||
|
notifier.Notify(rpcSub.ID, &log)
|
||||||
}
|
}
|
||||||
|
case <-rpcSub.Err(): // client send an unsubscribe request
|
||||||
// uninstall filter when subscription is unsubscribed/cancelled
|
logsSub.Unsubscribe()
|
||||||
if subscription, err = notifier.NewSubscription(func(string) {
|
return
|
||||||
s.UninstallFilter(externalId)
|
case <-notifier.Closed(): // connection dropped
|
||||||
}); err != nil {
|
logsSub.Unsubscribe()
|
||||||
return nil, err
|
return
|
||||||
}
|
|
||||||
|
|
||||||
notifySubscriber := func(log *vm.Log, removed bool) {
|
|
||||||
rpcLog := toRPCLogs(vm.Logs{log}, removed)
|
|
||||||
if err := subscription.Notify(rpcLog); err != nil {
|
|
||||||
subscription.Cancel()
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
}()
|
||||||
|
|
||||||
// from and to block number are not used since subscriptions don't allow you to travel to "time"
|
return rpcSub, nil
|
||||||
var id int
|
|
||||||
if len(args.Addresses) > 0 {
|
|
||||||
id, err = s.newLogFilter(-1, -1, args.Addresses, args.Topics, notifySubscriber)
|
|
||||||
} else {
|
|
||||||
id, err = s.newLogFilter(-1, -1, nil, args.Topics, notifySubscriber)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
if err != nil {
|
// FilterCriteria represents a request to create a new filter.
|
||||||
subscription.Cancel()
|
type FilterCriteria struct {
|
||||||
return nil, err
|
FromBlock *big.Int
|
||||||
}
|
ToBlock *big.Int
|
||||||
|
|
||||||
s.filterMapMu.Lock()
|
|
||||||
s.filterMapping[externalId] = id
|
|
||||||
s.filterMapMu.Unlock()
|
|
||||||
|
|
||||||
return subscription, err
|
|
||||||
}
|
|
||||||
|
|
||||||
// NewFilterArgs represents a request to create a new filter.
|
|
||||||
type NewFilterArgs struct {
|
|
||||||
FromBlock rpc.BlockNumber
|
|
||||||
ToBlock rpc.BlockNumber
|
|
||||||
Addresses []common.Address
|
Addresses []common.Address
|
||||||
Topics [][]common.Hash
|
Topics [][]common.Hash
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// NewFilter creates a new filter and returns the filter id. It can be
|
||||||
|
// used to retrieve logs when the state changes. This method cannot be
|
||||||
|
// used to fetch logs that are already stored in the state.
|
||||||
|
//
|
||||||
|
// https://github.com/ethereum/wiki/wiki/JSON-RPC#eth_newfilter
|
||||||
|
func (api *PublicFilterAPI) NewFilter(crit FilterCriteria) rpc.ID {
|
||||||
|
var (
|
||||||
|
logs = make(chan []Log)
|
||||||
|
logsSub = api.events.SubscribeLogs(crit, logs)
|
||||||
|
)
|
||||||
|
|
||||||
|
if crit.FromBlock == nil {
|
||||||
|
crit.FromBlock = big.NewInt(rpc.LatestBlockNumber.Int64())
|
||||||
|
}
|
||||||
|
if crit.ToBlock == nil {
|
||||||
|
crit.ToBlock = big.NewInt(rpc.LatestBlockNumber.Int64())
|
||||||
|
}
|
||||||
|
|
||||||
|
api.filtersMu.Lock()
|
||||||
|
api.filters[logsSub.ID] = &filter{typ: LogsSubscription, crit: crit, deadline: time.NewTimer(deadline), logs: make([]Log, 0), s: logsSub}
|
||||||
|
api.filtersMu.Unlock()
|
||||||
|
|
||||||
|
go func() {
|
||||||
|
for {
|
||||||
|
select {
|
||||||
|
case l := <-logs:
|
||||||
|
api.filtersMu.Lock()
|
||||||
|
if f, found := api.filters[logsSub.ID]; found {
|
||||||
|
f.logs = append(f.logs, l...)
|
||||||
|
}
|
||||||
|
api.filtersMu.Unlock()
|
||||||
|
case <-logsSub.Err():
|
||||||
|
api.filtersMu.Lock()
|
||||||
|
delete(api.filters, logsSub.ID)
|
||||||
|
api.filtersMu.Unlock()
|
||||||
|
return
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
|
||||||
|
return logsSub.ID
|
||||||
|
}
|
||||||
|
|
||||||
|
// GetLogs returns logs matching the given argument that are stored within the state.
|
||||||
|
//
|
||||||
|
// https://github.com/ethereum/wiki/wiki/JSON-RPC#eth_getlogs
|
||||||
|
func (api *PublicFilterAPI) GetLogs(crit FilterCriteria) []Log {
|
||||||
|
if crit.FromBlock == nil {
|
||||||
|
crit.FromBlock = big.NewInt(rpc.LatestBlockNumber.Int64())
|
||||||
|
}
|
||||||
|
if crit.ToBlock == nil {
|
||||||
|
crit.ToBlock = big.NewInt(rpc.LatestBlockNumber.Int64())
|
||||||
|
}
|
||||||
|
|
||||||
|
filter := New(api.chainDb)
|
||||||
|
filter.SetBeginBlock(crit.FromBlock.Int64())
|
||||||
|
filter.SetEndBlock(crit.ToBlock.Int64())
|
||||||
|
filter.SetAddresses(crit.Addresses)
|
||||||
|
filter.SetTopics(crit.Topics)
|
||||||
|
|
||||||
|
return returnLogs(filter.Find())
|
||||||
|
}
|
||||||
|
|
||||||
|
// UninstallFilter removes the filter with the given filter id.
|
||||||
|
//
|
||||||
|
// https://github.com/ethereum/wiki/wiki/JSON-RPC#eth_uninstallfilter
|
||||||
|
func (api *PublicFilterAPI) UninstallFilter(id rpc.ID) bool {
|
||||||
|
api.filtersMu.Lock()
|
||||||
|
f, found := api.filters[id]
|
||||||
|
if found {
|
||||||
|
delete(api.filters, id)
|
||||||
|
}
|
||||||
|
api.filtersMu.Unlock()
|
||||||
|
if found {
|
||||||
|
f.s.Unsubscribe()
|
||||||
|
}
|
||||||
|
|
||||||
|
return found
|
||||||
|
}
|
||||||
|
|
||||||
|
// GetFilterLogs returns the logs for the filter with the given id.
|
||||||
|
// If the filter could not be found an empty array of logs is returned.
|
||||||
|
//
|
||||||
|
// https://github.com/ethereum/wiki/wiki/JSON-RPC#eth_getfilterlogs
|
||||||
|
func (api *PublicFilterAPI) GetFilterLogs(id rpc.ID) []Log {
|
||||||
|
api.filtersMu.Lock()
|
||||||
|
f, found := api.filters[id]
|
||||||
|
api.filtersMu.Unlock()
|
||||||
|
|
||||||
|
if !found || f.typ != LogsSubscription {
|
||||||
|
return []Log{}
|
||||||
|
}
|
||||||
|
|
||||||
|
filter := New(api.chainDb)
|
||||||
|
filter.SetBeginBlock(f.crit.FromBlock.Int64())
|
||||||
|
filter.SetEndBlock(f.crit.ToBlock.Int64())
|
||||||
|
filter.SetAddresses(f.crit.Addresses)
|
||||||
|
filter.SetTopics(f.crit.Topics)
|
||||||
|
|
||||||
|
return returnLogs(filter.Find())
|
||||||
|
}
|
||||||
|
|
||||||
|
// GetFilterChanges returns the logs for the filter with the given id since
|
||||||
|
// last time is was called. This can be used for polling.
|
||||||
|
//
|
||||||
|
// For pending transaction and block filters the result is []common.Hash.
|
||||||
|
// (pending)Log filters return []Log. If the filter could not be found
|
||||||
|
// []interface{}{} is returned.
|
||||||
|
//
|
||||||
|
// https://github.com/ethereum/wiki/wiki/JSON-RPC#eth_getfilterchanges
|
||||||
|
func (api *PublicFilterAPI) GetFilterChanges(id rpc.ID) interface{} {
|
||||||
|
api.filtersMu.Lock()
|
||||||
|
defer api.filtersMu.Unlock()
|
||||||
|
|
||||||
|
if f, found := api.filters[id]; found {
|
||||||
|
if !f.deadline.Stop() {
|
||||||
|
// timer expired but filter is not yet removed in timeout loop
|
||||||
|
// receive timer value and reset timer
|
||||||
|
<-f.deadline.C
|
||||||
|
}
|
||||||
|
f.deadline.Reset(deadline)
|
||||||
|
|
||||||
|
switch f.typ {
|
||||||
|
case PendingTransactionsSubscription, BlocksSubscription:
|
||||||
|
hashes := f.hashes
|
||||||
|
f.hashes = nil
|
||||||
|
return returnHashes(hashes)
|
||||||
|
case PendingLogsSubscription, LogsSubscription:
|
||||||
|
logs := f.logs
|
||||||
|
f.logs = nil
|
||||||
|
return returnLogs(logs)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return []interface{}{}
|
||||||
|
}
|
||||||
|
|
||||||
|
// returnHashes is a helper that will return an empty hash array case the given hash array is nil,
|
||||||
|
// otherwise the given hashes array is returned.
|
||||||
|
func returnHashes(hashes []common.Hash) []common.Hash {
|
||||||
|
if hashes == nil {
|
||||||
|
return []common.Hash{}
|
||||||
|
}
|
||||||
|
return hashes
|
||||||
|
}
|
||||||
|
|
||||||
|
// returnLogs is a helper that will return an empty log array in case the given logs array is nil,
|
||||||
|
// otherwise the given logs array is returned.
|
||||||
|
func returnLogs(logs []Log) []Log {
|
||||||
|
if logs == nil {
|
||||||
|
return []Log{}
|
||||||
|
}
|
||||||
|
return logs
|
||||||
|
}
|
||||||
|
|
||||||
// UnmarshalJSON sets *args fields with given data.
|
// UnmarshalJSON sets *args fields with given data.
|
||||||
func (args *NewFilterArgs) UnmarshalJSON(data []byte) error {
|
func (args *FilterCriteria) UnmarshalJSON(data []byte) error {
|
||||||
type input struct {
|
type input struct {
|
||||||
From *rpc.BlockNumber `json:"fromBlock"`
|
From *rpc.BlockNumber `json:"fromBlock"`
|
||||||
ToBlock *rpc.BlockNumber `json:"toBlock"`
|
ToBlock *rpc.BlockNumber `json:"toBlock"`
|
||||||
@ -316,15 +438,15 @@ func (args *NewFilterArgs) UnmarshalJSON(data []byte) error {
|
|||||||
}
|
}
|
||||||
|
|
||||||
if raw.From == nil || raw.From.Int64() < 0 {
|
if raw.From == nil || raw.From.Int64() < 0 {
|
||||||
args.FromBlock = rpc.LatestBlockNumber
|
args.FromBlock = big.NewInt(rpc.LatestBlockNumber.Int64())
|
||||||
} else {
|
} else {
|
||||||
args.FromBlock = *raw.From
|
args.FromBlock = big.NewInt(raw.From.Int64())
|
||||||
}
|
}
|
||||||
|
|
||||||
if raw.ToBlock == nil || raw.ToBlock.Int64() < 0 {
|
if raw.ToBlock == nil || raw.ToBlock.Int64() < 0 {
|
||||||
args.ToBlock = rpc.LatestBlockNumber
|
args.ToBlock = big.NewInt(rpc.LatestBlockNumber.Int64())
|
||||||
} else {
|
} else {
|
||||||
args.ToBlock = *raw.ToBlock
|
args.ToBlock = big.NewInt(raw.ToBlock.Int64())
|
||||||
}
|
}
|
||||||
|
|
||||||
args.Addresses = []common.Address{}
|
args.Addresses = []common.Address{}
|
||||||
@ -414,255 +536,3 @@ func (args *NewFilterArgs) UnmarshalJSON(data []byte) error {
|
|||||||
|
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// NewFilter creates a new filter and returns the filter id. It can be uses to retrieve logs.
|
|
||||||
func (s *PublicFilterAPI) NewFilter(args NewFilterArgs) (string, error) {
|
|
||||||
externalId, err := newFilterId()
|
|
||||||
if err != nil {
|
|
||||||
return "", err
|
|
||||||
}
|
|
||||||
|
|
||||||
var id int
|
|
||||||
if len(args.Addresses) > 0 {
|
|
||||||
id, err = s.newLogFilter(args.FromBlock.Int64(), args.ToBlock.Int64(), args.Addresses, args.Topics, nil)
|
|
||||||
} else {
|
|
||||||
id, err = s.newLogFilter(args.FromBlock.Int64(), args.ToBlock.Int64(), nil, args.Topics, nil)
|
|
||||||
}
|
|
||||||
if err != nil {
|
|
||||||
return "", err
|
|
||||||
}
|
|
||||||
|
|
||||||
s.filterMapMu.Lock()
|
|
||||||
s.filterMapping[externalId] = id
|
|
||||||
s.filterMapMu.Unlock()
|
|
||||||
|
|
||||||
return externalId, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// GetLogs returns the logs matching the given argument.
|
|
||||||
func (s *PublicFilterAPI) GetLogs(args NewFilterArgs) []vmlog {
|
|
||||||
filter := New(s.chainDb)
|
|
||||||
filter.SetBeginBlock(args.FromBlock.Int64())
|
|
||||||
filter.SetEndBlock(args.ToBlock.Int64())
|
|
||||||
filter.SetAddresses(args.Addresses)
|
|
||||||
filter.SetTopics(args.Topics)
|
|
||||||
|
|
||||||
return toRPCLogs(filter.Find(), false)
|
|
||||||
}
|
|
||||||
|
|
||||||
// UninstallFilter removes the filter with the given filter id.
|
|
||||||
func (s *PublicFilterAPI) UninstallFilter(filterId string) bool {
|
|
||||||
s.filterManager.Lock()
|
|
||||||
defer s.filterManager.Unlock()
|
|
||||||
|
|
||||||
s.filterMapMu.Lock()
|
|
||||||
id, ok := s.filterMapping[filterId]
|
|
||||||
if !ok {
|
|
||||||
s.filterMapMu.Unlock()
|
|
||||||
return false
|
|
||||||
}
|
|
||||||
delete(s.filterMapping, filterId)
|
|
||||||
s.filterMapMu.Unlock()
|
|
||||||
|
|
||||||
s.filterManager.Remove(id)
|
|
||||||
|
|
||||||
s.logMu.Lock()
|
|
||||||
if _, ok := s.logQueue[id]; ok {
|
|
||||||
delete(s.logQueue, id)
|
|
||||||
s.logMu.Unlock()
|
|
||||||
return true
|
|
||||||
}
|
|
||||||
s.logMu.Unlock()
|
|
||||||
|
|
||||||
s.blockMu.Lock()
|
|
||||||
if _, ok := s.blockQueue[id]; ok {
|
|
||||||
delete(s.blockQueue, id)
|
|
||||||
s.blockMu.Unlock()
|
|
||||||
return true
|
|
||||||
}
|
|
||||||
s.blockMu.Unlock()
|
|
||||||
|
|
||||||
s.transactionMu.Lock()
|
|
||||||
if _, ok := s.transactionQueue[id]; ok {
|
|
||||||
delete(s.transactionQueue, id)
|
|
||||||
s.transactionMu.Unlock()
|
|
||||||
return true
|
|
||||||
}
|
|
||||||
s.transactionMu.Unlock()
|
|
||||||
|
|
||||||
return false
|
|
||||||
}
|
|
||||||
|
|
||||||
// getFilterType is a helper utility that determine the type of filter for the given filter id.
|
|
||||||
func (s *PublicFilterAPI) getFilterType(id int) byte {
|
|
||||||
if _, ok := s.blockQueue[id]; ok {
|
|
||||||
return blockFilterTy
|
|
||||||
} else if _, ok := s.transactionQueue[id]; ok {
|
|
||||||
return transactionFilterTy
|
|
||||||
} else if _, ok := s.logQueue[id]; ok {
|
|
||||||
return logFilterTy
|
|
||||||
}
|
|
||||||
|
|
||||||
return unknownFilterTy
|
|
||||||
}
|
|
||||||
|
|
||||||
// blockFilterChanged returns a collection of block hashes for the block filter with the given id.
|
|
||||||
func (s *PublicFilterAPI) blockFilterChanged(id int) []common.Hash {
|
|
||||||
s.blockMu.Lock()
|
|
||||||
defer s.blockMu.Unlock()
|
|
||||||
|
|
||||||
if s.blockQueue[id] != nil {
|
|
||||||
return s.blockQueue[id].get()
|
|
||||||
}
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// transactionFilterChanged returns a collection of transaction hashes for the pending
|
|
||||||
// transaction filter with the given id.
|
|
||||||
func (s *PublicFilterAPI) transactionFilterChanged(id int) []common.Hash {
|
|
||||||
s.blockMu.Lock()
|
|
||||||
defer s.blockMu.Unlock()
|
|
||||||
|
|
||||||
if s.transactionQueue[id] != nil {
|
|
||||||
return s.transactionQueue[id].get()
|
|
||||||
}
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// logFilterChanged returns a collection of logs for the log filter with the given id.
|
|
||||||
func (s *PublicFilterAPI) logFilterChanged(id int) []vmlog {
|
|
||||||
s.logMu.Lock()
|
|
||||||
defer s.logMu.Unlock()
|
|
||||||
|
|
||||||
if s.logQueue[id] != nil {
|
|
||||||
return s.logQueue[id].get()
|
|
||||||
}
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// GetFilterLogs returns the logs for the filter with the given id.
|
|
||||||
func (s *PublicFilterAPI) GetFilterLogs(filterId string) []vmlog {
|
|
||||||
s.filterMapMu.RLock()
|
|
||||||
id, ok := s.filterMapping[filterId]
|
|
||||||
s.filterMapMu.RUnlock()
|
|
||||||
if !ok {
|
|
||||||
return toRPCLogs(nil, false)
|
|
||||||
}
|
|
||||||
|
|
||||||
if filter := s.filterManager.Get(id); filter != nil {
|
|
||||||
return toRPCLogs(filter.Find(), false)
|
|
||||||
}
|
|
||||||
|
|
||||||
return toRPCLogs(nil, false)
|
|
||||||
}
|
|
||||||
|
|
||||||
// GetFilterChanges returns the logs for the filter with the given id since last time is was called.
|
|
||||||
// This can be used for polling.
|
|
||||||
func (s *PublicFilterAPI) GetFilterChanges(filterId string) interface{} {
|
|
||||||
s.filterMapMu.RLock()
|
|
||||||
id, ok := s.filterMapping[filterId]
|
|
||||||
s.filterMapMu.RUnlock()
|
|
||||||
|
|
||||||
if !ok { // filter not found
|
|
||||||
return []interface{}{}
|
|
||||||
}
|
|
||||||
|
|
||||||
switch s.getFilterType(id) {
|
|
||||||
case blockFilterTy:
|
|
||||||
return returnHashes(s.blockFilterChanged(id))
|
|
||||||
case transactionFilterTy:
|
|
||||||
return returnHashes(s.transactionFilterChanged(id))
|
|
||||||
case logFilterTy:
|
|
||||||
return s.logFilterChanged(id)
|
|
||||||
}
|
|
||||||
|
|
||||||
return []interface{}{}
|
|
||||||
}
|
|
||||||
|
|
||||||
type vmlog struct {
|
|
||||||
*vm.Log
|
|
||||||
Removed bool `json:"removed"`
|
|
||||||
}
|
|
||||||
|
|
||||||
type logQueue struct {
|
|
||||||
mu sync.Mutex
|
|
||||||
|
|
||||||
logs []vmlog
|
|
||||||
timeout time.Time
|
|
||||||
id int
|
|
||||||
}
|
|
||||||
|
|
||||||
func (l *logQueue) add(logs ...vmlog) {
|
|
||||||
l.mu.Lock()
|
|
||||||
defer l.mu.Unlock()
|
|
||||||
|
|
||||||
l.logs = append(l.logs, logs...)
|
|
||||||
}
|
|
||||||
|
|
||||||
func (l *logQueue) get() []vmlog {
|
|
||||||
l.mu.Lock()
|
|
||||||
defer l.mu.Unlock()
|
|
||||||
|
|
||||||
l.timeout = time.Now()
|
|
||||||
tmp := l.logs
|
|
||||||
l.logs = nil
|
|
||||||
return tmp
|
|
||||||
}
|
|
||||||
|
|
||||||
type hashQueue struct {
|
|
||||||
mu sync.Mutex
|
|
||||||
|
|
||||||
hashes []common.Hash
|
|
||||||
timeout time.Time
|
|
||||||
id int
|
|
||||||
}
|
|
||||||
|
|
||||||
func (l *hashQueue) add(hashes ...common.Hash) {
|
|
||||||
l.mu.Lock()
|
|
||||||
defer l.mu.Unlock()
|
|
||||||
|
|
||||||
l.hashes = append(l.hashes, hashes...)
|
|
||||||
}
|
|
||||||
|
|
||||||
func (l *hashQueue) get() []common.Hash {
|
|
||||||
l.mu.Lock()
|
|
||||||
defer l.mu.Unlock()
|
|
||||||
|
|
||||||
l.timeout = time.Now()
|
|
||||||
tmp := l.hashes
|
|
||||||
l.hashes = nil
|
|
||||||
return tmp
|
|
||||||
}
|
|
||||||
|
|
||||||
// newFilterId generates a new random filter identifier that can be exposed to the outer world. By publishing random
|
|
||||||
// identifiers it is not feasible for DApp's to guess filter id's for other DApp's and uninstall or poll for them
|
|
||||||
// causing the affected DApp to miss data.
|
|
||||||
func newFilterId() (string, error) {
|
|
||||||
var subid [16]byte
|
|
||||||
n, _ := rand.Read(subid[:])
|
|
||||||
if n != 16 {
|
|
||||||
return "", errors.New("Unable to generate filter id")
|
|
||||||
}
|
|
||||||
return "0x" + hex.EncodeToString(subid[:]), nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// toRPCLogs is a helper that will convert a vm.Logs array to an structure which
|
|
||||||
// can hold additional information about the logs such as whether it was deleted.
|
|
||||||
// Additionally when nil is given it will by default instead create an empty slice
|
|
||||||
// instead. This is required by the RPC specification.
|
|
||||||
func toRPCLogs(logs vm.Logs, removed bool) []vmlog {
|
|
||||||
convertedLogs := make([]vmlog, len(logs))
|
|
||||||
for i, log := range logs {
|
|
||||||
convertedLogs[i] = vmlog{Log: log, Removed: removed}
|
|
||||||
}
|
|
||||||
return convertedLogs
|
|
||||||
}
|
|
||||||
|
|
||||||
// returnHashes is a helper that will return an empty hash array case the given hash array is nil, otherwise is will
|
|
||||||
// return the given hashes. The RPC interfaces defines that always an array is returned.
|
|
||||||
func returnHashes(hashes []common.Hash) []common.Hash {
|
|
||||||
if hashes == nil {
|
|
||||||
return []common.Hash{}
|
|
||||||
}
|
|
||||||
return hashes
|
|
||||||
}
|
|
||||||
|
@ -14,7 +14,7 @@
|
|||||||
// You should have received a copy of the GNU Lesser General Public License
|
// You should have received a copy of the GNU Lesser General Public License
|
||||||
// along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
|
// along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
|
||||||
|
|
||||||
package filters_test
|
package filters
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
@ -22,7 +22,6 @@ import (
|
|||||||
"testing"
|
"testing"
|
||||||
|
|
||||||
"github.com/ethereum/go-ethereum/common"
|
"github.com/ethereum/go-ethereum/common"
|
||||||
"github.com/ethereum/go-ethereum/eth/filters"
|
|
||||||
"github.com/ethereum/go-ethereum/rpc"
|
"github.com/ethereum/go-ethereum/rpc"
|
||||||
)
|
)
|
||||||
|
|
||||||
@ -39,14 +38,14 @@ func TestUnmarshalJSONNewFilterArgs(t *testing.T) {
|
|||||||
)
|
)
|
||||||
|
|
||||||
// default values
|
// default values
|
||||||
var test0 filters.NewFilterArgs
|
var test0 FilterCriteria
|
||||||
if err := json.Unmarshal([]byte("{}"), &test0); err != nil {
|
if err := json.Unmarshal([]byte("{}"), &test0); err != nil {
|
||||||
t.Fatal(err)
|
t.Fatal(err)
|
||||||
}
|
}
|
||||||
if test0.FromBlock != rpc.LatestBlockNumber {
|
if test0.FromBlock.Int64() != rpc.LatestBlockNumber.Int64() {
|
||||||
t.Fatalf("expected %d, got %d", rpc.LatestBlockNumber, test0.FromBlock)
|
t.Fatalf("expected %d, got %d", rpc.LatestBlockNumber, test0.FromBlock)
|
||||||
}
|
}
|
||||||
if test0.ToBlock != rpc.LatestBlockNumber {
|
if test0.ToBlock.Int64() != rpc.LatestBlockNumber.Int64() {
|
||||||
t.Fatalf("expected %d, got %d", rpc.LatestBlockNumber, test0.ToBlock)
|
t.Fatalf("expected %d, got %d", rpc.LatestBlockNumber, test0.ToBlock)
|
||||||
}
|
}
|
||||||
if len(test0.Addresses) != 0 {
|
if len(test0.Addresses) != 0 {
|
||||||
@ -57,20 +56,20 @@ func TestUnmarshalJSONNewFilterArgs(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// from, to block number
|
// from, to block number
|
||||||
var test1 filters.NewFilterArgs
|
var test1 FilterCriteria
|
||||||
vector := fmt.Sprintf(`{"fromBlock":"0x%x","toBlock":"0x%x"}`, fromBlock, toBlock)
|
vector := fmt.Sprintf(`{"fromBlock":"0x%x","toBlock":"0x%x"}`, fromBlock, toBlock)
|
||||||
if err := json.Unmarshal([]byte(vector), &test1); err != nil {
|
if err := json.Unmarshal([]byte(vector), &test1); err != nil {
|
||||||
t.Fatal(err)
|
t.Fatal(err)
|
||||||
}
|
}
|
||||||
if test1.FromBlock != fromBlock {
|
if test1.FromBlock.Int64() != fromBlock.Int64() {
|
||||||
t.Fatalf("expected FromBlock %d, got %d", fromBlock, test1.FromBlock)
|
t.Fatalf("expected FromBlock %d, got %d", fromBlock, test1.FromBlock)
|
||||||
}
|
}
|
||||||
if test1.ToBlock != toBlock {
|
if test1.ToBlock.Int64() != toBlock.Int64() {
|
||||||
t.Fatalf("expected ToBlock %d, got %d", toBlock, test1.ToBlock)
|
t.Fatalf("expected ToBlock %d, got %d", toBlock, test1.ToBlock)
|
||||||
}
|
}
|
||||||
|
|
||||||
// single address
|
// single address
|
||||||
var test2 filters.NewFilterArgs
|
var test2 FilterCriteria
|
||||||
vector = fmt.Sprintf(`{"address": "%s"}`, address0.Hex())
|
vector = fmt.Sprintf(`{"address": "%s"}`, address0.Hex())
|
||||||
if err := json.Unmarshal([]byte(vector), &test2); err != nil {
|
if err := json.Unmarshal([]byte(vector), &test2); err != nil {
|
||||||
t.Fatal(err)
|
t.Fatal(err)
|
||||||
@ -83,7 +82,7 @@ func TestUnmarshalJSONNewFilterArgs(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// multiple address
|
// multiple address
|
||||||
var test3 filters.NewFilterArgs
|
var test3 FilterCriteria
|
||||||
vector = fmt.Sprintf(`{"address": ["%s", "%s"]}`, address0.Hex(), address1.Hex())
|
vector = fmt.Sprintf(`{"address": ["%s", "%s"]}`, address0.Hex(), address1.Hex())
|
||||||
if err := json.Unmarshal([]byte(vector), &test3); err != nil {
|
if err := json.Unmarshal([]byte(vector), &test3); err != nil {
|
||||||
t.Fatal(err)
|
t.Fatal(err)
|
||||||
@ -99,7 +98,7 @@ func TestUnmarshalJSONNewFilterArgs(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// single topic
|
// single topic
|
||||||
var test4 filters.NewFilterArgs
|
var test4 FilterCriteria
|
||||||
vector = fmt.Sprintf(`{"topics": ["%s"]}`, topic0.Hex())
|
vector = fmt.Sprintf(`{"topics": ["%s"]}`, topic0.Hex())
|
||||||
if err := json.Unmarshal([]byte(vector), &test4); err != nil {
|
if err := json.Unmarshal([]byte(vector), &test4); err != nil {
|
||||||
t.Fatal(err)
|
t.Fatal(err)
|
||||||
@ -115,7 +114,7 @@ func TestUnmarshalJSONNewFilterArgs(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// test multiple "AND" topics
|
// test multiple "AND" topics
|
||||||
var test5 filters.NewFilterArgs
|
var test5 FilterCriteria
|
||||||
vector = fmt.Sprintf(`{"topics": ["%s", "%s"]}`, topic0.Hex(), topic1.Hex())
|
vector = fmt.Sprintf(`{"topics": ["%s", "%s"]}`, topic0.Hex(), topic1.Hex())
|
||||||
if err := json.Unmarshal([]byte(vector), &test5); err != nil {
|
if err := json.Unmarshal([]byte(vector), &test5); err != nil {
|
||||||
t.Fatal(err)
|
t.Fatal(err)
|
||||||
@ -137,7 +136,7 @@ func TestUnmarshalJSONNewFilterArgs(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// test optional topic
|
// test optional topic
|
||||||
var test6 filters.NewFilterArgs
|
var test6 FilterCriteria
|
||||||
vector = fmt.Sprintf(`{"topics": ["%s", null, "%s"]}`, topic0.Hex(), topic2.Hex())
|
vector = fmt.Sprintf(`{"topics": ["%s", null, "%s"]}`, topic0.Hex(), topic2.Hex())
|
||||||
if err := json.Unmarshal([]byte(vector), &test6); err != nil {
|
if err := json.Unmarshal([]byte(vector), &test6); err != nil {
|
||||||
t.Fatal(err)
|
t.Fatal(err)
|
||||||
@ -165,7 +164,7 @@ func TestUnmarshalJSONNewFilterArgs(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// test OR topics
|
// test OR topics
|
||||||
var test7 filters.NewFilterArgs
|
var test7 FilterCriteria
|
||||||
vector = fmt.Sprintf(`{"topics": [["%s", "%s"], null, ["%s", null]]}`, topic0.Hex(), topic1.Hex(), topic2.Hex())
|
vector = fmt.Sprintf(`{"topics": [["%s", "%s"], null, ["%s", null]]}`, topic0.Hex(), topic1.Hex(), topic2.Hex())
|
||||||
if err := json.Unmarshal([]byte(vector), &test7); err != nil {
|
if err := json.Unmarshal([]byte(vector), &test7); err != nil {
|
||||||
t.Fatal(err)
|
t.Fatal(err)
|
||||||
|
@ -23,15 +23,10 @@ import (
|
|||||||
"github.com/ethereum/go-ethereum/common"
|
"github.com/ethereum/go-ethereum/common"
|
||||||
"github.com/ethereum/go-ethereum/core"
|
"github.com/ethereum/go-ethereum/core"
|
||||||
"github.com/ethereum/go-ethereum/core/types"
|
"github.com/ethereum/go-ethereum/core/types"
|
||||||
"github.com/ethereum/go-ethereum/core/vm"
|
|
||||||
"github.com/ethereum/go-ethereum/ethdb"
|
"github.com/ethereum/go-ethereum/ethdb"
|
||||||
)
|
)
|
||||||
|
|
||||||
type AccountChange struct {
|
// Filter can be used to retrieve and filter logs
|
||||||
Address, StateAddress []byte
|
|
||||||
}
|
|
||||||
|
|
||||||
// Filtering interface
|
|
||||||
type Filter struct {
|
type Filter struct {
|
||||||
created time.Time
|
created time.Time
|
||||||
|
|
||||||
@ -39,70 +34,72 @@ type Filter struct {
|
|||||||
begin, end int64
|
begin, end int64
|
||||||
addresses []common.Address
|
addresses []common.Address
|
||||||
topics [][]common.Hash
|
topics [][]common.Hash
|
||||||
|
|
||||||
BlockCallback func(*types.Block, vm.Logs)
|
|
||||||
TransactionCallback func(*types.Transaction)
|
|
||||||
LogCallback func(*vm.Log, bool)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// Create a new filter which uses a bloom filter on blocks to figure out whether a particular block
|
// New creates a new filter which uses a bloom filter on blocks to figure out whether
|
||||||
// is interesting or not.
|
// a particular block is interesting or not.
|
||||||
func New(db ethdb.Database) *Filter {
|
func New(db ethdb.Database) *Filter {
|
||||||
return &Filter{db: db}
|
return &Filter{db: db}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Set the earliest and latest block for filtering.
|
// SetBeginBlock sets the earliest block for filtering.
|
||||||
// -1 = latest block (i.e., the current block)
|
// -1 = latest block (i.e., the current block)
|
||||||
// hash = particular hash from-to
|
// hash = particular hash from-to
|
||||||
func (self *Filter) SetBeginBlock(begin int64) {
|
func (f *Filter) SetBeginBlock(begin int64) {
|
||||||
self.begin = begin
|
f.begin = begin
|
||||||
}
|
}
|
||||||
|
|
||||||
func (self *Filter) SetEndBlock(end int64) {
|
// SetEndBlock sets the latest block for filtering.
|
||||||
self.end = end
|
func (f *Filter) SetEndBlock(end int64) {
|
||||||
|
f.end = end
|
||||||
}
|
}
|
||||||
|
|
||||||
func (self *Filter) SetAddresses(addr []common.Address) {
|
// SetAddresses matches only logs that are generated from addresses that are included
|
||||||
self.addresses = addr
|
// in the given addresses.
|
||||||
|
func (f *Filter) SetAddresses(addr []common.Address) {
|
||||||
|
f.addresses = addr
|
||||||
}
|
}
|
||||||
|
|
||||||
func (self *Filter) SetTopics(topics [][]common.Hash) {
|
// SetTopics matches only logs that have topics matching the given topics.
|
||||||
self.topics = topics
|
func (f *Filter) SetTopics(topics [][]common.Hash) {
|
||||||
|
f.topics = topics
|
||||||
}
|
}
|
||||||
|
|
||||||
// Run filters logs with the current parameters set
|
// Run filters logs with the current parameters set
|
||||||
func (self *Filter) Find() vm.Logs {
|
func (f *Filter) Find() []Log {
|
||||||
latestHash := core.GetHeadBlockHash(self.db)
|
latestHash := core.GetHeadBlockHash(f.db)
|
||||||
latestBlock := core.GetBlock(self.db, latestHash, core.GetBlockNumber(self.db, latestHash))
|
latestBlock := core.GetBlock(f.db, latestHash, core.GetBlockNumber(f.db, latestHash))
|
||||||
if latestBlock == nil {
|
if latestBlock == nil {
|
||||||
return vm.Logs{}
|
return []Log{}
|
||||||
}
|
}
|
||||||
var beginBlockNo uint64 = uint64(self.begin)
|
|
||||||
if self.begin == -1 {
|
var beginBlockNo uint64 = uint64(f.begin)
|
||||||
|
if f.begin == -1 {
|
||||||
beginBlockNo = latestBlock.NumberU64()
|
beginBlockNo = latestBlock.NumberU64()
|
||||||
}
|
}
|
||||||
var endBlockNo uint64 = uint64(self.end)
|
|
||||||
if self.end == -1 {
|
endBlockNo := uint64(f.end)
|
||||||
|
if f.end == -1 {
|
||||||
endBlockNo = latestBlock.NumberU64()
|
endBlockNo = latestBlock.NumberU64()
|
||||||
}
|
}
|
||||||
|
|
||||||
// if no addresses are present we can't make use of fast search which
|
// if no addresses are present we can't make use of fast search which
|
||||||
// uses the mipmap bloom filters to check for fast inclusion and uses
|
// uses the mipmap bloom filters to check for fast inclusion and uses
|
||||||
// higher range probability in order to ensure at least a false positive
|
// higher range probability in order to ensure at least a false positive
|
||||||
if len(self.addresses) == 0 {
|
if len(f.addresses) == 0 {
|
||||||
return self.getLogs(beginBlockNo, endBlockNo)
|
return f.getLogs(beginBlockNo, endBlockNo)
|
||||||
}
|
}
|
||||||
return self.mipFind(beginBlockNo, endBlockNo, 0)
|
return f.mipFind(beginBlockNo, endBlockNo, 0)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (self *Filter) mipFind(start, end uint64, depth int) (logs vm.Logs) {
|
func (f *Filter) mipFind(start, end uint64, depth int) (logs []Log) {
|
||||||
level := core.MIPMapLevels[depth]
|
level := core.MIPMapLevels[depth]
|
||||||
// normalise numerator so we can work in level specific batches and
|
// normalise numerator so we can work in level specific batches and
|
||||||
// work with the proper range checks
|
// work with the proper range checks
|
||||||
for num := start / level * level; num <= end; num += level {
|
for num := start / level * level; num <= end; num += level {
|
||||||
// find addresses in bloom filters
|
// find addresses in bloom filters
|
||||||
bloom := core.GetMipmapBloom(self.db, num, level)
|
bloom := core.GetMipmapBloom(f.db, num, level)
|
||||||
for _, addr := range self.addresses {
|
for _, addr := range f.addresses {
|
||||||
if bloom.TestBytes(addr[:]) {
|
if bloom.TestBytes(addr[:]) {
|
||||||
// range check normalised values and make sure that
|
// range check normalised values and make sure that
|
||||||
// we're resolving the correct range instead of the
|
// we're resolving the correct range instead of the
|
||||||
@ -110,9 +107,9 @@ func (self *Filter) mipFind(start, end uint64, depth int) (logs vm.Logs) {
|
|||||||
start := uint64(math.Max(float64(num), float64(start)))
|
start := uint64(math.Max(float64(num), float64(start)))
|
||||||
end := uint64(math.Min(float64(num+level-1), float64(end)))
|
end := uint64(math.Min(float64(num+level-1), float64(end)))
|
||||||
if depth+1 == len(core.MIPMapLevels) {
|
if depth+1 == len(core.MIPMapLevels) {
|
||||||
logs = append(logs, self.getLogs(start, end)...)
|
logs = append(logs, f.getLogs(start, end)...)
|
||||||
} else {
|
} else {
|
||||||
logs = append(logs, self.mipFind(start, end, depth+1)...)
|
logs = append(logs, f.mipFind(start, end, depth+1)...)
|
||||||
}
|
}
|
||||||
// break so we don't check the same range for each
|
// break so we don't check the same range for each
|
||||||
// possible address. Checks on multiple addresses
|
// possible address. Checks on multiple addresses
|
||||||
@ -125,12 +122,15 @@ func (self *Filter) mipFind(start, end uint64, depth int) (logs vm.Logs) {
|
|||||||
return logs
|
return logs
|
||||||
}
|
}
|
||||||
|
|
||||||
func (self *Filter) getLogs(start, end uint64) (logs vm.Logs) {
|
func (f *Filter) getLogs(start, end uint64) (logs []Log) {
|
||||||
for i := start; i <= end; i++ {
|
|
||||||
var block *types.Block
|
var block *types.Block
|
||||||
hash := core.GetCanonicalHash(self.db, i)
|
|
||||||
|
for i := start; i <= end; i++ {
|
||||||
|
hash := core.GetCanonicalHash(f.db, i)
|
||||||
if hash != (common.Hash{}) {
|
if hash != (common.Hash{}) {
|
||||||
block = core.GetBlock(self.db, hash, i)
|
block = core.GetBlock(f.db, hash, i)
|
||||||
|
} else { // block not found
|
||||||
|
return logs
|
||||||
}
|
}
|
||||||
if block == nil { // block not found/written
|
if block == nil { // block not found/written
|
||||||
return logs
|
return logs
|
||||||
@ -138,16 +138,20 @@ func (self *Filter) getLogs(start, end uint64) (logs vm.Logs) {
|
|||||||
|
|
||||||
// Use bloom filtering to see if this block is interesting given the
|
// Use bloom filtering to see if this block is interesting given the
|
||||||
// current parameters
|
// current parameters
|
||||||
if self.bloomFilter(block) {
|
if f.bloomFilter(block) {
|
||||||
// Get the logs of the block
|
// Get the logs of the block
|
||||||
var (
|
var (
|
||||||
receipts = core.GetBlockReceipts(self.db, block.Hash(), i)
|
receipts = core.GetBlockReceipts(f.db, block.Hash(), i)
|
||||||
unfiltered vm.Logs
|
unfiltered []Log
|
||||||
)
|
)
|
||||||
for _, receipt := range receipts {
|
for _, receipt := range receipts {
|
||||||
unfiltered = append(unfiltered, receipt.Logs...)
|
rl := make([]Log, len(receipt.Logs))
|
||||||
|
for i, l := range receipt.Logs {
|
||||||
|
rl[i] = Log{l, false}
|
||||||
}
|
}
|
||||||
logs = append(logs, self.FilterLogs(unfiltered)...)
|
unfiltered = append(unfiltered, rl...)
|
||||||
|
}
|
||||||
|
logs = append(logs, filterLogs(unfiltered, f.addresses, f.topics)...)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -164,26 +168,25 @@ func includes(addresses []common.Address, a common.Address) bool {
|
|||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
|
|
||||||
func (self *Filter) FilterLogs(logs vm.Logs) vm.Logs {
|
func filterLogs(logs []Log, addresses []common.Address, topics [][]common.Hash) []Log {
|
||||||
var ret vm.Logs
|
var ret []Log
|
||||||
|
|
||||||
// Filter the logs for interesting stuff
|
// Filter the logs for interesting stuff
|
||||||
Logs:
|
Logs:
|
||||||
for _, log := range logs {
|
for _, log := range logs {
|
||||||
if len(self.addresses) > 0 && !includes(self.addresses, log.Address) {
|
if len(addresses) > 0 && !includes(addresses, log.Address) {
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
|
|
||||||
logTopics := make([]common.Hash, len(self.topics))
|
logTopics := make([]common.Hash, len(topics))
|
||||||
copy(logTopics, log.Topics)
|
copy(logTopics, log.Topics)
|
||||||
|
|
||||||
// If the to filtered topics is greater than the amount of topics in
|
// If the to filtered topics is greater than the amount of topics in logs, skip.
|
||||||
// logs, skip.
|
if len(topics) > len(log.Topics) {
|
||||||
if len(self.topics) > len(log.Topics) {
|
|
||||||
continue Logs
|
continue Logs
|
||||||
}
|
}
|
||||||
|
|
||||||
for i, topics := range self.topics {
|
for i, topics := range topics {
|
||||||
var match bool
|
var match bool
|
||||||
for _, topic := range topics {
|
for _, topic := range topics {
|
||||||
// common.Hash{} is a match all (wildcard)
|
// common.Hash{} is a match all (wildcard)
|
||||||
@ -196,7 +199,6 @@ Logs:
|
|||||||
if !match {
|
if !match {
|
||||||
continue Logs
|
continue Logs
|
||||||
}
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
ret = append(ret, log)
|
ret = append(ret, log)
|
||||||
@ -205,10 +207,10 @@ Logs:
|
|||||||
return ret
|
return ret
|
||||||
}
|
}
|
||||||
|
|
||||||
func (self *Filter) bloomFilter(block *types.Block) bool {
|
func (f *Filter) bloomFilter(block *types.Block) bool {
|
||||||
if len(self.addresses) > 0 {
|
if len(f.addresses) > 0 {
|
||||||
var included bool
|
var included bool
|
||||||
for _, addr := range self.addresses {
|
for _, addr := range f.addresses {
|
||||||
if types.BloomLookup(block.Bloom(), addr) {
|
if types.BloomLookup(block.Bloom(), addr) {
|
||||||
included = true
|
included = true
|
||||||
break
|
break
|
||||||
@ -220,7 +222,7 @@ func (self *Filter) bloomFilter(block *types.Block) bool {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
for _, sub := range self.topics {
|
for _, sub := range f.topics {
|
||||||
var included bool
|
var included bool
|
||||||
for _, topic := range sub {
|
for _, topic := range sub {
|
||||||
if (topic == common.Hash{}) || types.BloomLookup(block.Bloom(), topic) {
|
if (topic == common.Hash{}) || types.BloomLookup(block.Bloom(), topic) {
|
||||||
|
@ -14,179 +14,305 @@
|
|||||||
// You should have received a copy of the GNU Lesser General Public License
|
// You should have received a copy of the GNU Lesser General Public License
|
||||||
// along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
|
// along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
|
||||||
|
|
||||||
// package filters implements an ethereum filtering system for block,
|
// Package filters implements an ethereum filtering system for block,
|
||||||
// transactions and log events.
|
// transactions and log events.
|
||||||
package filters
|
package filters
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"encoding/json"
|
||||||
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
"sync"
|
"sync"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
|
"github.com/ethereum/go-ethereum/common"
|
||||||
"github.com/ethereum/go-ethereum/core"
|
"github.com/ethereum/go-ethereum/core"
|
||||||
|
"github.com/ethereum/go-ethereum/core/types"
|
||||||
"github.com/ethereum/go-ethereum/core/vm"
|
"github.com/ethereum/go-ethereum/core/vm"
|
||||||
"github.com/ethereum/go-ethereum/event"
|
"github.com/ethereum/go-ethereum/event"
|
||||||
|
"github.com/ethereum/go-ethereum/rpc"
|
||||||
)
|
)
|
||||||
|
|
||||||
// FilterType determines the type of filter and is used to put the filter in to
|
// Type determines the kind of filter and is used to put the filter in to
|
||||||
// the correct bucket when added.
|
// the correct bucket when added.
|
||||||
type FilterType byte
|
type Type byte
|
||||||
|
|
||||||
const (
|
const (
|
||||||
ChainFilter FilterType = iota // new block events filter
|
// UnknownSubscription indicates an unkown subscription type
|
||||||
PendingTxFilter // pending transaction filter
|
UnknownSubscription Type = iota
|
||||||
LogFilter // new or removed log filter
|
// LogsSubscription queries for new or removed (chain reorg) logs
|
||||||
PendingLogFilter // pending log filter
|
LogsSubscription
|
||||||
|
// PendingLogsSubscription queries for logs for the pending block
|
||||||
|
PendingLogsSubscription
|
||||||
|
// PendingTransactionsSubscription queries tx hashes for pending
|
||||||
|
// transactions entering the pending state
|
||||||
|
PendingTransactionsSubscription
|
||||||
|
// BlocksSubscription queries hashes for blocks that are imported
|
||||||
|
BlocksSubscription
|
||||||
)
|
)
|
||||||
|
|
||||||
// FilterSystem manages filters that filter specific events such as
|
var (
|
||||||
// block, transaction and log events. The Filtering system can be used to listen
|
ErrInvalidSubscriptionID = errors.New("invalid id")
|
||||||
// for specific LOG events fired by the EVM (Ethereum Virtual Machine).
|
)
|
||||||
type FilterSystem struct {
|
|
||||||
filterMu sync.RWMutex
|
|
||||||
filterId int
|
|
||||||
|
|
||||||
chainFilters map[int]*Filter
|
// Log is a helper that can hold additional information about vm.Log
|
||||||
pendingTxFilters map[int]*Filter
|
// necessary for the RPC interface.
|
||||||
logFilters map[int]*Filter
|
type Log struct {
|
||||||
pendingLogFilters map[int]*Filter
|
*vm.Log
|
||||||
|
Removed bool `json:"removed"`
|
||||||
|
}
|
||||||
|
|
||||||
// generic is an ugly hack for Get
|
func (l *Log) MarshalJSON() ([]byte, error) {
|
||||||
generic map[int]*Filter
|
fields := map[string]interface{}{
|
||||||
|
"address": l.Address,
|
||||||
|
"data": fmt.Sprintf("0x%x", l.Data),
|
||||||
|
"blockNumber": fmt.Sprintf("%#x", l.BlockNumber),
|
||||||
|
"logIndex": fmt.Sprintf("%#x", l.Index),
|
||||||
|
"blockHash": l.BlockHash,
|
||||||
|
"transactionHash": l.TxHash,
|
||||||
|
"transactionIndex": fmt.Sprintf("%#x", l.TxIndex),
|
||||||
|
"topics": l.Topics,
|
||||||
|
"removed": l.Removed,
|
||||||
|
}
|
||||||
|
|
||||||
|
return json.Marshal(fields)
|
||||||
|
}
|
||||||
|
|
||||||
|
type subscription struct {
|
||||||
|
id rpc.ID
|
||||||
|
typ Type
|
||||||
|
created time.Time
|
||||||
|
logsCrit FilterCriteria
|
||||||
|
logs chan []Log
|
||||||
|
hashes chan common.Hash
|
||||||
|
headers chan *types.Header
|
||||||
|
installed chan struct{} // closed when the filter is installed
|
||||||
|
err chan error // closed when the filter is uninstalled
|
||||||
|
}
|
||||||
|
|
||||||
|
// EventSystem creates subscriptions, processes events and broadcasts them to the
|
||||||
|
// subscription which match the subscription criteria.
|
||||||
|
type EventSystem struct {
|
||||||
|
mux *event.TypeMux
|
||||||
sub event.Subscription
|
sub event.Subscription
|
||||||
|
install chan *subscription // install filter for event notification
|
||||||
|
uninstall chan *subscription // remove filter for event notification
|
||||||
}
|
}
|
||||||
|
|
||||||
// NewFilterSystem returns a newly allocated filter manager
|
// NewEventSystem creates a new manager that listens for event on the given mux,
|
||||||
func NewFilterSystem(mux *event.TypeMux) *FilterSystem {
|
// parses and filters them. It uses the all map to retrieve filter changes. The
|
||||||
fs := &FilterSystem{
|
// work loop holds its own index that is used to forward events to filters.
|
||||||
chainFilters: make(map[int]*Filter),
|
//
|
||||||
pendingTxFilters: make(map[int]*Filter),
|
// The returned manager has a loop that needs to be stopped with the Stop function
|
||||||
logFilters: make(map[int]*Filter),
|
// or by stopping the given mux.
|
||||||
pendingLogFilters: make(map[int]*Filter),
|
func NewEventSystem(mux *event.TypeMux) *EventSystem {
|
||||||
generic: make(map[int]*Filter),
|
m := &EventSystem{
|
||||||
}
|
mux: mux,
|
||||||
fs.sub = mux.Subscribe(
|
install: make(chan *subscription),
|
||||||
core.PendingLogsEvent{},
|
uninstall: make(chan *subscription),
|
||||||
core.RemovedLogsEvent{},
|
|
||||||
core.ChainEvent{},
|
|
||||||
core.TxPreEvent{},
|
|
||||||
vm.Logs(nil),
|
|
||||||
)
|
|
||||||
go fs.filterLoop()
|
|
||||||
return fs
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// Stop quits the filter loop required for polling events
|
go m.eventLoop()
|
||||||
func (fs *FilterSystem) Stop() {
|
|
||||||
fs.sub.Unsubscribe()
|
return m
|
||||||
}
|
}
|
||||||
|
|
||||||
// Acquire filter system maps lock, required to force lock acquisition
|
// Subscription is created when the client registers itself for a particular event.
|
||||||
// sequence with filterMu acquired first to avoid deadlocks by callbacks
|
type Subscription struct {
|
||||||
func (fs *FilterSystem) Lock() {
|
ID rpc.ID
|
||||||
fs.filterMu.Lock()
|
f *subscription
|
||||||
|
es *EventSystem
|
||||||
|
unsubOnce sync.Once
|
||||||
}
|
}
|
||||||
|
|
||||||
// Release filter system maps lock
|
// Err returns a channel that is closed when unsubscribed.
|
||||||
func (fs *FilterSystem) Unlock() {
|
func (sub *Subscription) Err() <-chan error {
|
||||||
fs.filterMu.Unlock()
|
return sub.f.err
|
||||||
}
|
}
|
||||||
|
|
||||||
// Add adds a filter to the filter manager
|
// Unsubscribe uninstalls the subscription from the event broadcast loop.
|
||||||
// Expects filterMu to be locked.
|
func (sub *Subscription) Unsubscribe() {
|
||||||
func (fs *FilterSystem) Add(filter *Filter, filterType FilterType) (int, error) {
|
sub.unsubOnce.Do(func() {
|
||||||
id := fs.filterId
|
uninstallLoop:
|
||||||
filter.created = time.Now()
|
for {
|
||||||
|
// write uninstall request and consume logs/hashes. This prevents
|
||||||
switch filterType {
|
// the eventLoop broadcast method to deadlock when writing to the
|
||||||
case ChainFilter:
|
// filter event channel while the subscription loop is waiting for
|
||||||
fs.chainFilters[id] = filter
|
// this method to return (and thus not reading these events).
|
||||||
case PendingTxFilter:
|
select {
|
||||||
fs.pendingTxFilters[id] = filter
|
case sub.es.uninstall <- sub.f:
|
||||||
case LogFilter:
|
break uninstallLoop
|
||||||
fs.logFilters[id] = filter
|
case <-sub.f.logs:
|
||||||
case PendingLogFilter:
|
case <-sub.f.hashes:
|
||||||
fs.pendingLogFilters[id] = filter
|
case <-sub.f.headers:
|
||||||
default:
|
|
||||||
return 0, fmt.Errorf("unknown filter type %v", filterType)
|
|
||||||
}
|
|
||||||
fs.generic[id] = filter
|
|
||||||
|
|
||||||
fs.filterId++
|
|
||||||
|
|
||||||
return id, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// Remove removes a filter by filter id
|
|
||||||
// Expects filterMu to be locked.
|
|
||||||
func (fs *FilterSystem) Remove(id int) {
|
|
||||||
delete(fs.chainFilters, id)
|
|
||||||
delete(fs.pendingTxFilters, id)
|
|
||||||
delete(fs.logFilters, id)
|
|
||||||
delete(fs.pendingLogFilters, id)
|
|
||||||
delete(fs.generic, id)
|
|
||||||
}
|
|
||||||
|
|
||||||
func (fs *FilterSystem) Get(id int) *Filter {
|
|
||||||
fs.filterMu.RLock()
|
|
||||||
defer fs.filterMu.RUnlock()
|
|
||||||
|
|
||||||
return fs.generic[id]
|
|
||||||
}
|
|
||||||
|
|
||||||
// filterLoop waits for specific events from ethereum and fires their handlers
|
|
||||||
// when the filter matches the requirements.
|
|
||||||
func (fs *FilterSystem) filterLoop() {
|
|
||||||
for event := range fs.sub.Chan() {
|
|
||||||
switch ev := event.Data.(type) {
|
|
||||||
case core.ChainEvent:
|
|
||||||
fs.filterMu.RLock()
|
|
||||||
for _, filter := range fs.chainFilters {
|
|
||||||
if filter.BlockCallback != nil && !filter.created.After(event.Time) {
|
|
||||||
filter.BlockCallback(ev.Block, ev.Logs)
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
fs.filterMu.RUnlock()
|
|
||||||
case core.TxPreEvent:
|
|
||||||
fs.filterMu.RLock()
|
|
||||||
for _, filter := range fs.pendingTxFilters {
|
|
||||||
if filter.TransactionCallback != nil && !filter.created.After(event.Time) {
|
|
||||||
filter.TransactionCallback(ev.Tx)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
fs.filterMu.RUnlock()
|
|
||||||
|
|
||||||
|
// wait for filter to be uninstalled in work loop before returning
|
||||||
|
// this ensures that the manager won't use the event channel which
|
||||||
|
// will probably be closed by the client asap after this method returns.
|
||||||
|
<-sub.Err()
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
// subscribe installs the subscription in the event broadcast loop.
|
||||||
|
func (es *EventSystem) subscribe(sub *subscription) *Subscription {
|
||||||
|
es.install <- sub
|
||||||
|
<-sub.installed
|
||||||
|
return &Subscription{ID: sub.id, f: sub, es: es}
|
||||||
|
}
|
||||||
|
|
||||||
|
// SubscribeLogs creates a subscription that will write all logs matching the
|
||||||
|
// given criteria to the given logs channel.
|
||||||
|
func (es *EventSystem) SubscribeLogs(crit FilterCriteria, logs chan []Log) *Subscription {
|
||||||
|
sub := &subscription{
|
||||||
|
id: rpc.NewID(),
|
||||||
|
typ: LogsSubscription,
|
||||||
|
logsCrit: crit,
|
||||||
|
created: time.Now(),
|
||||||
|
logs: logs,
|
||||||
|
hashes: make(chan common.Hash),
|
||||||
|
headers: make(chan *types.Header),
|
||||||
|
installed: make(chan struct{}),
|
||||||
|
err: make(chan error),
|
||||||
|
}
|
||||||
|
|
||||||
|
return es.subscribe(sub)
|
||||||
|
}
|
||||||
|
|
||||||
|
// SubscribePendingLogs creates a subscription that will write pending logs matching the
|
||||||
|
// given criteria to the given channel.
|
||||||
|
func (es *EventSystem) SubscribePendingLogs(crit FilterCriteria, logs chan []Log) *Subscription {
|
||||||
|
sub := &subscription{
|
||||||
|
id: rpc.NewID(),
|
||||||
|
typ: PendingLogsSubscription,
|
||||||
|
logsCrit: crit,
|
||||||
|
created: time.Now(),
|
||||||
|
logs: logs,
|
||||||
|
hashes: make(chan common.Hash),
|
||||||
|
headers: make(chan *types.Header),
|
||||||
|
installed: make(chan struct{}),
|
||||||
|
err: make(chan error),
|
||||||
|
}
|
||||||
|
|
||||||
|
return es.subscribe(sub)
|
||||||
|
}
|
||||||
|
|
||||||
|
// SubscribePendingTxEvents creates a sbuscription that writes transaction hashes for
|
||||||
|
// transactions that enter the transaction pool.
|
||||||
|
func (es *EventSystem) SubscribePendingTxEvents(hashes chan common.Hash) *Subscription {
|
||||||
|
sub := &subscription{
|
||||||
|
id: rpc.NewID(),
|
||||||
|
typ: PendingTransactionsSubscription,
|
||||||
|
created: time.Now(),
|
||||||
|
logs: make(chan []Log),
|
||||||
|
hashes: hashes,
|
||||||
|
headers: make(chan *types.Header),
|
||||||
|
installed: make(chan struct{}),
|
||||||
|
err: make(chan error),
|
||||||
|
}
|
||||||
|
|
||||||
|
return es.subscribe(sub)
|
||||||
|
}
|
||||||
|
|
||||||
|
// SubscribeNewHeads creates a subscription that writes the header of a block that is
|
||||||
|
// imported in the chain.
|
||||||
|
func (es *EventSystem) SubscribeNewHeads(headers chan *types.Header) *Subscription {
|
||||||
|
sub := &subscription{
|
||||||
|
id: rpc.NewID(),
|
||||||
|
typ: BlocksSubscription,
|
||||||
|
created: time.Now(),
|
||||||
|
logs: make(chan []Log),
|
||||||
|
hashes: make(chan common.Hash),
|
||||||
|
headers: headers,
|
||||||
|
installed: make(chan struct{}),
|
||||||
|
err: make(chan error),
|
||||||
|
}
|
||||||
|
|
||||||
|
return es.subscribe(sub)
|
||||||
|
}
|
||||||
|
|
||||||
|
type filterIndex map[Type]map[rpc.ID]*subscription
|
||||||
|
|
||||||
|
// broadcast event to filters that match criteria.
|
||||||
|
func broadcast(filters filterIndex, ev *event.Event) {
|
||||||
|
if ev == nil {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
switch e := ev.Data.(type) {
|
||||||
case vm.Logs:
|
case vm.Logs:
|
||||||
fs.filterMu.RLock()
|
if len(e) > 0 {
|
||||||
for _, filter := range fs.logFilters {
|
for _, f := range filters[LogsSubscription] {
|
||||||
if filter.LogCallback != nil && !filter.created.After(event.Time) {
|
if ev.Time.After(f.created) {
|
||||||
for _, log := range filter.FilterLogs(ev) {
|
if matchedLogs := filterLogs(convertLogs(e, false), f.logsCrit.Addresses, f.logsCrit.Topics); len(matchedLogs) > 0 {
|
||||||
filter.LogCallback(log, false)
|
f.logs <- matchedLogs
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
fs.filterMu.RUnlock()
|
|
||||||
case core.RemovedLogsEvent:
|
case core.RemovedLogsEvent:
|
||||||
fs.filterMu.RLock()
|
for _, f := range filters[LogsSubscription] {
|
||||||
for _, filter := range fs.logFilters {
|
if ev.Time.After(f.created) {
|
||||||
if filter.LogCallback != nil && !filter.created.After(event.Time) {
|
if matchedLogs := filterLogs(convertLogs(e.Logs, true), f.logsCrit.Addresses, f.logsCrit.Topics); len(matchedLogs) > 0 {
|
||||||
for _, removedLog := range filter.FilterLogs(ev.Logs) {
|
f.logs <- matchedLogs
|
||||||
filter.LogCallback(removedLog, true)
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
fs.filterMu.RUnlock()
|
|
||||||
case core.PendingLogsEvent:
|
case core.PendingLogsEvent:
|
||||||
fs.filterMu.RLock()
|
for _, f := range filters[PendingLogsSubscription] {
|
||||||
for _, filter := range fs.pendingLogFilters {
|
if ev.Time.After(f.created) {
|
||||||
if filter.LogCallback != nil && !filter.created.After(event.Time) {
|
if matchedLogs := filterLogs(convertLogs(e.Logs, false), f.logsCrit.Addresses, f.logsCrit.Topics); len(matchedLogs) > 0 {
|
||||||
for _, pendingLog := range ev.Logs {
|
f.logs <- matchedLogs
|
||||||
filter.LogCallback(pendingLog, false)
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
fs.filterMu.RUnlock()
|
case core.TxPreEvent:
|
||||||
|
for _, f := range filters[PendingTransactionsSubscription] {
|
||||||
|
if ev.Time.After(f.created) {
|
||||||
|
f.hashes <- e.Tx.Hash()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
case core.ChainEvent:
|
||||||
|
for _, f := range filters[BlocksSubscription] {
|
||||||
|
if ev.Time.After(f.created) {
|
||||||
|
f.headers <- e.Block.Header()
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// eventLoop (un)installs filters and processes mux events.
|
||||||
|
func (es *EventSystem) eventLoop() {
|
||||||
|
var (
|
||||||
|
index = make(filterIndex)
|
||||||
|
sub = es.mux.Subscribe(core.PendingLogsEvent{}, core.RemovedLogsEvent{}, vm.Logs{}, core.TxPreEvent{}, core.ChainEvent{})
|
||||||
|
)
|
||||||
|
for {
|
||||||
|
select {
|
||||||
|
case ev, active := <-sub.Chan():
|
||||||
|
if !active { // system stopped
|
||||||
|
return
|
||||||
|
}
|
||||||
|
broadcast(index, ev)
|
||||||
|
case f := <-es.install:
|
||||||
|
if _, found := index[f.typ]; !found {
|
||||||
|
index[f.typ] = make(map[rpc.ID]*subscription)
|
||||||
|
}
|
||||||
|
index[f.typ][f.id] = f
|
||||||
|
close(f.installed)
|
||||||
|
case f := <-es.uninstall:
|
||||||
|
delete(index[f.typ], f.id)
|
||||||
|
close(f.err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// convertLogs is a helper utility that converts vm.Logs to []filter.Log.
|
||||||
|
func convertLogs(in vm.Logs, removed bool) []Log {
|
||||||
|
logs := make([]Log, len(in))
|
||||||
|
for i, l := range in {
|
||||||
|
logs[i] = Log{l, removed}
|
||||||
|
}
|
||||||
|
return logs
|
||||||
|
}
|
||||||
|
@ -17,101 +17,310 @@
|
|||||||
package filters
|
package filters
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"math/big"
|
||||||
|
"reflect"
|
||||||
"testing"
|
"testing"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
|
"github.com/ethereum/go-ethereum/common"
|
||||||
"github.com/ethereum/go-ethereum/core"
|
"github.com/ethereum/go-ethereum/core"
|
||||||
"github.com/ethereum/go-ethereum/core/types"
|
"github.com/ethereum/go-ethereum/core/types"
|
||||||
"github.com/ethereum/go-ethereum/core/vm"
|
"github.com/ethereum/go-ethereum/core/vm"
|
||||||
|
"github.com/ethereum/go-ethereum/ethdb"
|
||||||
"github.com/ethereum/go-ethereum/event"
|
"github.com/ethereum/go-ethereum/event"
|
||||||
|
"github.com/ethereum/go-ethereum/rpc"
|
||||||
)
|
)
|
||||||
|
|
||||||
func TestCallbacks(t *testing.T) {
|
|
||||||
var (
|
var (
|
||||||
mux event.TypeMux
|
mux = new(event.TypeMux)
|
||||||
fs = NewFilterSystem(&mux)
|
db, _ = ethdb.NewMemDatabase()
|
||||||
blockDone = make(chan struct{})
|
api = NewPublicFilterAPI(db, mux)
|
||||||
txDone = make(chan struct{})
|
|
||||||
logDone = make(chan struct{})
|
|
||||||
removedLogDone = make(chan struct{})
|
|
||||||
pendingLogDone = make(chan struct{})
|
|
||||||
)
|
)
|
||||||
|
|
||||||
blockFilter := &Filter{
|
// TestBlockSubscription tests if a block subscription returns block hashes for posted chain events.
|
||||||
BlockCallback: func(*types.Block, vm.Logs) {
|
// It creates multiple subscriptions:
|
||||||
close(blockDone)
|
// - one at the start and should receive all posted chain events and a second (blockHashes)
|
||||||
},
|
// - one that is created after a cutoff moment and uninstalled after a second cutoff moment (blockHashes[cutoff1:cutoff2])
|
||||||
}
|
// - one that is created after the second cutoff moment (blockHashes[cutoff2:])
|
||||||
txFilter := &Filter{
|
func TestBlockSubscription(t *testing.T) {
|
||||||
TransactionCallback: func(*types.Transaction) {
|
t.Parallel()
|
||||||
close(txDone)
|
|
||||||
},
|
var (
|
||||||
}
|
genesis = core.WriteGenesisBlockForTesting(db)
|
||||||
logFilter := &Filter{
|
chain, _ = core.GenerateChain(nil, genesis, db, 10, func(i int, gen *core.BlockGen) {})
|
||||||
LogCallback: func(l *vm.Log, oob bool) {
|
chainEvents = []core.ChainEvent{}
|
||||||
if !oob {
|
)
|
||||||
close(logDone)
|
|
||||||
}
|
for _, blk := range chain {
|
||||||
},
|
chainEvents = append(chainEvents, core.ChainEvent{Hash: blk.Hash(), Block: blk})
|
||||||
}
|
|
||||||
removedLogFilter := &Filter{
|
|
||||||
LogCallback: func(l *vm.Log, oob bool) {
|
|
||||||
if oob {
|
|
||||||
close(removedLogDone)
|
|
||||||
}
|
|
||||||
},
|
|
||||||
}
|
|
||||||
pendingLogFilter := &Filter{
|
|
||||||
LogCallback: func(*vm.Log, bool) {
|
|
||||||
close(pendingLogDone)
|
|
||||||
},
|
|
||||||
}
|
}
|
||||||
|
|
||||||
fs.Add(blockFilter, ChainFilter)
|
chan0 := make(chan *types.Header)
|
||||||
fs.Add(txFilter, PendingTxFilter)
|
sub0 := api.events.SubscribeNewHeads(chan0)
|
||||||
fs.Add(logFilter, LogFilter)
|
chan1 := make(chan *types.Header)
|
||||||
fs.Add(removedLogFilter, LogFilter)
|
sub1 := api.events.SubscribeNewHeads(chan1)
|
||||||
fs.Add(pendingLogFilter, PendingLogFilter)
|
|
||||||
|
|
||||||
mux.Post(core.ChainEvent{})
|
go func() { // simulate client
|
||||||
mux.Post(core.TxPreEvent{})
|
i1, i2 := 0, 0
|
||||||
mux.Post(vm.Logs{&vm.Log{}})
|
for i1 != len(chainEvents) || i2 != len(chainEvents) {
|
||||||
mux.Post(core.RemovedLogsEvent{Logs: vm.Logs{&vm.Log{}}})
|
|
||||||
mux.Post(core.PendingLogsEvent{Logs: vm.Logs{&vm.Log{}}})
|
|
||||||
|
|
||||||
const dura = 5 * time.Second
|
|
||||||
failTimer := time.NewTimer(dura)
|
|
||||||
select {
|
select {
|
||||||
case <-blockDone:
|
case header := <-chan0:
|
||||||
case <-failTimer.C:
|
if chainEvents[i1].Hash != header.Hash() {
|
||||||
t.Error("block filter failed to trigger (timeout)")
|
t.Errorf("sub0 received invalid hash on index %d, want %x, got %x", i1, chainEvents[i1].Hash, header.Hash())
|
||||||
|
}
|
||||||
|
i1++
|
||||||
|
case header := <-chan1:
|
||||||
|
if chainEvents[i2].Hash != header.Hash() {
|
||||||
|
t.Errorf("sub1 received invalid hash on index %d, want %x, got %x", i2, chainEvents[i2].Hash, header.Hash())
|
||||||
|
}
|
||||||
|
i2++
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
failTimer.Reset(dura)
|
sub0.Unsubscribe()
|
||||||
select {
|
sub1.Unsubscribe()
|
||||||
case <-txDone:
|
}()
|
||||||
case <-failTimer.C:
|
|
||||||
t.Error("transaction filter failed to trigger (timeout)")
|
time.Sleep(1 * time.Second)
|
||||||
|
for _, e := range chainEvents {
|
||||||
|
mux.Post(e)
|
||||||
}
|
}
|
||||||
|
|
||||||
failTimer.Reset(dura)
|
<-sub0.Err()
|
||||||
select {
|
<-sub1.Err()
|
||||||
case <-logDone:
|
|
||||||
case <-failTimer.C:
|
|
||||||
t.Error("log filter failed to trigger (timeout)")
|
|
||||||
}
|
}
|
||||||
|
|
||||||
failTimer.Reset(dura)
|
// TestPendingTxFilter tests whether pending tx filters retrieve all pending transactions that are posted to the event mux.
|
||||||
select {
|
func TestPendingTxFilter(t *testing.T) {
|
||||||
case <-removedLogDone:
|
t.Parallel()
|
||||||
case <-failTimer.C:
|
|
||||||
t.Error("removed log filter failed to trigger (timeout)")
|
var (
|
||||||
|
transactions = []*types.Transaction{
|
||||||
|
types.NewTransaction(0, common.HexToAddress("0xb794f5ea0ba39494ce83a213fffba74279579268"), new(big.Int), new(big.Int), new(big.Int), nil),
|
||||||
|
types.NewTransaction(1, common.HexToAddress("0xb794f5ea0ba39494ce83a213fffba74279579268"), new(big.Int), new(big.Int), new(big.Int), nil),
|
||||||
|
types.NewTransaction(2, common.HexToAddress("0xb794f5ea0ba39494ce83a213fffba74279579268"), new(big.Int), new(big.Int), new(big.Int), nil),
|
||||||
|
types.NewTransaction(3, common.HexToAddress("0xb794f5ea0ba39494ce83a213fffba74279579268"), new(big.Int), new(big.Int), new(big.Int), nil),
|
||||||
|
types.NewTransaction(4, common.HexToAddress("0xb794f5ea0ba39494ce83a213fffba74279579268"), new(big.Int), new(big.Int), new(big.Int), nil),
|
||||||
}
|
}
|
||||||
|
|
||||||
failTimer.Reset(dura)
|
hashes []common.Hash
|
||||||
select {
|
)
|
||||||
case <-pendingLogDone:
|
|
||||||
case <-failTimer.C:
|
fid0 := api.NewPendingTransactionFilter()
|
||||||
t.Error("pending log filter failed to trigger (timeout)")
|
|
||||||
|
time.Sleep(1 * time.Second)
|
||||||
|
for _, tx := range transactions {
|
||||||
|
ev := core.TxPreEvent{Tx: tx}
|
||||||
|
mux.Post(ev)
|
||||||
|
}
|
||||||
|
|
||||||
|
for {
|
||||||
|
h := api.GetFilterChanges(fid0).([]common.Hash)
|
||||||
|
hashes = append(hashes, h...)
|
||||||
|
|
||||||
|
if len(hashes) >= len(transactions) {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
|
||||||
|
time.Sleep(100 * time.Millisecond)
|
||||||
|
}
|
||||||
|
|
||||||
|
for i := range hashes {
|
||||||
|
if hashes[i] != transactions[i].Hash() {
|
||||||
|
t.Errorf("hashes[%d] invalid, want %x, got %x", i, transactions[i].Hash(), hashes[i])
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// TestLogFilter tests whether log filters match the correct logs that are posted to the event mux.
|
||||||
|
func TestLogFilter(t *testing.T) {
|
||||||
|
t.Parallel()
|
||||||
|
|
||||||
|
var (
|
||||||
|
firstAddr = common.HexToAddress("0x1111111111111111111111111111111111111111")
|
||||||
|
secondAddr = common.HexToAddress("0x2222222222222222222222222222222222222222")
|
||||||
|
thirdAddress = common.HexToAddress("0x3333333333333333333333333333333333333333")
|
||||||
|
notUsedAddress = common.HexToAddress("0x9999999999999999999999999999999999999999")
|
||||||
|
firstTopic = common.HexToHash("0x1111111111111111111111111111111111111111111111111111111111111111")
|
||||||
|
secondTopic = common.HexToHash("0x2222222222222222222222222222222222222222222222222222222222222222")
|
||||||
|
notUsedTopic = common.HexToHash("0x9999999999999999999999999999999999999999999999999999999999999999")
|
||||||
|
|
||||||
|
allLogs = vm.Logs{
|
||||||
|
// Note, these are used for comparison of the test cases.
|
||||||
|
vm.NewLog(firstAddr, []common.Hash{}, []byte(""), 0),
|
||||||
|
vm.NewLog(firstAddr, []common.Hash{firstTopic}, []byte(""), 1),
|
||||||
|
vm.NewLog(secondAddr, []common.Hash{firstTopic}, []byte(""), 1),
|
||||||
|
vm.NewLog(thirdAddress, []common.Hash{secondTopic}, []byte(""), 2),
|
||||||
|
vm.NewLog(thirdAddress, []common.Hash{secondTopic}, []byte(""), 3),
|
||||||
|
}
|
||||||
|
|
||||||
|
testCases = []struct {
|
||||||
|
crit FilterCriteria
|
||||||
|
expected vm.Logs
|
||||||
|
id rpc.ID
|
||||||
|
}{
|
||||||
|
// match all
|
||||||
|
{FilterCriteria{}, allLogs, ""},
|
||||||
|
// match none due to no matching addresses
|
||||||
|
{FilterCriteria{Addresses: []common.Address{common.Address{}, notUsedAddress}, Topics: [][]common.Hash{allLogs[0].Topics}}, vm.Logs{}, ""},
|
||||||
|
// match logs based on addresses, ignore topics
|
||||||
|
{FilterCriteria{Addresses: []common.Address{firstAddr}}, allLogs[:2], ""},
|
||||||
|
// match none due to no matching topics (match with address)
|
||||||
|
{FilterCriteria{Addresses: []common.Address{secondAddr}, Topics: [][]common.Hash{[]common.Hash{notUsedTopic}}}, vm.Logs{}, ""},
|
||||||
|
// match logs based on addresses and topics
|
||||||
|
{FilterCriteria{Addresses: []common.Address{thirdAddress}, Topics: [][]common.Hash{[]common.Hash{firstTopic, secondTopic}}}, allLogs[3:5], ""},
|
||||||
|
// match logs based on multiple addresses and "or" topics
|
||||||
|
{FilterCriteria{Addresses: []common.Address{secondAddr, thirdAddress}, Topics: [][]common.Hash{[]common.Hash{firstTopic, secondTopic}}}, allLogs[2:5], ""},
|
||||||
|
// block numbers are ignored for filters created with New***Filter, these return all logs that match the given criterias when the state changes
|
||||||
|
{FilterCriteria{Addresses: []common.Address{firstAddr}, FromBlock: big.NewInt(1), ToBlock: big.NewInt(2)}, allLogs[:2], ""},
|
||||||
|
}
|
||||||
|
|
||||||
|
err error
|
||||||
|
)
|
||||||
|
|
||||||
|
// create all filters
|
||||||
|
for i := range testCases {
|
||||||
|
testCases[i].id = api.NewFilter(testCases[i].crit)
|
||||||
|
}
|
||||||
|
|
||||||
|
// raise events
|
||||||
|
time.Sleep(1 * time.Second)
|
||||||
|
if err = mux.Post(allLogs); err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
for i, tt := range testCases {
|
||||||
|
var fetched []Log
|
||||||
|
for { // fetch all expected logs
|
||||||
|
fetched = append(fetched, api.GetFilterChanges(tt.id).([]Log)...)
|
||||||
|
if len(fetched) >= len(tt.expected) {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
|
||||||
|
time.Sleep(100 * time.Millisecond)
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(fetched) != len(tt.expected) {
|
||||||
|
t.Errorf("invalid number of logs for case %d, want %d log(s), got %d", i, len(tt.expected), len(fetched))
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
for l := range fetched {
|
||||||
|
if fetched[l].Removed {
|
||||||
|
t.Errorf("expected log not to be removed for log %d in case %d", l, i)
|
||||||
|
}
|
||||||
|
if !reflect.DeepEqual(fetched[l].Log, tt.expected[l]) {
|
||||||
|
t.Errorf("invalid log on index %d for case %d", l, i)
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// TestPendingLogsSubscription tests if a subscription receives the correct pending logs that are posted to the event mux.
|
||||||
|
func TestPendingLogsSubscription(t *testing.T) {
|
||||||
|
t.Parallel()
|
||||||
|
|
||||||
|
var (
|
||||||
|
firstAddr = common.HexToAddress("0x1111111111111111111111111111111111111111")
|
||||||
|
secondAddr = common.HexToAddress("0x2222222222222222222222222222222222222222")
|
||||||
|
thirdAddress = common.HexToAddress("0x3333333333333333333333333333333333333333")
|
||||||
|
notUsedAddress = common.HexToAddress("0x9999999999999999999999999999999999999999")
|
||||||
|
firstTopic = common.HexToHash("0x1111111111111111111111111111111111111111111111111111111111111111")
|
||||||
|
secondTopic = common.HexToHash("0x2222222222222222222222222222222222222222222222222222222222222222")
|
||||||
|
thirdTopic = common.HexToHash("0x3333333333333333333333333333333333333333333333333333333333333333")
|
||||||
|
forthTopic = common.HexToHash("0x4444444444444444444444444444444444444444444444444444444444444444")
|
||||||
|
notUsedTopic = common.HexToHash("0x9999999999999999999999999999999999999999999999999999999999999999")
|
||||||
|
|
||||||
|
allLogs = []core.PendingLogsEvent{
|
||||||
|
core.PendingLogsEvent{Logs: vm.Logs{vm.NewLog(firstAddr, []common.Hash{}, []byte(""), 0)}},
|
||||||
|
core.PendingLogsEvent{Logs: vm.Logs{vm.NewLog(firstAddr, []common.Hash{firstTopic}, []byte(""), 1)}},
|
||||||
|
core.PendingLogsEvent{Logs: vm.Logs{vm.NewLog(secondAddr, []common.Hash{firstTopic}, []byte(""), 2)}},
|
||||||
|
core.PendingLogsEvent{Logs: vm.Logs{vm.NewLog(thirdAddress, []common.Hash{secondTopic}, []byte(""), 3)}},
|
||||||
|
core.PendingLogsEvent{Logs: vm.Logs{vm.NewLog(thirdAddress, []common.Hash{secondTopic}, []byte(""), 4)}},
|
||||||
|
core.PendingLogsEvent{Logs: vm.Logs{
|
||||||
|
vm.NewLog(thirdAddress, []common.Hash{firstTopic}, []byte(""), 5),
|
||||||
|
vm.NewLog(thirdAddress, []common.Hash{thirdTopic}, []byte(""), 5),
|
||||||
|
vm.NewLog(thirdAddress, []common.Hash{forthTopic}, []byte(""), 5),
|
||||||
|
vm.NewLog(firstAddr, []common.Hash{firstTopic}, []byte(""), 5),
|
||||||
|
}},
|
||||||
|
}
|
||||||
|
|
||||||
|
convertLogs = func(pl []core.PendingLogsEvent) vm.Logs {
|
||||||
|
var logs vm.Logs
|
||||||
|
for _, l := range pl {
|
||||||
|
logs = append(logs, l.Logs...)
|
||||||
|
}
|
||||||
|
return logs
|
||||||
|
}
|
||||||
|
|
||||||
|
testCases = []struct {
|
||||||
|
crit FilterCriteria
|
||||||
|
expected vm.Logs
|
||||||
|
c chan []Log
|
||||||
|
sub *Subscription
|
||||||
|
}{
|
||||||
|
// match all
|
||||||
|
{FilterCriteria{}, convertLogs(allLogs), nil, nil},
|
||||||
|
// match none due to no matching addresses
|
||||||
|
{FilterCriteria{Addresses: []common.Address{common.Address{}, notUsedAddress}, Topics: [][]common.Hash{[]common.Hash{}}}, vm.Logs{}, nil, nil},
|
||||||
|
// match logs based on addresses, ignore topics
|
||||||
|
{FilterCriteria{Addresses: []common.Address{firstAddr}}, append(convertLogs(allLogs[:2]), allLogs[5].Logs[3]), nil, nil},
|
||||||
|
// match none due to no matching topics (match with address)
|
||||||
|
{FilterCriteria{Addresses: []common.Address{secondAddr}, Topics: [][]common.Hash{[]common.Hash{notUsedTopic}}}, vm.Logs{}, nil, nil},
|
||||||
|
// match logs based on addresses and topics
|
||||||
|
{FilterCriteria{Addresses: []common.Address{thirdAddress}, Topics: [][]common.Hash{[]common.Hash{firstTopic, secondTopic}}}, append(convertLogs(allLogs[3:5]), allLogs[5].Logs[0]), nil, nil},
|
||||||
|
// match logs based on multiple addresses and "or" topics
|
||||||
|
{FilterCriteria{Addresses: []common.Address{secondAddr, thirdAddress}, Topics: [][]common.Hash{[]common.Hash{firstTopic, secondTopic}}}, append(convertLogs(allLogs[2:5]), allLogs[5].Logs[0]), nil, nil},
|
||||||
|
// block numbers are ignored for filters created with New***Filter, these return all logs that match the given criterias when the state changes
|
||||||
|
{FilterCriteria{Addresses: []common.Address{firstAddr}, FromBlock: big.NewInt(2), ToBlock: big.NewInt(3)}, append(convertLogs(allLogs[:2]), allLogs[5].Logs[3]), nil, nil},
|
||||||
|
// multiple pending logs, should match only 2 topics from the logs in block 5
|
||||||
|
{FilterCriteria{Addresses: []common.Address{thirdAddress}, Topics: [][]common.Hash{[]common.Hash{firstTopic, forthTopic}}}, vm.Logs{allLogs[5].Logs[0], allLogs[5].Logs[2]}, nil, nil},
|
||||||
|
}
|
||||||
|
)
|
||||||
|
|
||||||
|
// create all subscriptions, this ensures all subscriptions are created before the events are posted.
|
||||||
|
// on slow machines this could otherwise lead to missing events when the subscription is created after
|
||||||
|
// (some) events are posted.
|
||||||
|
for i := range testCases {
|
||||||
|
testCases[i].c = make(chan []Log)
|
||||||
|
testCases[i].sub = api.events.SubscribePendingLogs(testCases[i].crit, testCases[i].c)
|
||||||
|
}
|
||||||
|
|
||||||
|
for n, test := range testCases {
|
||||||
|
i := n
|
||||||
|
tt := test
|
||||||
|
go func() {
|
||||||
|
var fetched []Log
|
||||||
|
fetchLoop:
|
||||||
|
for {
|
||||||
|
logs := <-tt.c
|
||||||
|
fetched = append(fetched, logs...)
|
||||||
|
if len(fetched) >= len(tt.expected) {
|
||||||
|
break fetchLoop
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(fetched) != len(tt.expected) {
|
||||||
|
t.Fatalf("invalid number of logs for case %d, want %d log(s), got %d", i, len(tt.expected), len(fetched))
|
||||||
|
}
|
||||||
|
|
||||||
|
for l := range fetched {
|
||||||
|
if fetched[l].Removed {
|
||||||
|
t.Errorf("expected log not to be removed for log %d in case %d", l, i)
|
||||||
|
}
|
||||||
|
if !reflect.DeepEqual(fetched[l].Log, tt.expected[l]) {
|
||||||
|
t.Errorf("invalid log on index %d for case %d", l, i)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
}
|
||||||
|
|
||||||
|
// raise events
|
||||||
|
time.Sleep(1 * time.Second)
|
||||||
|
for _, l := range allLogs {
|
||||||
|
if err := mux.Post(l); err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -24,7 +24,6 @@ import (
|
|||||||
"fmt"
|
"fmt"
|
||||||
"math/big"
|
"math/big"
|
||||||
"strings"
|
"strings"
|
||||||
"sync"
|
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/ethereum/ethash"
|
"github.com/ethereum/ethash"
|
||||||
@ -346,36 +345,11 @@ func (s *PrivateAccountAPI) SignAndSendTransaction(ctx context.Context, args Sen
|
|||||||
// It offers only methods that operate on public data that is freely available to anyone.
|
// It offers only methods that operate on public data that is freely available to anyone.
|
||||||
type PublicBlockChainAPI struct {
|
type PublicBlockChainAPI struct {
|
||||||
b Backend
|
b Backend
|
||||||
muNewBlockSubscriptions sync.Mutex // protects newBlocksSubscriptions
|
|
||||||
newBlockSubscriptions map[string]func(core.ChainEvent) error // callbacks for new block subscriptions
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// NewPublicBlockChainAPI creates a new Etheruem blockchain API.
|
// NewPublicBlockChainAPI creates a new Etheruem blockchain API.
|
||||||
func NewPublicBlockChainAPI(b Backend) *PublicBlockChainAPI {
|
func NewPublicBlockChainAPI(b Backend) *PublicBlockChainAPI {
|
||||||
api := &PublicBlockChainAPI{
|
return &PublicBlockChainAPI{b}
|
||||||
b: b,
|
|
||||||
newBlockSubscriptions: make(map[string]func(core.ChainEvent) error),
|
|
||||||
}
|
|
||||||
|
|
||||||
go api.subscriptionLoop()
|
|
||||||
|
|
||||||
return api
|
|
||||||
}
|
|
||||||
|
|
||||||
// subscriptionLoop reads events from the global event mux and creates notifications for the matched subscriptions.
|
|
||||||
func (s *PublicBlockChainAPI) subscriptionLoop() {
|
|
||||||
sub := s.b.EventMux().Subscribe(core.ChainEvent{})
|
|
||||||
for event := range sub.Chan() {
|
|
||||||
if chainEvent, ok := event.Data.(core.ChainEvent); ok {
|
|
||||||
s.muNewBlockSubscriptions.Lock()
|
|
||||||
for id, notifyOf := range s.newBlockSubscriptions {
|
|
||||||
if notifyOf(chainEvent) == rpc.ErrNotificationNotFound {
|
|
||||||
delete(s.newBlockSubscriptions, id)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
s.muNewBlockSubscriptions.Unlock()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// BlockNumber returns the block number of the chain head.
|
// BlockNumber returns the block number of the chain head.
|
||||||
@ -470,45 +444,6 @@ func (s *PublicBlockChainAPI) GetUncleCountByBlockHash(ctx context.Context, bloc
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// NewBlocksArgs allows the user to specify if the returned block should include transactions and in which format.
|
|
||||||
type NewBlocksArgs struct {
|
|
||||||
IncludeTransactions bool `json:"includeTransactions"`
|
|
||||||
TransactionDetails bool `json:"transactionDetails"`
|
|
||||||
}
|
|
||||||
|
|
||||||
// NewBlocks triggers a new block event each time a block is appended to the chain. It accepts an argument which allows
|
|
||||||
// the caller to specify whether the output should contain transactions and in what format.
|
|
||||||
func (s *PublicBlockChainAPI) NewBlocks(ctx context.Context, args NewBlocksArgs) (rpc.Subscription, error) {
|
|
||||||
notifier, supported := rpc.NotifierFromContext(ctx)
|
|
||||||
if !supported {
|
|
||||||
return nil, rpc.ErrNotificationsUnsupported
|
|
||||||
}
|
|
||||||
|
|
||||||
// create a subscription that will remove itself when unsubscribed/cancelled
|
|
||||||
subscription, err := notifier.NewSubscription(func(subId string) {
|
|
||||||
s.muNewBlockSubscriptions.Lock()
|
|
||||||
delete(s.newBlockSubscriptions, subId)
|
|
||||||
s.muNewBlockSubscriptions.Unlock()
|
|
||||||
})
|
|
||||||
|
|
||||||
if err != nil {
|
|
||||||
return nil, err
|
|
||||||
}
|
|
||||||
|
|
||||||
// add a callback that is called on chain events which will format the block and notify the client
|
|
||||||
s.muNewBlockSubscriptions.Lock()
|
|
||||||
s.newBlockSubscriptions[subscription.ID()] = func(e core.ChainEvent) error {
|
|
||||||
notification, err := s.rpcOutputBlock(e.Block, args.IncludeTransactions, args.TransactionDetails)
|
|
||||||
if err == nil {
|
|
||||||
return subscription.Notify(notification)
|
|
||||||
}
|
|
||||||
glog.V(logger.Warn).Info("unable to format block %v\n", err)
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
s.muNewBlockSubscriptions.Unlock()
|
|
||||||
return subscription, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// GetCode returns the code stored at the given address in the state for the given block number.
|
// GetCode returns the code stored at the given address in the state for the given block number.
|
||||||
func (s *PublicBlockChainAPI) GetCode(ctx context.Context, address common.Address, blockNr rpc.BlockNumber) (string, error) {
|
func (s *PublicBlockChainAPI) GetCode(ctx context.Context, address common.Address, blockNr rpc.BlockNumber) (string, error) {
|
||||||
state, _, err := s.b.StateAndHeaderByNumber(blockNr)
|
state, _, err := s.b.StateAndHeaderByNumber(blockNr)
|
||||||
@ -868,39 +803,11 @@ func newRPCTransaction(b *types.Block, txHash common.Hash) (*RPCTransaction, err
|
|||||||
// PublicTransactionPoolAPI exposes methods for the RPC interface
|
// PublicTransactionPoolAPI exposes methods for the RPC interface
|
||||||
type PublicTransactionPoolAPI struct {
|
type PublicTransactionPoolAPI struct {
|
||||||
b Backend
|
b Backend
|
||||||
muPendingTxSubs sync.Mutex
|
|
||||||
pendingTxSubs map[string]rpc.Subscription
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// NewPublicTransactionPoolAPI creates a new RPC service with methods specific for the transaction pool.
|
// NewPublicTransactionPoolAPI creates a new RPC service with methods specific for the transaction pool.
|
||||||
func NewPublicTransactionPoolAPI(b Backend) *PublicTransactionPoolAPI {
|
func NewPublicTransactionPoolAPI(b Backend) *PublicTransactionPoolAPI {
|
||||||
api := &PublicTransactionPoolAPI{
|
return &PublicTransactionPoolAPI{b}
|
||||||
b: b,
|
|
||||||
pendingTxSubs: make(map[string]rpc.Subscription),
|
|
||||||
}
|
|
||||||
|
|
||||||
go api.subscriptionLoop()
|
|
||||||
|
|
||||||
return api
|
|
||||||
}
|
|
||||||
|
|
||||||
// subscriptionLoop listens for events on the global event mux and creates notifications for subscriptions.
|
|
||||||
func (s *PublicTransactionPoolAPI) subscriptionLoop() {
|
|
||||||
sub := s.b.EventMux().Subscribe(core.TxPreEvent{})
|
|
||||||
for event := range sub.Chan() {
|
|
||||||
tx := event.Data.(core.TxPreEvent)
|
|
||||||
if from, err := tx.Tx.FromFrontier(); err == nil {
|
|
||||||
if s.b.AccountManager().HasAddress(from) {
|
|
||||||
s.muPendingTxSubs.Lock()
|
|
||||||
for id, sub := range s.pendingTxSubs {
|
|
||||||
if sub.Notify(tx.Tx.Hash()) == rpc.ErrNotificationNotFound {
|
|
||||||
delete(s.pendingTxSubs, id)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
s.muPendingTxSubs.Unlock()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func getTransaction(chainDb ethdb.Database, b Backend, txHash common.Hash) (*types.Transaction, bool, error) {
|
func getTransaction(chainDb ethdb.Database, b Backend, txHash common.Hash) (*types.Transaction, bool, error) {
|
||||||
@ -1353,31 +1260,6 @@ func (s *PublicTransactionPoolAPI) PendingTransactions() []*RPCTransaction {
|
|||||||
return transactions
|
return transactions
|
||||||
}
|
}
|
||||||
|
|
||||||
// NewPendingTransactions creates a subscription that is triggered each time a transaction enters the transaction pool
|
|
||||||
// and is send from one of the transactions this nodes manages.
|
|
||||||
func (s *PublicTransactionPoolAPI) NewPendingTransactions(ctx context.Context) (rpc.Subscription, error) {
|
|
||||||
notifier, supported := rpc.NotifierFromContext(ctx)
|
|
||||||
if !supported {
|
|
||||||
return nil, rpc.ErrNotificationsUnsupported
|
|
||||||
}
|
|
||||||
|
|
||||||
subscription, err := notifier.NewSubscription(func(id string) {
|
|
||||||
s.muPendingTxSubs.Lock()
|
|
||||||
delete(s.pendingTxSubs, id)
|
|
||||||
s.muPendingTxSubs.Unlock()
|
|
||||||
})
|
|
||||||
|
|
||||||
if err != nil {
|
|
||||||
return nil, err
|
|
||||||
}
|
|
||||||
|
|
||||||
s.muPendingTxSubs.Lock()
|
|
||||||
s.pendingTxSubs[subscription.ID()] = subscription
|
|
||||||
s.muPendingTxSubs.Unlock()
|
|
||||||
|
|
||||||
return subscription, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// Resend accepts an existing transaction and a new gas price and limit. It will remove the given transaction from the
|
// Resend accepts an existing transaction and a new gas price and limit. It will remove the given transaction from the
|
||||||
// pool and reinsert it with the new gas price and limit.
|
// pool and reinsert it with the new gas price and limit.
|
||||||
func (s *PublicTransactionPoolAPI) Resend(ctx context.Context, tx *Tx, gasPrice, gasLimit *rpc.HexNumber) (common.Hash, error) {
|
func (s *PublicTransactionPoolAPI) Resend(ctx context.Context, tx *Tx, gasPrice, gasLimit *rpc.HexNumber) (common.Hash, error) {
|
||||||
|
@ -1,297 +0,0 @@
|
|||||||
// Copyright 2016 The go-ethereum Authors
|
|
||||||
// This file is part of the go-ethereum library.
|
|
||||||
//
|
|
||||||
// The go-ethereum library is free software: you can redistribute it and/or modify
|
|
||||||
// it under the terms of the GNU Lesser General Public License as published by
|
|
||||||
// the Free Software Foundation, either version 3 of the License, or
|
|
||||||
// (at your option) any later version.
|
|
||||||
//
|
|
||||||
// The go-ethereum library is distributed in the hope that it will be useful,
|
|
||||||
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
||||||
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
||||||
// GNU Lesser General Public License for more details.
|
|
||||||
//
|
|
||||||
// You should have received a copy of the GNU Lesser General Public License
|
|
||||||
// along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
|
|
||||||
|
|
||||||
package rpc
|
|
||||||
|
|
||||||
import (
|
|
||||||
"errors"
|
|
||||||
"sync"
|
|
||||||
"time"
|
|
||||||
|
|
||||||
"github.com/ethereum/go-ethereum/logger"
|
|
||||||
"github.com/ethereum/go-ethereum/logger/glog"
|
|
||||||
"golang.org/x/net/context"
|
|
||||||
)
|
|
||||||
|
|
||||||
var (
|
|
||||||
// ErrNotificationsUnsupported is returned when the connection doesn't support notifications
|
|
||||||
ErrNotificationsUnsupported = errors.New("subscription notifications not supported by the current transport")
|
|
||||||
|
|
||||||
// ErrNotificationNotFound is returned when the notification for the given id is not found
|
|
||||||
ErrNotificationNotFound = errors.New("notification not found")
|
|
||||||
|
|
||||||
// errNotifierStopped is returned when the notifier is stopped (e.g. codec is closed)
|
|
||||||
errNotifierStopped = errors.New("unable to send notification")
|
|
||||||
|
|
||||||
// errNotificationQueueFull is returns when there are too many notifications in the queue
|
|
||||||
errNotificationQueueFull = errors.New("too many pending notifications")
|
|
||||||
)
|
|
||||||
|
|
||||||
// unsubSignal is a signal that the subscription is unsubscribed. It is used to flush buffered
|
|
||||||
// notifications that might be pending in the internal queue.
|
|
||||||
var unsubSignal = new(struct{})
|
|
||||||
|
|
||||||
// UnsubscribeCallback defines a callback that is called when a subcription ends.
|
|
||||||
// It receives the subscription id as argument.
|
|
||||||
type UnsubscribeCallback func(id string)
|
|
||||||
|
|
||||||
// notification is a helper object that holds event data for a subscription
|
|
||||||
type notification struct {
|
|
||||||
sub *bufferedSubscription // subscription id
|
|
||||||
data interface{} // event data
|
|
||||||
}
|
|
||||||
|
|
||||||
// A Notifier type describes the interface for objects that can send create subscriptions
|
|
||||||
type Notifier interface {
|
|
||||||
// Create a new subscription. The given callback is called when this subscription
|
|
||||||
// is cancelled (e.g. client send an unsubscribe, connection closed).
|
|
||||||
NewSubscription(UnsubscribeCallback) (Subscription, error)
|
|
||||||
// Cancel subscription
|
|
||||||
Unsubscribe(id string) error
|
|
||||||
}
|
|
||||||
|
|
||||||
type notifierKey struct{}
|
|
||||||
|
|
||||||
// NotifierFromContext returns the Notifier value stored in ctx, if any.
|
|
||||||
func NotifierFromContext(ctx context.Context) (Notifier, bool) {
|
|
||||||
n, ok := ctx.Value(notifierKey{}).(Notifier)
|
|
||||||
return n, ok
|
|
||||||
}
|
|
||||||
|
|
||||||
// Subscription defines the interface for objects that can notify subscribers
|
|
||||||
type Subscription interface {
|
|
||||||
// Inform client of an event
|
|
||||||
Notify(data interface{}) error
|
|
||||||
// Unique identifier
|
|
||||||
ID() string
|
|
||||||
// Cancel subscription
|
|
||||||
Cancel() error
|
|
||||||
}
|
|
||||||
|
|
||||||
// bufferedSubscription is a subscription that uses a bufferedNotifier to send
|
|
||||||
// notifications to subscribers.
|
|
||||||
type bufferedSubscription struct {
|
|
||||||
id string
|
|
||||||
unsubOnce sync.Once // call unsub method once
|
|
||||||
unsub UnsubscribeCallback // called on Unsubscribed
|
|
||||||
notifier *bufferedNotifier // forward notifications to
|
|
||||||
pending chan interface{} // closed when active
|
|
||||||
flushed chan interface{} // closed when all buffered notifications are send
|
|
||||||
lastNotification time.Time // last time a notification was send
|
|
||||||
}
|
|
||||||
|
|
||||||
// ID returns the subscription identifier that the client uses to refer to this instance.
|
|
||||||
func (s *bufferedSubscription) ID() string {
|
|
||||||
return s.id
|
|
||||||
}
|
|
||||||
|
|
||||||
// Cancel informs the notifier that this subscription is cancelled by the API
|
|
||||||
func (s *bufferedSubscription) Cancel() error {
|
|
||||||
return s.notifier.Unsubscribe(s.id)
|
|
||||||
}
|
|
||||||
|
|
||||||
// Notify the subscriber of a particular event.
|
|
||||||
func (s *bufferedSubscription) Notify(data interface{}) error {
|
|
||||||
return s.notifier.send(s.id, data)
|
|
||||||
}
|
|
||||||
|
|
||||||
// bufferedNotifier is a notifier that queues notifications in an internal queue and
|
|
||||||
// send them as fast as possible to the client from this queue. It will stop if the
|
|
||||||
// queue grows past a given size.
|
|
||||||
type bufferedNotifier struct {
|
|
||||||
codec ServerCodec // underlying connection
|
|
||||||
mu sync.Mutex // guard internal state
|
|
||||||
subscriptions map[string]*bufferedSubscription // keep track of subscriptions associated with codec
|
|
||||||
queueSize int // max number of items in queue
|
|
||||||
queue chan *notification // notification queue
|
|
||||||
stopped bool // indication if this notifier is ordered to stop
|
|
||||||
}
|
|
||||||
|
|
||||||
// newBufferedNotifier returns a notifier that queues notifications in an internal queue
|
|
||||||
// from which notifications are send as fast as possible to the client. If the queue size
|
|
||||||
// limit is reached (client is unable to keep up) it will stop and closes the codec.
|
|
||||||
func newBufferedNotifier(codec ServerCodec, size int) *bufferedNotifier {
|
|
||||||
notifier := &bufferedNotifier{
|
|
||||||
codec: codec,
|
|
||||||
subscriptions: make(map[string]*bufferedSubscription),
|
|
||||||
queue: make(chan *notification, size),
|
|
||||||
queueSize: size,
|
|
||||||
}
|
|
||||||
|
|
||||||
go notifier.run()
|
|
||||||
|
|
||||||
return notifier
|
|
||||||
}
|
|
||||||
|
|
||||||
// NewSubscription creates a new subscription that forwards events to this instance internal
|
|
||||||
// queue. The given callback is called when the subscription is unsubscribed/cancelled.
|
|
||||||
func (n *bufferedNotifier) NewSubscription(callback UnsubscribeCallback) (Subscription, error) {
|
|
||||||
id, err := newSubscriptionID()
|
|
||||||
if err != nil {
|
|
||||||
return nil, err
|
|
||||||
}
|
|
||||||
|
|
||||||
n.mu.Lock()
|
|
||||||
defer n.mu.Unlock()
|
|
||||||
|
|
||||||
if n.stopped {
|
|
||||||
return nil, errNotifierStopped
|
|
||||||
}
|
|
||||||
|
|
||||||
sub := &bufferedSubscription{
|
|
||||||
id: id,
|
|
||||||
unsub: callback,
|
|
||||||
notifier: n,
|
|
||||||
pending: make(chan interface{}),
|
|
||||||
flushed: make(chan interface{}),
|
|
||||||
lastNotification: time.Now(),
|
|
||||||
}
|
|
||||||
|
|
||||||
n.subscriptions[id] = sub
|
|
||||||
|
|
||||||
return sub, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// Remove the given subscription. If subscription is not found notificationNotFoundErr is returned.
|
|
||||||
func (n *bufferedNotifier) Unsubscribe(subid string) error {
|
|
||||||
n.mu.Lock()
|
|
||||||
sub, found := n.subscriptions[subid]
|
|
||||||
n.mu.Unlock()
|
|
||||||
|
|
||||||
if found {
|
|
||||||
// send the unsubscribe signal, this will cause the notifier not to accept new events
|
|
||||||
// for this subscription and will close the flushed channel after the last (buffered)
|
|
||||||
// notification was send to the client.
|
|
||||||
if err := n.send(subid, unsubSignal); err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
|
|
||||||
// wait for confirmation that all (buffered) events are send for this subscription.
|
|
||||||
// this ensures that the unsubscribe method response is not send before all buffered
|
|
||||||
// events for this subscription are send.
|
|
||||||
<-sub.flushed
|
|
||||||
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
return ErrNotificationNotFound
|
|
||||||
}
|
|
||||||
|
|
||||||
// Send enques the given data for the subscription with public ID on the internal queue. t returns
|
|
||||||
// an error when the notifier is stopped or the queue is full. If data is the unsubscribe signal it
|
|
||||||
// will remove the subscription with the given id from the subscription collection.
|
|
||||||
func (n *bufferedNotifier) send(id string, data interface{}) error {
|
|
||||||
n.mu.Lock()
|
|
||||||
defer n.mu.Unlock()
|
|
||||||
|
|
||||||
if n.stopped {
|
|
||||||
return errNotifierStopped
|
|
||||||
}
|
|
||||||
|
|
||||||
var (
|
|
||||||
subscription *bufferedSubscription
|
|
||||||
found bool
|
|
||||||
)
|
|
||||||
|
|
||||||
// check if subscription is associated with this connection, it might be cancelled
|
|
||||||
// (subscribe/connection closed)
|
|
||||||
if subscription, found = n.subscriptions[id]; !found {
|
|
||||||
glog.V(logger.Error).Infof("received notification for unknown subscription %s\n", id)
|
|
||||||
return ErrNotificationNotFound
|
|
||||||
}
|
|
||||||
|
|
||||||
// received the unsubscribe signal. Add it to the queue to make sure any pending notifications
|
|
||||||
// for this subscription are send. When the run loop receives this singal it will signal that
|
|
||||||
// all pending subscriptions are flushed and that the confirmation of the unsubscribe can be
|
|
||||||
// send to the user. Remove the subscriptions to make sure new notifications are not accepted.
|
|
||||||
if data == unsubSignal {
|
|
||||||
delete(n.subscriptions, id)
|
|
||||||
if subscription.unsub != nil {
|
|
||||||
subscription.unsubOnce.Do(func() { subscription.unsub(id) })
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
subscription.lastNotification = time.Now()
|
|
||||||
|
|
||||||
if len(n.queue) >= n.queueSize {
|
|
||||||
glog.V(logger.Warn).Infoln("too many buffered notifications -> close connection")
|
|
||||||
n.codec.Close()
|
|
||||||
return errNotificationQueueFull
|
|
||||||
}
|
|
||||||
|
|
||||||
n.queue <- ¬ification{subscription, data}
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// run reads notifications from the internal queue and sends them to the client. In case of an
|
|
||||||
// error, or when the codec is closed it will cancel all active subscriptions and returns.
|
|
||||||
func (n *bufferedNotifier) run() {
|
|
||||||
defer func() {
|
|
||||||
n.mu.Lock()
|
|
||||||
defer n.mu.Unlock()
|
|
||||||
|
|
||||||
n.stopped = true
|
|
||||||
close(n.queue)
|
|
||||||
|
|
||||||
// on exit call unsubscribe callback
|
|
||||||
for id, sub := range n.subscriptions {
|
|
||||||
if sub.unsub != nil {
|
|
||||||
sub.unsubOnce.Do(func() { sub.unsub(id) })
|
|
||||||
}
|
|
||||||
close(sub.flushed)
|
|
||||||
delete(n.subscriptions, id)
|
|
||||||
}
|
|
||||||
}()
|
|
||||||
|
|
||||||
for {
|
|
||||||
select {
|
|
||||||
case notification := <-n.queue:
|
|
||||||
// It can happen that an event is raised before the RPC server was able to send the sub
|
|
||||||
// id to the client. Therefore subscriptions are marked as pending until the sub id was
|
|
||||||
// send. The RPC server will activate the subscription by closing the pending chan.
|
|
||||||
<-notification.sub.pending
|
|
||||||
|
|
||||||
if notification.data == unsubSignal {
|
|
||||||
// unsubSignal is the last accepted message for this subscription. Raise the signal
|
|
||||||
// that all buffered notifications are sent by closing the flushed channel. This
|
|
||||||
// indicates that the response for the unsubscribe can be send to the client.
|
|
||||||
close(notification.sub.flushed)
|
|
||||||
} else {
|
|
||||||
msg := n.codec.CreateNotification(notification.sub.id, notification.data)
|
|
||||||
if err := n.codec.Write(msg); err != nil {
|
|
||||||
n.codec.Close()
|
|
||||||
// unable to send notification to client, unsubscribe all subscriptions
|
|
||||||
glog.V(logger.Warn).Infof("unable to send notification - %v\n", err)
|
|
||||||
return
|
|
||||||
}
|
|
||||||
}
|
|
||||||
case <-n.codec.Closed(): // connection was closed
|
|
||||||
glog.V(logger.Debug).Infoln("codec closed, stop subscriptions")
|
|
||||||
return
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// Marks the subscription as active. This will causes the notifications for this subscription to be
|
|
||||||
// forwarded to the client.
|
|
||||||
func (n *bufferedNotifier) activate(subid string) {
|
|
||||||
n.mu.Lock()
|
|
||||||
defer n.mu.Unlock()
|
|
||||||
|
|
||||||
if sub, found := n.subscriptions[subid]; found {
|
|
||||||
close(sub.pending)
|
|
||||||
}
|
|
||||||
}
|
|
@ -166,7 +166,7 @@ func (s *Server) serveRequest(codec ServerCodec, singleShot bool, options CodecO
|
|||||||
// to send notification to clients. It is thight to the codec/connection. If the
|
// to send notification to clients. It is thight to the codec/connection. If the
|
||||||
// connection is closed the notifier will stop and cancels all active subscriptions.
|
// connection is closed the notifier will stop and cancels all active subscriptions.
|
||||||
if options&OptionSubscriptions == OptionSubscriptions {
|
if options&OptionSubscriptions == OptionSubscriptions {
|
||||||
ctx = context.WithValue(ctx, notifierKey{}, newBufferedNotifier(codec, notificationBufferSize))
|
ctx = context.WithValue(ctx, notifierKey{}, newNotifier(codec))
|
||||||
}
|
}
|
||||||
s.codecsMu.Lock()
|
s.codecsMu.Lock()
|
||||||
if atomic.LoadInt32(&s.run) != 1 { // server stopped
|
if atomic.LoadInt32(&s.run) != 1 { // server stopped
|
||||||
@ -247,7 +247,7 @@ func (s *Server) Stop() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// createSubscription will call the subscription callback and returns the subscription id or error.
|
// createSubscription will call the subscription callback and returns the subscription id or error.
|
||||||
func (s *Server) createSubscription(ctx context.Context, c ServerCodec, req *serverRequest) (string, error) {
|
func (s *Server) createSubscription(ctx context.Context, c ServerCodec, req *serverRequest) (ID, error) {
|
||||||
// subscription have as first argument the context following optional arguments
|
// subscription have as first argument the context following optional arguments
|
||||||
args := []reflect.Value{req.callb.rcvr, reflect.ValueOf(ctx)}
|
args := []reflect.Value{req.callb.rcvr, reflect.ValueOf(ctx)}
|
||||||
args = append(args, req.args...)
|
args = append(args, req.args...)
|
||||||
@ -257,7 +257,7 @@ func (s *Server) createSubscription(ctx context.Context, c ServerCodec, req *ser
|
|||||||
return "", reply[1].Interface().(error)
|
return "", reply[1].Interface().(error)
|
||||||
}
|
}
|
||||||
|
|
||||||
return reply[0].Interface().(Subscription).ID(), nil
|
return reply[0].Interface().(*Subscription).ID, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// handle executes a request and returns the response from the callback.
|
// handle executes a request and returns the response from the callback.
|
||||||
@ -273,8 +273,8 @@ func (s *Server) handle(ctx context.Context, codec ServerCodec, req *serverReque
|
|||||||
return codec.CreateErrorResponse(&req.id, &callbackError{ErrNotificationsUnsupported.Error()}), nil
|
return codec.CreateErrorResponse(&req.id, &callbackError{ErrNotificationsUnsupported.Error()}), nil
|
||||||
}
|
}
|
||||||
|
|
||||||
subid := req.args[0].String()
|
subid := ID(req.args[0].String())
|
||||||
if err := notifier.Unsubscribe(subid); err != nil {
|
if err := notifier.unsubscribe(subid); err != nil {
|
||||||
return codec.CreateErrorResponse(&req.id, &callbackError{err.Error()}), nil
|
return codec.CreateErrorResponse(&req.id, &callbackError{err.Error()}), nil
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -292,7 +292,7 @@ func (s *Server) handle(ctx context.Context, codec ServerCodec, req *serverReque
|
|||||||
// active the subscription after the sub id was successfully sent to the client
|
// active the subscription after the sub id was successfully sent to the client
|
||||||
activateSub := func() {
|
activateSub := func() {
|
||||||
notifier, _ := NotifierFromContext(ctx)
|
notifier, _ := NotifierFromContext(ctx)
|
||||||
notifier.(*bufferedNotifier).activate(subid)
|
notifier.activate(subid)
|
||||||
}
|
}
|
||||||
|
|
||||||
return codec.CreateResponse(req.id, subid), activateSub
|
return codec.CreateResponse(req.id, subid), activateSub
|
||||||
|
@ -72,7 +72,7 @@ func (s *Service) InvalidRets3() (string, string, error) {
|
|||||||
return "", "", nil
|
return "", "", nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *Service) Subscription(ctx context.Context) (Subscription, error) {
|
func (s *Service) Subscription(ctx context.Context) (*Subscription, error) {
|
||||||
return nil, nil
|
return nil, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
135
rpc/subscription.go
Normal file
135
rpc/subscription.go
Normal file
@ -0,0 +1,135 @@
|
|||||||
|
// Copyright 2016 The go-ethereum Authors
|
||||||
|
// This file is part of the go-ethereum library.
|
||||||
|
//
|
||||||
|
// The go-ethereum library is free software: you can redistribute it and/or modify
|
||||||
|
// it under the terms of the GNU Lesser General Public License as published by
|
||||||
|
// the Free Software Foundation, either version 3 of the License, or
|
||||||
|
// (at your option) any later version.
|
||||||
|
//
|
||||||
|
// The go-ethereum library is distributed in the hope that it will be useful,
|
||||||
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||||
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||||
|
// GNU Lesser General Public License for more details.
|
||||||
|
//
|
||||||
|
// You should have received a copy of the GNU Lesser General Public License
|
||||||
|
// along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
|
||||||
|
|
||||||
|
package rpc
|
||||||
|
|
||||||
|
import (
|
||||||
|
"errors"
|
||||||
|
"sync"
|
||||||
|
|
||||||
|
"golang.org/x/net/context"
|
||||||
|
)
|
||||||
|
|
||||||
|
var (
|
||||||
|
// ErrNotificationsUnsupported is returned when the connection doesn't support notifications
|
||||||
|
ErrNotificationsUnsupported = errors.New("notifications not supported")
|
||||||
|
// ErrNotificationNotFound is returned when the notification for the given id is not found
|
||||||
|
ErrSubscriptionNotFound = errors.New("subscription not found")
|
||||||
|
)
|
||||||
|
|
||||||
|
// ID defines a psuedo random number that is used to identify RPC subscriptions.
|
||||||
|
type ID string
|
||||||
|
|
||||||
|
// a Subscription is created by a notifier and tight to that notifier. The client can use
|
||||||
|
// this subscription to wait for an unsubscribe request for the client, see Err().
|
||||||
|
type Subscription struct {
|
||||||
|
ID ID
|
||||||
|
err chan error // closed on unsubscribe
|
||||||
|
}
|
||||||
|
|
||||||
|
// Err returns a channel that is closed when the client send an unsubscribe request.
|
||||||
|
func (s *Subscription) Err() <-chan error {
|
||||||
|
return s.err
|
||||||
|
}
|
||||||
|
|
||||||
|
// notifierKey is used to store a notifier within the connection context.
|
||||||
|
type notifierKey struct{}
|
||||||
|
|
||||||
|
// Notifier is tight to a RPC connection that supports subscriptions.
|
||||||
|
// Server callbacks use the notifier to send notifications.
|
||||||
|
type Notifier struct {
|
||||||
|
codec ServerCodec
|
||||||
|
subMu sync.RWMutex // guards active and inactive maps
|
||||||
|
stopped bool
|
||||||
|
active map[ID]*Subscription
|
||||||
|
inactive map[ID]*Subscription
|
||||||
|
}
|
||||||
|
|
||||||
|
// newNotifier creates a new notifier that can be used to send subscription
|
||||||
|
// notifications to the client.
|
||||||
|
func newNotifier(codec ServerCodec) *Notifier {
|
||||||
|
return &Notifier{
|
||||||
|
codec: codec,
|
||||||
|
active: make(map[ID]*Subscription),
|
||||||
|
inactive: make(map[ID]*Subscription),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// NotifierFromContext returns the Notifier value stored in ctx, if any.
|
||||||
|
func NotifierFromContext(ctx context.Context) (*Notifier, bool) {
|
||||||
|
n, ok := ctx.Value(notifierKey{}).(*Notifier)
|
||||||
|
return n, ok
|
||||||
|
}
|
||||||
|
|
||||||
|
// CreateSubscription returns a new subscription that is coupled to the
|
||||||
|
// RPC connection. By default subscriptions are inactive and notifications
|
||||||
|
// are dropped until the subscription is marked as active. This is done
|
||||||
|
// by the RPC server after the subscription ID is send to the client.
|
||||||
|
func (n *Notifier) CreateSubscription() *Subscription {
|
||||||
|
s := &Subscription{NewID(), make(chan error)}
|
||||||
|
n.subMu.Lock()
|
||||||
|
n.inactive[s.ID] = s
|
||||||
|
n.subMu.Unlock()
|
||||||
|
return s
|
||||||
|
}
|
||||||
|
|
||||||
|
// Notify sends a notification to the client with the given data as payload.
|
||||||
|
// If an error occurs the RPC connection is closed and the error is returned.
|
||||||
|
func (n *Notifier) Notify(id ID, data interface{}) error {
|
||||||
|
n.subMu.RLock()
|
||||||
|
defer n.subMu.RUnlock()
|
||||||
|
|
||||||
|
_, active := n.active[id]
|
||||||
|
if active {
|
||||||
|
notification := n.codec.CreateNotification(string(id), data)
|
||||||
|
if err := n.codec.Write(notification); err != nil {
|
||||||
|
n.codec.Close()
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Closed returns a channel that is closed when the RPC connection is closed.
|
||||||
|
func (n *Notifier) Closed() <-chan interface{} {
|
||||||
|
return n.codec.Closed()
|
||||||
|
}
|
||||||
|
|
||||||
|
// unsubscribe a subscription.
|
||||||
|
// If the subscription could not be found ErrSubscriptionNotFound is returned.
|
||||||
|
func (n *Notifier) unsubscribe(id ID) error {
|
||||||
|
n.subMu.Lock()
|
||||||
|
defer n.subMu.Unlock()
|
||||||
|
if s, found := n.active[id]; found {
|
||||||
|
close(s.err)
|
||||||
|
delete(n.active, id)
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
return ErrSubscriptionNotFound
|
||||||
|
}
|
||||||
|
|
||||||
|
// activate enables a subscription. Until a subscription is enabled all
|
||||||
|
// notifications are dropped. This method is called by the RPC server after
|
||||||
|
// the subscription ID was sent to client. This prevents notifications being
|
||||||
|
// send to the client before the subscription ID is send to the client.
|
||||||
|
func (n *Notifier) activate(id ID) {
|
||||||
|
n.subMu.Lock()
|
||||||
|
defer n.subMu.Unlock()
|
||||||
|
if sub, found := n.inactive[id]; found {
|
||||||
|
n.active[id] = sub
|
||||||
|
delete(n.inactive, id)
|
||||||
|
}
|
||||||
|
}
|
@ -50,7 +50,7 @@ func (s *NotificationTestService) Unsubscribe(subid string) {
|
|||||||
s.mu.Unlock()
|
s.mu.Unlock()
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *NotificationTestService) SomeSubscription(ctx context.Context, n, val int) (Subscription, error) {
|
func (s *NotificationTestService) SomeSubscription(ctx context.Context, n, val int) (*Subscription, error) {
|
||||||
notifier, supported := NotifierFromContext(ctx)
|
notifier, supported := NotifierFromContext(ctx)
|
||||||
if !supported {
|
if !supported {
|
||||||
return nil, ErrNotificationsUnsupported
|
return nil, ErrNotificationsUnsupported
|
||||||
@ -59,17 +59,29 @@ func (s *NotificationTestService) SomeSubscription(ctx context.Context, n, val i
|
|||||||
// by explicitly creating an subscription we make sure that the subscription id is send back to the client
|
// by explicitly creating an subscription we make sure that the subscription id is send back to the client
|
||||||
// before the first subscription.Notify is called. Otherwise the events might be send before the response
|
// before the first subscription.Notify is called. Otherwise the events might be send before the response
|
||||||
// for the eth_subscribe method.
|
// for the eth_subscribe method.
|
||||||
subscription, err := notifier.NewSubscription(s.Unsubscribe)
|
subscription := notifier.CreateSubscription()
|
||||||
if err != nil {
|
|
||||||
return nil, err
|
|
||||||
}
|
|
||||||
|
|
||||||
go func() {
|
go func() {
|
||||||
|
// test expects n events, if we begin sending event immediatly some events
|
||||||
|
// will probably be dropped since the subscription ID might not be send to
|
||||||
|
// the client.
|
||||||
|
time.Sleep(5 * time.Second)
|
||||||
for i := 0; i < n; i++ {
|
for i := 0; i < n; i++ {
|
||||||
if err := subscription.Notify(val + i); err != nil {
|
if err := notifier.Notify(subscription.ID, val+i); err != nil {
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
select {
|
||||||
|
case <-notifier.Closed():
|
||||||
|
s.mu.Lock()
|
||||||
|
s.unsubscribed = true
|
||||||
|
s.mu.Unlock()
|
||||||
|
case <-subscription.Err():
|
||||||
|
s.mu.Lock()
|
||||||
|
s.unsubscribed = true
|
||||||
|
s.mu.Unlock()
|
||||||
|
}
|
||||||
}()
|
}()
|
||||||
|
|
||||||
return subscription, nil
|
return subscription, nil
|
||||||
@ -77,7 +89,7 @@ func (s *NotificationTestService) SomeSubscription(ctx context.Context, n, val i
|
|||||||
|
|
||||||
// HangSubscription blocks on s.unblockHangSubscription before
|
// HangSubscription blocks on s.unblockHangSubscription before
|
||||||
// sending anything.
|
// sending anything.
|
||||||
func (s *NotificationTestService) HangSubscription(ctx context.Context, val int) (Subscription, error) {
|
func (s *NotificationTestService) HangSubscription(ctx context.Context, val int) (*Subscription, error) {
|
||||||
notifier, supported := NotifierFromContext(ctx)
|
notifier, supported := NotifierFromContext(ctx)
|
||||||
if !supported {
|
if !supported {
|
||||||
return nil, ErrNotificationsUnsupported
|
return nil, ErrNotificationsUnsupported
|
||||||
@ -85,12 +97,10 @@ func (s *NotificationTestService) HangSubscription(ctx context.Context, val int)
|
|||||||
|
|
||||||
s.gotHangSubscriptionReq <- struct{}{}
|
s.gotHangSubscriptionReq <- struct{}{}
|
||||||
<-s.unblockHangSubscription
|
<-s.unblockHangSubscription
|
||||||
subscription, err := notifier.NewSubscription(s.Unsubscribe)
|
subscription := notifier.CreateSubscription()
|
||||||
if err != nil {
|
|
||||||
return nil, err
|
|
||||||
}
|
|
||||||
go func() {
|
go func() {
|
||||||
subscription.Notify(val)
|
notifier.Notify(subscription.ID, val)
|
||||||
}()
|
}()
|
||||||
return subscription, nil
|
return subscription, nil
|
||||||
}
|
}
|
@ -269,6 +269,6 @@ func (bn *BlockNumber) UnmarshalJSON(data []byte) error {
|
|||||||
return fmt.Errorf("blocknumber not in range [%d, %d]", earliestBlockNumber, maxBlockNumber)
|
return fmt.Errorf("blocknumber not in range [%d, %d]", earliestBlockNumber, maxBlockNumber)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (bn *BlockNumber) Int64() int64 {
|
func (bn BlockNumber) Int64() int64 {
|
||||||
return (int64)(*bn)
|
return (int64)(bn)
|
||||||
}
|
}
|
||||||
|
42
rpc/utils.go
42
rpc/utils.go
@ -17,17 +17,26 @@
|
|||||||
package rpc
|
package rpc
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"crypto/rand"
|
"bufio"
|
||||||
|
crand "crypto/rand"
|
||||||
|
"encoding/binary"
|
||||||
"encoding/hex"
|
"encoding/hex"
|
||||||
"errors"
|
|
||||||
"math/big"
|
"math/big"
|
||||||
|
"math/rand"
|
||||||
"reflect"
|
"reflect"
|
||||||
|
"sync"
|
||||||
|
"time"
|
||||||
"unicode"
|
"unicode"
|
||||||
"unicode/utf8"
|
"unicode/utf8"
|
||||||
|
|
||||||
"golang.org/x/net/context"
|
"golang.org/x/net/context"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
var (
|
||||||
|
subscriptionIDGenMu sync.Mutex
|
||||||
|
subscriptionIDGen = idGenerator()
|
||||||
|
)
|
||||||
|
|
||||||
// Is this an exported - upper case - name?
|
// Is this an exported - upper case - name?
|
||||||
func isExported(name string) bool {
|
func isExported(name string) bool {
|
||||||
rune, _ := utf8.DecodeRuneInString(name)
|
rune, _ := utf8.DecodeRuneInString(name)
|
||||||
@ -218,11 +227,28 @@ METHODS:
|
|||||||
return callbacks, subscriptions
|
return callbacks, subscriptions
|
||||||
}
|
}
|
||||||
|
|
||||||
func newSubscriptionID() (string, error) {
|
// idGenerator helper utility that generates a (pseudo) random sequence of
|
||||||
var subid [16]byte
|
// bytes that are used to generate identifiers.
|
||||||
n, _ := rand.Read(subid[:])
|
func idGenerator() *rand.Rand {
|
||||||
if n != 16 {
|
if seed, err := binary.ReadVarint(bufio.NewReader(crand.Reader)); err == nil {
|
||||||
return "", errors.New("Unable to generate subscription id")
|
return rand.New(rand.NewSource(seed))
|
||||||
}
|
}
|
||||||
return "0x" + hex.EncodeToString(subid[:]), nil
|
return rand.New(rand.NewSource(int64(time.Now().Nanosecond())))
|
||||||
|
}
|
||||||
|
|
||||||
|
// NewID generates a identifier that can be used as an identifier in the RPC interface.
|
||||||
|
// e.g. filter and subscription identifier.
|
||||||
|
func NewID() ID {
|
||||||
|
subscriptionIDGenMu.Lock()
|
||||||
|
defer subscriptionIDGenMu.Unlock()
|
||||||
|
|
||||||
|
id := make([]byte, 16)
|
||||||
|
for i := 0; i < len(id); i += 7 {
|
||||||
|
val := subscriptionIDGen.Int63()
|
||||||
|
for j := 0; i+j < len(id) && j < 7; j++ {
|
||||||
|
id[i+j] = byte(val)
|
||||||
|
val >>= 8
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return ID("0x" + hex.EncodeToString(id))
|
||||||
}
|
}
|
||||||
|
Loading…
Reference in New Issue
Block a user