2019-05-17 06:27:02 +00:00
// VulcanizeDB
// Copyright © 2019 Vulcanize
// This program is free software: you can redistribute it and/or modify
// it under the terms of the GNU Affero General Public License as published by
// the Free Software Foundation, either version 3 of the License, or
// (at your option) any later version.
// This program is distributed in the hope that it will be useful,
// but WITHOUT ANY WARRANTY; without even the implied warranty of
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
// GNU Affero General Public License for more details.
// You should have received a copy of the GNU Affero General Public License
// along with this program. If not, see <http://www.gnu.org/licenses/>.
2019-10-02 14:10:37 +00:00
package super_node
2019-05-17 06:27:02 +00:00
import (
2020-01-17 23:16:01 +00:00
"fmt"
2019-05-17 06:27:02 +00:00
"sync"
2019-06-18 17:28:57 +00:00
"github.com/ethereum/go-ethereum/common"
"github.com/ethereum/go-ethereum/crypto"
2019-05-17 06:27:02 +00:00
"github.com/ethereum/go-ethereum/node"
"github.com/ethereum/go-ethereum/p2p"
2019-06-18 17:28:57 +00:00
"github.com/ethereum/go-ethereum/rlp"
2019-05-17 06:27:02 +00:00
"github.com/ethereum/go-ethereum/rpc"
log "github.com/sirupsen/logrus"
2020-01-29 19:00:07 +00:00
"github.com/vulcanize/vulcanizedb/pkg/eth/core"
"github.com/vulcanize/vulcanizedb/pkg/eth/datastore/postgres"
2019-08-26 02:13:40 +00:00
"github.com/vulcanize/vulcanizedb/pkg/ipfs"
2020-01-17 23:16:01 +00:00
"github.com/vulcanize/vulcanizedb/pkg/super_node/config"
"github.com/vulcanize/vulcanizedb/pkg/super_node/shared"
2019-05-17 06:27:02 +00:00
)
2019-10-02 14:10:37 +00:00
const (
2020-01-17 23:16:01 +00:00
PayloadChanBufferSize = 20000
2019-10-02 14:10:37 +00:00
)
2019-05-17 06:27:02 +00:00
2020-01-17 23:16:01 +00:00
// SuperNode is the top level interface for streaming, converting to IPLDs, publishing,
2019-06-18 17:28:57 +00:00
// and indexing all Ethereum data; screening this data; and serving it up to subscribed clients
2019-06-07 02:09:27 +00:00
// This service is compatible with the Ethereum service interface (node.Service)
2020-01-17 23:16:01 +00:00
type SuperNode interface {
2019-05-17 06:27:02 +00:00
// APIs(), Protocols(), Start() and Stop()
node . Service
// Main event loop for syncAndPublish processes
2020-01-31 18:03:37 +00:00
SyncAndPublish ( wg * sync . WaitGroup , forwardPayloadChan chan <- shared . StreamedIPLDs , forwardQuitchan chan <- bool ) error
2019-05-17 06:27:02 +00:00
// Main event loop for handling client pub-sub
2020-01-31 18:03:37 +00:00
ScreenAndServe ( wg * sync . WaitGroup , screenAndServePayload <- chan shared . StreamedIPLDs , screenAndServeQuit <- chan bool )
2019-05-17 06:27:02 +00:00
// Method to subscribe to receive state diff processing output
2020-01-31 18:03:37 +00:00
Subscribe ( id rpc . ID , sub chan <- SubscriptionPayload , quitChan chan <- bool , params shared . SubscriptionSettings )
2019-05-17 06:27:02 +00:00
// Method to unsubscribe from state diff processing
2019-06-18 17:28:57 +00:00
Unsubscribe ( id rpc . ID )
2020-01-17 23:16:01 +00:00
// Method to access the node info for this service
2019-09-13 19:41:50 +00:00
Node ( ) core . Node
2019-05-17 06:27:02 +00:00
}
2019-10-02 14:10:37 +00:00
// Service is the underlying struct for the super node
2019-05-17 06:27:02 +00:00
type Service struct {
// Used to sync access to the Subscriptions
sync . Mutex
2020-01-17 23:16:01 +00:00
// Interface for streaming payloads over an rpc subscription
Streamer shared . PayloadStreamer
// Interface for converting raw payloads into IPLD object payloads
Converter shared . PayloadConverter
// Interface for publishing the IPLD payloads to IPFS
Publisher shared . IPLDPublisher
// Interface for indexing the CIDs of the published IPLDs in Postgres
Indexer shared . CIDIndexer
2019-05-21 19:27:24 +00:00
// Interface for filtering and serving data according to subscribed clients according to their specification
2020-01-17 23:16:01 +00:00
Filterer shared . ResponseFilterer
// Interface for fetching IPLD objects from IPFS
IPLDFetcher shared . IPLDFetcher
2019-05-21 19:27:24 +00:00
// Interface for searching and retrieving CIDs from Postgres index
2020-01-17 23:16:01 +00:00
Retriever shared . CIDRetriever
// Interface for resolving IPLDs to their data types
Resolver shared . IPLDResolver
// Chan the processor uses to subscribe to payloads from the Streamer
2020-01-31 18:03:37 +00:00
PayloadChan chan shared . RawChainData
2019-05-17 06:27:02 +00:00
// Used to signal shutdown of the service
QuitChan chan bool
2019-06-18 17:28:57 +00:00
// A mapping of rpc.IDs to their subscription channels, mapped to their subscription type (hash of the StreamFilters)
Subscriptions map [ common . Hash ] map [ rpc . ID ] Subscription
2020-01-17 23:16:01 +00:00
// A mapping of subscription params hash to the corresponding subscription params
2020-01-31 18:03:37 +00:00
SubscriptionTypes map [ common . Hash ] shared . SubscriptionSettings
2019-10-02 14:10:37 +00:00
// Info for the Geth node that this super node is working with
2020-01-17 23:16:01 +00:00
NodeInfo core . Node
// Number of publishAndIndex workers
WorkerPoolSize int
// chain type for this service
chain config . ChainType
// Path to ipfs data dir
ipfsPath string
// Underlying db
db * postgres . DB
2019-05-17 06:27:02 +00:00
}
2019-10-02 14:10:37 +00:00
// NewSuperNode creates a new super_node.Interface using an underlying super_node.Service struct
2020-01-17 23:16:01 +00:00
func NewSuperNode ( settings * config . SuperNode ) ( SuperNode , error ) {
if err := ipfs . InitIPFSPlugins ( ) ; err != nil {
return nil , err
2019-05-17 06:27:02 +00:00
}
2020-01-17 23:16:01 +00:00
sn := new ( Service )
var err error
// If we are syncing, initialize the needed interfaces
if settings . Sync {
sn . Streamer , sn . PayloadChan , err = NewPayloadStreamer ( settings . Chain , settings . WSClient )
if err != nil {
return nil , err
}
2020-01-31 18:03:37 +00:00
sn . Converter , err = NewPayloadConverter ( settings . Chain )
2020-01-17 23:16:01 +00:00
if err != nil {
return nil , err
}
sn . Publisher , err = NewIPLDPublisher ( settings . Chain , settings . IPFSPath )
if err != nil {
return nil , err
}
sn . Indexer , err = NewCIDIndexer ( settings . Chain , settings . DB )
if err != nil {
return nil , err
}
sn . Filterer , err = NewResponseFilterer ( settings . Chain )
if err != nil {
return nil , err
}
2019-10-08 19:51:38 +00:00
}
2020-01-17 23:16:01 +00:00
// If we are serving, initialize the needed interfaces
if settings . Serve {
sn . Retriever , err = NewCIDRetriever ( settings . Chain , settings . DB )
if err != nil {
return nil , err
}
sn . IPLDFetcher , err = NewIPLDFetcher ( settings . Chain , settings . IPFSPath )
if err != nil {
return nil , err
}
sn . Resolver , err = NewIPLDResolver ( settings . Chain )
if err != nil {
return nil , err
}
2019-05-21 19:27:24 +00:00
}
2020-01-17 23:16:01 +00:00
sn . QuitChan = settings . Quit
sn . Subscriptions = make ( map [ common . Hash ] map [ rpc . ID ] Subscription )
2020-01-31 18:03:37 +00:00
sn . SubscriptionTypes = make ( map [ common . Hash ] shared . SubscriptionSettings )
2020-01-17 23:16:01 +00:00
sn . WorkerPoolSize = settings . Workers
sn . NodeInfo = settings . NodeInfo
sn . ipfsPath = settings . IPFSPath
sn . chain = settings . Chain
sn . db = settings . DB
return sn , nil
2019-05-17 06:27:02 +00:00
}
// Protocols exports the services p2p protocols, this service has none
func ( sap * Service ) Protocols ( ) [ ] p2p . Protocol {
return [ ] p2p . Protocol { }
}
2019-10-02 14:10:37 +00:00
// APIs returns the RPC descriptors the super node service offers
2019-05-17 06:27:02 +00:00
func ( sap * Service ) APIs ( ) [ ] rpc . API {
2020-01-17 23:16:01 +00:00
apis := [ ] rpc . API {
2019-05-17 06:27:02 +00:00
{
Namespace : APIName ,
Version : APIVersion ,
2019-10-02 14:10:37 +00:00
Service : NewPublicSuperNodeAPI ( sap ) ,
2019-05-17 06:27:02 +00:00
Public : true ,
} ,
}
2020-01-17 23:16:01 +00:00
chainAPI , err := NewPublicAPI ( sap . chain , sap . db , sap . ipfsPath )
if err != nil {
log . Error ( err )
return apis
}
return append ( apis , chainAPI )
2019-05-17 06:27:02 +00:00
}
// SyncAndPublish is the backend processing loop which streams data from geth, converts it to iplds, publishes them to ipfs, and indexes their cids
2019-06-07 13:42:10 +00:00
// This continues on no matter if or how many subscribers there are, it then forwards the data to the ScreenAndServe() loop
// which filters and sends relevant data to client subscriptions, if there are any
2020-01-31 18:03:37 +00:00
func ( sap * Service ) SyncAndPublish ( wg * sync . WaitGroup , screenAndServePayload chan <- shared . StreamedIPLDs , screenAndServeQuit chan <- bool ) error {
2020-01-17 23:16:01 +00:00
sub , err := sap . Streamer . Stream ( sap . PayloadChan )
if err != nil {
return err
2019-05-17 06:27:02 +00:00
}
wg . Add ( 1 )
2019-06-25 20:31:14 +00:00
// Channels for forwarding data to the publishAndIndex workers
2020-01-31 18:03:37 +00:00
publishAndIndexPayload := make ( chan shared . StreamedIPLDs , PayloadChanBufferSize )
2019-08-26 02:13:40 +00:00
publishAndIndexQuit := make ( chan bool , sap . WorkerPoolSize )
2019-06-25 20:31:14 +00:00
// publishAndIndex worker pool to handle publishing and indexing concurrently, while
// limiting the number of Postgres connections we can possibly open so as to prevent error
2019-08-26 02:13:40 +00:00
for i := 0 ; i < sap . WorkerPoolSize ; i ++ {
2019-06-25 20:31:14 +00:00
sap . publishAndIndex ( i , publishAndIndexPayload , publishAndIndexQuit )
}
2019-05-17 06:27:02 +00:00
go func ( ) {
for {
select {
case payload := <- sap . PayloadChan :
2020-01-17 23:16:01 +00:00
ipldPayload , err := sap . Converter . Convert ( payload )
if err != nil {
log . Error ( err )
2019-05-17 06:27:02 +00:00
continue
}
2020-01-31 18:03:37 +00:00
// If we have a ScreenAndServe process running, forward the iplds to it
2019-05-17 06:27:02 +00:00
select {
2020-01-31 18:03:37 +00:00
case screenAndServePayload <- ipldPayload . Value ( ) :
2019-05-17 06:27:02 +00:00
default :
}
2019-06-25 20:31:14 +00:00
// Forward the payload to the publishAndIndex workers
select {
2020-01-31 18:03:37 +00:00
case publishAndIndexPayload <- ipldPayload . Value ( ) :
2019-06-25 20:31:14 +00:00
default :
2019-05-17 06:27:02 +00:00
}
2020-01-17 23:16:01 +00:00
case err := <- sub . Err ( ) :
log . Error ( err )
2019-05-17 06:27:02 +00:00
case <- sap . QuitChan :
2019-05-21 19:27:24 +00:00
// If we have a ScreenAndServe process running, forward the quit signal to it
2019-05-17 06:27:02 +00:00
select {
2019-06-25 20:31:14 +00:00
case screenAndServeQuit <- true :
2019-05-17 06:27:02 +00:00
default :
}
2020-01-17 23:16:01 +00:00
// Also forward a quit signal for each of the publishAndIndex workers
2019-08-26 02:13:40 +00:00
for i := 0 ; i < sap . WorkerPoolSize ; i ++ {
2019-06-25 20:31:14 +00:00
select {
case publishAndIndexQuit <- true :
default :
}
}
2019-05-17 06:27:02 +00:00
log . Info ( "quiting SyncAndPublish process" )
wg . Done ( )
return
}
}
} ( )
2019-10-08 19:51:38 +00:00
log . Info ( "syncAndPublish goroutine successfully spun up" )
2019-05-17 06:27:02 +00:00
return nil
}
2020-01-31 18:03:37 +00:00
func ( sap * Service ) publishAndIndex ( id int , publishAndIndexPayload <- chan shared . StreamedIPLDs , publishAndIndexQuit <- chan bool ) {
2019-06-25 20:31:14 +00:00
go func ( ) {
for {
select {
case payload := <- publishAndIndexPayload :
2020-01-17 23:16:01 +00:00
cidPayload , err := sap . Publisher . Publish ( payload )
if err != nil {
log . Errorf ( "worker %d error: %v" , id , err )
2019-06-25 20:31:14 +00:00
continue
}
2020-01-17 23:16:01 +00:00
if err := sap . Indexer . Index ( cidPayload ) ; err != nil {
log . Errorf ( "worker %d error: %v" , id , err )
2019-06-25 20:31:14 +00:00
}
case <- publishAndIndexQuit :
2019-08-28 18:41:49 +00:00
log . Infof ( "quiting publishAndIndex worker %d" , id )
2019-06-25 20:31:14 +00:00
return
}
}
} ( )
2019-10-08 19:51:38 +00:00
log . Info ( "publishAndIndex goroutine successfully spun up" )
2019-06-25 20:31:14 +00:00
}
2019-06-07 13:42:10 +00:00
// ScreenAndServe is the loop used to screen data streamed from the state diffing eth node
// and send the appropriate portions of it to a requesting client subscription, according to their subscription configuration
2020-01-31 18:03:37 +00:00
func ( sap * Service ) ScreenAndServe ( wg * sync . WaitGroup , screenAndServePayload <- chan shared . StreamedIPLDs , screenAndServeQuit <- chan bool ) {
2019-10-08 19:51:38 +00:00
wg . Add ( 1 )
2019-05-17 06:27:02 +00:00
go func ( ) {
for {
select {
2019-06-25 20:31:14 +00:00
case payload := <- screenAndServePayload :
2020-01-17 23:16:01 +00:00
sap . sendResponse ( payload )
2019-06-25 20:31:14 +00:00
case <- screenAndServeQuit :
2019-05-21 19:27:24 +00:00
log . Info ( "quiting ScreenAndServe process" )
2019-10-08 19:51:38 +00:00
wg . Done ( )
2019-05-17 06:27:02 +00:00
return
}
}
} ( )
2019-10-08 19:51:38 +00:00
log . Info ( "screenAndServe goroutine successfully spun up" )
2019-05-17 06:27:02 +00:00
}
2020-01-31 18:03:37 +00:00
func ( sap * Service ) sendResponse ( payload shared . StreamedIPLDs ) {
2019-06-25 20:31:14 +00:00
sap . Lock ( )
2019-06-18 17:28:57 +00:00
for ty , subs := range sap . Subscriptions {
2019-06-25 20:31:14 +00:00
// Retrieve the subscription parameters for this subscription type
2019-06-18 17:28:57 +00:00
subConfig , ok := sap . SubscriptionTypes [ ty ]
if ! ok {
2019-06-25 20:31:14 +00:00
log . Errorf ( "subscription configuration for subscription type %s not available" , ty . Hex ( ) )
2020-01-17 23:16:01 +00:00
sap . closeType ( ty )
2019-06-25 20:31:14 +00:00
continue
2019-06-18 17:28:57 +00:00
}
2020-01-17 23:16:01 +00:00
response , err := sap . Filterer . Filter ( subConfig , payload )
if err != nil {
log . Error ( err )
sap . closeType ( ty )
2019-06-25 20:31:14 +00:00
continue
2019-05-21 19:27:24 +00:00
}
2019-06-25 20:31:14 +00:00
for id , sub := range subs {
select {
2020-01-31 18:03:37 +00:00
case sub . PayloadChan <- SubscriptionPayload { response . Value ( ) , "" } :
2019-10-02 14:10:37 +00:00
log . Infof ( "sending super node payload to subscription %s" , id )
2019-06-25 20:31:14 +00:00
default :
log . Infof ( "unable to send payload to subscription %s; channel has no receiver" , id )
}
2019-06-18 17:28:57 +00:00
}
2019-05-21 19:27:24 +00:00
}
2019-06-25 20:31:14 +00:00
sap . Unlock ( )
2019-05-21 19:27:24 +00:00
}
// Subscribe is used by the API to subscribe to the service loop
2020-01-31 18:03:37 +00:00
// The params must be rlp serializable and satisfy the SubscriptionSettings() interface
func ( sap * Service ) Subscribe ( id rpc . ID , sub chan <- SubscriptionPayload , quitChan chan <- bool , params shared . SubscriptionSettings ) {
2019-10-02 14:10:37 +00:00
log . Info ( "Subscribing to the super node service" )
2019-06-07 16:01:29 +00:00
subscription := Subscription {
2020-01-17 23:16:01 +00:00
ID : id ,
2019-06-18 17:28:57 +00:00
PayloadChan : sub ,
QuitChan : quitChan ,
2019-05-17 06:27:02 +00:00
}
2020-01-17 23:16:01 +00:00
if params . ChainType ( ) != sap . chain {
sendNonBlockingErr ( subscription , fmt . Errorf ( "subscription %s is for chain %s, service supports chain %s" , id , params . ChainType ( ) . String ( ) , sap . chain . String ( ) ) )
sendNonBlockingQuit ( subscription )
return
}
// Subscription type is defined as the hash of the subscription settings
by , err := rlp . EncodeToBytes ( params )
if err != nil {
sendNonBlockingErr ( subscription , err )
sendNonBlockingQuit ( subscription )
return
}
subscriptionType := crypto . Keccak256Hash ( by )
2019-05-21 19:27:24 +00:00
// If the subscription requests a backfill, use the Postgres index to lookup and retrieve historical data
// Otherwise we only filter new data as it is streamed in from the state diffing geth node
2020-01-17 23:16:01 +00:00
if params . HistoricalData ( ) || params . HistoricalDataOnly ( ) {
if err := sap . backFill ( subscription , id , params ) ; err != nil {
sendNonBlockingErr ( subscription , err )
sendNonBlockingQuit ( subscription )
return
}
2019-06-07 16:01:29 +00:00
}
2020-01-17 23:16:01 +00:00
if ! params . HistoricalDataOnly ( ) {
// Add subscriber
2019-06-07 16:01:29 +00:00
sap . Lock ( )
2019-06-18 17:28:57 +00:00
if sap . Subscriptions [ subscriptionType ] == nil {
sap . Subscriptions [ subscriptionType ] = make ( map [ rpc . ID ] Subscription )
}
sap . Subscriptions [ subscriptionType ] [ id ] = subscription
2020-01-17 23:16:01 +00:00
sap . SubscriptionTypes [ subscriptionType ] = params
2019-06-07 16:01:29 +00:00
sap . Unlock ( )
}
}
2020-01-31 18:03:37 +00:00
func ( sap * Service ) backFill ( sub Subscription , id rpc . ID , params shared . SubscriptionSettings ) error {
2020-01-17 23:16:01 +00:00
log . Debug ( "sending historical data for subscriber" , id )
2019-06-07 16:01:29 +00:00
// Retrieve cached CIDs relevant to this subscriber
2019-07-02 17:38:12 +00:00
var endingBlock int64
var startingBlock int64
2020-01-17 23:16:01 +00:00
var err error
startingBlock , err = sap . Retriever . RetrieveFirstBlockNumber ( )
if err != nil {
return err
2019-06-07 16:01:29 +00:00
}
2020-01-17 23:16:01 +00:00
if startingBlock < params . StartingBlock ( ) . Int64 ( ) {
startingBlock = params . StartingBlock ( ) . Int64 ( )
2019-07-02 17:38:12 +00:00
}
2020-01-17 23:16:01 +00:00
endingBlock , err = sap . Retriever . RetrieveLastBlockNumber ( )
if err != nil {
return err
2019-07-02 17:38:12 +00:00
}
2020-01-17 23:16:01 +00:00
if endingBlock > params . EndingBlock ( ) . Int64 ( ) && params . EndingBlock ( ) . Int64 ( ) > 0 && params . EndingBlock ( ) . Int64 ( ) > startingBlock {
endingBlock = params . EndingBlock ( ) . Int64 ( )
2019-08-28 22:07:36 +00:00
}
2020-01-17 23:16:01 +00:00
log . Debug ( "historical data starting block:" , params . StartingBlock ( ) )
log . Debug ( "histocial data ending block:" , endingBlock )
2019-07-02 17:38:12 +00:00
go func ( ) {
2019-10-02 14:10:37 +00:00
for i := startingBlock ; i <= endingBlock ; i ++ {
2020-01-17 23:16:01 +00:00
cidWrapper , empty , err := sap . Retriever . Retrieve ( params , i )
if err != nil {
sendNonBlockingErr ( sub , fmt . Errorf ( "CID Retrieval error at block %d\r%s" , i , err . Error ( ) ) )
2019-07-02 17:38:12 +00:00
continue
}
2020-01-17 23:16:01 +00:00
if empty {
2019-07-02 17:38:12 +00:00
continue
}
2020-01-17 23:16:01 +00:00
blocksWrapper , err := sap . IPLDFetcher . Fetch ( cidWrapper )
if err != nil {
sendNonBlockingErr ( sub , fmt . Errorf ( "IPLD Fetching error at block %d\r%s" , i , err . Error ( ) ) )
continue
}
backFillIplds , err := sap . Resolver . Resolve ( blocksWrapper )
if err != nil {
sendNonBlockingErr ( sub , fmt . Errorf ( "IPLD Resolving error at block %d\r%s" , i , err . Error ( ) ) )
2019-07-02 17:38:12 +00:00
continue
}
select {
2020-01-31 18:03:37 +00:00
case sub . PayloadChan <- SubscriptionPayload { backFillIplds . Value ( ) , "" } :
2020-01-17 23:16:01 +00:00
log . Infof ( "sending super node historical data payload to subscription %s" , id )
2019-07-02 17:38:12 +00:00
default :
log . Infof ( "unable to send back-fill payload to subscription %s; channel has no receiver" , id )
2019-05-21 19:27:24 +00:00
}
}
2019-07-02 17:38:12 +00:00
} ( )
2020-01-17 23:16:01 +00:00
return nil
2019-05-17 06:27:02 +00:00
}
// Unsubscribe is used to unsubscribe to the StateDiffingService loop
2019-06-18 17:28:57 +00:00
func ( sap * Service ) Unsubscribe ( id rpc . ID ) {
2019-10-02 14:10:37 +00:00
log . Info ( "Unsubscribing from the super node service" )
2019-05-17 06:27:02 +00:00
sap . Lock ( )
2019-06-18 17:28:57 +00:00
for ty := range sap . Subscriptions {
delete ( sap . Subscriptions [ ty ] , id )
if len ( sap . Subscriptions [ ty ] ) == 0 {
// If we removed the last subscription of this type, remove the subscription type outright
delete ( sap . Subscriptions , ty )
delete ( sap . SubscriptionTypes , ty )
}
2019-05-17 06:27:02 +00:00
}
sap . Unlock ( )
}
2019-05-21 19:27:24 +00:00
// Start is used to begin the service
2019-05-17 06:27:02 +00:00
func ( sap * Service ) Start ( * p2p . Server ) error {
2019-10-02 14:10:37 +00:00
log . Info ( "Starting super node service" )
2019-05-17 06:27:02 +00:00
wg := new ( sync . WaitGroup )
2020-01-31 18:03:37 +00:00
payloadChan := make ( chan shared . StreamedIPLDs , PayloadChanBufferSize )
2019-06-07 02:09:27 +00:00
quitChan := make ( chan bool , 1 )
if err := sap . SyncAndPublish ( wg , payloadChan , quitChan ) ; err != nil {
return err
}
2019-10-08 19:51:38 +00:00
sap . ScreenAndServe ( wg , payloadChan , quitChan )
2019-05-17 06:27:02 +00:00
return nil
}
2019-05-21 19:27:24 +00:00
// Stop is used to close down the service
2019-05-17 06:27:02 +00:00
func ( sap * Service ) Stop ( ) error {
2019-10-02 14:10:37 +00:00
log . Info ( "Stopping super node service" )
2020-01-17 23:16:01 +00:00
sap . Lock ( )
2019-05-17 06:27:02 +00:00
close ( sap . QuitChan )
2020-01-17 23:16:01 +00:00
sap . close ( )
sap . Unlock ( )
2019-05-17 06:27:02 +00:00
return nil
}
2020-01-17 23:16:01 +00:00
// Node returns the node info for this service
2019-09-13 19:41:50 +00:00
func ( sap * Service ) Node ( ) core . Node {
2020-01-17 23:16:01 +00:00
return sap . NodeInfo
2019-09-13 19:41:50 +00:00
}
2019-05-17 06:27:02 +00:00
// close is used to close all listening subscriptions
2020-01-17 23:16:01 +00:00
// close needs to be called with subscription access locked
2019-05-17 06:27:02 +00:00
func ( sap * Service ) close ( ) {
2020-01-17 23:16:01 +00:00
for subType , subs := range sap . Subscriptions {
for _ , sub := range subs {
sendNonBlockingQuit ( sub )
2019-05-17 06:27:02 +00:00
}
2020-01-17 23:16:01 +00:00
delete ( sap . Subscriptions , subType )
delete ( sap . SubscriptionTypes , subType )
2019-05-17 06:27:02 +00:00
}
2020-01-17 23:16:01 +00:00
}
// closeType is used to close all subscriptions of given type
// closeType needs to be called with subscription access locked
func ( sap * Service ) closeType ( subType common . Hash ) {
subs := sap . Subscriptions [ subType ]
for _ , sub := range subs {
sendNonBlockingQuit ( sub )
}
delete ( sap . Subscriptions , subType )
delete ( sap . SubscriptionTypes , subType )
2019-05-17 06:27:02 +00:00
}