watcher configuration; super node versioning

This commit is contained in:
Ian Norden 2020-02-25 16:38:27 -06:00
parent 25aa4634e9
commit 330a083749
13 changed files with 381 additions and 77 deletions

View File

@ -18,12 +18,11 @@ package cmd
import ( import (
"sync" "sync"
"github.com/vulcanize/vulcanizedb/pkg/ipfs"
"github.com/ethereum/go-ethereum/rpc" "github.com/ethereum/go-ethereum/rpc"
log "github.com/sirupsen/logrus" log "github.com/sirupsen/logrus"
"github.com/spf13/cobra" "github.com/spf13/cobra"
"github.com/vulcanize/vulcanizedb/pkg/ipfs"
"github.com/vulcanize/vulcanizedb/pkg/super_node" "github.com/vulcanize/vulcanizedb/pkg/super_node"
"github.com/vulcanize/vulcanizedb/pkg/super_node/shared" "github.com/vulcanize/vulcanizedb/pkg/super_node/shared"
) )
@ -55,7 +54,7 @@ func init() {
} }
func superNode() { func superNode() {
superNodeConfigs, err := shared.NewSuperNodeConfigs() superNodeConfigs, err := super_node.NewSuperNodeConfigs()
if err != nil { if err != nil {
logWithCommand.Fatal(err) logWithCommand.Fatal(err)
} }
@ -92,7 +91,7 @@ func superNode() {
wg.Wait() wg.Wait()
} }
func startServers(superNode super_node.SuperNode, settings *shared.SuperNodeConfig) error { func startServers(superNode super_node.SuperNode, settings *super_node.Config) error {
_, _, err := rpc.StartIPCEndpoint(settings.IPCEndpoint, superNode.APIs()) _, _, err := rpc.StartIPCEndpoint(settings.IPCEndpoint, superNode.APIs())
if err != nil { if err != nil {
return err return err

View File

@ -83,6 +83,10 @@ func makePropertiesReader(client core.RPCClient) IPropertiesReader {
} }
func getNodeType(client core.RPCClient) core.NodeType { func getNodeType(client core.RPCClient) core.NodeType {
// TODO: fix this
// This heuristics for figuring out the node type are not usefull...
// for example we often port forward remote nodes to localhost
// and geth does not have to expose the admin api...
if strings.Contains(client.IpcPath(), "infura") { if strings.Contains(client.IpcPath(), "infura") {
return core.INFURA return core.INFURA
} }

View File

@ -19,8 +19,8 @@ package super_node
import ( import (
"context" "context"
"github.com/ethereum/go-ethereum/p2p"
"github.com/ethereum/go-ethereum/rlp" "github.com/ethereum/go-ethereum/rlp"
"github.com/ethereum/go-ethereum/rpc" "github.com/ethereum/go-ethereum/rpc"
log "github.com/sirupsen/logrus" log "github.com/sirupsen/logrus"
@ -105,6 +105,41 @@ func (api *PublicSuperNodeAPI) Stream(ctx context.Context, rlpParams []byte) (*r
} }
// Node is a public rpc method to allow transformers to fetch the node info for the super node // Node is a public rpc method to allow transformers to fetch the node info for the super node
func (api *PublicSuperNodeAPI) Node() core.Node { // NOTE: this is the node info for the node that the super node is syncing from, not the node info for the super node itself
func (api *PublicSuperNodeAPI) Node() *core.Node {
return api.sn.Node() return api.sn.Node()
} }
// Chain returns the chain type that this super node instance supports
func (api *PublicSuperNodeAPI) Chain() shared.ChainType {
return api.sn.Chain()
}
// Struct for holding super node meta data
type InfoAPI struct{}
// NewPublicSuperNodeAPI creates a new PublicSuperNodeAPI with the provided underlying SyncPublishScreenAndServe process
func NewInfoAPI() *InfoAPI {
return &InfoAPI{}
}
// Modules returns modules supported by this api
func (iapi *InfoAPI) Modules() map[string]string {
return map[string]string{
"vdb": "Stream",
}
}
// NodeInfo gathers and returns a collection of metadata for the super node
func (iapi *InfoAPI) NodeInfo() *p2p.NodeInfo {
return &p2p.NodeInfo{
// TODO: formalize this
ID: "vulcanizeDB",
Name: "superNode",
}
}
// Version returns the version of the super node
func (iapi *InfoAPI) Version() string {
return VersionWithMeta
}

View File

@ -17,7 +17,7 @@
package super_node package super_node
import ( import (
"errors" "fmt"
"sync" "sync"
"sync/atomic" "sync/atomic"
"time" "time"
@ -59,10 +59,12 @@ type BackFillService struct {
BatchSize uint64 BatchSize uint64
// Channel for receiving quit signal // Channel for receiving quit signal
QuitChan chan bool QuitChan chan bool
// Chain type
chain shared.ChainType
} }
// NewBackFillService returns a new BackFillInterface // NewBackFillService returns a new BackFillInterface
func NewBackFillService(settings *shared.SuperNodeConfig, screenAndServeChan chan shared.ConvertedData) (BackFillInterface, error) { func NewBackFillService(settings *Config, screenAndServeChan chan shared.ConvertedData) (BackFillInterface, error) {
publisher, err := NewIPLDPublisher(settings.Chain, settings.IPFSPath) publisher, err := NewIPLDPublisher(settings.Chain, settings.IPFSPath)
if err != nil { if err != nil {
return nil, err return nil, err
@ -97,6 +99,7 @@ func NewBackFillService(settings *shared.SuperNodeConfig, screenAndServeChan cha
BatchSize: batchSize, BatchSize: batchSize,
ScreenAndServeChan: screenAndServeChan, ScreenAndServeChan: screenAndServeChan,
QuitChan: settings.Quit, QuitChan: settings.Quit,
chain: settings.Chain,
}, nil }, nil
} }
@ -109,18 +112,18 @@ func (bfs *BackFillService) FillGapsInSuperNode(wg *sync.WaitGroup) {
for { for {
select { select {
case <-bfs.QuitChan: case <-bfs.QuitChan:
log.Info("quiting FillGaps process") log.Infof("quiting %s FillGaps process", bfs.chain.String())
wg.Done() wg.Done()
return return
case <-ticker.C: case <-ticker.C:
log.Info("searching for gaps in the super node database") log.Infof("searching for gaps in the %s super node database", bfs.chain.String())
startingBlock, err := bfs.Retriever.RetrieveFirstBlockNumber() startingBlock, err := bfs.Retriever.RetrieveFirstBlockNumber()
if err != nil { if err != nil {
log.Error(err) log.Error(err)
continue continue
} }
if startingBlock != 0 { if startingBlock != 0 {
log.Info("found gap at the beginning of the sync") log.Infof("found gap at the beginning of the %s sync", bfs.chain.String())
bfs.fillGaps(0, uint64(startingBlock-1)) bfs.fillGaps(0, uint64(startingBlock-1))
} }
gaps, err := bfs.Retriever.RetrieveGapsInData() gaps, err := bfs.Retriever.RetrieveGapsInData()
@ -136,11 +139,11 @@ func (bfs *BackFillService) FillGapsInSuperNode(wg *sync.WaitGroup) {
} }
} }
}() }()
log.Info("fillGaps goroutine successfully spun up") log.Infof("%s fillGaps goroutine successfully spun up", bfs.chain.String())
} }
func (bfs *BackFillService) fillGaps(startingBlock, endingBlock uint64) error { func (bfs *BackFillService) fillGaps(startingBlock, endingBlock uint64) error {
log.Infof("going to fill in gap from %d to %d", startingBlock, endingBlock) log.Infof("going to fill in %s gap from %d to %d", bfs.chain.String(), startingBlock, endingBlock)
errChan := make(chan error) errChan := make(chan error)
done := make(chan bool) done := make(chan bool)
err := bfs.backFill(startingBlock, endingBlock, errChan, done) err := bfs.backFill(startingBlock, endingBlock, errChan, done)
@ -152,7 +155,7 @@ func (bfs *BackFillService) fillGaps(startingBlock, endingBlock uint64) error {
case err := <-errChan: case err := <-errChan:
log.Error(err) log.Error(err)
case <-done: case <-done:
log.Infof("finished filling in gap from %d to %d", startingBlock, endingBlock) log.Infof("finished filling in %s gap from %d to %d", bfs.chain.String(), startingBlock, endingBlock)
return nil return nil
} }
} }
@ -162,7 +165,7 @@ func (bfs *BackFillService) fillGaps(startingBlock, endingBlock uint64) error {
// It splits a large range up into smaller chunks, batch fetching and processing those chunks concurrently // It splits a large range up into smaller chunks, batch fetching and processing those chunks concurrently
func (bfs *BackFillService) backFill(startingBlock, endingBlock uint64, errChan chan error, done chan bool) error { func (bfs *BackFillService) backFill(startingBlock, endingBlock uint64, errChan chan error, done chan bool) error {
if endingBlock < startingBlock { if endingBlock < startingBlock {
return errors.New("backfill: ending block number needs to be greater than starting block number") return fmt.Errorf("%s backfill: ending block number needs to be greater than starting block number", bfs.chain.String())
} }
// //
// break the range up into bins of smaller ranges // break the range up into bins of smaller ranges
@ -230,7 +233,7 @@ func (bfs *BackFillService) backFill(startingBlock, endingBlock uint64, errChan
case forwardDone <- true: case forwardDone <- true:
default: default:
} }
log.Infof("finished filling in gap sub-bin from %d to %d", doneWithHeights[0], doneWithHeights[1]) log.Infof("finished filling in %s gap sub-bin from %d to %d", bfs.chain.String(), doneWithHeights[0], doneWithHeights[1])
goroutinesFinished++ goroutinesFinished++
if goroutinesFinished >= len(blockRangeBins) { if goroutinesFinished >= len(blockRangeBins) {
done <- true done <- true

View File

@ -14,7 +14,7 @@
// You should have received a copy of the GNU Affero General Public License // You should have received a copy of the GNU Affero General Public License
// along with this program. If not, see <http://www.gnu.org/licenses/>. // along with this program. If not, see <http://www.gnu.org/licenses/>.
package shared package super_node
import ( import (
"fmt" "fmt"
@ -31,13 +31,14 @@ import (
"github.com/vulcanize/vulcanizedb/pkg/eth/core" "github.com/vulcanize/vulcanizedb/pkg/eth/core"
"github.com/vulcanize/vulcanizedb/pkg/eth/node" "github.com/vulcanize/vulcanizedb/pkg/eth/node"
"github.com/vulcanize/vulcanizedb/pkg/postgres" "github.com/vulcanize/vulcanizedb/pkg/postgres"
"github.com/vulcanize/vulcanizedb/pkg/super_node/shared"
"github.com/vulcanize/vulcanizedb/utils" "github.com/vulcanize/vulcanizedb/utils"
) )
// SuperNodeConfig struct // Config struct
type SuperNodeConfig struct { type Config struct {
// Ubiquitous fields // Ubiquitous fields
Chain ChainType Chain shared.ChainType
IPFSPath string IPFSPath string
DB *postgres.DB DB *postgres.DB
DBConfig config.Database DBConfig config.Database
@ -61,9 +62,9 @@ type SuperNodeConfig struct {
// NewSuperNodeConfigs is used to initialize multiple SuperNode configs from a single config .toml file // NewSuperNodeConfigs is used to initialize multiple SuperNode configs from a single config .toml file
// Separate chain supernode instances need to be ran in the same process in order to avoid lock contention on the ipfs repository // Separate chain supernode instances need to be ran in the same process in order to avoid lock contention on the ipfs repository
func NewSuperNodeConfigs() ([]*SuperNodeConfig, error) { func NewSuperNodeConfigs() ([]*Config, error) {
chains := viper.GetStringSlice("superNode.chains") chains := viper.GetStringSlice("superNode.chains")
configs := make([]*SuperNodeConfig, len(chains)) configs := make([]*Config, len(chains))
var err error var err error
ipfsPath := viper.GetString("superNode.ipfsPath") ipfsPath := viper.GetString("superNode.ipfsPath")
if ipfsPath == "" { if ipfsPath == "" {
@ -74,8 +75,8 @@ func NewSuperNodeConfigs() ([]*SuperNodeConfig, error) {
ipfsPath = filepath.Join(home, ".ipfs") ipfsPath = filepath.Join(home, ".ipfs")
} }
for i, chain := range chains { for i, chain := range chains {
sn := new(SuperNodeConfig) sn := new(Config)
sn.Chain, err = NewChainType(chain) sn.Chain, err = shared.NewChainType(chain)
if err != nil { if err != nil {
return nil, err return nil, err
} }
@ -96,9 +97,12 @@ func NewSuperNodeConfigs() ([]*SuperNodeConfig, error) {
} }
sn.Workers = workers sn.Workers = workers
switch sn.Chain { switch sn.Chain {
case Ethereum: case shared.Ethereum:
sn.NodeInfo, sn.WSClient, err = getEthNodeAndClient(viper.GetString("superNode.ethereum.sync.wsPath")) sn.NodeInfo, sn.WSClient, err = getEthNodeAndClient(viper.GetString("superNode.ethereum.sync.wsPath"))
case Bitcoin: if err != nil {
return nil, err
}
case shared.Bitcoin:
sn.NodeInfo = core.Node{ sn.NodeInfo = core.Node{
ID: viper.GetString("superNode.bitcoin.node.nodeID"), ID: viper.GetString("superNode.bitcoin.node.nodeID"),
ClientName: viper.GetString("superNode.bitcoin.node.clientName"), ClientName: viper.GetString("superNode.bitcoin.node.clientName"),
@ -146,21 +150,21 @@ func NewSuperNodeConfigs() ([]*SuperNodeConfig, error) {
} }
configs[i] = sn configs[i] = sn
} }
return configs, err return configs, nil
} }
// BackFillFields is used to fill in the BackFill fields of the config // BackFillFields is used to fill in the BackFill fields of the config
func (sn *SuperNodeConfig) BackFillFields(chain string) error { func (sn *Config) BackFillFields(chain string) error {
sn.BackFill = true sn.BackFill = true
var httpClient interface{} var httpClient interface{}
var err error var err error
switch sn.Chain { switch sn.Chain {
case Ethereum: case shared.Ethereum:
_, httpClient, err = getEthNodeAndClient(viper.GetString("superNode.ethereum.backFill.httpPath")) _, httpClient, err = getEthNodeAndClient(viper.GetString("superNode.ethereum.backFill.httpPath"))
if err != nil { if err != nil {
return err return err
} }
case Bitcoin: case shared.Bitcoin:
httpClient = &rpcclient.ConnConfig{ httpClient = &rpcclient.ConnConfig{
Host: viper.GetString("superNode.bitcoin.backFill.httpPath"), Host: viper.GetString("superNode.bitcoin.backFill.httpPath"),
HTTPPostMode: true, // Bitcoin core only supports HTTP POST mode HTTPPostMode: true, // Bitcoin core only supports HTTP POST mode

View File

@ -52,7 +52,7 @@ type SuperNode interface {
// Method to unsubscribe from the service // Method to unsubscribe from the service
Unsubscribe(id rpc.ID) Unsubscribe(id rpc.ID)
// Method to access the node info for the service // Method to access the node info for the service
Node() core.Node Node() *core.Node
// Method to access chain type // Method to access chain type
Chain() shared.ChainType Chain() shared.ChainType
} }
@ -84,7 +84,7 @@ type Service struct {
// A mapping of subscription params hash to the corresponding subscription params // A mapping of subscription params hash to the corresponding subscription params
SubscriptionTypes map[common.Hash]shared.SubscriptionSettings SubscriptionTypes map[common.Hash]shared.SubscriptionSettings
// Info for the Geth node that this super node is working with // Info for the Geth node that this super node is working with
NodeInfo core.Node NodeInfo *core.Node
// Number of publishAndIndex workers // Number of publishAndIndex workers
WorkerPoolSize int WorkerPoolSize int
// chain type for this service // chain type for this service
@ -96,7 +96,7 @@ type Service struct {
} }
// NewSuperNode creates a new super_node.Interface using an underlying super_node.Service struct // NewSuperNode creates a new super_node.Interface using an underlying super_node.Service struct
func NewSuperNode(settings *shared.SuperNodeConfig) (SuperNode, error) { func NewSuperNode(settings *Config) (SuperNode, error) {
sn := new(Service) sn := new(Service)
var err error var err error
// If we are syncing, initialize the needed interfaces // If we are syncing, initialize the needed interfaces
@ -137,7 +137,7 @@ func NewSuperNode(settings *shared.SuperNodeConfig) (SuperNode, error) {
sn.Subscriptions = make(map[common.Hash]map[rpc.ID]Subscription) sn.Subscriptions = make(map[common.Hash]map[rpc.ID]Subscription)
sn.SubscriptionTypes = make(map[common.Hash]shared.SubscriptionSettings) sn.SubscriptionTypes = make(map[common.Hash]shared.SubscriptionSettings)
sn.WorkerPoolSize = settings.Workers sn.WorkerPoolSize = settings.Workers
sn.NodeInfo = settings.NodeInfo sn.NodeInfo = &settings.NodeInfo
sn.ipfsPath = settings.IPFSPath sn.ipfsPath = settings.IPFSPath
sn.chain = settings.Chain sn.chain = settings.Chain
sn.db = settings.DB sn.db = settings.DB
@ -151,6 +151,7 @@ func (sap *Service) Protocols() []p2p.Protocol {
// APIs returns the RPC descriptors the super node service offers // APIs returns the RPC descriptors the super node service offers
func (sap *Service) APIs() []rpc.API { func (sap *Service) APIs() []rpc.API {
ifnoAPI := NewInfoAPI()
apis := []rpc.API{ apis := []rpc.API{
{ {
Namespace: APIName, Namespace: APIName,
@ -158,6 +159,24 @@ func (sap *Service) APIs() []rpc.API {
Service: NewPublicSuperNodeAPI(sap), Service: NewPublicSuperNodeAPI(sap),
Public: true, Public: true,
}, },
{
Namespace: "rpc",
Version: APIVersion,
Service: ifnoAPI,
Public: true,
},
{
Namespace: "net",
Version: APIVersion,
Service: ifnoAPI,
Public: true,
},
{
Namespace: "admin",
Version: APIVersion,
Service: ifnoAPI,
Public: true,
},
} }
chainAPI, err := NewPublicAPI(sap.chain, sap.db, sap.ipfsPath) chainAPI, err := NewPublicAPI(sap.chain, sap.db, sap.ipfsPath)
if err != nil { if err != nil {
@ -204,13 +223,13 @@ func (sap *Service) ProcessData(wg *sync.WaitGroup, screenAndServePayload chan<-
case err := <-sub.Err(): case err := <-sub.Err():
log.Error(err) log.Error(err)
case <-sap.QuitChan: case <-sap.QuitChan:
log.Info("quiting SyncAndPublish process") log.Infof("quiting %s SyncAndPublish process", sap.chain.String())
wg.Done() wg.Done()
return return
} }
} }
}() }()
log.Info("ProcessData goroutine successfully spun up") log.Infof("%s ProcessData goroutine successfully spun up", sap.chain.String())
return nil return nil
} }
@ -232,7 +251,7 @@ func (sap *Service) publishAndIndex(id int, publishAndIndexPayload <-chan shared
} }
} }
}() }()
log.Debugf("publishAndIndex goroutine successfully spun up") log.Debugf("%s publishAndIndex goroutine successfully spun up", sap.chain.String())
} }
// FilterAndServe listens for incoming converter data off the screenAndServePayload from the SyncAndConvert process // FilterAndServe listens for incoming converter data off the screenAndServePayload from the SyncAndConvert process
@ -247,24 +266,24 @@ func (sap *Service) FilterAndServe(wg *sync.WaitGroup, screenAndServePayload <-c
case payload := <-screenAndServePayload: case payload := <-screenAndServePayload:
sap.filterAndServe(payload) sap.filterAndServe(payload)
case <-sap.QuitChan: case <-sap.QuitChan:
log.Info("quiting ScreenAndServe process") log.Infof("quiting %s ScreenAndServe process", sap.chain.String())
wg.Done() wg.Done()
return return
} }
} }
}() }()
log.Info("FilterAndServe goroutine successfully spun up") log.Infof("%s FilterAndServe goroutine successfully spun up", sap.chain.String())
} }
// filterAndServe filters the payload according to each subscription type and sends to the subscriptions // filterAndServe filters the payload according to each subscription type and sends to the subscriptions
func (sap *Service) filterAndServe(payload shared.ConvertedData) { func (sap *Service) filterAndServe(payload shared.ConvertedData) {
log.Debugf("Sending payload to subscriptions") log.Debugf("Sending %s payload to subscriptions", sap.chain.String())
sap.Lock() sap.Lock()
for ty, subs := range sap.Subscriptions { for ty, subs := range sap.Subscriptions {
// Retrieve the subscription parameters for this subscription type // Retrieve the subscription parameters for this subscription type
subConfig, ok := sap.SubscriptionTypes[ty] subConfig, ok := sap.SubscriptionTypes[ty]
if !ok { if !ok {
log.Errorf("subscription configuration for subscription type %s not available", ty.Hex()) log.Errorf("%s subscription configuration for subscription type %s not available", sap.chain.String(), ty.Hex())
sap.closeType(ty) sap.closeType(ty)
continue continue
} }
@ -288,9 +307,9 @@ func (sap *Service) filterAndServe(payload shared.ConvertedData) {
for id, sub := range subs { for id, sub := range subs {
select { select {
case sub.PayloadChan <- SubscriptionPayload{Data: responseRLP, Err: "", Flag: EmptyFlag, Height: response.Height()}: case sub.PayloadChan <- SubscriptionPayload{Data: responseRLP, Err: "", Flag: EmptyFlag, Height: response.Height()}:
log.Debugf("sending super node payload to subscription %s", id) log.Debugf("sending super node %s payload to subscription %s", sap.chain.String(), id)
default: default:
log.Infof("unable to send payload to subscription %s; channel has no receiver", id) log.Infof("unable to send %s payload to subscription %s; channel has no receiver", sap.chain.String(), id)
} }
} }
} }
@ -300,7 +319,7 @@ func (sap *Service) filterAndServe(payload shared.ConvertedData) {
// Subscribe is used by the API to remotely subscribe to the service loop // Subscribe is used by the API to remotely subscribe to the service loop
// The params must be rlp serializable and satisfy the SubscriptionSettings() interface // The params must be rlp serializable and satisfy the SubscriptionSettings() interface
func (sap *Service) Subscribe(id rpc.ID, sub chan<- SubscriptionPayload, quitChan chan<- bool, params shared.SubscriptionSettings) { func (sap *Service) Subscribe(id rpc.ID, sub chan<- SubscriptionPayload, quitChan chan<- bool, params shared.SubscriptionSettings) {
log.Infof("New subscription %s", id) log.Infof("New %s subscription %s", sap.chain.String(), id)
subscription := Subscription{ subscription := Subscription{
ID: id, ID: id,
PayloadChan: sub, PayloadChan: sub,
@ -342,7 +361,7 @@ func (sap *Service) Subscribe(id rpc.ID, sub chan<- SubscriptionPayload, quitCha
// sendHistoricalData sends historical data to the requesting subscription // sendHistoricalData sends historical data to the requesting subscription
func (sap *Service) sendHistoricalData(sub Subscription, id rpc.ID, params shared.SubscriptionSettings) error { func (sap *Service) sendHistoricalData(sub Subscription, id rpc.ID, params shared.SubscriptionSettings) error {
log.Info("Sending historical data to subscription", id) log.Infof("Sending %s historical data to subscription %s", sap.chain.String(), id)
// Retrieve cached CIDs relevant to this subscriber // Retrieve cached CIDs relevant to this subscriber
var endingBlock int64 var endingBlock int64
var startingBlock int64 var startingBlock int64
@ -361,13 +380,13 @@ func (sap *Service) sendHistoricalData(sub Subscription, id rpc.ID, params share
if endingBlock > params.EndingBlock().Int64() && params.EndingBlock().Int64() > 0 && params.EndingBlock().Int64() > startingBlock { if endingBlock > params.EndingBlock().Int64() && params.EndingBlock().Int64() > 0 && params.EndingBlock().Int64() > startingBlock {
endingBlock = params.EndingBlock().Int64() endingBlock = params.EndingBlock().Int64()
} }
log.Debug("historical data starting block:", params.StartingBlock()) log.Debugf("%s historical data starting block: %d", sap.chain.String(), params.StartingBlock().Int64())
log.Debug("historical data ending block:", endingBlock) log.Debugf("%s historical data ending block: %d", sap.chain.String(), endingBlock)
go func() { go func() {
for i := startingBlock; i <= endingBlock; i++ { for i := startingBlock; i <= endingBlock; i++ {
cidWrappers, empty, err := sap.Retriever.Retrieve(params, i) cidWrappers, empty, err := sap.Retriever.Retrieve(params, i)
if err != nil { if err != nil {
sendNonBlockingErr(sub, fmt.Errorf("CID Retrieval error at block %d\r%s", i, err.Error())) sendNonBlockingErr(sub, fmt.Errorf("%s CID Retrieval error at block %d\r%s", sap.chain.String(), i, err.Error()))
continue continue
} }
if empty { if empty {
@ -376,7 +395,7 @@ func (sap *Service) sendHistoricalData(sub Subscription, id rpc.ID, params share
for _, cids := range cidWrappers { for _, cids := range cidWrappers {
response, err := sap.IPLDFetcher.Fetch(cids) response, err := sap.IPLDFetcher.Fetch(cids)
if err != nil { if err != nil {
sendNonBlockingErr(sub, fmt.Errorf("IPLD Fetching error at block %d\r%s", i, err.Error())) sendNonBlockingErr(sub, fmt.Errorf("%s IPLD Fetching error at block %d\r%s", sap.chain.String(), i, err.Error()))
continue continue
} }
responseRLP, err := rlp.EncodeToBytes(response) responseRLP, err := rlp.EncodeToBytes(response)
@ -386,18 +405,18 @@ func (sap *Service) sendHistoricalData(sub Subscription, id rpc.ID, params share
} }
select { select {
case sub.PayloadChan <- SubscriptionPayload{Data: responseRLP, Err: "", Flag: EmptyFlag, Height: response.Height()}: case sub.PayloadChan <- SubscriptionPayload{Data: responseRLP, Err: "", Flag: EmptyFlag, Height: response.Height()}:
log.Debugf("sending super node historical data payload to subscription %s", id) log.Debugf("sending super node historical data payload to %s subscription %s", sap.chain.String(), id)
default: default:
log.Infof("unable to send back-fill payload to subscription %s; channel has no receiver", id) log.Infof("unable to send back-fill payload to %s subscription %s; channel has no receiver", sap.chain.String(), id)
} }
} }
} }
// when we are done backfilling send an empty payload signifying so in the msg // when we are done backfilling send an empty payload signifying so in the msg
select { select {
case sub.PayloadChan <- SubscriptionPayload{Data: nil, Err: "", Flag: BackFillCompleteFlag}: case sub.PayloadChan <- SubscriptionPayload{Data: nil, Err: "", Flag: BackFillCompleteFlag}:
log.Debugf("sending backfill completion notice to subscription %s", id) log.Debugf("sending backfill completion notice to %s subscription %s", sap.chain.String(), id)
default: default:
log.Infof("unable to send backfill completion notice to subscription %s", id) log.Infof("unable to send backfill completion notice to %s subscription %s", sap.chain.String(), id)
} }
}() }()
return nil return nil
@ -405,7 +424,7 @@ func (sap *Service) sendHistoricalData(sub Subscription, id rpc.ID, params share
// Unsubscribe is used by the API to remotely unsubscribe to the StateDiffingService loop // Unsubscribe is used by the API to remotely unsubscribe to the StateDiffingService loop
func (sap *Service) Unsubscribe(id rpc.ID) { func (sap *Service) Unsubscribe(id rpc.ID) {
log.Infof("Unsubscribing %s from the super node service", id) log.Infof("Unsubscribing %s from the %s super node service", id, sap.chain.String())
sap.Lock() sap.Lock()
for ty := range sap.Subscriptions { for ty := range sap.Subscriptions {
delete(sap.Subscriptions[ty], id) delete(sap.Subscriptions[ty], id)
@ -421,7 +440,7 @@ func (sap *Service) Unsubscribe(id rpc.ID) {
// Start is used to begin the service // Start is used to begin the service
// This is mostly just to satisfy the node.Service interface // This is mostly just to satisfy the node.Service interface
func (sap *Service) Start(*p2p.Server) error { func (sap *Service) Start(*p2p.Server) error {
log.Info("Starting super node service") log.Infof("Starting %s super node service", sap.chain.String())
wg := new(sync.WaitGroup) wg := new(sync.WaitGroup)
payloadChan := make(chan shared.ConvertedData, PayloadChanBufferSize) payloadChan := make(chan shared.ConvertedData, PayloadChanBufferSize)
if err := sap.ProcessData(wg, payloadChan); err != nil { if err := sap.ProcessData(wg, payloadChan); err != nil {
@ -434,7 +453,7 @@ func (sap *Service) Start(*p2p.Server) error {
// Stop is used to close down the service // Stop is used to close down the service
// This is mostly just to satisfy the node.Service interface // This is mostly just to satisfy the node.Service interface
func (sap *Service) Stop() error { func (sap *Service) Stop() error {
log.Info("Stopping super node service") log.Infof("Stopping %s super node service", sap.chain.String())
sap.Lock() sap.Lock()
close(sap.QuitChan) close(sap.QuitChan)
sap.close() sap.close()
@ -443,7 +462,7 @@ func (sap *Service) Stop() error {
} }
// Node returns the node info for this service // Node returns the node info for this service
func (sap *Service) Node() core.Node { func (sap *Service) Node() *core.Node {
return sap.NodeInfo return sap.NodeInfo
} }
@ -455,7 +474,7 @@ func (sap *Service) Chain() shared.ChainType {
// close is used to close all listening subscriptions // close is used to close all listening subscriptions
// close needs to be called with subscription access locked // close needs to be called with subscription access locked
func (sap *Service) close() { func (sap *Service) close() {
log.Info("Closing all subscriptions") log.Infof("Closing all %s subscriptions", sap.chain.String())
for subType, subs := range sap.Subscriptions { for subType, subs := range sap.Subscriptions {
for _, sub := range subs { for _, sub := range subs {
sendNonBlockingQuit(sub) sendNonBlockingQuit(sub)
@ -468,7 +487,7 @@ func (sap *Service) close() {
// closeType is used to close all subscriptions of given type // closeType is used to close all subscriptions of given type
// closeType needs to be called with subscription access locked // closeType needs to be called with subscription access locked
func (sap *Service) closeType(subType common.Hash) { func (sap *Service) closeType(subType common.Hash) {
log.Infof("Closing all subscriptions of type %s", subType.String()) log.Infof("Closing all %s subscriptions of type %s", sap.chain.String(), subType.String())
subs := sap.Subscriptions[subType] subs := sap.Subscriptions[subType]
for _, sub := range subs { for _, sub := range subs {
sendNonBlockingQuit(sub) sendNonBlockingQuit(sub)

63
pkg/super_node/version.go Normal file
View File

@ -0,0 +1,63 @@
// VulcanizeDB
// Copyright © 2019 Vulcanize
// This program is free software: you can redistribute it and/or modify
// it under the terms of the GNU Affero General Public License as published by
// the Free Software Foundation, either version 3 of the License, or
// (at your option) any later version.
// This program is distributed in the hope that it will be useful,
// but WITHOUT ANY WARRANTY; without even the implied warranty of
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
// GNU Affero General Public License for more details.
// You should have received a copy of the GNU Affero General Public License
// along with this program. If not, see <http://www.gnu.org/licenses/>.
package super_node
import "fmt"
const (
VersionMajor = 0 // Major version component of the current release
VersionMinor = 1 // Minor version component of the current release
VersionPatch = 0 // Patch version component of the current release
VersionMeta = "alpha" // Version metadata to append to the version string
)
// Version holds the textual version string.
var Version = func() string {
return fmt.Sprintf("%d.%d.%d", VersionMajor, VersionMinor, VersionPatch)
}()
// VersionWithMeta holds the textual version string including the metadata.
var VersionWithMeta = func() string {
v := Version
if VersionMeta != "" {
v += "-" + VersionMeta
}
return v
}()
// ArchiveVersion holds the textual version string
func ArchiveVersion(gitCommit string) string {
vsn := Version
if VersionMeta != "stable" {
vsn += "-" + VersionMeta
}
if len(gitCommit) >= 8 {
vsn += "-" + gitCommit[:8]
}
return vsn
}
func VersionWithCommit(gitCommit, gitDate string) string {
vsn := VersionWithMeta
if len(gitCommit) >= 8 {
vsn += "-" + gitCommit[:8]
}
if (VersionMeta != "stable") && (gitDate != "") {
vsn += "-" + gitDate
}
return vsn
}

View File

@ -16,16 +16,23 @@
package wasm package wasm
import "github.com/vulcanize/vulcanizedb/pkg/postgres" import (
"github.com/vulcanize/vulcanizedb/pkg/postgres"
)
// Instantiator is used to instantiate WASM functions in Postgres // Instantiator is used to instantiate WASM functions in Postgres
type Instantiator struct { type Instantiator struct {
db *postgres.DB db *postgres.DB
instances [][2]string // list of WASM file paths and namespaces instances []WasmFunction // WASM file paths and namespaces
}
type WasmFunction struct {
BinaryPath string
Namespace string
} }
// NewWASMInstantiator returns a pointer to a new Instantiator // NewWASMInstantiator returns a pointer to a new Instantiator
func NewWASMInstantiator(db *postgres.DB, instances [][2]string) *Instantiator { func NewWASMInstantiator(db *postgres.DB, instances []WasmFunction) *Instantiator {
return &Instantiator{ return &Instantiator{
db: db, db: db,
instances: instances, instances: instances,
@ -39,7 +46,7 @@ func (i *Instantiator) Instantiate() error {
return err return err
} }
for _, pn := range i.instances { for _, pn := range i.instances {
_, err := tx.Exec(`SELECT wasm_new_instance('$1', '$2')`, pn[0], pn[1]) _, err := tx.Exec(`SELECT wasm_new_instance('$1', '$2')`, pn.BinaryPath, pn.Namespace)
if err != nil { if err != nil {
return err return err
} }

View File

@ -17,10 +17,24 @@
package watcher package watcher
import ( import (
"context"
"errors"
"fmt"
"github.com/vulcanize/vulcanizedb/pkg/wasm"
"github.com/ethereum/go-ethereum/rpc"
"github.com/spf13/viper"
"github.com/vulcanize/vulcanizedb/pkg/config" "github.com/vulcanize/vulcanizedb/pkg/config"
"github.com/vulcanize/vulcanizedb/pkg/eth/client"
"github.com/vulcanize/vulcanizedb/pkg/eth/core" "github.com/vulcanize/vulcanizedb/pkg/eth/core"
"github.com/vulcanize/vulcanizedb/pkg/postgres" "github.com/vulcanize/vulcanizedb/pkg/postgres"
"github.com/vulcanize/vulcanizedb/pkg/super_node/btc"
"github.com/vulcanize/vulcanizedb/pkg/super_node/eth"
"github.com/vulcanize/vulcanizedb/pkg/super_node/shared" "github.com/vulcanize/vulcanizedb/pkg/super_node/shared"
shared2 "github.com/vulcanize/vulcanizedb/pkg/watcher/shared"
"github.com/vulcanize/vulcanizedb/utils"
) )
// Config holds all of the parameters necessary for defining and running an instance of a watcher // Config holds all of the parameters necessary for defining and running an instance of a watcher
@ -32,9 +46,93 @@ type Config struct {
// DB itself // DB itself
DB *postgres.DB DB *postgres.DB
// Subscription client // Subscription client
Client core.RPCClient Client interface{}
// WASM instantiation paths and namespaces // WASM instantiation paths and namespaces
WASMInstances [][2]string WASMFunctions []wasm.WasmFunction
// Path and names for trigger functions (sql files) that (can) use the instantiated wasm namespaces // File paths for trigger functions (sql files) that (can) use the instantiated wasm namespaces
TriggerFunctions [][2]string TriggerFunctions []string
// Chain type used to specify what type of raw data we will be processing
Chain shared.ChainType
// Source type used to specify which streamer to use based on what API we will be interfacing with
Source shared2.SourceType
// Info for the node
NodeInfo core.Node
}
func NewWatcherConfig() (*Config, error) {
c := new(Config)
var err error
chain := viper.GetString("watcher.chain")
c.Chain, err = shared.NewChainType(chain)
if err != nil {
return nil, err
}
switch c.Chain {
case shared.Ethereum:
c.SubscriptionConfig, err = eth.NewEthSubscriptionConfig()
if err != nil {
return nil, err
}
case shared.Bitcoin:
c.SubscriptionConfig, err = btc.NewEthSubscriptionConfig()
if err != nil {
return nil, err
}
case shared.Omni:
return nil, errors.New("omni chain type currently not supported")
default:
return nil, fmt.Errorf("unexpected chain type %s", c.Chain.String())
}
sourcePath := viper.GetString("watcher.dataSource")
if sourcePath == "" {
sourcePath = "ws://127.0.0.1:8080" // default to and try the default ws url if no path is provided
}
sourceType := viper.GetString("watcher.dataPath")
c.Source, err = shared2.NewSourceType(sourceType)
if err != nil {
return nil, err
}
switch c.Source {
case shared2.Ethereum:
return nil, errors.New("ethereum data source currently not supported")
case shared2.Bitcoin:
return nil, errors.New("bitcoin data source currently not supported")
case shared2.VulcanizeDB:
rawRPCClient, err := rpc.Dial(sourcePath)
if err != nil {
return nil, err
}
cli := client.NewRPCClient(rawRPCClient, sourcePath)
var nodeInfo core.Node
if err := cli.CallContext(context.Background(), &nodeInfo, "vdb_node"); err != nil {
return nil, err
}
c.NodeInfo = nodeInfo
c.Client = cli
default:
return nil, fmt.Errorf("unexpected data source type %s", c.Source.String())
}
wasmBinaries := viper.GetStringSlice("watcher.wasmBinaries")
wasmNamespaces := viper.GetStringSlice("watcher.wasmNamespaces")
if len(wasmBinaries) != len(wasmNamespaces) {
return nil, fmt.Errorf("watcher config needs a namespace for every wasm binary\r\nhave %d binaries and %d namespaces", len(wasmBinaries), len(wasmNamespaces))
}
c.WASMFunctions = make([]wasm.WasmFunction, len(wasmBinaries))
for i, bin := range wasmBinaries {
c.WASMFunctions[i] = wasm.WasmFunction{
BinaryPath: bin,
Namespace: wasmNamespaces[i],
}
}
c.TriggerFunctions = viper.GetStringSlice("watcher.triggerFunctions")
c.DBConfig = config.Database{
Name: viper.GetString("watcher.database.name"),
Hostname: viper.GetString("watcher.database.hostname"),
Port: viper.GetInt("watcher.database.port"),
User: viper.GetString("watcher.database.user"),
Password: viper.GetString("watcher.database.password"),
}
db := utils.LoadPostgres(c.DBConfig, c.NodeInfo)
c.DB = &db
return c, nil
} }

View File

@ -28,12 +28,22 @@ import (
) )
// NewSuperNodeStreamer returns a new shared.SuperNodeStreamer // NewSuperNodeStreamer returns a new shared.SuperNodeStreamer
func NewSuperNodeStreamer(client core.RPCClient) shared.SuperNodeStreamer { func NewSuperNodeStreamer(source shared.SourceType, client interface{}) (shared.SuperNodeStreamer, error) {
return streamer.NewSuperNodeStreamer(client) switch source {
case shared.VulcanizeDB:
cli, ok := client.(core.RPCClient)
if !ok {
var expectedClientType core.RPCClient
return nil, fmt.Errorf("vulcanizedb NewSuperNodeStreamer construct expects client type %T got %T", expectedClientType, client)
}
return streamer.NewSuperNodeStreamer(cli), nil
default:
return nil, fmt.Errorf("NewSuperNodeStreamer constructor unexpected souce type %s", source.String())
}
} }
// NewRepository constructs and returns a new Repository that satisfies the shared.Repository interface for the specified chain // NewRepository constructs and returns a new Repository that satisfies the shared.Repository interface for the specified chain
func NewRepository(chain shared2.ChainType, db *postgres.DB, triggerFuncs [][2]string) (shared.Repository, error) { func NewRepository(chain shared2.ChainType, db *postgres.DB, triggerFuncs []string) (shared.Repository, error) {
switch chain { switch chain {
case shared2.Ethereum: case shared2.Ethereum:
return eth.NewRepository(db, triggerFuncs), nil return eth.NewRepository(db, triggerFuncs), nil

View File

@ -32,12 +32,12 @@ var (
// Repository is the underlying struct for satisfying the shared.Repository interface for eth // Repository is the underlying struct for satisfying the shared.Repository interface for eth
type Repository struct { type Repository struct {
db *postgres.DB db *postgres.DB
triggerFunctions [][2]string triggerFunctions []string
deleteCalls int64 deleteCalls int64
} }
// NewRepository returns a new eth.Repository that satisfies the shared.Repository interface // NewRepository returns a new eth.Repository that satisfies the shared.Repository interface
func NewRepository(db *postgres.DB, triggerFunctions [][2]string) shared.Repository { func NewRepository(db *postgres.DB, triggerFunctions []string) shared.Repository {
return &Repository{ return &Repository{
db: db, db: db,
triggerFunctions: triggerFunctions, triggerFunctions: triggerFunctions,

View File

@ -40,7 +40,7 @@ type Watcher interface {
// Service is the underlying struct for the SuperNodeWatcher // Service is the underlying struct for the SuperNodeWatcher
type Service struct { type Service struct {
// Config // Config
WatcherConfig Config WatcherConfig *Config
// Interface for streaming data from super node // Interface for streaming data from super node
SuperNodeStreamer shared.SuperNodeStreamer SuperNodeStreamer shared.SuperNodeStreamer
// Interface for db operations // Interface for db operations
@ -60,16 +60,20 @@ type Service struct {
} }
// NewWatcher returns a new Service which satisfies the Watcher interface // NewWatcher returns a new Service which satisfies the Watcher interface
func NewWatcher(c Config, quitChan chan bool) (Watcher, error) { func NewWatcher(c *Config, quitChan chan bool) (Watcher, error) {
repo, err := NewRepository(c.SubscriptionConfig.ChainType(), c.DB, c.TriggerFunctions) repo, err := NewRepository(c.SubscriptionConfig.ChainType(), c.DB, c.TriggerFunctions)
if err != nil { if err != nil {
return nil, err return nil, err
} }
streamer, err := NewSuperNodeStreamer(c.Source, c.Client)
if err != nil {
return nil, err
}
return &Service{ return &Service{
WatcherConfig: c, WatcherConfig: c,
SuperNodeStreamer: NewSuperNodeStreamer(c.Client), SuperNodeStreamer: streamer,
Repository: repo, Repository: repo,
WASMIniter: wasm.NewWASMInstantiator(c.DB, c.WASMInstances), WASMIniter: wasm.NewWASMInstantiator(c.DB, c.WASMFunctions),
PayloadChan: make(chan super_node.SubscriptionPayload, super_node.PayloadChanBufferSize), PayloadChan: make(chan super_node.SubscriptionPayload, super_node.PayloadChanBufferSize),
QuitChan: quitChan, QuitChan: quitChan,
}, nil }, nil
@ -85,7 +89,7 @@ func (s *Service) Init() error {
return s.Repository.LoadTriggers() return s.Repository.LoadTriggers()
} }
// Watch is the top level loop for watching super node // Watch is the top level loop for watching
func (s *Service) Watch(wg *sync.WaitGroup) error { func (s *Service) Watch(wg *sync.WaitGroup) error {
rlpConfig, err := rlp.EncodeToBytes(s.WatcherConfig.SubscriptionConfig) rlpConfig, err := rlp.EncodeToBytes(s.WatcherConfig.SubscriptionConfig)
if err != nil { if err != nil {

View File

@ -0,0 +1,58 @@
// VulcanizeDB
// Copyright © 2019 Vulcanize
// This program is free software: you can redistribute it and/or modify
// it under the terms of the GNU Affero General Public License as published by
// the Free Software Foundation, either version 3 of the License, or
// (at your option) any later version.
// This program is distributed in the hope that it will be useful,
// but WITHOUT ANY WARRANTY; without even the implied warranty of
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
// GNU Affero General Public License for more details.
// You should have received a copy of the GNU Affero General Public License
// along with this program. If not, see <http://www.gnu.org/licenses/>.
package shared
import (
"errors"
"strings"
)
// SourceType enum for specifying source type for raw chain data
type SourceType int
const (
Unknown SourceType = iota
VulcanizeDB
Ethereum
Bitcoin
)
func (c SourceType) String() string {
switch c {
case Ethereum:
return "Ethereum"
case Bitcoin:
return "Bitcoin"
case VulcanizeDB:
return "VulcanizeDB"
default:
return ""
}
}
func NewSourceType(name string) (SourceType, error) {
switch strings.ToLower(name) {
case "ethereum", "eth":
return Ethereum, nil
case "bitcoin", "btc", "xbt":
return Bitcoin, nil
case "vulcanizedb", "vdb":
return VulcanizeDB, nil
default:
return Unknown, errors.New("invalid name for data source")
}
}