misc fixes; config files for resyncing
This commit is contained in:
parent
d47ba24373
commit
c72dc273ba
@ -18,7 +18,10 @@ package cmd
|
|||||||
import (
|
import (
|
||||||
"fmt"
|
"fmt"
|
||||||
|
|
||||||
|
log "github.com/sirupsen/logrus"
|
||||||
"github.com/spf13/cobra"
|
"github.com/spf13/cobra"
|
||||||
|
|
||||||
|
"github.com/vulcanize/vulcanizedb/pkg/ipfs"
|
||||||
"github.com/vulcanize/vulcanizedb/pkg/super_node/resync"
|
"github.com/vulcanize/vulcanizedb/pkg/super_node/resync"
|
||||||
)
|
)
|
||||||
|
|
||||||
@ -28,15 +31,24 @@ var resyncCmd = &cobra.Command{
|
|||||||
Short: "Resync historical data",
|
Short: "Resync historical data",
|
||||||
Long: `Use this command to fill in sections of missing data in the super node`,
|
Long: `Use this command to fill in sections of missing data in the super node`,
|
||||||
Run: func(cmd *cobra.Command, args []string) {
|
Run: func(cmd *cobra.Command, args []string) {
|
||||||
|
subCommand = cmd.CalledAs()
|
||||||
|
logWithCommand = *log.WithField("SubCommand", subCommand)
|
||||||
rsyncCmdCommand()
|
rsyncCmdCommand()
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func init() {
|
||||||
|
rootCmd.AddCommand(resyncCmd)
|
||||||
|
}
|
||||||
|
|
||||||
func rsyncCmdCommand() {
|
func rsyncCmdCommand() {
|
||||||
rConfig, err := resync.NewReSyncConfig()
|
rConfig, err := resync.NewReSyncConfig()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
logWithCommand.Fatal(err)
|
logWithCommand.Fatal(err)
|
||||||
}
|
}
|
||||||
|
if err := ipfs.InitIPFSPlugins(); err != nil {
|
||||||
|
logWithCommand.Fatal(err)
|
||||||
|
}
|
||||||
rService, err := resync.NewResyncService(rConfig)
|
rService, err := resync.NewResyncService(rConfig)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
logWithCommand.Fatal(err)
|
logWithCommand.Fatal(err)
|
||||||
@ -46,7 +58,3 @@ func rsyncCmdCommand() {
|
|||||||
}
|
}
|
||||||
fmt.Printf("%s %s resync finished", rConfig.Chain.String(), rConfig.ResyncType.String())
|
fmt.Printf("%s %s resync finished", rConfig.Chain.String(), rConfig.ResyncType.String())
|
||||||
}
|
}
|
||||||
|
|
||||||
func init() {
|
|
||||||
rootCmd.AddCommand(resyncCmd)
|
|
||||||
}
|
|
||||||
|
24
environments/resyncBTC.toml
Normal file
24
environments/resyncBTC.toml
Normal file
@ -0,0 +1,24 @@
|
|||||||
|
[database]
|
||||||
|
name = "vulcanize_public"
|
||||||
|
hostname = "localhost"
|
||||||
|
port = 5432
|
||||||
|
user = "vdbm"
|
||||||
|
|
||||||
|
[resync]
|
||||||
|
chain = "bitcoin"
|
||||||
|
type = "full"
|
||||||
|
clearOldCache = true
|
||||||
|
ipfsPath = "/root/.ipfs"
|
||||||
|
batchSize = 1
|
||||||
|
batchNumber = 50
|
||||||
|
start = 0
|
||||||
|
stop = 0
|
||||||
|
|
||||||
|
[bitcoin]
|
||||||
|
httpPath = "127.0.0.1:8332"
|
||||||
|
pass = "password"
|
||||||
|
user = "username"
|
||||||
|
nodeID = "ocd0"
|
||||||
|
clientName = "Omnicore"
|
||||||
|
genesisBlock = "000000000019d6689c085ae165831e934ff763ae46a2a6c172b3f1b60a8ce26f"
|
||||||
|
networkID = "0xD9B4BEF9"
|
18
environments/resyncETH.toml
Normal file
18
environments/resyncETH.toml
Normal file
@ -0,0 +1,18 @@
|
|||||||
|
[database]
|
||||||
|
name = "vulcanize_public"
|
||||||
|
hostname = "localhost"
|
||||||
|
port = 5432
|
||||||
|
user = "vdbm"
|
||||||
|
|
||||||
|
[resync]
|
||||||
|
chain = "ethereum"
|
||||||
|
type = "state"
|
||||||
|
clearOldCache = true
|
||||||
|
ipfsPath = "/root/.ipfs"
|
||||||
|
batchSize = 5
|
||||||
|
batchNumber = 50
|
||||||
|
start = 0
|
||||||
|
stop = 0
|
||||||
|
|
||||||
|
[ethereum]
|
||||||
|
httpPath = "127.0.0.1:8545"
|
@ -14,8 +14,9 @@
|
|||||||
sync = true
|
sync = true
|
||||||
workers = 1
|
workers = 1
|
||||||
backFill = true
|
backFill = true
|
||||||
frequency = 15
|
frequency = 45
|
||||||
batchSize = 50
|
batchSize = 1
|
||||||
|
batchNumber = 50
|
||||||
|
|
||||||
[bitcoin]
|
[bitcoin]
|
||||||
wsPath = "127.0.0.1:8332"
|
wsPath = "127.0.0.1:8332"
|
||||||
|
@ -9,14 +9,15 @@
|
|||||||
ipfsPath = "/root/.ipfs"
|
ipfsPath = "/root/.ipfs"
|
||||||
server = true
|
server = true
|
||||||
ipcPath = "/root/.vulcanize/eth/vulcanize.ipc"
|
ipcPath = "/root/.vulcanize/eth/vulcanize.ipc"
|
||||||
wsPath = "127.0.0.1:8080"
|
wsPath = "127.0.0.1:8081"
|
||||||
httpPath = "127.0.0.1:8081"
|
httpPath = "127.0.0.1:8082"
|
||||||
sync = true
|
sync = true
|
||||||
workers = 1
|
workers = 1
|
||||||
backFill = true
|
backFill = true
|
||||||
frequency = 15
|
frequency = 15
|
||||||
batchSize = 50
|
batchSize = 5
|
||||||
|
batchNumber = 50
|
||||||
|
|
||||||
[ethereum]
|
[ethereum]
|
||||||
wsPath = "ws://127.0.0.1:8546"
|
wsPath = "127.0.0.1:8546"
|
||||||
httpPath = "http://127.0.0.1:8545"
|
httpPath = "127.0.0.1:8545"
|
@ -30,7 +30,7 @@ import (
|
|||||||
|
|
||||||
const (
|
const (
|
||||||
DefaultMaxBatchSize uint64 = 100
|
DefaultMaxBatchSize uint64 = 100
|
||||||
DefaultMaxBatchNumber int64 = 10
|
DefaultMaxBatchNumber int64 = 50
|
||||||
)
|
)
|
||||||
|
|
||||||
// BackFillInterface for filling in gaps in the super node
|
// BackFillInterface for filling in gaps in the super node
|
||||||
@ -57,6 +57,8 @@ type BackFillService struct {
|
|||||||
GapCheckFrequency time.Duration
|
GapCheckFrequency time.Duration
|
||||||
// Size of batch fetches
|
// Size of batch fetches
|
||||||
BatchSize uint64
|
BatchSize uint64
|
||||||
|
// Number of goroutines
|
||||||
|
BatchNumber int64
|
||||||
// Channel for receiving quit signal
|
// Channel for receiving quit signal
|
||||||
QuitChan chan bool
|
QuitChan chan bool
|
||||||
// Chain type
|
// Chain type
|
||||||
@ -89,6 +91,10 @@ func NewBackFillService(settings *Config, screenAndServeChan chan shared.Convert
|
|||||||
if batchSize == 0 {
|
if batchSize == 0 {
|
||||||
batchSize = DefaultMaxBatchSize
|
batchSize = DefaultMaxBatchSize
|
||||||
}
|
}
|
||||||
|
batchNumber := int64(settings.BatchNumber)
|
||||||
|
if batchNumber == 0 {
|
||||||
|
batchNumber = DefaultMaxBatchNumber
|
||||||
|
}
|
||||||
return &BackFillService{
|
return &BackFillService{
|
||||||
Indexer: indexer,
|
Indexer: indexer,
|
||||||
Converter: converter,
|
Converter: converter,
|
||||||
@ -97,6 +103,7 @@ func NewBackFillService(settings *Config, screenAndServeChan chan shared.Convert
|
|||||||
Fetcher: fetcher,
|
Fetcher: fetcher,
|
||||||
GapCheckFrequency: settings.Frequency,
|
GapCheckFrequency: settings.Frequency,
|
||||||
BatchSize: batchSize,
|
BatchSize: batchSize,
|
||||||
|
BatchNumber: int64(batchNumber),
|
||||||
ScreenAndServeChan: screenAndServeChan,
|
ScreenAndServeChan: screenAndServeChan,
|
||||||
QuitChan: settings.Quit,
|
QuitChan: settings.Quit,
|
||||||
chain: settings.Chain,
|
chain: settings.Chain,
|
||||||
@ -112,7 +119,7 @@ func (bfs *BackFillService) FillGapsInSuperNode(wg *sync.WaitGroup) {
|
|||||||
for {
|
for {
|
||||||
select {
|
select {
|
||||||
case <-bfs.QuitChan:
|
case <-bfs.QuitChan:
|
||||||
log.Infof("quiting %s FillGaps process", bfs.chain.String())
|
log.Infof("quiting %s FillGapsInSuperNode process", bfs.chain.String())
|
||||||
wg.Done()
|
wg.Done()
|
||||||
return
|
return
|
||||||
case <-ticker.C:
|
case <-ticker.C:
|
||||||
@ -124,7 +131,9 @@ func (bfs *BackFillService) FillGapsInSuperNode(wg *sync.WaitGroup) {
|
|||||||
}
|
}
|
||||||
if startingBlock != 0 {
|
if startingBlock != 0 {
|
||||||
log.Infof("found gap at the beginning of the %s sync", bfs.chain.String())
|
log.Infof("found gap at the beginning of the %s sync", bfs.chain.String())
|
||||||
bfs.fillGaps(0, uint64(startingBlock-1))
|
if err := bfs.backFill(0, uint64(startingBlock-1)); err != nil {
|
||||||
|
log.Error(err)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
gaps, err := bfs.Retriever.RetrieveGapsInData()
|
gaps, err := bfs.Retriever.RetrieveGapsInData()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@ -132,7 +141,7 @@ func (bfs *BackFillService) FillGapsInSuperNode(wg *sync.WaitGroup) {
|
|||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
for _, gap := range gaps {
|
for _, gap := range gaps {
|
||||||
if err := bfs.fillGaps(gap.Start, gap.Stop); err != nil {
|
if err := bfs.backFill(gap.Start, gap.Stop); err != nil {
|
||||||
log.Error(err)
|
log.Error(err)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@ -142,27 +151,10 @@ func (bfs *BackFillService) FillGapsInSuperNode(wg *sync.WaitGroup) {
|
|||||||
log.Infof("%s fillGaps goroutine successfully spun up", bfs.chain.String())
|
log.Infof("%s fillGaps goroutine successfully spun up", bfs.chain.String())
|
||||||
}
|
}
|
||||||
|
|
||||||
func (bfs *BackFillService) fillGaps(startingBlock, endingBlock uint64) error {
|
|
||||||
log.Infof("going to fill in %s gap from %d to %d", bfs.chain.String(), startingBlock, endingBlock)
|
|
||||||
errChan := make(chan error)
|
|
||||||
done := make(chan bool)
|
|
||||||
if err := bfs.backFill(startingBlock, endingBlock, errChan, done); err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
for {
|
|
||||||
select {
|
|
||||||
case err := <-errChan:
|
|
||||||
log.Errorf("super node db backfill error for chain %s: %v", bfs.chain.String(), err)
|
|
||||||
case <-done:
|
|
||||||
log.Infof("finished filling in %s gap from %d to %d", bfs.chain.String(), startingBlock, endingBlock)
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// backFill fetches, processes, and returns utils.StorageDiffs over a range of blocks
|
// backFill fetches, processes, and returns utils.StorageDiffs over a range of blocks
|
||||||
// It splits a large range up into smaller chunks, batch fetching and processing those chunks concurrently
|
// It splits a large range up into smaller chunks, batch fetching and processing those chunks concurrently
|
||||||
func (bfs *BackFillService) backFill(startingBlock, endingBlock uint64, errChan chan error, done chan bool) error {
|
func (bfs *BackFillService) backFill(startingBlock, endingBlock uint64) error {
|
||||||
|
log.Infof("filling in %s gap from %d to %d", bfs.chain.String(), startingBlock, endingBlock)
|
||||||
if endingBlock < startingBlock {
|
if endingBlock < startingBlock {
|
||||||
return fmt.Errorf("super node %s db backfill: ending block number needs to be greater than starting block number", bfs.chain.String())
|
return fmt.Errorf("super node %s db backfill: ending block number needs to be greater than starting block number", bfs.chain.String())
|
||||||
}
|
}
|
||||||
@ -175,28 +167,27 @@ func (bfs *BackFillService) backFill(startingBlock, endingBlock uint64, errChan
|
|||||||
// int64 for atomic incrementing and decrementing to track the number of active processing goroutines we have
|
// int64 for atomic incrementing and decrementing to track the number of active processing goroutines we have
|
||||||
var activeCount int64
|
var activeCount int64
|
||||||
// channel for processing goroutines to signal when they are done
|
// channel for processing goroutines to signal when they are done
|
||||||
processingDone := make(chan [2]uint64)
|
processingDone := make(chan bool)
|
||||||
forwardDone := make(chan bool)
|
forwardDone := make(chan bool)
|
||||||
|
|
||||||
// for each block range bin spin up a goroutine to batch fetch and process state diffs for that range
|
// for each block range bin spin up a goroutine to batch fetch and process data for that range
|
||||||
go func() {
|
go func() {
|
||||||
for _, blockHeights := range blockRangeBins {
|
for _, blockHeights := range blockRangeBins {
|
||||||
// if we have reached our limit of active goroutines
|
// if we have reached our limit of active goroutines
|
||||||
// wait for one to finish before starting the next
|
// wait for one to finish before starting the next
|
||||||
if atomic.AddInt64(&activeCount, 1) > DefaultMaxBatchNumber {
|
if atomic.AddInt64(&activeCount, 1) > bfs.BatchNumber {
|
||||||
// this blocks until a process signals it has finished
|
// this blocks until a process signals it has finished
|
||||||
<-forwardDone
|
<-forwardDone
|
||||||
}
|
}
|
||||||
go func(blockHeights []uint64) {
|
go func(blockHeights []uint64) {
|
||||||
payloads, err := bfs.Fetcher.FetchAt(blockHeights)
|
payloads, err := bfs.Fetcher.FetchAt(blockHeights)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
errChan <- err
|
log.Errorf("%s super node historical data fetcher error: %s", bfs.chain.String(), err.Error())
|
||||||
}
|
}
|
||||||
for _, payload := range payloads {
|
for _, payload := range payloads {
|
||||||
ipldPayload, err := bfs.Converter.Convert(payload)
|
ipldPayload, err := bfs.Converter.Convert(payload)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
errChan <- err
|
log.Errorf("%s super node historical data converter error: %s", bfs.chain.String(), err.Error())
|
||||||
continue
|
|
||||||
}
|
}
|
||||||
// If there is a ScreenAndServe process listening, forward payload to it
|
// If there is a ScreenAndServe process listening, forward payload to it
|
||||||
select {
|
select {
|
||||||
@ -205,42 +196,36 @@ func (bfs *BackFillService) backFill(startingBlock, endingBlock uint64, errChan
|
|||||||
}
|
}
|
||||||
cidPayload, err := bfs.Publisher.Publish(ipldPayload)
|
cidPayload, err := bfs.Publisher.Publish(ipldPayload)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
errChan <- err
|
log.Errorf("%s super node historical data publisher error: %s", bfs.chain.String(), err.Error())
|
||||||
continue
|
|
||||||
}
|
}
|
||||||
if err := bfs.Indexer.Index(cidPayload); err != nil {
|
if err := bfs.Indexer.Index(cidPayload); err != nil {
|
||||||
errChan <- err
|
log.Errorf("%s super node historical data indexer error: %s", bfs.chain.String(), err.Error())
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
// when this goroutine is done, send out a signal
|
// when this goroutine is done, send out a signal
|
||||||
processingDone <- [2]uint64{blockHeights[0], blockHeights[len(blockHeights)-1]}
|
log.Infof("finished filling in %s gap from %d to %d", bfs.chain.String(), blockHeights[0], blockHeights[len(blockHeights)-1])
|
||||||
|
processingDone <- true
|
||||||
}(blockHeights)
|
}(blockHeights)
|
||||||
}
|
}
|
||||||
}()
|
}()
|
||||||
|
|
||||||
// goroutine that listens on the processingDone chan
|
// listen on the processingDone chan
|
||||||
// keeps track of the number of processing goroutines that have finished
|
// keeps track of the number of processing goroutines that have finished
|
||||||
// when they have all finished, sends the final signal out
|
// when they have all finished, return
|
||||||
go func() {
|
|
||||||
goroutinesFinished := 0
|
goroutinesFinished := 0
|
||||||
for {
|
for {
|
||||||
select {
|
select {
|
||||||
case doneWithHeights := <-processingDone:
|
case <-processingDone:
|
||||||
atomic.AddInt64(&activeCount, -1)
|
atomic.AddInt64(&activeCount, -1)
|
||||||
select {
|
select {
|
||||||
// if we are waiting for a process to finish, signal that one has
|
// if we are waiting for a process to finish, signal that one has
|
||||||
case forwardDone <- true:
|
case forwardDone <- true:
|
||||||
default:
|
default:
|
||||||
}
|
}
|
||||||
log.Infof("finished filling in %s gap sub-bin from %d to %d", bfs.chain.String(), doneWithHeights[0], doneWithHeights[1])
|
|
||||||
goroutinesFinished++
|
goroutinesFinished++
|
||||||
if goroutinesFinished >= len(blockRangeBins) {
|
if goroutinesFinished >= len(blockRangeBins) {
|
||||||
done <- true
|
|
||||||
return
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}()
|
|
||||||
|
|
||||||
return nil
|
return nil
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
@ -67,6 +67,7 @@ var _ = Describe("BackFiller", func() {
|
|||||||
Retriever: mockRetriever,
|
Retriever: mockRetriever,
|
||||||
GapCheckFrequency: time.Second * 2,
|
GapCheckFrequency: time.Second * 2,
|
||||||
BatchSize: super_node.DefaultMaxBatchSize,
|
BatchSize: super_node.DefaultMaxBatchSize,
|
||||||
|
BatchNumber: super_node.DefaultMaxBatchNumber,
|
||||||
QuitChan: quitChan,
|
QuitChan: quitChan,
|
||||||
}
|
}
|
||||||
wg := &sync.WaitGroup{}
|
wg := &sync.WaitGroup{}
|
||||||
@ -121,6 +122,7 @@ var _ = Describe("BackFiller", func() {
|
|||||||
Retriever: mockRetriever,
|
Retriever: mockRetriever,
|
||||||
GapCheckFrequency: time.Second * 2,
|
GapCheckFrequency: time.Second * 2,
|
||||||
BatchSize: super_node.DefaultMaxBatchSize,
|
BatchSize: super_node.DefaultMaxBatchSize,
|
||||||
|
BatchNumber: super_node.DefaultMaxBatchNumber,
|
||||||
QuitChan: quitChan,
|
QuitChan: quitChan,
|
||||||
}
|
}
|
||||||
wg := &sync.WaitGroup{}
|
wg := &sync.WaitGroup{}
|
||||||
@ -169,6 +171,7 @@ var _ = Describe("BackFiller", func() {
|
|||||||
Retriever: mockRetriever,
|
Retriever: mockRetriever,
|
||||||
GapCheckFrequency: time.Second * 2,
|
GapCheckFrequency: time.Second * 2,
|
||||||
BatchSize: super_node.DefaultMaxBatchSize,
|
BatchSize: super_node.DefaultMaxBatchSize,
|
||||||
|
BatchNumber: super_node.DefaultMaxBatchNumber,
|
||||||
QuitChan: quitChan,
|
QuitChan: quitChan,
|
||||||
}
|
}
|
||||||
wg := &sync.WaitGroup{}
|
wg := &sync.WaitGroup{}
|
||||||
|
@ -68,12 +68,12 @@ func (ps *HTTPPayloadStreamer) Stream(payloadChan chan shared.RawChainData) (sha
|
|||||||
if bytes.Equal(blockHashBytes, ps.lastHash) {
|
if bytes.Equal(blockHashBytes, ps.lastHash) {
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
ps.lastHash = blockHashBytes
|
|
||||||
block, err := client.GetBlock(blockHash)
|
block, err := client.GetBlock(blockHash)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
errChan <- err
|
errChan <- err
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
|
ps.lastHash = blockHashBytes
|
||||||
payloadChan <- BlockPayload{
|
payloadChan <- BlockPayload{
|
||||||
Header: &block.Header,
|
Header: &block.Header,
|
||||||
BlockHeight: height,
|
BlockHeight: height,
|
||||||
|
@ -17,6 +17,7 @@
|
|||||||
package super_node
|
package super_node
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"fmt"
|
||||||
"os"
|
"os"
|
||||||
"path/filepath"
|
"path/filepath"
|
||||||
"time"
|
"time"
|
||||||
@ -57,6 +58,7 @@ type Config struct {
|
|||||||
HTTPClient interface{}
|
HTTPClient interface{}
|
||||||
Frequency time.Duration
|
Frequency time.Duration
|
||||||
BatchSize uint64
|
BatchSize uint64
|
||||||
|
BatchNumber uint64
|
||||||
}
|
}
|
||||||
|
|
||||||
// NewSuperNodeConfig is used to initialize a SuperNode config from a .toml file
|
// NewSuperNodeConfig is used to initialize a SuperNode config from a .toml file
|
||||||
@ -90,7 +92,7 @@ func NewSuperNodeConfig() (*Config, error) {
|
|||||||
c.Workers = workers
|
c.Workers = workers
|
||||||
switch c.Chain {
|
switch c.Chain {
|
||||||
case shared.Ethereum:
|
case shared.Ethereum:
|
||||||
c.NodeInfo, c.WSClient, err = getEthNodeAndClient(viper.GetString("ethereum.wsPath"))
|
c.NodeInfo, c.WSClient, err = getEthNodeAndClient(fmt.Sprintf("ws://%s", viper.GetString("ethereum.wsPath")))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
@ -116,7 +118,7 @@ func NewSuperNodeConfig() (*Config, error) {
|
|||||||
if c.Serve {
|
if c.Serve {
|
||||||
wsPath := viper.GetString("superNode.wsPath")
|
wsPath := viper.GetString("superNode.wsPath")
|
||||||
if wsPath == "" {
|
if wsPath == "" {
|
||||||
wsPath = "ws://127.0.0.1:8546"
|
wsPath = "127.0.0.1:8080"
|
||||||
}
|
}
|
||||||
c.WSEndpoint = wsPath
|
c.WSEndpoint = wsPath
|
||||||
ipcPath := viper.GetString("superNode.ipcPath")
|
ipcPath := viper.GetString("superNode.ipcPath")
|
||||||
@ -130,7 +132,7 @@ func NewSuperNodeConfig() (*Config, error) {
|
|||||||
c.IPCEndpoint = ipcPath
|
c.IPCEndpoint = ipcPath
|
||||||
httpPath := viper.GetString("superNode.httpPath")
|
httpPath := viper.GetString("superNode.httpPath")
|
||||||
if httpPath == "" {
|
if httpPath == "" {
|
||||||
httpPath = "http://127.0.0.1:8545"
|
httpPath = "127.0.0.1:8081"
|
||||||
}
|
}
|
||||||
c.HTTPEndpoint = httpPath
|
c.HTTPEndpoint = httpPath
|
||||||
}
|
}
|
||||||
@ -162,7 +164,7 @@ func (sn *Config) BackFillFields(chain string) error {
|
|||||||
var err error
|
var err error
|
||||||
switch sn.Chain {
|
switch sn.Chain {
|
||||||
case shared.Ethereum:
|
case shared.Ethereum:
|
||||||
_, httpClient, err = getEthNodeAndClient(viper.GetString("ethereum.httpPath"))
|
_, httpClient, err = getEthNodeAndClient(fmt.Sprintf("http://%s", viper.GetString("ethereum.httpPath")))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
@ -185,6 +187,7 @@ func (sn *Config) BackFillFields(chain string) error {
|
|||||||
}
|
}
|
||||||
sn.Frequency = frequency
|
sn.Frequency = frequency
|
||||||
sn.BatchSize = uint64(viper.GetInt64("superNode.batchSize"))
|
sn.BatchSize = uint64(viper.GetInt64("superNode.batchSize"))
|
||||||
|
sn.BatchNumber = uint64(viper.GetInt64("superNode.batchNumber"))
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -172,8 +172,9 @@ func NewCleaner(chain shared.ChainType, db *postgres.DB) (shared.Cleaner, error)
|
|||||||
switch chain {
|
switch chain {
|
||||||
case shared.Ethereum:
|
case shared.Ethereum:
|
||||||
return eth.NewCleaner(db), nil
|
return eth.NewCleaner(db), nil
|
||||||
// TODO: support BTC
|
case shared.Bitcoin:
|
||||||
|
return btc.NewCleaner(db), nil
|
||||||
default:
|
default:
|
||||||
return nil, fmt.Errorf("invalid chain %s for publisher constructor", chain.String())
|
return nil, fmt.Errorf("invalid chain %s for cleaner constructor", chain.String())
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -49,6 +49,7 @@ type Config struct {
|
|||||||
NodeInfo core.Node // Info for the associated node
|
NodeInfo core.Node // Info for the associated node
|
||||||
Ranges [][2]uint64 // The block height ranges to resync
|
Ranges [][2]uint64 // The block height ranges to resync
|
||||||
BatchSize uint64 // BatchSize for the resync http calls (client has to support batch sizing)
|
BatchSize uint64 // BatchSize for the resync http calls (client has to support batch sizing)
|
||||||
|
BatchNumber uint64
|
||||||
|
|
||||||
Quit chan bool // Channel for shutting down
|
Quit chan bool // Channel for shutting down
|
||||||
}
|
}
|
||||||
@ -57,6 +58,9 @@ type Config struct {
|
|||||||
func NewReSyncConfig() (*Config, error) {
|
func NewReSyncConfig() (*Config, error) {
|
||||||
c := new(Config)
|
c := new(Config)
|
||||||
var err error
|
var err error
|
||||||
|
start := uint64(viper.GetInt64("resync.start"))
|
||||||
|
stop := uint64(viper.GetInt64("resync.stop"))
|
||||||
|
c.Ranges = [][2]uint64{{start, stop}}
|
||||||
ipfsPath := viper.GetString("resync.ipfsPath")
|
ipfsPath := viper.GetString("resync.ipfsPath")
|
||||||
if ipfsPath == "" {
|
if ipfsPath == "" {
|
||||||
home, err := os.UserHomeDir()
|
home, err := os.UserHomeDir()
|
||||||
@ -90,11 +94,10 @@ func NewReSyncConfig() (*Config, error) {
|
|||||||
}
|
}
|
||||||
return nil, fmt.Errorf("chain type %s does not support data type %s", c.Chain.String(), c.ResyncType.String())
|
return nil, fmt.Errorf("chain type %s does not support data type %s", c.Chain.String(), c.ResyncType.String())
|
||||||
}
|
}
|
||||||
c.BatchSize = uint64(viper.GetInt64("resync.batchSize"))
|
|
||||||
|
|
||||||
switch c.Chain {
|
switch c.Chain {
|
||||||
case shared.Ethereum:
|
case shared.Ethereum:
|
||||||
c.NodeInfo, c.HTTPClient, err = getEthNodeAndClient(viper.GetString("ethereum.httpPath"))
|
c.NodeInfo, c.HTTPClient, err = getEthNodeAndClient(fmt.Sprintf("http://%s", viper.GetString("ethereum.httpPath")))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
@ -117,6 +120,8 @@ func NewReSyncConfig() (*Config, error) {
|
|||||||
db := utils.LoadPostgres(c.DBConfig, c.NodeInfo)
|
db := utils.LoadPostgres(c.DBConfig, c.NodeInfo)
|
||||||
c.DB = &db
|
c.DB = &db
|
||||||
c.Quit = make(chan bool)
|
c.Quit = make(chan bool)
|
||||||
|
c.BatchSize = uint64(viper.GetInt64("resync.batchSize"))
|
||||||
|
c.BatchNumber = uint64(viper.GetInt64("resync.batchNumber"))
|
||||||
return c, nil
|
return c, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -46,6 +46,8 @@ type Service struct {
|
|||||||
Cleaner shared.Cleaner
|
Cleaner shared.Cleaner
|
||||||
// Size of batch fetches
|
// Size of batch fetches
|
||||||
BatchSize uint64
|
BatchSize uint64
|
||||||
|
// Number of goroutines
|
||||||
|
BatchNumber int64
|
||||||
// Channel for receiving quit signal
|
// Channel for receiving quit signal
|
||||||
QuitChan chan bool
|
QuitChan chan bool
|
||||||
// Chain type
|
// Chain type
|
||||||
@ -88,6 +90,10 @@ func NewResyncService(settings *Config) (Resync, error) {
|
|||||||
if batchSize == 0 {
|
if batchSize == 0 {
|
||||||
batchSize = super_node.DefaultMaxBatchSize
|
batchSize = super_node.DefaultMaxBatchSize
|
||||||
}
|
}
|
||||||
|
batchNumber := int64(settings.BatchNumber)
|
||||||
|
if batchNumber == 0 {
|
||||||
|
batchNumber = super_node.DefaultMaxBatchNumber
|
||||||
|
}
|
||||||
return &Service{
|
return &Service{
|
||||||
Indexer: indexer,
|
Indexer: indexer,
|
||||||
Converter: converter,
|
Converter: converter,
|
||||||
@ -96,6 +102,7 @@ func NewResyncService(settings *Config) (Resync, error) {
|
|||||||
Fetcher: fetcher,
|
Fetcher: fetcher,
|
||||||
Cleaner: cleaner,
|
Cleaner: cleaner,
|
||||||
BatchSize: batchSize,
|
BatchSize: batchSize,
|
||||||
|
BatchNumber: int64(batchNumber),
|
||||||
QuitChan: settings.Quit,
|
QuitChan: settings.Quit,
|
||||||
chain: settings.Chain,
|
chain: settings.Chain,
|
||||||
ranges: settings.Ranges,
|
ranges: settings.Ranges,
|
||||||
@ -107,36 +114,19 @@ func NewResyncService(settings *Config) (Resync, error) {
|
|||||||
func (rs *Service) Resync() error {
|
func (rs *Service) Resync() error {
|
||||||
if rs.clearOldCache {
|
if rs.clearOldCache {
|
||||||
if err := rs.Cleaner.Clean(rs.ranges, rs.data); err != nil {
|
if err := rs.Cleaner.Clean(rs.ranges, rs.data); err != nil {
|
||||||
return fmt.Errorf("%s resync %s data cleaning error: %v", rs.chain.String(), rs.data.String(), err)
|
return fmt.Errorf("%s %s data resync cleaning error: %v", rs.chain.String(), rs.data.String(), err)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
for _, rng := range rs.ranges {
|
for _, rng := range rs.ranges {
|
||||||
if err := rs.resync(rng[0], rng[1]); err != nil {
|
if err := rs.resync(rng[0], rng[1]); err != nil {
|
||||||
return fmt.Errorf("%s resync %s data sync initialization error: %v", rs.chain.String(), rs.data.String(), err)
|
return fmt.Errorf("%s %s data resync initialization error: %v", rs.chain.String(), rs.data.String(), err)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (rs *Service) resync(startingBlock, endingBlock uint64) error {
|
func (rs *Service) resync(startingBlock, endingBlock uint64) error {
|
||||||
logrus.Infof("resyncing %s %s data from %d to %d", rs.chain.String(), rs.data.String(), startingBlock, endingBlock)
|
logrus.Infof("resyncing %s data from %d to %d", rs.chain.String(), startingBlock, endingBlock)
|
||||||
errChan := make(chan error)
|
|
||||||
done := make(chan bool)
|
|
||||||
if err := rs.refill(startingBlock, endingBlock, errChan, done); err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
for {
|
|
||||||
select {
|
|
||||||
case err := <-errChan:
|
|
||||||
logrus.Errorf("%s resync %s data sync error: %v", rs.chain.String(), rs.data.String(), err)
|
|
||||||
case <-done:
|
|
||||||
logrus.Infof("finished %s %s resync from %d to %d", rs.chain.String(), rs.data.String(), startingBlock, endingBlock)
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func (rs *Service) refill(startingBlock, endingBlock uint64, errChan chan error, done chan bool) error {
|
|
||||||
if endingBlock < startingBlock {
|
if endingBlock < startingBlock {
|
||||||
return fmt.Errorf("%s resync range ending block number needs to be greater than the starting block number", rs.chain.String())
|
return fmt.Errorf("%s resync range ending block number needs to be greater than the starting block number", rs.chain.String())
|
||||||
}
|
}
|
||||||
@ -148,7 +138,7 @@ func (rs *Service) refill(startingBlock, endingBlock uint64, errChan chan error,
|
|||||||
// int64 for atomic incrementing and decrementing to track the number of active processing goroutines we have
|
// int64 for atomic incrementing and decrementing to track the number of active processing goroutines we have
|
||||||
var activeCount int64
|
var activeCount int64
|
||||||
// channel for processing goroutines to signal when they are done
|
// channel for processing goroutines to signal when they are done
|
||||||
processingDone := make(chan [2]uint64)
|
processingDone := make(chan bool)
|
||||||
forwardDone := make(chan bool)
|
forwardDone := make(chan bool)
|
||||||
|
|
||||||
// for each block range bin spin up a goroutine to batch fetch and process state diffs for that range
|
// for each block range bin spin up a goroutine to batch fetch and process state diffs for that range
|
||||||
@ -156,59 +146,52 @@ func (rs *Service) refill(startingBlock, endingBlock uint64, errChan chan error,
|
|||||||
for _, blockHeights := range blockRangeBins {
|
for _, blockHeights := range blockRangeBins {
|
||||||
// if we have reached our limit of active goroutines
|
// if we have reached our limit of active goroutines
|
||||||
// wait for one to finish before starting the next
|
// wait for one to finish before starting the next
|
||||||
if atomic.AddInt64(&activeCount, 1) > super_node.DefaultMaxBatchNumber {
|
if atomic.AddInt64(&activeCount, 1) > rs.BatchNumber {
|
||||||
// this blocks until a process signals it has finished
|
// this blocks until a process signals it has finished
|
||||||
<-forwardDone
|
<-forwardDone
|
||||||
}
|
}
|
||||||
go func(blockHeights []uint64) {
|
go func(blockHeights []uint64) {
|
||||||
payloads, err := rs.Fetcher.FetchAt(blockHeights)
|
payloads, err := rs.Fetcher.FetchAt(blockHeights)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
errChan <- err
|
logrus.Errorf("%s resync fetcher error: %s", rs.chain.String(), err.Error())
|
||||||
}
|
}
|
||||||
for _, payload := range payloads {
|
for _, payload := range payloads {
|
||||||
ipldPayload, err := rs.Converter.Convert(payload)
|
ipldPayload, err := rs.Converter.Convert(payload)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
errChan <- err
|
logrus.Errorf("%s resync converter error: %s", rs.chain.String(), err.Error())
|
||||||
continue
|
|
||||||
}
|
}
|
||||||
cidPayload, err := rs.Publisher.Publish(ipldPayload)
|
cidPayload, err := rs.Publisher.Publish(ipldPayload)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
errChan <- err
|
logrus.Errorf("%s resync publisher error: %s", rs.chain.String(), err.Error())
|
||||||
continue
|
|
||||||
}
|
}
|
||||||
if err := rs.Indexer.Index(cidPayload); err != nil {
|
if err := rs.Indexer.Index(cidPayload); err != nil {
|
||||||
errChan <- err
|
logrus.Errorf("%s resync indexer error: %s", rs.chain.String(), err.Error())
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
// when this goroutine is done, send out a signal
|
// when this goroutine is done, send out a signal
|
||||||
processingDone <- [2]uint64{blockHeights[0], blockHeights[len(blockHeights)-1]}
|
logrus.Infof("finished %s resync section from %d to %d", rs.chain.String(), blockHeights[0], blockHeights[len(blockHeights)-1])
|
||||||
|
processingDone <- true
|
||||||
}(blockHeights)
|
}(blockHeights)
|
||||||
}
|
}
|
||||||
}()
|
}()
|
||||||
|
|
||||||
// goroutine that listens on the processingDone chan
|
// listen on the processingDone chan and
|
||||||
// keeps track of the number of processing goroutines that have finished
|
// keep track of the number of processing goroutines that have finished
|
||||||
// when they have all finished, sends the final signal out
|
// when they have all finished, sends the final signal out
|
||||||
go func() {
|
|
||||||
goroutinesFinished := 0
|
goroutinesFinished := 0
|
||||||
for {
|
for {
|
||||||
select {
|
select {
|
||||||
case doneWithHeights := <-processingDone:
|
case <-processingDone:
|
||||||
atomic.AddInt64(&activeCount, -1)
|
atomic.AddInt64(&activeCount, -1)
|
||||||
select {
|
select {
|
||||||
// if we are waiting for a process to finish, signal that one has
|
// if we are waiting for a process to finish, signal that one has
|
||||||
case forwardDone <- true:
|
case forwardDone <- true:
|
||||||
default:
|
default:
|
||||||
}
|
}
|
||||||
logrus.Infof("finished %s resync sub-bin from %d to %d", rs.chain.String(), doneWithHeights[0], doneWithHeights[1])
|
|
||||||
goroutinesFinished++
|
goroutinesFinished++
|
||||||
if goroutinesFinished >= len(blockRangeBins) {
|
if goroutinesFinished >= len(blockRangeBins) {
|
||||||
done <- true
|
|
||||||
return
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}()
|
|
||||||
|
|
||||||
return nil
|
return nil
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
Loading…
Reference in New Issue
Block a user