publisher unit test
This commit is contained in:
parent
d79cc90cb2
commit
35c8f3561a
@ -56,7 +56,7 @@ func init() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func syncAndPublish() {
|
func syncAndPublish() {
|
||||||
blockChain, ethClient, rpcClient := getBlockChainAndClients()
|
blockChain, rpcClient := getBlockChainAndClient()
|
||||||
|
|
||||||
db := utils.LoadPostgres(databaseConfig, blockChain.Node())
|
db := utils.LoadPostgres(databaseConfig, blockChain.Node())
|
||||||
quitChan := make(chan bool)
|
quitChan := make(chan bool)
|
||||||
@ -69,7 +69,7 @@ func syncAndPublish() {
|
|||||||
}
|
}
|
||||||
ipfsPath = filepath.Join(home, ".ipfs")
|
ipfsPath = filepath.Join(home, ".ipfs")
|
||||||
}
|
}
|
||||||
processor, err := seed_node.NewProcessor(ipfsPath, &db, ethClient, rpcClient, quitChan)
|
processor, err := seed_node.NewSeedNode(ipfsPath, &db, rpcClient, quitChan, 1)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
log.Fatal(err)
|
log.Fatal(err)
|
||||||
}
|
}
|
||||||
@ -82,7 +82,7 @@ func syncAndPublish() {
|
|||||||
wg.Wait() // If an error was thrown, wg.Add was never called and this will fall through
|
wg.Wait() // If an error was thrown, wg.Add was never called and this will fall through
|
||||||
}
|
}
|
||||||
|
|
||||||
func getBlockChainAndClients() (*geth.BlockChain, core.EthClient, core.RpcClient) {
|
func getBlockChainAndClient() (*geth.BlockChain, core.RpcClient) {
|
||||||
rawRpcClient, err := rpc.Dial(ipc)
|
rawRpcClient, err := rpc.Dial(ipc)
|
||||||
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@ -94,5 +94,5 @@ func getBlockChainAndClients() (*geth.BlockChain, core.EthClient, core.RpcClient
|
|||||||
vdbNode := node.MakeNode(rpcClient)
|
vdbNode := node.MakeNode(rpcClient)
|
||||||
transactionConverter := vRpc.NewRpcTransactionConverter(ethClient)
|
transactionConverter := vRpc.NewRpcTransactionConverter(ethClient)
|
||||||
blockChain := geth.NewBlockChain(vdbEthClient, rpcClient, vdbNode, transactionConverter)
|
blockChain := geth.NewBlockChain(vdbEthClient, rpcClient, vdbNode, transactionConverter)
|
||||||
return blockChain, ethClient, rpcClient
|
return blockChain, rpcClient
|
||||||
}
|
}
|
||||||
|
@ -50,7 +50,7 @@ func init() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func syncPublishScreenAndServe() {
|
func syncPublishScreenAndServe() {
|
||||||
blockChain, ethClient, rpcClient := getBlockChainAndClients()
|
blockChain, rpcClient := getBlockChainAndClient()
|
||||||
|
|
||||||
db := utils.LoadPostgres(databaseConfig, blockChain.Node())
|
db := utils.LoadPostgres(databaseConfig, blockChain.Node())
|
||||||
quitChan := make(chan bool, 1)
|
quitChan := make(chan bool, 1)
|
||||||
@ -63,7 +63,7 @@ func syncPublishScreenAndServe() {
|
|||||||
}
|
}
|
||||||
ipfsPath = filepath.Join(home, ".ipfs")
|
ipfsPath = filepath.Join(home, ".ipfs")
|
||||||
}
|
}
|
||||||
processor, err := seed_node.NewProcessor(ipfsPath, &db, ethClient, rpcClient, quitChan)
|
processor, err := seed_node.NewSeedNode(ipfsPath, &db, rpcClient, quitChan, 1)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
log.Fatal(err)
|
log.Fatal(err)
|
||||||
}
|
}
|
||||||
|
@ -1,14 +1,15 @@
|
|||||||
FROM golang:alpine as builder
|
FROM golang:alpine as builder
|
||||||
|
|
||||||
RUN apk --update --no-cache add dep make git g++ linux-headers
|
ENV GO111MODULE=on
|
||||||
|
|
||||||
|
RUN apk --update --no-cache add make git g++ linux-headers
|
||||||
# DEBUG
|
# DEBUG
|
||||||
RUN apk add busybox-extras
|
RUN apk add busybox-extras
|
||||||
|
|
||||||
# Get and build vulcanizedb syncAndPublish fork
|
# Get and build vulcanizedb ipfs_concurreny fork
|
||||||
RUN go get -u -d github.com/vulcanize/vulcanizedb
|
RUN go get -u -d github.com/vulcanize/vulcanizedb
|
||||||
WORKDIR /go/src/github.com/vulcanize/vulcanizedb
|
WORKDIR /go/src/github.com/vulcanize/vulcanizedb
|
||||||
RUN git checkout ipfs_concurrency
|
RUN git checkout ipfs_concurrency
|
||||||
RUN dep ensure
|
|
||||||
RUN GCO_ENABLED=0 GOOS=linux go build -a -installsuffix cgo -ldflags '-extldflags "-static"' -o vulcanizedb .
|
RUN GCO_ENABLED=0 GOOS=linux go build -a -installsuffix cgo -ldflags '-extldflags "-static"' -o vulcanizedb .
|
||||||
|
|
||||||
# Get and build vulcanize's go-ipfs fork
|
# Get and build vulcanize's go-ipfs fork
|
||||||
@ -17,7 +18,6 @@ WORKDIR /go/src/github.com/ipfs/go-ipfs
|
|||||||
RUN git remote add vulcanize https://github.com/vulcanize/go-ipfs.git
|
RUN git remote add vulcanize https://github.com/vulcanize/go-ipfs.git
|
||||||
RUN git fetch vulcanize
|
RUN git fetch vulcanize
|
||||||
RUN git checkout -b pg_ipfs vulcanize/postgres_update
|
RUN git checkout -b pg_ipfs vulcanize/postgres_update
|
||||||
RUN dep ensure
|
|
||||||
RUN GCO_ENABLED=0 GOOS=linux go build -a -installsuffix cgo -ldflags '-extldflags "-static"' -o ipfs ./cmd/ipfs
|
RUN GCO_ENABLED=0 GOOS=linux go build -a -installsuffix cgo -ldflags '-extldflags "-static"' -o ipfs ./cmd/ipfs
|
||||||
|
|
||||||
# Get and build vulcanize's geth fork
|
# Get and build vulcanize's geth fork
|
||||||
|
@ -63,32 +63,26 @@ func (f *EthIPLDFetcher) FetchCIDs(cids CIDWrapper) (*IPLDWrapper, error) {
|
|||||||
|
|
||||||
err := f.fetchHeaders(cids, blocks)
|
err := f.fetchHeaders(cids, blocks)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
println(1)
|
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
err = f.fetchUncles(cids, blocks)
|
err = f.fetchUncles(cids, blocks)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
println(2)
|
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
err = f.fetchTrxs(cids, blocks)
|
err = f.fetchTrxs(cids, blocks)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
println(3)
|
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
err = f.fetchRcts(cids, blocks)
|
err = f.fetchRcts(cids, blocks)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
println(4)
|
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
err = f.fetchStorage(cids, blocks)
|
err = f.fetchStorage(cids, blocks)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
println(5)
|
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
err = f.fetchState(cids, blocks)
|
err = f.fetchState(cids, blocks)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
println(6)
|
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -29,7 +29,6 @@ import (
|
|||||||
)
|
)
|
||||||
|
|
||||||
var (
|
var (
|
||||||
// these need to be actual typed objects so that the cid.Decode works
|
|
||||||
mockHeaderData = []byte{0, 1, 2, 3, 4}
|
mockHeaderData = []byte{0, 1, 2, 3, 4}
|
||||||
mockUncleData = []byte{1, 2, 3, 4, 5}
|
mockUncleData = []byte{1, 2, 3, 4, 5}
|
||||||
mockTrxData = []byte{2, 3, 4, 5, 6}
|
mockTrxData = []byte{2, 3, 4, 5, 6}
|
||||||
|
47
pkg/ipfs/mocks/dag_putters.go
Normal file
47
pkg/ipfs/mocks/dag_putters.go
Normal file
@ -0,0 +1,47 @@
|
|||||||
|
// VulcanizeDB
|
||||||
|
// Copyright © 2019 Vulcanize
|
||||||
|
|
||||||
|
// This program is free software: you can redistribute it and/or modify
|
||||||
|
// it under the terms of the GNU Affero General Public License as published by
|
||||||
|
// the Free Software Foundation, either version 3 of the License, or
|
||||||
|
// (at your option) any later version.
|
||||||
|
|
||||||
|
// This program is distributed in the hope that it will be useful,
|
||||||
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||||
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||||
|
// GNU Affero General Public License for more details.
|
||||||
|
|
||||||
|
// You should have received a copy of the GNU Affero General Public License
|
||||||
|
// along with this program. If not, see <http://www.gnu.org/licenses/>.
|
||||||
|
|
||||||
|
package mocks
|
||||||
|
|
||||||
|
import "errors"
|
||||||
|
|
||||||
|
// DagPutter is a mock for testing the ipfs publisher
|
||||||
|
type DagPutter struct {
|
||||||
|
CIDsToReturn []string
|
||||||
|
ErrToReturn error
|
||||||
|
}
|
||||||
|
|
||||||
|
// DagPut returns the pre-loaded CIDs or error
|
||||||
|
func (dp *DagPutter) DagPut(raw interface{}) ([]string, error) {
|
||||||
|
return dp.CIDsToReturn, dp.ErrToReturn
|
||||||
|
}
|
||||||
|
|
||||||
|
// IncrementingDagPutter is a mock for testing the ipfs publisher
|
||||||
|
type IncrementingDagPutter struct {
|
||||||
|
CIDsToReturn []string
|
||||||
|
iterator int
|
||||||
|
ErrToReturn error
|
||||||
|
}
|
||||||
|
|
||||||
|
// DagPut returns the pre-loaded CIDs or error
|
||||||
|
func (dp *IncrementingDagPutter) DagPut(raw interface{}) ([]string, error) {
|
||||||
|
if len(dp.CIDsToReturn) >= dp.iterator+1 {
|
||||||
|
cid := dp.CIDsToReturn[dp.iterator]
|
||||||
|
dp.iterator++
|
||||||
|
return []string{cid}, dp.ErrToReturn
|
||||||
|
}
|
||||||
|
return nil, errors.New("dag putter iterator is out of range")
|
||||||
|
}
|
@ -183,67 +183,92 @@ var (
|
|||||||
ReceiptsRlp: []byte{},
|
ReceiptsRlp: []byte{},
|
||||||
}
|
}
|
||||||
|
|
||||||
MockIPLDPayload = ipfs.IPLDPayload{
|
MockIPLDPayload = &ipfs.IPLDPayload{
|
||||||
BlockNumber: big.NewInt(1),
|
BlockNumber: big.NewInt(1),
|
||||||
BlockHash: MockBlock.Hash(),
|
BlockHash: MockBlock.Hash(),
|
||||||
Receipts: MockReceipts,
|
Receipts: MockReceipts,
|
||||||
HeaderRLP: MockHeaderRlp,
|
HeaderRLP: MockHeaderRlp,
|
||||||
BlockBody: MockBlock.Body(),
|
BlockBody: MockBlock.Body(),
|
||||||
TrxMetaData: MockTrxMeta,
|
TrxMetaData: []*ipfs.TrxMetaData{
|
||||||
ReceiptMetaData: MockRctMeta,
|
{
|
||||||
StorageNodes: MockStorageNodes,
|
CID: "",
|
||||||
StateNodes: MockStateNodes,
|
Src: senderAddr.Hex(),
|
||||||
|
Dst: "0x0000000000000000000000000000000000000000",
|
||||||
|
},
|
||||||
|
{
|
||||||
|
CID: "",
|
||||||
|
Src: senderAddr.Hex(),
|
||||||
|
Dst: "0x0000000000000000000000000000000000000001",
|
||||||
|
},
|
||||||
|
},
|
||||||
|
ReceiptMetaData: []*ipfs.ReceiptMetaData{
|
||||||
|
{
|
||||||
|
CID: "",
|
||||||
|
Topic0s: []string{
|
||||||
|
"0x0000000000000000000000000000000000000000000000000000000000000004",
|
||||||
|
},
|
||||||
|
ContractAddress: "0x0000000000000000000000000000000000000000",
|
||||||
|
},
|
||||||
|
{
|
||||||
|
CID: "",
|
||||||
|
Topic0s: []string{
|
||||||
|
"0x0000000000000000000000000000000000000000000000000000000000000005",
|
||||||
|
},
|
||||||
|
ContractAddress: "0x0000000000000000000000000000000000000001",
|
||||||
|
},
|
||||||
|
},
|
||||||
|
StorageNodes: MockStorageNodes,
|
||||||
|
StateNodes: MockStateNodes,
|
||||||
}
|
}
|
||||||
|
|
||||||
MockCIDPayload = ipfs.CIDPayload{
|
MockCIDPayload = &ipfs.CIDPayload{
|
||||||
BlockNumber: "1",
|
BlockNumber: "1",
|
||||||
BlockHash: common.HexToHash("0x0"),
|
BlockHash: MockBlock.Hash(),
|
||||||
HeaderCID: "mockHeaderCID",
|
HeaderCID: "mockHeaderCID",
|
||||||
|
UncleCIDS: make(map[common.Hash]string),
|
||||||
TransactionCIDs: map[common.Hash]*ipfs.TrxMetaData{
|
TransactionCIDs: map[common.Hash]*ipfs.TrxMetaData{
|
||||||
common.HexToHash("0x0"): {
|
MockTransactions[0].Hash(): {
|
||||||
CID: "mockTrxCID1",
|
CID: "mockTrxCID1",
|
||||||
Dst: "mockTo1",
|
Dst: "0x0000000000000000000000000000000000000000",
|
||||||
Src: "mockFrom1",
|
Src: senderAddr.Hex(),
|
||||||
},
|
},
|
||||||
common.HexToHash("0x1"): {
|
MockTransactions[1].Hash(): {
|
||||||
CID: "mockTrxCID2",
|
CID: "mockTrxCID2",
|
||||||
Dst: "mockTo2",
|
Dst: "0x0000000000000000000000000000000000000001",
|
||||||
Src: "mockFrom2",
|
Src: senderAddr.Hex(),
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
ReceiptCIDs: map[common.Hash]*ipfs.ReceiptMetaData{
|
ReceiptCIDs: map[common.Hash]*ipfs.ReceiptMetaData{
|
||||||
common.HexToHash("0x0"): {
|
MockTransactions[0].Hash(): {
|
||||||
CID: "mockReceiptCID1",
|
CID: "mockRctCID1",
|
||||||
Topic0s: []string{"mockTopic1"},
|
Topic0s: []string{"0x0000000000000000000000000000000000000000000000000000000000000004"},
|
||||||
|
ContractAddress: "0x0000000000000000000000000000000000000000",
|
||||||
},
|
},
|
||||||
common.HexToHash("0x1"): {
|
MockTransactions[1].Hash(): {
|
||||||
CID: "mockReceiptCID2",
|
CID: "mockRctCID2",
|
||||||
Topic0s: []string{"mockTopic1", "mockTopic2"},
|
Topic0s: []string{"0x0000000000000000000000000000000000000000000000000000000000000005"},
|
||||||
|
ContractAddress: "0x0000000000000000000000000000000000000001",
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
StateNodeCIDs: map[common.Hash]ipfs.StateNodeCID{
|
StateNodeCIDs: map[common.Hash]ipfs.StateNodeCID{
|
||||||
common.HexToHash("0x0"): {
|
ContractLeafKey: {
|
||||||
CID: "mockStateCID1",
|
CID: "mockStateCID1",
|
||||||
Leaf: true,
|
Leaf: true,
|
||||||
|
Key: "",
|
||||||
},
|
},
|
||||||
common.HexToHash("0x1"): {
|
AnotherContractLeafKey: {
|
||||||
CID: "mockStateCID2",
|
CID: "mockStateCID2",
|
||||||
Leaf: true,
|
Leaf: true,
|
||||||
|
Key: "",
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
StorageNodeCIDs: map[common.Hash][]ipfs.StorageNodeCID{
|
StorageNodeCIDs: map[common.Hash][]ipfs.StorageNodeCID{
|
||||||
common.HexToHash("0x0"): {
|
ContractLeafKey: {
|
||||||
{
|
{
|
||||||
CID: "mockStorageCID1",
|
CID: "mockStorageCID",
|
||||||
Key: "0x0",
|
Key: "0x0000000000000000000000000000000000000000000000000000000000000001",
|
||||||
Leaf: true,
|
Leaf: true,
|
||||||
},
|
StateKey: "",
|
||||||
},
|
|
||||||
common.HexToHash("0x1"): {
|
|
||||||
{
|
|
||||||
CID: "mockStorageCID2",
|
|
||||||
Key: "0x1",
|
|
||||||
Leaf: true,
|
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
@ -281,7 +306,7 @@ func createTransactionsAndReceipts() (types.Transactions, types.Receipts, common
|
|||||||
Topics: []common.Hash{mockTopic1},
|
Topics: []common.Hash{mockTopic1},
|
||||||
}
|
}
|
||||||
mockReceipt1.Logs = []*types.Log{mockLog1}
|
mockReceipt1.Logs = []*types.Log{mockLog1}
|
||||||
mockReceipt1.TxHash = trx1.Hash()
|
mockReceipt1.TxHash = signedTrx1.Hash()
|
||||||
mockTopic2 := common.HexToHash("0x05")
|
mockTopic2 := common.HexToHash("0x05")
|
||||||
mockReceipt2 := types.NewReceipt(common.HexToHash("0x1").Bytes(), false, 100)
|
mockReceipt2 := types.NewReceipt(common.HexToHash("0x1").Bytes(), false, 100)
|
||||||
mockReceipt2.ContractAddress = common.HexToAddress("0xaE9BEa628c4Ce503DcFD7E305CaB4e29E7476593")
|
mockReceipt2.ContractAddress = common.HexToAddress("0xaE9BEa628c4Ce503DcFD7E305CaB4e29E7476593")
|
||||||
@ -289,6 +314,6 @@ func createTransactionsAndReceipts() (types.Transactions, types.Receipts, common
|
|||||||
Topics: []common.Hash{mockTopic2},
|
Topics: []common.Hash{mockTopic2},
|
||||||
}
|
}
|
||||||
mockReceipt2.Logs = []*types.Log{mockLog2}
|
mockReceipt2.Logs = []*types.Log{mockLog2}
|
||||||
mockReceipt2.TxHash = trx2.Hash()
|
mockReceipt2.TxHash = signedTrx2.Hash()
|
||||||
return types.Transactions{signedTrx1, signedTrx2}, types.Receipts{mockReceipt1, mockReceipt2}, senderAddr
|
return types.Transactions{signedTrx1, signedTrx2}, types.Receipts{mockReceipt1, mockReceipt2}, senderAddr
|
||||||
}
|
}
|
||||||
|
@ -40,11 +40,11 @@ type IPLDPublisher interface {
|
|||||||
|
|
||||||
// Publisher is the underlying struct for the IPLDPublisher interface
|
// Publisher is the underlying struct for the IPLDPublisher interface
|
||||||
type Publisher struct {
|
type Publisher struct {
|
||||||
HeaderPutter *eth_block_header.BlockHeaderDagPutter
|
HeaderPutter ipfs.DagPutter
|
||||||
TransactionPutter *eth_block_transactions.BlockTransactionsDagPutter
|
TransactionPutter ipfs.DagPutter
|
||||||
ReceiptPutter *eth_block_receipts.EthBlockReceiptDagPutter
|
ReceiptPutter ipfs.DagPutter
|
||||||
StatePutter *eth_state_trie.StateTrieDagPutter
|
StatePutter ipfs.DagPutter
|
||||||
StoragePutter *eth_storage_trie.StorageTrieDagPutter
|
StoragePutter ipfs.DagPutter
|
||||||
}
|
}
|
||||||
|
|
||||||
// NewIPLDPublisher creates a pointer to a new Publisher which satisfies the IPLDPublisher interface
|
// NewIPLDPublisher creates a pointer to a new Publisher which satisfies the IPLDPublisher interface
|
||||||
|
62
pkg/ipfs/publisher_test.go
Normal file
62
pkg/ipfs/publisher_test.go
Normal file
@ -0,0 +1,62 @@
|
|||||||
|
// VulcanizeDB
|
||||||
|
// Copyright © 2019 Vulcanize
|
||||||
|
|
||||||
|
// This program is free software: you can redistribute it and/or modify
|
||||||
|
// it under the terms of the GNU Affero General Public License as published by
|
||||||
|
// the Free Software Foundation, either version 3 of the License, or
|
||||||
|
// (at your option) any later version.
|
||||||
|
|
||||||
|
// This program is distributed in the hope that it will be useful,
|
||||||
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||||
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||||
|
// GNU Affero General Public License for more details.
|
||||||
|
|
||||||
|
// You should have received a copy of the GNU Affero General Public License
|
||||||
|
// along with this program. If not, see <http://www.gnu.org/licenses/>.
|
||||||
|
|
||||||
|
package ipfs_test
|
||||||
|
|
||||||
|
import (
|
||||||
|
. "github.com/onsi/ginkgo"
|
||||||
|
. "github.com/onsi/gomega"
|
||||||
|
|
||||||
|
"github.com/vulcanize/vulcanizedb/pkg/ipfs"
|
||||||
|
"github.com/vulcanize/vulcanizedb/pkg/ipfs/mocks"
|
||||||
|
)
|
||||||
|
|
||||||
|
var (
|
||||||
|
mockHeaderDagPutter *mocks.DagPutter
|
||||||
|
mockTrxDagPutter *mocks.DagPutter
|
||||||
|
mockRctDagPutter *mocks.DagPutter
|
||||||
|
mockStateDagPutter *mocks.IncrementingDagPutter
|
||||||
|
mockStorageDagPutter *mocks.IncrementingDagPutter
|
||||||
|
)
|
||||||
|
|
||||||
|
var _ = Describe("Publisher", func() {
|
||||||
|
BeforeEach(func() {
|
||||||
|
mockHeaderDagPutter = new(mocks.DagPutter)
|
||||||
|
mockTrxDagPutter = new(mocks.DagPutter)
|
||||||
|
mockRctDagPutter = new(mocks.DagPutter)
|
||||||
|
mockStateDagPutter = new(mocks.IncrementingDagPutter)
|
||||||
|
mockStorageDagPutter = new(mocks.IncrementingDagPutter)
|
||||||
|
})
|
||||||
|
Describe("Publish", func() {
|
||||||
|
It("Publishes the passed IPLDPayload objects to IPFS and returns a CIDPayload for indexing", func() {
|
||||||
|
mockHeaderDagPutter.CIDsToReturn = []string{"mockHeaderCID"}
|
||||||
|
mockTrxDagPutter.CIDsToReturn = []string{"mockTrxCID1", "mockTrxCID2"}
|
||||||
|
mockRctDagPutter.CIDsToReturn = []string{"mockRctCID1", "mockRctCID2"}
|
||||||
|
mockStateDagPutter.CIDsToReturn = []string{"mockStateCID1", "mockStateCID2"}
|
||||||
|
mockStorageDagPutter.CIDsToReturn = []string{"mockStorageCID"}
|
||||||
|
publisher := ipfs.Publisher{
|
||||||
|
HeaderPutter: mockHeaderDagPutter,
|
||||||
|
TransactionPutter: mockTrxDagPutter,
|
||||||
|
ReceiptPutter: mockRctDagPutter,
|
||||||
|
StatePutter: mockStateDagPutter,
|
||||||
|
StoragePutter: mockStorageDagPutter,
|
||||||
|
}
|
||||||
|
cidPayload, err := publisher.Publish(mocks.MockIPLDPayload)
|
||||||
|
Expect(err).ToNot(HaveOccurred())
|
||||||
|
Expect(*cidPayload).To(Equal(*mocks.MockCIDPayload))
|
||||||
|
})
|
||||||
|
})
|
||||||
|
})
|
@ -34,13 +34,13 @@ const APIVersion = "0.0.1"
|
|||||||
|
|
||||||
// PublicSeedNodeAPI is the public api for the seed node
|
// PublicSeedNodeAPI is the public api for the seed node
|
||||||
type PublicSeedNodeAPI struct {
|
type PublicSeedNodeAPI struct {
|
||||||
snp Processor
|
sni NodeInterface
|
||||||
}
|
}
|
||||||
|
|
||||||
// NewPublicSeedNodeAPI creates a new PublicSeedNodeAPI with the provided underlying SyncPublishScreenAndServe process
|
// NewPublicSeedNodeAPI creates a new PublicSeedNodeAPI with the provided underlying SyncPublishScreenAndServe process
|
||||||
func NewPublicSeedNodeAPI(seedNodeProcessor Processor) *PublicSeedNodeAPI {
|
func NewPublicSeedNodeAPI(seedNodeInterface NodeInterface) *PublicSeedNodeAPI {
|
||||||
return &PublicSeedNodeAPI{
|
return &PublicSeedNodeAPI{
|
||||||
snp: seedNodeProcessor,
|
sni: seedNodeInterface,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -59,7 +59,7 @@ func (api *PublicSeedNodeAPI) Stream(ctx context.Context, streamFilters config.S
|
|||||||
// subscribe to events from the SyncPublishScreenAndServe service
|
// subscribe to events from the SyncPublishScreenAndServe service
|
||||||
payloadChannel := make(chan streamer.SeedNodePayload, payloadChanBufferSize)
|
payloadChannel := make(chan streamer.SeedNodePayload, payloadChanBufferSize)
|
||||||
quitChan := make(chan bool, 1)
|
quitChan := make(chan bool, 1)
|
||||||
go api.snp.Subscribe(rpcSub.ID, payloadChannel, quitChan, streamFilters)
|
go api.sni.Subscribe(rpcSub.ID, payloadChannel, quitChan, streamFilters)
|
||||||
|
|
||||||
// loop and await state diff payloads and relay them to the subscriber with then notifier
|
// loop and await state diff payloads and relay them to the subscriber with then notifier
|
||||||
for {
|
for {
|
||||||
@ -67,11 +67,11 @@ func (api *PublicSeedNodeAPI) Stream(ctx context.Context, streamFilters config.S
|
|||||||
case packet := <-payloadChannel:
|
case packet := <-payloadChannel:
|
||||||
if notifyErr := notifier.Notify(rpcSub.ID, packet); notifyErr != nil {
|
if notifyErr := notifier.Notify(rpcSub.ID, packet); notifyErr != nil {
|
||||||
log.Error("Failed to send state diff packet", "err", notifyErr)
|
log.Error("Failed to send state diff packet", "err", notifyErr)
|
||||||
api.snp.Unsubscribe(rpcSub.ID)
|
api.sni.Unsubscribe(rpcSub.ID)
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
case <-rpcSub.Err():
|
case <-rpcSub.Err():
|
||||||
api.snp.Unsubscribe(rpcSub.ID)
|
api.sni.Unsubscribe(rpcSub.ID)
|
||||||
return
|
return
|
||||||
case <-quitChan:
|
case <-quitChan:
|
||||||
// don't need to unsubscribe, SyncPublishScreenAndServe service does so before sending the quit signal
|
// don't need to unsubscribe, SyncPublishScreenAndServe service does so before sending the quit signal
|
||||||
|
@ -24,8 +24,8 @@ import (
|
|||||||
"github.com/ethereum/go-ethereum/rlp"
|
"github.com/ethereum/go-ethereum/rlp"
|
||||||
|
|
||||||
"github.com/vulcanize/vulcanizedb/libraries/shared/streamer"
|
"github.com/vulcanize/vulcanizedb/libraries/shared/streamer"
|
||||||
"github.com/vulcanize/vulcanizedb/pkg/ipfs"
|
|
||||||
"github.com/vulcanize/vulcanizedb/pkg/config"
|
"github.com/vulcanize/vulcanizedb/pkg/config"
|
||||||
|
"github.com/vulcanize/vulcanizedb/pkg/ipfs"
|
||||||
)
|
)
|
||||||
|
|
||||||
// ResponseFilterer is the inteface used to screen eth data and package appropriate data into a response payload
|
// ResponseFilterer is the inteface used to screen eth data and package appropriate data into a response payload
|
||||||
|
@ -20,8 +20,8 @@ import (
|
|||||||
"github.com/jmoiron/sqlx"
|
"github.com/jmoiron/sqlx"
|
||||||
"github.com/lib/pq"
|
"github.com/lib/pq"
|
||||||
|
|
||||||
"github.com/vulcanize/vulcanizedb/pkg/ipfs"
|
|
||||||
"github.com/vulcanize/vulcanizedb/pkg/datastore/postgres"
|
"github.com/vulcanize/vulcanizedb/pkg/datastore/postgres"
|
||||||
|
"github.com/vulcanize/vulcanizedb/pkg/ipfs"
|
||||||
)
|
)
|
||||||
|
|
||||||
// CIDRepository is an interface for indexing ipfs.CIDPayloads
|
// CIDRepository is an interface for indexing ipfs.CIDPayloads
|
||||||
|
@ -23,9 +23,9 @@ import (
|
|||||||
"github.com/lib/pq"
|
"github.com/lib/pq"
|
||||||
log "github.com/sirupsen/logrus"
|
log "github.com/sirupsen/logrus"
|
||||||
|
|
||||||
"github.com/vulcanize/vulcanizedb/pkg/ipfs"
|
|
||||||
"github.com/vulcanize/vulcanizedb/pkg/config"
|
"github.com/vulcanize/vulcanizedb/pkg/config"
|
||||||
"github.com/vulcanize/vulcanizedb/pkg/datastore/postgres"
|
"github.com/vulcanize/vulcanizedb/pkg/datastore/postgres"
|
||||||
|
"github.com/vulcanize/vulcanizedb/pkg/ipfs"
|
||||||
)
|
)
|
||||||
|
|
||||||
// CIDRetriever is the interface for retrieving CIDs from the Postgres cache
|
// CIDRetriever is the interface for retrieving CIDs from the Postgres cache
|
||||||
|
@ -20,9 +20,9 @@ import (
|
|||||||
"io/ioutil"
|
"io/ioutil"
|
||||||
"testing"
|
"testing"
|
||||||
|
|
||||||
"github.com/sirupsen/logrus"
|
|
||||||
. "github.com/onsi/ginkgo"
|
. "github.com/onsi/ginkgo"
|
||||||
. "github.com/onsi/gomega"
|
. "github.com/onsi/gomega"
|
||||||
|
"github.com/sirupsen/logrus"
|
||||||
)
|
)
|
||||||
|
|
||||||
func TestSeedNode(t *testing.T) {
|
func TestSeedNode(t *testing.T) {
|
||||||
|
@ -19,30 +19,29 @@ package seed_node
|
|||||||
import (
|
import (
|
||||||
"sync"
|
"sync"
|
||||||
|
|
||||||
"github.com/ethereum/go-ethereum/params"
|
|
||||||
"github.com/ethereum/go-ethereum/common"
|
"github.com/ethereum/go-ethereum/common"
|
||||||
"github.com/ethereum/go-ethereum/crypto"
|
"github.com/ethereum/go-ethereum/crypto"
|
||||||
"github.com/ethereum/go-ethereum/node"
|
"github.com/ethereum/go-ethereum/node"
|
||||||
"github.com/ethereum/go-ethereum/p2p"
|
"github.com/ethereum/go-ethereum/p2p"
|
||||||
|
"github.com/ethereum/go-ethereum/params"
|
||||||
"github.com/ethereum/go-ethereum/rlp"
|
"github.com/ethereum/go-ethereum/rlp"
|
||||||
"github.com/ethereum/go-ethereum/rpc"
|
"github.com/ethereum/go-ethereum/rpc"
|
||||||
"github.com/ethereum/go-ethereum/statediff"
|
"github.com/ethereum/go-ethereum/statediff"
|
||||||
log "github.com/sirupsen/logrus"
|
log "github.com/sirupsen/logrus"
|
||||||
|
|
||||||
"github.com/vulcanize/vulcanizedb/libraries/shared/streamer"
|
"github.com/vulcanize/vulcanizedb/libraries/shared/streamer"
|
||||||
"github.com/vulcanize/vulcanizedb/pkg/ipfs"
|
|
||||||
"github.com/vulcanize/vulcanizedb/pkg/config"
|
"github.com/vulcanize/vulcanizedb/pkg/config"
|
||||||
"github.com/vulcanize/vulcanizedb/pkg/core"
|
"github.com/vulcanize/vulcanizedb/pkg/core"
|
||||||
"github.com/vulcanize/vulcanizedb/pkg/datastore/postgres"
|
"github.com/vulcanize/vulcanizedb/pkg/datastore/postgres"
|
||||||
|
"github.com/vulcanize/vulcanizedb/pkg/ipfs"
|
||||||
)
|
)
|
||||||
|
|
||||||
const payloadChanBufferSize = 20000 // the max eth sub buffer size
|
const payloadChanBufferSize = 20000 // the max eth sub buffer size
|
||||||
const workerPoolSize = 1
|
|
||||||
|
|
||||||
// Processor is the top level interface for streaming, converting to IPLDs, publishing,
|
// NodeInterface is the top level interface for streaming, converting to IPLDs, publishing,
|
||||||
// and indexing all Ethereum data; screening this data; and serving it up to subscribed clients
|
// and indexing all Ethereum data; screening this data; and serving it up to subscribed clients
|
||||||
// This service is compatible with the Ethereum service interface (node.Service)
|
// This service is compatible with the Ethereum service interface (node.Service)
|
||||||
type Processor interface {
|
type NodeInterface interface {
|
||||||
// APIs(), Protocols(), Start() and Stop()
|
// APIs(), Protocols(), Start() and Stop()
|
||||||
node.Service
|
node.Service
|
||||||
// Main event loop for syncAndPublish processes
|
// Main event loop for syncAndPublish processes
|
||||||
@ -83,10 +82,12 @@ type Service struct {
|
|||||||
Subscriptions map[common.Hash]map[rpc.ID]Subscription
|
Subscriptions map[common.Hash]map[rpc.ID]Subscription
|
||||||
// A mapping of subscription hash type to the corresponding StreamFilters
|
// A mapping of subscription hash type to the corresponding StreamFilters
|
||||||
SubscriptionTypes map[common.Hash]config.Subscription
|
SubscriptionTypes map[common.Hash]config.Subscription
|
||||||
|
// Number of workers
|
||||||
|
WorkerPoolSize int
|
||||||
}
|
}
|
||||||
|
|
||||||
// NewProcessor creates a new Processor interface using an underlying Service struct
|
// NewSeedNode creates a new seed_node.Interface using an underlying seed_node.Service struct
|
||||||
func NewProcessor(ipfsPath string, db *postgres.DB, ethClient core.EthClient, rpcClient core.RpcClient, qc chan bool) (Processor, error) {
|
func NewSeedNode(ipfsPath string, db *postgres.DB, rpcClient core.RpcClient, qc chan bool, workers int) (NodeInterface, error) {
|
||||||
publisher, err := ipfs.NewIPLDPublisher(ipfsPath)
|
publisher, err := ipfs.NewIPLDPublisher(ipfsPath)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
@ -108,6 +109,7 @@ func NewProcessor(ipfsPath string, db *postgres.DB, ethClient core.EthClient, rp
|
|||||||
QuitChan: qc,
|
QuitChan: qc,
|
||||||
Subscriptions: make(map[common.Hash]map[rpc.ID]Subscription),
|
Subscriptions: make(map[common.Hash]map[rpc.ID]Subscription),
|
||||||
SubscriptionTypes: make(map[common.Hash]config.Subscription),
|
SubscriptionTypes: make(map[common.Hash]config.Subscription),
|
||||||
|
WorkerPoolSize: workers,
|
||||||
}, nil
|
}, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -140,10 +142,10 @@ func (sap *Service) SyncAndPublish(wg *sync.WaitGroup, screenAndServePayload cha
|
|||||||
|
|
||||||
// Channels for forwarding data to the publishAndIndex workers
|
// Channels for forwarding data to the publishAndIndex workers
|
||||||
publishAndIndexPayload := make(chan ipfs.IPLDPayload, payloadChanBufferSize)
|
publishAndIndexPayload := make(chan ipfs.IPLDPayload, payloadChanBufferSize)
|
||||||
publishAndIndexQuit := make(chan bool, workerPoolSize)
|
publishAndIndexQuit := make(chan bool, sap.WorkerPoolSize)
|
||||||
// publishAndIndex worker pool to handle publishing and indexing concurrently, while
|
// publishAndIndex worker pool to handle publishing and indexing concurrently, while
|
||||||
// limiting the number of Postgres connections we can possibly open so as to prevent error
|
// limiting the number of Postgres connections we can possibly open so as to prevent error
|
||||||
for i := 0; i < workerPoolSize; i++ {
|
for i := 0; i < sap.WorkerPoolSize; i++ {
|
||||||
sap.publishAndIndex(i, publishAndIndexPayload, publishAndIndexQuit)
|
sap.publishAndIndex(i, publishAndIndexPayload, publishAndIndexQuit)
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -175,7 +177,7 @@ func (sap *Service) SyncAndPublish(wg *sync.WaitGroup, screenAndServePayload cha
|
|||||||
default:
|
default:
|
||||||
}
|
}
|
||||||
// Also forward a quit signal for each of the workers
|
// Also forward a quit signal for each of the workers
|
||||||
for i := 0; i < workerPoolSize; i++ {
|
for i := 0; i < sap.WorkerPoolSize; i++ {
|
||||||
select {
|
select {
|
||||||
case publishAndIndexQuit <- true:
|
case publishAndIndexQuit <- true:
|
||||||
default:
|
default:
|
||||||
|
@ -25,15 +25,15 @@ import (
|
|||||||
. "github.com/onsi/ginkgo"
|
. "github.com/onsi/ginkgo"
|
||||||
. "github.com/onsi/gomega"
|
. "github.com/onsi/gomega"
|
||||||
|
|
||||||
"github.com/vulcanize/vulcanizedb/pkg/seed_node"
|
|
||||||
mocks2 "github.com/vulcanize/vulcanizedb/libraries/shared/mocks"
|
mocks2 "github.com/vulcanize/vulcanizedb/libraries/shared/mocks"
|
||||||
"github.com/vulcanize/vulcanizedb/pkg/ipfs/mocks"
|
"github.com/vulcanize/vulcanizedb/pkg/ipfs/mocks"
|
||||||
|
"github.com/vulcanize/vulcanizedb/pkg/seed_node"
|
||||||
mocks3 "github.com/vulcanize/vulcanizedb/pkg/seed_node/mocks"
|
mocks3 "github.com/vulcanize/vulcanizedb/pkg/seed_node/mocks"
|
||||||
)
|
)
|
||||||
|
|
||||||
var _ = Describe("Service", func() {
|
var _ = Describe("Service", func() {
|
||||||
|
|
||||||
Describe("Loop", func() {
|
Describe("SyncAndPublish", func() {
|
||||||
It("Streams statediff.Payloads, converts them to IPLDPayloads, publishes IPLDPayloads, and indexes CIDPayloads", func() {
|
It("Streams statediff.Payloads, converts them to IPLDPayloads, publishes IPLDPayloads, and indexes CIDPayloads", func() {
|
||||||
wg := new(sync.WaitGroup)
|
wg := new(sync.WaitGroup)
|
||||||
payloadChan := make(chan statediff.Payload, 1)
|
payloadChan := make(chan statediff.Payload, 1)
|
||||||
@ -42,36 +42,37 @@ var _ = Describe("Service", func() {
|
|||||||
ReturnErr: nil,
|
ReturnErr: nil,
|
||||||
}
|
}
|
||||||
mockPublisher := &mocks.IPLDPublisher{
|
mockPublisher := &mocks.IPLDPublisher{
|
||||||
ReturnCIDPayload: &mocks.MockCIDPayload,
|
ReturnCIDPayload: mocks.MockCIDPayload,
|
||||||
ReturnErr: nil,
|
ReturnErr: nil,
|
||||||
}
|
}
|
||||||
mockStreamer := &mocks2.StateDiffStreamer{
|
mockStreamer := &mocks2.StateDiffStreamer{
|
||||||
ReturnSub: &rpc.ClientSubscription{},
|
ReturnSub: &rpc.ClientSubscription{},
|
||||||
StreamPayloads: []statediff.Payload{
|
StreamPayloads: []statediff.Payload{
|
||||||
mocks.MockStatediffPayload,
|
mocks.MockStateDiffPayload,
|
||||||
},
|
},
|
||||||
ReturnErr: nil,
|
ReturnErr: nil,
|
||||||
}
|
}
|
||||||
mockConverter := &mocks.PayloadConverter{
|
mockConverter := &mocks.PayloadConverter{
|
||||||
ReturnIPLDPayload: &mocks.MockIPLDPayload,
|
ReturnIPLDPayload: mocks.MockIPLDPayload,
|
||||||
ReturnErr: nil,
|
ReturnErr: nil,
|
||||||
}
|
}
|
||||||
processor := &seed_node.Service{
|
processor := &seed_node.Service{
|
||||||
Repository: mockCidRepo,
|
Repository: mockCidRepo,
|
||||||
Publisher: mockPublisher,
|
Publisher: mockPublisher,
|
||||||
Streamer: mockStreamer,
|
Streamer: mockStreamer,
|
||||||
Converter: mockConverter,
|
Converter: mockConverter,
|
||||||
PayloadChan: payloadChan,
|
PayloadChan: payloadChan,
|
||||||
QuitChan: quitChan,
|
QuitChan: quitChan,
|
||||||
|
WorkerPoolSize: 1,
|
||||||
}
|
}
|
||||||
err := processor.SyncAndPublish(wg, nil, nil)
|
err := processor.SyncAndPublish(wg, nil, nil)
|
||||||
Expect(err).ToNot(HaveOccurred())
|
Expect(err).ToNot(HaveOccurred())
|
||||||
time.Sleep(2 * time.Second)
|
time.Sleep(2 * time.Second)
|
||||||
quitChan <- true
|
quitChan <- true
|
||||||
wg.Wait()
|
wg.Wait()
|
||||||
Expect(mockConverter.PassedStatediffPayload).To(Equal(mocks.MockStatediffPayload))
|
Expect(mockConverter.PassedStatediffPayload).To(Equal(mocks.MockStateDiffPayload))
|
||||||
Expect(mockCidRepo.PassedCIDPayload).To(Equal(&mocks.MockCIDPayload))
|
Expect(mockCidRepo.PassedCIDPayload).To(Equal(mocks.MockCIDPayload))
|
||||||
Expect(mockPublisher.PassedIPLDPayload).To(Equal(&mocks.MockIPLDPayload))
|
Expect(mockPublisher.PassedIPLDPayload).To(Equal(mocks.MockIPLDPayload))
|
||||||
Expect(mockStreamer.PassedPayloadChan).To(Equal(payloadChan))
|
Expect(mockStreamer.PassedPayloadChan).To(Equal(payloadChan))
|
||||||
})
|
})
|
||||||
})
|
})
|
||||||
|
Loading…
Reference in New Issue
Block a user