1186 lines
38 KiB
Go
1186 lines
38 KiB
Go
// VulcanizeDB
|
|
// Copyright © 2019 Vulcanize
|
|
|
|
// This program is free software: you can redistribute it and/or modify
|
|
// it under the terms of the GNU Affero General Public License as published by
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
// (at your option) any later version.
|
|
|
|
// This program is distributed in the hope that it will be useful,
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
// GNU Affero General Public License for more details.
|
|
|
|
// You should have received a copy of the GNU Affero General Public License
|
|
// along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
package sql_test
|
|
|
|
import (
|
|
"context"
|
|
"math/big"
|
|
"sort"
|
|
"testing"
|
|
|
|
"github.com/ipfs/go-cid"
|
|
blockstore "github.com/ipfs/go-ipfs-blockstore"
|
|
dshelp "github.com/ipfs/go-ipfs-ds-help"
|
|
"github.com/stretchr/testify/require"
|
|
|
|
"github.com/ethereum/go-ethereum/common"
|
|
"github.com/ethereum/go-ethereum/core/types"
|
|
"github.com/ethereum/go-ethereum/rlp"
|
|
"github.com/ethereum/go-ethereum/statediff/indexer/database/sql"
|
|
"github.com/ethereum/go-ethereum/statediff/indexer/database/sql/postgres"
|
|
"github.com/ethereum/go-ethereum/statediff/indexer/interfaces"
|
|
"github.com/ethereum/go-ethereum/statediff/indexer/mocks"
|
|
"github.com/ethereum/go-ethereum/statediff/indexer/models"
|
|
"github.com/ethereum/go-ethereum/statediff/indexer/shared"
|
|
"github.com/ethereum/go-ethereum/statediff/indexer/test_helpers"
|
|
sdtypes "github.com/ethereum/go-ethereum/statediff/types"
|
|
)
|
|
|
|
func setupPGXIndexer(t *testing.T) {
|
|
db, err = postgres.SetupPGXDB()
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
ind, err = sql.NewStateDiffIndexer(context.Background(), mocks.TestConfig, db)
|
|
require.NoError(t, err)
|
|
}
|
|
|
|
func setupPGX(t *testing.T) {
|
|
setupPGXIndexer(t)
|
|
var tx interfaces.Batch
|
|
tx, err = ind.PushBlock(
|
|
mockBlock,
|
|
mocks.MockReceipts,
|
|
mocks.MockBlock.Difficulty())
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
defer func() {
|
|
if err := tx.Submit(err); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
}()
|
|
for _, node := range mocks.StateDiffs {
|
|
err = ind.PushStateNode(tx, node, mockBlock.Hash().String())
|
|
require.NoError(t, err)
|
|
}
|
|
|
|
require.Equal(t, mocks.BlockNumber.String(), tx.(*sql.BatchTx).BlockNumber)
|
|
}
|
|
|
|
func setupPGXNonCanonical(t *testing.T) {
|
|
setupPGXIndexer(t)
|
|
var tx1 interfaces.Batch
|
|
|
|
tx1, err = ind.PushBlock(
|
|
mockBlock,
|
|
mocks.MockReceipts,
|
|
mocks.MockBlock.Difficulty())
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
for _, node := range mocks.StateDiffs {
|
|
err = ind.PushStateNode(tx1, node, mockBlock.Hash().String())
|
|
require.NoError(t, err)
|
|
}
|
|
|
|
require.Equal(t, mocks.BlockNumber.String(), tx1.(*sql.BatchTx).BlockNumber)
|
|
if err := tx1.Submit(err); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
var tx2 interfaces.Batch
|
|
tx2, err = ind.PushBlock(
|
|
mocks.MockNonCanonicalBlock,
|
|
mocks.MockNonCanonicalBlockReceipts,
|
|
mocks.MockNonCanonicalBlock.Difficulty())
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
require.Equal(t, mocks.BlockNumber.String(), tx2.(*sql.BatchTx).BlockNumber)
|
|
if err := tx2.Submit(err); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// TODO index state & storage nodes for second block
|
|
}
|
|
|
|
func TestPGXIndexer(t *testing.T) {
|
|
t.Run("Publish and index header IPLDs in a single tx", func(t *testing.T) {
|
|
setupPGX(t)
|
|
defer tearDown(t)
|
|
defer checkTxClosure(t, 1, 0, 1)
|
|
pgStr := `SELECT cid, cast(td AS TEXT), cast(reward AS TEXT), block_hash, coinbase
|
|
FROM eth.header_cids
|
|
WHERE block_number = $1`
|
|
// check header was properly indexed
|
|
type res struct {
|
|
CID string
|
|
TD string
|
|
Reward string
|
|
BlockHash string `db:"block_hash"`
|
|
Coinbase string `db:"coinbase"`
|
|
}
|
|
header := new(res)
|
|
err = db.QueryRow(context.Background(), pgStr, mocks.BlockNumber.Uint64()).Scan(
|
|
&header.CID,
|
|
&header.TD,
|
|
&header.Reward,
|
|
&header.BlockHash,
|
|
&header.Coinbase)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
require.Equal(t, headerCID.String(), header.CID)
|
|
require.Equal(t, mocks.MockBlock.Difficulty().String(), header.TD)
|
|
require.Equal(t, "2000000000000021250", header.Reward)
|
|
require.Equal(t, mocks.MockHeader.Coinbase.String(), header.Coinbase)
|
|
dc, err := cid.Decode(header.CID)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
mhKey := dshelp.MultihashToDsKey(dc.Hash())
|
|
prefixedKey := blockstore.BlockPrefix.String() + mhKey.String()
|
|
var data []byte
|
|
err = db.Get(context.Background(), &data, ipfsPgGet, prefixedKey, mocks.BlockNumber.Uint64())
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
require.Equal(t, mocks.MockHeaderRlp, data)
|
|
})
|
|
|
|
t.Run("Publish and index transaction IPLDs in a single tx", func(t *testing.T) {
|
|
setupPGX(t)
|
|
defer tearDown(t)
|
|
defer checkTxClosure(t, 1, 0, 1)
|
|
// check that txs were properly indexed and published
|
|
trxs := make([]string, 0)
|
|
pgStr := `SELECT transaction_cids.cid FROM eth.transaction_cids INNER JOIN eth.header_cids ON (transaction_cids.header_id = header_cids.block_hash)
|
|
WHERE header_cids.block_number = $1`
|
|
err = db.Select(context.Background(), &trxs, pgStr, mocks.BlockNumber.Uint64())
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
require.Equal(t, 5, len(trxs))
|
|
expectTrue(t, test_helpers.ListContainsString(trxs, trx1CID.String()))
|
|
expectTrue(t, test_helpers.ListContainsString(trxs, trx2CID.String()))
|
|
expectTrue(t, test_helpers.ListContainsString(trxs, trx3CID.String()))
|
|
expectTrue(t, test_helpers.ListContainsString(trxs, trx4CID.String()))
|
|
expectTrue(t, test_helpers.ListContainsString(trxs, trx5CID.String()))
|
|
|
|
transactions := mocks.MockBlock.Transactions()
|
|
type txResult struct {
|
|
TxType uint8 `db:"tx_type"`
|
|
Value string
|
|
}
|
|
for _, c := range trxs {
|
|
dc, err := cid.Decode(c)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
mhKey := dshelp.MultihashToDsKey(dc.Hash())
|
|
prefixedKey := blockstore.BlockPrefix.String() + mhKey.String()
|
|
var data []byte
|
|
err = db.Get(context.Background(), &data, ipfsPgGet, prefixedKey, mocks.BlockNumber.Uint64())
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
txTypeAndValueStr := `SELECT tx_type, CAST(value as TEXT) FROM eth.transaction_cids WHERE cid = $1`
|
|
switch c {
|
|
case trx1CID.String():
|
|
require.Equal(t, tx1, data)
|
|
txRes := new(txResult)
|
|
err = db.QueryRow(context.Background(), txTypeAndValueStr, c).Scan(&txRes.TxType, &txRes.Value)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if txRes.TxType != 0 {
|
|
t.Fatalf("expected LegacyTxType (0), got %d", txRes.TxType)
|
|
}
|
|
if txRes.Value != transactions[0].Value().String() {
|
|
t.Fatalf("expected tx value %s got %s", transactions[0].Value().String(), txRes.Value)
|
|
}
|
|
case trx2CID.String():
|
|
require.Equal(t, tx2, data)
|
|
txRes := new(txResult)
|
|
err = db.QueryRow(context.Background(), txTypeAndValueStr, c).Scan(&txRes.TxType, &txRes.Value)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if txRes.TxType != 0 {
|
|
t.Fatalf("expected LegacyTxType (0), got %d", txRes.TxType)
|
|
}
|
|
if txRes.Value != transactions[1].Value().String() {
|
|
t.Fatalf("expected tx value %s got %s", transactions[1].Value().String(), txRes.Value)
|
|
}
|
|
case trx3CID.String():
|
|
require.Equal(t, tx3, data)
|
|
txRes := new(txResult)
|
|
err = db.QueryRow(context.Background(), txTypeAndValueStr, c).Scan(&txRes.TxType, &txRes.Value)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if txRes.TxType != 0 {
|
|
t.Fatalf("expected LegacyTxType (0), got %d", txRes.TxType)
|
|
}
|
|
if txRes.Value != transactions[2].Value().String() {
|
|
t.Fatalf("expected tx value %s got %s", transactions[2].Value().String(), txRes.Value)
|
|
}
|
|
case trx4CID.String():
|
|
require.Equal(t, tx4, data)
|
|
txRes := new(txResult)
|
|
err = db.QueryRow(context.Background(), txTypeAndValueStr, c).Scan(&txRes.TxType, &txRes.Value)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if txRes.TxType != types.AccessListTxType {
|
|
t.Fatalf("expected AccessListTxType (1), got %d", txRes.TxType)
|
|
}
|
|
if txRes.Value != transactions[3].Value().String() {
|
|
t.Fatalf("expected tx value %s got %s", transactions[3].Value().String(), txRes.Value)
|
|
}
|
|
accessListElementModels := make([]models.AccessListElementModel, 0)
|
|
pgStr = "SELECT cast(access_list_elements.block_number AS TEXT), access_list_elements.index, access_list_elements.tx_id, " +
|
|
"access_list_elements.address, access_list_elements.storage_keys FROM eth.access_list_elements " +
|
|
"INNER JOIN eth.transaction_cids ON (tx_id = transaction_cids.tx_hash) WHERE cid = $1 ORDER BY access_list_elements.index ASC"
|
|
err = db.Select(context.Background(), &accessListElementModels, pgStr, c)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if len(accessListElementModels) != 2 {
|
|
t.Fatalf("expected two access list entries, got %d", len(accessListElementModels))
|
|
}
|
|
model1 := models.AccessListElementModel{
|
|
BlockNumber: mocks.BlockNumber.String(),
|
|
Index: accessListElementModels[0].Index,
|
|
Address: accessListElementModels[0].Address,
|
|
}
|
|
model2 := models.AccessListElementModel{
|
|
BlockNumber: mocks.BlockNumber.String(),
|
|
Index: accessListElementModels[1].Index,
|
|
Address: accessListElementModels[1].Address,
|
|
StorageKeys: accessListElementModels[1].StorageKeys,
|
|
}
|
|
require.Equal(t, mocks.AccessListEntry1Model, model1)
|
|
require.Equal(t, mocks.AccessListEntry2Model, model2)
|
|
case trx5CID.String():
|
|
require.Equal(t, tx5, data)
|
|
txRes := new(txResult)
|
|
err = db.QueryRow(context.Background(), txTypeAndValueStr, c).Scan(&txRes.TxType, &txRes.Value)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if txRes.TxType != types.DynamicFeeTxType {
|
|
t.Fatalf("expected DynamicFeeTxType (2), got %d", txRes.TxType)
|
|
}
|
|
if txRes.Value != transactions[4].Value().String() {
|
|
t.Fatalf("expected tx value %s got %s", transactions[4].Value().String(), txRes.Value)
|
|
}
|
|
}
|
|
}
|
|
})
|
|
|
|
t.Run("Publish and index log IPLDs for multiple receipt of a specific block", func(t *testing.T) {
|
|
setupPGX(t)
|
|
defer tearDown(t)
|
|
defer checkTxClosure(t, 1, 0, 1)
|
|
|
|
rcts := make([]string, 0)
|
|
rctsPgStr := `SELECT receipt_cids.leaf_cid FROM eth.receipt_cids, eth.transaction_cids, eth.header_cids
|
|
WHERE receipt_cids.tx_id = transaction_cids.tx_hash
|
|
AND transaction_cids.header_id = header_cids.block_hash
|
|
AND header_cids.block_number = $1
|
|
ORDER BY transaction_cids.index`
|
|
logsPgStr := `SELECT log_cids.index, log_cids.address, log_cids.topic0, log_cids.topic1, data FROM eth.log_cids
|
|
INNER JOIN eth.receipt_cids ON (log_cids.rct_id = receipt_cids.tx_id)
|
|
INNER JOIN public.blocks ON (log_cids.leaf_mh_key = blocks.key)
|
|
WHERE receipt_cids.leaf_cid = $1 ORDER BY eth.log_cids.index ASC`
|
|
err = db.Select(context.Background(), &rcts, rctsPgStr, mocks.BlockNumber.Uint64())
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if len(rcts) != len(mocks.MockReceipts) {
|
|
t.Fatalf("expected %d receipts, got %d", len(mocks.MockReceipts), len(rcts))
|
|
}
|
|
|
|
type logIPLD struct {
|
|
Index int `db:"index"`
|
|
Address string `db:"address"`
|
|
Data []byte `db:"data"`
|
|
Topic0 string `db:"topic0"`
|
|
Topic1 string `db:"topic1"`
|
|
}
|
|
for i := range rcts {
|
|
results := make([]logIPLD, 0)
|
|
err = db.Select(context.Background(), &results, logsPgStr, rcts[i])
|
|
require.NoError(t, err)
|
|
|
|
expectedLogs := mocks.MockReceipts[i].Logs
|
|
require.Equal(t, len(expectedLogs), len(results))
|
|
|
|
var nodeElements []interface{}
|
|
for idx, r := range results {
|
|
// Attempt to decode the log leaf node.
|
|
err = rlp.DecodeBytes(r.Data, &nodeElements)
|
|
require.NoError(t, err)
|
|
if len(nodeElements) == 2 {
|
|
logRaw, err := rlp.EncodeToBytes(&expectedLogs[idx])
|
|
require.NoError(t, err)
|
|
// 2nd element of the leaf node contains the encoded log data.
|
|
require.Equal(t, nodeElements[1].([]byte), logRaw)
|
|
} else {
|
|
logRaw, err := rlp.EncodeToBytes(&expectedLogs[idx])
|
|
require.NoError(t, err)
|
|
// raw log was IPLDized
|
|
require.Equal(t, r.Data, logRaw)
|
|
}
|
|
}
|
|
}
|
|
})
|
|
|
|
t.Run("Publish and index receipt IPLDs in a single tx", func(t *testing.T) {
|
|
setupPGX(t)
|
|
defer tearDown(t)
|
|
defer checkTxClosure(t, 1, 0, 1)
|
|
|
|
// check receipts were properly indexed and published
|
|
rcts := make([]string, 0)
|
|
pgStr := `SELECT receipt_cids.leaf_cid FROM eth.receipt_cids, eth.transaction_cids, eth.header_cids
|
|
WHERE receipt_cids.tx_id = transaction_cids.tx_hash
|
|
AND transaction_cids.header_id = header_cids.block_hash
|
|
AND header_cids.block_number = $1 order by transaction_cids.index`
|
|
err = db.Select(context.Background(), &rcts, pgStr, mocks.BlockNumber.Uint64())
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
require.Equal(t, 5, len(rcts))
|
|
expectTrue(t, test_helpers.ListContainsString(rcts, rct1CID.String()))
|
|
expectTrue(t, test_helpers.ListContainsString(rcts, rct2CID.String()))
|
|
expectTrue(t, test_helpers.ListContainsString(rcts, rct3CID.String()))
|
|
expectTrue(t, test_helpers.ListContainsString(rcts, rct4CID.String()))
|
|
expectTrue(t, test_helpers.ListContainsString(rcts, rct5CID.String()))
|
|
|
|
for idx, c := range rcts {
|
|
result := make([]models.IPLDModel, 0)
|
|
pgStr = `SELECT data
|
|
FROM eth.receipt_cids
|
|
INNER JOIN public.blocks ON (receipt_cids.leaf_mh_key = public.blocks.key)
|
|
WHERE receipt_cids.leaf_cid = $1`
|
|
err = db.Select(context.Background(), &result, pgStr, c)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
// Decode the log leaf node.
|
|
var nodeElements []interface{}
|
|
err = rlp.DecodeBytes(result[0].Data, &nodeElements)
|
|
require.NoError(t, err)
|
|
|
|
expectedRct, err := mocks.MockReceipts[idx].MarshalBinary()
|
|
require.NoError(t, err)
|
|
|
|
require.Equal(t, nodeElements[1].([]byte), expectedRct)
|
|
|
|
dc, err := cid.Decode(c)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
mhKey := dshelp.MultihashToDsKey(dc.Hash())
|
|
prefixedKey := blockstore.BlockPrefix.String() + mhKey.String()
|
|
var data []byte
|
|
err = db.Get(context.Background(), &data, ipfsPgGet, prefixedKey, mocks.BlockNumber.Uint64())
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
postStatePgStr := `SELECT post_state FROM eth.receipt_cids WHERE leaf_cid = $1`
|
|
switch c {
|
|
case rct1CID.String():
|
|
require.Equal(t, rctLeaf1, data)
|
|
var postStatus uint64
|
|
pgStr = `SELECT post_status FROM eth.receipt_cids WHERE leaf_cid = $1`
|
|
err = db.Get(context.Background(), &postStatus, pgStr, c)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
require.Equal(t, mocks.ExpectedPostStatus, postStatus)
|
|
case rct2CID.String():
|
|
require.Equal(t, rctLeaf2, data)
|
|
var postState string
|
|
err = db.Get(context.Background(), &postState, postStatePgStr, c)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
require.Equal(t, mocks.ExpectedPostState1, postState)
|
|
case rct3CID.String():
|
|
require.Equal(t, rctLeaf3, data)
|
|
var postState string
|
|
err = db.Get(context.Background(), &postState, postStatePgStr, c)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
require.Equal(t, mocks.ExpectedPostState2, postState)
|
|
case rct4CID.String():
|
|
require.Equal(t, rctLeaf4, data)
|
|
var postState string
|
|
err = db.Get(context.Background(), &postState, postStatePgStr, c)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
require.Equal(t, mocks.ExpectedPostState3, postState)
|
|
case rct5CID.String():
|
|
require.Equal(t, rctLeaf5, data)
|
|
var postState string
|
|
err = db.Get(context.Background(), &postState, postStatePgStr, c)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
require.Equal(t, mocks.ExpectedPostState3, postState)
|
|
}
|
|
}
|
|
})
|
|
|
|
t.Run("Publish and index state IPLDs in a single tx", func(t *testing.T) {
|
|
setupPGX(t)
|
|
defer tearDown(t)
|
|
defer checkTxClosure(t, 1, 0, 1)
|
|
// check that state nodes were properly indexed and published
|
|
stateNodes := make([]models.StateNodeModel, 0)
|
|
pgStr := `SELECT state_cids.cid, state_cids.state_leaf_key, state_cids.node_type, state_cids.state_path, state_cids.header_id
|
|
FROM eth.state_cids INNER JOIN eth.header_cids ON (state_cids.header_id = header_cids.block_hash)
|
|
WHERE header_cids.block_number = $1 AND node_type != 3`
|
|
err = db.Select(context.Background(), &stateNodes, pgStr, mocks.BlockNumber.Uint64())
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
require.Equal(t, 2, len(stateNodes))
|
|
for _, stateNode := range stateNodes {
|
|
var data []byte
|
|
dc, err := cid.Decode(stateNode.CID)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
mhKey := dshelp.MultihashToDsKey(dc.Hash())
|
|
prefixedKey := blockstore.BlockPrefix.String() + mhKey.String()
|
|
err = db.Get(context.Background(), &data, ipfsPgGet, prefixedKey, mocks.BlockNumber.Uint64())
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
pgStr = `SELECT cast(block_number AS TEXT), header_id, state_path, cast(balance AS TEXT), nonce, code_hash, storage_root from eth.state_accounts WHERE header_id = $1 AND state_path = $2`
|
|
var account models.StateAccountModel
|
|
err = db.Get(context.Background(), &account, pgStr, stateNode.HeaderID, stateNode.Path)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if stateNode.CID == state1CID.String() {
|
|
require.Equal(t, 2, stateNode.NodeType)
|
|
require.Equal(t, common.BytesToHash(mocks.ContractLeafKey).Hex(), stateNode.StateKey)
|
|
require.Equal(t, []byte{'\x06'}, stateNode.Path)
|
|
require.Equal(t, mocks.ContractLeafNode, data)
|
|
require.Equal(t, models.StateAccountModel{
|
|
BlockNumber: mocks.BlockNumber.String(),
|
|
HeaderID: account.HeaderID,
|
|
StatePath: stateNode.Path,
|
|
Balance: "0",
|
|
CodeHash: mocks.ContractCodeHash.Bytes(),
|
|
StorageRoot: mocks.ContractRoot,
|
|
Nonce: 1,
|
|
}, account)
|
|
}
|
|
if stateNode.CID == state2CID.String() {
|
|
require.Equal(t, 2, stateNode.NodeType)
|
|
require.Equal(t, common.BytesToHash(mocks.AccountLeafKey).Hex(), stateNode.StateKey)
|
|
require.Equal(t, []byte{'\x0c'}, stateNode.Path)
|
|
require.Equal(t, mocks.AccountLeafNode, data)
|
|
require.Equal(t, models.StateAccountModel{
|
|
BlockNumber: mocks.BlockNumber.String(),
|
|
HeaderID: account.HeaderID,
|
|
StatePath: stateNode.Path,
|
|
Balance: "1000",
|
|
CodeHash: mocks.AccountCodeHash.Bytes(),
|
|
StorageRoot: mocks.AccountRoot,
|
|
Nonce: 0,
|
|
}, account)
|
|
}
|
|
}
|
|
|
|
// check that Removed state nodes were properly indexed and published
|
|
stateNodes = make([]models.StateNodeModel, 0)
|
|
pgStr = `SELECT state_cids.cid, state_cids.state_leaf_key, state_cids.node_type, state_cids.state_path, state_cids.header_id
|
|
FROM eth.state_cids INNER JOIN eth.header_cids ON (state_cids.header_id = header_cids.block_hash)
|
|
WHERE header_cids.block_number = $1 AND node_type = 3`
|
|
err = db.Select(context.Background(), &stateNodes, pgStr, mocks.BlockNumber.Uint64())
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
require.Equal(t, 2, len(stateNodes))
|
|
for idx, stateNode := range stateNodes {
|
|
var data []byte
|
|
dc, err := cid.Decode(stateNode.CID)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
mhKey := dshelp.MultihashToDsKey(dc.Hash())
|
|
prefixedKey := blockstore.BlockPrefix.String() + mhKey.String()
|
|
require.Equal(t, shared.RemovedNodeMhKey, prefixedKey)
|
|
err = db.Get(context.Background(), &data, ipfsPgGet, prefixedKey, mocks.BlockNumber.Uint64())
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
if idx == 0 {
|
|
require.Equal(t, shared.RemovedNodeStateCID, stateNode.CID)
|
|
require.Equal(t, common.BytesToHash(mocks.RemovedLeafKey).Hex(), stateNode.StateKey)
|
|
require.Equal(t, []byte{'\x02'}, stateNode.Path)
|
|
require.Equal(t, []byte{}, data)
|
|
}
|
|
if idx == 1 {
|
|
require.Equal(t, shared.RemovedNodeStateCID, stateNode.CID)
|
|
require.Equal(t, common.BytesToHash(mocks.Contract2LeafKey).Hex(), stateNode.StateKey)
|
|
require.Equal(t, []byte{'\x07'}, stateNode.Path)
|
|
require.Equal(t, []byte{}, data)
|
|
}
|
|
}
|
|
})
|
|
|
|
t.Run("Publish and index storage IPLDs in a single tx", func(t *testing.T) {
|
|
setupPGX(t)
|
|
defer tearDown(t)
|
|
defer checkTxClosure(t, 1, 0, 1)
|
|
// check that storage nodes were properly indexed
|
|
storageNodes := make([]models.StorageNodeWithStateKeyModel, 0)
|
|
pgStr := `SELECT cast(storage_cids.block_number AS TEXT), storage_cids.cid, state_cids.state_leaf_key, storage_cids.storage_leaf_key, storage_cids.node_type, storage_cids.storage_path
|
|
FROM eth.storage_cids, eth.state_cids, eth.header_cids
|
|
WHERE (storage_cids.state_path, storage_cids.header_id) = (state_cids.state_path, state_cids.header_id)
|
|
AND state_cids.header_id = header_cids.block_hash
|
|
AND header_cids.block_number = $1
|
|
AND storage_cids.node_type != 3
|
|
ORDER BY storage_path`
|
|
err = db.Select(context.Background(), &storageNodes, pgStr, mocks.BlockNumber.Uint64())
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
require.Equal(t, 1, len(storageNodes))
|
|
require.Equal(t, models.StorageNodeWithStateKeyModel{
|
|
BlockNumber: mocks.BlockNumber.String(),
|
|
CID: storageCID.String(),
|
|
NodeType: 2,
|
|
StorageKey: common.BytesToHash(mocks.StorageLeafKey).Hex(),
|
|
StateKey: common.BytesToHash(mocks.ContractLeafKey).Hex(),
|
|
Path: []byte{},
|
|
}, storageNodes[0])
|
|
var data []byte
|
|
dc, err := cid.Decode(storageNodes[0].CID)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
mhKey := dshelp.MultihashToDsKey(dc.Hash())
|
|
prefixedKey := blockstore.BlockPrefix.String() + mhKey.String()
|
|
err = db.Get(context.Background(), &data, ipfsPgGet, prefixedKey, mocks.BlockNumber.Uint64())
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
require.Equal(t, mocks.StorageLeafNode, data)
|
|
|
|
// check that Removed storage nodes were properly indexed
|
|
storageNodes = make([]models.StorageNodeWithStateKeyModel, 0)
|
|
pgStr = `SELECT cast(storage_cids.block_number AS TEXT), storage_cids.cid, state_cids.state_leaf_key, storage_cids.storage_leaf_key, storage_cids.node_type, storage_cids.storage_path
|
|
FROM eth.storage_cids, eth.state_cids, eth.header_cids
|
|
WHERE (storage_cids.state_path, storage_cids.header_id) = (state_cids.state_path, state_cids.header_id)
|
|
AND state_cids.header_id = header_cids.block_hash
|
|
AND header_cids.block_number = $1
|
|
AND storage_cids.node_type = 3`
|
|
err = db.Select(context.Background(), &storageNodes, pgStr, mocks.BlockNumber.Uint64())
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
require.Equal(t, 3, len(storageNodes))
|
|
expectedStorageNodes := []models.StorageNodeWithStateKeyModel{
|
|
{
|
|
BlockNumber: mocks.BlockNumber.String(),
|
|
CID: shared.RemovedNodeStorageCID,
|
|
NodeType: 3,
|
|
StorageKey: common.BytesToHash(mocks.RemovedLeafKey).Hex(),
|
|
StateKey: common.BytesToHash(mocks.ContractLeafKey).Hex(),
|
|
Path: []byte{'\x03'},
|
|
},
|
|
{
|
|
BlockNumber: mocks.BlockNumber.String(),
|
|
CID: shared.RemovedNodeStorageCID,
|
|
NodeType: 3,
|
|
StorageKey: common.BytesToHash(mocks.Storage2LeafKey).Hex(),
|
|
StateKey: common.BytesToHash(mocks.Contract2LeafKey).Hex(),
|
|
Path: []byte{'\x0e'},
|
|
},
|
|
{
|
|
BlockNumber: mocks.BlockNumber.String(),
|
|
CID: shared.RemovedNodeStorageCID,
|
|
NodeType: 3,
|
|
StorageKey: common.BytesToHash(mocks.Storage3LeafKey).Hex(),
|
|
StateKey: common.BytesToHash(mocks.Contract2LeafKey).Hex(),
|
|
Path: []byte{'\x0f'},
|
|
},
|
|
}
|
|
for idx, storageNode := range storageNodes {
|
|
require.Equal(t, expectedStorageNodes[idx], storageNode)
|
|
dc, err = cid.Decode(storageNode.CID)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
mhKey = dshelp.MultihashToDsKey(dc.Hash())
|
|
prefixedKey = blockstore.BlockPrefix.String() + mhKey.String()
|
|
require.Equal(t, shared.RemovedNodeMhKey, prefixedKey)
|
|
err = db.Get(context.Background(), &data, ipfsPgGet, prefixedKey, mocks.BlockNumber.Uint64())
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
require.Equal(t, []byte{}, data)
|
|
}
|
|
})
|
|
}
|
|
|
|
func XTestPGXIndexerNonCanonical(t *testing.T) {
|
|
t.Run("Publish and index header", func(t *testing.T) {
|
|
setupPGXNonCanonical(t)
|
|
defer tearDown(t)
|
|
defer checkTxClosure(t, 1, 0, 1)
|
|
|
|
// check indexed headers
|
|
pgStr := `SELECT block_hash, cid, cast(td AS TEXT), cast(reward AS TEXT),
|
|
tx_root, receipt_root, uncle_root, coinbase
|
|
FROM eth.header_cids
|
|
WHERE block_number = $1`
|
|
headerRes := make([]models.HeaderModel, 0)
|
|
err = db.Select(context.Background(), &headerRes, pgStr, mocks.BlockNumber.Uint64())
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
expectedRes := []models.HeaderModel{
|
|
{
|
|
BlockHash: mockBlock.Hash().String(),
|
|
CID: headerCID.String(),
|
|
TotalDifficulty: mockBlock.Difficulty().String(),
|
|
TxRoot: mockBlock.TxHash().String(),
|
|
RctRoot: mockBlock.ReceiptHash().String(),
|
|
UncleRoot: mockBlock.UncleHash().String(),
|
|
Coinbase: mocks.MockHeader.Coinbase.String(),
|
|
},
|
|
{
|
|
BlockHash: mockNonCanonicalBlock.Hash().String(),
|
|
CID: mockNonCanonicalHeaderCID.String(),
|
|
TotalDifficulty: mockNonCanonicalBlock.Difficulty().String(),
|
|
TxRoot: mockNonCanonicalBlock.TxHash().String(),
|
|
RctRoot: mockNonCanonicalBlock.ReceiptHash().String(),
|
|
UncleRoot: mockNonCanonicalBlock.UncleHash().String(),
|
|
Coinbase: mocks.MockNonCanonicalHeader.Coinbase.String(),
|
|
},
|
|
}
|
|
expectedRes[0].Reward = shared.CalcEthBlockReward(mockBlock.Header(), mockBlock.Uncles(), mockBlock.Transactions(), mocks.MockReceipts).String()
|
|
expectedRes[1].Reward = shared.CalcEthBlockReward(mockNonCanonicalBlock.Header(), mockNonCanonicalBlock.Uncles(), mockNonCanonicalBlock.Transactions(), mocks.MockNonCanonicalBlockReceipts).String()
|
|
|
|
require.Equal(t, len(expectedRes), len(headerRes))
|
|
require.ElementsMatch(t,
|
|
[]string{mockBlock.Hash().String(), mocks.MockNonCanonicalBlock.Hash().String()},
|
|
[]string{headerRes[0].BlockHash, headerRes[1].BlockHash},
|
|
)
|
|
|
|
if headerRes[0].BlockHash == mockBlock.Hash().String() {
|
|
require.Equal(t, expectedRes[0], headerRes[0])
|
|
require.Equal(t, expectedRes[1], headerRes[1])
|
|
} else {
|
|
require.Equal(t, expectedRes[1], headerRes[0])
|
|
require.Equal(t, expectedRes[0], headerRes[1])
|
|
}
|
|
|
|
// check indexed IPLD blocks
|
|
var data []byte
|
|
var prefixedKey string
|
|
|
|
prefixedKey = shared.MultihashKeyFromCID(headerCID)
|
|
err = db.Get(context.Background(), &data, ipfsPgGet, prefixedKey, mocks.BlockNumber.Uint64())
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
require.Equal(t, mocks.MockHeaderRlp, data)
|
|
|
|
prefixedKey = shared.MultihashKeyFromCID(mockNonCanonicalHeaderCID)
|
|
err = db.Get(context.Background(), &data, ipfsPgGet, prefixedKey, mocks.BlockNumber.Uint64())
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
require.Equal(t, mocks.MockNonCanonicalHeaderRlp, data)
|
|
})
|
|
|
|
t.Run("Publish and index transactions", func(t *testing.T) {
|
|
setupPGXNonCanonical(t)
|
|
defer tearDown(t)
|
|
defer checkTxClosure(t, 1, 0, 1)
|
|
|
|
// check indexed transactions
|
|
pgStr := `SELECT header_id, tx_hash, cid, dst, src, index,
|
|
tx_data, tx_type, CAST(value as TEXT)
|
|
FROM eth.transaction_cids
|
|
WHERE block_number = $1`
|
|
txRes := make([]models.TxModel, 0)
|
|
err = db.Select(context.Background(), &txRes, pgStr, mocks.BlockNumber.Uint64())
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
mockBlockTxs := mocks.MockBlock.Transactions()
|
|
expectedBlockTxs := []models.TxModel{
|
|
{
|
|
HeaderID: mockBlock.Hash().String(),
|
|
TxHash: mockBlockTxs[0].Hash().String(),
|
|
CID: trx1CID.String(),
|
|
Dst: shared.HandleZeroAddrPointer(mockBlockTxs[0].To()),
|
|
Src: mocks.SenderAddr.String(),
|
|
Index: 0,
|
|
Data: mockBlockTxs[0].Data(),
|
|
Type: mockBlockTxs[0].Type(),
|
|
Value: mockBlockTxs[0].Value().String(),
|
|
},
|
|
{
|
|
HeaderID: mockBlock.Hash().String(),
|
|
TxHash: mockBlockTxs[1].Hash().String(),
|
|
CID: trx2CID.String(),
|
|
Dst: shared.HandleZeroAddrPointer(mockBlockTxs[1].To()),
|
|
Src: mocks.SenderAddr.String(),
|
|
Index: 1,
|
|
Data: mockBlockTxs[1].Data(),
|
|
Type: mockBlockTxs[1].Type(),
|
|
Value: mockBlockTxs[1].Value().String(),
|
|
},
|
|
{
|
|
HeaderID: mockBlock.Hash().String(),
|
|
TxHash: mockBlockTxs[2].Hash().String(),
|
|
CID: trx3CID.String(),
|
|
Dst: shared.HandleZeroAddrPointer(mockBlockTxs[2].To()),
|
|
Src: mocks.SenderAddr.String(),
|
|
Index: 2,
|
|
Data: mockBlockTxs[2].Data(),
|
|
Type: mockBlockTxs[2].Type(),
|
|
Value: mockBlockTxs[2].Value().String(),
|
|
},
|
|
{
|
|
HeaderID: mockBlock.Hash().String(),
|
|
TxHash: mockBlockTxs[3].Hash().String(),
|
|
CID: trx4CID.String(),
|
|
Dst: shared.HandleZeroAddrPointer(mockBlockTxs[3].To()),
|
|
Src: mocks.SenderAddr.String(),
|
|
Index: 3,
|
|
Data: mockBlockTxs[3].Data(),
|
|
Type: mockBlockTxs[3].Type(),
|
|
Value: mockBlockTxs[3].Value().String(),
|
|
},
|
|
{
|
|
HeaderID: mockBlock.Hash().String(),
|
|
TxHash: mockBlockTxs[4].Hash().String(),
|
|
CID: trx5CID.String(),
|
|
Dst: shared.HandleZeroAddrPointer(mockBlockTxs[4].To()),
|
|
Src: mocks.SenderAddr.String(),
|
|
Index: 4,
|
|
Data: mockBlockTxs[4].Data(),
|
|
Type: mockBlockTxs[4].Type(),
|
|
Value: mockBlockTxs[4].Value().String(),
|
|
},
|
|
}
|
|
|
|
mockNonCanonicalBlockTxs := mocks.MockNonCanonicalBlock.Transactions()
|
|
expectedNonCanonicalBlockTxs := []models.TxModel{
|
|
{
|
|
HeaderID: mockNonCanonicalBlock.Hash().String(),
|
|
TxHash: mockNonCanonicalBlockTxs[0].Hash().String(),
|
|
CID: trx2CID.String(),
|
|
Dst: mockNonCanonicalBlockTxs[0].To().String(),
|
|
Src: mocks.SenderAddr.String(),
|
|
Index: 0,
|
|
Data: mockNonCanonicalBlockTxs[0].Data(),
|
|
Type: mockNonCanonicalBlockTxs[0].Type(),
|
|
Value: mockNonCanonicalBlockTxs[0].Value().String(),
|
|
},
|
|
{
|
|
HeaderID: mockNonCanonicalBlock.Hash().String(),
|
|
TxHash: mockNonCanonicalBlockTxs[1].Hash().String(),
|
|
CID: trx5CID.String(),
|
|
Dst: mockNonCanonicalBlockTxs[1].To().String(),
|
|
Src: mocks.SenderAddr.String(),
|
|
Index: 1,
|
|
Data: mockNonCanonicalBlockTxs[1].Data(),
|
|
Type: mockNonCanonicalBlockTxs[1].Type(),
|
|
Value: mockNonCanonicalBlockTxs[1].Value().String(),
|
|
},
|
|
}
|
|
|
|
require.Equal(t, len(expectedBlockTxs)+len(expectedNonCanonicalBlockTxs), len(txRes))
|
|
sort.Slice(txRes, func(i, j int) bool {
|
|
if txRes[i].HeaderID == txRes[j].HeaderID {
|
|
return txRes[i].Index < txRes[j].Index
|
|
} else if txRes[i].HeaderID == mockBlock.Hash().String() {
|
|
return true
|
|
} else {
|
|
return false
|
|
}
|
|
})
|
|
|
|
for i, expectedTx := range expectedBlockTxs {
|
|
require.Equal(t, expectedTx, txRes[i])
|
|
}
|
|
for i, expectedTx := range expectedNonCanonicalBlockTxs {
|
|
require.Equal(t, expectedTx, txRes[len(expectedBlockTxs)+i])
|
|
}
|
|
|
|
// check indexed IPLD blocks
|
|
var data []byte
|
|
var prefixedKey string
|
|
|
|
txCIDs := []cid.Cid{trx1CID, trx2CID, trx3CID, trx4CID, trx5CID}
|
|
txRLPs := [][]byte{tx1, tx2, tx3, tx4, tx5}
|
|
for i, txCID := range txCIDs {
|
|
prefixedKey = shared.MultihashKeyFromCID(txCID)
|
|
err = db.Get(context.Background(), &data, ipfsPgGet, prefixedKey, mocks.BlockNumber.Uint64())
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
require.Equal(t, txRLPs[i], data)
|
|
}
|
|
})
|
|
}
|
|
|
|
func TestPGXWatchAddressMethods(t *testing.T) {
|
|
setupPGXIndexer(t)
|
|
defer tearDown(t)
|
|
defer checkTxClosure(t, 1, 0, 1)
|
|
|
|
type res struct {
|
|
Address string `db:"address"`
|
|
CreatedAt uint64 `db:"created_at"`
|
|
WatchedAt uint64 `db:"watched_at"`
|
|
LastFilledAt uint64 `db:"last_filled_at"`
|
|
}
|
|
pgStr := "SELECT * FROM eth_meta.watched_addresses"
|
|
|
|
t.Run("Load watched addresses (empty table)", func(t *testing.T) {
|
|
expectedData := []common.Address{}
|
|
|
|
rows, err := ind.LoadWatchedAddresses()
|
|
require.NoError(t, err)
|
|
|
|
expectTrue(t, len(rows) == len(expectedData))
|
|
for idx, row := range rows {
|
|
require.Equal(t, expectedData[idx], row)
|
|
}
|
|
})
|
|
|
|
t.Run("Insert watched addresses", func(t *testing.T) {
|
|
args := []sdtypes.WatchAddressArg{
|
|
{
|
|
Address: contract1Address,
|
|
CreatedAt: contract1CreatedAt,
|
|
},
|
|
{
|
|
Address: contract2Address,
|
|
CreatedAt: contract2CreatedAt,
|
|
},
|
|
}
|
|
expectedData := []res{
|
|
{
|
|
Address: contract1Address,
|
|
CreatedAt: contract1CreatedAt,
|
|
WatchedAt: watchedAt1,
|
|
LastFilledAt: lastFilledAt,
|
|
},
|
|
{
|
|
Address: contract2Address,
|
|
CreatedAt: contract2CreatedAt,
|
|
WatchedAt: watchedAt1,
|
|
LastFilledAt: lastFilledAt,
|
|
},
|
|
}
|
|
|
|
err = ind.InsertWatchedAddresses(args, big.NewInt(int64(watchedAt1)))
|
|
require.NoError(t, err)
|
|
|
|
rows := []res{}
|
|
err = db.Select(context.Background(), &rows, pgStr)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
expectTrue(t, len(rows) == len(expectedData))
|
|
for idx, row := range rows {
|
|
require.Equal(t, expectedData[idx], row)
|
|
}
|
|
})
|
|
|
|
t.Run("Insert watched addresses (some already watched)", func(t *testing.T) {
|
|
args := []sdtypes.WatchAddressArg{
|
|
{
|
|
Address: contract3Address,
|
|
CreatedAt: contract3CreatedAt,
|
|
},
|
|
{
|
|
Address: contract2Address,
|
|
CreatedAt: contract2CreatedAt,
|
|
},
|
|
}
|
|
expectedData := []res{
|
|
{
|
|
Address: contract1Address,
|
|
CreatedAt: contract1CreatedAt,
|
|
WatchedAt: watchedAt1,
|
|
LastFilledAt: lastFilledAt,
|
|
},
|
|
{
|
|
Address: contract2Address,
|
|
CreatedAt: contract2CreatedAt,
|
|
WatchedAt: watchedAt1,
|
|
LastFilledAt: lastFilledAt,
|
|
},
|
|
{
|
|
Address: contract3Address,
|
|
CreatedAt: contract3CreatedAt,
|
|
WatchedAt: watchedAt2,
|
|
LastFilledAt: lastFilledAt,
|
|
},
|
|
}
|
|
|
|
err = ind.InsertWatchedAddresses(args, big.NewInt(int64(watchedAt2)))
|
|
require.NoError(t, err)
|
|
|
|
rows := []res{}
|
|
err = db.Select(context.Background(), &rows, pgStr)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
expectTrue(t, len(rows) == len(expectedData))
|
|
for idx, row := range rows {
|
|
require.Equal(t, expectedData[idx], row)
|
|
}
|
|
})
|
|
|
|
t.Run("Remove watched addresses", func(t *testing.T) {
|
|
args := []sdtypes.WatchAddressArg{
|
|
{
|
|
Address: contract3Address,
|
|
CreatedAt: contract3CreatedAt,
|
|
},
|
|
{
|
|
Address: contract2Address,
|
|
CreatedAt: contract2CreatedAt,
|
|
},
|
|
}
|
|
expectedData := []res{
|
|
{
|
|
Address: contract1Address,
|
|
CreatedAt: contract1CreatedAt,
|
|
WatchedAt: watchedAt1,
|
|
LastFilledAt: lastFilledAt,
|
|
},
|
|
}
|
|
|
|
err = ind.RemoveWatchedAddresses(args)
|
|
require.NoError(t, err)
|
|
|
|
rows := []res{}
|
|
err = db.Select(context.Background(), &rows, pgStr)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
expectTrue(t, len(rows) == len(expectedData))
|
|
for idx, row := range rows {
|
|
require.Equal(t, expectedData[idx], row)
|
|
}
|
|
})
|
|
|
|
t.Run("Remove watched addresses (some non-watched)", func(t *testing.T) {
|
|
args := []sdtypes.WatchAddressArg{
|
|
{
|
|
Address: contract1Address,
|
|
CreatedAt: contract1CreatedAt,
|
|
},
|
|
{
|
|
Address: contract2Address,
|
|
CreatedAt: contract2CreatedAt,
|
|
},
|
|
}
|
|
expectedData := []res{}
|
|
|
|
err = ind.RemoveWatchedAddresses(args)
|
|
require.NoError(t, err)
|
|
|
|
rows := []res{}
|
|
err = db.Select(context.Background(), &rows, pgStr)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
expectTrue(t, len(rows) == len(expectedData))
|
|
for idx, row := range rows {
|
|
require.Equal(t, expectedData[idx], row)
|
|
}
|
|
})
|
|
|
|
t.Run("Set watched addresses", func(t *testing.T) {
|
|
args := []sdtypes.WatchAddressArg{
|
|
{
|
|
Address: contract1Address,
|
|
CreatedAt: contract1CreatedAt,
|
|
},
|
|
{
|
|
Address: contract2Address,
|
|
CreatedAt: contract2CreatedAt,
|
|
},
|
|
{
|
|
Address: contract3Address,
|
|
CreatedAt: contract3CreatedAt,
|
|
},
|
|
}
|
|
expectedData := []res{
|
|
{
|
|
Address: contract1Address,
|
|
CreatedAt: contract1CreatedAt,
|
|
WatchedAt: watchedAt2,
|
|
LastFilledAt: lastFilledAt,
|
|
},
|
|
{
|
|
Address: contract2Address,
|
|
CreatedAt: contract2CreatedAt,
|
|
WatchedAt: watchedAt2,
|
|
LastFilledAt: lastFilledAt,
|
|
},
|
|
{
|
|
Address: contract3Address,
|
|
CreatedAt: contract3CreatedAt,
|
|
WatchedAt: watchedAt2,
|
|
LastFilledAt: lastFilledAt,
|
|
},
|
|
}
|
|
|
|
err = ind.SetWatchedAddresses(args, big.NewInt(int64(watchedAt2)))
|
|
require.NoError(t, err)
|
|
|
|
rows := []res{}
|
|
err = db.Select(context.Background(), &rows, pgStr)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
expectTrue(t, len(rows) == len(expectedData))
|
|
for idx, row := range rows {
|
|
require.Equal(t, expectedData[idx], row)
|
|
}
|
|
})
|
|
|
|
t.Run("Set watched addresses (some already watched)", func(t *testing.T) {
|
|
args := []sdtypes.WatchAddressArg{
|
|
{
|
|
Address: contract4Address,
|
|
CreatedAt: contract4CreatedAt,
|
|
},
|
|
{
|
|
Address: contract2Address,
|
|
CreatedAt: contract2CreatedAt,
|
|
},
|
|
{
|
|
Address: contract3Address,
|
|
CreatedAt: contract3CreatedAt,
|
|
},
|
|
}
|
|
expectedData := []res{
|
|
{
|
|
Address: contract4Address,
|
|
CreatedAt: contract4CreatedAt,
|
|
WatchedAt: watchedAt3,
|
|
LastFilledAt: lastFilledAt,
|
|
},
|
|
{
|
|
Address: contract2Address,
|
|
CreatedAt: contract2CreatedAt,
|
|
WatchedAt: watchedAt3,
|
|
LastFilledAt: lastFilledAt,
|
|
},
|
|
{
|
|
Address: contract3Address,
|
|
CreatedAt: contract3CreatedAt,
|
|
WatchedAt: watchedAt3,
|
|
LastFilledAt: lastFilledAt,
|
|
},
|
|
}
|
|
|
|
err = ind.SetWatchedAddresses(args, big.NewInt(int64(watchedAt3)))
|
|
require.NoError(t, err)
|
|
|
|
rows := []res{}
|
|
err = db.Select(context.Background(), &rows, pgStr)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
expectTrue(t, len(rows) == len(expectedData))
|
|
for idx, row := range rows {
|
|
require.Equal(t, expectedData[idx], row)
|
|
}
|
|
})
|
|
|
|
t.Run("Load watched addresses", func(t *testing.T) {
|
|
expectedData := []common.Address{
|
|
common.HexToAddress(contract4Address),
|
|
common.HexToAddress(contract2Address),
|
|
common.HexToAddress(contract3Address),
|
|
}
|
|
|
|
rows, err := ind.LoadWatchedAddresses()
|
|
require.NoError(t, err)
|
|
|
|
expectTrue(t, len(rows) == len(expectedData))
|
|
for idx, row := range rows {
|
|
require.Equal(t, expectedData[idx], row)
|
|
}
|
|
})
|
|
|
|
t.Run("Clear watched addresses", func(t *testing.T) {
|
|
expectedData := []res{}
|
|
|
|
err = ind.ClearWatchedAddresses()
|
|
require.NoError(t, err)
|
|
|
|
rows := []res{}
|
|
err = db.Select(context.Background(), &rows, pgStr)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
expectTrue(t, len(rows) == len(expectedData))
|
|
for idx, row := range rows {
|
|
require.Equal(t, expectedData[idx], row)
|
|
}
|
|
})
|
|
|
|
t.Run("Clear watched addresses (empty table)", func(t *testing.T) {
|
|
expectedData := []res{}
|
|
|
|
err = ind.ClearWatchedAddresses()
|
|
require.NoError(t, err)
|
|
|
|
rows := []res{}
|
|
err = db.Select(context.Background(), &rows, pgStr)
|
|
if err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
|
|
expectTrue(t, len(rows) == len(expectedData))
|
|
for idx, row := range rows {
|
|
require.Equal(t, expectedData[idx], row)
|
|
}
|
|
})
|
|
}
|