* Statediff Geth Handle conflicts (#244) * Handle conflicts * Update go mod file versions * Make lint changes Disassociate block number from the indexer object Update ipld-eth-db ref Refactor builder code to make it reusable Use prefix comparison for account selective statediffing Update builder unit tests Add mode to write to CSV files in statediff file writer (#249) * Change file writing mode to csv files * Implement writer interface for file indexer * Implement option for csv or sql in file mode * Close files in CSV writer * Add tests for CSV file mode * Implement CSV file for watched addresses * Separate test configs for CSV and SQL * Refactor common code for file indexer tests Update indexer to include block hash in receipts and logs (#256) * Update indexer to include block hash in receipts and logs * Upgrade ipld-eth-db image in docker-compose to run tests Use watched addresses from direct indexing params by default while serving statediff APIs (#262) * Use watched addresses from direct indexing params in statediff APIs by default * Avoid using indexer object when direct indexing is off * Add nil check before accessing watched addresses from direct indexing params Rebase missed these changes needed at 1.10.20 Flags cleanup for CLI changes and linter complaints Linter appeasements to achieve perfection enforce go 1.18 for check (#267) * enforce go 1.18 for check * tests on 1.18 as well * adding db yml for possible change in docker-compose behavior in yml parsing Add indexer tests for handling non canonical blocks (#254) * Add indexer tests for header and transactions in a non canonical block * Add indexer tests for receipts in a non-canonical block and refactor * Add indexer tests for logs in a non-canonical block * Add indexer tests for state and storage nodes in a non-canonical block * Add indexer tests for non-canonical block at another height * Avoid passing address of a pointer * Update refs in GitHub workflow * Add genesis file path to stack-orchestrator config in GitHub workflow * Add descriptive comments fix non-deterministic ordering in unit tests Refactor indexer tests to avoid duplicate code (#270) * Refactor indexer tests to avoid duplicate code * Refactor file mode indexer tests * Fix expected db stats for sqlx after tx closure * Refactor indexer tests for legacy block * Refactor mainnet indexer tests * Refactor tests for watched addressess methods * Fix query in legacy indexer test rebase and resolve onto 1.10.23... still error out of index related to GetLeafKeys changed trie.Commit behavior was subtle about not not flushing to disk without an Update * no merge nodeset throws nil * linter appeasement Co-authored-by: Abdul Rabbani <abdulrabbani00@gmail.com>
430 lines
15 KiB
Go
430 lines
15 KiB
Go
// VulcanizeDB
|
|
// Copyright © 2019 Vulcanize
|
|
|
|
// This program is free software: you can redistribute it and/or modify
|
|
// it under the terms of the GNU Affero General Public License as published by
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
// (at your option) any later version.
|
|
|
|
// This program is distributed in the hope that it will be useful,
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
// GNU Affero General Public License for more details.
|
|
|
|
// You should have received a copy of the GNU Affero General Public License
|
|
// along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
package file
|
|
|
|
import (
|
|
"bufio"
|
|
"errors"
|
|
"fmt"
|
|
"io"
|
|
"math/big"
|
|
"os"
|
|
|
|
blockstore "github.com/ipfs/go-ipfs-blockstore"
|
|
dshelp "github.com/ipfs/go-ipfs-ds-help"
|
|
node "github.com/ipfs/go-ipld-format"
|
|
pg_query "github.com/pganalyze/pg_query_go/v2"
|
|
"github.com/thoas/go-funk"
|
|
|
|
"github.com/ethereum/go-ethereum/common"
|
|
"github.com/ethereum/go-ethereum/statediff/indexer/ipld"
|
|
"github.com/ethereum/go-ethereum/statediff/indexer/models"
|
|
nodeinfo "github.com/ethereum/go-ethereum/statediff/indexer/node"
|
|
"github.com/ethereum/go-ethereum/statediff/types"
|
|
)
|
|
|
|
var (
|
|
nullHash = common.HexToHash("0x0000000000000000000000000000000000000000000000000000000000000000")
|
|
pipeSize = 65336 // min(linuxPipeSize, macOSPipeSize)
|
|
writeBufferSize = pipeSize * 16 * 96
|
|
)
|
|
|
|
// SQLWriter writes sql statements to a file
|
|
type SQLWriter struct {
|
|
wc io.WriteCloser
|
|
stmts chan []byte
|
|
collatedStmt []byte
|
|
collationIndex int
|
|
|
|
flushChan chan struct{}
|
|
flushFinished chan struct{}
|
|
quitChan chan struct{}
|
|
doneChan chan struct{}
|
|
|
|
watchedAddressesFilePath string
|
|
}
|
|
|
|
// NewSQLWriter creates a new pointer to a Writer
|
|
func NewSQLWriter(wc io.WriteCloser, watchedAddressesFilePath string) *SQLWriter {
|
|
return &SQLWriter{
|
|
wc: wc,
|
|
stmts: make(chan []byte),
|
|
collatedStmt: make([]byte, writeBufferSize),
|
|
flushChan: make(chan struct{}),
|
|
flushFinished: make(chan struct{}),
|
|
quitChan: make(chan struct{}),
|
|
doneChan: make(chan struct{}),
|
|
watchedAddressesFilePath: watchedAddressesFilePath,
|
|
}
|
|
}
|
|
|
|
// Loop enables concurrent writes to the underlying os.File
|
|
// since os.File does not buffer, it utilizes an internal buffer that is the size of a unix pipe
|
|
// by using copy() and tracking the index/size of the buffer, we require only the initial memory allocation
|
|
func (sqw *SQLWriter) Loop() {
|
|
sqw.collationIndex = 0
|
|
go func() {
|
|
defer close(sqw.doneChan)
|
|
var l int
|
|
for {
|
|
select {
|
|
case stmt := <-sqw.stmts:
|
|
l = len(stmt)
|
|
if sqw.collationIndex+l > writeBufferSize {
|
|
if err := sqw.flush(); err != nil {
|
|
panic(fmt.Sprintf("error writing sql stmts buffer to file: %v", err))
|
|
}
|
|
if l > writeBufferSize {
|
|
if _, err := sqw.wc.Write(stmt); err != nil {
|
|
panic(fmt.Sprintf("error writing large sql stmt to file: %v", err))
|
|
}
|
|
continue
|
|
}
|
|
}
|
|
copy(sqw.collatedStmt[sqw.collationIndex:sqw.collationIndex+l], stmt)
|
|
sqw.collationIndex += l
|
|
case <-sqw.quitChan:
|
|
if err := sqw.flush(); err != nil {
|
|
panic(fmt.Sprintf("error writing sql stmts buffer to file: %v", err))
|
|
}
|
|
return
|
|
case <-sqw.flushChan:
|
|
if err := sqw.flush(); err != nil {
|
|
panic(fmt.Sprintf("error writing sql stmts buffer to file: %v", err))
|
|
}
|
|
sqw.flushFinished <- struct{}{}
|
|
}
|
|
}
|
|
}()
|
|
}
|
|
|
|
// Close satisfies io.Closer
|
|
func (sqw *SQLWriter) Close() error {
|
|
close(sqw.quitChan)
|
|
<-sqw.doneChan
|
|
close(sqw.stmts)
|
|
close(sqw.flushChan)
|
|
close(sqw.flushFinished)
|
|
return sqw.wc.Close()
|
|
}
|
|
|
|
// Flush sends a flush signal to the looping process
|
|
func (sqw *SQLWriter) Flush() {
|
|
sqw.flushChan <- struct{}{}
|
|
<-sqw.flushFinished
|
|
}
|
|
|
|
func (sqw *SQLWriter) flush() error {
|
|
if _, err := sqw.wc.Write(sqw.collatedStmt[0:sqw.collationIndex]); err != nil {
|
|
return err
|
|
}
|
|
sqw.collationIndex = 0
|
|
return nil
|
|
}
|
|
|
|
const (
|
|
nodeInsert = "INSERT INTO nodes (genesis_block, network_id, node_id, client_name, chain_id) VALUES " +
|
|
"('%s', '%s', '%s', '%s', %d);\n"
|
|
|
|
ipldInsert = "INSERT INTO public.blocks (block_number, key, data) VALUES ('%s', '%s', '\\x%x');\n"
|
|
|
|
headerInsert = "INSERT INTO eth.header_cids (block_number, block_hash, parent_hash, cid, td, node_id, reward, " +
|
|
"state_root, tx_root, receipt_root, uncle_root, bloom, timestamp, mh_key, times_validated, coinbase) VALUES " +
|
|
"('%s', '%s', '%s', '%s', '%s', '%s', '%s', '%s', '%s', '%s', '%s', '\\x%x', %d, '%s', %d, '%s');\n"
|
|
|
|
uncleInsert = "INSERT INTO eth.uncle_cids (block_number, block_hash, header_id, parent_hash, cid, reward, mh_key) VALUES " +
|
|
"('%s', '%s', '%s', '%s', '%s', '%s', '%s');\n"
|
|
|
|
txInsert = "INSERT INTO eth.transaction_cids (block_number, header_id, tx_hash, cid, dst, src, index, mh_key, tx_data, tx_type, " +
|
|
"value) VALUES ('%s', '%s', '%s', '%s', '%s', '%s', %d, '%s', '\\x%x', %d, '%s');\n"
|
|
|
|
alInsert = "INSERT INTO eth.access_list_elements (block_number, tx_id, index, address, storage_keys) VALUES " +
|
|
"('%s', '%s', %d, '%s', '%s');\n"
|
|
|
|
rctInsert = "INSERT INTO eth.receipt_cids (block_number, header_id, tx_id, leaf_cid, contract, contract_hash, leaf_mh_key, post_state, " +
|
|
"post_status, log_root) VALUES ('%s', '%s', '%s', '%s', '%s', '%s', '%s', '%s', %d, '%s');\n"
|
|
|
|
logInsert = "INSERT INTO eth.log_cids (block_number, header_id, leaf_cid, leaf_mh_key, rct_id, address, index, topic0, topic1, topic2, " +
|
|
"topic3, log_data) VALUES ('%s', '%s', '%s', '%s', '%s', '%s', %d, '%s', '%s', '%s', '%s', '\\x%x');\n"
|
|
|
|
stateInsert = "INSERT INTO eth.state_cids (block_number, header_id, state_leaf_key, cid, state_path, node_type, diff, mh_key) " +
|
|
"VALUES ('%s', '%s', '%s', '%s', '\\x%x', %d, %t, '%s');\n"
|
|
|
|
accountInsert = "INSERT INTO eth.state_accounts (block_number, header_id, state_path, balance, nonce, code_hash, storage_root) " +
|
|
"VALUES ('%s', '%s', '\\x%x', '%s', %d, '\\x%x', '%s');\n"
|
|
|
|
storageInsert = "INSERT INTO eth.storage_cids (block_number, header_id, state_path, storage_leaf_key, cid, storage_path, " +
|
|
"node_type, diff, mh_key) VALUES ('%s', '%s', '\\x%x', '%s', '%s', '\\x%x', %d, %t, '%s');\n"
|
|
)
|
|
|
|
func (sqw *SQLWriter) upsertNode(node nodeinfo.Info) {
|
|
sqw.stmts <- []byte(fmt.Sprintf(nodeInsert, node.GenesisBlock, node.NetworkID, node.ID, node.ClientName, node.ChainID))
|
|
}
|
|
|
|
func (sqw *SQLWriter) upsertIPLD(ipld models.IPLDModel) {
|
|
sqw.stmts <- []byte(fmt.Sprintf(ipldInsert, ipld.BlockNumber, ipld.Key, ipld.Data))
|
|
}
|
|
|
|
func (sqw *SQLWriter) upsertIPLDDirect(blockNumber, key string, value []byte) {
|
|
sqw.upsertIPLD(models.IPLDModel{
|
|
BlockNumber: blockNumber,
|
|
Key: key,
|
|
Data: value,
|
|
})
|
|
}
|
|
|
|
func (sqw *SQLWriter) upsertIPLDNode(blockNumber string, i node.Node) {
|
|
sqw.upsertIPLD(models.IPLDModel{
|
|
BlockNumber: blockNumber,
|
|
Key: blockstore.BlockPrefix.String() + dshelp.MultihashToDsKey(i.Cid().Hash()).String(),
|
|
Data: i.RawData(),
|
|
})
|
|
}
|
|
|
|
func (sqw *SQLWriter) upsertIPLDRaw(blockNumber string, codec, mh uint64, raw []byte) (string, string, error) {
|
|
c, err := ipld.RawdataToCid(codec, raw, mh)
|
|
if err != nil {
|
|
return "", "", err
|
|
}
|
|
prefixedKey := blockstore.BlockPrefix.String() + dshelp.MultihashToDsKey(c.Hash()).String()
|
|
sqw.upsertIPLD(models.IPLDModel{
|
|
BlockNumber: blockNumber,
|
|
Key: prefixedKey,
|
|
Data: raw,
|
|
})
|
|
return c.String(), prefixedKey, err
|
|
}
|
|
|
|
func (sqw *SQLWriter) upsertHeaderCID(header models.HeaderModel) {
|
|
stmt := fmt.Sprintf(headerInsert, header.BlockNumber, header.BlockHash, header.ParentHash, header.CID,
|
|
header.TotalDifficulty, header.NodeID, header.Reward, header.StateRoot, header.TxRoot,
|
|
header.RctRoot, header.UncleRoot, header.Bloom, header.Timestamp, header.MhKey, 1, header.Coinbase)
|
|
sqw.stmts <- []byte(stmt)
|
|
indexerMetrics.blocks.Inc(1)
|
|
}
|
|
|
|
func (sqw *SQLWriter) upsertUncleCID(uncle models.UncleModel) {
|
|
sqw.stmts <- []byte(fmt.Sprintf(uncleInsert, uncle.BlockNumber, uncle.BlockHash, uncle.HeaderID, uncle.ParentHash, uncle.CID,
|
|
uncle.Reward, uncle.MhKey))
|
|
}
|
|
|
|
func (sqw *SQLWriter) upsertTransactionCID(transaction models.TxModel) {
|
|
sqw.stmts <- []byte(fmt.Sprintf(txInsert, transaction.BlockNumber, transaction.HeaderID, transaction.TxHash, transaction.CID, transaction.Dst,
|
|
transaction.Src, transaction.Index, transaction.MhKey, transaction.Data, transaction.Type, transaction.Value))
|
|
indexerMetrics.transactions.Inc(1)
|
|
}
|
|
|
|
func (sqw *SQLWriter) upsertAccessListElement(accessListElement models.AccessListElementModel) {
|
|
sqw.stmts <- []byte(fmt.Sprintf(alInsert, accessListElement.BlockNumber, accessListElement.TxID, accessListElement.Index, accessListElement.Address,
|
|
formatPostgresStringArray(accessListElement.StorageKeys)))
|
|
indexerMetrics.accessListEntries.Inc(1)
|
|
}
|
|
|
|
func (sqw *SQLWriter) upsertReceiptCID(rct *models.ReceiptModel) {
|
|
sqw.stmts <- []byte(fmt.Sprintf(rctInsert, rct.BlockNumber, rct.HeaderID, rct.TxID, rct.LeafCID, rct.Contract, rct.ContractHash, rct.LeafMhKey,
|
|
rct.PostState, rct.PostStatus, rct.LogRoot))
|
|
indexerMetrics.receipts.Inc(1)
|
|
}
|
|
|
|
func (sqw *SQLWriter) upsertLogCID(logs []*models.LogsModel) {
|
|
for _, l := range logs {
|
|
sqw.stmts <- []byte(fmt.Sprintf(logInsert, l.BlockNumber, l.HeaderID, l.LeafCID, l.LeafMhKey, l.ReceiptID, l.Address, l.Index, l.Topic0,
|
|
l.Topic1, l.Topic2, l.Topic3, l.Data))
|
|
indexerMetrics.logs.Inc(1)
|
|
}
|
|
}
|
|
|
|
func (sqw *SQLWriter) upsertStateCID(stateNode models.StateNodeModel) {
|
|
var stateKey string
|
|
if stateNode.StateKey != nullHash.String() {
|
|
stateKey = stateNode.StateKey
|
|
}
|
|
sqw.stmts <- []byte(fmt.Sprintf(stateInsert, stateNode.BlockNumber, stateNode.HeaderID, stateKey, stateNode.CID, stateNode.Path,
|
|
stateNode.NodeType, true, stateNode.MhKey))
|
|
}
|
|
|
|
func (sqw *SQLWriter) upsertStateAccount(stateAccount models.StateAccountModel) {
|
|
sqw.stmts <- []byte(fmt.Sprintf(accountInsert, stateAccount.BlockNumber, stateAccount.HeaderID, stateAccount.StatePath, stateAccount.Balance,
|
|
stateAccount.Nonce, stateAccount.CodeHash, stateAccount.StorageRoot))
|
|
}
|
|
|
|
func (sqw *SQLWriter) upsertStorageCID(storageCID models.StorageNodeModel) {
|
|
var storageKey string
|
|
if storageCID.StorageKey != nullHash.String() {
|
|
storageKey = storageCID.StorageKey
|
|
}
|
|
sqw.stmts <- []byte(fmt.Sprintf(storageInsert, storageCID.BlockNumber, storageCID.HeaderID, storageCID.StatePath, storageKey, storageCID.CID,
|
|
storageCID.Path, storageCID.NodeType, true, storageCID.MhKey))
|
|
}
|
|
|
|
// LoadWatchedAddresses loads watched addresses from a file
|
|
func (sqw *SQLWriter) loadWatchedAddresses() ([]common.Address, error) {
|
|
// load sql statements from watched addresses file
|
|
stmts, err := loadWatchedAddressesStatements(sqw.watchedAddressesFilePath)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
// extract addresses from the sql statements
|
|
watchedAddresses := []common.Address{}
|
|
for _, stmt := range stmts {
|
|
addressString, err := parseWatchedAddressStatement(stmt)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
watchedAddresses = append(watchedAddresses, common.HexToAddress(addressString))
|
|
}
|
|
|
|
return watchedAddresses, nil
|
|
}
|
|
|
|
// InsertWatchedAddresses inserts the given addresses in a file
|
|
func (sqw *SQLWriter) insertWatchedAddresses(args []types.WatchAddressArg, currentBlockNumber *big.Int) error {
|
|
// load sql statements from watched addresses file
|
|
stmts, err := loadWatchedAddressesStatements(sqw.watchedAddressesFilePath)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// get already watched addresses
|
|
var watchedAddresses []string
|
|
for _, stmt := range stmts {
|
|
addressString, err := parseWatchedAddressStatement(stmt)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
watchedAddresses = append(watchedAddresses, addressString)
|
|
}
|
|
|
|
// append statements for new addresses to existing statements
|
|
for _, arg := range args {
|
|
// ignore if already watched
|
|
if funk.Contains(watchedAddresses, arg.Address) {
|
|
continue
|
|
}
|
|
|
|
stmt := fmt.Sprintf(watchedAddressesInsert, arg.Address, arg.CreatedAt, currentBlockNumber.Uint64())
|
|
stmts = append(stmts, stmt)
|
|
}
|
|
|
|
return dumpWatchedAddressesStatements(sqw.watchedAddressesFilePath, stmts)
|
|
}
|
|
|
|
// RemoveWatchedAddresses removes the given watched addresses from a file
|
|
func (sqw *SQLWriter) removeWatchedAddresses(args []types.WatchAddressArg) error {
|
|
// load sql statements from watched addresses file
|
|
stmts, err := loadWatchedAddressesStatements(sqw.watchedAddressesFilePath)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
// get rid of statements having addresses to be removed
|
|
var filteredStmts []string
|
|
for _, stmt := range stmts {
|
|
addressString, err := parseWatchedAddressStatement(stmt)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
|
|
toRemove := funk.Contains(args, func(arg types.WatchAddressArg) bool {
|
|
return arg.Address == addressString
|
|
})
|
|
|
|
if !toRemove {
|
|
filteredStmts = append(filteredStmts, stmt)
|
|
}
|
|
}
|
|
|
|
return dumpWatchedAddressesStatements(sqw.watchedAddressesFilePath, filteredStmts)
|
|
}
|
|
|
|
// SetWatchedAddresses clears and inserts the given addresses in a file
|
|
func (sqw *SQLWriter) setWatchedAddresses(args []types.WatchAddressArg, currentBlockNumber *big.Int) error {
|
|
var stmts []string
|
|
for _, arg := range args {
|
|
stmt := fmt.Sprintf(watchedAddressesInsert, arg.Address, arg.CreatedAt, currentBlockNumber.Uint64())
|
|
stmts = append(stmts, stmt)
|
|
}
|
|
|
|
return dumpWatchedAddressesStatements(sqw.watchedAddressesFilePath, stmts)
|
|
}
|
|
|
|
// loadWatchedAddressesStatements loads sql statements from the given file in a string slice
|
|
func loadWatchedAddressesStatements(filePath string) ([]string, error) {
|
|
file, err := os.Open(filePath)
|
|
if err != nil {
|
|
if errors.Is(err, os.ErrNotExist) {
|
|
return []string{}, nil
|
|
}
|
|
|
|
return nil, fmt.Errorf("error opening watched addresses file: %v", err)
|
|
}
|
|
defer file.Close()
|
|
|
|
stmts := []string{}
|
|
scanner := bufio.NewScanner(file)
|
|
for scanner.Scan() {
|
|
stmts = append(stmts, scanner.Text())
|
|
}
|
|
|
|
if err := scanner.Err(); err != nil {
|
|
return nil, fmt.Errorf("error loading watched addresses: %v", err)
|
|
}
|
|
|
|
return stmts, nil
|
|
}
|
|
|
|
// dumpWatchedAddressesStatements dumps sql statements to the given file
|
|
func dumpWatchedAddressesStatements(filePath string, stmts []string) error {
|
|
file, err := os.Create(filePath)
|
|
if err != nil {
|
|
return fmt.Errorf("error creating watched addresses file: %v", err)
|
|
}
|
|
defer file.Close()
|
|
|
|
for _, stmt := range stmts {
|
|
_, err := file.Write([]byte(stmt + "\n"))
|
|
if err != nil {
|
|
return fmt.Errorf("error inserting watched_addresses entry: %v", err)
|
|
}
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
// parseWatchedAddressStatement parses given sql insert statement to extract the address argument
|
|
func parseWatchedAddressStatement(stmt string) (string, error) {
|
|
parseResult, err := pg_query.Parse(stmt)
|
|
if err != nil {
|
|
return "", fmt.Errorf("error parsing sql stmt: %v", err)
|
|
}
|
|
|
|
// extract address argument from parse output for a SQL statement of form
|
|
// "INSERT INTO eth_meta.watched_addresses (address, created_at, watched_at)
|
|
// VALUES ('0xabc', '123', '130') ON CONFLICT (address) DO NOTHING;"
|
|
addressString := parseResult.Stmts[0].Stmt.GetInsertStmt().
|
|
SelectStmt.GetSelectStmt().
|
|
ValuesLists[0].GetList().
|
|
Items[0].GetAConst().
|
|
GetVal().
|
|
GetString_().
|
|
Str
|
|
|
|
return addressString, nil
|
|
}
|