mirror of
https://github.com/cerc-io/watcher-ts
synced 2025-01-08 12:28:05 +00:00
Refactor code to initialize clients (#290)
* Refactor CLI code * Refactor code to initialize clients * Don't return config from function to initialize clients
This commit is contained in:
parent
a846917f1d
commit
bb1345c696
@ -2,15 +2,11 @@
|
|||||||
// Copyright 2021 Vulcanize, Inc.
|
// Copyright 2021 Vulcanize, Inc.
|
||||||
//
|
//
|
||||||
|
|
||||||
import assert from 'assert';
|
|
||||||
import yargs from 'yargs';
|
import yargs from 'yargs';
|
||||||
import 'reflect-metadata';
|
import 'reflect-metadata';
|
||||||
import debug from 'debug';
|
import debug from 'debug';
|
||||||
import { getDefaultProvider } from 'ethers';
|
|
||||||
|
|
||||||
import { Config, DEFAULT_CONFIG_PATH, getConfig } from '@vulcanize/util';
|
import { Config, DEFAULT_CONFIG_PATH, getConfig, initClients } from '@vulcanize/util';
|
||||||
import { getCache } from '@vulcanize/cache';
|
|
||||||
import { EthClient } from '@vulcanize/ipld-eth-client';
|
|
||||||
|
|
||||||
import { Database } from '../database';
|
import { Database } from '../database';
|
||||||
import { Indexer } from '../indexer';
|
import { Indexer } from '../indexer';
|
||||||
@ -42,36 +38,12 @@ const main = async (): Promise<void> => {
|
|||||||
}).argv;
|
}).argv;
|
||||||
|
|
||||||
const config: Config = await getConfig(argv.configFile);
|
const config: Config = await getConfig(argv.configFile);
|
||||||
|
const { ethClient, postgraphileClient, ethProvider } = await initClients(config);
|
||||||
|
|
||||||
const { upstream, database: dbConfig, server: serverConfig } = config;
|
const db = new Database(config.database);
|
||||||
|
|
||||||
assert(upstream, 'Missing upstream config');
|
|
||||||
assert(dbConfig, 'Missing database config');
|
|
||||||
assert(serverConfig, 'Missing server config');
|
|
||||||
|
|
||||||
const db = new Database(dbConfig);
|
|
||||||
await db.init();
|
await db.init();
|
||||||
|
|
||||||
const { ethServer: { gqlApiEndpoint, gqlPostgraphileEndpoint, rpcProviderEndpoint }, cache: cacheConfig } = upstream;
|
const indexer = new Indexer(config.server, db, ethClient, postgraphileClient, ethProvider);
|
||||||
assert(gqlApiEndpoint, 'Missing upstream ethServer.gqlApiEndpoint');
|
|
||||||
assert(gqlPostgraphileEndpoint, 'Missing upstream ethServer.gqlPostgraphileEndpoint');
|
|
||||||
|
|
||||||
const cache = await getCache(cacheConfig);
|
|
||||||
|
|
||||||
const ethClient = new EthClient({
|
|
||||||
gqlEndpoint: gqlApiEndpoint,
|
|
||||||
gqlSubscriptionEndpoint: gqlPostgraphileEndpoint,
|
|
||||||
cache
|
|
||||||
});
|
|
||||||
|
|
||||||
const postgraphileClient = new EthClient({
|
|
||||||
gqlEndpoint: gqlPostgraphileEndpoint,
|
|
||||||
cache
|
|
||||||
});
|
|
||||||
|
|
||||||
const ethProvider = getDefaultProvider(rpcProviderEndpoint);
|
|
||||||
|
|
||||||
const indexer = new Indexer(serverConfig, db, ethClient, postgraphileClient, ethProvider);
|
|
||||||
const blockHash = await indexer.processCLICheckpoint(argv.address, argv.blockHash);
|
const blockHash = await indexer.processCLICheckpoint(argv.address, argv.blockHash);
|
||||||
|
|
||||||
log(`Created a checkpoint for contract ${argv.address} at block-hash ${blockHash}`);
|
log(`Created a checkpoint for contract ${argv.address} at block-hash ${blockHash}`);
|
||||||
|
@ -6,13 +6,10 @@ import assert from 'assert';
|
|||||||
import yargs from 'yargs';
|
import yargs from 'yargs';
|
||||||
import 'reflect-metadata';
|
import 'reflect-metadata';
|
||||||
import debug from 'debug';
|
import debug from 'debug';
|
||||||
import { getDefaultProvider } from 'ethers';
|
|
||||||
import fs from 'fs';
|
import fs from 'fs';
|
||||||
import path from 'path';
|
import path from 'path';
|
||||||
|
|
||||||
import { Config, DEFAULT_CONFIG_PATH, getConfig } from '@vulcanize/util';
|
import { Config, DEFAULT_CONFIG_PATH, getConfig, initClients } from '@vulcanize/util';
|
||||||
import { getCache } from '@vulcanize/cache';
|
|
||||||
import { EthClient } from '@vulcanize/ipld-eth-client';
|
|
||||||
import * as codec from '@ipld/dag-cbor';
|
import * as codec from '@ipld/dag-cbor';
|
||||||
|
|
||||||
import { Database } from '../database';
|
import { Database } from '../database';
|
||||||
@ -40,36 +37,12 @@ const main = async (): Promise<void> => {
|
|||||||
}).argv;
|
}).argv;
|
||||||
|
|
||||||
const config: Config = await getConfig(argv.configFile);
|
const config: Config = await getConfig(argv.configFile);
|
||||||
|
const { ethClient, postgraphileClient, ethProvider } = await initClients(config);
|
||||||
|
|
||||||
const { upstream, database: dbConfig, server: serverConfig } = config;
|
const db = new Database(config.database);
|
||||||
|
|
||||||
assert(upstream, 'Missing upstream config');
|
|
||||||
assert(dbConfig, 'Missing database config');
|
|
||||||
assert(serverConfig, 'Missing server config');
|
|
||||||
|
|
||||||
const db = new Database(dbConfig);
|
|
||||||
await db.init();
|
await db.init();
|
||||||
|
|
||||||
const { ethServer: { gqlApiEndpoint, gqlPostgraphileEndpoint, rpcProviderEndpoint }, cache: cacheConfig } = upstream;
|
const indexer = new Indexer(config.server, db, ethClient, postgraphileClient, ethProvider);
|
||||||
assert(gqlApiEndpoint, 'Missing upstream ethServer.gqlApiEndpoint');
|
|
||||||
assert(gqlPostgraphileEndpoint, 'Missing upstream ethServer.gqlPostgraphileEndpoint');
|
|
||||||
|
|
||||||
const cache = await getCache(cacheConfig);
|
|
||||||
|
|
||||||
const ethClient = new EthClient({
|
|
||||||
gqlEndpoint: gqlApiEndpoint,
|
|
||||||
gqlSubscriptionEndpoint: gqlPostgraphileEndpoint,
|
|
||||||
cache
|
|
||||||
});
|
|
||||||
|
|
||||||
const postgraphileClient = new EthClient({
|
|
||||||
gqlEndpoint: gqlPostgraphileEndpoint,
|
|
||||||
cache
|
|
||||||
});
|
|
||||||
|
|
||||||
const ethProvider = getDefaultProvider(rpcProviderEndpoint);
|
|
||||||
|
|
||||||
const indexer = new Indexer(serverConfig, db, ethClient, postgraphileClient, ethProvider);
|
|
||||||
|
|
||||||
const exportData: any = {
|
const exportData: any = {
|
||||||
snapshotBlock: {},
|
snapshotBlock: {},
|
||||||
|
@ -9,9 +9,7 @@ import { hideBin } from 'yargs/helpers';
|
|||||||
import debug from 'debug';
|
import debug from 'debug';
|
||||||
import { PubSub } from 'apollo-server-express';
|
import { PubSub } from 'apollo-server-express';
|
||||||
|
|
||||||
import { getCache } from '@vulcanize/cache';
|
import { Config, getConfig, fillBlocks, JobQueue, DEFAULT_CONFIG_PATH, initClients } from '@vulcanize/util';
|
||||||
import { EthClient } from '@vulcanize/ipld-eth-client';
|
|
||||||
import { getConfig, fillBlocks, JobQueue, DEFAULT_CONFIG_PATH, getCustomProvider } from '@vulcanize/util';
|
|
||||||
|
|
||||||
import { Database } from './database';
|
import { Database } from './database';
|
||||||
import { Indexer } from './indexer';
|
import { Indexer } from './indexer';
|
||||||
@ -42,40 +40,19 @@ export const main = async (): Promise<any> => {
|
|||||||
}
|
}
|
||||||
}).argv;
|
}).argv;
|
||||||
|
|
||||||
const config = await getConfig(argv.configFile);
|
const config: Config = await getConfig(argv.configFile);
|
||||||
|
const { ethClient, postgraphileClient, ethProvider } = await initClients(config);
|
||||||
|
|
||||||
const { upstream, database: dbConfig, jobQueue: jobQueueConfig, server: serverConfig } = config;
|
const db = new Database(config.database);
|
||||||
|
|
||||||
assert(upstream, 'Missing upstream config');
|
|
||||||
assert(dbConfig, 'Missing database config');
|
|
||||||
assert(serverConfig, 'Missing server config');
|
|
||||||
|
|
||||||
const db = new Database(dbConfig);
|
|
||||||
await db.init();
|
await db.init();
|
||||||
|
|
||||||
assert(upstream, 'Missing upstream config');
|
|
||||||
const { ethServer: { gqlApiEndpoint, gqlPostgraphileEndpoint, rpcProviderEndpoint, blockDelayInMilliSecs }, cache: cacheConfig } = upstream;
|
|
||||||
assert(gqlPostgraphileEndpoint, 'Missing upstream ethServer.gqlPostgraphileEndpoint');
|
|
||||||
|
|
||||||
const cache = await getCache(cacheConfig);
|
|
||||||
|
|
||||||
const ethClient = new EthClient({
|
|
||||||
gqlEndpoint: gqlApiEndpoint,
|
|
||||||
gqlSubscriptionEndpoint: gqlPostgraphileEndpoint,
|
|
||||||
cache
|
|
||||||
});
|
|
||||||
|
|
||||||
const postgraphileClient = new EthClient({
|
|
||||||
gqlEndpoint: gqlPostgraphileEndpoint,
|
|
||||||
cache
|
|
||||||
});
|
|
||||||
|
|
||||||
const ethProvider = getCustomProvider(rpcProviderEndpoint);
|
|
||||||
|
|
||||||
// Note: In-memory pubsub works fine for now, as each watcher is a single process anyway.
|
// Note: In-memory pubsub works fine for now, as each watcher is a single process anyway.
|
||||||
// Later: https://www.apollographql.com/docs/apollo-server/data/subscriptions/#production-pubsub-libraries
|
// Later: https://www.apollographql.com/docs/apollo-server/data/subscriptions/#production-pubsub-libraries
|
||||||
const pubsub = new PubSub();
|
const pubsub = new PubSub();
|
||||||
const indexer = new Indexer(serverConfig, db, ethClient, postgraphileClient, ethProvider);
|
const indexer = new Indexer(config.server, db, ethClient, postgraphileClient, ethProvider);
|
||||||
|
|
||||||
|
const jobQueueConfig = config.jobQueue;
|
||||||
|
assert(jobQueueConfig, 'Missing job queue config');
|
||||||
|
|
||||||
const { dbConnectionString, maxCompletionLagInSecs } = jobQueueConfig;
|
const { dbConnectionString, maxCompletionLagInSecs } = jobQueueConfig;
|
||||||
assert(dbConnectionString, 'Missing job queue db connection string');
|
assert(dbConnectionString, 'Missing job queue db connection string');
|
||||||
@ -83,11 +60,9 @@ export const main = async (): Promise<any> => {
|
|||||||
const jobQueue = new JobQueue({ dbConnectionString, maxCompletionLag: maxCompletionLagInSecs });
|
const jobQueue = new JobQueue({ dbConnectionString, maxCompletionLag: maxCompletionLagInSecs });
|
||||||
await jobQueue.start();
|
await jobQueue.start();
|
||||||
|
|
||||||
const eventWatcher = new EventWatcher(upstream, ethClient, postgraphileClient, indexer, pubsub, jobQueue);
|
const eventWatcher = new EventWatcher(config.upstream, ethClient, postgraphileClient, indexer, pubsub, jobQueue);
|
||||||
|
|
||||||
assert(jobQueueConfig, 'Missing job queue config');
|
await fillBlocks(jobQueue, indexer, postgraphileClient, eventWatcher, argv);
|
||||||
|
|
||||||
await fillBlocks(jobQueue, indexer, postgraphileClient, eventWatcher, blockDelayInMilliSecs, argv);
|
|
||||||
};
|
};
|
||||||
|
|
||||||
main().catch(err => {
|
main().catch(err => {
|
||||||
|
@ -8,13 +8,10 @@ import yargs from 'yargs';
|
|||||||
import { hideBin } from 'yargs/helpers';
|
import { hideBin } from 'yargs/helpers';
|
||||||
import debug from 'debug';
|
import debug from 'debug';
|
||||||
import { PubSub } from 'apollo-server-express';
|
import { PubSub } from 'apollo-server-express';
|
||||||
import { getDefaultProvider } from 'ethers';
|
|
||||||
import fs from 'fs';
|
import fs from 'fs';
|
||||||
import path from 'path';
|
import path from 'path';
|
||||||
|
|
||||||
import { getCache } from '@vulcanize/cache';
|
import { getConfig, fillBlocks, JobQueue, DEFAULT_CONFIG_PATH, Config, initClients } from '@vulcanize/util';
|
||||||
import { EthClient } from '@vulcanize/ipld-eth-client';
|
|
||||||
import { getConfig, fillBlocks, JobQueue, DEFAULT_CONFIG_PATH } from '@vulcanize/util';
|
|
||||||
import * as codec from '@ipld/dag-cbor';
|
import * as codec from '@ipld/dag-cbor';
|
||||||
|
|
||||||
import { Database } from '../database';
|
import { Database } from '../database';
|
||||||
@ -43,40 +40,19 @@ export const main = async (): Promise<any> => {
|
|||||||
}
|
}
|
||||||
}).argv;
|
}).argv;
|
||||||
|
|
||||||
const config = await getConfig(argv.configFile);
|
const config: Config = await getConfig(argv.configFile);
|
||||||
|
const { ethClient, postgraphileClient, ethProvider } = await initClients(config);
|
||||||
|
|
||||||
const { upstream, database: dbConfig, jobQueue: jobQueueConfig, server: serverConfig } = config;
|
const db = new Database(config.database);
|
||||||
|
|
||||||
assert(upstream, 'Missing upstream config');
|
|
||||||
assert(dbConfig, 'Missing database config');
|
|
||||||
assert(serverConfig, 'Missing server config');
|
|
||||||
|
|
||||||
const db = new Database(dbConfig);
|
|
||||||
await db.init();
|
await db.init();
|
||||||
|
|
||||||
assert(upstream, 'Missing upstream config');
|
|
||||||
const { ethServer: { gqlPostgraphileEndpoint, rpcProviderEndpoint }, cache: cacheConfig } = upstream;
|
|
||||||
assert(gqlPostgraphileEndpoint, 'Missing upstream ethServer.gqlPostgraphileEndpoint');
|
|
||||||
|
|
||||||
const cache = await getCache(cacheConfig);
|
|
||||||
|
|
||||||
const ethClient = new EthClient({
|
|
||||||
gqlEndpoint: gqlPostgraphileEndpoint,
|
|
||||||
gqlSubscriptionEndpoint: gqlPostgraphileEndpoint,
|
|
||||||
cache
|
|
||||||
});
|
|
||||||
|
|
||||||
const postgraphileClient = new EthClient({
|
|
||||||
gqlEndpoint: gqlPostgraphileEndpoint,
|
|
||||||
cache
|
|
||||||
});
|
|
||||||
|
|
||||||
const ethProvider = getDefaultProvider(rpcProviderEndpoint);
|
|
||||||
|
|
||||||
// Note: In-memory pubsub works fine for now, as each watcher is a single process anyway.
|
// Note: In-memory pubsub works fine for now, as each watcher is a single process anyway.
|
||||||
// Later: https://www.apollographql.com/docs/apollo-server/data/subscriptions/#production-pubsub-libraries
|
// Later: https://www.apollographql.com/docs/apollo-server/data/subscriptions/#production-pubsub-libraries
|
||||||
const pubsub = new PubSub();
|
const pubsub = new PubSub();
|
||||||
const indexer = new Indexer(serverConfig, db, ethClient, postgraphileClient, ethProvider);
|
const indexer = new Indexer(config.server, db, ethClient, postgraphileClient, ethProvider);
|
||||||
|
|
||||||
|
const jobQueueConfig = config.jobQueue;
|
||||||
|
assert(jobQueueConfig, 'Missing job queue config');
|
||||||
|
|
||||||
const { dbConnectionString, maxCompletionLagInSecs } = jobQueueConfig;
|
const { dbConnectionString, maxCompletionLagInSecs } = jobQueueConfig;
|
||||||
assert(dbConnectionString, 'Missing job queue db connection string');
|
assert(dbConnectionString, 'Missing job queue db connection string');
|
||||||
@ -86,8 +62,6 @@ export const main = async (): Promise<any> => {
|
|||||||
|
|
||||||
const eventWatcher = new EventWatcher(ethClient, indexer, pubsub, jobQueue);
|
const eventWatcher = new EventWatcher(ethClient, indexer, pubsub, jobQueue);
|
||||||
|
|
||||||
assert(jobQueueConfig, 'Missing job queue config');
|
|
||||||
|
|
||||||
// Import data.
|
// Import data.
|
||||||
const importFilePath = path.resolve(argv.importFile);
|
const importFilePath = path.resolve(argv.importFile);
|
||||||
const encodedImportData = fs.readFileSync(importFilePath);
|
const encodedImportData = fs.readFileSync(importFilePath);
|
||||||
|
@ -6,12 +6,9 @@ import assert from 'assert';
|
|||||||
import yargs from 'yargs';
|
import yargs from 'yargs';
|
||||||
import 'reflect-metadata';
|
import 'reflect-metadata';
|
||||||
import debug from 'debug';
|
import debug from 'debug';
|
||||||
import { getDefaultProvider } from 'ethers';
|
|
||||||
import util from 'util';
|
import util from 'util';
|
||||||
|
|
||||||
import { Config, DEFAULT_CONFIG_PATH, getConfig } from '@vulcanize/util';
|
import { Config, DEFAULT_CONFIG_PATH, getConfig, initClients } from '@vulcanize/util';
|
||||||
import { getCache } from '@vulcanize/cache';
|
|
||||||
import { EthClient } from '@vulcanize/ipld-eth-client';
|
|
||||||
|
|
||||||
import { Database } from '../database';
|
import { Database } from '../database';
|
||||||
import { Indexer } from '../indexer';
|
import { Indexer } from '../indexer';
|
||||||
@ -39,36 +36,12 @@ const main = async (): Promise<void> => {
|
|||||||
}).argv;
|
}).argv;
|
||||||
|
|
||||||
const config: Config = await getConfig(argv.configFile);
|
const config: Config = await getConfig(argv.configFile);
|
||||||
|
const { ethClient, postgraphileClient, ethProvider } = await initClients(config);
|
||||||
|
|
||||||
const { upstream, database: dbConfig, server: serverConfig } = config;
|
const db = new Database(config.database);
|
||||||
|
|
||||||
assert(upstream, 'Missing upstream config');
|
|
||||||
assert(dbConfig, 'Missing database config');
|
|
||||||
assert(serverConfig, 'Missing server config');
|
|
||||||
|
|
||||||
const db = new Database(dbConfig);
|
|
||||||
await db.init();
|
await db.init();
|
||||||
|
|
||||||
const { ethServer: { gqlApiEndpoint, gqlPostgraphileEndpoint, rpcProviderEndpoint }, cache: cacheConfig } = upstream;
|
const indexer = new Indexer(config.server, db, ethClient, postgraphileClient, ethProvider);
|
||||||
assert(gqlApiEndpoint, 'Missing upstream ethServer.gqlApiEndpoint');
|
|
||||||
assert(gqlPostgraphileEndpoint, 'Missing upstream ethServer.gqlPostgraphileEndpoint');
|
|
||||||
|
|
||||||
const cache = await getCache(cacheConfig);
|
|
||||||
|
|
||||||
const ethClient = new EthClient({
|
|
||||||
gqlEndpoint: gqlApiEndpoint,
|
|
||||||
gqlSubscriptionEndpoint: gqlPostgraphileEndpoint,
|
|
||||||
cache
|
|
||||||
});
|
|
||||||
|
|
||||||
const postgraphileClient = new EthClient({
|
|
||||||
gqlEndpoint: gqlPostgraphileEndpoint,
|
|
||||||
cache
|
|
||||||
});
|
|
||||||
|
|
||||||
const ethProvider = getDefaultProvider(rpcProviderEndpoint);
|
|
||||||
|
|
||||||
const indexer = new Indexer(serverConfig, db, ethClient, postgraphileClient, ethProvider);
|
|
||||||
|
|
||||||
const ipldBlock = await indexer.getIPLDBlockByCid(argv.cid);
|
const ipldBlock = await indexer.getIPLDBlockByCid(argv.cid);
|
||||||
assert(ipldBlock, 'IPLDBlock for the provided CID doesn\'t exist.');
|
assert(ipldBlock, 'IPLDBlock for the provided CID doesn\'t exist.');
|
||||||
|
@ -7,11 +7,9 @@ import 'reflect-metadata';
|
|||||||
import yargs from 'yargs';
|
import yargs from 'yargs';
|
||||||
import { hideBin } from 'yargs/helpers';
|
import { hideBin } from 'yargs/helpers';
|
||||||
import debug from 'debug';
|
import debug from 'debug';
|
||||||
|
|
||||||
import { getCache } from '@vulcanize/cache';
|
|
||||||
import { EthClient } from '@vulcanize/ipld-eth-client';
|
|
||||||
import {
|
import {
|
||||||
getConfig,
|
getConfig,
|
||||||
|
Config,
|
||||||
JobQueue,
|
JobQueue,
|
||||||
JobRunner as BaseJobRunner,
|
JobRunner as BaseJobRunner,
|
||||||
QUEUE_BLOCK_PROCESSING,
|
QUEUE_BLOCK_PROCESSING,
|
||||||
@ -21,7 +19,7 @@ import {
|
|||||||
QUEUE_IPFS,
|
QUEUE_IPFS,
|
||||||
JobQueueConfig,
|
JobQueueConfig,
|
||||||
DEFAULT_CONFIG_PATH,
|
DEFAULT_CONFIG_PATH,
|
||||||
getCustomProvider
|
initClients
|
||||||
} from '@vulcanize/util';
|
} from '@vulcanize/util';
|
||||||
|
|
||||||
import { Indexer } from './indexer';
|
import { Indexer } from './indexer';
|
||||||
@ -123,37 +121,15 @@ export const main = async (): Promise<any> => {
|
|||||||
})
|
})
|
||||||
.argv;
|
.argv;
|
||||||
|
|
||||||
const config = await getConfig(argv.f);
|
const config: Config = await getConfig(argv.f);
|
||||||
|
const { ethClient, postgraphileClient, ethProvider } = await initClients(config);
|
||||||
|
|
||||||
const { upstream, database: dbConfig, jobQueue: jobQueueConfig, server: serverConfig } = config;
|
const db = new Database(config.database);
|
||||||
|
|
||||||
assert(upstream, 'Missing upstream config');
|
|
||||||
assert(dbConfig, 'Missing database config');
|
|
||||||
assert(serverConfig, 'Missing server config');
|
|
||||||
|
|
||||||
const db = new Database(dbConfig);
|
|
||||||
await db.init();
|
await db.init();
|
||||||
|
|
||||||
const { ethServer: { gqlApiEndpoint, gqlPostgraphileEndpoint, rpcProviderEndpoint }, cache: cacheConfig } = upstream;
|
const indexer = new Indexer(config.server, db, ethClient, postgraphileClient, ethProvider);
|
||||||
assert(gqlApiEndpoint, 'Missing upstream ethServer.gqlApiEndpoint');
|
|
||||||
assert(gqlPostgraphileEndpoint, 'Missing upstream ethServer.gqlPostgraphileEndpoint');
|
|
||||||
|
|
||||||
const cache = await getCache(cacheConfig);
|
|
||||||
|
|
||||||
const ethClient = new EthClient({
|
|
||||||
gqlEndpoint: gqlApiEndpoint,
|
|
||||||
gqlSubscriptionEndpoint: gqlPostgraphileEndpoint,
|
|
||||||
cache
|
|
||||||
});
|
|
||||||
|
|
||||||
const postgraphileClient = new EthClient({
|
|
||||||
gqlEndpoint: gqlPostgraphileEndpoint,
|
|
||||||
cache
|
|
||||||
});
|
|
||||||
|
|
||||||
const ethProvider = getCustomProvider(rpcProviderEndpoint);
|
|
||||||
const indexer = new Indexer(db, ethClient, postgraphileClient, ethProvider);
|
|
||||||
|
|
||||||
|
const jobQueueConfig = config.jobQueue;
|
||||||
assert(jobQueueConfig, 'Missing job queue config');
|
assert(jobQueueConfig, 'Missing job queue config');
|
||||||
|
|
||||||
const { dbConnectionString, maxCompletionLagInSecs } = jobQueueConfig;
|
const { dbConnectionString, maxCompletionLagInSecs } = jobQueueConfig;
|
||||||
@ -162,7 +138,7 @@ export const main = async (): Promise<any> => {
|
|||||||
const jobQueue = new JobQueue({ dbConnectionString, maxCompletionLag: maxCompletionLagInSecs });
|
const jobQueue = new JobQueue({ dbConnectionString, maxCompletionLag: maxCompletionLagInSecs });
|
||||||
await jobQueue.start();
|
await jobQueue.start();
|
||||||
|
|
||||||
const jobRunner = new JobRunner(jobQueueConfig, serverConfig, indexer, jobQueue);
|
const jobRunner = new JobRunner(jobQueueConfig, config.server, indexer, jobQueue);
|
||||||
await jobRunner.start();
|
await jobRunner.start();
|
||||||
};
|
};
|
||||||
|
|
||||||
|
@ -6,7 +6,7 @@ import debug from 'debug';
|
|||||||
import { MoreThan } from 'typeorm';
|
import { MoreThan } from 'typeorm';
|
||||||
import assert from 'assert';
|
import assert from 'assert';
|
||||||
|
|
||||||
import { getConfig, getResetConfig, resetJobs } from '@vulcanize/util';
|
import { getConfig, initClients, resetJobs } from '@vulcanize/util';
|
||||||
|
|
||||||
import { Database } from '../../database';
|
import { Database } from '../../database';
|
||||||
import { Indexer } from '../../indexer';
|
import { Indexer } from '../../indexer';
|
||||||
@ -31,13 +31,13 @@ export const builder = {
|
|||||||
export const handler = async (argv: any): Promise<void> => {
|
export const handler = async (argv: any): Promise<void> => {
|
||||||
const config = await getConfig(argv.configFile);
|
const config = await getConfig(argv.configFile);
|
||||||
await resetJobs(config);
|
await resetJobs(config);
|
||||||
const { dbConfig, serverConfig, ethClient, postgraphileClient, ethProvider } = await getResetConfig(config);
|
const { ethClient, postgraphileClient, ethProvider } = await initClients(config);
|
||||||
|
|
||||||
// Initialize database.
|
// Initialize database.
|
||||||
const db = new Database(dbConfig);
|
const db = new Database(config.database);
|
||||||
await db.init();
|
await db.init();
|
||||||
|
|
||||||
const indexer = new Indexer(serverConfig, db, ethClient, postgraphileClient, ethProvider);
|
const indexer = new Indexer(config.server, db, ethClient, postgraphileClient, ethProvider);
|
||||||
|
|
||||||
const syncStatus = await indexer.getSyncStatus();
|
const syncStatus = await indexer.getSyncStatus();
|
||||||
assert(syncStatus, 'Missing syncStatus');
|
assert(syncStatus, 'Missing syncStatus');
|
||||||
|
@ -14,9 +14,7 @@ import debug from 'debug';
|
|||||||
import 'graphql-import-node';
|
import 'graphql-import-node';
|
||||||
import { createServer } from 'http';
|
import { createServer } from 'http';
|
||||||
|
|
||||||
import { getCache } from '@vulcanize/cache';
|
import { DEFAULT_CONFIG_PATH, getConfig, Config, JobQueue, KIND_ACTIVE, initClients } from '@vulcanize/util';
|
||||||
import { EthClient } from '@vulcanize/ipld-eth-client';
|
|
||||||
import { DEFAULT_CONFIG_PATH, getConfig, JobQueue, KIND_ACTIVE, getCustomProvider } from '@vulcanize/util';
|
|
||||||
|
|
||||||
import { createResolvers } from './resolvers';
|
import { createResolvers } from './resolvers';
|
||||||
import { Indexer } from './indexer';
|
import { Indexer } from './indexer';
|
||||||
@ -36,51 +34,28 @@ export const main = async (): Promise<any> => {
|
|||||||
})
|
})
|
||||||
.argv;
|
.argv;
|
||||||
|
|
||||||
const config = await getConfig(argv.f);
|
const config: Config = await getConfig(argv.f);
|
||||||
|
const { ethClient, postgraphileClient, ethProvider } = await initClients(config);
|
||||||
|
|
||||||
const { upstream, database: dbConfig, jobQueue: jobQueueConfig, server: serverConfig } = config;
|
const { host, port, kind: watcherKind } = config.server;
|
||||||
|
|
||||||
assert(upstream, 'Missing upstream config');
|
const db = new Database(config.database);
|
||||||
assert(dbConfig, 'Missing database config');
|
|
||||||
assert(serverConfig, 'Missing server config');
|
|
||||||
|
|
||||||
const { host, port, kind: watcherKind } = serverConfig;
|
|
||||||
|
|
||||||
const db = new Database(dbConfig);
|
|
||||||
await db.init();
|
await db.init();
|
||||||
|
|
||||||
const { ethServer: { gqlApiEndpoint, gqlPostgraphileEndpoint, rpcProviderEndpoint }, cache: cacheConfig } = upstream;
|
|
||||||
assert(gqlApiEndpoint, 'Missing upstream ethServer.gqlApiEndpoint');
|
|
||||||
assert(gqlPostgraphileEndpoint, 'Missing upstream ethServer.gqlPostgraphileEndpoint');
|
|
||||||
|
|
||||||
const cache = await getCache(cacheConfig);
|
|
||||||
|
|
||||||
const ethClient = new EthClient({
|
|
||||||
gqlEndpoint: gqlApiEndpoint,
|
|
||||||
gqlSubscriptionEndpoint: gqlPostgraphileEndpoint,
|
|
||||||
cache
|
|
||||||
});
|
|
||||||
|
|
||||||
const postgraphileClient = new EthClient({
|
|
||||||
gqlEndpoint: gqlPostgraphileEndpoint,
|
|
||||||
cache
|
|
||||||
});
|
|
||||||
|
|
||||||
const ethProvider = getCustomProvider(rpcProviderEndpoint);
|
|
||||||
|
|
||||||
const indexer = new Indexer(serverConfig, db, ethClient, postgraphileClient, ethProvider);
|
|
||||||
|
|
||||||
// Note: In-memory pubsub works fine for now, as each watcher is a single process anyway.
|
// Note: In-memory pubsub works fine for now, as each watcher is a single process anyway.
|
||||||
// Later: https://www.apollographql.com/docs/apollo-server/data/subscriptions/#production-pubsub-libraries
|
// Later: https://www.apollographql.com/docs/apollo-server/data/subscriptions/#production-pubsub-libraries
|
||||||
const pubsub = new PubSub();
|
const pubsub = new PubSub();
|
||||||
|
const indexer = new Indexer(config.server, db, ethClient, ethProvider);
|
||||||
|
|
||||||
|
const jobQueueConfig = config.jobQueue;
|
||||||
assert(jobQueueConfig, 'Missing job queue config');
|
assert(jobQueueConfig, 'Missing job queue config');
|
||||||
|
|
||||||
const { dbConnectionString, maxCompletionLagInSecs } = jobQueueConfig;
|
const { dbConnectionString, maxCompletionLagInSecs } = jobQueueConfig;
|
||||||
assert(dbConnectionString, 'Missing job queue db connection string');
|
assert(dbConnectionString, 'Missing job queue db connection string');
|
||||||
|
|
||||||
const jobQueue = new JobQueue({ dbConnectionString, maxCompletionLag: maxCompletionLagInSecs });
|
const jobQueue = new JobQueue({ dbConnectionString, maxCompletionLag: maxCompletionLagInSecs });
|
||||||
|
|
||||||
const eventWatcher = new EventWatcher(upstream, ethClient, postgraphileClient, indexer, pubsub, jobQueue);
|
const eventWatcher = new EventWatcher(config.upstream, ethClient, postgraphileClient, indexer, pubsub, jobQueue);
|
||||||
|
|
||||||
if (watcherKind === KIND_ACTIVE) {
|
if (watcherKind === KIND_ACTIVE) {
|
||||||
await jobQueue.start();
|
await jobQueue.start();
|
||||||
|
@ -2,15 +2,11 @@
|
|||||||
// Copyright 2021 Vulcanize, Inc.
|
// Copyright 2021 Vulcanize, Inc.
|
||||||
//
|
//
|
||||||
|
|
||||||
import assert from 'assert';
|
|
||||||
import yargs from 'yargs';
|
import yargs from 'yargs';
|
||||||
import 'reflect-metadata';
|
import 'reflect-metadata';
|
||||||
import debug from 'debug';
|
import debug from 'debug';
|
||||||
import { getDefaultProvider } from 'ethers';
|
|
||||||
|
|
||||||
import { Config, DEFAULT_CONFIG_PATH, getConfig } from '@vulcanize/util';
|
import { Config, DEFAULT_CONFIG_PATH, getConfig, initClients } from '@vulcanize/util';
|
||||||
import { getCache } from '@vulcanize/cache';
|
|
||||||
import { EthClient } from '@vulcanize/ipld-eth-client';
|
|
||||||
|
|
||||||
import { Database } from '../database';
|
import { Database } from '../database';
|
||||||
import { Indexer } from '../indexer';
|
import { Indexer } from '../indexer';
|
||||||
@ -54,36 +50,12 @@ const main = async (): Promise<void> => {
|
|||||||
}).argv;
|
}).argv;
|
||||||
|
|
||||||
const config: Config = await getConfig(argv.configFile);
|
const config: Config = await getConfig(argv.configFile);
|
||||||
|
const { ethClient, postgraphileClient, ethProvider } = await initClients(config);
|
||||||
|
|
||||||
const { upstream, database: dbConfig, server: serverConfig } = config;
|
const db = new Database(config.database);
|
||||||
|
|
||||||
assert(upstream, 'Missing upstream config');
|
|
||||||
assert(dbConfig, 'Missing database config');
|
|
||||||
assert(serverConfig, 'Missing server config');
|
|
||||||
|
|
||||||
const db = new Database(dbConfig);
|
|
||||||
await db.init();
|
await db.init();
|
||||||
|
|
||||||
const { ethServer: { gqlApiEndpoint, gqlPostgraphileEndpoint, rpcProviderEndpoint }, cache: cacheConfig } = upstream;
|
const indexer = new Indexer(config.server, db, ethClient, postgraphileClient, ethProvider);
|
||||||
assert(gqlApiEndpoint, 'Missing upstream ethServer.gqlApiEndpoint');
|
|
||||||
assert(gqlPostgraphileEndpoint, 'Missing upstream ethServer.gqlPostgraphileEndpoint');
|
|
||||||
|
|
||||||
const cache = await getCache(cacheConfig);
|
|
||||||
|
|
||||||
const ethClient = new EthClient({
|
|
||||||
gqlEndpoint: gqlApiEndpoint,
|
|
||||||
gqlSubscriptionEndpoint: gqlPostgraphileEndpoint,
|
|
||||||
cache
|
|
||||||
});
|
|
||||||
|
|
||||||
const postgraphileClient = new EthClient({
|
|
||||||
gqlEndpoint: gqlPostgraphileEndpoint,
|
|
||||||
cache
|
|
||||||
});
|
|
||||||
|
|
||||||
const ethProvider = getDefaultProvider(rpcProviderEndpoint);
|
|
||||||
|
|
||||||
const indexer = new Indexer(serverConfig, db, ethClient, postgraphileClient, ethProvider);
|
|
||||||
await indexer.watchContract(argv.address, argv.kind, argv.checkpoint, argv.startingBlock);
|
await indexer.watchContract(argv.address, argv.kind, argv.checkpoint, argv.startingBlock);
|
||||||
|
|
||||||
await db.close();
|
await db.close();
|
||||||
|
@ -6,7 +6,7 @@ import debug from 'debug';
|
|||||||
import { MoreThan } from 'typeorm';
|
import { MoreThan } from 'typeorm';
|
||||||
import assert from 'assert';
|
import assert from 'assert';
|
||||||
|
|
||||||
import { getConfig, getResetConfig, JobQueue, resetJobs } from '@vulcanize/util';
|
import { getConfig, initClients, JobQueue, resetJobs } from '@vulcanize/util';
|
||||||
|
|
||||||
import { Database } from '../../database';
|
import { Database } from '../../database';
|
||||||
import { Indexer } from '../../indexer';
|
import { Indexer } from '../../indexer';
|
||||||
@ -29,13 +29,13 @@ export const builder = {
|
|||||||
export const handler = async (argv: any): Promise<void> => {
|
export const handler = async (argv: any): Promise<void> => {
|
||||||
const config = await getConfig(argv.configFile);
|
const config = await getConfig(argv.configFile);
|
||||||
await resetJobs(config);
|
await resetJobs(config);
|
||||||
const { jobQueue: jobQueueConfig } = config;
|
const { ethClient, postgraphileClient, ethProvider } = await initClients(config);
|
||||||
const { dbConfig, serverConfig, ethClient, postgraphileClient, ethProvider } = await getResetConfig(config);
|
|
||||||
|
|
||||||
// Initialize database.
|
// Initialize database.
|
||||||
const db = new Database(dbConfig);
|
const db = new Database(config.database);
|
||||||
await db.init();
|
await db.init();
|
||||||
|
|
||||||
|
const { jobQueue: jobQueueConfig } = config;
|
||||||
assert(jobQueueConfig, 'Missing job queue config');
|
assert(jobQueueConfig, 'Missing job queue config');
|
||||||
|
|
||||||
const { dbConnectionString, maxCompletionLagInSecs } = jobQueueConfig;
|
const { dbConnectionString, maxCompletionLagInSecs } = jobQueueConfig;
|
||||||
@ -43,7 +43,7 @@ export const handler = async (argv: any): Promise<void> => {
|
|||||||
|
|
||||||
const jobQueue = new JobQueue({ dbConnectionString, maxCompletionLag: maxCompletionLagInSecs });
|
const jobQueue = new JobQueue({ dbConnectionString, maxCompletionLag: maxCompletionLagInSecs });
|
||||||
|
|
||||||
const indexer = new Indexer(db, ethClient, postgraphileClient, ethProvider, jobQueue, serverConfig.mode);
|
const indexer = new Indexer(db, ethClient, postgraphileClient, ethProvider, jobQueue, config.server.mode);
|
||||||
|
|
||||||
const syncStatus = await indexer.getSyncStatus();
|
const syncStatus = await indexer.getSyncStatus();
|
||||||
assert(syncStatus, 'Missing syncStatus');
|
assert(syncStatus, 'Missing syncStatus');
|
||||||
|
@ -6,7 +6,7 @@ import assert from 'assert';
|
|||||||
import yargs from 'yargs';
|
import yargs from 'yargs';
|
||||||
import 'reflect-metadata';
|
import 'reflect-metadata';
|
||||||
|
|
||||||
import { Config, DEFAULT_CONFIG_PATH, getConfig, getResetConfig, JobQueue } from '@vulcanize/util';
|
import { Config, DEFAULT_CONFIG_PATH, getConfig, initClients, JobQueue } from '@vulcanize/util';
|
||||||
|
|
||||||
import { Database } from '../database';
|
import { Database } from '../database';
|
||||||
import { Indexer } from '../indexer';
|
import { Indexer } from '../indexer';
|
||||||
@ -43,7 +43,7 @@ import { Indexer } from '../indexer';
|
|||||||
|
|
||||||
const config: Config = await getConfig(argv.configFile);
|
const config: Config = await getConfig(argv.configFile);
|
||||||
const { database: dbConfig, server: { mode }, jobQueue: jobQueueConfig } = config;
|
const { database: dbConfig, server: { mode }, jobQueue: jobQueueConfig } = config;
|
||||||
const { ethClient, postgraphileClient, ethProvider } = await getResetConfig(config);
|
const { ethClient, postgraphileClient, ethProvider } = await initClients(config);
|
||||||
|
|
||||||
assert(dbConfig);
|
assert(dbConfig);
|
||||||
|
|
||||||
|
@ -9,9 +9,7 @@ import { hideBin } from 'yargs/helpers';
|
|||||||
import debug from 'debug';
|
import debug from 'debug';
|
||||||
import { PubSub } from 'apollo-server-express';
|
import { PubSub } from 'apollo-server-express';
|
||||||
|
|
||||||
import { getCache } from '@vulcanize/cache';
|
import { Config, getConfig, fillBlocks, JobQueue, DEFAULT_CONFIG_PATH, initClients } from '@vulcanize/util';
|
||||||
import { EthClient } from '@vulcanize/ipld-eth-client';
|
|
||||||
import { getConfig, fillBlocks, JobQueue, DEFAULT_CONFIG_PATH, getCustomProvider } from '@vulcanize/util';
|
|
||||||
|
|
||||||
import { Database } from './database';
|
import { Database } from './database';
|
||||||
import { Indexer } from './indexer';
|
import { Indexer } from './indexer';
|
||||||
@ -55,53 +53,30 @@ export const main = async (): Promise<any> => {
|
|||||||
}
|
}
|
||||||
}).argv;
|
}).argv;
|
||||||
|
|
||||||
const config = await getConfig(argv.configFile);
|
const config: Config = await getConfig(argv.configFile);
|
||||||
|
const { ethClient, postgraphileClient, ethProvider } = await initClients(config);
|
||||||
|
|
||||||
assert(config.server, 'Missing server config');
|
const db = new Database(config.database);
|
||||||
|
|
||||||
const { upstream, database: dbConfig, jobQueue: jobQueueConfig, server: { mode } } = config;
|
|
||||||
|
|
||||||
assert(dbConfig, 'Missing database config');
|
|
||||||
|
|
||||||
const db = new Database(dbConfig);
|
|
||||||
await db.init();
|
await db.init();
|
||||||
|
|
||||||
assert(upstream, 'Missing upstream config');
|
|
||||||
const { ethServer: { gqlApiEndpoint, gqlPostgraphileEndpoint, rpcProviderEndpoint, blockDelayInMilliSecs }, cache: cacheConfig } = upstream;
|
|
||||||
assert(gqlPostgraphileEndpoint, 'Missing upstream ethServer.gqlPostgraphileEndpoint');
|
|
||||||
|
|
||||||
const cache = await getCache(cacheConfig);
|
|
||||||
|
|
||||||
const ethClient = new EthClient({
|
|
||||||
gqlEndpoint: gqlApiEndpoint,
|
|
||||||
gqlSubscriptionEndpoint: gqlPostgraphileEndpoint,
|
|
||||||
cache
|
|
||||||
});
|
|
||||||
|
|
||||||
const postgraphileClient = new EthClient({
|
|
||||||
gqlEndpoint: gqlPostgraphileEndpoint,
|
|
||||||
cache
|
|
||||||
});
|
|
||||||
|
|
||||||
const ethProvider = getCustomProvider(rpcProviderEndpoint);
|
|
||||||
|
|
||||||
// Note: In-memory pubsub works fine for now, as each watcher is a single process anyway.
|
// Note: In-memory pubsub works fine for now, as each watcher is a single process anyway.
|
||||||
// Later: https://www.apollographql.com/docs/apollo-server/data/subscriptions/#production-pubsub-libraries
|
// Later: https://www.apollographql.com/docs/apollo-server/data/subscriptions/#production-pubsub-libraries
|
||||||
const pubsub = new PubSub();
|
const pubsub = new PubSub();
|
||||||
|
|
||||||
|
const jobQueueConfig = config.jobQueue;
|
||||||
|
assert(jobQueueConfig, 'Missing job queue config');
|
||||||
|
|
||||||
const { dbConnectionString, maxCompletionLagInSecs } = jobQueueConfig;
|
const { dbConnectionString, maxCompletionLagInSecs } = jobQueueConfig;
|
||||||
assert(dbConnectionString, 'Missing job queue db connection string');
|
assert(dbConnectionString, 'Missing job queue db connection string');
|
||||||
|
|
||||||
const jobQueue = new JobQueue({ dbConnectionString, maxCompletionLag: maxCompletionLagInSecs });
|
const jobQueue = new JobQueue({ dbConnectionString, maxCompletionLag: maxCompletionLagInSecs });
|
||||||
await jobQueue.start();
|
await jobQueue.start();
|
||||||
|
|
||||||
const indexer = new Indexer(db, ethClient, postgraphileClient, ethProvider, jobQueue, mode);
|
const indexer = new Indexer(db, ethClient, postgraphileClient, ethProvider, jobQueue, config.server.mode);
|
||||||
|
|
||||||
const eventWatcher = new EventWatcher(upstream, ethClient, postgraphileClient, indexer, pubsub, jobQueue);
|
const eventWatcher = new EventWatcher(config.upstream, ethClient, postgraphileClient, indexer, pubsub, jobQueue);
|
||||||
|
|
||||||
assert(jobQueueConfig, 'Missing job queue config');
|
await fillBlocks(jobQueue, indexer, eventWatcher, config.upstream.ethServer.blockDelayInMilliSecs, argv);
|
||||||
|
|
||||||
await fillBlocks(jobQueue, indexer, eventWatcher, blockDelayInMilliSecs, argv);
|
|
||||||
};
|
};
|
||||||
|
|
||||||
main().catch(err => {
|
main().catch(err => {
|
||||||
|
@ -8,18 +8,16 @@ import yargs from 'yargs';
|
|||||||
import { hideBin } from 'yargs/helpers';
|
import { hideBin } from 'yargs/helpers';
|
||||||
import debug from 'debug';
|
import debug from 'debug';
|
||||||
|
|
||||||
import { getCache } from '@vulcanize/cache';
|
|
||||||
import { EthClient } from '@vulcanize/ipld-eth-client';
|
|
||||||
import {
|
import {
|
||||||
getConfig,
|
getConfig,
|
||||||
|
Config,
|
||||||
JobQueue,
|
JobQueue,
|
||||||
JobRunner as BaseJobRunner,
|
JobRunner as BaseJobRunner,
|
||||||
QUEUE_BLOCK_PROCESSING,
|
QUEUE_BLOCK_PROCESSING,
|
||||||
QUEUE_EVENT_PROCESSING,
|
QUEUE_EVENT_PROCESSING,
|
||||||
JobQueueConfig,
|
JobQueueConfig,
|
||||||
DEFAULT_CONFIG_PATH,
|
DEFAULT_CONFIG_PATH,
|
||||||
getCustomProvider,
|
initClients
|
||||||
ServerConfig
|
|
||||||
} from '@vulcanize/util';
|
} from '@vulcanize/util';
|
||||||
|
|
||||||
import { Indexer } from './indexer';
|
import { Indexer } from './indexer';
|
||||||
@ -32,14 +30,12 @@ export class JobRunner {
|
|||||||
_jobQueue: JobQueue
|
_jobQueue: JobQueue
|
||||||
_baseJobRunner: BaseJobRunner
|
_baseJobRunner: BaseJobRunner
|
||||||
_jobQueueConfig: JobQueueConfig
|
_jobQueueConfig: JobQueueConfig
|
||||||
_serverConfig: ServerConfig
|
|
||||||
|
|
||||||
constructor (jobQueueConfig: JobQueueConfig, serverConfig: ServerConfig, indexer: Indexer, jobQueue: JobQueue) {
|
constructor (jobQueueConfig: JobQueueConfig, indexer: Indexer, jobQueue: JobQueue) {
|
||||||
this._jobQueueConfig = jobQueueConfig;
|
this._jobQueueConfig = jobQueueConfig;
|
||||||
this._indexer = indexer;
|
this._indexer = indexer;
|
||||||
this._jobQueue = jobQueue;
|
this._jobQueue = jobQueue;
|
||||||
this._serverConfig = serverConfig;
|
this._baseJobRunner = new BaseJobRunner(this._jobQueueConfig, this._indexer, this._jobQueue);
|
||||||
this._baseJobRunner = new BaseJobRunner(this._jobQueueConfig, this._serverConfig, this._indexer, this._jobQueue);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
async start (): Promise<void> {
|
async start (): Promise<void> {
|
||||||
@ -71,36 +67,13 @@ export const main = async (): Promise<any> => {
|
|||||||
})
|
})
|
||||||
.argv;
|
.argv;
|
||||||
|
|
||||||
const config = await getConfig(argv.f);
|
const config: Config = await getConfig(argv.f);
|
||||||
|
const { ethClient, postgraphileClient, ethProvider } = await initClients(config);
|
||||||
|
|
||||||
const { upstream, database: dbConfig, jobQueue: jobQueueConfig, server: serverConfig } = config;
|
const db = new Database(config.database);
|
||||||
|
|
||||||
assert(upstream, 'Missing upstream config');
|
|
||||||
assert(dbConfig, 'Missing database config');
|
|
||||||
assert(serverConfig, 'Missing server config');
|
|
||||||
|
|
||||||
const db = new Database(dbConfig);
|
|
||||||
await db.init();
|
await db.init();
|
||||||
|
|
||||||
const { ethServer: { gqlApiEndpoint, gqlPostgraphileEndpoint, rpcProviderEndpoint }, cache: cacheConfig } = upstream;
|
const jobQueueConfig = config.jobQueue;
|
||||||
assert(gqlApiEndpoint, 'Missing upstream ethServer.gqlApiEndpoint');
|
|
||||||
assert(gqlPostgraphileEndpoint, 'Missing upstream ethServer.gqlPostgraphileEndpoint');
|
|
||||||
|
|
||||||
const cache = await getCache(cacheConfig);
|
|
||||||
|
|
||||||
const ethClient = new EthClient({
|
|
||||||
gqlEndpoint: gqlApiEndpoint,
|
|
||||||
gqlSubscriptionEndpoint: gqlPostgraphileEndpoint,
|
|
||||||
cache
|
|
||||||
});
|
|
||||||
|
|
||||||
const postgraphileClient = new EthClient({
|
|
||||||
gqlEndpoint: gqlPostgraphileEndpoint,
|
|
||||||
cache
|
|
||||||
});
|
|
||||||
|
|
||||||
const ethProvider = getCustomProvider(rpcProviderEndpoint);
|
|
||||||
|
|
||||||
assert(jobQueueConfig, 'Missing job queue config');
|
assert(jobQueueConfig, 'Missing job queue config');
|
||||||
|
|
||||||
const { dbConnectionString, maxCompletionLagInSecs } = jobQueueConfig;
|
const { dbConnectionString, maxCompletionLagInSecs } = jobQueueConfig;
|
||||||
@ -109,9 +82,9 @@ export const main = async (): Promise<any> => {
|
|||||||
const jobQueue = new JobQueue({ dbConnectionString, maxCompletionLag: maxCompletionLagInSecs });
|
const jobQueue = new JobQueue({ dbConnectionString, maxCompletionLag: maxCompletionLagInSecs });
|
||||||
await jobQueue.start();
|
await jobQueue.start();
|
||||||
|
|
||||||
const indexer = new Indexer(db, ethClient, postgraphileClient, ethProvider, jobQueue, serverConfig.mode);
|
const indexer = new Indexer(db, ethClient, postgraphileClient, ethProvider, jobQueue, config.server.mode);
|
||||||
|
|
||||||
const jobRunner = new JobRunner(jobQueueConfig, serverConfig, indexer, jobQueue);
|
const jobRunner = new JobRunner(jobQueueConfig, indexer, jobQueue);
|
||||||
await jobRunner.start();
|
await jobRunner.start();
|
||||||
};
|
};
|
||||||
|
|
||||||
|
@ -12,9 +12,7 @@ import debug from 'debug';
|
|||||||
import 'graphql-import-node';
|
import 'graphql-import-node';
|
||||||
import { createServer } from 'http';
|
import { createServer } from 'http';
|
||||||
|
|
||||||
import { getCache } from '@vulcanize/cache';
|
import { DEFAULT_CONFIG_PATH, getConfig, Config, JobQueue, KIND_ACTIVE, initClients } from '@vulcanize/util';
|
||||||
import { EthClient } from '@vulcanize/ipld-eth-client';
|
|
||||||
import { DEFAULT_CONFIG_PATH, getConfig, getCustomProvider, JobQueue, KIND_ACTIVE } from '@vulcanize/util';
|
|
||||||
|
|
||||||
import typeDefs from './schema';
|
import typeDefs from './schema';
|
||||||
|
|
||||||
@ -37,43 +35,19 @@ export const main = async (): Promise<any> => {
|
|||||||
})
|
})
|
||||||
.argv;
|
.argv;
|
||||||
|
|
||||||
const config = await getConfig(argv.f);
|
const config: Config = await getConfig(argv.f);
|
||||||
|
const { ethClient, postgraphileClient, ethProvider } = await initClients(config);
|
||||||
assert(config.server, 'Missing server config');
|
|
||||||
|
|
||||||
const { host, port, mode, kind: watcherKind } = config.server;
|
const { host, port, mode, kind: watcherKind } = config.server;
|
||||||
|
|
||||||
const { upstream, database: dbConfig, jobQueue: jobQueueConfig } = config;
|
const db = new Database(config.database);
|
||||||
|
|
||||||
assert(dbConfig, 'Missing database config');
|
|
||||||
|
|
||||||
const db = new Database(dbConfig);
|
|
||||||
await db.init();
|
await db.init();
|
||||||
|
|
||||||
assert(upstream, 'Missing upstream config');
|
|
||||||
const { ethServer: { gqlApiEndpoint, gqlPostgraphileEndpoint, rpcProviderEndpoint }, cache: cacheConfig } = upstream;
|
|
||||||
assert(gqlApiEndpoint, 'Missing upstream ethServer.gqlApiEndpoint');
|
|
||||||
assert(gqlPostgraphileEndpoint, 'Missing upstream ethServer.gqlPostgraphileEndpoint');
|
|
||||||
|
|
||||||
const cache = await getCache(cacheConfig);
|
|
||||||
|
|
||||||
const ethClient = new EthClient({
|
|
||||||
gqlEndpoint: gqlApiEndpoint,
|
|
||||||
gqlSubscriptionEndpoint: gqlPostgraphileEndpoint,
|
|
||||||
cache
|
|
||||||
});
|
|
||||||
|
|
||||||
const postgraphileClient = new EthClient({
|
|
||||||
gqlEndpoint: gqlPostgraphileEndpoint,
|
|
||||||
cache
|
|
||||||
});
|
|
||||||
|
|
||||||
const ethProvider = getCustomProvider(rpcProviderEndpoint);
|
|
||||||
|
|
||||||
// Note: In-memory pubsub works fine for now, as each watcher is a single process anyway.
|
// Note: In-memory pubsub works fine for now, as each watcher is a single process anyway.
|
||||||
// Later: https://www.apollographql.com/docs/apollo-server/data/subscriptions/#production-pubsub-libraries
|
// Later: https://www.apollographql.com/docs/apollo-server/data/subscriptions/#production-pubsub-libraries
|
||||||
const pubsub = new PubSub();
|
const pubsub = new PubSub();
|
||||||
|
|
||||||
|
const jobQueueConfig = config.jobQueue;
|
||||||
assert(jobQueueConfig, 'Missing job queue config');
|
assert(jobQueueConfig, 'Missing job queue config');
|
||||||
|
|
||||||
const { dbConnectionString, maxCompletionLagInSecs } = jobQueueConfig;
|
const { dbConnectionString, maxCompletionLagInSecs } = jobQueueConfig;
|
||||||
@ -83,7 +57,7 @@ export const main = async (): Promise<any> => {
|
|||||||
|
|
||||||
const indexer = new Indexer(db, ethClient, postgraphileClient, ethProvider, jobQueue, mode);
|
const indexer = new Indexer(db, ethClient, postgraphileClient, ethProvider, jobQueue, mode);
|
||||||
|
|
||||||
const eventWatcher = new EventWatcher(upstream, ethClient, postgraphileClient, indexer, pubsub, jobQueue);
|
const eventWatcher = new EventWatcher(config.upstream, ethClient, postgraphileClient, indexer, pubsub, jobQueue);
|
||||||
|
|
||||||
if (watcherKind === KIND_ACTIVE) {
|
if (watcherKind === KIND_ACTIVE) {
|
||||||
await jobQueue.start();
|
await jobQueue.start();
|
||||||
|
@ -6,7 +6,7 @@ import debug from 'debug';
|
|||||||
import { MoreThan } from 'typeorm';
|
import { MoreThan } from 'typeorm';
|
||||||
import assert from 'assert';
|
import assert from 'assert';
|
||||||
|
|
||||||
import { getConfig, getResetConfig, JobQueue, resetJobs } from '@vulcanize/util';
|
import { getConfig, initClients, JobQueue, resetJobs } from '@vulcanize/util';
|
||||||
import { Client as ERC20Client } from '@vulcanize/erc20-watcher';
|
import { Client as ERC20Client } from '@vulcanize/erc20-watcher';
|
||||||
import { Client as UniClient } from '@vulcanize/uni-watcher';
|
import { Client as UniClient } from '@vulcanize/uni-watcher';
|
||||||
|
|
||||||
@ -39,16 +39,16 @@ export const handler = async (argv: any): Promise<void> => {
|
|||||||
const config = await getConfig(argv.configFile);
|
const config = await getConfig(argv.configFile);
|
||||||
await resetJobs(config);
|
await resetJobs(config);
|
||||||
const { jobQueue: jobQueueConfig } = config;
|
const { jobQueue: jobQueueConfig } = config;
|
||||||
const { dbConfig, serverConfig, upstreamConfig, ethClient, postgraphileClient, ethProvider } = await getResetConfig(config);
|
const { ethClient, postgraphileClient, ethProvider } = await initClients(config);
|
||||||
|
|
||||||
// Initialize database.
|
// Initialize database.
|
||||||
const db = new Database(dbConfig);
|
const db = new Database(config.database);
|
||||||
await db.init();
|
await db.init();
|
||||||
|
|
||||||
const {
|
const {
|
||||||
uniWatcher,
|
uniWatcher,
|
||||||
tokenWatcher
|
tokenWatcher
|
||||||
} = upstreamConfig;
|
} = config.upstream;
|
||||||
|
|
||||||
const uniClient = new UniClient(uniWatcher);
|
const uniClient = new UniClient(uniWatcher);
|
||||||
const erc20Client = new ERC20Client(tokenWatcher);
|
const erc20Client = new ERC20Client(tokenWatcher);
|
||||||
@ -61,7 +61,7 @@ export const handler = async (argv: any): Promise<void> => {
|
|||||||
const jobQueue = new JobQueue({ dbConnectionString, maxCompletionLag: maxCompletionLagInSecs });
|
const jobQueue = new JobQueue({ dbConnectionString, maxCompletionLag: maxCompletionLagInSecs });
|
||||||
await jobQueue.start();
|
await jobQueue.start();
|
||||||
|
|
||||||
const indexer = new Indexer(db, uniClient, erc20Client, ethClient, postgraphileClient, ethProvider, jobQueue, serverConfig.mode);
|
const indexer = new Indexer(db, uniClient, erc20Client, ethClient, postgraphileClient, ethProvider, jobQueue, config.server.mode);
|
||||||
|
|
||||||
const syncStatus = await indexer.getSyncStatus();
|
const syncStatus = await indexer.getSyncStatus();
|
||||||
assert(syncStatus, 'Missing syncStatus');
|
assert(syncStatus, 'Missing syncStatus');
|
||||||
|
@ -8,9 +8,7 @@ import yargs from 'yargs';
|
|||||||
import { hideBin } from 'yargs/helpers';
|
import { hideBin } from 'yargs/helpers';
|
||||||
import debug from 'debug';
|
import debug from 'debug';
|
||||||
|
|
||||||
import { getCache } from '@vulcanize/cache';
|
import { getConfig, Config, fillBlocks, JobQueue, DEFAULT_CONFIG_PATH, initClients } from '@vulcanize/util';
|
||||||
import { EthClient } from '@vulcanize/ipld-eth-client';
|
|
||||||
import { getConfig, fillBlocks, JobQueue, DEFAULT_CONFIG_PATH, getCustomProvider } from '@vulcanize/util';
|
|
||||||
import { Client as UniClient } from '@vulcanize/uni-watcher';
|
import { Client as UniClient } from '@vulcanize/uni-watcher';
|
||||||
import { Client as ERC20Client } from '@vulcanize/erc20-watcher';
|
import { Client as ERC20Client } from '@vulcanize/erc20-watcher';
|
||||||
|
|
||||||
@ -57,52 +55,33 @@ export const main = async (): Promise<any> => {
|
|||||||
}
|
}
|
||||||
}).argv;
|
}).argv;
|
||||||
|
|
||||||
const config = await getConfig(argv.configFile);
|
const config: Config = await getConfig(argv.configFile);
|
||||||
|
const { ethClient, postgraphileClient, ethProvider } = await initClients(config);
|
||||||
|
|
||||||
assert(config.server, 'Missing server config');
|
const db = new Database(config.database);
|
||||||
|
|
||||||
const { upstream, database: dbConfig, jobQueue: jobQueueConfig, server: { mode } } = config;
|
|
||||||
|
|
||||||
assert(dbConfig, 'Missing database config');
|
|
||||||
|
|
||||||
const db = new Database(dbConfig);
|
|
||||||
await db.init();
|
await db.init();
|
||||||
|
|
||||||
assert(upstream, 'Missing upstream config');
|
const { uniWatcher, tokenWatcher, ethServer: { blockDelayInMilliSecs } } = config.upstream;
|
||||||
const { ethServer: { gqlApiEndpoint, gqlPostgraphileEndpoint, rpcProviderEndpoint, blockDelayInMilliSecs }, cache: cacheConfig, uniWatcher, tokenWatcher } = upstream;
|
|
||||||
assert(gqlPostgraphileEndpoint, 'Missing upstream ethServer.gqlPostgraphileEndpoint');
|
|
||||||
|
|
||||||
const cache = await getCache(cacheConfig);
|
|
||||||
|
|
||||||
const ethClient = new EthClient({
|
|
||||||
gqlEndpoint: gqlApiEndpoint,
|
|
||||||
gqlSubscriptionEndpoint: gqlPostgraphileEndpoint,
|
|
||||||
cache
|
|
||||||
});
|
|
||||||
|
|
||||||
const postgraphileClient = new EthClient({
|
|
||||||
gqlEndpoint: gqlPostgraphileEndpoint,
|
|
||||||
cache
|
|
||||||
});
|
|
||||||
|
|
||||||
const uniClient = new UniClient(uniWatcher);
|
const uniClient = new UniClient(uniWatcher);
|
||||||
const erc20Client = new ERC20Client(tokenWatcher);
|
const erc20Client = new ERC20Client(tokenWatcher);
|
||||||
const ethProvider = getCustomProvider(rpcProviderEndpoint);
|
|
||||||
|
|
||||||
// Note: In-memory pubsub works fine for now, as each watcher is a single process anyway.
|
// Note: In-memory pubsub works fine for now, as each watcher is a single process anyway.
|
||||||
// Later: https://www.apollographql.com/docs/apollo-server/data/subscriptions/#production-pubsub-libraries
|
// Later: https://www.apollographql.com/docs/apollo-server/data/subscriptions/#production-pubsub-libraries
|
||||||
const pubsub = new PubSub();
|
const pubsub = new PubSub();
|
||||||
|
|
||||||
|
const jobQueueConfig = config.jobQueue;
|
||||||
assert(jobQueueConfig, 'Missing job queue config');
|
assert(jobQueueConfig, 'Missing job queue config');
|
||||||
|
|
||||||
const { dbConnectionString, maxCompletionLagInSecs } = jobQueueConfig;
|
const { dbConnectionString, maxCompletionLagInSecs } = jobQueueConfig;
|
||||||
assert(dbConnectionString, 'Missing job queue db connection string');
|
assert(dbConnectionString, 'Missing job queue db connection string');
|
||||||
|
|
||||||
const jobQueue = new JobQueue({ dbConnectionString, maxCompletionLag: maxCompletionLagInSecs });
|
const jobQueue = new JobQueue({ dbConnectionString, maxCompletionLag: maxCompletionLagInSecs });
|
||||||
await jobQueue.start();
|
await jobQueue.start();
|
||||||
|
|
||||||
const indexer = new Indexer(db, uniClient, erc20Client, ethClient, postgraphileClient, ethProvider, jobQueue, mode);
|
const indexer = new Indexer(db, uniClient, erc20Client, ethClient, postgraphileClient, ethProvider, jobQueue, config.server.mode);
|
||||||
|
|
||||||
const eventWatcher = new EventWatcher(upstream, ethClient, postgraphileClient, indexer, pubsub, jobQueue);
|
const eventWatcher = new EventWatcher(config.upstream, ethClient, postgraphileClient, indexer, pubsub, jobQueue);
|
||||||
|
|
||||||
await fillBlocks(jobQueue, indexer, eventWatcher, blockDelayInMilliSecs, argv);
|
await fillBlocks(jobQueue, indexer, eventWatcher, blockDelayInMilliSecs, argv);
|
||||||
};
|
};
|
||||||
|
@ -10,18 +10,17 @@ import debug from 'debug';
|
|||||||
|
|
||||||
import { Client as ERC20Client } from '@vulcanize/erc20-watcher';
|
import { Client as ERC20Client } from '@vulcanize/erc20-watcher';
|
||||||
import { Client as UniClient } from '@vulcanize/uni-watcher';
|
import { Client as UniClient } from '@vulcanize/uni-watcher';
|
||||||
import { getCache } from '@vulcanize/cache';
|
|
||||||
import { EthClient } from '@vulcanize/ipld-eth-client';
|
|
||||||
import {
|
import {
|
||||||
getConfig,
|
getConfig,
|
||||||
|
Config,
|
||||||
JobQueue,
|
JobQueue,
|
||||||
QUEUE_BLOCK_PROCESSING,
|
QUEUE_BLOCK_PROCESSING,
|
||||||
QUEUE_EVENT_PROCESSING,
|
QUEUE_EVENT_PROCESSING,
|
||||||
JobRunner as BaseJobRunner,
|
JobRunner as BaseJobRunner,
|
||||||
JobQueueConfig,
|
JobQueueConfig,
|
||||||
DEFAULT_CONFIG_PATH,
|
DEFAULT_CONFIG_PATH,
|
||||||
getCustomProvider,
|
initClients
|
||||||
ServerConfig
|
|
||||||
} from '@vulcanize/util';
|
} from '@vulcanize/util';
|
||||||
|
|
||||||
import { Indexer } from './indexer';
|
import { Indexer } from './indexer';
|
||||||
@ -34,14 +33,12 @@ export class JobRunner {
|
|||||||
_jobQueue: JobQueue
|
_jobQueue: JobQueue
|
||||||
_baseJobRunner: BaseJobRunner
|
_baseJobRunner: BaseJobRunner
|
||||||
_jobQueueConfig: JobQueueConfig
|
_jobQueueConfig: JobQueueConfig
|
||||||
_serverConfig: ServerConfig
|
|
||||||
|
|
||||||
constructor (jobQueueConfig: JobQueueConfig, serverConfig: ServerConfig, indexer: Indexer, jobQueue: JobQueue) {
|
constructor (jobQueueConfig: JobQueueConfig, indexer: Indexer, jobQueue: JobQueue) {
|
||||||
this._jobQueueConfig = jobQueueConfig;
|
this._jobQueueConfig = jobQueueConfig;
|
||||||
this._indexer = indexer;
|
this._indexer = indexer;
|
||||||
this._jobQueue = jobQueue;
|
this._jobQueue = jobQueue;
|
||||||
this._serverConfig = serverConfig;
|
this._baseJobRunner = new BaseJobRunner(this._jobQueueConfig, this._indexer, this._jobQueue);
|
||||||
this._baseJobRunner = new BaseJobRunner(this._jobQueueConfig, this._serverConfig, this._indexer, this._jobQueue);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
async start (): Promise<void> {
|
async start (): Promise<void> {
|
||||||
@ -73,58 +70,21 @@ export const main = async (): Promise<any> => {
|
|||||||
})
|
})
|
||||||
.argv;
|
.argv;
|
||||||
|
|
||||||
const config = await getConfig(argv.f);
|
const config: Config = await getConfig(argv.f);
|
||||||
|
const { ethClient, postgraphileClient, ethProvider } = await initClients(config);
|
||||||
|
|
||||||
const { upstream, database: dbConfig, jobQueue: jobQueueConfig, server: serverConfig } = config;
|
const db = new Database(config.database);
|
||||||
|
|
||||||
assert(upstream, 'Missing upstream config');
|
|
||||||
assert(dbConfig, 'Missing database config');
|
|
||||||
assert(serverConfig, 'Missing server config');
|
|
||||||
|
|
||||||
const db = new Database(dbConfig);
|
|
||||||
await db.init();
|
await db.init();
|
||||||
|
|
||||||
assert(upstream, 'Missing upstream config');
|
|
||||||
|
|
||||||
const {
|
const {
|
||||||
uniWatcher: {
|
uniWatcher,
|
||||||
gqlEndpoint,
|
tokenWatcher
|
||||||
gqlSubscriptionEndpoint
|
} = config.upstream;
|
||||||
},
|
|
||||||
tokenWatcher,
|
|
||||||
cache: cacheConfig,
|
|
||||||
ethServer: {
|
|
||||||
gqlApiEndpoint,
|
|
||||||
gqlPostgraphileEndpoint,
|
|
||||||
rpcProviderEndpoint
|
|
||||||
}
|
|
||||||
} = upstream;
|
|
||||||
|
|
||||||
assert(gqlApiEndpoint, 'Missing upstream ethServer.gqlApiEndpoint');
|
|
||||||
assert(gqlEndpoint, 'Missing upstream uniWatcher.gqlEndpoint');
|
|
||||||
assert(gqlSubscriptionEndpoint, 'Missing upstream uniWatcher.gqlSubscriptionEndpoint');
|
|
||||||
|
|
||||||
const cache = await getCache(cacheConfig);
|
|
||||||
|
|
||||||
const ethClient = new EthClient({
|
|
||||||
gqlEndpoint: gqlApiEndpoint,
|
|
||||||
gqlSubscriptionEndpoint: gqlPostgraphileEndpoint,
|
|
||||||
cache
|
|
||||||
});
|
|
||||||
|
|
||||||
const postgraphileClient = new EthClient({
|
|
||||||
gqlEndpoint: gqlPostgraphileEndpoint,
|
|
||||||
cache
|
|
||||||
});
|
|
||||||
|
|
||||||
const uniClient = new UniClient({
|
|
||||||
gqlEndpoint,
|
|
||||||
gqlSubscriptionEndpoint
|
|
||||||
});
|
|
||||||
|
|
||||||
|
const uniClient = new UniClient(uniWatcher);
|
||||||
const erc20Client = new ERC20Client(tokenWatcher);
|
const erc20Client = new ERC20Client(tokenWatcher);
|
||||||
const ethProvider = getCustomProvider(rpcProviderEndpoint);
|
|
||||||
|
|
||||||
|
const jobQueueConfig = config.jobQueue;
|
||||||
assert(jobQueueConfig, 'Missing job queue config');
|
assert(jobQueueConfig, 'Missing job queue config');
|
||||||
|
|
||||||
const { dbConnectionString, maxCompletionLagInSecs } = jobQueueConfig;
|
const { dbConnectionString, maxCompletionLagInSecs } = jobQueueConfig;
|
||||||
@ -133,9 +93,9 @@ export const main = async (): Promise<any> => {
|
|||||||
const jobQueue = new JobQueue({ dbConnectionString, maxCompletionLag: maxCompletionLagInSecs });
|
const jobQueue = new JobQueue({ dbConnectionString, maxCompletionLag: maxCompletionLagInSecs });
|
||||||
await jobQueue.start();
|
await jobQueue.start();
|
||||||
|
|
||||||
const indexer = new Indexer(db, uniClient, erc20Client, ethClient, postgraphileClient, ethProvider, jobQueue, serverConfig.mode);
|
const indexer = new Indexer(db, uniClient, erc20Client, ethClient, postgraphileClient, ethProvider, jobQueue, config.server.mode);
|
||||||
|
|
||||||
const jobRunner = new JobRunner(jobQueueConfig, serverConfig, indexer, jobQueue);
|
const jobRunner = new JobRunner(jobQueueConfig, indexer, jobQueue);
|
||||||
await jobRunner.start();
|
await jobRunner.start();
|
||||||
};
|
};
|
||||||
|
|
||||||
|
@ -14,9 +14,7 @@ import { createServer } from 'http';
|
|||||||
|
|
||||||
import { Client as ERC20Client } from '@vulcanize/erc20-watcher';
|
import { Client as ERC20Client } from '@vulcanize/erc20-watcher';
|
||||||
import { Client as UniClient } from '@vulcanize/uni-watcher';
|
import { Client as UniClient } from '@vulcanize/uni-watcher';
|
||||||
import { EthClient } from '@vulcanize/ipld-eth-client';
|
import { DEFAULT_CONFIG_PATH, getConfig, Config, getCustomProvider, JobQueue, initClients } from '@vulcanize/util';
|
||||||
import { DEFAULT_CONFIG_PATH, getConfig, getCustomProvider, JobQueue } from '@vulcanize/util';
|
|
||||||
import { getCache } from '@vulcanize/cache';
|
|
||||||
|
|
||||||
import typeDefs from './schema';
|
import typeDefs from './schema';
|
||||||
|
|
||||||
@ -39,51 +37,21 @@ export const main = async (): Promise<any> => {
|
|||||||
})
|
})
|
||||||
.argv;
|
.argv;
|
||||||
|
|
||||||
const config = await getConfig(argv.f);
|
const config: Config = await getConfig(argv.f);
|
||||||
|
const { ethClient, postgraphileClient } = await initClients(config);
|
||||||
assert(config.server, 'Missing server config');
|
|
||||||
|
|
||||||
const { host, port, mode } = config.server;
|
const { host, port, mode } = config.server;
|
||||||
|
|
||||||
const { upstream, database: dbConfig, jobQueue: jobQueueConfig } = config;
|
const db = new Database(config.database);
|
||||||
|
|
||||||
assert(dbConfig, 'Missing database config');
|
|
||||||
|
|
||||||
const db = new Database(dbConfig);
|
|
||||||
await db.init();
|
await db.init();
|
||||||
|
|
||||||
assert(upstream, 'Missing upstream config');
|
const { uniWatcher, tokenWatcher, ethServer: { rpcProviderEndpoint } } = config.upstream;
|
||||||
const {
|
|
||||||
ethServer: {
|
|
||||||
gqlApiEndpoint,
|
|
||||||
gqlPostgraphileEndpoint,
|
|
||||||
rpcProviderEndpoint
|
|
||||||
},
|
|
||||||
uniWatcher,
|
|
||||||
tokenWatcher,
|
|
||||||
cache: cacheConfig
|
|
||||||
} = upstream;
|
|
||||||
|
|
||||||
assert(gqlApiEndpoint, 'Missing upstream ethServer.gqlApiEndpoint');
|
|
||||||
assert(gqlPostgraphileEndpoint, 'Missing upstream ethServer.gqlPostgraphileEndpoint');
|
|
||||||
|
|
||||||
const cache = await getCache(cacheConfig);
|
|
||||||
|
|
||||||
const ethClient = new EthClient({
|
|
||||||
gqlEndpoint: gqlApiEndpoint,
|
|
||||||
gqlSubscriptionEndpoint: gqlPostgraphileEndpoint,
|
|
||||||
cache
|
|
||||||
});
|
|
||||||
|
|
||||||
const postgraphileClient = new EthClient({
|
|
||||||
gqlEndpoint: gqlPostgraphileEndpoint,
|
|
||||||
cache
|
|
||||||
});
|
|
||||||
|
|
||||||
const uniClient = new UniClient(uniWatcher);
|
const uniClient = new UniClient(uniWatcher);
|
||||||
const erc20Client = new ERC20Client(tokenWatcher);
|
const erc20Client = new ERC20Client(tokenWatcher);
|
||||||
const ethProvider = getCustomProvider(rpcProviderEndpoint);
|
const ethProvider = getCustomProvider(rpcProviderEndpoint);
|
||||||
|
|
||||||
|
const jobQueueConfig = config.jobQueue;
|
||||||
assert(jobQueueConfig, 'Missing job queue config');
|
assert(jobQueueConfig, 'Missing job queue config');
|
||||||
|
|
||||||
const { dbConnectionString, maxCompletionLagInSecs } = jobQueueConfig;
|
const { dbConnectionString, maxCompletionLagInSecs } = jobQueueConfig;
|
||||||
@ -95,7 +63,7 @@ export const main = async (): Promise<any> => {
|
|||||||
const indexer = new Indexer(db, uniClient, erc20Client, ethClient, postgraphileClient, ethProvider, jobQueue, mode);
|
const indexer = new Indexer(db, uniClient, erc20Client, ethClient, postgraphileClient, ethProvider, jobQueue, mode);
|
||||||
|
|
||||||
const pubSub = new PubSub();
|
const pubSub = new PubSub();
|
||||||
const eventWatcher = new EventWatcher(upstream, ethClient, postgraphileClient, indexer, pubSub, jobQueue);
|
const eventWatcher = new EventWatcher(config.upstream, ethClient, postgraphileClient, indexer, pubSub, jobQueue);
|
||||||
await eventWatcher.start();
|
await eventWatcher.start();
|
||||||
|
|
||||||
const resolvers = process.env.MOCK ? await createMockResolvers() : await createResolvers(indexer, eventWatcher);
|
const resolvers = process.env.MOCK ? await createMockResolvers() : await createResolvers(indexer, eventWatcher);
|
||||||
|
@ -6,7 +6,7 @@ import debug from 'debug';
|
|||||||
import { MoreThan } from 'typeorm';
|
import { MoreThan } from 'typeorm';
|
||||||
import assert from 'assert';
|
import assert from 'assert';
|
||||||
|
|
||||||
import { getConfig, getResetConfig, JobQueue, resetJobs } from '@vulcanize/util';
|
import { getConfig, initClients, resetJobs, JobQueue } from '@vulcanize/util';
|
||||||
|
|
||||||
import { Database } from '../../database';
|
import { Database } from '../../database';
|
||||||
import { Indexer } from '../../indexer';
|
import { Indexer } from '../../indexer';
|
||||||
@ -27,16 +27,15 @@ export const builder = {
|
|||||||
export const handler = async (argv: any): Promise<void> => {
|
export const handler = async (argv: any): Promise<void> => {
|
||||||
const config = await getConfig(argv.configFile);
|
const config = await getConfig(argv.configFile);
|
||||||
await resetJobs(config);
|
await resetJobs(config);
|
||||||
const { jobQueue: jobQueueConfig } = config;
|
const { ethClient, postgraphileClient, ethProvider } = await initClients(config);
|
||||||
const { dbConfig, ethClient, postgraphileClient, ethProvider } = await getResetConfig(config);
|
|
||||||
|
|
||||||
// Initialize database.
|
// Initialize database.
|
||||||
const db = new Database(dbConfig);
|
const db = new Database(config.database);
|
||||||
await db.init();
|
await db.init();
|
||||||
|
|
||||||
assert(jobQueueConfig, 'Missing job queue config');
|
assert(config.jobQueue, 'Missing job queue config');
|
||||||
|
|
||||||
const { dbConnectionString, maxCompletionLagInSecs } = jobQueueConfig;
|
const { dbConnectionString, maxCompletionLagInSecs } = config.jobQueue;
|
||||||
assert(dbConnectionString, 'Missing job queue db connection string');
|
assert(dbConnectionString, 'Missing job queue db connection string');
|
||||||
|
|
||||||
const jobQueue = new JobQueue({ dbConnectionString, maxCompletionLag: maxCompletionLagInSecs });
|
const jobQueue = new JobQueue({ dbConnectionString, maxCompletionLag: maxCompletionLagInSecs });
|
||||||
|
@ -6,7 +6,7 @@ import assert from 'assert';
|
|||||||
import yargs from 'yargs';
|
import yargs from 'yargs';
|
||||||
import 'reflect-metadata';
|
import 'reflect-metadata';
|
||||||
|
|
||||||
import { Config, DEFAULT_CONFIG_PATH, getConfig, getResetConfig, JobQueue } from '@vulcanize/util';
|
import { Config, DEFAULT_CONFIG_PATH, getConfig, initClients, JobQueue } from '@vulcanize/util';
|
||||||
|
|
||||||
import { Database } from '../database';
|
import { Database } from '../database';
|
||||||
import { Indexer } from '../indexer';
|
import { Indexer } from '../indexer';
|
||||||
@ -49,7 +49,7 @@ import { Indexer } from '../indexer';
|
|||||||
|
|
||||||
const config: Config = await getConfig(argv.configFile);
|
const config: Config = await getConfig(argv.configFile);
|
||||||
const { database: dbConfig, jobQueue: jobQueueConfig } = config;
|
const { database: dbConfig, jobQueue: jobQueueConfig } = config;
|
||||||
const { ethClient, postgraphileClient, ethProvider } = await getResetConfig(config);
|
const { ethClient, postgraphileClient, ethProvider } = await initClients(config);
|
||||||
|
|
||||||
assert(dbConfig);
|
assert(dbConfig);
|
||||||
|
|
||||||
|
@ -9,9 +9,7 @@ import { hideBin } from 'yargs/helpers';
|
|||||||
import debug from 'debug';
|
import debug from 'debug';
|
||||||
import { PubSub } from 'apollo-server-express';
|
import { PubSub } from 'apollo-server-express';
|
||||||
|
|
||||||
import { getCache } from '@vulcanize/cache';
|
import { getConfig, Config, fillBlocks, JobQueue, DEFAULT_CONFIG_PATH, initClients } from '@vulcanize/util';
|
||||||
import { EthClient } from '@vulcanize/ipld-eth-client';
|
|
||||||
import { getConfig, fillBlocks, JobQueue, DEFAULT_CONFIG_PATH, getCustomProvider } from '@vulcanize/util';
|
|
||||||
|
|
||||||
import { Database } from './database';
|
import { Database } from './database';
|
||||||
import { Indexer } from './indexer';
|
import { Indexer } from './indexer';
|
||||||
@ -55,39 +53,19 @@ export const main = async (): Promise<any> => {
|
|||||||
}
|
}
|
||||||
}).argv;
|
}).argv;
|
||||||
|
|
||||||
const config = await getConfig(argv.configFile);
|
const config: Config = await getConfig(argv.configFile);
|
||||||
|
const { ethClient, postgraphileClient, ethProvider } = await initClients(config);
|
||||||
|
|
||||||
assert(config.server, 'Missing server config');
|
const db = new Database(config.database);
|
||||||
|
|
||||||
const { upstream, database: dbConfig, jobQueue: jobQueueConfig } = config;
|
|
||||||
|
|
||||||
assert(dbConfig, 'Missing database config');
|
|
||||||
|
|
||||||
const db = new Database(dbConfig);
|
|
||||||
await db.init();
|
await db.init();
|
||||||
|
|
||||||
assert(upstream, 'Missing upstream config');
|
|
||||||
const { ethServer: { gqlApiEndpoint, gqlPostgraphileEndpoint, rpcProviderEndpoint, blockDelayInMilliSecs }, cache: cacheConfig } = upstream;
|
|
||||||
assert(gqlPostgraphileEndpoint, 'Missing upstream ethServer.gqlPostgraphileEndpoint');
|
|
||||||
|
|
||||||
const cache = await getCache(cacheConfig);
|
|
||||||
const ethClient = new EthClient({
|
|
||||||
gqlEndpoint: gqlApiEndpoint,
|
|
||||||
gqlSubscriptionEndpoint: gqlPostgraphileEndpoint,
|
|
||||||
cache
|
|
||||||
});
|
|
||||||
|
|
||||||
const postgraphileClient = new EthClient({
|
|
||||||
gqlEndpoint: gqlPostgraphileEndpoint,
|
|
||||||
cache
|
|
||||||
});
|
|
||||||
|
|
||||||
const ethProvider = getCustomProvider(rpcProviderEndpoint);
|
|
||||||
|
|
||||||
// Note: In-memory pubsub works fine for now, as each watcher is a single process anyway.
|
// Note: In-memory pubsub works fine for now, as each watcher is a single process anyway.
|
||||||
// Later: https://www.apollographql.com/docs/apollo-server/data/subscriptions/#production-pubsub-libraries
|
// Later: https://www.apollographql.com/docs/apollo-server/data/subscriptions/#production-pubsub-libraries
|
||||||
const pubsub = new PubSub();
|
const pubsub = new PubSub();
|
||||||
|
|
||||||
|
const jobQueueConfig = config.jobQueue;
|
||||||
|
assert(jobQueueConfig, 'Missing job queue config');
|
||||||
|
|
||||||
const { dbConnectionString, maxCompletionLagInSecs } = jobQueueConfig;
|
const { dbConnectionString, maxCompletionLagInSecs } = jobQueueConfig;
|
||||||
assert(dbConnectionString, 'Missing job queue db connection string');
|
assert(dbConnectionString, 'Missing job queue db connection string');
|
||||||
|
|
||||||
@ -97,11 +75,9 @@ export const main = async (): Promise<any> => {
|
|||||||
const indexer = new Indexer(db, ethClient, postgraphileClient, ethProvider, jobQueue);
|
const indexer = new Indexer(db, ethClient, postgraphileClient, ethProvider, jobQueue);
|
||||||
await indexer.init();
|
await indexer.init();
|
||||||
|
|
||||||
const eventWatcher = new EventWatcher(upstream, ethClient, postgraphileClient, indexer, pubsub, jobQueue);
|
const eventWatcher = new EventWatcher(config.upstream, ethClient, postgraphileClient, indexer, pubsub, jobQueue);
|
||||||
|
|
||||||
assert(jobQueueConfig, 'Missing job queue config');
|
await fillBlocks(jobQueue, indexer, eventWatcher, config.upstream.ethServer.blockDelayInMilliSecs, argv);
|
||||||
|
|
||||||
await fillBlocks(jobQueue, indexer, eventWatcher, blockDelayInMilliSecs, argv);
|
|
||||||
};
|
};
|
||||||
|
|
||||||
main().catch(err => {
|
main().catch(err => {
|
||||||
|
@ -8,18 +8,16 @@ import yargs from 'yargs';
|
|||||||
import { hideBin } from 'yargs/helpers';
|
import { hideBin } from 'yargs/helpers';
|
||||||
import debug from 'debug';
|
import debug from 'debug';
|
||||||
|
|
||||||
import { getCache } from '@vulcanize/cache';
|
|
||||||
import { EthClient } from '@vulcanize/ipld-eth-client';
|
|
||||||
import {
|
import {
|
||||||
getConfig,
|
getConfig,
|
||||||
|
Config,
|
||||||
JobQueue,
|
JobQueue,
|
||||||
JobRunner as BaseJobRunner,
|
JobRunner as BaseJobRunner,
|
||||||
QUEUE_BLOCK_PROCESSING,
|
QUEUE_BLOCK_PROCESSING,
|
||||||
QUEUE_EVENT_PROCESSING,
|
QUEUE_EVENT_PROCESSING,
|
||||||
JobQueueConfig,
|
JobQueueConfig,
|
||||||
DEFAULT_CONFIG_PATH,
|
DEFAULT_CONFIG_PATH,
|
||||||
getCustomProvider,
|
initClients
|
||||||
ServerConfig
|
|
||||||
} from '@vulcanize/util';
|
} from '@vulcanize/util';
|
||||||
|
|
||||||
import { Indexer } from './indexer';
|
import { Indexer } from './indexer';
|
||||||
@ -32,14 +30,12 @@ export class JobRunner {
|
|||||||
_jobQueue: JobQueue
|
_jobQueue: JobQueue
|
||||||
_baseJobRunner: BaseJobRunner
|
_baseJobRunner: BaseJobRunner
|
||||||
_jobQueueConfig: JobQueueConfig
|
_jobQueueConfig: JobQueueConfig
|
||||||
_serverConfig: ServerConfig
|
|
||||||
|
|
||||||
constructor (jobQueueConfig: JobQueueConfig, serverConfig: ServerConfig, indexer: Indexer, jobQueue: JobQueue) {
|
constructor (jobQueueConfig: JobQueueConfig, indexer: Indexer, jobQueue: JobQueue) {
|
||||||
this._jobQueueConfig = jobQueueConfig;
|
this._jobQueueConfig = jobQueueConfig;
|
||||||
this._indexer = indexer;
|
this._indexer = indexer;
|
||||||
this._jobQueue = jobQueue;
|
this._jobQueue = jobQueue;
|
||||||
this._serverConfig = serverConfig;
|
this._baseJobRunner = new BaseJobRunner(this._jobQueueConfig, this._indexer, this._jobQueue);
|
||||||
this._baseJobRunner = new BaseJobRunner(this._jobQueueConfig, this._serverConfig, this._indexer, this._jobQueue);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
async start (): Promise<void> {
|
async start (): Promise<void> {
|
||||||
@ -71,36 +67,13 @@ export const main = async (): Promise<any> => {
|
|||||||
})
|
})
|
||||||
.argv;
|
.argv;
|
||||||
|
|
||||||
const config = await getConfig(argv.f);
|
const config: Config = await getConfig(argv.f);
|
||||||
|
const { ethClient, postgraphileClient, ethProvider } = await initClients(config);
|
||||||
|
|
||||||
const { upstream, database: dbConfig, jobQueue: jobQueueConfig, server: serverConfig } = config;
|
const db = new Database(config.database);
|
||||||
|
|
||||||
assert(upstream, 'Missing upstream config');
|
|
||||||
assert(dbConfig, 'Missing database config');
|
|
||||||
assert(serverConfig, 'Missing server config');
|
|
||||||
|
|
||||||
const db = new Database(dbConfig);
|
|
||||||
await db.init();
|
await db.init();
|
||||||
|
|
||||||
assert(upstream, 'Missing upstream config');
|
const jobQueueConfig = config.jobQueue;
|
||||||
const { ethServer: { gqlApiEndpoint, gqlPostgraphileEndpoint, rpcProviderEndpoint }, cache: cacheConfig } = upstream;
|
|
||||||
assert(gqlApiEndpoint, 'Missing upstream ethServer.gqlApiEndpoint');
|
|
||||||
assert(gqlPostgraphileEndpoint, 'Missing upstream ethServer.gqlPostgraphileEndpoint');
|
|
||||||
|
|
||||||
const cache = await getCache(cacheConfig);
|
|
||||||
const ethClient = new EthClient({
|
|
||||||
gqlEndpoint: gqlApiEndpoint,
|
|
||||||
gqlSubscriptionEndpoint: gqlPostgraphileEndpoint,
|
|
||||||
cache
|
|
||||||
});
|
|
||||||
|
|
||||||
const postgraphileClient = new EthClient({
|
|
||||||
gqlEndpoint: gqlPostgraphileEndpoint,
|
|
||||||
cache
|
|
||||||
});
|
|
||||||
|
|
||||||
const ethProvider = getCustomProvider(rpcProviderEndpoint);
|
|
||||||
|
|
||||||
assert(jobQueueConfig, 'Missing job queue config');
|
assert(jobQueueConfig, 'Missing job queue config');
|
||||||
|
|
||||||
const { dbConnectionString, maxCompletionLagInSecs } = jobQueueConfig;
|
const { dbConnectionString, maxCompletionLagInSecs } = jobQueueConfig;
|
||||||
@ -112,7 +85,7 @@ export const main = async (): Promise<any> => {
|
|||||||
const indexer = new Indexer(db, ethClient, postgraphileClient, ethProvider, jobQueue);
|
const indexer = new Indexer(db, ethClient, postgraphileClient, ethProvider, jobQueue);
|
||||||
await indexer.init();
|
await indexer.init();
|
||||||
|
|
||||||
const jobRunner = new JobRunner(jobQueueConfig, serverConfig, indexer, jobQueue);
|
const jobRunner = new JobRunner(jobQueueConfig, indexer, jobQueue);
|
||||||
await jobRunner.start();
|
await jobRunner.start();
|
||||||
};
|
};
|
||||||
|
|
||||||
|
@ -12,9 +12,7 @@ import debug from 'debug';
|
|||||||
import 'graphql-import-node';
|
import 'graphql-import-node';
|
||||||
import { createServer } from 'http';
|
import { createServer } from 'http';
|
||||||
|
|
||||||
import { getCache } from '@vulcanize/cache';
|
import { DEFAULT_CONFIG_PATH, getConfig, Config, JobQueue, initClients } from '@vulcanize/util';
|
||||||
import { EthClient } from '@vulcanize/ipld-eth-client';
|
|
||||||
import { DEFAULT_CONFIG_PATH, getConfig, getCustomProvider, JobQueue } from '@vulcanize/util';
|
|
||||||
|
|
||||||
import typeDefs from './schema';
|
import typeDefs from './schema';
|
||||||
|
|
||||||
@ -37,42 +35,19 @@ export const main = async (): Promise<any> => {
|
|||||||
})
|
})
|
||||||
.argv;
|
.argv;
|
||||||
|
|
||||||
const config = await getConfig(argv.f);
|
const config: Config = await getConfig(argv.f);
|
||||||
|
const { ethClient, postgraphileClient, ethProvider } = await initClients(config);
|
||||||
assert(config.server, 'Missing server config');
|
|
||||||
|
|
||||||
const { host, port } = config.server;
|
const { host, port } = config.server;
|
||||||
|
|
||||||
const { upstream, database: dbConfig, jobQueue: jobQueueConfig } = config;
|
const db = new Database(config.database);
|
||||||
|
|
||||||
assert(dbConfig, 'Missing database config');
|
|
||||||
|
|
||||||
const db = new Database(dbConfig);
|
|
||||||
await db.init();
|
await db.init();
|
||||||
|
|
||||||
assert(upstream, 'Missing upstream config');
|
|
||||||
const { ethServer: { gqlApiEndpoint, gqlPostgraphileEndpoint, rpcProviderEndpoint }, cache: cacheConfig } = upstream;
|
|
||||||
assert(gqlApiEndpoint, 'Missing upstream ethServer.gqlApiEndpoint');
|
|
||||||
assert(gqlPostgraphileEndpoint, 'Missing upstream ethServer.gqlPostgraphileEndpoint');
|
|
||||||
|
|
||||||
const cache = await getCache(cacheConfig);
|
|
||||||
const ethClient = new EthClient({
|
|
||||||
gqlEndpoint: gqlApiEndpoint,
|
|
||||||
gqlSubscriptionEndpoint: gqlPostgraphileEndpoint,
|
|
||||||
cache
|
|
||||||
});
|
|
||||||
|
|
||||||
const postgraphileClient = new EthClient({
|
|
||||||
gqlEndpoint: gqlPostgraphileEndpoint,
|
|
||||||
cache
|
|
||||||
});
|
|
||||||
|
|
||||||
const ethProvider = getCustomProvider(rpcProviderEndpoint);
|
|
||||||
|
|
||||||
// Note: In-memory pubsub works fine for now, as each watcher is a single process anyway.
|
// Note: In-memory pubsub works fine for now, as each watcher is a single process anyway.
|
||||||
// Later: https://www.apollographql.com/docs/apollo-server/data/subscriptions/#production-pubsub-libraries
|
// Later: https://www.apollographql.com/docs/apollo-server/data/subscriptions/#production-pubsub-libraries
|
||||||
const pubsub = new PubSub();
|
const pubsub = new PubSub();
|
||||||
|
|
||||||
|
const jobQueueConfig = config.jobQueue;
|
||||||
assert(jobQueueConfig, 'Missing job queue config');
|
assert(jobQueueConfig, 'Missing job queue config');
|
||||||
|
|
||||||
const { dbConnectionString, maxCompletionLagInSecs } = jobQueueConfig;
|
const { dbConnectionString, maxCompletionLagInSecs } = jobQueueConfig;
|
||||||
@ -84,7 +59,7 @@ export const main = async (): Promise<any> => {
|
|||||||
const indexer = new Indexer(db, ethClient, postgraphileClient, ethProvider, jobQueue);
|
const indexer = new Indexer(db, ethClient, postgraphileClient, ethProvider, jobQueue);
|
||||||
await indexer.init();
|
await indexer.init();
|
||||||
|
|
||||||
const eventWatcher = new EventWatcher(upstream, ethClient, postgraphileClient, indexer, pubsub, jobQueue);
|
const eventWatcher = new EventWatcher(config.upstream, ethClient, postgraphileClient, indexer, pubsub, jobQueue);
|
||||||
await eventWatcher.start();
|
await eventWatcher.start();
|
||||||
|
|
||||||
const resolvers = process.env.MOCK ? await createMockResolvers() : await createResolvers(indexer, eventWatcher);
|
const resolvers = process.env.MOCK ? await createMockResolvers() : await createResolvers(indexer, eventWatcher);
|
||||||
|
@ -12,5 +12,5 @@ export * from './src/events';
|
|||||||
export * from './src/types';
|
export * from './src/types';
|
||||||
export * from './src/indexer';
|
export * from './src/indexer';
|
||||||
export * from './src/job-runner';
|
export * from './src/job-runner';
|
||||||
|
export * from './src/ipld-helper';
|
||||||
export * from './src/graph-decimal';
|
export * from './src/graph-decimal';
|
||||||
export * from './src/ipldHelper';
|
|
||||||
|
@ -9,9 +9,9 @@ import toml from 'toml';
|
|||||||
import debug from 'debug';
|
import debug from 'debug';
|
||||||
import { ConnectionOptions } from 'typeorm';
|
import { ConnectionOptions } from 'typeorm';
|
||||||
|
|
||||||
import { BaseProvider } from '@ethersproject/providers';
|
|
||||||
import { Config as CacheConfig, getCache } from '@vulcanize/cache';
|
import { Config as CacheConfig, getCache } from '@vulcanize/cache';
|
||||||
import { EthClient } from '@vulcanize/ipld-eth-client';
|
import { EthClient } from '@vulcanize/ipld-eth-client';
|
||||||
|
import { BaseProvider } from '@ethersproject/providers';
|
||||||
|
|
||||||
import { getCustomProvider } from './misc';
|
import { getCustomProvider } from './misc';
|
||||||
|
|
||||||
@ -73,10 +73,7 @@ export const getConfig = async (configFile: string): Promise<Config> => {
|
|||||||
return config;
|
return config;
|
||||||
};
|
};
|
||||||
|
|
||||||
export const getResetConfig = async (config: Config): Promise<{
|
export const initClients = async (config: Config): Promise<{
|
||||||
dbConfig: ConnectionOptions,
|
|
||||||
serverConfig: ServerConfig,
|
|
||||||
upstreamConfig: UpstreamConfig,
|
|
||||||
ethClient: EthClient,
|
ethClient: EthClient,
|
||||||
postgraphileClient: EthClient,
|
postgraphileClient: EthClient,
|
||||||
ethProvider: BaseProvider
|
ethProvider: BaseProvider
|
||||||
@ -85,9 +82,10 @@ export const getResetConfig = async (config: Config): Promise<{
|
|||||||
|
|
||||||
assert(serverConfig, 'Missing server config');
|
assert(serverConfig, 'Missing server config');
|
||||||
assert(dbConfig, 'Missing database config');
|
assert(dbConfig, 'Missing database config');
|
||||||
|
|
||||||
assert(upstreamConfig, 'Missing upstream config');
|
assert(upstreamConfig, 'Missing upstream config');
|
||||||
|
|
||||||
const { ethServer: { gqlApiEndpoint, gqlPostgraphileEndpoint, rpcProviderEndpoint }, cache: cacheConfig } = upstreamConfig;
|
const { ethServer: { gqlApiEndpoint, gqlPostgraphileEndpoint, rpcProviderEndpoint }, cache: cacheConfig } = upstreamConfig;
|
||||||
|
|
||||||
assert(gqlApiEndpoint, 'Missing upstream ethServer.gqlApiEndpoint');
|
assert(gqlApiEndpoint, 'Missing upstream ethServer.gqlApiEndpoint');
|
||||||
assert(gqlPostgraphileEndpoint, 'Missing upstream ethServer.gqlPostgraphileEndpoint');
|
assert(gqlPostgraphileEndpoint, 'Missing upstream ethServer.gqlPostgraphileEndpoint');
|
||||||
assert(rpcProviderEndpoint, 'Missing upstream ethServer.rpcProviderEndpoint');
|
assert(rpcProviderEndpoint, 'Missing upstream ethServer.rpcProviderEndpoint');
|
||||||
@ -108,9 +106,6 @@ export const getResetConfig = async (config: Config): Promise<{
|
|||||||
const ethProvider = getCustomProvider(rpcProviderEndpoint);
|
const ethProvider = getCustomProvider(rpcProviderEndpoint);
|
||||||
|
|
||||||
return {
|
return {
|
||||||
dbConfig,
|
|
||||||
serverConfig,
|
|
||||||
upstreamConfig,
|
|
||||||
ethClient,
|
ethClient,
|
||||||
postgraphileClient,
|
postgraphileClient,
|
||||||
ethProvider
|
ethProvider
|
||||||
|
@ -6,7 +6,7 @@ import assert from 'assert';
|
|||||||
import debug from 'debug';
|
import debug from 'debug';
|
||||||
import { In } from 'typeorm';
|
import { In } from 'typeorm';
|
||||||
|
|
||||||
import { JobQueueConfig, ServerConfig } from './config';
|
import { JobQueueConfig } from './config';
|
||||||
import {
|
import {
|
||||||
JOB_KIND_INDEX,
|
JOB_KIND_INDEX,
|
||||||
JOB_KIND_PRUNE,
|
JOB_KIND_PRUNE,
|
||||||
@ -32,13 +32,11 @@ export class JobRunner {
|
|||||||
_jobQueue: JobQueue
|
_jobQueue: JobQueue
|
||||||
_jobQueueConfig: JobQueueConfig
|
_jobQueueConfig: JobQueueConfig
|
||||||
_blockProcessStartTime?: Date
|
_blockProcessStartTime?: Date
|
||||||
_serverConfig: ServerConfig
|
|
||||||
|
|
||||||
constructor (jobQueueConfig: JobQueueConfig, serverConfig: ServerConfig, indexer: IndexerInterface, jobQueue: JobQueue) {
|
constructor (jobQueueConfig: JobQueueConfig, indexer: IndexerInterface, jobQueue: JobQueue) {
|
||||||
this._indexer = indexer;
|
this._indexer = indexer;
|
||||||
this._jobQueue = jobQueue;
|
this._jobQueue = jobQueue;
|
||||||
this._jobQueueConfig = jobQueueConfig;
|
this._jobQueueConfig = jobQueueConfig;
|
||||||
this._serverConfig = serverConfig;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
async processBlock (job: any): Promise<void> {
|
async processBlock (job: any): Promise<void> {
|
||||||
|
Loading…
Reference in New Issue
Block a user