Server backend for indexed ETH IPLD objects
Go to file
2019-02-08 13:36:30 -06:00
bin update supfile and deploy script for production 2018-09-24 10:24:06 -05:00
cmd (VDB-280) Consume vat contract storage diffs 2019-02-08 13:36:30 -06:00
db (VDB-280) Consume vat contract storage diffs 2019-02-08 13:36:30 -06:00
dockerfiles/rinkeby dockerfiles/rinkeby: added named volume for pg container 2018-06-25 23:34:49 +03:00
environments (VDB-298) Consume Pit contract storage diffs 2019-02-06 10:31:46 -06:00
integration_test change omniwatcher database test helpers so that it doesn't interfere with downstream maker tests 2019-01-23 14:55:20 -06:00
libraries/shared * update transformer to able to recheck headers (#4) 2019-02-08 10:35:46 -06:00
pkg (VDB-280) Consume vat contract storage diffs 2019-02-08 13:36:30 -06:00
postgraphile update license to AGPL-3.0 2019-01-24 14:41:30 -06:00
scripts Delete old migration script 2019-01-25 11:44:05 +01:00
test_config * update transformer to able to recheck headers (#4) 2019-02-08 10:35:46 -06:00
utils merge public vulcanizedb/master 2019-01-23 00:37:26 -06:00
vendor merge public vulcanizedb/master 2019-01-23 00:37:26 -06:00
.gitignore use logrus for logging 2018-11-21 10:14:11 -06:00
.private_blockchain_password Add integration test 2017-10-24 15:36:50 -05:00
.travis.yml Merge branch 'staging' into VDB-281-Migrations-remake 2019-02-01 17:29:56 +01:00
Gopkg.lock merge public vulcanizedb/master 2019-01-23 00:37:26 -06:00
Gopkg.toml Bump geth to 1.8.21 (#137) 2019-01-16 10:54:01 +01:00
LICENSE Edits to address PR issues; change license from apache to AGPL; and work 2018-11-15 12:32:52 -06:00
main.go use logrus for logging 2018-11-21 10:14:11 -06:00
Makefile Merge branch 'staging' into VDB-281-Migrations-remake 2019-02-01 17:29:56 +01:00
README.md Merge branch 'staging' into VDB-281-Migrations-remake 2019-02-01 17:29:56 +01:00
Supfile * update transformer to able to recheck headers (#4) 2019-02-08 10:35:46 -06:00

Vulcanize DB

Join the chat at https://gitter.im/vulcanizeio/VulcanizeDB

Build Status

About

Vulcanize DB is a set of tools that make it easier for developers to write application-specific indexes and caches for dapps built on Ethereum.

Dependencies

Project Setup

Using Vulcanize for the first time requires several steps be done in order to allow use of the software. The following instructions will offer a guide through the steps of the process:

  1. Fetching the project
  2. Installing dependencies
  3. Configuring shell environment
  4. Database setup
  5. Configuring synced Ethereum node integration
  6. Data syncing

Installation

In order to fetch the project codebase for local use or modification, install it to your GOPATH via:

go get github.com/vulcanize/vulcanizedb go get gopkg.in/DataDog/dd-trace-go.v1/ddtrace

Once fetched, dependencies can be installed via go get or (the preferred method) at specific versions via golang/dep, the prototype golang pakcage manager. Installation instructions are here.

In order to install packages with dep, ensure you are in the project directory now within your GOPATH (default location is ~/go/src/github.com/vulcanize/vulcanizedb/) and run:

dep ensure

After dep finishes, dependencies should be installed within your GOPATH at the versions specified in Gopkg.toml.

Lastly, ensure that GOPATH is defined in your shell. If necessary, GOPATH can be set in ~/.bashrc or ~/.bash_profile, depending upon your system. It can be additionally helpful to add $GOPATH/bin to your shell's $PATH.

Setting up the Database

  1. Install Postgres

  2. Create a superuser for yourself and make sure psql --list works without prompting for a password.

  3. createdb vulcanize_public

  4. cd $GOPATH/src/github.com/vulcanize/vulcanizedb

  5. Run the migrations: make migrate HOST_NAME=localhost NAME=vulcanize_public PORT=5432

    • To rollback a single step: make rollback NAME=vulcanize_public
    • To rollback to a certain migration: make rollback_to MIGRATION=n NAME=vulcanize_public
    • To see status of migrations: make migration_status NAME=vulcanize_public
    • See below for configuring additional environments

Create a migration file

  1. make new_migration NAME=add_columnA_to_table1
    • This will create a new timestamped migration file in db/migrations
  2. Write the migration code in the created file, under the respective goose pragma
    • Goose automatically runs each migration in a transaction; don't add BEGIN and COMMIT statements.

Configuration

  • To use a local Ethereum node, copy environments/public.toml.example to environments/public.toml and update the ipcPath and levelDbPath.

    • ipcPath should match the local node's IPC filepath:

      • For Geth:

        • The IPC file is called geth.ipc.
        • The geth IPC file path is printed to the console when you start geth.
        • The default location is:
          • Mac: <full home path>/Library/Ethereum
          • Linux: <full home path>/ethereum/geth.ipc
      • For Parity:

        • The IPC file is called jsonrpc.ipc.
        • The default location is:
          • Mac: <full home path>/Library/Application\ Support/io.parity.ethereum/
          • Linux: <full home path>/local/share/io.parity.ethereum/
    • levelDbPath should match Geth's chaindata directory path.

      • The geth LevelDB chaindata path is printed to the console when you start geth.
      • The default location is:
        • Mac: <full home path>/Library/Ethereum/geth/chaindata
        • Linux: <full home path>/ethereum/geth/chaindata
      • levelDbPath is irrelevant (and coldImport is currently unavailable) if only running parity.
  • See environments/infura.toml to configure commands to run against infura, if a local node is unavailable.

  • Copy environments/local.toml.example to environments/local.toml to configure commands to run against a local node such as Ganache or ganache-cli.

Start syncing with postgres

Syncs VulcanizeDB with the configured Ethereum node, populating blocks, transactions, receipts, and logs. This command is useful when you want to maintain a broad cache of what's happening on the blockchain.

  1. Start Ethereum node (if fast syncing your Ethereum node, wait for initial sync to finish)
  2. In a separate terminal start VulcanizeDB:
    • ./vulcanizedb sync --config <config.toml> --starting-block-number <block-number>

Alternatively, sync from Geth's underlying LevelDB

Sync VulcanizeDB from the LevelDB underlying a Geth node.

  1. Assure node is not running, and that it has synced to the desired block height.
  2. Start vulcanize_db
    • ./vulcanizedb coldImport --config <config.toml>
  3. Optional flags:
    • --starting-block-number <block number>/-s <block number>: block number to start syncing from
    • --ending-block-number <block number>/-e <block number>: block number to sync to
    • --all/-a: sync all missing blocks

Alternatively, sync in "light" mode

Syncs VulcanizeDB with the configured Ethereum node, populating only block headers. This command is useful when you want a minimal baseline from which to track targeted data on the blockchain (e.g. individual smart contract storage values).

  1. Start Ethereum node
  2. In a separate terminal start VulcanizeDB:
    • ./vulcanizedb lightSync --config <config.toml> --starting-block-number <block-number>

Continuously sync Maker event logs from light sync

Continuously syncs Maker event logs from the configured Ethereum node based on the populated block headers. This includes logs related to auctions, multi-collateral dai, and price feeds. This command requires that the lightSync process is also being run so as to be able to sync in real time.

  1. Start Ethereum node (or plan to configure the commands to point to a remote IPC path).
  2. In a separate terminal run the lightSync command (see above).
  3. In another terminal window run the continuousLogSync command:
  • ./vulcanizedb continuousLogSync --config <config.toml>
  • An option --transformers flag may be passed to the command to specific which transformers to execute, this will default to all transformers if the flag is not passed.
    • ./vulcanizedb continuousLogSync --config environments/private.toml --transformers="priceFeed"
    • see the buildTransformerInitializerMap method in cmd/continuousLogSync.go for available transformers

Backfill Maker event logs from light sync

Backfills Maker event logs from the configured Ethereum node based on the populated block headers. This includes logs related to auctions, multi-collateral dai, and price feeds. This command requires that a light sync (see command above) has previously been run.

Since auction/mcd contracts have not yet been deployed, this command will need to be run a local blockchain at the moment. As such, a new environment file will need to be added. See environments/local.toml.example.

  1. Start Ethereum node
  2. In a separate terminal run the backfill command:
  • ./vulcanizedb backfillMakerLogs --config <config.toml>

Start full environment in docker by single command

Geth Rinkeby

make command description
rinkeby_env_up start geth, postgres and rolling migrations, after migrations done starting vulcanizedb container
rinkeby_env_deploy build and run vulcanizedb container in rinkeby environment
rinkeby_env_migrate build and run rinkeby env migrations
rinkeby_env_down stop and remove all rinkeby env containers

Success run of the VulcanizeDB container require full geth state sync, attach to geth console and check sync state:

$ docker exec -it rinkeby_vulcanizedb_geth geth --rinkeby attach
...
> eth.syncing
false

If you have full rinkeby chaindata you can move it to rinkeby_vulcanizedb_geth_data docker volume to skip long wait of sync.

Running the Tests

  • make test will run the unit tests and skip the integration tests
  • make integrationtest will run the just the integration tests
  • Note: requires Ganache chain setup and seeded with flip-kick.js and frob.js (in that order)

Deploying

  1. you will need to make sure you have ssh agent running and your ssh key added to it. instructions here
  2. go get -u github.com/pressly/sup/cmd/sup
  3. sup staging deploy

omniWatcher

These commands require a pre-synced (full or light) vulcanizeDB (see above sections)

To watch all events of a contract using a light synced vDB:
- Execute ./vulcanizedb omniWatcher --config <path to config.toml> --contract-address <contract address>

Or if you are using a full synced vDB, change the mode to full:
- Execute ./vulcanizedb omniWatcher --mode full --config <path to config.toml> --contract-address <contract address>

To watch contracts on a network other than mainnet, use the network flag:
- Execute ./vulcanizedb omniWatcher --config <path to config.toml> --contract-address <contract address> --network <ropsten, kovan, or rinkeby>

To watch events starting at a certain block use the starting block flag: - Execute ./vulcanizedb omniWatcher --config <path to config.toml> --contract-address <contract address> --starting-block-number <#>

To watch only specified events use the events flag:
- Execute ./vulcanizedb omniWatcher --config <path to config.toml> --contract-address <contract address> --events <EventName1> --events <EventName2>

To watch events and poll the specified methods with any addresses and hashes emitted by the watched events utilize the methods flag:
- Execute ./vulcanizedb omniWatcher --config <path to config.toml> --contract-address <contract address> --methods <methodName1> --methods <methodName2>

To watch specified events and poll the specified method with any addresses and hashes emitted by the watched events:
- Execute ./vulcanizedb omniWatcher --config <path to config.toml> --contract-address <contract address> --events <EventName1> --events <EventName2> --methods <methodName>

To turn on method piping so that values returned from previous method calls are cached and used as arguments in subsequent method calls:
- Execute ./vulcanizedb omniWatcher --config <path to config.toml> --piping true --contract-address <contract address> --events <EventName1> --events <EventName2> --methods <methodName>

To watch all types of events of the contract but only persist the ones that emit one of the filtered-for argument values:
- Execute ./vulcanizedb omniWatcher --config <path to config.toml> --contract-address <contract address> --event-args <arg1> --event-args <arg2>

To watch all events of the contract but only poll the specified method with specified argument values (if they are emitted from the watched events):
- Execute ./vulcanizedb omniWatcher --config <path to config.toml> --contract-address <contract address> --methods <methodName> --method-args <arg1> --method-args <arg2>