Compare commits
89 Commits
cerc_refac
...
v5
Author | SHA1 | Date | |
---|---|---|---|
89d08ac1b9 | |||
d24142a301 | |||
9e0b69d23c | |||
825a0bc235 | |||
e90bc38fdb | |||
ccfc2dbc84 | |||
fdd56e9803 | |||
68a347e38d | |||
097804b1e9 | |||
925abd314b | |||
|
aaa4459655 | ||
|
8f360133ed | ||
|
f6da2ce571 | ||
|
245e6d07b2 | ||
|
402d2790e0 | ||
29c5d04159 | |||
|
9196c449f0 | ||
|
146282fb7f | ||
323ed2da00 | |||
5b92e02ebd | |||
dcd911a8ff | |||
59d01ad30b | |||
d2d8020856 | |||
9618c0710e | |||
8779bb2b86 | |||
1b922dbff3 | |||
66cd1d9e69 | |||
af9910d381 | |||
e4e1b0ea1f | |||
050c1ae3df | |||
2afa6e9fa5 | |||
3a475b4de4 | |||
bf7b4eb627 | |||
|
167cfbfb20 | ||
|
fcb93d17ed | ||
|
1bb4fe04f0 | ||
|
10c84b589f | ||
|
527ff11328 | ||
a91f44773c | |||
0595f3dc15 | |||
|
9f07d4f4d4 | ||
|
c9edf6c832 | ||
|
67dc84205a | ||
|
22dcf5c72e | ||
|
be28c2ab79 | ||
|
42803af51a | ||
|
afc47af045 | ||
|
2695d9e353 | ||
71dede2031 | |||
|
df352ffd1a | ||
|
dd35277c86 | ||
|
92cc8fbea3 | ||
|
05db3c697f | ||
|
a710db0284 | ||
|
5e153c601f | ||
|
802cfe7180 | ||
|
b06b4f2cfb | ||
|
3fd1638ff6 | ||
|
1e5cbfd184 | ||
|
40b1709c2c | ||
|
92a9f5856b | ||
|
9f060ff0bf | ||
|
a8440e4ded | ||
|
c402e5c285 | ||
|
85bc243896 | ||
|
f67f03481b | ||
|
27e923f70d | ||
|
73a66dae8b | ||
|
29e2bd4e7b | ||
986ce1ead8 | |||
268a282eac | |||
53461a0996 | |||
|
26d970ed2f | ||
|
c35cda7b5e | ||
|
f3c58e39ca | ||
|
2165b316fa | ||
|
713f6a9208 | ||
|
241bb281eb | ||
|
bb57b4a033 | ||
|
a9755c6ecc | ||
|
7f8247cb4f | ||
|
701f9c2729 | ||
|
475ead282b | ||
|
42f46fc397 | ||
|
2bf5c82150 | ||
|
0c62ccc552 | ||
|
13f0ff3933 | ||
|
0856168b92 | ||
|
a8395d1413 |
23
.gitea/workflows/on-pr.yaml
Normal file
23
.gitea/workflows/on-pr.yaml
Normal file
@ -0,0 +1,23 @@
|
||||
name: Basic test
|
||||
on: [pull_request]
|
||||
|
||||
jobs:
|
||||
basic-test:
|
||||
name: Build and sanity check
|
||||
runs-on: ubuntu-latest
|
||||
steps:
|
||||
- uses: actions/checkout@v4
|
||||
- name: Build Docker images
|
||||
run: docker compose build
|
||||
- name: Run Docker containers
|
||||
run: docker compose up -d
|
||||
- name: Check migration version
|
||||
timeout-minutes: 1
|
||||
run: |
|
||||
MIGRATION_VERSION=$(ls db/migrations/*.sql | wc -l)
|
||||
while
|
||||
version=$(docker compose run --rm migrations version 2>&1 | tail -1 | awk '{print $(NF)}')
|
||||
[[ $version != $MIGRATION_VERSION ]]; do
|
||||
echo "Incorrect version: $version"
|
||||
echo "Retrying..."
|
||||
done
|
26
.gitea/workflows/publish.yaml
Normal file
26
.gitea/workflows/publish.yaml
Normal file
@ -0,0 +1,26 @@
|
||||
name: Publish Docker image
|
||||
on:
|
||||
release:
|
||||
types: [published, edited]
|
||||
jobs:
|
||||
build:
|
||||
name: Build and publish image
|
||||
runs-on: ubuntu-latest
|
||||
steps:
|
||||
- uses: actions/checkout@v3
|
||||
- id: vars
|
||||
name: Output SHA and version tag
|
||||
run: |
|
||||
echo "sha=${GITHUB_SHA:0:7}" >> $GITHUB_OUTPUT
|
||||
echo "tag=${GITHUB_REF#refs/tags/}" >> $GITHUB_OUTPUT
|
||||
- name: Build and tag image
|
||||
run: |
|
||||
docker build . \
|
||||
-t cerc-io/ipld-eth-db \
|
||||
-t git.vdb.to/cerc-io/ipld-eth-db/ipld-eth-db:${{steps.vars.outputs.sha}} \
|
||||
-t git.vdb.to/cerc-io/ipld-eth-db/ipld-eth-db:${{steps.vars.outputs.tag}}
|
||||
- name: Push image tags
|
||||
run: |
|
||||
echo ${{ secrets.CICD_PUBLISH_TOKEN }} | docker login https://git.vdb.to -u cerccicd --password-stdin
|
||||
docker push git.vdb.to/cerc-io/ipld-eth-db/ipld-eth-db:${{steps.vars.outputs.sha}}
|
||||
docker push git.vdb.to/cerc-io/ipld-eth-db/ipld-eth-db:${{steps.vars.outputs.tag}}
|
78
.github/workflows/on-pr.yaml
vendored
78
.github/workflows/on-pr.yaml
vendored
@ -1,78 +0,0 @@
|
||||
name: Docker Build
|
||||
|
||||
on: [pull_request]
|
||||
|
||||
jobs:
|
||||
build:
|
||||
name: Run docker build
|
||||
runs-on: ubuntu-latest
|
||||
steps:
|
||||
- uses: actions/checkout@v2
|
||||
- name: Run docker build
|
||||
run: make docker-build
|
||||
test:
|
||||
# Add a dummy job to avoid failing GitHub CI checks.
|
||||
# Other checks to be added later.
|
||||
name: Dummy job
|
||||
runs-on: ubuntu-latest
|
||||
steps:
|
||||
- name: Run dummy job
|
||||
run: echo "Empty dummy job"
|
||||
# concise_migration_diff:
|
||||
# name: Verify concise migration and generated schema
|
||||
# runs-on: ubuntu-latest
|
||||
# steps:
|
||||
# - uses: actions/checkout@v2
|
||||
# - name: Run docker concise migration build
|
||||
# run: make docker-concise-migration-build
|
||||
# - name: Run database
|
||||
# run: docker-compose -f docker-compose.test.yml up -d test-db
|
||||
# - name: Test concise migration
|
||||
# run: |
|
||||
# sleep 10
|
||||
# docker run --rm --network host -e DATABASE_USER=vdbm -e DATABASE_PASSWORD=password \
|
||||
# -e DATABASE_HOSTNAME=127.0.0.1 -e DATABASE_PORT=8066 -e DATABASE_NAME=vulcanize_testing \
|
||||
# vulcanize/concise-migration-build
|
||||
# - name: Verify schema is latest
|
||||
# run: |
|
||||
# PGPASSWORD="password" pg_dump -h localhost -p 8066 -U vdbm vulcanize_testing --no-owner --schema-only > ./db/migration_schema.sql
|
||||
# ./scripts/check_diff.sh ./db/migration_schema.sql db/schema.sql
|
||||
|
||||
# incremental_migration_diff:
|
||||
# name: Compare conscise migration schema with incremental migration.
|
||||
# runs-on: ubuntu-latest
|
||||
# steps:
|
||||
# - uses: actions/checkout@v2
|
||||
# - name: Run database
|
||||
# run: docker-compose -f docker-compose.test.yml up -d test-db statediff-migrations
|
||||
# - name: Test incremental migration
|
||||
# run: |
|
||||
# sleep 10
|
||||
# docker run --rm --network host -e DATABASE_USER=vdbm -e DATABASE_PASSWORD=password \
|
||||
# -e DATABASE_HOSTNAME=127.0.0.1 -e DATABASE_PORT=8066 -e DATABASE_NAME=vulcanize_testing \
|
||||
# vulcanize/statediff-migrations:v0.9.0
|
||||
# - name: Verify schema is latest
|
||||
# run: |
|
||||
# PGPASSWORD="password" pg_dump -h localhost -p 8066 -U vdbm vulcanize_testing --no-owner --schema-only > ./db/migration_schema.sql
|
||||
# ./scripts/check_diff.sh db/schema.sql ./db/migration_schema.sql
|
||||
|
||||
# migration:
|
||||
# name: Compare up and down migration
|
||||
# env:
|
||||
# GOPATH: /tmp/go
|
||||
# strategy:
|
||||
# matrix:
|
||||
# go-version: [ 1.16.x ]
|
||||
# os: [ ubuntu-latest ]
|
||||
# runs-on: ${{ matrix.os }}
|
||||
# steps:
|
||||
# - name: Create GOPATH
|
||||
# run: mkdir -p /tmp/go
|
||||
# - name: Install Go
|
||||
# uses: actions/setup-go@v2
|
||||
# with:
|
||||
# go-version: ${{ matrix.go-version }}
|
||||
# - uses: actions/checkout@v2
|
||||
# - name: Test migration
|
||||
# run: |
|
||||
# timeout 5m make test-migrations
|
49
.github/workflows/publish.yaml
vendored
49
.github/workflows/publish.yaml
vendored
@ -1,49 +0,0 @@
|
||||
name: Publish Docker image
|
||||
on:
|
||||
release:
|
||||
types: [published, edited]
|
||||
jobs:
|
||||
build:
|
||||
name: Run docker build
|
||||
runs-on: ubuntu-latest
|
||||
steps:
|
||||
- uses: actions/checkout@v2
|
||||
- name: Get the version
|
||||
id: vars
|
||||
run: |
|
||||
echo ::set-output name=sha::$(echo ${GITHUB_SHA:0:7})
|
||||
echo ::set-output name=tag::$(echo ${GITHUB_REF#refs/tags/})
|
||||
- name: Run docker build
|
||||
run: make docker-build
|
||||
- name: Tag docker image
|
||||
run: docker tag cerc-io/ipld-eth-db git.vdb.to/cerc-io/ipld-eth-db/ipld-eth-db:${{steps.vars.outputs.sha}}
|
||||
- name: Tag docker image
|
||||
run: docker tag git.vdb.to/cerc-io/ipld-eth-db/ipld-eth-db:${{steps.vars.outputs.sha}} git.vdb.to/cerc-io/ipld-eth-db/ipld-eth-db:${{steps.vars.outputs.tag}}
|
||||
- name: Docker Login
|
||||
run: echo ${{ secrets.GITEA_TOKEN }} | docker login https://git.vdb.to -u cerccicd --password-stdin
|
||||
- name: Docker Push
|
||||
run: docker push git.vdb.to/cerc-io/ipld-eth-db/ipld-eth-db:${{steps.vars.outputs.sha}}
|
||||
- name: Docker Push with TAG to git.vdb.to
|
||||
run: docker push git.vdb.to/cerc-io/ipld-eth-db/ipld-eth-db:${{steps.vars.outputs.tag}}
|
||||
#push_to_registries:
|
||||
# name: Push Docker image to Docker Hub
|
||||
# runs-on: ubuntu-latest
|
||||
# needs: build
|
||||
# steps:
|
||||
# - name: Get the version
|
||||
# id: vars
|
||||
# run: |
|
||||
# echo ::set-output name=sha::$(echo ${GITHUB_SHA:0:7})
|
||||
# echo ::set-output name=tag::$(echo ${GITHUB_REF#refs/tags/})
|
||||
# - name: Docker Login to Github Registry
|
||||
# run: echo ${{ secrets.GITEA_TOKEN }} | docker login https://git.vdb.to -u cerccicd --password-stdin
|
||||
# - name: Docker Pull
|
||||
# run: docker pull git.vdb.to/cerc-io/ipld-eth-db/ipld-eth-db:${{steps.vars.outputs.sha}}
|
||||
# - name: Docker Push to Docker Hub
|
||||
# run: docker push cerc-io/ipld-eth-db:${{steps.vars.outputs.tag}}
|
||||
# - name: Docker Login to Docker Registry
|
||||
# run: echo ${{ secrets.VULCANIZEJENKINS_PAT }} | docker login -u cerccicd --password-stdin
|
||||
# - name: Tag docker image
|
||||
# run: docker tag git.vdb.to/cerc-io/ipld-eth-db/ipld-eth-db:${{steps.vars.outputs.sha}} cerc-io/ipld-eth-db:${{steps.vars.outputs.tag}}
|
||||
# - name: Docker Push to Docker Hub
|
||||
# run: docker push cerc-io/ipld-eth-db:${{steps.vars.outputs.tag}}
|
25
Dockerfile
25
Dockerfile
@ -1,24 +1,19 @@
|
||||
FROM golang:1.18-alpine as builder
|
||||
FROM alpine as builder
|
||||
|
||||
RUN apk --update --no-cache add make git g++ linux-headers
|
||||
|
||||
ADD . /go/src/github.com/cerc-io/ipld-eth-db
|
||||
|
||||
# Build migration tool
|
||||
WORKDIR /go/src/github.com/pressly
|
||||
ARG GOOSE_VER="v3.6.1"
|
||||
RUN git clone --depth 1 --branch ${GOOSE_VER} https://github.com/pressly/goose.git
|
||||
WORKDIR /go/src/github.com/pressly/goose/cmd/goose
|
||||
RUN GCO_ENABLED=0 GOOS=linux go build -a -installsuffix cgo -ldflags '-extldflags "-static"' -tags='no_sqlite3' -o goose .
|
||||
# Get migration tool
|
||||
WORKDIR /
|
||||
ARG GOOSE_VERSION="v3.6.1"
|
||||
RUN arch=$(arch | sed s/aarch64/arm64/) && \
|
||||
wget -O ./goose https://github.com/pressly/goose/releases/download/${GOOSE_VERSION}/goose_linux_${arch}
|
||||
RUN chmod +x ./goose
|
||||
|
||||
# app container
|
||||
FROM alpine
|
||||
|
||||
WORKDIR /app
|
||||
|
||||
COPY --from=builder /go/src/github.com/cerc-io/ipld-eth-db/scripts/startup_script.sh .
|
||||
|
||||
COPY --from=builder /go/src/github.com/pressly/goose/cmd/goose/goose goose
|
||||
COPY --from=builder /go/src/github.com/cerc-io/ipld-eth-db/db/migrations migrations/vulcanizedb
|
||||
COPY --from=builder /goose goose
|
||||
ADD scripts/startup_script.sh .
|
||||
ADD db/migrations migrations
|
||||
|
||||
ENTRYPOINT ["/app/startup_script.sh"]
|
||||
|
127
Makefile
127
Makefile
@ -1,127 +0,0 @@
|
||||
ifndef GOPATH
|
||||
override GOPATH = $(HOME)/go
|
||||
endif
|
||||
|
||||
BIN = $(GOPATH)/bin
|
||||
|
||||
# Tools
|
||||
## Migration tool
|
||||
GOOSE = $(BIN)/goose
|
||||
$(BIN)/goose:
|
||||
go get -u github.com/pressly/goose/cmd/goose
|
||||
|
||||
.PHONY: installtools
|
||||
installtools: | $(GOOSE)
|
||||
echo "Installing tools"
|
||||
|
||||
#Database
|
||||
HOST_NAME = localhost
|
||||
PORT = 5432
|
||||
NAME =
|
||||
USER = postgres
|
||||
PASSWORD = password
|
||||
CONNECT_STRING=postgresql://$(USER):$(PASSWORD)@$(HOST_NAME):$(PORT)/$(NAME)?sslmode=disable
|
||||
|
||||
# Parameter checks
|
||||
## Check that DB variables are provided
|
||||
.PHONY: checkdbvars
|
||||
checkdbvars:
|
||||
test -n "$(HOST_NAME)" # $$HOST_NAME
|
||||
test -n "$(PORT)" # $$PORT
|
||||
test -n "$(NAME)" # $$NAME
|
||||
@echo $(CONNECT_STRING)
|
||||
|
||||
## Check that the migration variable (id/timestamp) is provided
|
||||
.PHONY: checkmigration
|
||||
checkmigration:
|
||||
test -n "$(MIGRATION)" # $$MIGRATION
|
||||
|
||||
# Check that the migration name is provided
|
||||
.PHONY: checkmigname
|
||||
checkmigname:
|
||||
test -n "$(NAME)" # $$NAME
|
||||
|
||||
# Migration operations
|
||||
## Rollback the last migration
|
||||
.PHONY: rollback
|
||||
rollback: $(GOOSE) checkdbvars
|
||||
$(GOOSE) -dir db/migrations postgres "$(CONNECT_STRING)" down
|
||||
pg_dump -O -s $(CONNECT_STRING) > schema.sql
|
||||
|
||||
## Rollback to a select migration (id/timestamp)
|
||||
.PHONY: rollback_to
|
||||
rollback_to: $(GOOSE) checkmigration checkdbvars
|
||||
$(GOOSE) -dir db/migrations postgres "$(CONNECT_STRING)" down-to "$(MIGRATION)"
|
||||
|
||||
## Rollback pre_batch_set
|
||||
.PHONY: `rollback_pre_batch_set`
|
||||
rollback_pre_batch_set: $(GOOSE) checkdbvars
|
||||
$(GOOSE) -dir db/pre_batch_processing_migrations postgres "$(CONNECT_STRING)" down
|
||||
|
||||
## Rollback post_batch_set
|
||||
.PHONY: rollback_post_batch_set
|
||||
rollback_post_batch_set: $(GOOSE) checkdbvars
|
||||
$(GOOSE) -dir db/post_batch_processing_migrations postgres "$(CONNECT_STRING)" down
|
||||
|
||||
## Apply the next up migration
|
||||
.PHONY: migrate_up_by_one
|
||||
migrate_up_by_one: $(GOOSE) checkdbvars
|
||||
$(GOOSE) -dir db/migrations postgres "$(CONNECT_STRING)" up-by-one
|
||||
|
||||
## Apply all migrations not already run
|
||||
.PHONY: migrate
|
||||
migrate: $(GOOSE) checkdbvars
|
||||
$(GOOSE) -dir db/migrations postgres "$(CONNECT_STRING)" up
|
||||
pg_dump -O -s $(CONNECT_STRING) > schema.sql
|
||||
|
||||
## Apply migrations to be ran before a batch processing
|
||||
.PHONY: migrate_pre_batch_set
|
||||
migrate_pre_batch_set: $(GOOSE) checkdbvars
|
||||
$(GOOSE) -dir db/pre_batch_processing_migrations postgres "$(CONNECT_STRING)" up
|
||||
|
||||
## Apply migrations to be ran after a batch processing, one-by-one
|
||||
.PHONY: migrate_post_batch_set_up_by_one
|
||||
migrate_post_batch_set_up_by_one: $(GOOSE) checkdbvars
|
||||
$(GOOSE) -dir db/post_batch_processing_migrations postgres "$(CONNECT_STRING)" up-by-one
|
||||
|
||||
## Apply migrations to be ran after a batch processing
|
||||
.PHONY: migrate_post_batch_set
|
||||
migrate_post_batch_set: $(GOOSE) checkdbvars
|
||||
$(GOOSE) -dir db/post_batch_processing_migrations postgres "$(CONNECT_STRING)" up
|
||||
|
||||
## Create a new migration file
|
||||
.PHONY: new_migration
|
||||
new_migration: $(GOOSE) checkmigname
|
||||
$(GOOSE) -dir db/migrations create $(NAME) sql
|
||||
|
||||
## Check which migrations are applied at the moment
|
||||
.PHONY: migration_status
|
||||
migration_status: $(GOOSE) checkdbvars
|
||||
$(GOOSE) -dir db/migrations postgres "$(CONNECT_STRING)" status
|
||||
|
||||
# Convert timestamped migrations to versioned (to be run in CI);
|
||||
# merge timestamped files to prevent conflict
|
||||
.PHONY: version_migrations
|
||||
version_migrations:
|
||||
$(GOOSE) -dir db/migrations fix
|
||||
|
||||
# Import a psql schema to the database
|
||||
.PHONY: import
|
||||
import:
|
||||
test -n "$(NAME)" # $$NAME
|
||||
psql $(NAME) < schema.sql
|
||||
|
||||
|
||||
## Build docker image with schema
|
||||
.PHONY: docker-build
|
||||
docker-build:
|
||||
docker-compose -f docker-compose.test.yml build
|
||||
|
||||
# ## Build docker image for migration
|
||||
# .PHONY: docker-concise-migration-build
|
||||
# docker-concise-migration-build:
|
||||
# docker build -t vulcanize/concise-migration-build -f ./db/Dockerfile .
|
||||
|
||||
.PHONY: test-migrations
|
||||
test-migrations: $(GOOSE)
|
||||
./scripts/test_migration.sh
|
228
README.md
228
README.md
@ -33,3 +33,231 @@ Schemas and utils for IPLD ETH Postgres database
|
||||
```
|
||||
docker-compose -f docker-compose.test.yml up --build
|
||||
```
|
||||
|
||||
## Example queries
|
||||
|
||||
Note that searching by block_number in addition to block_hash is optional in the below queries where both are provided,
|
||||
but since the tables are partitioned by block_number doing so will improve query performance by informing the query
|
||||
planner which partition it needs to search.
|
||||
|
||||
### Headers
|
||||
|
||||
Retrieve header RLP (IPLD block) and CID for a given block hash
|
||||
|
||||
```sql
|
||||
SELECT header_cids.cid,
|
||||
blocks.data
|
||||
FROM ipld.blocks,
|
||||
eth.header_cids
|
||||
WHERE header_cids.block_hash = {block_hash}
|
||||
AND header_cids.block_number = {block_number}
|
||||
AND header_cids.canonical
|
||||
AND blocks.key = header_cids.cid
|
||||
AND blocks.block_number = header_cids.block_number
|
||||
LIMIT 1
|
||||
```
|
||||
|
||||
### Uncles
|
||||
Retrieve the uncle list RLP (IPLD block) and CID for a given block hash
|
||||
|
||||
```sql
|
||||
SELECT uncle_cids.cid,
|
||||
blocks.data
|
||||
FROM eth.uncle_cids
|
||||
INNER JOIN eth.header_cids ON (
|
||||
uncle_cids.header_id = header_cids.block_hash
|
||||
AND uncle_cids.block_number = header_cids.block_number)
|
||||
INNER JOIN ipld.blocks ON (
|
||||
uncle_cids.cid = blocks.key
|
||||
AND uncle_cids.block_number = blocks.block_number)
|
||||
WHERE header_cids.block_hash = {block_hash}
|
||||
AND header_cids.block_number = {block_number}
|
||||
ORDER BY uncle_cids.parent_hash
|
||||
LIMIT 1
|
||||
```
|
||||
|
||||
### Transactions
|
||||
|
||||
Retrieve an ordered list of all the RLP encoded transactions (IPLD blocks) and their CIDs for a given block hash
|
||||
|
||||
```sql
|
||||
SELECT transaction_cids.cid,
|
||||
blocks.data
|
||||
FROM eth.transaction_cids,
|
||||
eth.header_cids,
|
||||
ipld.blocks
|
||||
WHERE header_cids.block_hash = {block_hash}
|
||||
AND header_cids.block_number = {block_number}
|
||||
AND header_cids.canonical
|
||||
AND transaction_cids.block_number = header_cids.block_number
|
||||
AND transaction_cids.header_id = header_cids.block_hash
|
||||
AND blocks.block_number = header_cids.block_number
|
||||
AND blocks.key = transaction_cids.cid
|
||||
ORDER BY eth.transaction_cids.index ASC
|
||||
```
|
||||
|
||||
Retrieve an RLP encoded transaction (IPLD block), the block hash and block number for the block it belongs to, and its position in the transaction
|
||||
for that block for a provided transaction hash
|
||||
|
||||
```sql
|
||||
SELECT blocks.data,
|
||||
transaction_cids.header_id,
|
||||
transaction_cids.block_number,
|
||||
transaction_cids.index
|
||||
FROM eth.transaction_cids,
|
||||
ipld.blocks,
|
||||
eth.header_cids
|
||||
WHERE transaction_cids.tx_hash = {transaction_hash}
|
||||
AND header_cids.block_hash = transaction_cids.header_id
|
||||
AND header_cids.block_number = transaction_cids.block_number
|
||||
AND header_cids.canonical
|
||||
AND blocks.key = transaction_cids.cid
|
||||
AND blocks.block_number = transaction_cids.block_number
|
||||
```
|
||||
|
||||
### Receipts
|
||||
|
||||
Retrieve an ordered list of all the RLP encoded receipts (IPLD blocks), their CIDs, and their corresponding transaction
|
||||
hashes for a given block hash
|
||||
|
||||
```sql
|
||||
SELECT receipt_cids.cid,
|
||||
blocks.data,
|
||||
eth.transaction_cids.tx_hash
|
||||
FROM eth.receipt_cids,
|
||||
eth.transaction_cids,
|
||||
eth.header_cids,
|
||||
ipld.blocks
|
||||
WHERE header_cids.block_hash = {block_hash}
|
||||
AND header_cids.block_number = {block_number}
|
||||
AND header_cids.canonical
|
||||
AND receipt_cids.block_number = header_cids.block_number
|
||||
AND receipt_cids.header_id = header_cids.block_hash
|
||||
AND receipt_cids.TX_ID = transaction_cids.TX_HASH
|
||||
AND transaction_cids.block_number = header_cids.block_number
|
||||
AND transaction_cids.header_id = header_cids.block_hash
|
||||
AND blocks.block_number = header_cids.block_number
|
||||
AND blocks.key = receipt_cids.cid
|
||||
ORDER BY eth.transaction_cids.index ASC
|
||||
```
|
||||
|
||||
Retrieve the RLP encoded receipt (IPLD) and CID corresponding to a provided transaction hash
|
||||
|
||||
```sql
|
||||
SELECT receipt_cids.cid,
|
||||
blocks.data
|
||||
FROM eth.receipt_cids
|
||||
INNER JOIN eth.transaction_cids ON (
|
||||
receipt_cids.tx_id = transaction_cids.tx_hash
|
||||
AND receipt_cids.block_number = transaction_cids.block_number)
|
||||
INNER JOIN ipld.blocks ON (
|
||||
receipt_cids.cid = blocks.key
|
||||
AND receipt_cids.block_number = blocks.block_number)
|
||||
WHERE transaction_cids.tx_hash = {transaction_hash}
|
||||
```
|
||||
|
||||
### Logs
|
||||
|
||||
Retrieve all the logs and their associated transaction hashes at a given block with that were emitted from
|
||||
any of the provided contract addresses and which match on any of the provided topics
|
||||
|
||||
```sql
|
||||
SELECT blocks.data,
|
||||
eth.transaction_cids.tx_hash
|
||||
FROM eth.log_cids
|
||||
INNER JOIN eth.transaction_cids ON (
|
||||
log_cids.rct_id = transaction_cids.tx_hash
|
||||
AND log_cids.header_id = transaction_cids.header_id
|
||||
AND log_cids.block_number = transaction_cids.block_number)
|
||||
INNER JOIN ipld.blocks ON (
|
||||
log_cids.cid = blocks.key
|
||||
AND log_cids.block_number = blocks.block_number)
|
||||
WHERE log_cids.header_id = {block_hash}
|
||||
AND log_cids.block_number = {block_number}
|
||||
AND eth.log_cids.address = ANY ({list,of,addresses})
|
||||
AND eth.log_cids.topic0 = ANY ({list,of,topic0s})
|
||||
AND eth.log_cids.topic1 = ANY ({list,of,topic1s})
|
||||
AND eth.log_cids.topic2 = ANY ({list,of,topic2s})
|
||||
AND eth.log_cids.topic3 = ANY ({list,of,topic3s})
|
||||
ORDER BY eth.transaction_cids.index, eth.log_cids.index
|
||||
```
|
||||
|
||||
Retrieve all the logs and their associated transaction hashes within a provided block range that were emitted from
|
||||
any of the provided contract addresses and which match on any of the provided topics
|
||||
|
||||
```sql
|
||||
SELECT blocks.data,
|
||||
eth.transaction_cids.tx_hash
|
||||
FROM eth.log_cids
|
||||
INNER JOIN eth.transaction_cids ON (
|
||||
log_cids.rct_id = transaction_cids.tx_hash
|
||||
AND log_cids.header_id = transaction_cids.header_id
|
||||
AND log_cids.block_number = transaction_cids.block_number)
|
||||
INNER JOIN eth.header_cids ON (
|
||||
transaction_cids.header_id = header_cids.block_hash
|
||||
AND transaction_cids.block_number = header_cids.block_number)
|
||||
INNER JOIN ipld.blocks ON (
|
||||
log_cids.cid = blocks.key
|
||||
AND log_cids.block_number = blocks.block_number)
|
||||
WHERE eth.header_cids.block_number >= {range_start} AND eth.header_cids.block_number <= {range_stop}
|
||||
AND eth.header_cids.canonical
|
||||
AND eth.log_cids.address = ANY ({list,of,addresses})
|
||||
AND eth.log_cids.topic0 = ANY ({list,of,topic0s})
|
||||
AND eth.log_cids.topic1 = ANY ({list,of,topic1s})
|
||||
AND eth.log_cids.topic2 = ANY ({list,of,topic2s})
|
||||
AND eth.log_cids.topic3 = ANY ({list,of,topic3s})
|
||||
ORDER BY eth.header_cids.block_number, eth.transaction_cids.index, eth.log_cids.index
|
||||
```
|
||||
|
||||
### State and storage
|
||||
|
||||
Retrieve the state account for a given address hash at a provided block hash. If `state_cids.removed == true` then
|
||||
the account is empty.
|
||||
|
||||
```sql
|
||||
SELECT state_cids.nonce,
|
||||
state_cids.balance,
|
||||
state_cids.storage_root,
|
||||
state_cids.code_hash,
|
||||
state_cids.removed
|
||||
FROM eth.state_cids,
|
||||
eth.header_cids
|
||||
WHERE state_cids.state_leaf_key = {address_hash}
|
||||
AND state_cids.block_number <=
|
||||
(SELECT block_number
|
||||
FROM eth.header_cids
|
||||
WHERE block_hash = {block_hash}
|
||||
LIMIT 1)
|
||||
AND header_cids.canonical
|
||||
AND state_cids.header_id = header_cids.block_hash
|
||||
AND state_cids.block_number = header_cids.block_number
|
||||
ORDER BY state_cids.block_number DESC
|
||||
LIMIT 1
|
||||
```
|
||||
|
||||
Retrieve a storage value, as well as the RLP encoded leaf node that stores it, for a given contract address hash and
|
||||
storage leaf key (storage slot hash) at a provided block hash. If `state_leaf_removed == true`
|
||||
or `storage_cids.removed == true` then the slot is empty
|
||||
|
||||
```sql
|
||||
SELECT storage_cids.cid,
|
||||
storage_cids.val,
|
||||
storage_cids.block_number,
|
||||
storage_cids.removed,
|
||||
was_state_leaf_removed_by_number(storage_cids.state_leaf_key, storage_cids.block_number) AS state_leaf_removed,
|
||||
blocks.data
|
||||
FROM eth.storage_cids,
|
||||
eth.header_cids,
|
||||
ipld.blocks
|
||||
WHERE header_cids.block_number <= (SELECT block_number from eth.header_cids where block_hash = $3 LIMIT 1)
|
||||
AND header_cids.canonical
|
||||
AND storage_cids.block_number = header_cids.block_number
|
||||
AND storage_cids.header_id = header_cids.block_hash
|
||||
AND storage_cids.storage_leaf_key = {storage_slot_hash}
|
||||
AND storage_cids.state_leaf_key = {contract_address_hash}
|
||||
AND blocks.key = storage_cids.cid
|
||||
AND blocks.block_number = storage_cids.block_number
|
||||
ORDER BY storage_cids.block_number DESC LIMIT 1
|
||||
```
|
||||
|
||||
|
||||
|
@ -1,15 +1,18 @@
|
||||
version: '3.2'
|
||||
|
||||
services:
|
||||
migrations:
|
||||
restart: on-failure
|
||||
depends_on:
|
||||
- ipld-eth-db
|
||||
ipld-eth-db:
|
||||
condition: service_healthy
|
||||
# Use local build
|
||||
build:
|
||||
context: .
|
||||
dockerfile: Dockerfile
|
||||
# Use an existing image
|
||||
image: cerc-io/ipld-eth-db:v4.2.1-alpha-unstable
|
||||
image: cerc/ipld-eth-db:local
|
||||
environment:
|
||||
DATABASE_USER: "vdbm"
|
||||
DATABASE_NAME: "vulcanize_testing"
|
||||
DATABASE_NAME: "cerc_testing"
|
||||
DATABASE_PASSWORD: "password"
|
||||
DATABASE_HOSTNAME: "ipld-eth-db"
|
||||
DATABASE_PORT: 5432
|
||||
@ -20,7 +23,12 @@ services:
|
||||
command: ["postgres", "-c", "log_statement=all"]
|
||||
environment:
|
||||
POSTGRES_USER: "vdbm"
|
||||
POSTGRES_DB: "vulcanize_testing"
|
||||
POSTGRES_DB: "cerc_testing"
|
||||
POSTGRES_PASSWORD: "password"
|
||||
ports:
|
||||
- "127.0.0.1:8077:5432"
|
||||
healthcheck:
|
||||
test: ["CMD", "pg_isready", "-U", "vdbm"]
|
||||
interval: 2s
|
||||
timeout: 1s
|
||||
retries: 3
|
@ -1,5 +1,7 @@
|
||||
-- +goose Up
|
||||
CREATE TABLE IF NOT EXISTS public.blocks (
|
||||
CREATE SCHEMA ipld;
|
||||
|
||||
CREATE TABLE IF NOT EXISTS ipld.blocks (
|
||||
block_number BIGINT NOT NULL,
|
||||
key TEXT NOT NULL,
|
||||
data BYTEA NOT NULL,
|
||||
@ -7,4 +9,5 @@ CREATE TABLE IF NOT EXISTS public.blocks (
|
||||
);
|
||||
|
||||
-- +goose Down
|
||||
DROP TABLE public.blocks;
|
||||
DROP TABLE ipld.blocks;
|
||||
DROP SCHEMA ipld;
|
||||
|
@ -5,17 +5,17 @@ CREATE TABLE IF NOT EXISTS eth.header_cids (
|
||||
parent_hash VARCHAR(66) NOT NULL,
|
||||
cid TEXT NOT NULL,
|
||||
td NUMERIC NOT NULL,
|
||||
node_id VARCHAR(128) NOT NULL,
|
||||
node_ids VARCHAR(128)[] NOT NULL,
|
||||
reward NUMERIC NOT NULL,
|
||||
state_root VARCHAR(66) NOT NULL,
|
||||
tx_root VARCHAR(66) NOT NULL,
|
||||
receipt_root VARCHAR(66) NOT NULL,
|
||||
uncle_root VARCHAR(66) NOT NULL,
|
||||
uncles_hash VARCHAR(66) NOT NULL,
|
||||
bloom BYTEA NOT NULL,
|
||||
timestamp BIGINT NOT NULL,
|
||||
mh_key TEXT NOT NULL,
|
||||
times_validated INTEGER NOT NULL DEFAULT 1,
|
||||
coinbase VARCHAR(66) NOT NULL,
|
||||
canonical BOOLEAN NOT NULL DEFAULT TRUE,
|
||||
withdrawals_root VARCHAR(66) NOT NULL,
|
||||
PRIMARY KEY (block_hash, block_number)
|
||||
);
|
||||
|
||||
|
@ -6,7 +6,7 @@ CREATE TABLE IF NOT EXISTS eth.uncle_cids (
|
||||
parent_hash VARCHAR(66) NOT NULL,
|
||||
cid TEXT NOT NULL,
|
||||
reward NUMERIC NOT NULL,
|
||||
mh_key TEXT NOT NULL,
|
||||
index INT NOT NULL,
|
||||
PRIMARY KEY (block_hash, block_number)
|
||||
);
|
||||
|
||||
|
@ -4,11 +4,9 @@ CREATE TABLE IF NOT EXISTS eth.transaction_cids (
|
||||
header_id VARCHAR(66) NOT NULL,
|
||||
tx_hash VARCHAR(66) NOT NULL,
|
||||
cid TEXT NOT NULL,
|
||||
dst VARCHAR(66) NOT NULL,
|
||||
dst VARCHAR(66),
|
||||
src VARCHAR(66) NOT NULL,
|
||||
index INTEGER NOT NULL,
|
||||
mh_key TEXT NOT NULL,
|
||||
tx_data BYTEA,
|
||||
tx_type INTEGER,
|
||||
value NUMERIC,
|
||||
PRIMARY KEY (tx_hash, header_id, block_number)
|
||||
|
@ -3,13 +3,10 @@ CREATE TABLE IF NOT EXISTS eth.receipt_cids (
|
||||
block_number BIGINT NOT NULL,
|
||||
header_id VARCHAR(66) NOT NULL,
|
||||
tx_id VARCHAR(66) NOT NULL,
|
||||
leaf_cid TEXT NOT NULL,
|
||||
cid TEXT NOT NULL,
|
||||
contract VARCHAR(66),
|
||||
contract_hash VARCHAR(66),
|
||||
leaf_mh_key TEXT NOT NULL,
|
||||
post_state VARCHAR(66),
|
||||
post_status INTEGER,
|
||||
log_root VARCHAR(66),
|
||||
post_status SMALLINT,
|
||||
PRIMARY KEY (tx_id, header_id, block_number)
|
||||
);
|
||||
|
||||
|
@ -2,13 +2,15 @@
|
||||
CREATE TABLE IF NOT EXISTS eth.state_cids (
|
||||
block_number BIGINT NOT NULL,
|
||||
header_id VARCHAR(66) NOT NULL,
|
||||
state_leaf_key VARCHAR(66),
|
||||
state_leaf_key VARCHAR(66) NOT NULL,
|
||||
cid TEXT NOT NULL,
|
||||
state_path BYTEA NOT NULL,
|
||||
node_type INTEGER NOT NULL,
|
||||
diff BOOLEAN NOT NULL DEFAULT FALSE,
|
||||
mh_key TEXT NOT NULL,
|
||||
PRIMARY KEY (state_path, header_id, block_number)
|
||||
balance NUMERIC, -- NULL if "removed"
|
||||
nonce BIGINT, -- NULL if "removed"
|
||||
code_hash VARCHAR(66), -- NULL if "removed"
|
||||
storage_root VARCHAR(66), -- NULL if "removed"
|
||||
removed BOOLEAN NOT NULL,
|
||||
PRIMARY KEY (state_leaf_key, header_id, block_number)
|
||||
);
|
||||
|
||||
-- +goose Down
|
||||
|
@ -2,14 +2,13 @@
|
||||
CREATE TABLE IF NOT EXISTS eth.storage_cids (
|
||||
block_number BIGINT NOT NULL,
|
||||
header_id VARCHAR(66) NOT NULL,
|
||||
state_path BYTEA NOT NULL,
|
||||
storage_leaf_key VARCHAR(66),
|
||||
state_leaf_key VARCHAR(66) NOT NULL,
|
||||
storage_leaf_key VARCHAR(66) NOT NULL,
|
||||
cid TEXT NOT NULL,
|
||||
storage_path BYTEA NOT NULL,
|
||||
node_type INTEGER NOT NULL,
|
||||
diff BOOLEAN NOT NULL DEFAULT FALSE,
|
||||
mh_key TEXT NOT NULL,
|
||||
PRIMARY KEY (storage_path, state_path, header_id, block_number)
|
||||
val BYTEA, -- NULL if "removed"
|
||||
removed BOOLEAN NOT NULL,
|
||||
PRIMARY KEY (storage_leaf_key, state_leaf_key, header_id, block_number)
|
||||
);
|
||||
|
||||
-- +goose Down
|
||||
|
@ -2,8 +2,7 @@
|
||||
CREATE TABLE IF NOT EXISTS eth.log_cids (
|
||||
block_number BIGINT NOT NULL,
|
||||
header_id VARCHAR(66) NOT NULL,
|
||||
leaf_cid TEXT NOT NULL,
|
||||
leaf_mh_key TEXT NOT NULL,
|
||||
cid TEXT NOT NULL,
|
||||
rct_id VARCHAR(66) NOT NULL,
|
||||
address VARCHAR(66) NOT NULL,
|
||||
index INTEGER NOT NULL,
|
||||
@ -11,7 +10,6 @@ CREATE TABLE IF NOT EXISTS eth.log_cids (
|
||||
topic1 VARCHAR(66),
|
||||
topic2 VARCHAR(66),
|
||||
topic3 VARCHAR(66),
|
||||
log_data BYTEA,
|
||||
PRIMARY KEY (rct_id, index, header_id, block_number)
|
||||
);
|
||||
|
@ -1,14 +0,0 @@
|
||||
-- +goose Up
|
||||
CREATE TABLE IF NOT EXISTS eth.state_accounts (
|
||||
block_number BIGINT NOT NULL,
|
||||
header_id VARCHAR(66) NOT NULL,
|
||||
state_path BYTEA NOT NULL,
|
||||
balance NUMERIC NOT NULL,
|
||||
nonce BIGINT NOT NULL,
|
||||
code_hash BYTEA NOT NULL,
|
||||
storage_root VARCHAR(66) NOT NULL,
|
||||
PRIMARY KEY (state_path, header_id, block_number)
|
||||
);
|
||||
|
||||
-- +goose Down
|
||||
DROP TABLE eth.state_accounts;
|
@ -1,12 +0,0 @@
|
||||
-- +goose Up
|
||||
CREATE TABLE IF NOT EXISTS eth.access_list_elements (
|
||||
block_number BIGINT NOT NULL,
|
||||
tx_id VARCHAR(66) NOT NULL,
|
||||
index INTEGER NOT NULL,
|
||||
address VARCHAR(66),
|
||||
storage_keys VARCHAR(66)[],
|
||||
PRIMARY KEY (tx_id, index, block_number)
|
||||
);
|
||||
|
||||
-- +goose Down
|
||||
DROP TABLE eth.access_list_elements;
|
@ -3,7 +3,7 @@ COMMENT ON TABLE public.nodes IS E'@name NodeInfo';
|
||||
COMMENT ON TABLE eth.transaction_cids IS E'@name EthTransactionCids';
|
||||
COMMENT ON TABLE eth.header_cids IS E'@name EthHeaderCids';
|
||||
COMMENT ON COLUMN public.nodes.node_id IS E'@name ChainNodeID';
|
||||
COMMENT ON COLUMN eth.header_cids.node_id IS E'@name EthNodeID';
|
||||
COMMENT ON COLUMN eth.header_cids.node_ids IS E'@name EthNodeIDs';
|
||||
|
||||
-- +goose Down
|
||||
|
||||
@ -11,4 +11,4 @@ COMMENT ON TABLE public.nodes IS NULL;
|
||||
COMMENT ON TABLE eth.transaction_cids IS NULL;
|
||||
COMMENT ON TABLE eth.header_cids IS NULL;
|
||||
COMMENT ON COLUMN public.nodes.node_id IS NULL;
|
||||
COMMENT ON COLUMN eth.header_cids.node_id IS NULL;
|
||||
COMMENT ON COLUMN eth.header_cids.node_ids IS NULL;
|
101
db/migrations/00012_create_cid_indexes.sql
Normal file
101
db/migrations/00012_create_cid_indexes.sql
Normal file
@ -0,0 +1,101 @@
|
||||
-- +goose Up
|
||||
-- header indexes
|
||||
CREATE INDEX header_block_number_index ON eth.header_cids USING btree (block_number);
|
||||
CREATE UNIQUE INDEX header_cid_block_number_index ON eth.header_cids USING btree (cid, block_number);
|
||||
CREATE INDEX state_root_index ON eth.header_cids USING btree (state_root);
|
||||
CREATE INDEX timestamp_index ON eth.header_cids USING btree (timestamp);
|
||||
|
||||
-- uncle indexes
|
||||
CREATE INDEX uncle_block_number_index ON eth.uncle_cids USING btree (block_number);
|
||||
CREATE UNIQUE INDEX uncle_cid_block_number_index ON eth.uncle_cids USING btree (cid, block_number, index);
|
||||
CREATE INDEX uncle_header_id_index ON eth.uncle_cids USING btree (header_id);
|
||||
|
||||
-- transaction indexes
|
||||
CREATE INDEX tx_block_number_index ON eth.transaction_cids USING btree (block_number);
|
||||
CREATE INDEX tx_header_id_index ON eth.transaction_cids USING btree (header_id);
|
||||
CREATE INDEX tx_cid_block_number_index ON eth.transaction_cids USING btree (cid, block_number);
|
||||
CREATE INDEX tx_dst_index ON eth.transaction_cids USING btree (dst);
|
||||
CREATE INDEX tx_src_index ON eth.transaction_cids USING btree (src);
|
||||
|
||||
-- receipt indexes
|
||||
CREATE INDEX rct_block_number_index ON eth.receipt_cids USING btree (block_number);
|
||||
CREATE INDEX rct_header_id_index ON eth.receipt_cids USING btree (header_id);
|
||||
CREATE INDEX rct_cid_block_number_index ON eth.receipt_cids USING btree (cid, block_number);
|
||||
CREATE INDEX rct_contract_index ON eth.receipt_cids USING btree (contract);
|
||||
|
||||
-- state node indexes
|
||||
CREATE INDEX state_block_number_index ON eth.state_cids USING btree (block_number);
|
||||
CREATE INDEX state_cid_block_number_index ON eth.state_cids USING btree (cid, block_number);
|
||||
CREATE INDEX state_header_id_index ON eth.state_cids USING btree (header_id);
|
||||
CREATE INDEX state_removed_index ON eth.state_cids USING btree (removed);
|
||||
CREATE INDEX state_code_hash_index ON eth.state_cids USING btree (code_hash); -- could be useful for e.g. selecting all the state accounts with the same contract bytecode deployed
|
||||
CREATE INDEX state_leaf_key_block_number_index ON eth.state_cids(state_leaf_key, block_number DESC);
|
||||
|
||||
-- storage node indexes
|
||||
CREATE INDEX storage_block_number_index ON eth.storage_cids USING btree (block_number);
|
||||
CREATE INDEX storage_state_leaf_key_index ON eth.storage_cids USING btree (state_leaf_key);
|
||||
CREATE INDEX storage_cid_block_number_index ON eth.storage_cids USING btree (cid, block_number);
|
||||
CREATE INDEX storage_header_id_index ON eth.storage_cids USING btree (header_id);
|
||||
CREATE INDEX storage_removed_index ON eth.storage_cids USING btree (removed);
|
||||
CREATE INDEX storage_leaf_key_block_number_index ON eth.storage_cids(storage_leaf_key, block_number DESC);
|
||||
|
||||
-- log indexes
|
||||
CREATE INDEX log_block_number_index ON eth.log_cids USING btree (block_number);
|
||||
CREATE INDEX log_header_id_index ON eth.log_cids USING btree (header_id);
|
||||
CREATE INDEX log_cid_block_number_index ON eth.log_cids USING btree (cid, block_number);
|
||||
CREATE INDEX log_address_index ON eth.log_cids USING btree (address);
|
||||
CREATE INDEX log_topic0_index ON eth.log_cids USING btree (topic0);
|
||||
CREATE INDEX log_topic1_index ON eth.log_cids USING btree (topic1);
|
||||
CREATE INDEX log_topic2_index ON eth.log_cids USING btree (topic2);
|
||||
CREATE INDEX log_topic3_index ON eth.log_cids USING btree (topic3);
|
||||
|
||||
-- +goose Down
|
||||
-- log indexes
|
||||
DROP INDEX eth.log_topic3_index;
|
||||
DROP INDEX eth.log_topic2_index;
|
||||
DROP INDEX eth.log_topic1_index;
|
||||
DROP INDEX eth.log_topic0_index;
|
||||
DROP INDEX eth.log_address_index;
|
||||
DROP INDEX eth.log_cid_block_number_index;
|
||||
DROP INDEX eth.log_header_id_index;
|
||||
DROP INDEX eth.log_block_number_index;
|
||||
|
||||
-- storage node indexes
|
||||
DROP INDEX eth.storage_removed_index;
|
||||
DROP INDEX eth.storage_header_id_index;
|
||||
DROP INDEX eth.storage_cid_block_number_index;
|
||||
DROP INDEX eth.storage_state_leaf_key_index;
|
||||
DROP INDEX eth.storage_block_number_index;
|
||||
DROP INDEX eth.storage_leaf_key_block_number_index;
|
||||
|
||||
-- state node indexes
|
||||
DROP INDEX eth.state_code_hash_index;
|
||||
DROP INDEX eth.state_removed_index;
|
||||
DROP INDEX eth.state_header_id_index;
|
||||
DROP INDEX eth.state_cid_block_number_index;
|
||||
DROP INDEX eth.state_block_number_index;
|
||||
DROP INDEX eth.state_leaf_key_block_number_index;
|
||||
|
||||
-- receipt indexes
|
||||
DROP INDEX eth.rct_contract_index;
|
||||
DROP INDEX eth.rct_cid_block_number_index;
|
||||
DROP INDEX eth.rct_header_id_index;
|
||||
DROP INDEX eth.rct_block_number_index;
|
||||
|
||||
-- transaction indexes
|
||||
DROP INDEX eth.tx_src_index;
|
||||
DROP INDEX eth.tx_dst_index;
|
||||
DROP INDEX eth.tx_cid_block_number_index;
|
||||
DROP INDEX eth.tx_header_id_index;
|
||||
DROP INDEX eth.tx_block_number_index;
|
||||
|
||||
-- uncle indexes
|
||||
DROP INDEX eth.uncle_block_number_index;
|
||||
DROP INDEX eth.uncle_cid_block_number_index;
|
||||
DROP INDEX eth.uncle_header_id_index;
|
||||
|
||||
-- header indexes
|
||||
DROP INDEX eth.timestamp_index;
|
||||
DROP INDEX eth.state_root_index;
|
||||
DROP INDEX eth.header_cid_block_number_index;
|
||||
DROP INDEX eth.header_block_number_index;
|
@ -5,5 +5,8 @@ CREATE TABLE IF NOT EXISTS public.db_version (
|
||||
tstamp TIMESTAMP WITHOUT TIME ZONE DEFAULT NOW()
|
||||
);
|
||||
|
||||
INSERT INTO public.db_version (singleton, version) VALUES (true, 'v5.0.0')
|
||||
ON CONFLICT (singleton) DO UPDATE SET (version, tstamp) = ('v5.0.0', NOW());
|
||||
|
||||
-- +goose Down
|
||||
DROP TABLE public.db_version;
|
@ -1,133 +0,0 @@
|
||||
-- +goose Up
|
||||
-- header indexes
|
||||
CREATE INDEX header_block_number_index ON eth.header_cids USING brin (block_number);
|
||||
CREATE UNIQUE INDEX header_cid_index ON eth.header_cids USING btree (cid, block_number);
|
||||
CREATE UNIQUE INDEX header_mh_block_number_index ON eth.header_cids USING btree (mh_key, block_number);
|
||||
CREATE INDEX state_root_index ON eth.header_cids USING btree (state_root);
|
||||
CREATE INDEX timestamp_index ON eth.header_cids USING brin (timestamp);
|
||||
|
||||
-- uncle indexes
|
||||
CREATE INDEX uncle_block_number_index ON eth.uncle_cids USING brin (block_number);
|
||||
CREATE UNIQUE INDEX uncle_mh_block_number_index ON eth.uncle_cids USING btree (mh_key, block_number);
|
||||
CREATE INDEX uncle_header_id_index ON eth.uncle_cids USING btree (header_id);
|
||||
|
||||
-- transaction indexes
|
||||
CREATE INDEX tx_block_number_index ON eth.transaction_cids USING brin (block_number);
|
||||
CREATE INDEX tx_header_id_index ON eth.transaction_cids USING btree (header_id);
|
||||
CREATE INDEX tx_cid_index ON eth.transaction_cids USING btree (cid, block_number);
|
||||
CREATE INDEX tx_mh_block_number_index ON eth.transaction_cids USING btree (mh_key, block_number);
|
||||
CREATE INDEX tx_dst_index ON eth.transaction_cids USING btree (dst);
|
||||
CREATE INDEX tx_src_index ON eth.transaction_cids USING btree (src);
|
||||
|
||||
-- receipt indexes
|
||||
CREATE INDEX rct_block_number_index ON eth.receipt_cids USING brin (block_number);
|
||||
CREATE INDEX rct_header_id_index ON eth.receipt_cids USING btree (header_id);
|
||||
CREATE INDEX rct_leaf_cid_index ON eth.receipt_cids USING btree (leaf_cid);
|
||||
CREATE INDEX rct_leaf_mh_block_number_index ON eth.receipt_cids USING btree (leaf_mh_key, block_number);
|
||||
CREATE INDEX rct_contract_index ON eth.receipt_cids USING btree (contract);
|
||||
CREATE INDEX rct_contract_hash_index ON eth.receipt_cids USING btree (contract_hash);
|
||||
|
||||
-- state node indexes
|
||||
CREATE INDEX state_block_number_index ON eth.state_cids USING brin (block_number);
|
||||
CREATE INDEX state_leaf_key_index ON eth.state_cids USING btree (state_leaf_key);
|
||||
CREATE INDEX state_cid_index ON eth.state_cids USING btree (cid);
|
||||
CREATE INDEX state_mh_block_number_index ON eth.state_cids USING btree (mh_key, block_number);
|
||||
CREATE INDEX state_header_id_index ON eth.state_cids USING btree (header_id);
|
||||
CREATE INDEX state_node_type_index ON eth.state_cids USING btree (node_type);
|
||||
|
||||
-- storage node indexes
|
||||
CREATE INDEX storage_block_number_index ON eth.storage_cids USING brin (block_number);
|
||||
CREATE INDEX storage_state_path_index ON eth.storage_cids USING btree (state_path);
|
||||
CREATE INDEX storage_leaf_key_index ON eth.storage_cids USING btree (storage_leaf_key);
|
||||
CREATE INDEX storage_cid_index ON eth.storage_cids USING btree (cid);
|
||||
CREATE INDEX storage_mh_block_number_index ON eth.storage_cids USING btree (mh_key, block_number);
|
||||
CREATE INDEX storage_header_id_index ON eth.storage_cids USING btree (header_id);
|
||||
CREATE INDEX storage_node_type_index ON eth.storage_cids USING btree (node_type);
|
||||
|
||||
-- state accounts indexes
|
||||
CREATE INDEX account_block_number_index ON eth.state_accounts USING brin (block_number);
|
||||
CREATE INDEX account_header_id_index ON eth.state_accounts USING btree (header_id);
|
||||
CREATE INDEX account_storage_root_index ON eth.state_accounts USING btree (storage_root);
|
||||
|
||||
-- access list indexes
|
||||
CREATE INDEX access_list_block_number_index ON eth.access_list_elements USING brin (block_number);
|
||||
CREATE INDEX access_list_element_address_index ON eth.access_list_elements USING btree (address);
|
||||
CREATE INDEX access_list_storage_keys_index ON eth.access_list_elements USING gin (storage_keys);
|
||||
|
||||
-- log indexes
|
||||
CREATE INDEX log_block_number_index ON eth.log_cids USING brin (block_number);
|
||||
CREATE INDEX log_header_id_index ON eth.log_cids USING btree (header_id);
|
||||
CREATE INDEX log_leaf_mh_block_number_index ON eth.log_cids USING btree (leaf_mh_key, block_number);
|
||||
CREATE INDEX log_cid_index ON eth.log_cids USING btree (leaf_cid);
|
||||
CREATE INDEX log_address_index ON eth.log_cids USING btree (address);
|
||||
CREATE INDEX log_topic0_index ON eth.log_cids USING btree (topic0);
|
||||
CREATE INDEX log_topic1_index ON eth.log_cids USING btree (topic1);
|
||||
CREATE INDEX log_topic2_index ON eth.log_cids USING btree (topic2);
|
||||
CREATE INDEX log_topic3_index ON eth.log_cids USING btree (topic3);
|
||||
|
||||
-- +goose Down
|
||||
-- log indexes
|
||||
DROP INDEX eth.log_topic3_index;
|
||||
DROP INDEX eth.log_topic2_index;
|
||||
DROP INDEX eth.log_topic1_index;
|
||||
DROP INDEX eth.log_topic0_index;
|
||||
DROP INDEX eth.log_address_index;
|
||||
DROP INDEX eth.log_cid_index;
|
||||
DROP INDEX eth.log_leaf_mh_block_number_index;
|
||||
DROP INDEX eth.log_header_id_index;
|
||||
DROP INDEX eth.log_block_number_index;
|
||||
|
||||
-- access list indexes
|
||||
DROP INDEX eth.access_list_storage_keys_index;
|
||||
DROP INDEX eth.access_list_element_address_index;
|
||||
DROP INDEX eth.access_list_block_number_index;
|
||||
|
||||
-- state account indexes
|
||||
DROP INDEX eth.account_storage_root_index;
|
||||
DROP index eth.account_header_id_index;
|
||||
DROP INDEX eth.account_block_number_index;
|
||||
|
||||
-- storage node indexes
|
||||
DROP INDEX eth.storage_node_type_index;
|
||||
DROP INDEX eth.storage_header_id_index;
|
||||
DROP INDEX eth.storage_mh_block_number_index;
|
||||
DROP INDEX eth.storage_cid_index;
|
||||
DROP INDEX eth.storage_leaf_key_index;
|
||||
DROP INDEX eth.storage_state_path_index;
|
||||
DROP INDEX eth.storage_block_number_index;
|
||||
|
||||
-- state node indexes
|
||||
DROP INDEX eth.state_node_type_index;
|
||||
DROP INDEX eth.state_header_id_index;
|
||||
DROP INDEX eth.state_mh_block_number_index;
|
||||
DROP INDEX eth.state_cid_index;
|
||||
DROP INDEX eth.state_leaf_key_index;
|
||||
DROP INDEX eth.state_block_number_index;
|
||||
|
||||
-- receipt indexes
|
||||
DROP INDEX eth.rct_contract_hash_index;
|
||||
DROP INDEX eth.rct_contract_index;
|
||||
DROP INDEX eth.rct_leaf_mh_block_number_index;
|
||||
DROP INDEX eth.rct_leaf_cid_index;
|
||||
DROP INDEX eth.rct_header_id_index;
|
||||
DROP INDEX eth.rct_block_number_index;
|
||||
|
||||
-- transaction indexes
|
||||
DROP INDEX eth.tx_src_index;
|
||||
DROP INDEX eth.tx_dst_index;
|
||||
DROP INDEX eth.tx_mh_block_number_index;
|
||||
DROP INDEX eth.tx_cid_index;
|
||||
DROP INDEX eth.tx_header_id_index;
|
||||
DROP INDEX eth.tx_block_number_index;
|
||||
|
||||
-- uncle indexes
|
||||
DROP INDEX eth.uncle_block_number_index;
|
||||
DROP INDEX eth.uncle_mh_block_number_index;
|
||||
DROP INDEX eth.uncle_header_id_index;
|
||||
|
||||
-- header indexes
|
||||
DROP INDEX eth.timestamp_index;
|
||||
DROP INDEX eth.state_root_index;
|
||||
DROP INDEX eth.header_mh_block_number_index;
|
||||
DROP INDEX eth.header_cid_index;
|
||||
DROP INDEX eth.header_block_number_index;
|
@ -1,119 +0,0 @@
|
||||
-- +goose Up
|
||||
-- Name: graphql_subscription(); Type: FUNCTION; Schema: eth; Owner: -
|
||||
|
||||
-- +goose StatementBegin
|
||||
CREATE FUNCTION eth.graphql_subscription() RETURNS TRIGGER AS $$
|
||||
DECLARE
|
||||
obj jsonb;
|
||||
BEGIN
|
||||
IF (TG_TABLE_NAME = 'state_cids') OR (TG_TABLE_NAME = 'state_accounts') THEN
|
||||
obj := json_build_array(
|
||||
TG_TABLE_NAME,
|
||||
NEW.header_id,
|
||||
NEW.state_path
|
||||
);
|
||||
ELSIF (TG_TABLE_NAME = 'storage_cids') THEN
|
||||
obj := json_build_array(
|
||||
TG_TABLE_NAME,
|
||||
NEW.header_id,
|
||||
NEW.state_path,
|
||||
NEW.storage_path
|
||||
);
|
||||
ELSIF (TG_TABLE_NAME = 'log_cids') THEN
|
||||
obj := json_build_array(
|
||||
TG_TABLE_NAME,
|
||||
NEW.header_id,
|
||||
NEW.rct_id,
|
||||
NEW.index
|
||||
);
|
||||
ELSIF (TG_TABLE_NAME = 'receipt_cids') THEN
|
||||
obj := json_build_array(
|
||||
TG_TABLE_NAME,
|
||||
NEW.header_id,
|
||||
NEW.tx_id
|
||||
);
|
||||
ELSIF (TG_TABLE_NAME = 'transaction_cids') THEN
|
||||
obj := json_build_array(
|
||||
TG_TABLE_NAME,
|
||||
NEW.header_id,
|
||||
NEW.tx_hash
|
||||
);
|
||||
ELSIF (TG_TABLE_NAME = 'access_list_elements') THEN
|
||||
obj := json_build_array(
|
||||
TG_TABLE_NAME,
|
||||
NEW.tx_id,
|
||||
NEW.index
|
||||
);
|
||||
ELSIF (TG_TABLE_NAME = 'uncle_cids') OR (TG_TABLE_NAME = 'header_cids') THEN
|
||||
obj := json_build_array(
|
||||
TG_TABLE_NAME,
|
||||
NEW.block_hash
|
||||
);
|
||||
END IF;
|
||||
|
||||
perform pg_notify('postgraphile:' || TG_RELNAME , json_build_object(
|
||||
'__node__', obj
|
||||
)::text
|
||||
);
|
||||
RETURN NEW;
|
||||
END;
|
||||
$$ language plpgsql;
|
||||
-- +goose StatementEnd
|
||||
|
||||
CREATE TRIGGER trg_eth_header_cids
|
||||
AFTER INSERT ON eth.header_cids
|
||||
FOR EACH ROW
|
||||
EXECUTE PROCEDURE eth.graphql_subscription();
|
||||
|
||||
CREATE TRIGGER trg_eth_uncle_cids
|
||||
AFTER INSERT ON eth.uncle_cids
|
||||
FOR EACH ROW
|
||||
EXECUTE PROCEDURE eth.graphql_subscription();
|
||||
|
||||
CREATE TRIGGER trg_eth_transaction_cids
|
||||
AFTER INSERT ON eth.transaction_cids
|
||||
FOR EACH ROW
|
||||
EXECUTE PROCEDURE eth.graphql_subscription();
|
||||
|
||||
CREATE TRIGGER trg_eth_receipt_cids
|
||||
AFTER INSERT ON eth.receipt_cids
|
||||
FOR EACH ROW
|
||||
EXECUTE PROCEDURE eth.graphql_subscription();
|
||||
|
||||
CREATE TRIGGER trg_eth_state_cids
|
||||
AFTER INSERT ON eth.state_cids
|
||||
FOR EACH ROW
|
||||
EXECUTE PROCEDURE eth.graphql_subscription();
|
||||
|
||||
CREATE TRIGGER trg_eth_log_cids
|
||||
AFTER INSERT ON eth.log_cids
|
||||
FOR EACH ROW
|
||||
EXECUTE PROCEDURE eth.graphql_subscription();
|
||||
|
||||
CREATE TRIGGER trg_eth_storage_cids
|
||||
AFTER INSERT ON eth.storage_cids
|
||||
FOR EACH ROW
|
||||
EXECUTE PROCEDURE eth.graphql_subscription();
|
||||
|
||||
CREATE TRIGGER trg_eth_state_accounts
|
||||
AFTER INSERT ON eth.state_accounts
|
||||
FOR EACH ROW
|
||||
EXECUTE PROCEDURE eth.graphql_subscription();
|
||||
|
||||
CREATE TRIGGER trg_eth_access_list_elements
|
||||
AFTER INSERT ON eth.access_list_elements
|
||||
FOR EACH ROW
|
||||
EXECUTE PROCEDURE eth.graphql_subscription();
|
||||
|
||||
-- +goose Down
|
||||
DROP TRIGGER trg_eth_uncle_cids ON eth.uncle_cids;
|
||||
DROP TRIGGER trg_eth_transaction_cids ON eth.transaction_cids;
|
||||
DROP TRIGGER trg_eth_storage_cids ON eth.storage_cids;
|
||||
DROP TRIGGER trg_eth_state_cids ON eth.state_cids;
|
||||
DROP TRIGGER trg_eth_state_accounts ON eth.state_accounts;
|
||||
DROP TRIGGER trg_eth_receipt_cids ON eth.receipt_cids;
|
||||
DROP TRIGGER trg_eth_header_cids ON eth.header_cids;
|
||||
DROP TRIGGER trg_eth_log_cids ON eth.log_cids;
|
||||
DROP TRIGGER trg_eth_access_list_elements ON eth.access_list_elements;
|
||||
|
||||
DROP FUNCTION eth.graphql_subscription();
|
43
db/migrations/00016_create_stored_procedures.sql
Normal file
43
db/migrations/00016_create_stored_procedures.sql
Normal file
@ -0,0 +1,43 @@
|
||||
-- +goose Up
|
||||
-- +goose StatementBegin
|
||||
-- returns whether the state leaf key is vacated (previously existed but now is empty) at the provided block hash
|
||||
CREATE OR REPLACE FUNCTION was_state_leaf_removed(v_key VARCHAR(66), v_hash VARCHAR)
|
||||
RETURNS boolean AS $$
|
||||
SELECT state_cids.removed = true
|
||||
FROM eth.state_cids
|
||||
INNER JOIN eth.header_cids ON (state_cids.header_id = header_cids.block_hash)
|
||||
WHERE state_leaf_key = v_key
|
||||
AND state_cids.block_number <= (SELECT block_number
|
||||
FROM eth.header_cids
|
||||
WHERE block_hash = v_hash)
|
||||
ORDER BY state_cids.block_number DESC LIMIT 1;
|
||||
$$
|
||||
language sql;
|
||||
-- +goose StatementEnd
|
||||
|
||||
-- +goose StatementBegin
|
||||
-- returns whether the state leaf key is vacated (previously existed but now is empty) at the provided block height
|
||||
CREATE OR REPLACE FUNCTION public.was_state_leaf_removed_by_number(v_key VARCHAR(66), v_block_no BIGINT)
|
||||
RETURNS BOOLEAN AS $$
|
||||
SELECT state_cids.removed = true
|
||||
FROM eth.state_cids
|
||||
INNER JOIN eth.header_cids ON (state_cids.header_id = header_cids.block_hash)
|
||||
WHERE state_leaf_key = v_key
|
||||
AND state_cids.block_number <= v_block_no
|
||||
ORDER BY state_cids.block_number DESC LIMIT 1;
|
||||
$$
|
||||
language sql;
|
||||
-- +goose StatementEnd
|
||||
|
||||
-- +goose StatementBegin
|
||||
CREATE OR REPLACE FUNCTION canonical_header_hash(height BIGINT) RETURNS character varying AS
|
||||
$BODY$
|
||||
SELECT block_hash from eth.header_cids WHERE block_number = height AND canonical = true LIMIT 1;
|
||||
$BODY$
|
||||
LANGUAGE sql;
|
||||
-- +goose StatementEnd
|
||||
|
||||
-- +goose Down
|
||||
DROP FUNCTION was_state_leaf_removed;
|
||||
DROP FUNCTION was_state_leaf_removed_by_number;
|
||||
DROP FUNCTION canonical_header_hash;
|
110
db/migrations/00017_create_get_storage_at_functions.sql
Normal file
110
db/migrations/00017_create_get_storage_at_functions.sql
Normal file
@ -0,0 +1,110 @@
|
||||
-- +goose Up
|
||||
-- +goose StatementBegin
|
||||
CREATE OR REPLACE FUNCTION public.get_storage_at_by_number(v_state_leaf_key text, v_storage_leaf_key text, v_block_no bigint)
|
||||
RETURNS TABLE
|
||||
(
|
||||
cid TEXT,
|
||||
val BYTEA,
|
||||
block_number BIGINT,
|
||||
removed BOOL,
|
||||
state_leaf_removed BOOL
|
||||
)
|
||||
AS
|
||||
$BODY$
|
||||
DECLARE
|
||||
v_state_path BYTEA;
|
||||
v_header TEXT;
|
||||
v_canonical_header TEXT;
|
||||
BEGIN
|
||||
CREATE TEMP TABLE tmp_tt_stg2
|
||||
(
|
||||
header_id TEXT,
|
||||
cid TEXT,
|
||||
val BYTEA,
|
||||
block_number BIGINT,
|
||||
removed BOOL,
|
||||
state_leaf_removed BOOL
|
||||
) ON COMMIT DROP;
|
||||
|
||||
-- in best case scenario, the latest record we find for the provided keys is for a canonical block
|
||||
INSERT INTO tmp_tt_stg2
|
||||
SELECT storage_cids.header_id,
|
||||
storage_cids.cid,
|
||||
storage_cids.val,
|
||||
storage_cids.block_number,
|
||||
storage_cids.removed,
|
||||
was_state_leaf_removed_by_number(v_state_leaf_key, v_block_no) AS state_leaf_removed
|
||||
FROM eth.storage_cids
|
||||
WHERE storage_leaf_key = v_storage_leaf_key
|
||||
AND storage_cids.state_leaf_key = v_state_leaf_key -- can lookup directly on the leaf key in v5
|
||||
AND storage_cids.block_number <= v_block_no
|
||||
ORDER BY storage_cids.block_number DESC LIMIT 1;
|
||||
|
||||
-- check if result is from canonical state
|
||||
SELECT header_id, canonical_header_hash(tmp_tt_stg2.block_number)
|
||||
INTO v_header, v_canonical_header
|
||||
FROM tmp_tt_stg2 LIMIT 1;
|
||||
|
||||
IF v_header IS NULL OR v_header != v_canonical_header THEN
|
||||
RAISE NOTICE 'get_storage_at_by_number: chosen header NULL OR % != canonical header % for block number %, trying again.', v_header, v_canonical_header, v_block_no;
|
||||
TRUNCATE tmp_tt_stg2;
|
||||
-- If we hit on a non-canonical block, we need to go back and do a comprehensive check.
|
||||
-- We try to avoid this to avoid joining between storage_cids and header_cids
|
||||
INSERT INTO tmp_tt_stg2
|
||||
SELECT storage_cids.header_id,
|
||||
storage_cids.cid,
|
||||
storage_cids.val,
|
||||
storage_cids.block_number,
|
||||
storage_cids.removed,
|
||||
was_state_leaf_removed_by_number(
|
||||
v_state_leaf_key,
|
||||
v_block_no
|
||||
) AS state_leaf_removed
|
||||
FROM eth.storage_cids
|
||||
INNER JOIN eth.header_cids ON (
|
||||
storage_cids.header_id = header_cids.block_hash
|
||||
AND storage_cids.block_number = header_cids.block_number
|
||||
)
|
||||
WHERE state_leaf_key = v_state_leaf_key
|
||||
AND storage_leaf_key = v_storage_leaf_key
|
||||
AND storage_cids.block_number <= v_block_no
|
||||
AND header_cids.block_number <= v_block_no
|
||||
AND header_cids.block_hash = (SELECT canonical_header_hash(header_cids.block_number))
|
||||
ORDER BY header_cids.block_number DESC LIMIT 1;
|
||||
END IF;
|
||||
|
||||
RETURN QUERY SELECT t.cid, t.val, t.block_number, t.removed, t.state_leaf_removed
|
||||
FROM tmp_tt_stg2 AS t LIMIT 1;
|
||||
END
|
||||
$BODY$
|
||||
language 'plpgsql';
|
||||
-- +goose StatementEnd
|
||||
|
||||
-- +goose StatementBegin
|
||||
CREATE OR REPLACE FUNCTION public.get_storage_at_by_hash(v_state_leaf_key TEXT, v_storage_leaf_key text, v_block_hash text)
|
||||
RETURNS TABLE
|
||||
(
|
||||
cid TEXT,
|
||||
val BYTEA,
|
||||
block_number BIGINT,
|
||||
removed BOOL,
|
||||
state_leaf_removed BOOL
|
||||
)
|
||||
AS
|
||||
$BODY$
|
||||
DECLARE
|
||||
v_block_no BIGINT;
|
||||
BEGIN
|
||||
SELECT h.block_number INTO v_block_no FROM eth.header_cids AS h WHERE block_hash = v_block_hash LIMIT 1;
|
||||
IF v_block_no IS NULL THEN
|
||||
RETURN;
|
||||
END IF;
|
||||
RETURN QUERY SELECT * FROM get_storage_at_by_number(v_state_leaf_key, v_storage_leaf_key, v_block_no);
|
||||
END
|
||||
$BODY$
|
||||
LANGUAGE 'plpgsql';
|
||||
-- +goose StatementEnd
|
||||
|
||||
-- +goose Down
|
||||
DROP FUNCTION get_storage_at_by_hash;
|
||||
DROP FUNCTION get_storage_at_by_number;
|
105
db/migrations/00018_create_foreign_keys.sql
Normal file
105
db/migrations/00018_create_foreign_keys.sql
Normal file
@ -0,0 +1,105 @@
|
||||
-- +goose Up
|
||||
ALTER TABLE eth.header_cids
|
||||
ADD CONSTRAINT header_cids_ipld_blocks_fkey
|
||||
FOREIGN KEY (cid, block_number)
|
||||
REFERENCES ipld.blocks (key, block_number);
|
||||
|
||||
ALTER TABLE eth.uncle_cids
|
||||
ADD CONSTRAINT uncle_cids_ipld_blocks_fkey
|
||||
FOREIGN KEY (cid, block_number)
|
||||
REFERENCES ipld.blocks (key, block_number);
|
||||
|
||||
ALTER TABLE eth.uncle_cids
|
||||
ADD CONSTRAINT uncle_cids_header_cids_fkey
|
||||
FOREIGN KEY (header_id, block_number)
|
||||
REFERENCES eth.header_cids (block_hash, block_number);
|
||||
|
||||
ALTER TABLE eth.transaction_cids
|
||||
ADD CONSTRAINT transaction_cids_ipld_blocks_fkey
|
||||
FOREIGN KEY (cid, block_number)
|
||||
REFERENCES ipld.blocks (key, block_number);
|
||||
|
||||
ALTER TABLE eth.transaction_cids
|
||||
ADD CONSTRAINT transaction_cids_header_cids_fkey
|
||||
FOREIGN KEY (header_id, block_number)
|
||||
REFERENCES eth.header_cids (block_hash, block_number);
|
||||
|
||||
ALTER TABLE eth.receipt_cids
|
||||
ADD CONSTRAINT receipt_cids_ipld_blocks_fkey
|
||||
FOREIGN KEY (cid, block_number)
|
||||
REFERENCES ipld.blocks (key, block_number);
|
||||
|
||||
ALTER TABLE eth.receipt_cids
|
||||
ADD CONSTRAINT receipt_cids_transaction_cids_fkey
|
||||
FOREIGN KEY (tx_id, header_id, block_number)
|
||||
REFERENCES eth.transaction_cids (tx_hash, header_id, block_number);
|
||||
|
||||
ALTER TABLE eth.state_cids
|
||||
ADD CONSTRAINT state_cids_ipld_blocks_fkey
|
||||
FOREIGN KEY (cid, block_number)
|
||||
REFERENCES ipld.blocks (key, block_number);
|
||||
|
||||
ALTER TABLE eth.state_cids
|
||||
ADD CONSTRAINT state_cids_header_cids_fkey
|
||||
FOREIGN KEY (header_id, block_number)
|
||||
REFERENCES eth.header_cids (block_hash, block_number);
|
||||
|
||||
ALTER TABLE eth.storage_cids
|
||||
ADD CONSTRAINT storage_cids_ipld_blocks_fkey
|
||||
FOREIGN KEY (cid, block_number)
|
||||
REFERENCES ipld.blocks (key, block_number);
|
||||
|
||||
ALTER TABLE eth.storage_cids
|
||||
ADD CONSTRAINT storage_cids_state_cids_fkey
|
||||
FOREIGN KEY (state_leaf_key, header_id, block_number)
|
||||
REFERENCES eth.state_cids (state_leaf_key, header_id, block_number);
|
||||
|
||||
ALTER TABLE eth.log_cids
|
||||
ADD CONSTRAINT log_cids_ipld_blocks_fkey
|
||||
FOREIGN KEY (cid, block_number)
|
||||
REFERENCES ipld.blocks (key, block_number);
|
||||
|
||||
ALTER TABLE eth.log_cids
|
||||
ADD CONSTRAINT log_cids_receipt_cids_fkey
|
||||
FOREIGN KEY (rct_id, header_id, block_number)
|
||||
REFERENCES eth.receipt_cids (tx_id, header_id, block_number);
|
||||
|
||||
-- +goose Down
|
||||
ALTER TABLE eth.log_cids
|
||||
DROP CONSTRAINT log_cids_receipt_cids_fkey;
|
||||
|
||||
ALTER TABLE eth.log_cids
|
||||
DROP CONSTRAINT log_cids_ipld_blocks_fkey;
|
||||
|
||||
ALTER TABLE eth.storage_cids
|
||||
DROP CONSTRAINT storage_cids_state_cids_fkey;
|
||||
|
||||
ALTER TABLE eth.storage_cids
|
||||
DROP CONSTRAINT storage_cids_ipld_blocks_fkey;
|
||||
|
||||
ALTER TABLE eth.state_cids
|
||||
DROP CONSTRAINT state_cids_header_cids_fkey;
|
||||
|
||||
ALTER TABLE eth.state_cids
|
||||
DROP CONSTRAINT state_cids_ipld_blocks_fkey;
|
||||
|
||||
ALTER TABLE eth.receipt_cids
|
||||
DROP CONSTRAINT receipt_cids_transaction_cids_fkey;
|
||||
|
||||
ALTER TABLE eth.receipt_cids
|
||||
DROP CONSTRAINT receipt_cids_ipld_blocks_fkey;
|
||||
|
||||
ALTER TABLE eth.transaction_cids
|
||||
DROP CONSTRAINT transaction_cids_header_cids_fkey;
|
||||
|
||||
ALTER TABLE eth.transaction_cids
|
||||
DROP CONSTRAINT transaction_cids_ipld_blocks_fkey;
|
||||
|
||||
ALTER TABLE eth.uncle_cids
|
||||
DROP CONSTRAINT uncle_cids_header_cids_fkey;
|
||||
|
||||
ALTER TABLE eth.uncle_cids
|
||||
DROP CONSTRAINT uncle_cids_ipld_blocks_fkey;
|
||||
|
||||
ALTER TABLE eth.header_cids
|
||||
DROP CONSTRAINT header_cids_ipld_blocks_fkey;
|
@ -1,10 +0,0 @@
|
||||
-- +goose Up
|
||||
CREATE TABLE eth_meta.known_gaps (
|
||||
starting_block_number bigint PRIMARY KEY,
|
||||
ending_block_number bigint,
|
||||
checked_out boolean,
|
||||
processing_key bigint
|
||||
);
|
||||
|
||||
-- +goose Down
|
||||
DROP TABLE eth_meta.known_gaps;
|
105
db/migrations/00019_drop_foreign_keys.sql
Normal file
105
db/migrations/00019_drop_foreign_keys.sql
Normal file
@ -0,0 +1,105 @@
|
||||
-- +goose Up
|
||||
ALTER TABLE eth.log_cids
|
||||
DROP CONSTRAINT log_cids_receipt_cids_fkey;
|
||||
|
||||
ALTER TABLE eth.log_cids
|
||||
DROP CONSTRAINT log_cids_ipld_blocks_fkey;
|
||||
|
||||
ALTER TABLE eth.storage_cids
|
||||
DROP CONSTRAINT storage_cids_state_cids_fkey;
|
||||
|
||||
ALTER TABLE eth.storage_cids
|
||||
DROP CONSTRAINT storage_cids_ipld_blocks_fkey;
|
||||
|
||||
ALTER TABLE eth.state_cids
|
||||
DROP CONSTRAINT state_cids_header_cids_fkey;
|
||||
|
||||
ALTER TABLE eth.state_cids
|
||||
DROP CONSTRAINT state_cids_ipld_blocks_fkey;
|
||||
|
||||
ALTER TABLE eth.receipt_cids
|
||||
DROP CONSTRAINT receipt_cids_transaction_cids_fkey;
|
||||
|
||||
ALTER TABLE eth.receipt_cids
|
||||
DROP CONSTRAINT receipt_cids_ipld_blocks_fkey;
|
||||
|
||||
ALTER TABLE eth.transaction_cids
|
||||
DROP CONSTRAINT transaction_cids_header_cids_fkey;
|
||||
|
||||
ALTER TABLE eth.transaction_cids
|
||||
DROP CONSTRAINT transaction_cids_ipld_blocks_fkey;
|
||||
|
||||
ALTER TABLE eth.uncle_cids
|
||||
DROP CONSTRAINT uncle_cids_header_cids_fkey;
|
||||
|
||||
ALTER TABLE eth.uncle_cids
|
||||
DROP CONSTRAINT uncle_cids_ipld_blocks_fkey;
|
||||
|
||||
ALTER TABLE eth.header_cids
|
||||
DROP CONSTRAINT header_cids_ipld_blocks_fkey;
|
||||
|
||||
-- +goose Down
|
||||
ALTER TABLE eth.header_cids
|
||||
ADD CONSTRAINT header_cids_ipld_blocks_fkey
|
||||
FOREIGN KEY (cid, block_number)
|
||||
REFERENCES ipld.blocks (key, block_number);
|
||||
|
||||
ALTER TABLE eth.uncle_cids
|
||||
ADD CONSTRAINT uncle_cids_ipld_blocks_fkey
|
||||
FOREIGN KEY (cid, block_number)
|
||||
REFERENCES ipld.blocks (key, block_number);
|
||||
|
||||
ALTER TABLE eth.uncle_cids
|
||||
ADD CONSTRAINT uncle_cids_header_cids_fkey
|
||||
FOREIGN KEY (header_id, block_number)
|
||||
REFERENCES eth.header_cids (block_hash, block_number);
|
||||
|
||||
ALTER TABLE eth.transaction_cids
|
||||
ADD CONSTRAINT transaction_cids_ipld_blocks_fkey
|
||||
FOREIGN KEY (cid, block_number)
|
||||
REFERENCES ipld.blocks (key, block_number);
|
||||
|
||||
ALTER TABLE eth.transaction_cids
|
||||
ADD CONSTRAINT transaction_cids_header_cids_fkey
|
||||
FOREIGN KEY (header_id, block_number)
|
||||
REFERENCES eth.header_cids (block_hash, block_number);
|
||||
|
||||
ALTER TABLE eth.receipt_cids
|
||||
ADD CONSTRAINT receipt_cids_ipld_blocks_fkey
|
||||
FOREIGN KEY (cid, block_number)
|
||||
REFERENCES ipld.blocks (key, block_number);
|
||||
|
||||
ALTER TABLE eth.receipt_cids
|
||||
ADD CONSTRAINT receipt_cids_transaction_cids_fkey
|
||||
FOREIGN KEY (tx_id, header_id, block_number)
|
||||
REFERENCES eth.transaction_cids (tx_hash, header_id, block_number);
|
||||
|
||||
ALTER TABLE eth.state_cids
|
||||
ADD CONSTRAINT state_cids_ipld_blocks_fkey
|
||||
FOREIGN KEY (cid, block_number)
|
||||
REFERENCES ipld.blocks (key, block_number);
|
||||
|
||||
ALTER TABLE eth.state_cids
|
||||
ADD CONSTRAINT state_cids_header_cids_fkey
|
||||
FOREIGN KEY (header_id, block_number)
|
||||
REFERENCES eth.header_cids (block_hash, block_number);
|
||||
|
||||
ALTER TABLE eth.storage_cids
|
||||
ADD CONSTRAINT storage_cids_ipld_blocks_fkey
|
||||
FOREIGN KEY (cid, block_number)
|
||||
REFERENCES ipld.blocks (key, block_number);
|
||||
|
||||
ALTER TABLE eth.storage_cids
|
||||
ADD CONSTRAINT storage_cids_state_cids_fkey
|
||||
FOREIGN KEY (state_leaf_key, header_id, block_number)
|
||||
REFERENCES eth.state_cids (state_leaf_key, header_id, block_number);
|
||||
|
||||
ALTER TABLE eth.log_cids
|
||||
ADD CONSTRAINT log_cids_ipld_blocks_fkey
|
||||
FOREIGN KEY (cid, block_number)
|
||||
REFERENCES ipld.blocks (key, block_number);
|
||||
|
||||
ALTER TABLE eth.log_cids
|
||||
ADD CONSTRAINT log_cids_receipt_cids_fkey
|
||||
FOREIGN KEY (rct_id, header_id, block_number)
|
||||
REFERENCES eth.receipt_cids (tx_id, header_id, block_number);
|
98
db/migrations/00020_convert_to_hypertables.sql
Normal file
98
db/migrations/00020_convert_to_hypertables.sql
Normal file
@ -0,0 +1,98 @@
|
||||
-- +goose Up
|
||||
SELECT create_hypertable('ipld.blocks', 'block_number', migrate_data => true, chunk_time_interval => 32768);
|
||||
SELECT create_hypertable('eth.uncle_cids', 'block_number', migrate_data => true, chunk_time_interval => 32768);
|
||||
SELECT create_hypertable('eth.transaction_cids', 'block_number', migrate_data => true, chunk_time_interval => 32768);
|
||||
SELECT create_hypertable('eth.receipt_cids', 'block_number', migrate_data => true, chunk_time_interval => 32768);
|
||||
SELECT create_hypertable('eth.state_cids', 'block_number', migrate_data => true, chunk_time_interval => 32768);
|
||||
SELECT create_hypertable('eth.storage_cids', 'block_number', migrate_data => true, chunk_time_interval => 32768);
|
||||
SELECT create_hypertable('eth.log_cids', 'block_number', migrate_data => true, chunk_time_interval => 32768);
|
||||
|
||||
-- update version
|
||||
INSERT INTO public.db_version (singleton, version) VALUES (true, 'v5.0.0-h')
|
||||
ON CONFLICT (singleton) DO UPDATE SET (version, tstamp) = ('v5.0.0-h', NOW());
|
||||
|
||||
-- +goose Down
|
||||
INSERT INTO public.db_version (singleton, version) VALUES (true, 'v5.0.0')
|
||||
ON CONFLICT (singleton) DO UPDATE SET (version, tstamp) = ('v5.0.0', NOW());
|
||||
|
||||
-- reversing conversion to hypertable requires migrating all data from every chunk back to a single table
|
||||
-- create new regular tables
|
||||
CREATE TABLE eth.log_cids_i (LIKE eth.log_cids INCLUDING ALL);
|
||||
CREATE TABLE eth.storage_cids_i (LIKE eth.storage_cids INCLUDING ALL);
|
||||
CREATE TABLE eth.state_cids_i (LIKE eth.state_cids INCLUDING ALL);
|
||||
CREATE TABLE eth.receipt_cids_i (LIKE eth.receipt_cids INCLUDING ALL);
|
||||
CREATE TABLE eth.transaction_cids_i (LIKE eth.transaction_cids INCLUDING ALL);
|
||||
CREATE TABLE eth.uncle_cids_i (LIKE eth.uncle_cids INCLUDING ALL);
|
||||
CREATE TABLE ipld.blocks_i (LIKE ipld.blocks INCLUDING ALL);
|
||||
|
||||
-- migrate data
|
||||
INSERT INTO eth.log_cids_i (SELECT * FROM eth.log_cids);
|
||||
INSERT INTO eth.storage_cids_i (SELECT * FROM eth.storage_cids);
|
||||
INSERT INTO eth.state_cids_i (SELECT * FROM eth.state_cids);
|
||||
INSERT INTO eth.receipt_cids_i (SELECT * FROM eth.receipt_cids);
|
||||
INSERT INTO eth.transaction_cids_i (SELECT * FROM eth.transaction_cids);
|
||||
INSERT INTO eth.uncle_cids_i (SELECT * FROM eth.uncle_cids);
|
||||
INSERT INTO ipld.blocks_i (SELECT * FROM ipld.blocks);
|
||||
|
||||
-- drop hypertables
|
||||
DROP TABLE eth.log_cids;
|
||||
DROP TABLE eth.storage_cids;
|
||||
DROP TABLE eth.state_cids;
|
||||
DROP TABLE eth.receipt_cids;
|
||||
DROP TABLE eth.transaction_cids;
|
||||
DROP TABLE eth.uncle_cids;
|
||||
DROP TABLE ipld.blocks;
|
||||
|
||||
-- rename new tables
|
||||
ALTER TABLE eth.log_cids_i RENAME TO log_cids;
|
||||
ALTER TABLE eth.storage_cids_i RENAME TO storage_cids;
|
||||
ALTER TABLE eth.state_cids_i RENAME TO state_cids;
|
||||
ALTER TABLE eth.receipt_cids_i RENAME TO receipt_cids;
|
||||
ALTER TABLE eth.transaction_cids_i RENAME TO transaction_cids;
|
||||
ALTER TABLE eth.uncle_cids_i RENAME TO uncle_cids;
|
||||
ALTER TABLE ipld.blocks_i RENAME TO blocks;
|
||||
|
||||
-- rename indexes:
|
||||
-- log indexes
|
||||
ALTER INDEX eth.log_cids_i_topic3_idx RENAME TO log_topic3_index;
|
||||
ALTER INDEX eth.log_cids_i_topic2_idx RENAME TO log_topic2_index;
|
||||
ALTER INDEX eth.log_cids_i_topic1_idx RENAME TO log_topic1_index;
|
||||
ALTER INDEX eth.log_cids_i_topic0_idx RENAME TO log_topic0_index;
|
||||
ALTER INDEX eth.log_cids_i_address_idx RENAME TO log_address_index;
|
||||
ALTER INDEX eth.log_cids_i_cid_block_number_idx RENAME TO log_cid_block_number_index;
|
||||
ALTER INDEX eth.log_cids_i_header_id_idx RENAME TO log_header_id_index;
|
||||
ALTER INDEX eth.log_cids_i_block_number_idx RENAME TO log_block_number_index;
|
||||
|
||||
-- storage node indexes -- storage node indexes
|
||||
ALTER INDEX eth.storage_cids_i_removed_idx RENAME TO storage_removed_index;
|
||||
ALTER INDEX eth.storage_cids_i_header_id_idx RENAME TO storage_header_id_index;
|
||||
ALTER INDEX eth.storage_cids_i_cid_block_number_idx RENAME TO storage_cid_block_number_index;
|
||||
ALTER INDEX eth.storage_cids_i_state_leaf_key_idx RENAME TO storage_state_leaf_key_index;
|
||||
ALTER INDEX eth.storage_cids_i_block_number_idx RENAME TO storage_block_number_index;
|
||||
ALTER INDEX eth.storage_cids_i_storage_leaf_key_block_number_idx RENAME TO storage_leaf_key_block_number_index;
|
||||
|
||||
-- state node indexes -- state node indexes
|
||||
ALTER INDEX eth.state_cids_i_code_hash_idx RENAME TO state_code_hash_index;
|
||||
ALTER INDEX eth.state_cids_i_removed_idx RENAME TO state_removed_index;
|
||||
ALTER INDEX eth.state_cids_i_header_id_idx RENAME TO state_header_id_index;
|
||||
ALTER INDEX eth.state_cids_i_cid_block_number_idx RENAME TO state_cid_block_number_index;
|
||||
ALTER INDEX eth.state_cids_i_block_number_idx RENAME TO state_block_number_index;
|
||||
ALTER INDEX eth.state_cids_i_state_leaf_key_block_number_idx RENAME TO state_leaf_key_block_number_index;
|
||||
|
||||
-- receipt indexes -- receipt indexes
|
||||
ALTER INDEX eth.receipt_cids_i_contract_idx RENAME TO rct_contract_index;
|
||||
ALTER INDEX eth.receipt_cids_i_cid_block_number_idx RENAME TO rct_cid_block_number_index;
|
||||
ALTER INDEX eth.receipt_cids_i_header_id_idx RENAME TO rct_header_id_index;
|
||||
ALTER INDEX eth.receipt_cids_i_block_number_idx RENAME TO rct_block_number_index;
|
||||
|
||||
-- transaction indexes -- transaction indexes
|
||||
ALTER INDEX eth.transaction_cids_i_src_idx RENAME TO tx_src_index;
|
||||
ALTER INDEX eth.transaction_cids_i_dst_idx RENAME TO tx_dst_index;
|
||||
ALTER INDEX eth.transaction_cids_i_cid_block_number_idx RENAME TO tx_cid_block_number_index;
|
||||
ALTER INDEX eth.transaction_cids_i_header_id_idx RENAME TO tx_header_id_index;
|
||||
ALTER INDEX eth.transaction_cids_i_block_number_idx RENAME TO tx_block_number_index;
|
||||
|
||||
-- uncle indexes -- uncle indexes
|
||||
ALTER INDEX eth.uncle_cids_i_block_number_idx RENAME TO uncle_block_number_index;
|
||||
ALTER INDEX eth.uncle_cids_i_cid_block_number_index_idx RENAME TO uncle_cid_block_number_index;
|
||||
ALTER INDEX eth.uncle_cids_i_header_id_idx RENAME TO uncle_header_id_index;
|
@ -1,6 +0,0 @@
|
||||
-- +goose Up
|
||||
INSERT INTO public.db_version (singleton, version) VALUES (true, 'v4.0.0')
|
||||
ON CONFLICT (singleton) DO UPDATE SET (version, tstamp) = ('v4.0.0', NOW());
|
||||
|
||||
-- +goose Down
|
||||
DELETE FROM public.db_version WHERE version = 'v4.0.0';
|
@ -1,68 +0,0 @@
|
||||
-- +goose Up
|
||||
SELECT create_hypertable('public.blocks', 'block_number', migrate_data => true, chunk_time_interval => 32768);
|
||||
SELECT create_hypertable('eth.header_cids', 'block_number', migrate_data => true, chunk_time_interval => 32768);
|
||||
SELECT create_hypertable('eth.uncle_cids', 'block_number', migrate_data => true, chunk_time_interval => 32768);
|
||||
SELECT create_hypertable('eth.transaction_cids', 'block_number', migrate_data => true, chunk_time_interval => 32768);
|
||||
SELECT create_hypertable('eth.receipt_cids', 'block_number', migrate_data => true, chunk_time_interval => 32768);
|
||||
SELECT create_hypertable('eth.state_cids', 'block_number', migrate_data => true, chunk_time_interval => 32768);
|
||||
SELECT create_hypertable('eth.storage_cids', 'block_number', migrate_data => true, chunk_time_interval => 32768);
|
||||
SELECT create_hypertable('eth.state_accounts', 'block_number', migrate_data => true, chunk_time_interval => 32768);
|
||||
SELECT create_hypertable('eth.access_list_elements', 'block_number', migrate_data => true, chunk_time_interval => 32768);
|
||||
SELECT create_hypertable('eth.log_cids', 'block_number', migrate_data => true, chunk_time_interval => 32768);
|
||||
|
||||
-- update version
|
||||
INSERT INTO public.db_version (singleton, version) VALUES (true, 'v4.0.0-h')
|
||||
ON CONFLICT (singleton) DO UPDATE SET (version, tstamp) = ('v4.0.0-h', NOW());
|
||||
|
||||
-- +goose Down
|
||||
INSERT INTO public.db_version (singleton, version) VALUES (true, 'v4.0.0')
|
||||
ON CONFLICT (singleton) DO UPDATE SET (version, tstamp) = ('v4.0.0', NOW());
|
||||
|
||||
-- reversing conversion to hypertable requires migrating all data from every chunk back to a single table
|
||||
-- create new regular tables
|
||||
CREATE TABLE eth.log_cids_i (LIKE eth.log_cids INCLUDING ALL);
|
||||
CREATE TABLE eth.access_list_elements_i (LIKE eth.access_list_elements INCLUDING ALL);
|
||||
CREATE TABLE eth.state_accounts_i (LIKE eth.state_accounts INCLUDING ALL);
|
||||
CREATE TABLE eth.storage_cids_i (LIKE eth.storage_cids INCLUDING ALL);
|
||||
CREATE TABLE eth.state_cids_i (LIKE eth.state_cids INCLUDING ALL);
|
||||
CREATE TABLE eth.receipt_cids_i (LIKE eth.receipt_cids INCLUDING ALL);
|
||||
CREATE TABLE eth.transaction_cids_i (LIKE eth.transaction_cids INCLUDING ALL);
|
||||
CREATE TABLE eth.uncle_cids_i (LIKE eth.uncle_cids INCLUDING ALL);
|
||||
CREATE TABLE eth.header_cids_i (LIKE eth.header_cids INCLUDING ALL);
|
||||
CREATE TABLE public.blocks_i (LIKE public.blocks INCLUDING ALL);
|
||||
|
||||
-- migrate data
|
||||
INSERT INTO eth.log_cids_i (SELECT * FROM eth.log_cids);
|
||||
INSERT INTO eth.access_list_elements_i (SELECT * FROM eth.access_list_elements);
|
||||
INSERT INTO eth.state_accounts_i (SELECT * FROM eth.state_accounts);
|
||||
INSERT INTO eth.storage_cids_i (SELECT * FROM eth.storage_cids);
|
||||
INSERT INTO eth.state_cids_i (SELECT * FROM eth.state_cids);
|
||||
INSERT INTO eth.receipt_cids_i (SELECT * FROM eth.receipt_cids);
|
||||
INSERT INTO eth.transaction_cids_i (SELECT * FROM eth.transaction_cids);
|
||||
INSERT INTO eth.uncle_cids_i (SELECT * FROM eth.uncle_cids);
|
||||
INSERT INTO eth.header_cids_i (SELECT * FROM eth.header_cids);
|
||||
INSERT INTO public.blocks_i (SELECT * FROM public.blocks);
|
||||
|
||||
-- drop hypertables
|
||||
DROP TABLE eth.log_cids;
|
||||
DROP TABLE eth.access_list_elements;
|
||||
DROP TABLE eth.state_accounts;
|
||||
DROP TABLE eth.storage_cids;
|
||||
DROP TABLE eth.state_cids;
|
||||
DROP TABLE eth.receipt_cids;
|
||||
DROP TABLE eth.transaction_cids;
|
||||
DROP TABLE eth.uncle_cids;
|
||||
DROP TABLE eth.header_cids;
|
||||
DROP TABLE public.blocks;
|
||||
|
||||
-- rename new tables
|
||||
ALTER TABLE eth.log_cids_i RENAME TO log_cids;
|
||||
ALTER TABLE eth.access_list_elements_i RENAME TO access_list_elements;
|
||||
ALTER TABLE eth.state_accounts_i RENAME TO state_accounts;
|
||||
ALTER TABLE eth.storage_cids_i RENAME TO storage_cids;
|
||||
ALTER TABLE eth.state_cids_i RENAME TO state_cids;
|
||||
ALTER TABLE eth.receipt_cids_i RENAME TO receipt_cids;
|
||||
ALTER TABLE eth.transaction_cids_i RENAME TO transaction_cids;
|
||||
ALTER TABLE eth.uncle_cids_i RENAME TO uncle_cids;
|
||||
ALTER TABLE eth.header_cids_i RENAME TO header_cids;
|
||||
ALTER TABLE public.blocks_i RENAME TO blocks;
|
16
db/migrations/00021_create_eth_withdrawal_cids_table.sql
Normal file
16
db/migrations/00021_create_eth_withdrawal_cids_table.sql
Normal file
@ -0,0 +1,16 @@
|
||||
-- +goose Up
|
||||
CREATE TABLE IF NOT EXISTS eth.withdrawal_cids (
|
||||
block_number BIGINT NOT NULL,
|
||||
header_id VARCHAR(66) NOT NULL,
|
||||
cid TEXT NOT NULL,
|
||||
index INTEGER NOT NULL,
|
||||
validator INTEGER NOT NULL,
|
||||
address VARCHAR(66) NOT NULL,
|
||||
amount NUMERIC NOT NULL,
|
||||
PRIMARY KEY (index, header_id, block_number)
|
||||
);
|
||||
|
||||
SELECT create_hypertable('eth.withdrawal_cids', 'block_number', migrate_data => true, chunk_time_interval => 32768);
|
||||
|
||||
-- +goose Down
|
||||
DROP TABLE eth.withdrawal_cids;
|
12
db/migrations/00022_create_eth_blob_hash_cids_table.sql
Normal file
12
db/migrations/00022_create_eth_blob_hash_cids_table.sql
Normal file
@ -0,0 +1,12 @@
|
||||
-- +goose Up
|
||||
CREATE TABLE eth.blob_hashes (
|
||||
tx_hash VARCHAR(66) NOT NULL,
|
||||
index INTEGER NOT NULL,
|
||||
blob_hash BYTEA NOT NULL
|
||||
);
|
||||
|
||||
CREATE UNIQUE INDEX blob_hashes_tx_hash_index ON eth.blob_hashes(tx_hash, index);
|
||||
|
||||
-- +goose Down
|
||||
DROP INDEX eth.blob_hashes_tx_hash_index;
|
||||
DROP TABLE eth.blob_hashes;
|
@ -1,248 +0,0 @@
|
||||
-- +goose Up
|
||||
-- +goose StatementBegin
|
||||
-- returns if a state leaf node was removed within the provided block number
|
||||
CREATE OR REPLACE FUNCTION was_state_leaf_removed(key character varying, hash character varying)
|
||||
RETURNS boolean AS $$
|
||||
SELECT state_cids.node_type = 3
|
||||
FROM eth.state_cids
|
||||
INNER JOIN eth.header_cids ON (state_cids.header_id = header_cids.block_hash)
|
||||
WHERE state_leaf_key = key
|
||||
AND state_cids.block_number <= (SELECT block_number
|
||||
FROM eth.header_cids
|
||||
WHERE block_hash = hash)
|
||||
ORDER BY state_cids.block_number DESC LIMIT 1;
|
||||
$$
|
||||
language sql;
|
||||
-- +goose StatementEnd
|
||||
|
||||
-- +goose StatementBegin
|
||||
CREATE TYPE child_result AS (
|
||||
has_child BOOLEAN,
|
||||
children eth.header_cids[]
|
||||
);
|
||||
|
||||
CREATE OR REPLACE FUNCTION has_child(hash VARCHAR(66), height BIGINT) RETURNS child_result AS
|
||||
$BODY$
|
||||
DECLARE
|
||||
child_height INT;
|
||||
temp_child eth.header_cids;
|
||||
new_child_result child_result;
|
||||
BEGIN
|
||||
child_height = height + 1;
|
||||
-- short circuit if there are no children
|
||||
SELECT exists(SELECT 1
|
||||
FROM eth.header_cids
|
||||
WHERE parent_hash = hash
|
||||
AND block_number = child_height
|
||||
LIMIT 1)
|
||||
INTO new_child_result.has_child;
|
||||
-- collect all the children for this header
|
||||
IF new_child_result.has_child THEN
|
||||
FOR temp_child IN
|
||||
SELECT * FROM eth.header_cids WHERE parent_hash = hash AND block_number = child_height
|
||||
LOOP
|
||||
new_child_result.children = array_append(new_child_result.children, temp_child);
|
||||
END LOOP;
|
||||
END IF;
|
||||
RETURN new_child_result;
|
||||
END
|
||||
$BODY$
|
||||
LANGUAGE 'plpgsql';
|
||||
-- +goose StatementEnd
|
||||
|
||||
-- +goose StatementBegin
|
||||
CREATE OR REPLACE FUNCTION canonical_header_from_array(headers eth.header_cids[]) RETURNS eth.header_cids AS
|
||||
$BODY$
|
||||
DECLARE
|
||||
canonical_header eth.header_cids;
|
||||
canonical_child eth.header_cids;
|
||||
header eth.header_cids;
|
||||
current_child_result child_result;
|
||||
child_headers eth.header_cids[];
|
||||
current_header_with_child eth.header_cids;
|
||||
has_children_count INT DEFAULT 0;
|
||||
BEGIN
|
||||
-- for each header in the provided set
|
||||
FOREACH header IN ARRAY headers
|
||||
LOOP
|
||||
-- check if it has any children
|
||||
current_child_result = has_child(header.block_hash, header.block_number);
|
||||
IF current_child_result.has_child THEN
|
||||
-- if it does, take note
|
||||
has_children_count = has_children_count + 1;
|
||||
current_header_with_child = header;
|
||||
-- and add the children to the growing set of child headers
|
||||
child_headers = array_cat(child_headers, current_child_result.children);
|
||||
END IF;
|
||||
END LOOP;
|
||||
-- if none of the headers had children, none is more canonical than the other
|
||||
IF has_children_count = 0 THEN
|
||||
-- return the first one selected
|
||||
SELECT * INTO canonical_header FROM unnest(headers) LIMIT 1;
|
||||
-- if only one header had children, it can be considered the heaviest/canonical header of the set
|
||||
ELSIF has_children_count = 1 THEN
|
||||
-- return the only header with a child
|
||||
canonical_header = current_header_with_child;
|
||||
-- if there are multiple headers with children
|
||||
ELSE
|
||||
-- find the canonical header from the child set
|
||||
canonical_child = canonical_header_from_array(child_headers);
|
||||
-- the header that is parent to this header, is the canonical header at this level
|
||||
SELECT * INTO canonical_header FROM unnest(headers)
|
||||
WHERE block_hash = canonical_child.parent_hash;
|
||||
END IF;
|
||||
RETURN canonical_header;
|
||||
END
|
||||
$BODY$
|
||||
LANGUAGE 'plpgsql';
|
||||
-- +goose StatementEnd
|
||||
|
||||
-- +goose StatementBegin
|
||||
CREATE OR REPLACE FUNCTION canonical_header_hash(height BIGINT) RETURNS character varying AS
|
||||
$BODY$
|
||||
DECLARE
|
||||
canonical_header eth.header_cids;
|
||||
headers eth.header_cids[];
|
||||
header_count INT;
|
||||
temp_header eth.header_cids;
|
||||
BEGIN
|
||||
-- collect all headers at this height
|
||||
FOR temp_header IN
|
||||
SELECT * FROM eth.header_cids WHERE block_number = height
|
||||
LOOP
|
||||
headers = array_append(headers, temp_header);
|
||||
END LOOP;
|
||||
-- count the number of headers collected
|
||||
header_count = array_length(headers, 1);
|
||||
-- if we have less than 1 header, return NULL
|
||||
IF header_count IS NULL OR header_count < 1 THEN
|
||||
RETURN NULL;
|
||||
-- if we have one header, return its hash
|
||||
ELSIF header_count = 1 THEN
|
||||
RETURN headers[1].block_hash;
|
||||
-- if we have multiple headers we need to determine which one is canonical
|
||||
ELSE
|
||||
canonical_header = canonical_header_from_array(headers);
|
||||
RETURN canonical_header.block_hash;
|
||||
END IF;
|
||||
END
|
||||
$BODY$
|
||||
LANGUAGE 'plpgsql';
|
||||
-- +goose StatementEnd
|
||||
|
||||
-- +goose StatementBegin
|
||||
CREATE TYPE state_node_result AS (
|
||||
data BYTEA,
|
||||
state_leaf_key VARCHAR(66),
|
||||
cid TEXT,
|
||||
state_path BYTEA,
|
||||
node_type INTEGER,
|
||||
mh_key TEXT
|
||||
);
|
||||
-- +goose StatementEnd
|
||||
|
||||
-- +goose StatementBegin
|
||||
CREATE OR REPLACE FUNCTION state_snapshot(starting_height BIGINT, ending_height BIGINT) RETURNS void AS
|
||||
$BODY$
|
||||
DECLARE
|
||||
canonical_hash VARCHAR(66);
|
||||
results state_node_result[];
|
||||
BEGIN
|
||||
-- get the canonical hash for the header at ending_height
|
||||
canonical_hash = canonical_header_hash(ending_height);
|
||||
IF canonical_hash IS NULL THEN
|
||||
RAISE EXCEPTION 'cannot create state snapshot, no header can be found at height %', ending_height;
|
||||
END IF;
|
||||
|
||||
-- select all of the state nodes for this snapshot: the latest state node record at every unique path
|
||||
SELECT ARRAY (SELECT DISTINCT ON (state_path) ROW (blocks.data, state_cids.state_leaf_key, state_cids.cid, state_cids.state_path,
|
||||
state_cids.node_type, state_cids.mh_key)
|
||||
FROM eth.state_cids
|
||||
INNER JOIN public.blocks
|
||||
ON (state_cids.mh_key, state_cids.block_number) = (blocks.key, blocks.block_number)
|
||||
WHERE state_cids.block_number BETWEEN starting_height AND ending_height
|
||||
ORDER BY state_path, state_cids.block_number DESC)
|
||||
INTO results;
|
||||
|
||||
-- from the set returned above, insert public.block records at the ending_height block number
|
||||
INSERT INTO public.blocks (block_number, key, data)
|
||||
SELECT ending_height, r.mh_key, r.data
|
||||
FROM unnest(results) r;
|
||||
|
||||
-- from the set returned above, insert eth.state_cids records at the ending_height block number
|
||||
-- anchoring all the records to the canonical header found at ending_height
|
||||
INSERT INTO eth.state_cids (block_number, header_id, state_leaf_key, cid, state_path, node_type, diff, mh_key)
|
||||
SELECT ending_height, canonical_hash, r.state_leaf_key, r.cid, r.state_path, r.node_type, false, r.mh_key
|
||||
FROM unnest(results) r
|
||||
ON CONFLICT (state_path, header_id, block_number) DO NOTHING;
|
||||
END
|
||||
$BODY$
|
||||
LANGUAGE 'plpgsql';
|
||||
-- +goose StatementEnd
|
||||
|
||||
-- +goose StatementBegin
|
||||
CREATE TYPE storage_node_result AS (
|
||||
data BYTEA,
|
||||
state_path BYTEA,
|
||||
storage_leaf_key VARCHAR(66),
|
||||
cid TEXT,
|
||||
storage_path BYTEA,
|
||||
node_type INTEGER,
|
||||
mh_key TEXT
|
||||
);
|
||||
-- +goose StatementEnd
|
||||
|
||||
-- +goose StatementBegin
|
||||
-- this should only be ran after a state_snapshot has been completed
|
||||
-- this should probably be rolled together with state_snapshot into a single procedure...
|
||||
CREATE OR REPLACE FUNCTION storage_snapshot(starting_height BIGINT, ending_height BIGINT) RETURNS void AS
|
||||
$BODY$
|
||||
DECLARE
|
||||
canonical_hash VARCHAR(66);
|
||||
results storage_node_result[];
|
||||
BEGIN
|
||||
-- get the canonical hash for the header at ending_height
|
||||
SELECT canonical_header_hash(ending_height) INTO canonical_hash;
|
||||
IF canonical_hash IS NULL THEN
|
||||
RAISE EXCEPTION 'cannot create state snapshot, no header can be found at height %', ending_height;
|
||||
END IF;
|
||||
|
||||
-- select all of the storage nodes for this snapshot: the latest storage node record at every unique state leaf key
|
||||
SELECT ARRAY (SELECT DISTINCT ON (state_leaf_key, storage_path) ROW (blocks.data, storage_cids.state_path, storage_cids.storage_leaf_key,
|
||||
storage_cids.cid, storage_cids.storage_path, storage_cids.node_type, storage_cids.mh_key)
|
||||
FROM eth.storage_cids
|
||||
INNER JOIN public.blocks
|
||||
ON (storage_cids.mh_key, storage_cids.block_number) = (blocks.key, blocks.block_number)
|
||||
INNER JOIN eth.state_cids
|
||||
ON (storage_cids.state_path, storage_cids.header_id) = (state_cids.state_path, state_cids.header_id)
|
||||
WHERE storage_cids.block_number BETWEEN starting_height AND ending_height
|
||||
ORDER BY state_leaf_key, storage_path, storage_cids.state_path, storage_cids.block_number DESC)
|
||||
INTO results;
|
||||
|
||||
-- from the set returned above, insert public.block records at the ending_height block number
|
||||
INSERT INTO public.blocks (block_number, key, data)
|
||||
SELECT ending_height, r.mh_key, r.data
|
||||
FROM unnest(results) r;
|
||||
|
||||
-- from the set returned above, insert eth.state_cids records at the ending_height block number
|
||||
-- anchoring all the records to the canonical header found at ending_height
|
||||
INSERT INTO eth.storage_cids (block_number, header_id, state_path, storage_leaf_key, cid, storage_path,
|
||||
node_type, diff, mh_key)
|
||||
SELECT ending_height, canonical_hash, r.state_path, r.storage_leaf_key, r.cid, r.storage_path, r.node_type, false, r.mh_key
|
||||
FROM unnest(results) r
|
||||
ON CONFLICT (storage_path, state_path, header_id, block_number) DO NOTHING;
|
||||
END
|
||||
$BODY$
|
||||
LANGUAGE 'plpgsql';
|
||||
-- +goose StatementEnd
|
||||
|
||||
-- +goose Down
|
||||
DROP FUNCTION storage_snapshot;
|
||||
DROP TYPE storage_node_result;
|
||||
DROP FUNCTION state_snapshot;
|
||||
DROP TYPE state_node_result;
|
||||
DROP FUNCTION was_state_leaf_removed;
|
||||
DROP FUNCTION canonical_header_hash;
|
||||
DROP FUNCTION canonical_header_from_array;
|
||||
DROP FUNCTION has_child;
|
||||
DROP TYPE child_result;
|
@ -1,27 +0,0 @@
|
||||
version: '3.2'
|
||||
|
||||
services:
|
||||
migrations:
|
||||
restart: on-failure
|
||||
depends_on:
|
||||
- ipld-eth-db
|
||||
image: cerc-io/ipld-eth-db
|
||||
# Build image using local context
|
||||
build: .
|
||||
environment:
|
||||
DATABASE_USER: "vdbm"
|
||||
DATABASE_NAME: "vulcanize_testing"
|
||||
DATABASE_PASSWORD: "password"
|
||||
DATABASE_HOSTNAME: "ipld-eth-db"
|
||||
DATABASE_PORT: 5432
|
||||
|
||||
ipld-eth-db:
|
||||
image: timescale/timescaledb:latest-pg14
|
||||
restart: always
|
||||
command: ["postgres", "-c", "log_statement=all"]
|
||||
environment:
|
||||
POSTGRES_USER: "vdbm"
|
||||
POSTGRES_DB: "vulcanize_testing"
|
||||
POSTGRES_PASSWORD: "password"
|
||||
ports:
|
||||
- "127.0.0.1:8077:5432"
|
820
schema.sql
820
schema.sql
File diff suppressed because it is too large
Load Diff
@ -1,20 +1,14 @@
|
||||
#!/bin/sh
|
||||
# Runs the db migrations
|
||||
set +x
|
||||
set -e
|
||||
|
||||
# Default command is "goose up"
|
||||
if [[ $# -eq 0 ]]; then
|
||||
set -- "up"
|
||||
fi
|
||||
|
||||
# Construct the connection string for postgres
|
||||
VDB_PG_CONNECT=postgresql://$DATABASE_USER:$DATABASE_PASSWORD@$DATABASE_HOSTNAME:$DATABASE_PORT/$DATABASE_NAME?sslmode=disable
|
||||
|
||||
# Run the DB migrations
|
||||
echo "Connecting with: $VDB_PG_CONNECT"
|
||||
echo "Running database migrations"
|
||||
./goose -dir migrations/vulcanizedb postgres "$VDB_PG_CONNECT" up
|
||||
|
||||
# If the db migrations ran without err
|
||||
if [[ $? -eq 0 ]]; then
|
||||
echo "Migration process ran successfully"
|
||||
tail -f /dev/null
|
||||
else
|
||||
echo "Could not run migrations. Are the database details correct?"
|
||||
exit 1
|
||||
fi
|
||||
set -x
|
||||
exec ./goose -dir migrations postgres "$VDB_PG_CONNECT" "$@"
|
||||
|
@ -10,7 +10,7 @@ sleep 5s
|
||||
export HOST_NAME=localhost
|
||||
export PORT=8066
|
||||
export USER=vdbm
|
||||
export TEST_DB=vulcanize_testing
|
||||
export TEST_DB=cerc_testing
|
||||
export TEST_CONNECT_STRING=postgresql://$USER@$HOST_NAME:$PORT/$TEST_DB?sslmode=disable
|
||||
export PGPASSWORD=password
|
||||
|
||||
|
BIN
vulcanize_db.png
BIN
vulcanize_db.png
Binary file not shown.
Before Width: | Height: | Size: 516 KiB After Width: | Height: | Size: 508 KiB |
211
vulcanize_db.uml
211
vulcanize_db.uml
@ -1,148 +1,113 @@
|
||||
<?xml version="1.0" encoding="UTF-8"?>
|
||||
<Diagram>
|
||||
<ID>DATABASE</ID>
|
||||
<OriginalElement>407978cb-39e6-453c-b9b7-c4183a4e26ef</OriginalElement>
|
||||
<OriginalElement>86a0461b-ec84-4911-9aa2-e562b5d7b24c</OriginalElement>
|
||||
<nodes>
|
||||
<node x="553.96484375" y="196.0">407978cb-39e6-453c-b9b7-c4183a4e26ef.TABLE:vulcanize_test.eth.header_cids</node>
|
||||
<node x="93.5" y="944.0">407978cb-39e6-453c-b9b7-c4183a4e26ef.TABLE:vulcanize_test.eth.receipt_cids</node>
|
||||
<node x="884.80078125" y="966.0">407978cb-39e6-453c-b9b7-c4183a4e26ef.TABLE:vulcanize_test.eth.state_accounts</node>
|
||||
<node x="849.705078125" y="0.0">407978cb-39e6-453c-b9b7-c4183a4e26ef.TABLE:vulcanize_test.public.nodes</node>
|
||||
<node x="127.30078125" y="680.0">407978cb-39e6-453c-b9b7-c4183a4e26ef.TABLE:vulcanize_test.eth.uncle_cids</node>
|
||||
<node x="102.67578125" y="33.0">407978cb-39e6-453c-b9b7-c4183a4e26ef.TABLE:vulcanize_test.public.blocks</node>
|
||||
<node x="341.30078125" y="636.0">407978cb-39e6-453c-b9b7-c4183a4e26ef.TABLE:vulcanize_test.eth.transaction_cids</node>
|
||||
<node x="357.365234375" y="988.0">407978cb-39e6-453c-b9b7-c4183a4e26ef.TABLE:vulcanize_test.eth.access_list_elements</node>
|
||||
<node x="832.365234375" y="669.0">407978cb-39e6-453c-b9b7-c4183a4e26ef.TABLE:vulcanize_test.eth.state_cids</node>
|
||||
<node x="1433.705078125" y="0.0">407978cb-39e6-453c-b9b7-c4183a4e26ef.TABLE:vulcanize_test.public.db_version</node>
|
||||
<node x="58.0" y="1206.0">407978cb-39e6-453c-b9b7-c4183a4e26ef.TABLE:vulcanize_test.eth.log_cids</node>
|
||||
<node x="1200.705078125" y="0.0">407978cb-39e6-453c-b9b7-c4183a4e26ef.TABLE:vulcanize_test.public.goose_db_version</node>
|
||||
<node x="592.365234375" y="944.0">407978cb-39e6-453c-b9b7-c4183a4e26ef.TABLE:vulcanize_test.eth.storage_cids</node>
|
||||
<node x="561.75" y="152.0">86a0461b-ec84-4911-9aa2-e562b5d7b24c.TABLE:uml_diagram.eth.header_cids</node>
|
||||
<node x="0.0" y="1439.0">86a0461b-ec84-4911-9aa2-e562b5d7b24c.TABLE:uml_diagram.public.goose_db_version</node>
|
||||
<node x="1133.0" y="0.0">86a0461b-ec84-4911-9aa2-e562b5d7b24c.TABLE:uml_diagram.public.nodes</node>
|
||||
<node x="203.25" y="1140.0">86a0461b-ec84-4911-9aa2-e562b5d7b24c.TABLE:uml_diagram.eth.log_cids</node>
|
||||
<node x="729.5" y="603.0">86a0461b-ec84-4911-9aa2-e562b5d7b24c.TABLE:uml_diagram.eth.uncle_cids</node>
|
||||
<node x="467.25" y="0.0">86a0461b-ec84-4911-9aa2-e562b5d7b24c.TABLE:uml_diagram.ipld.blocks</node>
|
||||
<node x="500.5" y="570.0">86a0461b-ec84-4911-9aa2-e562b5d7b24c.TABLE:uml_diagram.eth.state_cids</node>
|
||||
<node x="1384.0" y="0.0">86a0461b-ec84-4911-9aa2-e562b5d7b24c.TABLE:uml_diagram.eth_meta.watched_addresses</node>
|
||||
<node x="660.6941964285713" y="878.0">86a0461b-ec84-4911-9aa2-e562b5d7b24c.TABLE:uml_diagram.eth.storage_cids</node>
|
||||
<node x="303.5" y="581.0">86a0461b-ec84-4911-9aa2-e562b5d7b24c.TABLE:uml_diagram.eth.transaction_cids</node>
|
||||
<node x="233.0" y="1439.0">86a0461b-ec84-4911-9aa2-e562b5d7b24c.TABLE:uml_diagram.public.db_version</node>
|
||||
<node x="251.5" y="889.0">86a0461b-ec84-4911-9aa2-e562b5d7b24c.TABLE:uml_diagram.eth.receipt_cids</node>
|
||||
</nodes>
|
||||
<notes />
|
||||
<edges>
|
||||
<edge source="407978cb-39e6-453c-b9b7-c4183a4e26ef.TABLE:vulcanize_test.eth.access_list_elements" target="407978cb-39e6-453c-b9b7-c4183a4e26ef.TABLE:vulcanize_test.eth.transaction_cids" relationship="REFERENCES">
|
||||
<point x="0.0" y="-61.0" />
|
||||
<point x="464.865234375" y="922.0" />
|
||||
<point x="511.30078125" y="922.0" />
|
||||
<point x="0.0" y="127.0" />
|
||||
<edge source="86a0461b-ec84-4911-9aa2-e562b5d7b24c.TABLE:uml_diagram.eth.state_cids" target="86a0461b-ec84-4911-9aa2-e562b5d7b24c.TABLE:uml_diagram.ipld.blocks" relationship="REFERENCES">
|
||||
<point x="-52.25" y="-127.0" />
|
||||
<point x="552.75" y="538.0" />
|
||||
<point x="551.25" y="538.0" />
|
||||
<point x="0.0" y="50.0" />
|
||||
</edge>
|
||||
<edge source="407978cb-39e6-453c-b9b7-c4183a4e26ef.TABLE:vulcanize_test.eth.state_cids" target="407978cb-39e6-453c-b9b7-c4183a4e26ef.TABLE:vulcanize_test.eth.header_cids" relationship="REFERENCES">
|
||||
<point x="-52.25" y="-94.0" />
|
||||
<point x="884.615234375" y="614.0" />
|
||||
<point x="700.96484375" y="614.0" />
|
||||
<point x="0.0" y="193.0" />
|
||||
<edge source="86a0461b-ec84-4911-9aa2-e562b5d7b24c.TABLE:uml_diagram.eth.receipt_cids" target="86a0461b-ec84-4911-9aa2-e562b5d7b24c.TABLE:uml_diagram.ipld.blocks" relationship="REFERENCES">
|
||||
<point x="-45.5" y="-94.0" />
|
||||
<point x="297.0" y="851.0" />
|
||||
<point x="293.0" y="851.0" />
|
||||
<point x="293.0" y="126.0" />
|
||||
<point x="551.25" y="126.0" />
|
||||
<point x="0.0" y="50.0" />
|
||||
</edge>
|
||||
<edge source="407978cb-39e6-453c-b9b7-c4183a4e26ef.TABLE:vulcanize_test.eth.receipt_cids" target="407978cb-39e6-453c-b9b7-c4183a4e26ef.TABLE:vulcanize_test.public.blocks" relationship="REFERENCES">
|
||||
<point x="-51.75" y="-105.0" />
|
||||
<point x="145.25" y="922.0" />
|
||||
<point x="59.365234375" y="922.0" />
|
||||
<point x="59.365234375" y="175.0" />
|
||||
<point x="52.75" y="175.0" />
|
||||
<point x="52.75" y="165.0" />
|
||||
<point x="127.92578125" y="165.0" />
|
||||
<point x="-25.25" y="39.0" />
|
||||
<edge source="86a0461b-ec84-4911-9aa2-e562b5d7b24c.TABLE:uml_diagram.eth.transaction_cids" target="86a0461b-ec84-4911-9aa2-e562b5d7b24c.TABLE:uml_diagram.eth.header_cids" relationship="REFERENCES">
|
||||
<point x="44.25" y="-116.0" />
|
||||
<point x="436.25" y="548.0" />
|
||||
<point x="658.75" y="548.0" />
|
||||
<point x="0.0" y="182.0" />
|
||||
</edge>
|
||||
<edge source="407978cb-39e6-453c-b9b7-c4183a4e26ef.TABLE:vulcanize_test.eth.state_accounts" target="407978cb-39e6-453c-b9b7-c4183a4e26ef.TABLE:vulcanize_test.eth.state_cids" relationship="REFERENCES">
|
||||
<point x="0.0" y="-83.0" />
|
||||
<point x="983.30078125" y="922.0" />
|
||||
<point x="936.865234375" y="922.0" />
|
||||
<edge source="86a0461b-ec84-4911-9aa2-e562b5d7b24c.TABLE:uml_diagram.eth.log_cids" target="86a0461b-ec84-4911-9aa2-e562b5d7b24c.TABLE:uml_diagram.eth.receipt_cids" relationship="REFERENCES">
|
||||
<point x="44.25" y="-127.0" />
|
||||
<point x="336.0" y="1114.0" />
|
||||
<point x="342.5" y="1114.0" />
|
||||
<point x="0.0" y="94.0" />
|
||||
</edge>
|
||||
<edge source="407978cb-39e6-453c-b9b7-c4183a4e26ef.TABLE:vulcanize_test.eth.log_cids" target="407978cb-39e6-453c-b9b7-c4183a4e26ef.TABLE:vulcanize_test.public.blocks" relationship="REFERENCES">
|
||||
<point x="-41.0" y="-127.0" />
|
||||
<point x="99.0" y="1180.0" />
|
||||
<point x="47.5" y="1180.0" />
|
||||
<point x="47.5" y="175.0" />
|
||||
<point x="52.75" y="175.0" />
|
||||
<point x="52.75" y="165.0" />
|
||||
<point x="127.92578125" y="165.0" />
|
||||
<point x="-25.25" y="39.0" />
|
||||
<edge source="86a0461b-ec84-4911-9aa2-e562b5d7b24c.TABLE:uml_diagram.eth.uncle_cids" target="86a0461b-ec84-4911-9aa2-e562b5d7b24c.TABLE:uml_diagram.eth.header_cids" relationship="REFERENCES">
|
||||
<point x="-48.5" y="-94.0" />
|
||||
<point x="778.0" y="548.0" />
|
||||
<point x="658.75" y="548.0" />
|
||||
<point x="0.0" y="182.0" />
|
||||
</edge>
|
||||
<edge source="407978cb-39e6-453c-b9b7-c4183a4e26ef.TABLE:vulcanize_test.eth.receipt_cids" target="407978cb-39e6-453c-b9b7-c4183a4e26ef.TABLE:vulcanize_test.eth.transaction_cids" relationship="REFERENCES">
|
||||
<point x="51.75" y="-105.0" />
|
||||
<point x="248.75" y="922.0" />
|
||||
<point x="511.30078125" y="922.0" />
|
||||
<point x="0.0" y="127.0" />
|
||||
<edge source="86a0461b-ec84-4911-9aa2-e562b5d7b24c.TABLE:uml_diagram.eth.log_cids" target="86a0461b-ec84-4911-9aa2-e562b5d7b24c.TABLE:uml_diagram.ipld.blocks" relationship="REFERENCES">
|
||||
<point x="-44.25" y="-127.0" />
|
||||
<point x="247.5" y="1114.0" />
|
||||
<point x="241.0" y="1114.0" />
|
||||
<point x="241.0" y="126.0" />
|
||||
<point x="551.25" y="126.0" />
|
||||
<point x="0.0" y="50.0" />
|
||||
</edge>
|
||||
<edge source="407978cb-39e6-453c-b9b7-c4183a4e26ef.TABLE:vulcanize_test.eth.storage_cids" target="407978cb-39e6-453c-b9b7-c4183a4e26ef.TABLE:vulcanize_test.public.blocks" relationship="REFERENCES">
|
||||
<edge source="86a0461b-ec84-4911-9aa2-e562b5d7b24c.TABLE:uml_diagram.eth.receipt_cids" target="86a0461b-ec84-4911-9aa2-e562b5d7b24c.TABLE:uml_diagram.eth.transaction_cids" relationship="REFERENCES">
|
||||
<point x="45.5" y="-94.0" />
|
||||
<point x="388.0" y="851.0" />
|
||||
<point x="392.0" y="851.0" />
|
||||
<point x="0.0" y="116.0" />
|
||||
</edge>
|
||||
<edge source="86a0461b-ec84-4911-9aa2-e562b5d7b24c.TABLE:uml_diagram.eth.state_cids" target="86a0461b-ec84-4911-9aa2-e562b5d7b24c.TABLE:uml_diagram.eth.header_cids" relationship="REFERENCES">
|
||||
<point x="52.25" y="-127.0" />
|
||||
<point x="657.25" y="548.0" />
|
||||
<point x="658.75" y="548.0" />
|
||||
<point x="0.0" y="182.0" />
|
||||
</edge>
|
||||
<edge source="86a0461b-ec84-4911-9aa2-e562b5d7b24c.TABLE:uml_diagram.eth.storage_cids" target="86a0461b-ec84-4911-9aa2-e562b5d7b24c.TABLE:uml_diagram.eth.state_cids" relationship="REFERENCES">
|
||||
<point x="-56.5" y="-105.0" />
|
||||
<point x="648.865234375" y="912.0" />
|
||||
<point x="116.80078125" y="912.0" />
|
||||
<point x="116.80078125" y="175.0" />
|
||||
<point x="437.205078125" y="175.0" />
|
||||
<point x="437.205078125" y="165.0" />
|
||||
<point x="178.42578125" y="165.0" />
|
||||
<point x="25.25" y="39.0" />
|
||||
<point x="717.1941964285713" y="851.0" />
|
||||
<point x="605.0" y="851.0" />
|
||||
<point x="0.0" y="127.0" />
|
||||
</edge>
|
||||
<edge source="407978cb-39e6-453c-b9b7-c4183a4e26ef.TABLE:vulcanize_test.eth.transaction_cids" target="407978cb-39e6-453c-b9b7-c4183a4e26ef.TABLE:vulcanize_test.eth.header_cids" relationship="REFERENCES">
|
||||
<point x="85.0" y="-127.0" />
|
||||
<point x="596.30078125" y="614.0" />
|
||||
<point x="700.96484375" y="614.0" />
|
||||
<point x="0.0" y="193.0" />
|
||||
<edge source="86a0461b-ec84-4911-9aa2-e562b5d7b24c.TABLE:uml_diagram.eth.transaction_cids" target="86a0461b-ec84-4911-9aa2-e562b5d7b24c.TABLE:uml_diagram.ipld.blocks" relationship="REFERENCES">
|
||||
<point x="-44.25" y="-116.0" />
|
||||
<point x="347.75" y="548.0" />
|
||||
<point x="391.1941964285714" y="548.0" />
|
||||
<point x="391.1941964285714" y="126.0" />
|
||||
<point x="551.25" y="126.0" />
|
||||
<point x="0.0" y="50.0" />
|
||||
</edge>
|
||||
<edge source="407978cb-39e6-453c-b9b7-c4183a4e26ef.TABLE:vulcanize_test.eth.state_cids" target="407978cb-39e6-453c-b9b7-c4183a4e26ef.TABLE:vulcanize_test.public.blocks" relationship="REFERENCES">
|
||||
<point x="52.25" y="-94.0" />
|
||||
<point x="989.115234375" y="614.0" />
|
||||
<point x="970.279296875" y="614.0" />
|
||||
<point x="970.279296875" y="175.0" />
|
||||
<point x="437.205078125" y="175.0" />
|
||||
<point x="437.205078125" y="165.0" />
|
||||
<point x="178.42578125" y="165.0" />
|
||||
<point x="25.25" y="39.0" />
|
||||
</edge>
|
||||
<edge source="407978cb-39e6-453c-b9b7-c4183a4e26ef.TABLE:vulcanize_test.eth.storage_cids" target="407978cb-39e6-453c-b9b7-c4183a4e26ef.TABLE:vulcanize_test.eth.state_cids" relationship="REFERENCES">
|
||||
<edge source="86a0461b-ec84-4911-9aa2-e562b5d7b24c.TABLE:uml_diagram.eth.storage_cids" target="86a0461b-ec84-4911-9aa2-e562b5d7b24c.TABLE:uml_diagram.ipld.blocks" relationship="REFERENCES">
|
||||
<point x="56.5" y="-105.0" />
|
||||
<point x="761.865234375" y="922.0" />
|
||||
<point x="936.865234375" y="922.0" />
|
||||
<point x="0.0" y="94.0" />
|
||||
<point x="830.1941964285713" y="851.0" />
|
||||
<point x="934.0" y="851.0" />
|
||||
<point x="934.0" y="126.0" />
|
||||
<point x="551.25" y="126.0" />
|
||||
<point x="0.0" y="50.0" />
|
||||
</edge>
|
||||
<edge source="407978cb-39e6-453c-b9b7-c4183a4e26ef.TABLE:vulcanize_test.eth.log_cids" target="407978cb-39e6-453c-b9b7-c4183a4e26ef.TABLE:vulcanize_test.eth.receipt_cids" relationship="REFERENCES">
|
||||
<point x="41.0" y="-127.0" />
|
||||
<point x="181.0" y="1180.0" />
|
||||
<point x="197.0" y="1180.0" />
|
||||
<point x="0.0" y="105.0" />
|
||||
<edge source="86a0461b-ec84-4911-9aa2-e562b5d7b24c.TABLE:uml_diagram.eth.uncle_cids" target="86a0461b-ec84-4911-9aa2-e562b5d7b24c.TABLE:uml_diagram.ipld.blocks" relationship="REFERENCES">
|
||||
<point x="48.5" y="-94.0" />
|
||||
<point x="875.0" y="548.0" />
|
||||
<point x="835.6941964285714" y="548.0" />
|
||||
<point x="835.6941964285714" y="126.0" />
|
||||
<point x="551.25" y="126.0" />
|
||||
<point x="0.0" y="50.0" />
|
||||
</edge>
|
||||
<edge source="407978cb-39e6-453c-b9b7-c4183a4e26ef.TABLE:vulcanize_test.eth.header_cids" target="407978cb-39e6-453c-b9b7-c4183a4e26ef.TABLE:vulcanize_test.public.nodes" relationship="REFERENCES">
|
||||
<point x="73.5" y="-193.0" />
|
||||
<point x="774.46484375" y="165.0" />
|
||||
<point x="1002.705078125" y="165.0" />
|
||||
<point x="0.0" y="72.0" />
|
||||
</edge>
|
||||
<edge source="407978cb-39e6-453c-b9b7-c4183a4e26ef.TABLE:vulcanize_test.eth.uncle_cids" target="407978cb-39e6-453c-b9b7-c4183a4e26ef.TABLE:vulcanize_test.eth.header_cids" relationship="REFERENCES">
|
||||
<point x="48.5" y="-83.0" />
|
||||
<point x="272.80078125" y="614.0" />
|
||||
<point x="700.96484375" y="614.0" />
|
||||
<point x="0.0" y="193.0" />
|
||||
</edge>
|
||||
<edge source="407978cb-39e6-453c-b9b7-c4183a4e26ef.TABLE:vulcanize_test.eth.header_cids" target="407978cb-39e6-453c-b9b7-c4183a4e26ef.TABLE:vulcanize_test.public.blocks" relationship="REFERENCES">
|
||||
<point x="-73.5" y="-193.0" />
|
||||
<point x="627.46484375" y="175.0" />
|
||||
<point x="437.205078125" y="175.0" />
|
||||
<point x="437.205078125" y="165.0" />
|
||||
<point x="178.42578125" y="165.0" />
|
||||
<point x="25.25" y="39.0" />
|
||||
</edge>
|
||||
<edge source="407978cb-39e6-453c-b9b7-c4183a4e26ef.TABLE:vulcanize_test.eth.uncle_cids" target="407978cb-39e6-453c-b9b7-c4183a4e26ef.TABLE:vulcanize_test.public.blocks" relationship="REFERENCES">
|
||||
<point x="-48.5" y="-83.0" />
|
||||
<point x="175.80078125" y="614.0" />
|
||||
<point x="214.30078125" y="614.0" />
|
||||
<point x="214.30078125" y="175.0" />
|
||||
<point x="437.205078125" y="175.0" />
|
||||
<point x="437.205078125" y="165.0" />
|
||||
<point x="178.42578125" y="165.0" />
|
||||
<point x="25.25" y="39.0" />
|
||||
</edge>
|
||||
<edge source="407978cb-39e6-453c-b9b7-c4183a4e26ef.TABLE:vulcanize_test.eth.transaction_cids" target="407978cb-39e6-453c-b9b7-c4183a4e26ef.TABLE:vulcanize_test.public.blocks" relationship="REFERENCES">
|
||||
<point x="-85.0" y="-127.0" />
|
||||
<point x="426.30078125" y="604.0" />
|
||||
<point x="344.55078125" y="604.0" />
|
||||
<point x="344.55078125" y="175.0" />
|
||||
<point x="437.205078125" y="175.0" />
|
||||
<point x="437.205078125" y="165.0" />
|
||||
<point x="178.42578125" y="165.0" />
|
||||
<point x="25.25" y="39.0" />
|
||||
<edge source="86a0461b-ec84-4911-9aa2-e562b5d7b24c.TABLE:uml_diagram.eth.header_cids" target="86a0461b-ec84-4911-9aa2-e562b5d7b24c.TABLE:uml_diagram.ipld.blocks" relationship="REFERENCES">
|
||||
<point x="0.0" y="-182.0" />
|
||||
<point x="658.75" y="126.0" />
|
||||
<point x="551.25" y="126.0" />
|
||||
<point x="0.0" y="50.0" />
|
||||
</edge>
|
||||
</edges>
|
||||
<settings layout="Hierarchic" zoom="0.40337837837837837" showDependencies="false" x="791.0" y="730.0" />
|
||||
<settings layout="Hierarchic" zoom="0.40290955091714103" showDependencies="false" x="793.0" y="780.5" />
|
||||
<SelectedNodes />
|
||||
<Categories>
|
||||
<Category>Columns</Category>
|
||||
|
Loading…
Reference in New Issue
Block a user