* Start testing blob pruning * Get rid of unnecessary orphaned blob column * Make random blob tests deterministic * Test for pruning being blocked by finality * Fix bugs and test fork boundary * A few more tweaks to pruning conditions * Tweak oldest_blob_slot semantics * Test margin pruning * Clean up some terminology and lints * Schema migrations for v18 * Remove FIXME * Prune blobs on finalization not every slot * Fix more bugs + tests * Address review comments |
||
|---|---|---|
| .. | ||
| migration_schema_v12.rs | ||
| migration_schema_v13.rs | ||
| migration_schema_v14.rs | ||
| migration_schema_v15.rs | ||
| migration_schema_v16.rs | ||
| migration_schema_v17.rs | ||
| migration_schema_v18.rs | ||
| README.md | ||
Database Schema Migrations
This document is an attempt to record some best practices and design conventions for applying database schema migrations within Lighthouse.
General Structure
If you make a breaking change to an on-disk data structure you need to increment the
SCHEMA_VERSION in beacon_node/store/src/metadata.rs and add a migration from the previous
version to the new version.
The entry-point for database migrations is in schema_change.rs, not migrate.rs (which deals
with finalization). Supporting code for a specific migration may be added in
schema_change/migration_schema_vX.rs, where X is the version being migrated to.
Combining Schema Changes
Schema changes may be combined if they are part of the same pull request to
unstable. Once a schema version is defined in unstable we should not apply changes to it
without incrementing the version. This prevents conflicts between versions that appear to be the
same. This allows us to deploy unstable to nodes without having to worry about needing to resync
because of a sneaky schema change.
Changing the on-disk structure for a version before it is merged to unstable is OK. You will
just have to handle manually resyncing any test nodes (use checkpoint sync).
Naming Conventions
Prefer to name versions of structs by the version at which the change was introduced. For example
if you add a field to Foo in v9, call the previous version FooV1 (assuming this is Foo's first
migration) and write a schema change that migrates from FooV1 to FooV9.
Prefer to use explicit version names in schema_change.rs and the schema_change module. To
interface with the outside either:
- Define a type alias to the latest version, e.g.
pub type Foo = FooV9, or - Define a mapping from the latest version to the version used elsewhere, e.g.
impl From<FooV9> for Foo {}
Avoid names like:
LegacyFooOldFooFooWithoutX
First-version vs Last-version
Previously the schema migration code would name types by the last version at which they were
valid. For example if Foo changed in V9 then we would name the two variants FooV8 and FooV9.
The problem with this scheme is that if Foo changes again in the future at say v12 then FooV9 would
need to be renamed to FooV11, which is annoying. Using the first valid version as described
above does not have this issue.
Using SuperStruct
If possible, consider using superstruct to handle data
structure changes between versions.
- Use
superstruct(no_enum)to avoid generating an unnecessary top-level enum.
Example
A field is added to Foo in v9, and there are two variants: FooV1 and FooV9. There is a
migration from FooV1 to FooV9. Foo is aliased to FooV9.
Some time later another field is added to Foo in v12. A new FooV12 is created, along with a
migration from FooV9 to FooV12. The primary Foo type gets re-aliased to FooV12. The previous
migration from V1 to V9 shouldn't break because the schema migration refers to FooV9 explicitly
rather than Foo. Due to the re-aliasing (or re-mapping) the compiler will check every usage
of Foo to make sure that it still makes sense with FooV12.