Fix rebase conflicts

This commit is contained in:
Emilia Hane 2023-02-01 17:49:48 +01:00
parent 04f635c0ac
commit 625980e484
No known key found for this signature in database
GPG Key ID: E73394F9C09206FA
11 changed files with 92 additions and 127 deletions

View File

@ -1072,7 +1072,7 @@ impl<T: BeaconChainTypes> BeaconChain<T> {
block_root: &Hash256, block_root: &Hash256,
data_availability_boundary: Epoch, data_availability_boundary: Epoch,
) -> Result<Option<BlobsSidecar<T::EthSpec>>, Error> { ) -> Result<Option<BlobsSidecar<T::EthSpec>>, Error> {
match self.store.get_blobs(block_root, slot)? { match self.store.get_blobs(block_root)? {
Some(blobs) => Ok(Some(blobs)), Some(blobs) => Ok(Some(blobs)),
None => { None => {
// Check for the corresponding block to understand whether we *should* have blobs. // Check for the corresponding block to understand whether we *should* have blobs.
@ -3021,6 +3021,7 @@ impl<T: BeaconChainTypes> BeaconChain<T> {
ops.push(StoreOp::PutBlock(block_root, signed_block.clone())); ops.push(StoreOp::PutBlock(block_root, signed_block.clone()));
ops.push(StoreOp::PutState(block.state_root(), &state)); ops.push(StoreOp::PutState(block.state_root(), &state));
<<<<<<< HEAD
// Only consider blobs if the eip4844 fork is enabled. // Only consider blobs if the eip4844 fork is enabled.
if let Some(data_availability_boundary) = self.data_availability_boundary() { if let Some(data_availability_boundary) = self.data_availability_boundary() {
let block_epoch = block.slot().epoch(T::EthSpec::slots_per_epoch()); let block_epoch = block.slot().epoch(T::EthSpec::slots_per_epoch());
@ -3040,6 +3041,13 @@ impl<T: BeaconChainTypes> BeaconChain<T> {
ops.push(StoreOp::PutBlobs(block_root, blobs)); ops.push(StoreOp::PutBlobs(block_root, blobs));
} }
} }
=======
if let Some(blobs) = blobs {
if blobs.blobs.len() > 0 {
//FIXME(sean) using this for debugging for now
info!(self.log, "Writing blobs to store"; "block_root" => ?block_root);
self.store.put_blobs(&block_root, (&*blobs).clone())?;
>>>>>>> 43dc3a9a4 (Fix rebase conflicts)
} }
} }

View File

@ -68,7 +68,7 @@ pub struct ClientBuilder<T: BeaconChainTypes> {
gossipsub_registry: Option<Registry>, gossipsub_registry: Option<Registry>,
db_path: Option<PathBuf>, db_path: Option<PathBuf>,
freezer_db_path: Option<PathBuf>, freezer_db_path: Option<PathBuf>,
blobs_freezer_db_path: Option<PathBuf>, blobs_db_path: Option<PathBuf>,
http_api_config: http_api::Config, http_api_config: http_api::Config,
http_metrics_config: http_metrics::Config, http_metrics_config: http_metrics::Config,
slasher: Option<Arc<Slasher<T::EthSpec>>>, slasher: Option<Arc<Slasher<T::EthSpec>>>,
@ -101,7 +101,7 @@ where
gossipsub_registry: None, gossipsub_registry: None,
db_path: None, db_path: None,
freezer_db_path: None, freezer_db_path: None,
blobs_freezer_db_path: None, blobs_db_path: None,
http_api_config: <_>::default(), http_api_config: <_>::default(),
http_metrics_config: <_>::default(), http_metrics_config: <_>::default(),
slasher: None, slasher: None,
@ -894,7 +894,7 @@ where
mut self, mut self,
hot_path: &Path, hot_path: &Path,
cold_path: &Path, cold_path: &Path,
cold_blobs_path: Option<PathBuf>, blobs_path: Option<PathBuf>,
config: StoreConfig, config: StoreConfig,
log: Logger, log: Logger,
) -> Result<Self, String> { ) -> Result<Self, String> {
@ -910,7 +910,7 @@ where
self.db_path = Some(hot_path.into()); self.db_path = Some(hot_path.into());
self.freezer_db_path = Some(cold_path.into()); self.freezer_db_path = Some(cold_path.into());
self.blobs_freezer_db_path = cold_blobs_path.clone(); self.blobs_db_path = blobs_path.clone();
let inner_spec = spec.clone(); let inner_spec = spec.clone();
let deposit_contract_deploy_block = context let deposit_contract_deploy_block = context
@ -933,7 +933,7 @@ where
let store = HotColdDB::open( let store = HotColdDB::open(
hot_path, hot_path,
cold_path, cold_path,
cold_blobs_path, blobs_path,
schema_upgrade, schema_upgrade,
config, config,
spec, spec,

View File

@ -49,9 +49,8 @@ pub struct Config {
pub db_name: String, pub db_name: String,
/// Path where the freezer database will be located. /// Path where the freezer database will be located.
pub freezer_db_path: Option<PathBuf>, pub freezer_db_path: Option<PathBuf>,
/// Path where the blobs freezer database will be located if it should be separate from the /// Path where the blobs database will be located if blobs should be in a separate database.
/// historical state freezer. pub blobs_db_path: Option<PathBuf>,
pub blobs_freezer_db_path: Option<PathBuf>,
pub log_file: PathBuf, pub log_file: PathBuf,
/// If true, the node will use co-ordinated junk for eth1 values. /// If true, the node will use co-ordinated junk for eth1 values.
/// ///
@ -92,7 +91,7 @@ impl Default for Config {
data_dir: PathBuf::from(DEFAULT_ROOT_DIR), data_dir: PathBuf::from(DEFAULT_ROOT_DIR),
db_name: "chain_db".to_string(), db_name: "chain_db".to_string(),
freezer_db_path: None, freezer_db_path: None,
blobs_freezer_db_path: None, blobs_db_path: None,
log_file: PathBuf::from(""), log_file: PathBuf::from(""),
genesis: <_>::default(), genesis: <_>::default(),
store: <_>::default(), store: <_>::default(),
@ -153,12 +152,12 @@ impl Config {
.unwrap_or_else(|| self.default_freezer_db_path()) .unwrap_or_else(|| self.default_freezer_db_path())
} }
/// Returns the path to which the client may initialize the on-disk blobs freezer database. /// Returns the path to which the client may initialize the on-disk blobs database.
/// ///
/// Will attempt to use the user-supplied path from e.g. the CLI, or will default /// Will attempt to use the user-supplied path from e.g. the CLI, or will default
/// to None. /// to None.
pub fn get_blobs_freezer_db_path(&self) -> Option<PathBuf> { pub fn get_blobs_db_path(&self) -> Option<PathBuf> {
self.blobs_freezer_db_path.clone() self.blobs_db_path.clone()
} }
/// Get the freezer DB path, creating it if necessary. /// Get the freezer DB path, creating it if necessary.
@ -166,10 +165,10 @@ impl Config {
ensure_dir_exists(self.get_freezer_db_path()) ensure_dir_exists(self.get_freezer_db_path())
} }
/// Get the blobs freezer DB path, creating it if necessary. /// Get the blobs DB path, creating it if necessary.
pub fn create_blobs_freezer_db_path(&self) -> Result<Option<PathBuf>, String> { pub fn create_blobs_db_path(&self) -> Result<Option<PathBuf>, String> {
match self.get_blobs_freezer_db_path() { match self.get_blobs_db_path() {
Some(blobs_freezer_path) => Ok(Some(ensure_dir_exists(blobs_freezer_path)?)), Some(blobs_db_path) => Ok(Some(ensure_dir_exists(blobs_db_path)?)),
None => Ok(None), None => Ok(None),
} }
} }

View File

@ -218,7 +218,7 @@ impl BlockId {
chain: &BeaconChain<T>, chain: &BeaconChain<T>,
) -> Result<Arc<BlobsSidecar<T::EthSpec>>, warp::Rejection> { ) -> Result<Arc<BlobsSidecar<T::EthSpec>>, warp::Rejection> {
let root = self.root(chain)?.0; let root = self.root(chain)?.0;
match chain.get_blobs(&root, None) { match chain.get_blobs(&root) {
Ok(Some(blob)) => Ok(Arc::new(blob)), Ok(Some(blob)) => Ok(Arc::new(blob)),
Ok(None) => Err(warp_utils::reject::custom_not_found(format!( Ok(None) => Err(warp_utils::reject::custom_not_found(format!(
"Blob with block root {} is not in the store", "Blob with block root {} is not in the store",

View File

@ -795,14 +795,12 @@ impl<T: BeaconChainTypes> Worker<T> {
// remove all skip slots // remove all skip slots
let block_roots = block_roots.into_iter().flatten().collect::<Vec<_>>(); let block_roots = block_roots.into_iter().flatten().collect::<Vec<_>>();
let mut slot_hint: Option<Slot> = None;
let mut blobs_sent = 0; let mut blobs_sent = 0;
let mut send_response = true; let mut send_response = true;
for root in block_roots { for root in block_roots {
match self.chain.get_blobs(&root, data_availability_boundary) { match self.chain.get_blobs(&root, data_availability_boundary) {
Ok(Some(blobs)) => { Ok(Some(blobs)) => {
slot_hint = Some(blobs.beacon_block_slot + 1);
blobs_sent += 1; blobs_sent += 1;
self.send_network_message(NetworkMessage::SendResponse { self.send_network_message(NetworkMessage::SendResponse {
peer_id, peer_id,

View File

@ -29,10 +29,10 @@ pub fn cli_app<'a, 'b>() -> App<'a, 'b> {
.takes_value(true) .takes_value(true)
) )
.arg( .arg(
Arg::with_name("blobs-freezer-dir") Arg::with_name("blobs-dir")
.long("blobs-freezer-dir") .long("blobs-dir")
.value_name("DIR") .value_name("DIR")
.help("Data directory for the blobs freezer database.") .help("Data directory for the blobs database.")
.takes_value(true) .takes_value(true)
) )
/* /*

View File

@ -390,8 +390,8 @@ pub fn get_config<E: EthSpec>(
client_config.freezer_db_path = Some(PathBuf::from(freezer_dir)); client_config.freezer_db_path = Some(PathBuf::from(freezer_dir));
} }
if let Some(blobs_freezer_dir) = cli_args.value_of("blobs-freezer-dir") { if let Some(blobs_db_dir) = cli_args.value_of("blobs-dir") {
client_config.blobs_freezer_db_path = Some(PathBuf::from(blobs_freezer_dir)); client_config.blobs_db_path = Some(PathBuf::from(blobs_db_dir));
} }
let (sprp, sprp_explicit) = get_slots_per_restore_point::<E>(cli_args)?; let (sprp, sprp_explicit) = get_slots_per_restore_point::<E>(cli_args)?;

View File

@ -64,7 +64,7 @@ impl<E: EthSpec> ProductionBeaconNode<E> {
let _datadir = client_config.create_data_dir()?; let _datadir = client_config.create_data_dir()?;
let db_path = client_config.create_db_path()?; let db_path = client_config.create_db_path()?;
let freezer_db_path = client_config.create_freezer_db_path()?; let freezer_db_path = client_config.create_freezer_db_path()?;
let blobs_freezer_db_path = client_config.create_blobs_freezer_db_path()?; let blobs_db_path = client_config.create_blobs_db_path()?;
let executor = context.executor.clone(); let executor = context.executor.clone();
if let Some(legacy_dir) = client_config.get_existing_legacy_data_dir() { if let Some(legacy_dir) = client_config.get_existing_legacy_data_dir() {
@ -88,7 +88,7 @@ impl<E: EthSpec> ProductionBeaconNode<E> {
.disk_store( .disk_store(
&db_path, &db_path,
&freezer_db_path, &freezer_db_path,
blobs_freezer_db_path, blobs_db_path,
store_config, store_config,
log.clone(), log.clone(),
)?; )?;

View File

@ -59,8 +59,8 @@ pub struct HotColdDB<E: EthSpec, Hot: ItemStore<E>, Cold: ItemStore<E>> {
pub(crate) config: StoreConfig, pub(crate) config: StoreConfig,
/// Cold database containing compact historical data. /// Cold database containing compact historical data.
pub cold_db: Cold, pub cold_db: Cold,
/// Cold database containing blob data with slots less than `split.slot`. /// Database containing blobs.
pub cold_blobs_db: Option<Cold>, pub blobs_db: Option<Cold>,
/// Hot database containing duplicated but quick-to-access recent data. /// Hot database containing duplicated but quick-to-access recent data.
/// ///
/// The hot database also contains all blocks. /// The hot database also contains all blocks.
@ -101,7 +101,7 @@ pub enum HotColdDBError {
MissingExecutionPayload(Hash256), MissingExecutionPayload(Hash256),
MissingFullBlockExecutionPayloadPruned(Hash256, Slot), MissingFullBlockExecutionPayloadPruned(Hash256, Slot),
MissingAnchorInfo, MissingAnchorInfo,
MissingPathToBlobsFreezer, MissingPathToBlobsDatabase,
HotStateSummaryError(BeaconStateError), HotStateSummaryError(BeaconStateError),
RestorePointDecodeError(ssz::DecodeError), RestorePointDecodeError(ssz::DecodeError),
BlockReplayBeaconError(BeaconStateError), BlockReplayBeaconError(BeaconStateError),
@ -138,7 +138,7 @@ impl<E: EthSpec> HotColdDB<E, MemoryStore<E>, MemoryStore<E>> {
anchor_info: RwLock::new(None), anchor_info: RwLock::new(None),
blob_info: RwLock::new(BlobInfo::default()), blob_info: RwLock::new(BlobInfo::default()),
cold_db: MemoryStore::open(), cold_db: MemoryStore::open(),
cold_blobs_db: Some(MemoryStore::open()), blobs_db: Some(MemoryStore::open()),
hot_db: MemoryStore::open(), hot_db: MemoryStore::open(),
block_cache: Mutex::new(LruCache::new(config.block_cache_size)), block_cache: Mutex::new(LruCache::new(config.block_cache_size)),
blob_cache: Mutex::new(LruCache::new(config.blob_cache_size)), blob_cache: Mutex::new(LruCache::new(config.blob_cache_size)),
@ -162,7 +162,7 @@ impl<E: EthSpec> HotColdDB<E, LevelDB<E>, LevelDB<E>> {
pub fn open( pub fn open(
hot_path: &Path, hot_path: &Path,
cold_path: &Path, cold_path: &Path,
cold_blobs_path: Option<PathBuf>, blobs_db_path: Option<PathBuf>,
migrate_schema: impl FnOnce(Arc<Self>, SchemaVersion, SchemaVersion) -> Result<(), Error>, migrate_schema: impl FnOnce(Arc<Self>, SchemaVersion, SchemaVersion) -> Result<(), Error>,
config: StoreConfig, config: StoreConfig,
spec: ChainSpec, spec: ChainSpec,
@ -175,7 +175,7 @@ impl<E: EthSpec> HotColdDB<E, LevelDB<E>, LevelDB<E>> {
anchor_info: RwLock::new(None), anchor_info: RwLock::new(None),
blob_info: RwLock::new(BlobInfo::default()), blob_info: RwLock::new(BlobInfo::default()),
cold_db: LevelDB::open(cold_path)?, cold_db: LevelDB::open(cold_path)?,
cold_blobs_db: None, blobs_db: None,
hot_db: LevelDB::open(hot_path)?, hot_db: LevelDB::open(hot_path)?,
block_cache: Mutex::new(LruCache::new(config.block_cache_size)), block_cache: Mutex::new(LruCache::new(config.block_cache_size)),
blob_cache: Mutex::new(LruCache::new(config.blob_cache_size)), blob_cache: Mutex::new(LruCache::new(config.blob_cache_size)),
@ -220,23 +220,29 @@ impl<E: EthSpec> HotColdDB<E, LevelDB<E>, LevelDB<E>> {
); );
} }
if db.spec.eip4844_fork_epoch.is_some() { let blob_info_on_disk = db.load_blob_info()?;
let blob_info = match db.load_blob_info()? {
Some(mut blob_info) => { if let Some(ref blob_info) = blob_info_on_disk {
if blob_info.blobs_freezer { let prev_blobs_db = blob_info.blobs_db;
cold_blobs_path if prev_blobs_db {
blobs_db_path
.as_ref() .as_ref()
.ok_or(HotColdDBError::MissingPathToBlobsFreezer)?; .ok_or(HotColdDBError::MissingPathToBlobsDatabase)?;
} }
if let Some(path) = cold_blobs_path {
db.cold_blobs_db = Some(LevelDB::open(path.as_path())?);
blob_info.blobs_freezer = true;
} }
Some(blob_info)
if let Some(path) = blobs_db_path {
if db.spec.eip4844_fork_epoch.is_some() {
db.blobs_db = Some(LevelDB::open(path.as_path())?);
db.compare_and_set_blob_info_with_write(
blob_info_on_disk,
Some(BlobInfo { blobs_db: true }),
)?;
info!(
db.log,
"Blobs DB initialized";
);
} }
None => Some(BlobInfo::default()),
};
*db.blob_info.write() = blob_info;
} }
// Ensure that the schema version of the on-disk database matches the software. // Ensure that the schema version of the on-disk database matches the software.
@ -547,29 +553,6 @@ impl<E: EthSpec, Hot: ItemStore<E>, Cold: ItemStore<E>> HotColdDB<E, Hot, Cold>
Ok(()) Ok(())
} }
/// Fetch a blobs sidecar from the store.
///
/// If `slot` is provided then it will be used as a hint as to which database should
/// be checked first.
pub fn get_blobs(
&self,
block_root: &Hash256,
slot: Option<Slot>,
) -> Result<Option<BlobsSidecar<E>>, Error> {
if let Some(slot) = slot {
if slot < self.get_split_slot() {
return match self.load_cold_blobs(block_root)? {
Some(blobs) => Ok(Some(blobs)),
None => self.load_hot_blobs(block_root),
};
}
}
match self.load_hot_blobs(block_root)? {
Some(blobs) => Ok(Some(blobs)),
None => self.load_cold_blobs(block_root),
}
}
pub fn blobs_as_kv_store_ops( pub fn blobs_as_kv_store_ops(
&self, &self,
key: &Hash256, key: &Hash256,
@ -910,6 +893,7 @@ impl<E: EthSpec, Hot: ItemStore<E>, Cold: ItemStore<E>> HotColdDB<E, Hot, Cold>
self.hot_db self.hot_db
.do_atomically(self.convert_to_kv_batch(batch)?)?; .do_atomically(self.convert_to_kv_batch(batch)?)?;
drop(guard); drop(guard);
drop(guard_blob); drop(guard_blob);
@ -1246,35 +1230,22 @@ impl<E: EthSpec, Hot: ItemStore<E>, Cold: ItemStore<E>> HotColdDB<E, Hot, Cold>
}) })
} }
/// Load a blobs sidecar from the hot database. /// Fetch a blobs sidecar from the store.
pub fn load_hot_blobs(&self, block_root: &Hash256) -> Result<Option<BlobsSidecar<E>>, Error> { pub fn get_blobs(&self, block_root: &Hash256) -> Result<Option<BlobsSidecar<E>>, Error> {
// FIXME(sean) I was attempting to use a blob cache here but was getting deadlocks, let blobs_db = if let Some(ref blobs_db) = self.blobs_db {
// may want to attempt to use one again blobs_db
match self
.hot_db
.get_bytes(DBColumn::BeaconBlob.into(), block_root.as_bytes())?
{
Some(bytes) => {
let ret = BlobsSidecar::from_ssz_bytes(&bytes)?;
self.blob_cache.lock().put(*block_root, ret.clone());
Ok(Some(ret))
}
None => Ok(None),
}
}
/// Try to load a blobs from the freezer database.
///
/// Return `None` if no blobs sidecar with `block_root` lies in the freezer.
pub fn load_cold_blobs(&self, block_root: &Hash256) -> Result<Option<BlobsSidecar<E>>, Error> {
let blobs_freezer = if let Some(ref cold_blobs_db) = self.cold_blobs_db {
cold_blobs_db
} else { } else {
&self.cold_db &self.cold_db
}; };
match blobs_freezer.get_bytes(DBColumn::BeaconBlob.into(), block_root.as_bytes())? { match blobs_db.get_bytes(DBColumn::BeaconBlob.into(), block_root.as_bytes())? {
Some(ref blobs_bytes) => Ok(Some(BlobsSidecar::from_ssz_bytes(blobs_bytes)?)), Some(ref blobs_bytes) => {
let blobs = BlobsSidecar::from_ssz_bytes(blobs_bytes)?;
// FIXME(sean) I was attempting to use a blob cache here but was getting deadlocks,
// may want to attempt to use one again
self.blob_cache.lock().put(*block_root, blobs.clone());
Ok(Some(blobs))
}
None => Ok(None), None => Ok(None),
} }
} }
@ -1985,12 +1956,6 @@ pub fn migrate_database<E: EthSpec, Hot: ItemStore<E>, Cold: ItemStore<E>>(
} }
let mut hot_db_ops: Vec<StoreOp<E>> = Vec::new(); let mut hot_db_ops: Vec<StoreOp<E>> = Vec::new();
let mut cold_blobs_db_ops: Vec<StoreOp<E>> = Vec::new();
let blobs_freezer = match store.cold_blobs_db {
Some(ref cold_blobs_db) => cold_blobs_db,
None => &store.cold_db,
};
// 1. Copy all of the states between the head and the split slot, from the hot DB // 1. Copy all of the states between the head and the split slot, from the hot DB
// to the cold DB. Delete the execution payloads of these now-finalized blocks. // to the cold DB. Delete the execution payloads of these now-finalized blocks.
@ -2034,16 +1999,7 @@ pub fn migrate_database<E: EthSpec, Hot: ItemStore<E>, Cold: ItemStore<E>>(
if store.config.prune_payloads { if store.config.prune_payloads {
hot_db_ops.push(StoreOp::DeleteExecutionPayload(block_root)); hot_db_ops.push(StoreOp::DeleteExecutionPayload(block_root));
} }
// Prepare migration of blobs to freezer.
if let Some(blobs) = store.get_blobs(&block_root, Some(slot))? {
hot_db_ops.push(StoreOp::DeleteBlobs(block_root));
cold_blobs_db_ops.push(StoreOp::PutBlobs(block_root, Arc::new(blobs)));
} }
}
// Migrate blobs to freezer.
blobs_freezer.do_atomically(store.convert_to_kv_batch(cold_blobs_db_ops)?)?;
// Warning: Critical section. We have to take care not to put any of the two databases in an // Warning: Critical section. We have to take care not to put any of the two databases in an
// inconsistent state if the OS process dies at any point during the freezing // inconsistent state if the OS process dies at any point during the freezing
@ -2057,9 +2013,6 @@ pub fn migrate_database<E: EthSpec, Hot: ItemStore<E>, Cold: ItemStore<E>>(
// Flush to disk all the states that have just been migrated to the cold store. // Flush to disk all the states that have just been migrated to the cold store.
store.cold_db.sync()?; store.cold_db.sync()?;
if let Some(ref cold_blobs_db) = store.cold_blobs_db {
cold_blobs_db.sync()?;
}
{ {
let mut split_guard = store.split.write(); let mut split_guard = store.split.write();

View File

@ -124,6 +124,8 @@ impl StoreItem for AnchorInfo {
pub struct BlobInfo { pub struct BlobInfo {
/// The slot after which blobs are available (>=). /// The slot after which blobs are available (>=).
pub oldest_blob_slot: Option<Slot>, pub oldest_blob_slot: Option<Slot>,
/// A separate blobs database is in use.
pub blobs_db: bool,
} }
impl StoreItem for BlobInfo { impl StoreItem for BlobInfo {

View File

@ -95,6 +95,7 @@ pub fn cli_app<'a, 'b>() -> App<'a, 'b> {
.takes_value(true), .takes_value(true),
) )
.arg( .arg(
<<<<<<< HEAD
Arg::with_name("blob-prune-margin-epochs") Arg::with_name("blob-prune-margin-epochs")
.long("blob-prune-margin-epochs") .long("blob-prune-margin-epochs")
.help( .help(
@ -105,8 +106,12 @@ pub fn cli_app<'a, 'b>() -> App<'a, 'b> {
.default_value("0"), .default_value("0"),
Arg::with_name("blobs-freezer-dir") Arg::with_name("blobs-freezer-dir")
.long("blobs-freezer-dir") .long("blobs-freezer-dir")
=======
Arg::with_name("blobs-dir")
.long("blobs-dir")
>>>>>>> 43dc3a9a4 (Fix rebase conflicts)
.value_name("DIR") .value_name("DIR")
.help("Data directory for the blobs freezer database.") .help("Data directory for the blobs database.")
.takes_value(true), .takes_value(true),
) )
.subcommand(migrate_cli_app()) .subcommand(migrate_cli_app())
@ -128,8 +133,8 @@ fn parse_client_config<E: EthSpec>(
client_config.freezer_db_path = Some(freezer_dir); client_config.freezer_db_path = Some(freezer_dir);
} }
if let Some(blobs_freezer_dir) = clap_utils::parse_optional(cli_args, "blobs-freezer-dir")? { if let Some(blobs_db_dir) = clap_utils::parse_optional(cli_args, "blobs-dir")? {
client_config.blobs_freezer_db_path = Some(blobs_freezer_dir); client_config.blobs_db_path = Some(blobs_db_dir);
} }
let (sprp, sprp_explicit) = get_slots_per_restore_point::<E>(cli_args)?; let (sprp, sprp_explicit) = get_slots_per_restore_point::<E>(cli_args)?;
@ -153,13 +158,13 @@ pub fn display_db_version<E: EthSpec>(
let spec = runtime_context.eth2_config.spec.clone(); let spec = runtime_context.eth2_config.spec.clone();
let hot_path = client_config.get_db_path(); let hot_path = client_config.get_db_path();
let cold_path = client_config.get_freezer_db_path(); let cold_path = client_config.get_freezer_db_path();
let cold_blobs_path = client_config.get_blobs_freezer_db_path(); let blobs_path = client_config.get_blobs_db_path();
let mut version = CURRENT_SCHEMA_VERSION; let mut version = CURRENT_SCHEMA_VERSION;
HotColdDB::<E, LevelDB<E>, LevelDB<E>>::open( HotColdDB::<E, LevelDB<E>, LevelDB<E>>::open(
&hot_path, &hot_path,
&cold_path, &cold_path,
&cold_blobs_path, &blobs_path,
|_, from, _| { |_, from, _| {
version = from; version = from;
Ok(()) Ok(())
@ -211,12 +216,12 @@ pub fn inspect_db<E: EthSpec>(
let spec = runtime_context.eth2_config.spec.clone(); let spec = runtime_context.eth2_config.spec.clone();
let hot_path = client_config.get_db_path(); let hot_path = client_config.get_db_path();
let cold_path = client_config.get_freezer_db_path(); let cold_path = client_config.get_freezer_db_path();
let cold_blobs_path = client_config.get_blobs_freezer_db_path(); let blobs_path = client_config.get_blobs_db_path();
let db = HotColdDB::<E, LevelDB<E>, LevelDB<E>>::open( let db = HotColdDB::<E, LevelDB<E>, LevelDB<E>>::open(
&hot_path, &hot_path,
&cold_path, &cold_path,
&cold_blobs_path, &blobs_path,
|_, _, _| Ok(()), |_, _, _| Ok(()),
client_config.store, client_config.store,
spec, spec,
@ -267,14 +272,14 @@ pub fn migrate_db<E: EthSpec>(
let spec = &runtime_context.eth2_config.spec; let spec = &runtime_context.eth2_config.spec;
let hot_path = client_config.get_db_path(); let hot_path = client_config.get_db_path();
let cold_path = client_config.get_freezer_db_path(); let cold_path = client_config.get_freezer_db_path();
let cold_blobs_path = client_config.get_blobs_freezer_db_path(); let blobs_path = client_config.get_blobs_db_path();
let mut from = CURRENT_SCHEMA_VERSION; let mut from = CURRENT_SCHEMA_VERSION;
let to = migrate_config.to; let to = migrate_config.to;
let db = HotColdDB::<E, LevelDB<E>, LevelDB<E>>::open( let db = HotColdDB::<E, LevelDB<E>, LevelDB<E>>::open(
&hot_path, &hot_path,
&cold_path, &cold_path,
&cold_blobs_path, &blobs_path,
|_, db_initial_version, _| { |_, db_initial_version, _| {
from = db_initial_version; from = db_initial_version;
Ok(()) Ok(())
@ -309,12 +314,12 @@ pub fn prune_payloads<E: EthSpec>(
let spec = &runtime_context.eth2_config.spec; let spec = &runtime_context.eth2_config.spec;
let hot_path = client_config.get_db_path(); let hot_path = client_config.get_db_path();
let cold_path = client_config.get_freezer_db_path(); let cold_path = client_config.get_freezer_db_path();
let cold_blobs_path = client_config.get_blobs_freezer_db_path(); let blobs_path = client_config.get_blobs_db_path();
let db = HotColdDB::<E, LevelDB<E>, LevelDB<E>>::open( let db = HotColdDB::<E, LevelDB<E>, LevelDB<E>>::open(
&hot_path, &hot_path,
&cold_path, &cold_path,
&cold_blobs_path, &blobs_path,
|_, _, _| Ok(()), |_, _, _| Ok(()),
client_config.store, client_config.store,
spec.clone(), spec.clone(),