Fix rebase conflicts

This commit is contained in:
Emilia Hane 2023-02-01 17:51:01 +01:00
parent f8c3e7fc91
commit f971f3a3a2
No known key found for this signature in database
GPG Key ID: E73394F9C09206FA
6 changed files with 60 additions and 57 deletions

View File

@ -3021,7 +3021,6 @@ impl<T: BeaconChainTypes> BeaconChain<T> {
ops.push(StoreOp::PutBlock(block_root, signed_block.clone())); ops.push(StoreOp::PutBlock(block_root, signed_block.clone()));
ops.push(StoreOp::PutState(block.state_root(), &state)); ops.push(StoreOp::PutState(block.state_root(), &state));
<<<<<<< HEAD
// Only consider blobs if the eip4844 fork is enabled. // Only consider blobs if the eip4844 fork is enabled.
if let Some(data_availability_boundary) = self.data_availability_boundary() { if let Some(data_availability_boundary) = self.data_availability_boundary() {
let block_epoch = block.slot().epoch(T::EthSpec::slots_per_epoch()); let block_epoch = block.slot().epoch(T::EthSpec::slots_per_epoch());
@ -3032,7 +3031,7 @@ impl<T: BeaconChainTypes> BeaconChain<T> {
// margin, or younger (of higher epoch number). // margin, or younger (of higher epoch number).
if block_epoch >= import_boundary { if block_epoch >= import_boundary {
if let Some(blobs) = blobs { if let Some(blobs) = blobs {
if blobs.blobs.len() > 0 { if !blobs.blobs.is_empty() {
//FIXME(sean) using this for debugging for now //FIXME(sean) using this for debugging for now
info!( info!(
self.log, "Writing blobs to store"; self.log, "Writing blobs to store";
@ -3041,16 +3040,8 @@ impl<T: BeaconChainTypes> BeaconChain<T> {
ops.push(StoreOp::PutBlobs(block_root, blobs)); ops.push(StoreOp::PutBlobs(block_root, blobs));
} }
} }
=======
if let Some(blobs) = blobs {
if blobs.blobs.len() > 0 {
//FIXME(sean) using this for debugging for now
info!(self.log, "Writing blobs to store"; "block_root" => ?block_root);
self.store.put_blobs(&block_root, (&*blobs).clone())?;
>>>>>>> 43dc3a9a4 (Fix rebase conflicts)
} }
} }
let txn_lock = self.store.hot_db.begin_rw_transaction(); let txn_lock = self.store.hot_db.begin_rw_transaction();
if let Err(e) = self.store.do_atomically(ops) { if let Err(e) = self.store.do_atomically(ops) {
@ -3089,6 +3080,17 @@ impl<T: BeaconChainTypes> BeaconChain<T> {
return Err(e.into()); return Err(e.into());
} }
if let Some(blobs) = blobs? {
if blobs.blobs.len() > 0 {
//FIXME(sean) using this for debugging for now
info!(self.log, "Writing blobs to store"; "block_root" => ?block_root);
// WARNING! Deadlocks if the alternative to a separate blobs db is
// changed from the cold db to the hot db.
self.store.put_blobs(&block_root, (&*blobs).clone())?;
}
};
drop(txn_lock); drop(txn_lock);
// The fork choice write-lock is dropped *after* the on-disk database has been updated. // The fork choice write-lock is dropped *after* the on-disk database has been updated.

View File

@ -1382,7 +1382,7 @@ impl<T: BeaconChainTypes> ExecutionPendingBlock<T> {
}; };
chain chain
.store .store
.do_atomically_and_update_cache(state_batch, None)?; .do_atomically_with_block_and_blobs_cache(state_batch)?;
drop(txn_lock); drop(txn_lock);
confirmed_state_roots.push(state_root); confirmed_state_roots.push(state_root);

View File

@ -31,7 +31,7 @@ where
"Garbage collecting {} temporary states", "Garbage collecting {} temporary states",
delete_ops.len() / 2 delete_ops.len() / 2
); );
self.do_atomically_and_update_cache(delete_ops, None)?; self.do_atomically_with_block_and_blobs_cache(delete_ops)?;
} }
Ok(()) Ok(())

View File

@ -544,7 +544,8 @@ impl<E: EthSpec, Hot: ItemStore<E>, Cold: ItemStore<E>> HotColdDB<E, Hot, Cold>
} }
pub fn put_blobs(&self, block_root: &Hash256, blobs: BlobsSidecar<E>) -> Result<(), Error> { pub fn put_blobs(&self, block_root: &Hash256, blobs: BlobsSidecar<E>) -> Result<(), Error> {
self.hot_db.put_bytes( let blobs_db = self.blobs_db.as_ref().unwrap_or(&self.cold_db);
blobs_db.put_bytes(
DBColumn::BeaconBlob.into(), DBColumn::BeaconBlob.into(),
block_root.as_bytes(), block_root.as_bytes(),
&blobs.as_ssz_bytes(), &blobs.as_ssz_bytes(),
@ -849,19 +850,38 @@ impl<E: EthSpec, Hot: ItemStore<E>, Cold: ItemStore<E>> HotColdDB<E, Hot, Cold>
Ok(key_value_batch) Ok(key_value_batch)
} }
pub fn do_atomically_and_update_cache( pub fn do_atomically_with_block_and_blobs_cache(
&self, &self,
batch: Vec<StoreOp<E>>, batch: Vec<StoreOp<E>>,
blobs_batch: Option<Vec<StoreOp<E>>>,
) -> Result<(), Error> { ) -> Result<(), Error> {
// Update the block cache whilst holding a lock, to ensure that the cache updates atomically let mut hot_db_cache_ops = Vec::new();
// with the database.
let mut guard = self.block_cache.lock();
for op in &batch { let (blobs_ops, hot_db_ops) = batch.into_iter().partition(|store_op| match store_op {
StoreOp::PutBlobs(_, _) | StoreOp::DeleteBlobs(_) => true,
StoreOp::PutBlock(_, _) | StoreOp::DeleteBlock(_) => {
hot_db_cache_ops.push(store_op.clone());
false
}
_ => false,
});
// Update database whilst holding a lock on cache, to ensure that the cache updates
// atomically with the database.
let mut guard = self.block_cache.lock();
let mut guard_blob = self.blob_cache.lock();
self.hot_db
.do_atomically(self.convert_to_kv_batch(hot_db_ops)?)?;
let blob_cache_ops = blobs_ops.clone();
let blobs_db = self.blobs_db.as_ref().unwrap_or(&self.cold_db);
// todo(emhane): do we want to restore the hot db writes if this fails?
blobs_db.do_atomically(self.convert_to_kv_batch(blobs_ops)?)?;
for op in hot_db_cache_ops {
match op { match op {
StoreOp::PutBlock(block_root, block) => { StoreOp::PutBlock(block_root, block) => {
guard.put(*block_root, (**block).clone()); guard.put(block_root, (*block).clone());
} }
StoreOp::PutBlobs(_, _) => (), StoreOp::PutBlobs(_, _) => (),
@ -875,7 +895,7 @@ impl<E: EthSpec, Hot: ItemStore<E>, Cold: ItemStore<E>> HotColdDB<E, Hot, Cold>
StoreOp::DeleteStateTemporaryFlag(_) => (), StoreOp::DeleteStateTemporaryFlag(_) => (),
StoreOp::DeleteBlock(block_root) => { StoreOp::DeleteBlock(block_root) => {
guard.pop(block_root); guard.pop(&block_root);
} }
StoreOp::DeleteBlobs(_) => (), StoreOp::DeleteBlobs(_) => (),
@ -890,39 +910,22 @@ impl<E: EthSpec, Hot: ItemStore<E>, Cold: ItemStore<E>> HotColdDB<E, Hot, Cold>
} }
} }
let guard_blob = match blobs_batch { for op in blob_cache_ops {
Some(blob_ops) => {
let blobs_db = self.blobs_db.as_ref().unwrap_or(&self.cold_db);
// Update the blob cache whilst holding a lock, while holding a lock on the block
// cache, to ensure they and their databases all update atomically.
let mut guard_blob = self.blob_cache.lock();
for op in &blob_ops {
match op { match op {
StoreOp::PutBlobs(block_root, blobs) => { StoreOp::PutBlobs(block_root, blobs) => {
guard_blob.put(*block_root, (**blobs).clone()); guard_blob.put(block_root, (*blobs).clone());
} }
StoreOp::DeleteBlobs(block_root) => { StoreOp::DeleteBlobs(block_root) => {
guard_blob.pop(block_root); guard_blob.pop(&block_root);
} }
_ => (), _ => (),
} }
} }
blobs_db.do_atomically(self.convert_to_kv_batch(blob_ops)?)?;
Some(guard_blob)
}
None => None,
};
self.hot_db
.do_atomically(self.convert_to_kv_batch(batch)?)?;
drop(guard); drop(guard);
if let Some(guard_blob) = guard_blob {
drop(guard_blob); drop(guard_blob);
}
Ok(()) Ok(())
} }
@ -1774,7 +1777,7 @@ impl<E: EthSpec, Hot: ItemStore<E>, Cold: ItemStore<E>> HotColdDB<E, Hot, Cold>
} }
} }
let payloads_pruned = ops.len(); let payloads_pruned = ops.len();
self.do_atomically_and_update_cache(ops, None)?; self.do_atomically_with_block_and_blobs_cache(ops)?;
info!( info!(
self.log, self.log,
"Execution payload pruning complete"; "Execution payload pruning complete";
@ -2073,7 +2076,7 @@ pub fn migrate_database<E: EthSpec, Hot: ItemStore<E>, Cold: ItemStore<E>>(
} }
// Delete the states from the hot database if we got this far. // Delete the states from the hot database if we got this far.
store.do_atomically_and_update_cache(hot_db_ops, None)?; store.do_atomically_with_block_and_blobs_cache(hot_db_ops)?;
debug!( debug!(
store.log, store.log,

View File

@ -154,6 +154,7 @@ pub trait ItemStore<E: EthSpec>: KeyValueStore<E> + Sync + Send + Sized + 'stati
/// Reified key-value storage operation. Helps in modifying the storage atomically. /// Reified key-value storage operation. Helps in modifying the storage atomically.
/// See also https://github.com/sigp/lighthouse/issues/692 /// See also https://github.com/sigp/lighthouse/issues/692
#[derive(Clone)]
pub enum StoreOp<'a, E: EthSpec> { pub enum StoreOp<'a, E: EthSpec> {
PutBlock(Hash256, Arc<SignedBeaconBlock<E>>), PutBlock(Hash256, Arc<SignedBeaconBlock<E>>),
PutState(Hash256, &'a BeaconState<E>), PutState(Hash256, &'a BeaconState<E>),

View File

@ -95,7 +95,6 @@ pub fn cli_app<'a, 'b>() -> App<'a, 'b> {
.takes_value(true), .takes_value(true),
) )
.arg( .arg(
<<<<<<< HEAD
Arg::with_name("blob-prune-margin-epochs") Arg::with_name("blob-prune-margin-epochs")
.long("blob-prune-margin-epochs") .long("blob-prune-margin-epochs")
.help( .help(
@ -104,12 +103,10 @@ pub fn cli_app<'a, 'b>() -> App<'a, 'b> {
) )
.takes_value(true) .takes_value(true)
.default_value("0"), .default_value("0"),
Arg::with_name("blobs-freezer-dir") )
.long("blobs-freezer-dir") .arg(
=======
Arg::with_name("blobs-dir") Arg::with_name("blobs-dir")
.long("blobs-dir") .long("blobs-dir")
>>>>>>> 43dc3a9a4 (Fix rebase conflicts)
.value_name("DIR") .value_name("DIR")
.help("Data directory for the blobs database.") .help("Data directory for the blobs database.")
.takes_value(true), .takes_value(true),