775d222299
## Proposed Changes With proposer boosting implemented (#2822) we have an opportunity to re-org out late blocks. This PR adds three flags to the BN to control this behaviour: * `--disable-proposer-reorgs`: turn aggressive re-orging off (it's on by default). * `--proposer-reorg-threshold N`: attempt to orphan blocks with less than N% of the committee vote. If this parameter isn't set then N defaults to 20% when the feature is enabled. * `--proposer-reorg-epochs-since-finalization N`: only attempt to re-org late blocks when the number of epochs since finalization is less than or equal to N. The default is 2 epochs, meaning re-orgs will only be attempted when the chain is finalizing optimally. For safety Lighthouse will only attempt a re-org under very specific conditions: 1. The block being proposed is 1 slot after the canonical head, and the canonical head is 1 slot after its parent. i.e. at slot `n + 1` rather than building on the block from slot `n` we build on the block from slot `n - 1`. 2. The current canonical head received less than N% of the committee vote. N should be set depending on the proposer boost fraction itself, the fraction of the network that is believed to be applying it, and the size of the largest entity that could be hoarding votes. 3. The current canonical head arrived after the attestation deadline from our perspective. This condition was only added to support suppression of forkchoiceUpdated messages, but makes intuitive sense. 4. The block is being proposed in the first 2 seconds of the slot. This gives it time to propagate and receive the proposer boost. ## Additional Info For the initial idea and background, see: https://github.com/ethereum/consensus-specs/pull/2353#issuecomment-950238004 There is also a specification for this feature here: https://github.com/ethereum/consensus-specs/pull/3034 Co-authored-by: Michael Sproul <micsproul@gmail.com> Co-authored-by: pawan <pawandhananjay@gmail.com>
212 lines
6.1 KiB
Rust
212 lines
6.1 KiB
Rust
#![cfg(not(debug_assertions))] // Tests run too slow in debug.
|
|
|
|
use beacon_chain::test_utils::BeaconChainHarness;
|
|
use execution_layer::test_utils::{generate_pow_block, Block, DEFAULT_TERMINAL_BLOCK};
|
|
use types::*;
|
|
|
|
const VALIDATOR_COUNT: usize = 32;
|
|
|
|
type E = MainnetEthSpec;
|
|
|
|
fn verify_execution_payload_chain<T: EthSpec>(chain: &[FullPayload<T>]) {
|
|
let mut prev_ep: Option<FullPayload<T>> = None;
|
|
|
|
for ep in chain {
|
|
assert!(*ep != FullPayload::default());
|
|
assert!(ep.block_hash() != ExecutionBlockHash::zero());
|
|
|
|
// Check against previous `ExecutionPayload`.
|
|
if let Some(prev_ep) = prev_ep {
|
|
assert_eq!(prev_ep.block_hash(), ep.execution_payload.parent_hash);
|
|
assert_eq!(
|
|
prev_ep.execution_payload.block_number + 1,
|
|
ep.execution_payload.block_number
|
|
);
|
|
assert!(ep.execution_payload.timestamp > prev_ep.execution_payload.timestamp);
|
|
}
|
|
prev_ep = Some(ep.clone());
|
|
}
|
|
}
|
|
|
|
#[tokio::test]
|
|
// TODO(merge): This isn't working cause the non-zero values in `initialize_beacon_state_from_eth1`
|
|
// are causing failed lookups to the execution node. I need to come back to this.
|
|
#[should_panic]
|
|
async fn merge_with_terminal_block_hash_override() {
|
|
let altair_fork_epoch = Epoch::new(0);
|
|
let bellatrix_fork_epoch = Epoch::new(0);
|
|
|
|
let mut spec = E::default_spec();
|
|
spec.altair_fork_epoch = Some(altair_fork_epoch);
|
|
spec.bellatrix_fork_epoch = Some(bellatrix_fork_epoch);
|
|
|
|
let genesis_pow_block_hash = generate_pow_block(
|
|
spec.terminal_total_difficulty,
|
|
DEFAULT_TERMINAL_BLOCK,
|
|
0,
|
|
ExecutionBlockHash::zero(),
|
|
)
|
|
.unwrap()
|
|
.block_hash;
|
|
|
|
spec.terminal_block_hash = genesis_pow_block_hash;
|
|
|
|
let harness = BeaconChainHarness::builder(E::default())
|
|
.spec(spec)
|
|
.logger(logging::test_logger())
|
|
.deterministic_keypairs(VALIDATOR_COUNT)
|
|
.fresh_ephemeral_store()
|
|
.mock_execution_layer()
|
|
.build();
|
|
|
|
assert_eq!(
|
|
harness
|
|
.execution_block_generator()
|
|
.latest_block()
|
|
.unwrap()
|
|
.block_hash(),
|
|
genesis_pow_block_hash,
|
|
"pre-condition"
|
|
);
|
|
|
|
assert!(
|
|
harness
|
|
.chain
|
|
.head_snapshot()
|
|
.beacon_block
|
|
.as_merge()
|
|
.is_ok(),
|
|
"genesis block should be a merge block"
|
|
);
|
|
|
|
let mut execution_payloads = vec![];
|
|
for i in 0..E::slots_per_epoch() * 3 {
|
|
harness.extend_slots(1).await;
|
|
|
|
let block = &harness.chain.head_snapshot().beacon_block;
|
|
|
|
let execution_payload = block.message().body().execution_payload().unwrap().clone();
|
|
if i == 0 {
|
|
assert_eq!(execution_payload.block_hash(), genesis_pow_block_hash);
|
|
}
|
|
execution_payloads.push(execution_payload);
|
|
}
|
|
|
|
verify_execution_payload_chain(execution_payloads.as_slice());
|
|
}
|
|
|
|
#[tokio::test]
|
|
async fn base_altair_merge_with_terminal_block_after_fork() {
|
|
let altair_fork_epoch = Epoch::new(4);
|
|
let altair_fork_slot = altair_fork_epoch.start_slot(E::slots_per_epoch());
|
|
let bellatrix_fork_epoch = Epoch::new(8);
|
|
let merge_fork_slot = bellatrix_fork_epoch.start_slot(E::slots_per_epoch());
|
|
|
|
let mut spec = E::default_spec();
|
|
spec.altair_fork_epoch = Some(altair_fork_epoch);
|
|
spec.bellatrix_fork_epoch = Some(bellatrix_fork_epoch);
|
|
|
|
let mut execution_payloads = vec![];
|
|
|
|
let harness = BeaconChainHarness::builder(E::default())
|
|
.spec(spec)
|
|
.logger(logging::test_logger())
|
|
.deterministic_keypairs(VALIDATOR_COUNT)
|
|
.fresh_ephemeral_store()
|
|
.mock_execution_layer()
|
|
.build();
|
|
|
|
/*
|
|
* Start with the base fork.
|
|
*/
|
|
|
|
assert!(harness.chain.head_snapshot().beacon_block.as_base().is_ok());
|
|
|
|
/*
|
|
* Do the Altair fork.
|
|
*/
|
|
|
|
harness.extend_to_slot(altair_fork_slot).await;
|
|
|
|
let altair_head = &harness.chain.head_snapshot().beacon_block;
|
|
assert!(altair_head.as_altair().is_ok());
|
|
assert_eq!(altair_head.slot(), altair_fork_slot);
|
|
|
|
/*
|
|
* Do the merge fork, without a terminal PoW block.
|
|
*/
|
|
|
|
harness.extend_to_slot(merge_fork_slot).await;
|
|
|
|
let merge_head = &harness.chain.head_snapshot().beacon_block;
|
|
assert!(merge_head.as_merge().is_ok());
|
|
assert_eq!(merge_head.slot(), merge_fork_slot);
|
|
assert_eq!(
|
|
*merge_head.message().body().execution_payload().unwrap(),
|
|
FullPayload::default()
|
|
);
|
|
|
|
/*
|
|
* Next merge block shouldn't include an exec payload.
|
|
*/
|
|
|
|
harness.extend_slots(1).await;
|
|
|
|
let one_after_merge_head = &harness.chain.head_snapshot().beacon_block;
|
|
assert_eq!(
|
|
*one_after_merge_head
|
|
.message()
|
|
.body()
|
|
.execution_payload()
|
|
.unwrap(),
|
|
FullPayload::default()
|
|
);
|
|
assert_eq!(one_after_merge_head.slot(), merge_fork_slot + 1);
|
|
|
|
/*
|
|
* Trigger the terminal PoW block.
|
|
*/
|
|
|
|
harness
|
|
.execution_block_generator()
|
|
.move_to_terminal_block()
|
|
.unwrap();
|
|
|
|
// Add a slot duration to get to the next slot
|
|
let timestamp = harness.get_timestamp_at_slot() + harness.spec.seconds_per_slot;
|
|
|
|
harness
|
|
.execution_block_generator()
|
|
.modify_last_block(|block| {
|
|
if let Block::PoW(terminal_block) = block {
|
|
terminal_block.timestamp = timestamp;
|
|
}
|
|
});
|
|
|
|
harness.extend_slots(1).await;
|
|
|
|
let one_after_merge_head = &harness.chain.head_snapshot().beacon_block;
|
|
assert_eq!(
|
|
*one_after_merge_head
|
|
.message()
|
|
.body()
|
|
.execution_payload()
|
|
.unwrap(),
|
|
FullPayload::default()
|
|
);
|
|
assert_eq!(one_after_merge_head.slot(), merge_fork_slot + 2);
|
|
|
|
/*
|
|
* Next merge block should include an exec payload.
|
|
*/
|
|
|
|
for _ in 0..4 {
|
|
harness.extend_slots(1).await;
|
|
|
|
let block = &harness.chain.head_snapshot().beacon_block;
|
|
execution_payloads.push(block.message().body().execution_payload().unwrap().clone());
|
|
}
|
|
|
|
verify_execution_payload_chain(execution_payloads.as_slice());
|
|
}
|