95c8e476bc
* Remove ping protocol * Initial renaming of network services * Correct rebasing relative to latest master * Start updating types * Adds HashMapDelay struct to utils * Initial network restructure * Network restructure. Adds new types for v0.2.0 * Removes build artefacts * Shift validation to beacon chain * Temporarily remove gossip validation This is to be updated to match current optimisation efforts. * Adds AggregateAndProof * Begin rebuilding pubsub encoding/decoding * Signature hacking * Shift gossipsup decoding into eth2_libp2p * Existing EF tests passing with fake_crypto * Shifts block encoding/decoding into RPC * Delete outdated API spec * All release tests passing bar genesis state parsing * Update and test YamlConfig * Update to spec v0.10 compatible BLS * Updates to BLS EF tests * Add EF test for AggregateVerify And delete unused hash2curve tests for uncompressed points * Update EF tests to v0.10.1 * Use optional block root correctly in block proc * Use genesis fork in deposit domain. All tests pass * Fast aggregate verify test * Update REST API docs * Fix unused import * Bump spec tags to v0.10.1 * Add `seconds_per_eth1_block` to chainspec * Update to timestamp based eth1 voting scheme * Return None from `get_votes_to_consider` if block cache is empty * Handle overflows in `is_candidate_block` * Revert to failing tests * Fix eth1 data sets test * Choose default vote according to spec * Fix collect_valid_votes tests * Fix `get_votes_to_consider` to choose all eligible blocks * Uncomment winning_vote tests * Add comments; remove unused code * Reduce seconds_per_eth1_block for simulation * Addressed review comments * Add test for default vote case * Fix logs * Remove unused functions * Meter default eth1 votes * Fix comments * Progress on attestation service * Address review comments; remove unused dependency * Initial work on removing libp2p lock * Add LRU caches to store (rollup) * Update attestation validation for DB changes (WIP) * Initial version of should_forward_block * Scaffold * Progress on attestation validation Also, consolidate prod+testing slot clocks so that they share much of the same implementation and can both handle sub-slot time changes. * Removes lock from libp2p service * Completed network lock removal * Finish(?) attestation processing * Correct network termination future * Add slot check to block check * Correct fmt issues * Remove Drop implementation for network service * Add first attempt at attestation proc. re-write * Add version 2 of attestation processing * Minor fixes * Add validator pubkey cache * Make get_indexed_attestation take a committee * Link signature processing into new attn verification * First working version * Ensure pubkey cache is updated * Add more metrics, slight optimizations * Clone committee cache during attestation processing * Update shuffling cache during block processing * Remove old commented-out code * Fix shuffling cache insert bug * Used indexed attestation in fork choice * Restructure attn processing, add metrics * Add more detailed metrics * Tidy, fix failing tests * Fix failing tests, tidy * Address reviewers suggestions * Disable/delete two outdated tests * Modification of validator for subscriptions * Add slot signing to validator client * Further progress on validation subscription * Adds necessary validator subscription functionality * Add new Pubkeys struct to signature_sets * Refactor with functional approach * Update beacon chain * Clean up validator <-> beacon node http types * Add aggregator status to ValidatorDuty * Impl Clone for manual slot clock * Fix minor errors * Further progress validator client subscription * Initial subscription and aggregation handling * Remove decompressed member from pubkey bytes * Progress to modifying val client for attestation aggregation * First draft of validator client upgrade for aggregate attestations * Add hashmap for indices lookup * Add state cache, remove store cache * Only build the head committee cache * Removes lock on a network channel * Partially implement beacon node subscription http api * Correct compilation issues * Change `get_attesting_indices` to use Vec * Fix failing test * Partial implementation of timer * Adds timer, removes exit_future, http api to op pool * Partial multiple aggregate attestation handling * Permits bulk messages accross gossipsub network channel * Correct compile issues * Improve gosispsub messaging and correct rest api helpers * Added global gossipsub subscriptions * Update validator subscriptions data structs * Tidy * Re-structure validator subscriptions * Initial handling of subscriptions * Re-structure network service * Add pubkey cache persistence file * Add more comments * Integrate persistence file into builder * Add pubkey cache tests * Add HashSetDelay and introduce into attestation service * Handles validator subscriptions * Add data_dir to beacon chain builder * Remove Option in pubkey cache persistence file * Ensure consistency between datadir/data_dir * Fix failing network test * Peer subnet discovery gets queued for future subscriptions * Reorganise attestation service functions * Initial wiring of attestation service * First draft of attestation service timing logic * Correct minor typos * Tidy * Fix todos * Improve tests * Add PeerInfo to connected peers mapping * Fix compile error * Fix compile error from merge * Split up block processing metrics * Tidy * Refactor get_pubkey_from_state * Remove commented-out code * Rename state_cache -> checkpoint_cache * Rename Checkpoint -> Snapshot * Tidy, add comments * Tidy up find_head function * Change some checkpoint -> snapshot * Add tests * Expose max_len * Remove dead code * Tidy * Fix bug * Add sync-speed metric * Add first attempt at VerifiableBlock * Start integrating into beacon chain * Integrate VerifiableBlock * Rename VerifableBlock -> PartialBlockVerification * Add start of typed methods * Add progress * Add further progress * Rename structs * Add full block verification to block_processing.rs * Further beacon chain integration * Update checks for gossip * Add todo * Start adding segement verification * Add passing chain segement test * Initial integration with batch sync * Minor changes * Tidy, add more error checking * Start adding chain_segment tests * Finish invalid signature tests * Include single and gossip verified blocks in tests * Add gossip verification tests * Start adding docs * Finish adding comments to block_processing.rs * Rename block_processing.rs -> block_verification * Start removing old block processing code * Fixes beacon_chain compilation * Fix project-wide compile errors * Remove old code * Correct code to pass all tests * Fix bug with beacon proposer index * Fix shim for BlockProcessingError * Only process one epoch at a time * Fix loop in chain segment processing * Correct tests from master merge * Add caching for state.eth1_data_votes * Add BeaconChain::validator_pubkey * Revert "Add caching for state.eth1_data_votes" This reverts commit cd73dcd6434fb8d8e6bf30c5356355598ea7b78e. Co-authored-by: Grant Wuerker <gwuerker@gmail.com> Co-authored-by: Michael Sproul <michael@sigmaprime.io> Co-authored-by: Michael Sproul <micsproul@gmail.com> Co-authored-by: pawan <pawandhananjay@gmail.com> Co-authored-by: Paul Hauner <paul@paulhauner.com>
593 lines
17 KiB
Rust
593 lines
17 KiB
Rust
#![cfg(not(debug_assertions))]
|
|
|
|
#[macro_use]
|
|
extern crate lazy_static;
|
|
|
|
use beacon_chain::test_utils::{
|
|
AttestationStrategy, BeaconChainHarness, BlockStrategy, HarnessType, OP_POOL_DB_KEY,
|
|
};
|
|
use beacon_chain::AttestationProcessingOutcome;
|
|
use operation_pool::PersistedOperationPool;
|
|
use state_processing::{
|
|
per_slot_processing, per_slot_processing::Error as SlotProcessingError, EpochProcessingError,
|
|
};
|
|
use store::Store;
|
|
use types::{BeaconStateError, EthSpec, Hash256, Keypair, MinimalEthSpec, RelativeEpoch, Slot};
|
|
|
|
// Should ideally be divisible by 3.
|
|
pub const VALIDATOR_COUNT: usize = 24;
|
|
|
|
lazy_static! {
|
|
/// A cached set of keys.
|
|
static ref KEYPAIRS: Vec<Keypair> = types::test_utils::generate_deterministic_keypairs(VALIDATOR_COUNT);
|
|
}
|
|
|
|
fn get_harness(validator_count: usize) -> BeaconChainHarness<HarnessType<MinimalEthSpec>> {
|
|
let harness = BeaconChainHarness::new(MinimalEthSpec, KEYPAIRS[0..validator_count].to_vec());
|
|
|
|
harness.advance_slot();
|
|
|
|
harness
|
|
}
|
|
|
|
#[test]
|
|
fn massive_skips() {
|
|
let harness = get_harness(8);
|
|
let spec = &MinimalEthSpec::default_spec();
|
|
let mut state = harness.chain.head().expect("should get head").beacon_state;
|
|
|
|
// Run per_slot_processing until it returns an error.
|
|
let error = loop {
|
|
match per_slot_processing(&mut state, None, spec) {
|
|
Ok(_) => continue,
|
|
Err(e) => break e,
|
|
}
|
|
};
|
|
|
|
assert!(state.slot > 1, "the state should skip at least one slot");
|
|
assert_eq!(
|
|
error,
|
|
SlotProcessingError::EpochProcessingError(EpochProcessingError::BeaconStateError(
|
|
BeaconStateError::InsufficientValidators
|
|
)),
|
|
"should return error indicating that validators have been slashed out"
|
|
)
|
|
}
|
|
|
|
#[test]
|
|
fn iterators() {
|
|
let num_blocks_produced = MinimalEthSpec::slots_per_epoch() * 2 - 1;
|
|
|
|
let harness = get_harness(VALIDATOR_COUNT);
|
|
|
|
harness.extend_chain(
|
|
num_blocks_produced as usize,
|
|
BlockStrategy::OnCanonicalHead,
|
|
// No need to produce attestations for this test.
|
|
AttestationStrategy::SomeValidators(vec![]),
|
|
);
|
|
|
|
let block_roots: Vec<(Hash256, Slot)> = harness
|
|
.chain
|
|
.rev_iter_block_roots()
|
|
.expect("should get iter")
|
|
.collect();
|
|
let state_roots: Vec<(Hash256, Slot)> = harness
|
|
.chain
|
|
.rev_iter_state_roots()
|
|
.expect("should get iter")
|
|
.collect();
|
|
|
|
assert_eq!(
|
|
block_roots.len(),
|
|
state_roots.len(),
|
|
"should be an equal amount of block and state roots"
|
|
);
|
|
|
|
assert!(
|
|
block_roots.iter().any(|(_root, slot)| *slot == 0),
|
|
"should contain genesis block root"
|
|
);
|
|
assert!(
|
|
state_roots.iter().any(|(_root, slot)| *slot == 0),
|
|
"should contain genesis state root"
|
|
);
|
|
|
|
assert_eq!(
|
|
block_roots.len(),
|
|
num_blocks_produced as usize + 1,
|
|
"should contain all produced blocks, plus the genesis block"
|
|
);
|
|
|
|
block_roots.windows(2).for_each(|x| {
|
|
assert_eq!(
|
|
x[1].1,
|
|
x[0].1 - 1,
|
|
"block root slots should be decreasing by one"
|
|
)
|
|
});
|
|
state_roots.windows(2).for_each(|x| {
|
|
assert_eq!(
|
|
x[1].1,
|
|
x[0].1 - 1,
|
|
"state root slots should be decreasing by one"
|
|
)
|
|
});
|
|
|
|
let head = &harness.chain.head().expect("should get head");
|
|
|
|
assert_eq!(
|
|
*block_roots.first().expect("should have some block roots"),
|
|
(head.beacon_block_root, head.beacon_block.slot()),
|
|
"first block root and slot should be for the head block"
|
|
);
|
|
|
|
assert_eq!(
|
|
*state_roots.first().expect("should have some state roots"),
|
|
(head.beacon_state_root, head.beacon_state.slot),
|
|
"first state root and slot should be for the head state"
|
|
);
|
|
}
|
|
|
|
#[test]
|
|
fn chooses_fork() {
|
|
let harness = get_harness(VALIDATOR_COUNT);
|
|
|
|
let two_thirds = (VALIDATOR_COUNT / 3) * 2;
|
|
let delay = MinimalEthSpec::default_spec().min_attestation_inclusion_delay as usize;
|
|
|
|
let honest_validators: Vec<usize> = (0..two_thirds).collect();
|
|
let faulty_validators: Vec<usize> = (two_thirds..VALIDATOR_COUNT).collect();
|
|
|
|
let initial_blocks = delay + 1;
|
|
let honest_fork_blocks = delay + 1;
|
|
let faulty_fork_blocks = delay + 2;
|
|
|
|
// Build an initial chain where all validators agree.
|
|
harness.extend_chain(
|
|
initial_blocks,
|
|
BlockStrategy::OnCanonicalHead,
|
|
AttestationStrategy::AllValidators,
|
|
);
|
|
|
|
let (honest_head, faulty_head) = harness.generate_two_forks_by_skipping_a_block(
|
|
&honest_validators,
|
|
&faulty_validators,
|
|
honest_fork_blocks,
|
|
faulty_fork_blocks,
|
|
);
|
|
|
|
assert!(honest_head != faulty_head, "forks should be distinct");
|
|
|
|
let state = &harness.chain.head().expect("should get head").beacon_state;
|
|
|
|
assert_eq!(
|
|
state.slot,
|
|
Slot::from(initial_blocks + honest_fork_blocks),
|
|
"head should be at the current slot"
|
|
);
|
|
|
|
assert_eq!(
|
|
harness
|
|
.chain
|
|
.head()
|
|
.expect("should get head")
|
|
.beacon_block_root,
|
|
honest_head,
|
|
"the honest chain should be the canonical chain"
|
|
);
|
|
}
|
|
|
|
#[test]
|
|
fn finalizes_with_full_participation() {
|
|
let num_blocks_produced = MinimalEthSpec::slots_per_epoch() * 5;
|
|
|
|
let harness = get_harness(VALIDATOR_COUNT);
|
|
|
|
harness.extend_chain(
|
|
num_blocks_produced as usize,
|
|
BlockStrategy::OnCanonicalHead,
|
|
AttestationStrategy::AllValidators,
|
|
);
|
|
|
|
let state = &harness.chain.head().expect("should get head").beacon_state;
|
|
|
|
assert_eq!(
|
|
state.slot, num_blocks_produced,
|
|
"head should be at the current slot"
|
|
);
|
|
assert_eq!(
|
|
state.current_epoch(),
|
|
num_blocks_produced / MinimalEthSpec::slots_per_epoch(),
|
|
"head should be at the expected epoch"
|
|
);
|
|
assert_eq!(
|
|
state.current_justified_checkpoint.epoch,
|
|
state.current_epoch() - 1,
|
|
"the head should be justified one behind the current epoch"
|
|
);
|
|
assert_eq!(
|
|
state.finalized_checkpoint.epoch,
|
|
state.current_epoch() - 2,
|
|
"the head should be finalized two behind the current epoch"
|
|
);
|
|
}
|
|
|
|
#[test]
|
|
fn finalizes_with_two_thirds_participation() {
|
|
let num_blocks_produced = MinimalEthSpec::slots_per_epoch() * 5;
|
|
|
|
let harness = get_harness(VALIDATOR_COUNT);
|
|
|
|
let two_thirds = (VALIDATOR_COUNT / 3) * 2;
|
|
let attesters = (0..two_thirds).collect();
|
|
|
|
harness.extend_chain(
|
|
num_blocks_produced as usize,
|
|
BlockStrategy::OnCanonicalHead,
|
|
AttestationStrategy::SomeValidators(attesters),
|
|
);
|
|
|
|
let state = &harness.chain.head().expect("should get head").beacon_state;
|
|
|
|
assert_eq!(
|
|
state.slot, num_blocks_produced,
|
|
"head should be at the current slot"
|
|
);
|
|
assert_eq!(
|
|
state.current_epoch(),
|
|
num_blocks_produced / MinimalEthSpec::slots_per_epoch(),
|
|
"head should be at the expected epoch"
|
|
);
|
|
|
|
// Note: the 2/3rds tests are not justifying the immediately prior epochs because the
|
|
// `MIN_ATTESTATION_INCLUSION_DELAY` is preventing an adequate number of attestations being
|
|
// included in blocks during that epoch.
|
|
|
|
assert_eq!(
|
|
state.current_justified_checkpoint.epoch,
|
|
state.current_epoch() - 2,
|
|
"the head should be justified two behind the current epoch"
|
|
);
|
|
assert_eq!(
|
|
state.finalized_checkpoint.epoch,
|
|
state.current_epoch() - 4,
|
|
"the head should be finalized three behind the current epoch"
|
|
);
|
|
}
|
|
|
|
#[test]
|
|
fn does_not_finalize_with_less_than_two_thirds_participation() {
|
|
let num_blocks_produced = MinimalEthSpec::slots_per_epoch() * 5;
|
|
|
|
let harness = get_harness(VALIDATOR_COUNT);
|
|
|
|
let two_thirds = (VALIDATOR_COUNT / 3) * 2;
|
|
let less_than_two_thirds = two_thirds - 1;
|
|
let attesters = (0..less_than_two_thirds).collect();
|
|
|
|
harness.extend_chain(
|
|
num_blocks_produced as usize,
|
|
BlockStrategy::OnCanonicalHead,
|
|
AttestationStrategy::SomeValidators(attesters),
|
|
);
|
|
|
|
let state = &harness.chain.head().expect("should get head").beacon_state;
|
|
|
|
assert_eq!(
|
|
state.slot, num_blocks_produced,
|
|
"head should be at the current slot"
|
|
);
|
|
assert_eq!(
|
|
state.current_epoch(),
|
|
num_blocks_produced / MinimalEthSpec::slots_per_epoch(),
|
|
"head should be at the expected epoch"
|
|
);
|
|
assert_eq!(
|
|
state.current_justified_checkpoint.epoch, 0,
|
|
"no epoch should have been justified"
|
|
);
|
|
assert_eq!(
|
|
state.finalized_checkpoint.epoch, 0,
|
|
"no epoch should have been finalized"
|
|
);
|
|
}
|
|
|
|
#[test]
|
|
fn does_not_finalize_without_attestation() {
|
|
let num_blocks_produced = MinimalEthSpec::slots_per_epoch() * 5;
|
|
|
|
let harness = get_harness(VALIDATOR_COUNT);
|
|
|
|
harness.extend_chain(
|
|
num_blocks_produced as usize,
|
|
BlockStrategy::OnCanonicalHead,
|
|
AttestationStrategy::SomeValidators(vec![]),
|
|
);
|
|
|
|
let state = &harness.chain.head().expect("should get head").beacon_state;
|
|
|
|
assert_eq!(
|
|
state.slot, num_blocks_produced,
|
|
"head should be at the current slot"
|
|
);
|
|
assert_eq!(
|
|
state.current_epoch(),
|
|
num_blocks_produced / MinimalEthSpec::slots_per_epoch(),
|
|
"head should be at the expected epoch"
|
|
);
|
|
assert_eq!(
|
|
state.current_justified_checkpoint.epoch, 0,
|
|
"no epoch should have been justified"
|
|
);
|
|
assert_eq!(
|
|
state.finalized_checkpoint.epoch, 0,
|
|
"no epoch should have been finalized"
|
|
);
|
|
}
|
|
|
|
#[test]
|
|
fn roundtrip_operation_pool() {
|
|
let num_blocks_produced = MinimalEthSpec::slots_per_epoch() * 5;
|
|
|
|
let harness = get_harness(VALIDATOR_COUNT);
|
|
|
|
// Add some attestations
|
|
harness.extend_chain(
|
|
num_blocks_produced as usize,
|
|
BlockStrategy::OnCanonicalHead,
|
|
AttestationStrategy::AllValidators,
|
|
);
|
|
assert!(harness.chain.op_pool.num_attestations() > 0);
|
|
|
|
// TODO: could add some other operations
|
|
harness
|
|
.chain
|
|
.persist_op_pool()
|
|
.expect("should persist op pool");
|
|
|
|
let head_state = harness.chain.head().expect("should get head").beacon_state;
|
|
|
|
let key = Hash256::from_slice(&OP_POOL_DB_KEY);
|
|
let restored_op_pool = harness
|
|
.chain
|
|
.store
|
|
.get::<PersistedOperationPool<MinimalEthSpec>>(&key)
|
|
.expect("should read db")
|
|
.expect("should find op pool")
|
|
.into_operation_pool(&head_state, &harness.spec);
|
|
|
|
assert_eq!(harness.chain.op_pool, restored_op_pool);
|
|
}
|
|
|
|
#[test]
|
|
fn free_attestations_added_to_fork_choice_some_none() {
|
|
let num_blocks_produced = MinimalEthSpec::slots_per_epoch() / 2;
|
|
|
|
let harness = get_harness(VALIDATOR_COUNT);
|
|
|
|
harness.extend_chain(
|
|
num_blocks_produced as usize,
|
|
BlockStrategy::OnCanonicalHead,
|
|
AttestationStrategy::AllValidators,
|
|
);
|
|
|
|
let state = &harness.chain.head().expect("should get head").beacon_state;
|
|
let fork_choice = &harness.chain.fork_choice;
|
|
|
|
let validator_slots: Vec<(usize, Slot)> = (0..VALIDATOR_COUNT)
|
|
.into_iter()
|
|
.map(|validator_index| {
|
|
let slot = state
|
|
.get_attestation_duties(validator_index, RelativeEpoch::Current)
|
|
.expect("should get attester duties")
|
|
.unwrap()
|
|
.slot;
|
|
|
|
(validator_index, slot)
|
|
})
|
|
.collect();
|
|
|
|
for (validator, slot) in validator_slots.clone() {
|
|
let latest_message = fork_choice.latest_message(validator);
|
|
|
|
if slot <= num_blocks_produced && slot != 0 {
|
|
assert_eq!(
|
|
latest_message.unwrap().1,
|
|
slot.epoch(MinimalEthSpec::slots_per_epoch()),
|
|
"Latest message slot for {} should be equal to slot {}.",
|
|
validator,
|
|
slot
|
|
)
|
|
} else {
|
|
assert!(
|
|
latest_message.is_none(),
|
|
"Latest message slot should be None."
|
|
)
|
|
}
|
|
}
|
|
}
|
|
|
|
#[test]
|
|
fn attestations_with_increasing_slots() {
|
|
let num_blocks_produced = MinimalEthSpec::slots_per_epoch() * 5;
|
|
|
|
let harness = get_harness(VALIDATOR_COUNT);
|
|
|
|
let mut attestations = vec![];
|
|
|
|
for _ in 0..num_blocks_produced {
|
|
harness.extend_chain(
|
|
2,
|
|
BlockStrategy::OnCanonicalHead,
|
|
// Don't produce & include any attestations (we'll collect them later).
|
|
AttestationStrategy::SomeValidators(vec![]),
|
|
);
|
|
|
|
attestations.append(
|
|
&mut harness.get_free_attestations(
|
|
&AttestationStrategy::AllValidators,
|
|
&harness.chain.head().expect("should get head").beacon_state,
|
|
harness
|
|
.chain
|
|
.head()
|
|
.expect("should get head")
|
|
.beacon_block_root,
|
|
harness
|
|
.chain
|
|
.head()
|
|
.expect("should get head")
|
|
.beacon_block
|
|
.slot(),
|
|
),
|
|
);
|
|
|
|
harness.advance_slot();
|
|
}
|
|
|
|
let current_epoch = harness.chain.epoch().expect("should get epoch");
|
|
|
|
for attestation in attestations {
|
|
let attestation_epoch = attestation.data.target.epoch;
|
|
let res = harness.chain.process_attestation(attestation);
|
|
|
|
if attestation_epoch + 1 < current_epoch {
|
|
assert_eq!(
|
|
res,
|
|
Ok(AttestationProcessingOutcome::PastEpoch {
|
|
attestation_epoch,
|
|
current_epoch,
|
|
})
|
|
)
|
|
} else {
|
|
assert_eq!(res, Ok(AttestationProcessingOutcome::Processed))
|
|
}
|
|
}
|
|
}
|
|
|
|
#[test]
|
|
fn free_attestations_added_to_fork_choice_all_updated() {
|
|
let num_blocks_produced = MinimalEthSpec::slots_per_epoch() * 2 - 1;
|
|
|
|
let harness = get_harness(VALIDATOR_COUNT);
|
|
|
|
harness.extend_chain(
|
|
num_blocks_produced as usize,
|
|
BlockStrategy::OnCanonicalHead,
|
|
AttestationStrategy::AllValidators,
|
|
);
|
|
|
|
let state = &harness.chain.head().expect("should get head").beacon_state;
|
|
let fork_choice = &harness.chain.fork_choice;
|
|
|
|
let validators: Vec<usize> = (0..VALIDATOR_COUNT).collect();
|
|
let slots: Vec<Slot> = validators
|
|
.iter()
|
|
.map(|&v| {
|
|
state
|
|
.get_attestation_duties(v, RelativeEpoch::Current)
|
|
.expect("should get attester duties")
|
|
.unwrap()
|
|
.slot
|
|
})
|
|
.collect();
|
|
let validator_slots: Vec<(&usize, Slot)> = validators.iter().zip(slots).collect();
|
|
|
|
for (validator, slot) in validator_slots {
|
|
let latest_message = fork_choice.latest_message(*validator);
|
|
|
|
assert_eq!(
|
|
latest_message.unwrap().1,
|
|
slot.epoch(MinimalEthSpec::slots_per_epoch()),
|
|
"Latest message slot should be equal to attester duty."
|
|
);
|
|
|
|
if slot != num_blocks_produced {
|
|
let block_root = state
|
|
.get_block_root(slot)
|
|
.expect("Should get block root at slot");
|
|
|
|
assert_eq!(
|
|
latest_message.unwrap().0,
|
|
*block_root,
|
|
"Latest message block root should be equal to block at slot."
|
|
);
|
|
}
|
|
}
|
|
}
|
|
|
|
fn run_skip_slot_test(skip_slots: u64) {
|
|
let num_validators = 8;
|
|
let harness_a = get_harness(num_validators);
|
|
let harness_b = get_harness(num_validators);
|
|
|
|
for _ in 0..skip_slots {
|
|
harness_a.advance_slot();
|
|
harness_b.advance_slot();
|
|
}
|
|
|
|
harness_a.extend_chain(
|
|
1,
|
|
BlockStrategy::OnCanonicalHead,
|
|
// No attestation required for test.
|
|
AttestationStrategy::SomeValidators(vec![]),
|
|
);
|
|
|
|
assert_eq!(
|
|
harness_a
|
|
.chain
|
|
.head()
|
|
.expect("should get head")
|
|
.beacon_block
|
|
.slot(),
|
|
Slot::new(skip_slots + 1)
|
|
);
|
|
assert_eq!(
|
|
harness_b
|
|
.chain
|
|
.head()
|
|
.expect("should get head")
|
|
.beacon_block
|
|
.slot(),
|
|
Slot::new(0)
|
|
);
|
|
|
|
assert_eq!(
|
|
harness_b.chain.process_block(
|
|
harness_a
|
|
.chain
|
|
.head()
|
|
.expect("should get head")
|
|
.beacon_block
|
|
.clone(),
|
|
),
|
|
Ok(harness_a
|
|
.chain
|
|
.head()
|
|
.expect("should get head")
|
|
.beacon_block_root)
|
|
);
|
|
|
|
harness_b
|
|
.chain
|
|
.fork_choice()
|
|
.expect("should run fork choice");
|
|
|
|
assert_eq!(
|
|
harness_b
|
|
.chain
|
|
.head()
|
|
.expect("should get head")
|
|
.beacon_block
|
|
.slot(),
|
|
Slot::new(skip_slots + 1)
|
|
);
|
|
}
|
|
|
|
#[test]
|
|
fn produces_and_processes_with_genesis_skip_slots() {
|
|
for i in 0..MinimalEthSpec::slots_per_epoch() * 4 {
|
|
run_skip_slot_test(i)
|
|
}
|
|
}
|