Shift networking configuration (#4426)

## Issue Addressed
Addresses [#4401](https://github.com/sigp/lighthouse/issues/4401)

## Proposed Changes
Shift some constants into ```ChainSpec``` and remove the constant values from code space.

## Additional Info

I mostly used ```MainnetEthSpec::default_spec()``` for getting ```ChainSpec```. I wonder Did I make a mistake about that.


Co-authored-by: armaganyildirak <armaganyildirak@gmail.com>
Co-authored-by: Paul Hauner <paul@paulhauner.com>
Co-authored-by: Age Manning <Age@AgeManning.com>
Co-authored-by: Diva M <divma@protonmail.com>
This commit is contained in:
Armağan Yıldırak 2023-08-03 01:51:47 +00:00
parent 7399a54ca3
commit 3397612160
36 changed files with 523 additions and 213 deletions

1
Cargo.lock generated
View File

@ -8391,6 +8391,7 @@ dependencies = [
"merkle_proof", "merkle_proof",
"metastruct", "metastruct",
"parking_lot 0.12.1", "parking_lot 0.12.1",
"paste",
"rand 0.8.5", "rand 0.8.5",
"rand_xorshift", "rand_xorshift",
"rayon", "rayon",

View File

@ -35,10 +35,8 @@
mod batch; mod batch;
use crate::{ use crate::{
beacon_chain::{MAXIMUM_GOSSIP_CLOCK_DISPARITY, VALIDATOR_PUBKEY_CACHE_LOCK_TIMEOUT}, beacon_chain::VALIDATOR_PUBKEY_CACHE_LOCK_TIMEOUT, metrics,
metrics, observed_aggregates::ObserveOutcome, observed_attesters::Error as ObservedAttestersError,
observed_aggregates::ObserveOutcome,
observed_attesters::Error as ObservedAttestersError,
BeaconChain, BeaconChainError, BeaconChainTypes, BeaconChain, BeaconChainError, BeaconChainTypes,
}; };
use bls::verify_signature_sets; use bls::verify_signature_sets;
@ -57,8 +55,8 @@ use std::borrow::Cow;
use strum::AsRefStr; use strum::AsRefStr;
use tree_hash::TreeHash; use tree_hash::TreeHash;
use types::{ use types::{
Attestation, BeaconCommittee, CommitteeIndex, Epoch, EthSpec, Hash256, IndexedAttestation, Attestation, BeaconCommittee, ChainSpec, CommitteeIndex, Epoch, EthSpec, Hash256,
SelectionProof, SignedAggregateAndProof, Slot, SubnetId, IndexedAttestation, SelectionProof, SignedAggregateAndProof, Slot, SubnetId,
}; };
pub use batch::{batch_verify_aggregated_attestations, batch_verify_unaggregated_attestations}; pub use batch::{batch_verify_aggregated_attestations, batch_verify_unaggregated_attestations};
@ -454,7 +452,7 @@ impl<'a, T: BeaconChainTypes> IndexedAggregatedAttestation<'a, T> {
// MAXIMUM_GOSSIP_CLOCK_DISPARITY allowance). // MAXIMUM_GOSSIP_CLOCK_DISPARITY allowance).
// //
// We do not queue future attestations for later processing. // We do not queue future attestations for later processing.
verify_propagation_slot_range(&chain.slot_clock, attestation)?; verify_propagation_slot_range(&chain.slot_clock, attestation, &chain.spec)?;
// Check the attestation's epoch matches its target. // Check the attestation's epoch matches its target.
if attestation.data.slot.epoch(T::EthSpec::slots_per_epoch()) if attestation.data.slot.epoch(T::EthSpec::slots_per_epoch())
@ -722,7 +720,7 @@ impl<'a, T: BeaconChainTypes> IndexedUnaggregatedAttestation<'a, T> {
// MAXIMUM_GOSSIP_CLOCK_DISPARITY allowance). // MAXIMUM_GOSSIP_CLOCK_DISPARITY allowance).
// //
// We do not queue future attestations for later processing. // We do not queue future attestations for later processing.
verify_propagation_slot_range(&chain.slot_clock, attestation)?; verify_propagation_slot_range(&chain.slot_clock, attestation, &chain.spec)?;
// Check to ensure that the attestation is "unaggregated". I.e., it has exactly one // Check to ensure that the attestation is "unaggregated". I.e., it has exactly one
// aggregation bit set. // aggregation bit set.
@ -1037,11 +1035,11 @@ fn verify_head_block_is_known<T: BeaconChainTypes>(
pub fn verify_propagation_slot_range<S: SlotClock, E: EthSpec>( pub fn verify_propagation_slot_range<S: SlotClock, E: EthSpec>(
slot_clock: &S, slot_clock: &S,
attestation: &Attestation<E>, attestation: &Attestation<E>,
spec: &ChainSpec,
) -> Result<(), Error> { ) -> Result<(), Error> {
let attestation_slot = attestation.data.slot; let attestation_slot = attestation.data.slot;
let latest_permissible_slot = slot_clock let latest_permissible_slot = slot_clock
.now_with_future_tolerance(MAXIMUM_GOSSIP_CLOCK_DISPARITY) .now_with_future_tolerance(spec.maximum_gossip_clock_disparity())
.ok_or(BeaconChainError::UnableToReadSlot)?; .ok_or(BeaconChainError::UnableToReadSlot)?;
if attestation_slot > latest_permissible_slot { if attestation_slot > latest_permissible_slot {
return Err(Error::FutureSlot { return Err(Error::FutureSlot {
@ -1052,7 +1050,7 @@ pub fn verify_propagation_slot_range<S: SlotClock, E: EthSpec>(
// Taking advantage of saturating subtraction on `Slot`. // Taking advantage of saturating subtraction on `Slot`.
let earliest_permissible_slot = slot_clock let earliest_permissible_slot = slot_clock
.now_with_past_tolerance(MAXIMUM_GOSSIP_CLOCK_DISPARITY) .now_with_past_tolerance(spec.maximum_gossip_clock_disparity())
.ok_or(BeaconChainError::UnableToReadSlot)? .ok_or(BeaconChainError::UnableToReadSlot)?
- E::slots_per_epoch(); - E::slots_per_epoch();
if attestation_slot < earliest_permissible_slot { if attestation_slot < earliest_permissible_slot {

View File

@ -217,11 +217,6 @@ pub enum OverrideForkchoiceUpdate {
AlreadyApplied, AlreadyApplied,
} }
/// The accepted clock drift for nodes gossiping blocks and attestations. See:
///
/// https://github.com/ethereum/eth2.0-specs/blob/v0.12.1/specs/phase0/p2p-interface.md#configuration
pub const MAXIMUM_GOSSIP_CLOCK_DISPARITY: Duration = Duration::from_millis(500);
#[derive(Debug, PartialEq)] #[derive(Debug, PartialEq)]
pub enum AttestationProcessingOutcome { pub enum AttestationProcessingOutcome {
Processed, Processed,

View File

@ -59,7 +59,7 @@ use crate::validator_pubkey_cache::ValidatorPubkeyCache;
use crate::{ use crate::{
beacon_chain::{ beacon_chain::{
BeaconForkChoice, ForkChoiceError, BLOCK_PROCESSING_CACHE_LOCK_TIMEOUT, BeaconForkChoice, ForkChoiceError, BLOCK_PROCESSING_CACHE_LOCK_TIMEOUT,
MAXIMUM_GOSSIP_CLOCK_DISPARITY, VALIDATOR_PUBKEY_CACHE_LOCK_TIMEOUT, VALIDATOR_PUBKEY_CACHE_LOCK_TIMEOUT,
}, },
metrics, BeaconChain, BeaconChainError, BeaconChainTypes, metrics, BeaconChain, BeaconChainError, BeaconChainTypes,
}; };
@ -730,7 +730,7 @@ impl<T: BeaconChainTypes> GossipVerifiedBlock<T> {
// Do not gossip or process blocks from future slots. // Do not gossip or process blocks from future slots.
let present_slot_with_tolerance = chain let present_slot_with_tolerance = chain
.slot_clock .slot_clock
.now_with_future_tolerance(MAXIMUM_GOSSIP_CLOCK_DISPARITY) .now_with_future_tolerance(chain.spec.maximum_gossip_clock_disparity())
.ok_or(BeaconChainError::UnableToReadSlot)?; .ok_or(BeaconChainError::UnableToReadSlot)?;
if block.slot() > present_slot_with_tolerance { if block.slot() > present_slot_with_tolerance {
return Err(BlockError::FutureSlot { return Err(BlockError::FutureSlot {

View File

@ -54,7 +54,7 @@ pub use self::beacon_chain::{
AttestationProcessingOutcome, BeaconChain, BeaconChainTypes, BeaconStore, ChainSegmentResult, AttestationProcessingOutcome, BeaconChain, BeaconChainTypes, BeaconStore, ChainSegmentResult,
ForkChoiceError, OverrideForkchoiceUpdate, ProduceBlockVerification, StateSkipConfig, ForkChoiceError, OverrideForkchoiceUpdate, ProduceBlockVerification, StateSkipConfig,
WhenSlotSkipped, INVALID_FINALIZED_MERGE_TRANSITION_BLOCK_SHUTDOWN_REASON, WhenSlotSkipped, INVALID_FINALIZED_MERGE_TRANSITION_BLOCK_SHUTDOWN_REASON,
INVALID_JUSTIFIED_PAYLOAD_SHUTDOWN_REASON, MAXIMUM_GOSSIP_CLOCK_DISPARITY, INVALID_JUSTIFIED_PAYLOAD_SHUTDOWN_REASON,
}; };
pub use self::beacon_snapshot::BeaconSnapshot; pub use self::beacon_snapshot::BeaconSnapshot;
pub use self::chain_config::ChainConfig; pub use self::chain_config::ChainConfig;

View File

@ -1,6 +1,4 @@
use crate::{ use crate::{BeaconChain, BeaconChainError, BeaconChainTypes};
beacon_chain::MAXIMUM_GOSSIP_CLOCK_DISPARITY, BeaconChain, BeaconChainError, BeaconChainTypes,
};
use derivative::Derivative; use derivative::Derivative;
use slot_clock::SlotClock; use slot_clock::SlotClock;
use std::time::Duration; use std::time::Duration;
@ -103,7 +101,8 @@ impl<T: BeaconChainTypes> VerifiedLightClientFinalityUpdate<T> {
// verify that enough time has passed for the block to have been propagated // verify that enough time has passed for the block to have been propagated
match start_time { match start_time {
Some(time) => { Some(time) => {
if seen_timestamp + MAXIMUM_GOSSIP_CLOCK_DISPARITY < time + one_third_slot_duration if seen_timestamp + chain.spec.maximum_gossip_clock_disparity()
< time + one_third_slot_duration
{ {
return Err(Error::TooEarly); return Err(Error::TooEarly);
} }

View File

@ -1,6 +1,4 @@
use crate::{ use crate::{BeaconChain, BeaconChainError, BeaconChainTypes};
beacon_chain::MAXIMUM_GOSSIP_CLOCK_DISPARITY, BeaconChain, BeaconChainError, BeaconChainTypes,
};
use derivative::Derivative; use derivative::Derivative;
use eth2::types::Hash256; use eth2::types::Hash256;
use slot_clock::SlotClock; use slot_clock::SlotClock;
@ -103,7 +101,8 @@ impl<T: BeaconChainTypes> VerifiedLightClientOptimisticUpdate<T> {
// verify that enough time has passed for the block to have been propagated // verify that enough time has passed for the block to have been propagated
match start_time { match start_time {
Some(time) => { Some(time) => {
if seen_timestamp + MAXIMUM_GOSSIP_CLOCK_DISPARITY < time + one_third_slot_duration if seen_timestamp + chain.spec.maximum_gossip_clock_disparity()
< time + one_third_slot_duration
{ {
return Err(Error::TooEarly); return Err(Error::TooEarly);
} }

View File

@ -28,10 +28,8 @@
use crate::observed_attesters::SlotSubcommitteeIndex; use crate::observed_attesters::SlotSubcommitteeIndex;
use crate::{ use crate::{
beacon_chain::{MAXIMUM_GOSSIP_CLOCK_DISPARITY, VALIDATOR_PUBKEY_CACHE_LOCK_TIMEOUT}, beacon_chain::VALIDATOR_PUBKEY_CACHE_LOCK_TIMEOUT, metrics,
metrics, observed_aggregates::ObserveOutcome, BeaconChain, BeaconChainError, BeaconChainTypes,
observed_aggregates::ObserveOutcome,
BeaconChain, BeaconChainError, BeaconChainTypes,
}; };
use bls::{verify_signature_sets, PublicKeyBytes}; use bls::{verify_signature_sets, PublicKeyBytes};
use derivative::Derivative; use derivative::Derivative;
@ -52,6 +50,7 @@ use tree_hash_derive::TreeHash;
use types::consts::altair::SYNC_COMMITTEE_SUBNET_COUNT; use types::consts::altair::SYNC_COMMITTEE_SUBNET_COUNT;
use types::slot_data::SlotData; use types::slot_data::SlotData;
use types::sync_committee::Error as SyncCommitteeError; use types::sync_committee::Error as SyncCommitteeError;
use types::ChainSpec;
use types::{ use types::{
sync_committee_contribution::Error as ContributionError, AggregateSignature, BeaconStateError, sync_committee_contribution::Error as ContributionError, AggregateSignature, BeaconStateError,
EthSpec, Hash256, SignedContributionAndProof, Slot, SyncCommitteeContribution, EthSpec, Hash256, SignedContributionAndProof, Slot, SyncCommitteeContribution,
@ -297,7 +296,7 @@ impl<T: BeaconChainTypes> VerifiedSyncContribution<T> {
let subcommittee_index = contribution.subcommittee_index as usize; let subcommittee_index = contribution.subcommittee_index as usize;
// Ensure sync committee contribution is within the MAXIMUM_GOSSIP_CLOCK_DISPARITY allowance. // Ensure sync committee contribution is within the MAXIMUM_GOSSIP_CLOCK_DISPARITY allowance.
verify_propagation_slot_range(&chain.slot_clock, contribution)?; verify_propagation_slot_range(&chain.slot_clock, contribution, &chain.spec)?;
// Validate subcommittee index. // Validate subcommittee index.
if contribution.subcommittee_index >= SYNC_COMMITTEE_SUBNET_COUNT { if contribution.subcommittee_index >= SYNC_COMMITTEE_SUBNET_COUNT {
@ -460,7 +459,7 @@ impl VerifiedSyncCommitteeMessage {
// MAXIMUM_GOSSIP_CLOCK_DISPARITY allowance). // MAXIMUM_GOSSIP_CLOCK_DISPARITY allowance).
// //
// We do not queue future sync committee messages for later processing. // We do not queue future sync committee messages for later processing.
verify_propagation_slot_range(&chain.slot_clock, &sync_message)?; verify_propagation_slot_range(&chain.slot_clock, &sync_message, &chain.spec)?;
// Ensure the `subnet_id` is valid for the given validator. // Ensure the `subnet_id` is valid for the given validator.
let pubkey = chain let pubkey = chain
@ -576,11 +575,11 @@ impl VerifiedSyncCommitteeMessage {
pub fn verify_propagation_slot_range<S: SlotClock, U: SlotData>( pub fn verify_propagation_slot_range<S: SlotClock, U: SlotData>(
slot_clock: &S, slot_clock: &S,
sync_contribution: &U, sync_contribution: &U,
spec: &ChainSpec,
) -> Result<(), Error> { ) -> Result<(), Error> {
let message_slot = sync_contribution.get_slot(); let message_slot = sync_contribution.get_slot();
let latest_permissible_slot = slot_clock let latest_permissible_slot = slot_clock
.now_with_future_tolerance(MAXIMUM_GOSSIP_CLOCK_DISPARITY) .now_with_future_tolerance(spec.maximum_gossip_clock_disparity())
.ok_or(BeaconChainError::UnableToReadSlot)?; .ok_or(BeaconChainError::UnableToReadSlot)?;
if message_slot > latest_permissible_slot { if message_slot > latest_permissible_slot {
return Err(Error::FutureSlot { return Err(Error::FutureSlot {
@ -590,7 +589,7 @@ pub fn verify_propagation_slot_range<S: SlotClock, U: SlotData>(
} }
let earliest_permissible_slot = slot_clock let earliest_permissible_slot = slot_clock
.now_with_past_tolerance(MAXIMUM_GOSSIP_CLOCK_DISPARITY) .now_with_past_tolerance(spec.maximum_gossip_clock_disparity())
.ok_or(BeaconChainError::UnableToReadSlot)?; .ok_or(BeaconChainError::UnableToReadSlot)?;
if message_slot < earliest_permissible_slot { if message_slot < earliest_permissible_slot {

View File

@ -661,7 +661,8 @@ impl<E: EthSpec> BeaconProcessor<E> {
work_reprocessing_rx: mpsc::Receiver<ReprocessQueueMessage>, work_reprocessing_rx: mpsc::Receiver<ReprocessQueueMessage>,
work_journal_tx: Option<mpsc::Sender<&'static str>>, work_journal_tx: Option<mpsc::Sender<&'static str>>,
slot_clock: S, slot_clock: S,
) { maximum_gossip_clock_disparity: Duration,
) -> Result<(), String> {
// Used by workers to communicate that they are finished a task. // Used by workers to communicate that they are finished a task.
let (idle_tx, idle_rx) = mpsc::channel::<()>(MAX_IDLE_QUEUE_LEN); let (idle_tx, idle_rx) = mpsc::channel::<()>(MAX_IDLE_QUEUE_LEN);
@ -717,13 +718,15 @@ impl<E: EthSpec> BeaconProcessor<E> {
// receive them back once they are ready (`ready_work_rx`). // receive them back once they are ready (`ready_work_rx`).
let (ready_work_tx, ready_work_rx) = let (ready_work_tx, ready_work_rx) =
mpsc::channel::<ReadyWork>(MAX_SCHEDULED_WORK_QUEUE_LEN); mpsc::channel::<ReadyWork>(MAX_SCHEDULED_WORK_QUEUE_LEN);
spawn_reprocess_scheduler( spawn_reprocess_scheduler(
ready_work_tx, ready_work_tx,
work_reprocessing_rx, work_reprocessing_rx,
&self.executor, &self.executor,
slot_clock, slot_clock,
self.log.clone(), self.log.clone(),
); maximum_gossip_clock_disparity,
)?;
let executor = self.executor.clone(); let executor = self.executor.clone();
@ -1203,6 +1206,7 @@ impl<E: EthSpec> BeaconProcessor<E> {
// Spawn on the core executor. // Spawn on the core executor.
executor.spawn(manager_future, MANAGER_TASK_NAME); executor.spawn(manager_future, MANAGER_TASK_NAME);
Ok(())
} }
/// Spawns a blocking worker thread to process some `Work`. /// Spawns a blocking worker thread to process some `Work`.

View File

@ -361,7 +361,12 @@ pub fn spawn_reprocess_scheduler<S: SlotClock + 'static>(
executor: &TaskExecutor, executor: &TaskExecutor,
slot_clock: S, slot_clock: S,
log: Logger, log: Logger,
) { maximum_gossip_clock_disparity: Duration,
) -> Result<(), String> {
// Sanity check
if ADDITIONAL_QUEUED_BLOCK_DELAY >= maximum_gossip_clock_disparity {
return Err("The block delay and gossip disparity don't match.".to_string());
}
let mut queue = ReprocessQueue { let mut queue = ReprocessQueue {
work_reprocessing_rx, work_reprocessing_rx,
ready_work_tx, ready_work_tx,
@ -400,6 +405,7 @@ pub fn spawn_reprocess_scheduler<S: SlotClock + 'static>(
}, },
TASK_NAME, TASK_NAME,
); );
Ok(())
} }
impl<S: SlotClock> ReprocessQueue<S> { impl<S: SlotClock> ReprocessQueue<S> {

View File

@ -795,7 +795,8 @@ where
self.work_reprocessing_rx, self.work_reprocessing_rx,
None, None,
beacon_chain.slot_clock.clone(), beacon_chain.slot_clock.clone(),
); beacon_chain.spec.maximum_gossip_clock_disparity(),
)?;
} }
let state_advance_context = runtime_context.service_context("state_advance".into()); let state_advance_context = runtime_context.service_context("state_advance".into());

View File

@ -1,9 +1,7 @@
//! Contains the handler for the `GET validator/duties/attester/{epoch}` endpoint. //! Contains the handler for the `GET validator/duties/attester/{epoch}` endpoint.
use crate::state_id::StateId; use crate::state_id::StateId;
use beacon_chain::{ use beacon_chain::{BeaconChain, BeaconChainError, BeaconChainTypes};
BeaconChain, BeaconChainError, BeaconChainTypes, MAXIMUM_GOSSIP_CLOCK_DISPARITY,
};
use eth2::types::{self as api_types}; use eth2::types::{self as api_types};
use slot_clock::SlotClock; use slot_clock::SlotClock;
use state_processing::state_advance::partial_state_advance; use state_processing::state_advance::partial_state_advance;
@ -32,7 +30,7 @@ pub fn attester_duties<T: BeaconChainTypes>(
// will equal `current_epoch + 1` // will equal `current_epoch + 1`
let tolerant_current_epoch = chain let tolerant_current_epoch = chain
.slot_clock .slot_clock
.now_with_future_tolerance(MAXIMUM_GOSSIP_CLOCK_DISPARITY) .now_with_future_tolerance(chain.spec.maximum_gossip_clock_disparity())
.ok_or_else(|| warp_utils::reject::custom_server_error("unable to read slot clock".into()))? .ok_or_else(|| warp_utils::reject::custom_server_error("unable to read slot clock".into()))?
.epoch(T::EthSpec::slots_per_epoch()); .epoch(T::EthSpec::slots_per_epoch());

View File

@ -3,7 +3,7 @@
use crate::state_id::StateId; use crate::state_id::StateId;
use beacon_chain::{ use beacon_chain::{
beacon_proposer_cache::{compute_proposer_duties_from_head, ensure_state_is_in_epoch}, beacon_proposer_cache::{compute_proposer_duties_from_head, ensure_state_is_in_epoch},
BeaconChain, BeaconChainError, BeaconChainTypes, MAXIMUM_GOSSIP_CLOCK_DISPARITY, BeaconChain, BeaconChainError, BeaconChainTypes,
}; };
use eth2::types::{self as api_types}; use eth2::types::{self as api_types};
use safe_arith::SafeArith; use safe_arith::SafeArith;
@ -33,7 +33,7 @@ pub fn proposer_duties<T: BeaconChainTypes>(
// will equal `current_epoch + 1` // will equal `current_epoch + 1`
let tolerant_current_epoch = chain let tolerant_current_epoch = chain
.slot_clock .slot_clock
.now_with_future_tolerance(MAXIMUM_GOSSIP_CLOCK_DISPARITY) .now_with_future_tolerance(chain.spec.maximum_gossip_clock_disparity())
.ok_or_else(|| warp_utils::reject::custom_server_error("unable to read slot clock".into()))? .ok_or_else(|| warp_utils::reject::custom_server_error("unable to read slot clock".into()))?
.epoch(T::EthSpec::slots_per_epoch()); .epoch(T::EthSpec::slots_per_epoch());

View File

@ -6,7 +6,7 @@ use beacon_chain::sync_committee_verification::{
}; };
use beacon_chain::{ use beacon_chain::{
validator_monitor::timestamp_now, BeaconChain, BeaconChainError, BeaconChainTypes, validator_monitor::timestamp_now, BeaconChain, BeaconChainError, BeaconChainTypes,
StateSkipConfig, MAXIMUM_GOSSIP_CLOCK_DISPARITY, StateSkipConfig,
}; };
use eth2::types::{self as api_types}; use eth2::types::{self as api_types};
use lighthouse_network::PubsubMessage; use lighthouse_network::PubsubMessage;
@ -85,7 +85,7 @@ fn duties_from_state_load<T: BeaconChainTypes>(
let current_epoch = chain.epoch()?; let current_epoch = chain.epoch()?;
let tolerant_current_epoch = chain let tolerant_current_epoch = chain
.slot_clock .slot_clock
.now_with_future_tolerance(MAXIMUM_GOSSIP_CLOCK_DISPARITY) .now_with_future_tolerance(chain.spec.maximum_gossip_clock_disparity())
.ok_or(BeaconChainError::UnableToReadSlot)? .ok_or(BeaconChainError::UnableToReadSlot)?
.epoch(T::EthSpec::slots_per_epoch()); .epoch(T::EthSpec::slots_per_epoch());

View File

@ -1,7 +1,7 @@
use beacon_chain::test_utils::RelativeSyncCommittee; use beacon_chain::test_utils::RelativeSyncCommittee;
use beacon_chain::{ use beacon_chain::{
test_utils::{AttestationStrategy, BeaconChainHarness, BlockStrategy, EphemeralHarnessType}, test_utils::{AttestationStrategy, BeaconChainHarness, BlockStrategy, EphemeralHarnessType},
BeaconChain, StateSkipConfig, WhenSlotSkipped, MAXIMUM_GOSSIP_CLOCK_DISPARITY, BeaconChain, StateSkipConfig, WhenSlotSkipped,
}; };
use environment::null_logger; use environment::null_logger;
use eth2::{ use eth2::{
@ -2313,7 +2313,9 @@ impl ApiTester {
.unwrap(); .unwrap();
self.chain.slot_clock.set_current_time( self.chain.slot_clock.set_current_time(
current_epoch_start - MAXIMUM_GOSSIP_CLOCK_DISPARITY - Duration::from_millis(1), current_epoch_start
- self.chain.spec.maximum_gossip_clock_disparity()
- Duration::from_millis(1),
); );
let dependent_root = self let dependent_root = self
@ -2350,9 +2352,9 @@ impl ApiTester {
"should not get attester duties outside of tolerance" "should not get attester duties outside of tolerance"
); );
self.chain self.chain.slot_clock.set_current_time(
.slot_clock current_epoch_start - self.chain.spec.maximum_gossip_clock_disparity(),
.set_current_time(current_epoch_start - MAXIMUM_GOSSIP_CLOCK_DISPARITY); );
self.client self.client
.get_validator_duties_proposer(current_epoch) .get_validator_duties_proposer(current_epoch)

View File

@ -16,11 +16,6 @@ use std::sync::Arc;
use std::time::Duration; use std::time::Duration;
use types::{ForkContext, ForkName}; use types::{ForkContext, ForkName};
/// The maximum transmit size of gossip messages in bytes pre-merge.
const GOSSIP_MAX_SIZE: usize = 1_048_576; // 1M
/// The maximum transmit size of gossip messages in bytes post-merge.
const GOSSIP_MAX_SIZE_POST_MERGE: usize = 10 * 1_048_576; // 10M
/// The cache time is set to accommodate the circulation time of an attestation. /// The cache time is set to accommodate the circulation time of an attestation.
/// ///
/// The p2p spec declares that we accept attestations within the following range: /// The p2p spec declares that we accept attestations within the following range:
@ -35,20 +30,20 @@ const GOSSIP_MAX_SIZE_POST_MERGE: usize = 10 * 1_048_576; // 10M
/// another 500ms for "fudge factor". /// another 500ms for "fudge factor".
pub const DUPLICATE_CACHE_TIME: Duration = Duration::from_secs(33 * 12 + 1); pub const DUPLICATE_CACHE_TIME: Duration = Duration::from_secs(33 * 12 + 1);
// We treat uncompressed messages as invalid and never use the INVALID_SNAPPY_DOMAIN as in the
// specification. We leave it here for posterity.
// const MESSAGE_DOMAIN_INVALID_SNAPPY: [u8; 4] = [0, 0, 0, 0];
const MESSAGE_DOMAIN_VALID_SNAPPY: [u8; 4] = [1, 0, 0, 0];
/// The maximum size of gossip messages. /// The maximum size of gossip messages.
pub fn gossip_max_size(is_merge_enabled: bool) -> usize { pub fn gossip_max_size(is_merge_enabled: bool, gossip_max_size: usize) -> usize {
if is_merge_enabled { if is_merge_enabled {
GOSSIP_MAX_SIZE_POST_MERGE gossip_max_size
} else { } else {
GOSSIP_MAX_SIZE gossip_max_size / 10
} }
} }
pub struct GossipsubConfigParams {
pub message_domain_valid_snappy: [u8; 4],
pub gossip_max_size: usize,
}
#[derive(Clone, Debug, Serialize, Deserialize)] #[derive(Clone, Debug, Serialize, Deserialize)]
#[serde(default)] #[serde(default)]
/// Network configuration for lighthouse. /// Network configuration for lighthouse.
@ -413,7 +408,11 @@ impl From<u8> for NetworkLoad {
} }
/// Return a Lighthouse specific `GossipsubConfig` where the `message_id_fn` depends on the current fork. /// Return a Lighthouse specific `GossipsubConfig` where the `message_id_fn` depends on the current fork.
pub fn gossipsub_config(network_load: u8, fork_context: Arc<ForkContext>) -> gossipsub::Config { pub fn gossipsub_config(
network_load: u8,
fork_context: Arc<ForkContext>,
gossipsub_config_params: GossipsubConfigParams,
) -> gossipsub::Config {
// The function used to generate a gossipsub message id // The function used to generate a gossipsub message id
// We use the first 8 bytes of SHA256(topic, data) for content addressing // We use the first 8 bytes of SHA256(topic, data) for content addressing
let fast_gossip_message_id = |message: &gossipsub::RawMessage| { let fast_gossip_message_id = |message: &gossipsub::RawMessage| {
@ -446,12 +445,12 @@ pub fn gossipsub_config(network_load: u8, fork_context: Arc<ForkContext>) -> gos
} }
} }
} }
let message_domain_valid_snappy = gossipsub_config_params.message_domain_valid_snappy;
let is_merge_enabled = fork_context.fork_exists(ForkName::Merge); let is_merge_enabled = fork_context.fork_exists(ForkName::Merge);
let gossip_message_id = move |message: &gossipsub::Message| { let gossip_message_id = move |message: &gossipsub::Message| {
gossipsub::MessageId::from( gossipsub::MessageId::from(
&Sha256::digest( &Sha256::digest(
prefix(MESSAGE_DOMAIN_VALID_SNAPPY, message, fork_context.clone()).as_slice(), prefix(message_domain_valid_snappy, message, fork_context.clone()).as_slice(),
)[..20], )[..20],
) )
}; };
@ -459,7 +458,10 @@ pub fn gossipsub_config(network_load: u8, fork_context: Arc<ForkContext>) -> gos
let load = NetworkLoad::from(network_load); let load = NetworkLoad::from(network_load);
gossipsub::ConfigBuilder::default() gossipsub::ConfigBuilder::default()
.max_transmit_size(gossip_max_size(is_merge_enabled)) .max_transmit_size(gossip_max_size(
is_merge_enabled,
gossipsub_config_params.gossip_max_size,
))
.heartbeat_interval(load.heartbeat_interval) .heartbeat_interval(load.heartbeat_interval)
.mesh_n(load.mesh_n) .mesh_n(load.mesh_n)
.mesh_n_low(load.mesh_n_low) .mesh_n_low(load.mesh_n_low)

View File

@ -217,9 +217,12 @@ mod tests {
let snappy_protocol_id = ProtocolId::new(SupportedProtocol::StatusV1, Encoding::SSZSnappy); let snappy_protocol_id = ProtocolId::new(SupportedProtocol::StatusV1, Encoding::SSZSnappy);
let fork_context = Arc::new(fork_context(ForkName::Base)); let fork_context = Arc::new(fork_context(ForkName::Base));
let chain_spec = Spec::default_spec();
let mut snappy_outbound_codec = SSZSnappyOutboundCodec::<Spec>::new( let mut snappy_outbound_codec = SSZSnappyOutboundCodec::<Spec>::new(
snappy_protocol_id, snappy_protocol_id,
max_rpc_size(&fork_context), max_rpc_size(&fork_context, chain_spec.max_chunk_size as usize),
fork_context, fork_context,
); );
@ -251,9 +254,12 @@ mod tests {
let snappy_protocol_id = ProtocolId::new(SupportedProtocol::StatusV1, Encoding::SSZSnappy); let snappy_protocol_id = ProtocolId::new(SupportedProtocol::StatusV1, Encoding::SSZSnappy);
let fork_context = Arc::new(fork_context(ForkName::Base)); let fork_context = Arc::new(fork_context(ForkName::Base));
let chain_spec = Spec::default_spec();
let mut snappy_outbound_codec = SSZSnappyOutboundCodec::<Spec>::new( let mut snappy_outbound_codec = SSZSnappyOutboundCodec::<Spec>::new(
snappy_protocol_id, snappy_protocol_id,
max_rpc_size(&fork_context), max_rpc_size(&fork_context, chain_spec.max_chunk_size as usize),
fork_context, fork_context,
); );
@ -279,7 +285,10 @@ mod tests {
// Response limits // Response limits
let fork_context = Arc::new(fork_context(ForkName::Base)); let fork_context = Arc::new(fork_context(ForkName::Base));
let max_rpc_size = max_rpc_size(&fork_context);
let chain_spec = Spec::default_spec();
let max_rpc_size = max_rpc_size(&fork_context, chain_spec.max_chunk_size as usize);
let limit = protocol_id.rpc_response_limits::<Spec>(&fork_context); let limit = protocol_id.rpc_response_limits::<Spec>(&fork_context);
let mut max = encode_len(limit.max + 1); let mut max = encode_len(limit.max + 1);
let mut codec = SSZSnappyOutboundCodec::<Spec>::new( let mut codec = SSZSnappyOutboundCodec::<Spec>::new(

View File

@ -615,8 +615,8 @@ mod tests {
}; };
use std::sync::Arc; use std::sync::Arc;
use types::{ use types::{
BeaconBlock, BeaconBlockAltair, BeaconBlockBase, BeaconBlockMerge, EmptyBlock, Epoch, BeaconBlock, BeaconBlockAltair, BeaconBlockBase, BeaconBlockMerge, ChainSpec, EmptyBlock,
ForkContext, FullPayload, Hash256, Signature, SignedBeaconBlock, Slot, Epoch, ForkContext, FullPayload, Hash256, Signature, SignedBeaconBlock, Slot,
}; };
use snap::write::FrameEncoder; use snap::write::FrameEncoder;
@ -658,7 +658,7 @@ mod tests {
} }
/// Merge block with length < max_rpc_size. /// Merge block with length < max_rpc_size.
fn merge_block_small(fork_context: &ForkContext) -> SignedBeaconBlock<Spec> { fn merge_block_small(fork_context: &ForkContext, spec: &ChainSpec) -> SignedBeaconBlock<Spec> {
let mut block: BeaconBlockMerge<_, FullPayload<Spec>> = let mut block: BeaconBlockMerge<_, FullPayload<Spec>> =
BeaconBlockMerge::empty(&Spec::default_spec()); BeaconBlockMerge::empty(&Spec::default_spec());
let tx = VariableList::from(vec![0; 1024]); let tx = VariableList::from(vec![0; 1024]);
@ -667,14 +667,14 @@ mod tests {
block.body.execution_payload.execution_payload.transactions = txs; block.body.execution_payload.execution_payload.transactions = txs;
let block = BeaconBlock::Merge(block); let block = BeaconBlock::Merge(block);
assert!(block.ssz_bytes_len() <= max_rpc_size(fork_context)); assert!(block.ssz_bytes_len() <= max_rpc_size(fork_context, spec.max_chunk_size as usize));
SignedBeaconBlock::from_block(block, Signature::empty()) SignedBeaconBlock::from_block(block, Signature::empty())
} }
/// Merge block with length > MAX_RPC_SIZE. /// Merge block with length > MAX_RPC_SIZE.
/// The max limit for a merge block is in the order of ~16GiB which wouldn't fit in memory. /// The max limit for a merge block is in the order of ~16GiB which wouldn't fit in memory.
/// Hence, we generate a merge block just greater than `MAX_RPC_SIZE` to test rejection on the rpc layer. /// Hence, we generate a merge block just greater than `MAX_RPC_SIZE` to test rejection on the rpc layer.
fn merge_block_large(fork_context: &ForkContext) -> SignedBeaconBlock<Spec> { fn merge_block_large(fork_context: &ForkContext, spec: &ChainSpec) -> SignedBeaconBlock<Spec> {
let mut block: BeaconBlockMerge<_, FullPayload<Spec>> = let mut block: BeaconBlockMerge<_, FullPayload<Spec>> =
BeaconBlockMerge::empty(&Spec::default_spec()); BeaconBlockMerge::empty(&Spec::default_spec());
let tx = VariableList::from(vec![0; 1024]); let tx = VariableList::from(vec![0; 1024]);
@ -683,7 +683,7 @@ mod tests {
block.body.execution_payload.execution_payload.transactions = txs; block.body.execution_payload.execution_payload.transactions = txs;
let block = BeaconBlock::Merge(block); let block = BeaconBlock::Merge(block);
assert!(block.ssz_bytes_len() > max_rpc_size(fork_context)); assert!(block.ssz_bytes_len() > max_rpc_size(fork_context, spec.max_chunk_size as usize));
SignedBeaconBlock::from_block(block, Signature::empty()) SignedBeaconBlock::from_block(block, Signature::empty())
} }
@ -737,10 +737,11 @@ mod tests {
protocol: SupportedProtocol, protocol: SupportedProtocol,
message: RPCCodedResponse<Spec>, message: RPCCodedResponse<Spec>,
fork_name: ForkName, fork_name: ForkName,
spec: &ChainSpec,
) -> Result<BytesMut, RPCError> { ) -> Result<BytesMut, RPCError> {
let snappy_protocol_id = ProtocolId::new(protocol, Encoding::SSZSnappy); let snappy_protocol_id = ProtocolId::new(protocol, Encoding::SSZSnappy);
let fork_context = Arc::new(fork_context(fork_name)); let fork_context = Arc::new(fork_context(fork_name));
let max_packet_size = max_rpc_size(&fork_context); let max_packet_size = max_rpc_size(&fork_context, spec.max_chunk_size as usize);
let mut buf = BytesMut::new(); let mut buf = BytesMut::new();
let mut snappy_inbound_codec = let mut snappy_inbound_codec =
@ -783,10 +784,11 @@ mod tests {
protocol: SupportedProtocol, protocol: SupportedProtocol,
message: &mut BytesMut, message: &mut BytesMut,
fork_name: ForkName, fork_name: ForkName,
spec: &ChainSpec,
) -> Result<Option<RPCResponse<Spec>>, RPCError> { ) -> Result<Option<RPCResponse<Spec>>, RPCError> {
let snappy_protocol_id = ProtocolId::new(protocol, Encoding::SSZSnappy); let snappy_protocol_id = ProtocolId::new(protocol, Encoding::SSZSnappy);
let fork_context = Arc::new(fork_context(fork_name)); let fork_context = Arc::new(fork_context(fork_name));
let max_packet_size = max_rpc_size(&fork_context); let max_packet_size = max_rpc_size(&fork_context, spec.max_chunk_size as usize);
let mut snappy_outbound_codec = let mut snappy_outbound_codec =
SSZSnappyOutboundCodec::<Spec>::new(snappy_protocol_id, max_packet_size, fork_context); SSZSnappyOutboundCodec::<Spec>::new(snappy_protocol_id, max_packet_size, fork_context);
// decode message just as snappy message // decode message just as snappy message
@ -798,15 +800,20 @@ mod tests {
protocol: SupportedProtocol, protocol: SupportedProtocol,
message: RPCCodedResponse<Spec>, message: RPCCodedResponse<Spec>,
fork_name: ForkName, fork_name: ForkName,
spec: &ChainSpec,
) -> Result<Option<RPCResponse<Spec>>, RPCError> { ) -> Result<Option<RPCResponse<Spec>>, RPCError> {
let mut encoded = encode_response(protocol, message, fork_name)?; let mut encoded = encode_response(protocol, message, fork_name, spec)?;
decode_response(protocol, &mut encoded, fork_name) decode_response(protocol, &mut encoded, fork_name, spec)
} }
/// Verifies that requests we send are encoded in a way that we would correctly decode too. /// Verifies that requests we send are encoded in a way that we would correctly decode too.
fn encode_then_decode_request(req: OutboundRequest<Spec>, fork_name: ForkName) { fn encode_then_decode_request(
req: OutboundRequest<Spec>,
fork_name: ForkName,
spec: &ChainSpec,
) {
let fork_context = Arc::new(fork_context(fork_name)); let fork_context = Arc::new(fork_context(fork_name));
let max_packet_size = max_rpc_size(&fork_context); let max_packet_size = max_rpc_size(&fork_context, spec.max_chunk_size as usize);
let protocol = ProtocolId::new(req.versioned_protocol(), Encoding::SSZSnappy); let protocol = ProtocolId::new(req.versioned_protocol(), Encoding::SSZSnappy);
// Encode a request we send // Encode a request we send
let mut buf = BytesMut::new(); let mut buf = BytesMut::new();
@ -851,11 +858,14 @@ mod tests {
// Test RPCResponse encoding/decoding for V1 messages // Test RPCResponse encoding/decoding for V1 messages
#[test] #[test]
fn test_encode_then_decode_v1() { fn test_encode_then_decode_v1() {
let chain_spec = Spec::default_spec();
assert_eq!( assert_eq!(
encode_then_decode_response( encode_then_decode_response(
SupportedProtocol::StatusV1, SupportedProtocol::StatusV1,
RPCCodedResponse::Success(RPCResponse::Status(status_message())), RPCCodedResponse::Success(RPCResponse::Status(status_message())),
ForkName::Base, ForkName::Base,
&chain_spec,
), ),
Ok(Some(RPCResponse::Status(status_message()))) Ok(Some(RPCResponse::Status(status_message())))
); );
@ -865,6 +875,7 @@ mod tests {
SupportedProtocol::PingV1, SupportedProtocol::PingV1,
RPCCodedResponse::Success(RPCResponse::Pong(ping_message())), RPCCodedResponse::Success(RPCResponse::Pong(ping_message())),
ForkName::Base, ForkName::Base,
&chain_spec,
), ),
Ok(Some(RPCResponse::Pong(ping_message()))) Ok(Some(RPCResponse::Pong(ping_message())))
); );
@ -874,6 +885,7 @@ mod tests {
SupportedProtocol::BlocksByRangeV1, SupportedProtocol::BlocksByRangeV1,
RPCCodedResponse::Success(RPCResponse::BlocksByRange(Arc::new(empty_base_block()))), RPCCodedResponse::Success(RPCResponse::BlocksByRange(Arc::new(empty_base_block()))),
ForkName::Base, ForkName::Base,
&chain_spec,
), ),
Ok(Some(RPCResponse::BlocksByRange(Arc::new( Ok(Some(RPCResponse::BlocksByRange(Arc::new(
empty_base_block() empty_base_block()
@ -886,6 +898,7 @@ mod tests {
SupportedProtocol::BlocksByRangeV1, SupportedProtocol::BlocksByRangeV1,
RPCCodedResponse::Success(RPCResponse::BlocksByRange(Arc::new(altair_block()))), RPCCodedResponse::Success(RPCResponse::BlocksByRange(Arc::new(altair_block()))),
ForkName::Altair, ForkName::Altair,
&chain_spec,
) )
.unwrap_err(), .unwrap_err(),
RPCError::SSZDecodeError(_) RPCError::SSZDecodeError(_)
@ -898,6 +911,7 @@ mod tests {
SupportedProtocol::BlocksByRootV1, SupportedProtocol::BlocksByRootV1,
RPCCodedResponse::Success(RPCResponse::BlocksByRoot(Arc::new(empty_base_block()))), RPCCodedResponse::Success(RPCResponse::BlocksByRoot(Arc::new(empty_base_block()))),
ForkName::Base, ForkName::Base,
&chain_spec,
), ),
Ok(Some(RPCResponse::BlocksByRoot( Ok(Some(RPCResponse::BlocksByRoot(
Arc::new(empty_base_block()) Arc::new(empty_base_block())
@ -910,6 +924,7 @@ mod tests {
SupportedProtocol::BlocksByRootV1, SupportedProtocol::BlocksByRootV1,
RPCCodedResponse::Success(RPCResponse::BlocksByRoot(Arc::new(altair_block()))), RPCCodedResponse::Success(RPCResponse::BlocksByRoot(Arc::new(altair_block()))),
ForkName::Altair, ForkName::Altair,
&chain_spec,
) )
.unwrap_err(), .unwrap_err(),
RPCError::SSZDecodeError(_) RPCError::SSZDecodeError(_)
@ -922,6 +937,7 @@ mod tests {
SupportedProtocol::MetaDataV1, SupportedProtocol::MetaDataV1,
RPCCodedResponse::Success(RPCResponse::MetaData(metadata())), RPCCodedResponse::Success(RPCResponse::MetaData(metadata())),
ForkName::Base, ForkName::Base,
&chain_spec,
), ),
Ok(Some(RPCResponse::MetaData(metadata()))), Ok(Some(RPCResponse::MetaData(metadata()))),
); );
@ -932,6 +948,7 @@ mod tests {
SupportedProtocol::MetaDataV1, SupportedProtocol::MetaDataV1,
RPCCodedResponse::Success(RPCResponse::MetaData(metadata_v2())), RPCCodedResponse::Success(RPCResponse::MetaData(metadata_v2())),
ForkName::Base, ForkName::Base,
&chain_spec,
), ),
Ok(Some(RPCResponse::MetaData(metadata()))), Ok(Some(RPCResponse::MetaData(metadata()))),
); );
@ -940,11 +957,14 @@ mod tests {
// Test RPCResponse encoding/decoding for V1 messages // Test RPCResponse encoding/decoding for V1 messages
#[test] #[test]
fn test_encode_then_decode_v2() { fn test_encode_then_decode_v2() {
let chain_spec = Spec::default_spec();
assert_eq!( assert_eq!(
encode_then_decode_response( encode_then_decode_response(
SupportedProtocol::BlocksByRangeV2, SupportedProtocol::BlocksByRangeV2,
RPCCodedResponse::Success(RPCResponse::BlocksByRange(Arc::new(empty_base_block()))), RPCCodedResponse::Success(RPCResponse::BlocksByRange(Arc::new(empty_base_block()))),
ForkName::Base, ForkName::Base,
&chain_spec,
), ),
Ok(Some(RPCResponse::BlocksByRange(Arc::new( Ok(Some(RPCResponse::BlocksByRange(Arc::new(
empty_base_block() empty_base_block()
@ -959,6 +979,7 @@ mod tests {
SupportedProtocol::BlocksByRangeV2, SupportedProtocol::BlocksByRangeV2,
RPCCodedResponse::Success(RPCResponse::BlocksByRange(Arc::new(empty_base_block()))), RPCCodedResponse::Success(RPCResponse::BlocksByRange(Arc::new(empty_base_block()))),
ForkName::Altair, ForkName::Altair,
&chain_spec,
), ),
Ok(Some(RPCResponse::BlocksByRange(Arc::new( Ok(Some(RPCResponse::BlocksByRange(Arc::new(
empty_base_block() empty_base_block()
@ -970,12 +991,13 @@ mod tests {
SupportedProtocol::BlocksByRangeV2, SupportedProtocol::BlocksByRangeV2,
RPCCodedResponse::Success(RPCResponse::BlocksByRange(Arc::new(altair_block()))), RPCCodedResponse::Success(RPCResponse::BlocksByRange(Arc::new(altair_block()))),
ForkName::Altair, ForkName::Altair,
&chain_spec,
), ),
Ok(Some(RPCResponse::BlocksByRange(Arc::new(altair_block())))) Ok(Some(RPCResponse::BlocksByRange(Arc::new(altair_block()))))
); );
let merge_block_small = merge_block_small(&fork_context(ForkName::Merge)); let merge_block_small = merge_block_small(&fork_context(ForkName::Merge), &chain_spec);
let merge_block_large = merge_block_large(&fork_context(ForkName::Merge)); let merge_block_large = merge_block_large(&fork_context(ForkName::Merge), &chain_spec);
assert_eq!( assert_eq!(
encode_then_decode_response( encode_then_decode_response(
@ -984,6 +1006,7 @@ mod tests {
merge_block_small.clone() merge_block_small.clone()
))), ))),
ForkName::Merge, ForkName::Merge,
&chain_spec,
), ),
Ok(Some(RPCResponse::BlocksByRange(Arc::new( Ok(Some(RPCResponse::BlocksByRange(Arc::new(
merge_block_small.clone() merge_block_small.clone()
@ -1000,6 +1023,7 @@ mod tests {
SupportedProtocol::BlocksByRangeV2, SupportedProtocol::BlocksByRangeV2,
&mut encoded, &mut encoded,
ForkName::Merge, ForkName::Merge,
&chain_spec,
) )
.unwrap_err(), .unwrap_err(),
RPCError::InvalidData(_) RPCError::InvalidData(_)
@ -1012,6 +1036,7 @@ mod tests {
SupportedProtocol::BlocksByRootV2, SupportedProtocol::BlocksByRootV2,
RPCCodedResponse::Success(RPCResponse::BlocksByRoot(Arc::new(empty_base_block()))), RPCCodedResponse::Success(RPCResponse::BlocksByRoot(Arc::new(empty_base_block()))),
ForkName::Base, ForkName::Base,
&chain_spec,
), ),
Ok(Some(RPCResponse::BlocksByRoot( Ok(Some(RPCResponse::BlocksByRoot(
Arc::new(empty_base_block()) Arc::new(empty_base_block())
@ -1026,6 +1051,7 @@ mod tests {
SupportedProtocol::BlocksByRootV2, SupportedProtocol::BlocksByRootV2,
RPCCodedResponse::Success(RPCResponse::BlocksByRoot(Arc::new(empty_base_block()))), RPCCodedResponse::Success(RPCResponse::BlocksByRoot(Arc::new(empty_base_block()))),
ForkName::Altair, ForkName::Altair,
&chain_spec,
), ),
Ok(Some(RPCResponse::BlocksByRoot( Ok(Some(RPCResponse::BlocksByRoot(
Arc::new(empty_base_block()) Arc::new(empty_base_block())
@ -1037,6 +1063,7 @@ mod tests {
SupportedProtocol::BlocksByRootV2, SupportedProtocol::BlocksByRootV2,
RPCCodedResponse::Success(RPCResponse::BlocksByRoot(Arc::new(altair_block()))), RPCCodedResponse::Success(RPCResponse::BlocksByRoot(Arc::new(altair_block()))),
ForkName::Altair, ForkName::Altair,
&chain_spec,
), ),
Ok(Some(RPCResponse::BlocksByRoot(Arc::new(altair_block())))) Ok(Some(RPCResponse::BlocksByRoot(Arc::new(altair_block()))))
); );
@ -1048,6 +1075,7 @@ mod tests {
merge_block_small.clone() merge_block_small.clone()
))), ))),
ForkName::Merge, ForkName::Merge,
&chain_spec,
), ),
Ok(Some(RPCResponse::BlocksByRoot(Arc::new(merge_block_small)))) Ok(Some(RPCResponse::BlocksByRoot(Arc::new(merge_block_small))))
); );
@ -1062,6 +1090,7 @@ mod tests {
SupportedProtocol::BlocksByRootV2, SupportedProtocol::BlocksByRootV2,
&mut encoded, &mut encoded,
ForkName::Merge, ForkName::Merge,
&chain_spec,
) )
.unwrap_err(), .unwrap_err(),
RPCError::InvalidData(_) RPCError::InvalidData(_)
@ -1075,6 +1104,7 @@ mod tests {
SupportedProtocol::MetaDataV2, SupportedProtocol::MetaDataV2,
RPCCodedResponse::Success(RPCResponse::MetaData(metadata())), RPCCodedResponse::Success(RPCResponse::MetaData(metadata())),
ForkName::Base, ForkName::Base,
&chain_spec,
), ),
Ok(Some(RPCResponse::MetaData(metadata_v2()))) Ok(Some(RPCResponse::MetaData(metadata_v2())))
); );
@ -1084,6 +1114,7 @@ mod tests {
SupportedProtocol::MetaDataV2, SupportedProtocol::MetaDataV2,
RPCCodedResponse::Success(RPCResponse::MetaData(metadata_v2())), RPCCodedResponse::Success(RPCResponse::MetaData(metadata_v2())),
ForkName::Altair, ForkName::Altair,
&chain_spec,
), ),
Ok(Some(RPCResponse::MetaData(metadata_v2()))) Ok(Some(RPCResponse::MetaData(metadata_v2())))
); );
@ -1094,11 +1125,14 @@ mod tests {
fn test_context_bytes_v2() { fn test_context_bytes_v2() {
let fork_context = fork_context(ForkName::Altair); let fork_context = fork_context(ForkName::Altair);
let chain_spec = Spec::default_spec();
// Removing context bytes for v2 messages should error // Removing context bytes for v2 messages should error
let mut encoded_bytes = encode_response( let mut encoded_bytes = encode_response(
SupportedProtocol::BlocksByRangeV2, SupportedProtocol::BlocksByRangeV2,
RPCCodedResponse::Success(RPCResponse::BlocksByRange(Arc::new(empty_base_block()))), RPCCodedResponse::Success(RPCResponse::BlocksByRange(Arc::new(empty_base_block()))),
ForkName::Base, ForkName::Base,
&chain_spec,
) )
.unwrap(); .unwrap();
@ -1108,7 +1142,8 @@ mod tests {
decode_response( decode_response(
SupportedProtocol::BlocksByRangeV2, SupportedProtocol::BlocksByRangeV2,
&mut encoded_bytes, &mut encoded_bytes,
ForkName::Base ForkName::Base,
&chain_spec,
) )
.unwrap_err(), .unwrap_err(),
RPCError::ErrorResponse(RPCResponseErrorCode::InvalidRequest, _), RPCError::ErrorResponse(RPCResponseErrorCode::InvalidRequest, _),
@ -1118,6 +1153,7 @@ mod tests {
SupportedProtocol::BlocksByRootV2, SupportedProtocol::BlocksByRootV2,
RPCCodedResponse::Success(RPCResponse::BlocksByRoot(Arc::new(empty_base_block()))), RPCCodedResponse::Success(RPCResponse::BlocksByRoot(Arc::new(empty_base_block()))),
ForkName::Base, ForkName::Base,
&chain_spec,
) )
.unwrap(); .unwrap();
@ -1127,7 +1163,8 @@ mod tests {
decode_response( decode_response(
SupportedProtocol::BlocksByRangeV2, SupportedProtocol::BlocksByRangeV2,
&mut encoded_bytes, &mut encoded_bytes,
ForkName::Base ForkName::Base,
&chain_spec,
) )
.unwrap_err(), .unwrap_err(),
RPCError::ErrorResponse(RPCResponseErrorCode::InvalidRequest, _), RPCError::ErrorResponse(RPCResponseErrorCode::InvalidRequest, _),
@ -1138,6 +1175,7 @@ mod tests {
SupportedProtocol::BlocksByRangeV2, SupportedProtocol::BlocksByRangeV2,
RPCCodedResponse::Success(RPCResponse::BlocksByRange(Arc::new(empty_base_block()))), RPCCodedResponse::Success(RPCResponse::BlocksByRange(Arc::new(empty_base_block()))),
ForkName::Altair, ForkName::Altair,
&chain_spec,
) )
.unwrap(); .unwrap();
@ -1150,7 +1188,8 @@ mod tests {
decode_response( decode_response(
SupportedProtocol::BlocksByRangeV2, SupportedProtocol::BlocksByRangeV2,
&mut wrong_fork_bytes, &mut wrong_fork_bytes,
ForkName::Altair ForkName::Altair,
&chain_spec,
) )
.unwrap_err(), .unwrap_err(),
RPCError::SSZDecodeError(_), RPCError::SSZDecodeError(_),
@ -1161,6 +1200,7 @@ mod tests {
SupportedProtocol::BlocksByRootV2, SupportedProtocol::BlocksByRootV2,
RPCCodedResponse::Success(RPCResponse::BlocksByRoot(Arc::new(altair_block()))), RPCCodedResponse::Success(RPCResponse::BlocksByRoot(Arc::new(altair_block()))),
ForkName::Altair, ForkName::Altair,
&chain_spec,
) )
.unwrap(); .unwrap();
@ -1172,7 +1212,8 @@ mod tests {
decode_response( decode_response(
SupportedProtocol::BlocksByRangeV2, SupportedProtocol::BlocksByRangeV2,
&mut wrong_fork_bytes, &mut wrong_fork_bytes,
ForkName::Altair ForkName::Altair,
&chain_spec,
) )
.unwrap_err(), .unwrap_err(),
RPCError::SSZDecodeError(_), RPCError::SSZDecodeError(_),
@ -1186,6 +1227,7 @@ mod tests {
SupportedProtocol::MetaDataV2, SupportedProtocol::MetaDataV2,
RPCCodedResponse::Success(RPCResponse::MetaData(metadata())), RPCCodedResponse::Success(RPCResponse::MetaData(metadata())),
ForkName::Altair, ForkName::Altair,
&chain_spec,
) )
.unwrap(), .unwrap(),
); );
@ -1193,7 +1235,8 @@ mod tests {
assert!(decode_response( assert!(decode_response(
SupportedProtocol::MetaDataV2, SupportedProtocol::MetaDataV2,
&mut encoded_bytes, &mut encoded_bytes,
ForkName::Altair ForkName::Altair,
&chain_spec,
) )
.is_err()); .is_err());
@ -1202,6 +1245,7 @@ mod tests {
SupportedProtocol::BlocksByRootV2, SupportedProtocol::BlocksByRootV2,
RPCCodedResponse::Success(RPCResponse::BlocksByRoot(Arc::new(empty_base_block()))), RPCCodedResponse::Success(RPCResponse::BlocksByRoot(Arc::new(empty_base_block()))),
ForkName::Altair, ForkName::Altair,
&chain_spec,
) )
.unwrap(); .unwrap();
@ -1213,7 +1257,8 @@ mod tests {
decode_response( decode_response(
SupportedProtocol::BlocksByRangeV2, SupportedProtocol::BlocksByRangeV2,
&mut wrong_fork_bytes, &mut wrong_fork_bytes,
ForkName::Altair ForkName::Altair,
&chain_spec,
) )
.unwrap_err(), .unwrap_err(),
RPCError::ErrorResponse(RPCResponseErrorCode::InvalidRequest, _), RPCError::ErrorResponse(RPCResponseErrorCode::InvalidRequest, _),
@ -1224,6 +1269,7 @@ mod tests {
SupportedProtocol::BlocksByRootV2, SupportedProtocol::BlocksByRootV2,
RPCCodedResponse::Success(RPCResponse::BlocksByRoot(Arc::new(empty_base_block()))), RPCCodedResponse::Success(RPCResponse::BlocksByRoot(Arc::new(empty_base_block()))),
ForkName::Altair, ForkName::Altair,
&chain_spec,
) )
.unwrap(); .unwrap();
@ -1233,7 +1279,8 @@ mod tests {
decode_response( decode_response(
SupportedProtocol::BlocksByRangeV2, SupportedProtocol::BlocksByRangeV2,
&mut part, &mut part,
ForkName::Altair ForkName::Altair,
&chain_spec,
), ),
Ok(None) Ok(None)
) )
@ -1252,9 +1299,12 @@ mod tests {
OutboundRequest::MetaData(MetadataRequest::new_v1()), OutboundRequest::MetaData(MetadataRequest::new_v1()),
OutboundRequest::MetaData(MetadataRequest::new_v2()), OutboundRequest::MetaData(MetadataRequest::new_v2()),
]; ];
let chain_spec = Spec::default_spec();
for req in requests.iter() { for req in requests.iter() {
for fork_name in ForkName::list_all() { for fork_name in ForkName::list_all() {
encode_then_decode_request(req.clone(), fork_name); encode_then_decode_request(req.clone(), fork_name, &chain_spec);
} }
} }
} }
@ -1308,9 +1358,16 @@ mod tests {
assert_eq!(writer.get_ref().len(), 42); assert_eq!(writer.get_ref().len(), 42);
dst.extend_from_slice(writer.get_ref()); dst.extend_from_slice(writer.get_ref());
let chain_spec = Spec::default_spec();
// 10 (for stream identifier) + 80 + 42 = 132 > `max_compressed_len`. Hence, decoding should fail with `InvalidData`. // 10 (for stream identifier) + 80 + 42 = 132 > `max_compressed_len`. Hence, decoding should fail with `InvalidData`.
assert!(matches!( assert!(matches!(
decode_response(SupportedProtocol::StatusV1, &mut dst, ForkName::Base).unwrap_err(), decode_response(
SupportedProtocol::StatusV1,
&mut dst,
ForkName::Base,
&chain_spec
)
.unwrap_err(),
RPCError::InvalidData(_) RPCError::InvalidData(_)
)); ));
} }
@ -1365,12 +1422,15 @@ mod tests {
assert_eq!(writer.get_ref().len(), 8103); assert_eq!(writer.get_ref().len(), 8103);
dst.extend_from_slice(writer.get_ref()); dst.extend_from_slice(writer.get_ref());
let chain_spec = Spec::default_spec();
// 10 (for stream identifier) + 176156 + 8103 = 184269 > `max_compressed_len`. Hence, decoding should fail with `InvalidData`. // 10 (for stream identifier) + 176156 + 8103 = 184269 > `max_compressed_len`. Hence, decoding should fail with `InvalidData`.
assert!(matches!( assert!(matches!(
decode_response( decode_response(
SupportedProtocol::BlocksByRangeV2, SupportedProtocol::BlocksByRangeV2,
&mut dst, &mut dst,
ForkName::Altair ForkName::Altair,
&chain_spec,
) )
.unwrap_err(), .unwrap_err(),
RPCError::InvalidData(_) RPCError::InvalidData(_)
@ -1398,8 +1458,12 @@ mod tests {
let mut uvi_codec: Uvi<usize> = Uvi::default(); let mut uvi_codec: Uvi<usize> = Uvi::default();
let mut dst = BytesMut::with_capacity(1024); let mut dst = BytesMut::with_capacity(1024);
let chain_spec = Spec::default_spec();
// Insert length-prefix // Insert length-prefix
uvi_codec.encode(MAX_RPC_SIZE + 1, &mut dst).unwrap(); uvi_codec
.encode(chain_spec.max_chunk_size as usize + 1, &mut dst)
.unwrap();
// Insert snappy stream identifier // Insert snappy stream identifier
dst.extend_from_slice(stream_identifier); dst.extend_from_slice(stream_identifier);
@ -1411,7 +1475,13 @@ mod tests {
dst.extend_from_slice(writer.get_ref()); dst.extend_from_slice(writer.get_ref());
assert!(matches!( assert!(matches!(
decode_response(SupportedProtocol::StatusV1, &mut dst, ForkName::Base).unwrap_err(), decode_response(
SupportedProtocol::StatusV1,
&mut dst,
ForkName::Base,
&chain_spec
)
.unwrap_err(),
RPCError::InvalidData(_) RPCError::InvalidData(_)
)); ));
} }

View File

@ -3,9 +3,7 @@
use super::methods::{GoodbyeReason, RPCCodedResponse, RPCResponseErrorCode, ResponseTermination}; use super::methods::{GoodbyeReason, RPCCodedResponse, RPCResponseErrorCode, ResponseTermination};
use super::outbound::OutboundRequestContainer; use super::outbound::OutboundRequestContainer;
use super::protocol::{ use super::protocol::{InboundOutput, InboundRequest, Protocol, RPCError, RPCProtocol};
max_rpc_size, InboundOutput, InboundRequest, Protocol, RPCError, RPCProtocol,
};
use super::{RPCReceived, RPCSend, ReqId}; use super::{RPCReceived, RPCSend, ReqId};
use crate::rpc::outbound::{OutboundFramed, OutboundRequest}; use crate::rpc::outbound::{OutboundFramed, OutboundRequest};
use crate::rpc::protocol::InboundFramed; use crate::rpc::protocol::InboundFramed;
@ -31,9 +29,6 @@ use tokio::time::{sleep_until, Instant as TInstant, Sleep};
use tokio_util::time::{delay_queue, DelayQueue}; use tokio_util::time::{delay_queue, DelayQueue};
use types::{EthSpec, ForkContext}; use types::{EthSpec, ForkContext};
/// The time (in seconds) before a substream that is awaiting a response from the user times out.
pub const RESPONSE_TIMEOUT: u64 = 10;
/// The number of times to retry an outbound upgrade in the case of IO errors. /// The number of times to retry an outbound upgrade in the case of IO errors.
const IO_ERROR_RETRIES: u8 = 3; const IO_ERROR_RETRIES: u8 = 3;
@ -131,6 +126,9 @@ where
/// Logger for handling RPC streams /// Logger for handling RPC streams
log: slog::Logger, log: slog::Logger,
/// Timeout that will me used for inbound and outbound responses.
resp_timeout: Duration,
} }
enum HandlerState { enum HandlerState {
@ -212,7 +210,8 @@ where
pub fn new( pub fn new(
listen_protocol: SubstreamProtocol<RPCProtocol<TSpec>, ()>, listen_protocol: SubstreamProtocol<RPCProtocol<TSpec>, ()>,
fork_context: Arc<ForkContext>, fork_context: Arc<ForkContext>,
log: slog::Logger, log: &slog::Logger,
resp_timeout: Duration,
) -> Self { ) -> Self {
RPCHandler { RPCHandler {
listen_protocol, listen_protocol,
@ -230,7 +229,8 @@ where
outbound_io_error_retries: 0, outbound_io_error_retries: 0,
fork_context, fork_context,
waker: None, waker: None,
log, log: log.clone(),
resp_timeout,
} }
} }
@ -554,7 +554,7 @@ where
// Each chunk is allowed RESPONSE_TIMEOUT to be sent. // Each chunk is allowed RESPONSE_TIMEOUT to be sent.
if let Some(ref delay_key) = info.delay_key { if let Some(ref delay_key) = info.delay_key {
self.inbound_substreams_delay self.inbound_substreams_delay
.reset(delay_key, Duration::from_secs(RESPONSE_TIMEOUT)); .reset(delay_key, self.resp_timeout);
} }
// The stream may be currently idle. Attempt to process more // The stream may be currently idle. Attempt to process more
@ -688,7 +688,7 @@ where
}; };
substream_entry.remaining_chunks = Some(remaining_chunks); substream_entry.remaining_chunks = Some(remaining_chunks);
self.outbound_substreams_delay self.outbound_substreams_delay
.reset(delay_key, Duration::from_secs(RESPONSE_TIMEOUT)); .reset(delay_key, self.resp_timeout);
} }
} else { } else {
// either this is a single response request or this response closes the // either this is a single response request or this response closes the
@ -811,7 +811,7 @@ where
OutboundRequestContainer { OutboundRequestContainer {
req: req.clone(), req: req.clone(),
fork_context: self.fork_context.clone(), fork_context: self.fork_context.clone(),
max_rpc_size: max_rpc_size(&self.fork_context), max_rpc_size: self.listen_protocol().upgrade().max_rpc_size,
}, },
(), (),
) )
@ -896,10 +896,9 @@ where
if expected_responses > 0 { if expected_responses > 0 {
if self.inbound_substreams.len() < MAX_INBOUND_SUBSTREAMS { if self.inbound_substreams.len() < MAX_INBOUND_SUBSTREAMS {
// Store the stream and tag the output. // Store the stream and tag the output.
let delay_key = self.inbound_substreams_delay.insert( let delay_key = self
self.current_inbound_substream_id, .inbound_substreams_delay
Duration::from_secs(RESPONSE_TIMEOUT), .insert(self.current_inbound_substream_id, self.resp_timeout);
);
let awaiting_stream = InboundState::Idle(substream); let awaiting_stream = InboundState::Idle(substream);
self.inbound_substreams.insert( self.inbound_substreams.insert(
self.current_inbound_substream_id, self.current_inbound_substream_id,
@ -961,10 +960,9 @@ where
let expected_responses = request.expected_responses(); let expected_responses = request.expected_responses();
if expected_responses > 0 { if expected_responses > 0 {
// new outbound request. Store the stream and tag the output. // new outbound request. Store the stream and tag the output.
let delay_key = self.outbound_substreams_delay.insert( let delay_key = self
self.current_outbound_substream_id, .outbound_substreams_delay
Duration::from_secs(RESPONSE_TIMEOUT), .insert(self.current_outbound_substream_id, self.resp_timeout);
);
let awaiting_stream = OutboundSubstreamState::RequestPendingResponse { let awaiting_stream = OutboundSubstreamState::RequestPendingResponse {
substream: Box::new(substream), substream: Box::new(substream),
request, request,

View File

@ -17,6 +17,7 @@ use slog::{crit, debug, o};
use std::marker::PhantomData; use std::marker::PhantomData;
use std::sync::Arc; use std::sync::Arc;
use std::task::{Context, Poll}; use std::task::{Context, Poll};
use std::time::Duration;
use types::{EthSpec, ForkContext}; use types::{EthSpec, ForkContext};
pub(crate) use handler::HandlerErr; pub(crate) use handler::HandlerErr;
@ -107,6 +108,12 @@ pub struct RPCMessage<Id, TSpec: EthSpec> {
type BehaviourAction<Id, TSpec> = ToSwarm<RPCMessage<Id, TSpec>, RPCSend<Id, TSpec>>; type BehaviourAction<Id, TSpec> = ToSwarm<RPCMessage<Id, TSpec>, RPCSend<Id, TSpec>>;
pub struct NetworkParams {
pub max_chunk_size: usize,
pub ttfb_timeout: Duration,
pub resp_timeout: Duration,
}
/// Implements the libp2p `NetworkBehaviour` trait and therefore manages network-level /// Implements the libp2p `NetworkBehaviour` trait and therefore manages network-level
/// logic. /// logic.
pub struct RPC<Id: ReqId, TSpec: EthSpec> { pub struct RPC<Id: ReqId, TSpec: EthSpec> {
@ -120,6 +127,8 @@ pub struct RPC<Id: ReqId, TSpec: EthSpec> {
enable_light_client_server: bool, enable_light_client_server: bool,
/// Slog logger for RPC behaviour. /// Slog logger for RPC behaviour.
log: slog::Logger, log: slog::Logger,
/// Networking constant values
network_params: NetworkParams,
} }
impl<Id: ReqId, TSpec: EthSpec> RPC<Id, TSpec> { impl<Id: ReqId, TSpec: EthSpec> RPC<Id, TSpec> {
@ -129,6 +138,7 @@ impl<Id: ReqId, TSpec: EthSpec> RPC<Id, TSpec> {
inbound_rate_limiter_config: Option<InboundRateLimiterConfig>, inbound_rate_limiter_config: Option<InboundRateLimiterConfig>,
outbound_rate_limiter_config: Option<OutboundRateLimiterConfig>, outbound_rate_limiter_config: Option<OutboundRateLimiterConfig>,
log: slog::Logger, log: slog::Logger,
network_params: NetworkParams,
) -> Self { ) -> Self {
let log = log.new(o!("service" => "libp2p_rpc")); let log = log.new(o!("service" => "libp2p_rpc"));
@ -149,6 +159,7 @@ impl<Id: ReqId, TSpec: EthSpec> RPC<Id, TSpec> {
fork_context, fork_context,
enable_light_client_server, enable_light_client_server,
log, log,
network_params,
} }
} }
@ -220,16 +231,22 @@ where
let protocol = SubstreamProtocol::new( let protocol = SubstreamProtocol::new(
RPCProtocol { RPCProtocol {
fork_context: self.fork_context.clone(), fork_context: self.fork_context.clone(),
max_rpc_size: max_rpc_size(&self.fork_context), max_rpc_size: max_rpc_size(&self.fork_context, self.network_params.max_chunk_size),
enable_light_client_server: self.enable_light_client_server, enable_light_client_server: self.enable_light_client_server,
phantom: PhantomData, phantom: PhantomData,
ttfb_timeout: self.network_params.ttfb_timeout,
}, },
(), (),
); );
// NOTE: this is needed because PeerIds have interior mutability. // NOTE: this is needed because PeerIds have interior mutability.
let peer_repr = peer_id.to_string(); let peer_repr = peer_id.to_string();
let log = self.log.new(slog::o!("peer_id" => peer_repr)); let log = self.log.new(slog::o!("peer_id" => peer_repr));
let handler = RPCHandler::new(protocol, self.fork_context.clone(), log); let handler = RPCHandler::new(
protocol,
self.fork_context.clone(),
&log,
self.network_params.resp_timeout,
);
Ok(handler) Ok(handler)
} }
@ -244,9 +261,10 @@ where
let protocol = SubstreamProtocol::new( let protocol = SubstreamProtocol::new(
RPCProtocol { RPCProtocol {
fork_context: self.fork_context.clone(), fork_context: self.fork_context.clone(),
max_rpc_size: max_rpc_size(&self.fork_context), max_rpc_size: max_rpc_size(&self.fork_context, self.network_params.max_chunk_size),
enable_light_client_server: self.enable_light_client_server, enable_light_client_server: self.enable_light_client_server,
phantom: PhantomData, phantom: PhantomData,
ttfb_timeout: self.network_params.ttfb_timeout,
}, },
(), (),
); );
@ -254,7 +272,12 @@ where
// NOTE: this is needed because PeerIds have interior mutability. // NOTE: this is needed because PeerIds have interior mutability.
let peer_repr = peer_id.to_string(); let peer_repr = peer_id.to_string();
let log = self.log.new(slog::o!("peer_id" => peer_repr)); let log = self.log.new(slog::o!("peer_id" => peer_repr));
let handler = RPCHandler::new(protocol, self.fork_context.clone(), log); let handler = RPCHandler::new(
protocol,
self.fork_context.clone(),
&log,
self.network_params.resp_timeout,
);
Ok(handler) Ok(handler)
} }

View File

@ -72,7 +72,7 @@ lazy_static! {
/// The `BeaconBlockMerge` block has an `ExecutionPayload` field which has a max size ~16 GiB for future proofing. /// The `BeaconBlockMerge` block has an `ExecutionPayload` field which has a max size ~16 GiB for future proofing.
/// We calculate the value from its fields instead of constructing the block and checking the length. /// We calculate the value from its fields instead of constructing the block and checking the length.
/// Note: This is only the theoretical upper bound. We further bound the max size we receive over the network /// Note: This is only the theoretical upper bound. We further bound the max size we receive over the network
/// with `MAX_RPC_SIZE_POST_MERGE`. /// with `max_chunk_size`.
pub static ref SIGNED_BEACON_BLOCK_MERGE_MAX: usize = pub static ref SIGNED_BEACON_BLOCK_MERGE_MAX: usize =
// Size of a full altair block // Size of a full altair block
*SIGNED_BEACON_BLOCK_ALTAIR_MAX *SIGNED_BEACON_BLOCK_ALTAIR_MAX
@ -109,25 +109,18 @@ lazy_static! {
.len(); .len();
} }
/// The maximum bytes that can be sent across the RPC pre-merge.
pub(crate) const MAX_RPC_SIZE: usize = 1_048_576; // 1M
/// The maximum bytes that can be sent across the RPC post-merge.
pub(crate) const MAX_RPC_SIZE_POST_MERGE: usize = 10 * 1_048_576; // 10M
pub(crate) const MAX_RPC_SIZE_POST_CAPELLA: usize = 10 * 1_048_576; // 10M
/// The protocol prefix the RPC protocol id. /// The protocol prefix the RPC protocol id.
const PROTOCOL_PREFIX: &str = "/eth2/beacon_chain/req"; const PROTOCOL_PREFIX: &str = "/eth2/beacon_chain/req";
/// Time allowed for the first byte of a request to arrive before we time out (Time To First Byte).
const TTFB_TIMEOUT: u64 = 5;
/// The number of seconds to wait for the first bytes of a request once a protocol has been /// The number of seconds to wait for the first bytes of a request once a protocol has been
/// established before the stream is terminated. /// established before the stream is terminated.
const REQUEST_TIMEOUT: u64 = 15; const REQUEST_TIMEOUT: u64 = 15;
/// Returns the maximum bytes that can be sent across the RPC. /// Returns the maximum bytes that can be sent across the RPC.
pub fn max_rpc_size(fork_context: &ForkContext) -> usize { pub fn max_rpc_size(fork_context: &ForkContext, max_chunk_size: usize) -> usize {
match fork_context.current_fork() { match fork_context.current_fork() {
ForkName::Altair | ForkName::Base => MAX_RPC_SIZE, ForkName::Altair | ForkName::Base => max_chunk_size / 10,
ForkName::Merge => MAX_RPC_SIZE_POST_MERGE, ForkName::Merge => max_chunk_size,
ForkName::Capella => MAX_RPC_SIZE_POST_CAPELLA, ForkName::Capella => max_chunk_size,
} }
} }
@ -262,6 +255,7 @@ pub struct RPCProtocol<TSpec: EthSpec> {
pub max_rpc_size: usize, pub max_rpc_size: usize,
pub enable_light_client_server: bool, pub enable_light_client_server: bool,
pub phantom: PhantomData<TSpec>, pub phantom: PhantomData<TSpec>,
pub ttfb_timeout: Duration,
} }
impl<TSpec: EthSpec> UpgradeInfo for RPCProtocol<TSpec> { impl<TSpec: EthSpec> UpgradeInfo for RPCProtocol<TSpec> {
@ -447,7 +441,7 @@ where
} }
}; };
let mut timed_socket = TimeoutStream::new(socket); let mut timed_socket = TimeoutStream::new(socket);
timed_socket.set_read_timeout(Some(Duration::from_secs(TTFB_TIMEOUT))); timed_socket.set_read_timeout(Some(self.ttfb_timeout));
let socket = Framed::new(Box::pin(timed_socket), codec); let socket = Framed::new(Box::pin(timed_socket), codec);

View File

@ -1,6 +1,6 @@
use self::behaviour::Behaviour; use self::behaviour::Behaviour;
use self::gossip_cache::GossipCache; use self::gossip_cache::GossipCache;
use crate::config::{gossipsub_config, NetworkLoad}; use crate::config::{gossipsub_config, GossipsubConfigParams, NetworkLoad};
use crate::discovery::{ use crate::discovery::{
subnet_predicate, DiscoveredPeers, Discovery, FIND_NODE_QUERY_CLOSEST_PEERS, subnet_predicate, DiscoveredPeers, Discovery, FIND_NODE_QUERY_CLOSEST_PEERS,
}; };
@ -232,7 +232,15 @@ impl<AppReqId: ReqId, TSpec: EthSpec> Network<AppReqId, TSpec> {
max_subscriptions_per_request: 150, // 148 in theory = (64 attestation + 4 sync committee + 6 core topics) * 2 max_subscriptions_per_request: 150, // 148 in theory = (64 attestation + 4 sync committee + 6 core topics) * 2
}; };
config.gs_config = gossipsub_config(config.network_load, ctx.fork_context.clone()); let gossipsub_config_params = GossipsubConfigParams {
message_domain_valid_snappy: ctx.chain_spec.message_domain_valid_snappy,
gossip_max_size: ctx.chain_spec.gossip_max_size as usize,
};
config.gs_config = gossipsub_config(
config.network_load,
ctx.fork_context.clone(),
gossipsub_config_params,
);
// If metrics are enabled for gossipsub build the configuration // If metrics are enabled for gossipsub build the configuration
let gossipsub_metrics = ctx let gossipsub_metrics = ctx
@ -256,12 +264,18 @@ impl<AppReqId: ReqId, TSpec: EthSpec> Network<AppReqId, TSpec> {
(gossipsub, update_gossipsub_scores) (gossipsub, update_gossipsub_scores)
}; };
let network_params = NetworkParams {
max_chunk_size: ctx.chain_spec.max_chunk_size as usize,
ttfb_timeout: ctx.chain_spec.ttfb_timeout(),
resp_timeout: ctx.chain_spec.resp_timeout(),
};
let eth2_rpc = RPC::new( let eth2_rpc = RPC::new(
ctx.fork_context.clone(), ctx.fork_context.clone(),
config.enable_light_client_server, config.enable_light_client_server,
config.inbound_rate_limiter_config.clone(), config.inbound_rate_limiter_config.clone(),
config.outbound_rate_limiter_config.clone(), config.outbound_rate_limiter_config.clone(),
log.clone(), log.clone(),
network_params,
); );
let discovery = { let discovery = {

View File

@ -94,6 +94,7 @@ pub async fn build_libp2p_instance(
boot_nodes: Vec<Enr>, boot_nodes: Vec<Enr>,
log: slog::Logger, log: slog::Logger,
fork_name: ForkName, fork_name: ForkName,
spec: &ChainSpec,
) -> Libp2pInstance { ) -> Libp2pInstance {
let port = unused_tcp4_port().unwrap(); let port = unused_tcp4_port().unwrap();
let config = build_config(port, boot_nodes); let config = build_config(port, boot_nodes);
@ -106,7 +107,7 @@ pub async fn build_libp2p_instance(
config: &config, config: &config,
enr_fork_id: EnrForkId::default(), enr_fork_id: EnrForkId::default(),
fork_context: Arc::new(fork_context(fork_name)), fork_context: Arc::new(fork_context(fork_name)),
chain_spec: &ChainSpec::minimal(), chain_spec: spec,
gossipsub_registry: None, gossipsub_registry: None,
}; };
Libp2pInstance( Libp2pInstance(
@ -130,12 +131,13 @@ pub async fn build_node_pair(
rt: Weak<Runtime>, rt: Weak<Runtime>,
log: &slog::Logger, log: &slog::Logger,
fork_name: ForkName, fork_name: ForkName,
spec: &ChainSpec,
) -> (Libp2pInstance, Libp2pInstance) { ) -> (Libp2pInstance, Libp2pInstance) {
let sender_log = log.new(o!("who" => "sender")); let sender_log = log.new(o!("who" => "sender"));
let receiver_log = log.new(o!("who" => "receiver")); let receiver_log = log.new(o!("who" => "receiver"));
let mut sender = build_libp2p_instance(rt.clone(), vec![], sender_log, fork_name).await; let mut sender = build_libp2p_instance(rt.clone(), vec![], sender_log, fork_name, spec).await;
let mut receiver = build_libp2p_instance(rt, vec![], receiver_log, fork_name).await; let mut receiver = build_libp2p_instance(rt, vec![], receiver_log, fork_name, spec).await;
let receiver_multiaddr = receiver.local_enr().multiaddr()[1].clone(); let receiver_multiaddr = receiver.local_enr().multiaddr()[1].clone();
@ -180,10 +182,11 @@ pub async fn build_linear(
log: slog::Logger, log: slog::Logger,
n: usize, n: usize,
fork_name: ForkName, fork_name: ForkName,
spec: &ChainSpec,
) -> Vec<Libp2pInstance> { ) -> Vec<Libp2pInstance> {
let mut nodes = Vec::with_capacity(n); let mut nodes = Vec::with_capacity(n);
for _ in 0..n { for _ in 0..n {
nodes.push(build_libp2p_instance(rt.clone(), vec![], log.clone(), fork_name).await); nodes.push(build_libp2p_instance(rt.clone(), vec![], log.clone(), fork_name, spec).await);
} }
let multiaddrs: Vec<Multiaddr> = nodes let multiaddrs: Vec<Multiaddr> = nodes

View File

@ -9,8 +9,9 @@ use std::time::Duration;
use tokio::runtime::Runtime; use tokio::runtime::Runtime;
use tokio::time::sleep; use tokio::time::sleep;
use types::{ use types::{
BeaconBlock, BeaconBlockAltair, BeaconBlockBase, BeaconBlockMerge, EmptyBlock, Epoch, EthSpec, BeaconBlock, BeaconBlockAltair, BeaconBlockBase, BeaconBlockMerge, ChainSpec, EmptyBlock,
ForkContext, ForkName, Hash256, MinimalEthSpec, Signature, SignedBeaconBlock, Slot, Epoch, EthSpec, ForkContext, ForkName, Hash256, MinimalEthSpec, Signature, SignedBeaconBlock,
Slot,
}; };
mod common; mod common;
@ -18,30 +19,30 @@ mod common;
type E = MinimalEthSpec; type E = MinimalEthSpec;
/// Merge block with length < max_rpc_size. /// Merge block with length < max_rpc_size.
fn merge_block_small(fork_context: &ForkContext) -> BeaconBlock<E> { fn merge_block_small(fork_context: &ForkContext, spec: &ChainSpec) -> BeaconBlock<E> {
let mut block = BeaconBlockMerge::<E>::empty(&E::default_spec()); let mut block = BeaconBlockMerge::<E>::empty(spec);
let tx = VariableList::from(vec![0; 1024]); let tx = VariableList::from(vec![0; 1024]);
let txs = VariableList::from(std::iter::repeat(tx).take(5000).collect::<Vec<_>>()); let txs = VariableList::from(std::iter::repeat(tx).take(5000).collect::<Vec<_>>());
block.body.execution_payload.execution_payload.transactions = txs; block.body.execution_payload.execution_payload.transactions = txs;
let block = BeaconBlock::Merge(block); let block = BeaconBlock::Merge(block);
assert!(block.ssz_bytes_len() <= max_rpc_size(fork_context)); assert!(block.ssz_bytes_len() <= max_rpc_size(fork_context, spec.max_chunk_size as usize));
block block
} }
/// Merge block with length > MAX_RPC_SIZE. /// Merge block with length > MAX_RPC_SIZE.
/// The max limit for a merge block is in the order of ~16GiB which wouldn't fit in memory. /// The max limit for a merge block is in the order of ~16GiB which wouldn't fit in memory.
/// Hence, we generate a merge block just greater than `MAX_RPC_SIZE` to test rejection on the rpc layer. /// Hence, we generate a merge block just greater than `MAX_RPC_SIZE` to test rejection on the rpc layer.
fn merge_block_large(fork_context: &ForkContext) -> BeaconBlock<E> { fn merge_block_large(fork_context: &ForkContext, spec: &ChainSpec) -> BeaconBlock<E> {
let mut block = BeaconBlockMerge::<E>::empty(&E::default_spec()); let mut block = BeaconBlockMerge::<E>::empty(spec);
let tx = VariableList::from(vec![0; 1024]); let tx = VariableList::from(vec![0; 1024]);
let txs = VariableList::from(std::iter::repeat(tx).take(100000).collect::<Vec<_>>()); let txs = VariableList::from(std::iter::repeat(tx).take(100000).collect::<Vec<_>>());
block.body.execution_payload.execution_payload.transactions = txs; block.body.execution_payload.execution_payload.transactions = txs;
let block = BeaconBlock::Merge(block); let block = BeaconBlock::Merge(block);
assert!(block.ssz_bytes_len() > max_rpc_size(fork_context)); assert!(block.ssz_bytes_len() > max_rpc_size(fork_context, spec.max_chunk_size as usize));
block block
} }
@ -57,10 +58,12 @@ fn test_status_rpc() {
let log = common::build_log(log_level, enable_logging); let log = common::build_log(log_level, enable_logging);
let spec = E::default_spec();
rt.block_on(async { rt.block_on(async {
// get sender/receiver // get sender/receiver
let (mut sender, mut receiver) = let (mut sender, mut receiver) =
common::build_node_pair(Arc::downgrade(&rt), &log, ForkName::Base).await; common::build_node_pair(Arc::downgrade(&rt), &log, ForkName::Base, &spec).await;
// Dummy STATUS RPC message // Dummy STATUS RPC message
let rpc_request = Request::Status(StatusMessage { let rpc_request = Request::Status(StatusMessage {
@ -149,10 +152,12 @@ fn test_blocks_by_range_chunked_rpc() {
let rt = Arc::new(Runtime::new().unwrap()); let rt = Arc::new(Runtime::new().unwrap());
let spec = E::default_spec();
rt.block_on(async { rt.block_on(async {
// get sender/receiver // get sender/receiver
let (mut sender, mut receiver) = let (mut sender, mut receiver) =
common::build_node_pair(Arc::downgrade(&rt), &log, ForkName::Merge).await; common::build_node_pair(Arc::downgrade(&rt), &log, ForkName::Merge, &spec).await;
// BlocksByRange Request // BlocksByRange Request
let rpc_request = Request::BlocksByRange(BlocksByRangeRequest::new(0, messages_to_send)); let rpc_request = Request::BlocksByRange(BlocksByRangeRequest::new(0, messages_to_send));
@ -168,7 +173,7 @@ fn test_blocks_by_range_chunked_rpc() {
let signed_full_block = SignedBeaconBlock::from_block(full_block, Signature::empty()); let signed_full_block = SignedBeaconBlock::from_block(full_block, Signature::empty());
let rpc_response_altair = Response::BlocksByRange(Some(Arc::new(signed_full_block))); let rpc_response_altair = Response::BlocksByRange(Some(Arc::new(signed_full_block)));
let full_block = merge_block_small(&common::fork_context(ForkName::Merge)); let full_block = merge_block_small(&common::fork_context(ForkName::Merge), &spec);
let signed_full_block = SignedBeaconBlock::from_block(full_block, Signature::empty()); let signed_full_block = SignedBeaconBlock::from_block(full_block, Signature::empty());
let rpc_response_merge_small = Response::BlocksByRange(Some(Arc::new(signed_full_block))); let rpc_response_merge_small = Response::BlocksByRange(Some(Arc::new(signed_full_block)));
@ -273,16 +278,18 @@ fn test_blocks_by_range_over_limit() {
let rt = Arc::new(Runtime::new().unwrap()); let rt = Arc::new(Runtime::new().unwrap());
let spec = E::default_spec();
rt.block_on(async { rt.block_on(async {
// get sender/receiver // get sender/receiver
let (mut sender, mut receiver) = let (mut sender, mut receiver) =
common::build_node_pair(Arc::downgrade(&rt), &log, ForkName::Merge).await; common::build_node_pair(Arc::downgrade(&rt), &log, ForkName::Merge, &spec).await;
// BlocksByRange Request // BlocksByRange Request
let rpc_request = Request::BlocksByRange(BlocksByRangeRequest::new(0, messages_to_send)); let rpc_request = Request::BlocksByRange(BlocksByRangeRequest::new(0, messages_to_send));
// BlocksByRange Response // BlocksByRange Response
let full_block = merge_block_large(&common::fork_context(ForkName::Merge)); let full_block = merge_block_large(&common::fork_context(ForkName::Merge), &spec);
let signed_full_block = SignedBeaconBlock::from_block(full_block, Signature::empty()); let signed_full_block = SignedBeaconBlock::from_block(full_block, Signature::empty());
let rpc_response_merge_large = Response::BlocksByRange(Some(Arc::new(signed_full_block))); let rpc_response_merge_large = Response::BlocksByRange(Some(Arc::new(signed_full_block)));
@ -355,10 +362,12 @@ fn test_blocks_by_range_chunked_rpc_terminates_correctly() {
let rt = Arc::new(Runtime::new().unwrap()); let rt = Arc::new(Runtime::new().unwrap());
let spec = E::default_spec();
rt.block_on(async { rt.block_on(async {
// get sender/receiver // get sender/receiver
let (mut sender, mut receiver) = let (mut sender, mut receiver) =
common::build_node_pair(Arc::downgrade(&rt), &log, ForkName::Base).await; common::build_node_pair(Arc::downgrade(&rt), &log, ForkName::Base, &spec).await;
// BlocksByRange Request // BlocksByRange Request
let rpc_request = Request::BlocksByRange(BlocksByRangeRequest::new(0, messages_to_send)); let rpc_request = Request::BlocksByRange(BlocksByRangeRequest::new(0, messages_to_send));
@ -475,10 +484,12 @@ fn test_blocks_by_range_single_empty_rpc() {
let log = common::build_log(log_level, enable_logging); let log = common::build_log(log_level, enable_logging);
let rt = Arc::new(Runtime::new().unwrap()); let rt = Arc::new(Runtime::new().unwrap());
let spec = E::default_spec();
rt.block_on(async { rt.block_on(async {
// get sender/receiver // get sender/receiver
let (mut sender, mut receiver) = let (mut sender, mut receiver) =
common::build_node_pair(Arc::downgrade(&rt), &log, ForkName::Base).await; common::build_node_pair(Arc::downgrade(&rt), &log, ForkName::Base, &spec).await;
// BlocksByRange Request // BlocksByRange Request
let rpc_request = Request::BlocksByRange(BlocksByRangeRequest::new(0, 10)); let rpc_request = Request::BlocksByRange(BlocksByRangeRequest::new(0, 10));
@ -579,7 +590,7 @@ fn test_blocks_by_root_chunked_rpc() {
// get sender/receiver // get sender/receiver
rt.block_on(async { rt.block_on(async {
let (mut sender, mut receiver) = let (mut sender, mut receiver) =
common::build_node_pair(Arc::downgrade(&rt), &log, ForkName::Merge).await; common::build_node_pair(Arc::downgrade(&rt), &log, ForkName::Merge, &spec).await;
// BlocksByRoot Request // BlocksByRoot Request
let rpc_request = let rpc_request =
@ -601,7 +612,7 @@ fn test_blocks_by_root_chunked_rpc() {
let signed_full_block = SignedBeaconBlock::from_block(full_block, Signature::empty()); let signed_full_block = SignedBeaconBlock::from_block(full_block, Signature::empty());
let rpc_response_altair = Response::BlocksByRoot(Some(Arc::new(signed_full_block))); let rpc_response_altair = Response::BlocksByRoot(Some(Arc::new(signed_full_block)));
let full_block = merge_block_small(&common::fork_context(ForkName::Merge)); let full_block = merge_block_small(&common::fork_context(ForkName::Merge), &spec);
let signed_full_block = SignedBeaconBlock::from_block(full_block, Signature::empty()); let signed_full_block = SignedBeaconBlock::from_block(full_block, Signature::empty());
let rpc_response_merge_small = Response::BlocksByRoot(Some(Arc::new(signed_full_block))); let rpc_response_merge_small = Response::BlocksByRoot(Some(Arc::new(signed_full_block)));
@ -706,7 +717,7 @@ fn test_blocks_by_root_chunked_rpc_terminates_correctly() {
// get sender/receiver // get sender/receiver
rt.block_on(async { rt.block_on(async {
let (mut sender, mut receiver) = let (mut sender, mut receiver) =
common::build_node_pair(Arc::downgrade(&rt), &log, ForkName::Base).await; common::build_node_pair(Arc::downgrade(&rt), &log, ForkName::Base, &spec).await;
// BlocksByRoot Request // BlocksByRoot Request
let rpc_request = let rpc_request =
@ -833,10 +844,13 @@ fn test_goodbye_rpc() {
let log = common::build_log(log_level, enable_logging); let log = common::build_log(log_level, enable_logging);
let rt = Arc::new(Runtime::new().unwrap()); let rt = Arc::new(Runtime::new().unwrap());
let spec = E::default_spec();
// get sender/receiver // get sender/receiver
rt.block_on(async { rt.block_on(async {
let (mut sender, mut receiver) = let (mut sender, mut receiver) =
common::build_node_pair(Arc::downgrade(&rt), &log, ForkName::Base).await; common::build_node_pair(Arc::downgrade(&rt), &log, ForkName::Base, &spec).await;
// build the sender future // build the sender future
let sender_future = async { let sender_future = async {

View File

@ -1634,6 +1634,7 @@ impl<T: BeaconChainTypes> NetworkBeaconProcessor<T> {
attestation_verification::verify_propagation_slot_range( attestation_verification::verify_propagation_slot_range(
seen_clock, seen_clock,
failed_att.attestation(), failed_att.attestation(),
&self.chain.spec,
); );
// Only penalize the peer if it would have been invalid at the moment we received // Only penalize the peer if it would have been invalid at the moment we received
@ -2182,6 +2183,7 @@ impl<T: BeaconChainTypes> NetworkBeaconProcessor<T> {
sync_committee_verification::verify_propagation_slot_range( sync_committee_verification::verify_propagation_slot_range(
seen_clock, seen_clock,
&sync_committee_message_slot, &sync_committee_message_slot,
&self.chain.spec,
); );
hindsight_verification.is_err() hindsight_verification.is_err()
}; };
@ -2494,6 +2496,7 @@ impl<T: BeaconChainTypes> NetworkBeaconProcessor<T> {
let is_timely = attestation_verification::verify_propagation_slot_range( let is_timely = attestation_verification::verify_propagation_slot_range(
&self.chain.slot_clock, &self.chain.slot_clock,
attestation, attestation,
&self.chain.spec,
) )
.is_ok(); .is_ok();

View File

@ -577,14 +577,3 @@ impl<E: EthSpec> NetworkBeaconProcessor<TestBeaconChainType<E>> {
(network_beacon_processor, beacon_processor_receive) (network_beacon_processor, beacon_processor_receive)
} }
} }
#[cfg(test)]
mod test {
#[test]
fn queued_block_delay_is_sane() {
assert!(
beacon_processor::work_reprocessing_queue::ADDITIONAL_QUEUED_BLOCK_DELAY
< beacon_chain::MAXIMUM_GOSSIP_CLOCK_DISPARITY
);
}
}

View File

@ -11,7 +11,7 @@ use crate::{
use beacon_chain::test_utils::{ use beacon_chain::test_utils::{
AttestationStrategy, BeaconChainHarness, BlockStrategy, EphemeralHarnessType, AttestationStrategy, BeaconChainHarness, BlockStrategy, EphemeralHarnessType,
}; };
use beacon_chain::{BeaconChain, ChainConfig, MAXIMUM_GOSSIP_CLOCK_DISPARITY}; use beacon_chain::{BeaconChain, ChainConfig};
use beacon_processor::{work_reprocessing_queue::*, *}; use beacon_processor::{work_reprocessing_queue::*, *};
use lighthouse_network::{ use lighthouse_network::{
discv5::enr::{CombinedKey, EnrBuilder}, discv5::enr::{CombinedKey, EnrBuilder},
@ -215,7 +215,7 @@ impl TestRig {
}; };
let network_beacon_processor = Arc::new(network_beacon_processor); let network_beacon_processor = Arc::new(network_beacon_processor);
BeaconProcessor { let beacon_processor = BeaconProcessor {
network_globals, network_globals,
executor, executor,
max_workers: cmp::max(1, num_cpus::get()), max_workers: cmp::max(1, num_cpus::get()),
@ -229,8 +229,11 @@ impl TestRig {
work_reprocessing_rx, work_reprocessing_rx,
Some(work_journal_tx), Some(work_journal_tx),
harness.chain.slot_clock.clone(), harness.chain.slot_clock.clone(),
chain.spec.maximum_gossip_clock_disparity(),
); );
assert!(!beacon_processor.is_err());
Self { Self {
chain, chain,
next_block: Arc::new(next_block), next_block: Arc::new(next_block),
@ -505,7 +508,7 @@ async fn import_gossip_block_acceptably_early() {
rig.chain rig.chain
.slot_clock .slot_clock
.set_current_time(slot_start - MAXIMUM_GOSSIP_CLOCK_DISPARITY); .set_current_time(slot_start - rig.chain.spec.maximum_gossip_clock_disparity());
assert_eq!( assert_eq!(
rig.chain.slot().unwrap(), rig.chain.slot().unwrap(),
@ -552,9 +555,9 @@ async fn import_gossip_block_unacceptably_early() {
.start_of(rig.next_block.slot()) .start_of(rig.next_block.slot())
.unwrap(); .unwrap();
rig.chain rig.chain.slot_clock.set_current_time(
.slot_clock slot_start - rig.chain.spec.maximum_gossip_clock_disparity() - Duration::from_millis(1),
.set_current_time(slot_start - MAXIMUM_GOSSIP_CLOCK_DISPARITY - Duration::from_millis(1)); );
assert_eq!( assert_eq!(
rig.chain.slot().unwrap(), rig.chain.slot().unwrap(),

View File

@ -581,8 +581,10 @@ pub fn get_config<E: EthSpec>(
}; };
} }
client_config.chain.max_network_size = client_config.chain.max_network_size = lighthouse_network::gossip_max_size(
lighthouse_network::gossip_max_size(spec.bellatrix_fork_epoch.is_some()); spec.bellatrix_fork_epoch.is_some(),
spec.gossip_max_size as usize,
);
if cli_args.is_present("slasher") { if cli_args.is_present("slasher") {
let slasher_dir = if let Some(slasher_dir) = cli_args.value_of("slasher-dir") { let slasher_dir = if let Some(slasher_dir) = cli_args.value_of("slasher-dir") {

View File

@ -89,4 +89,14 @@ DEPOSIT_CONTRACT_ADDRESS: 0x0B98057eA310F4d31F2a452B414647007d1645d9
# Network # Network
# --------------------------------------------------------------- # ---------------------------------------------------------------
SUBNETS_PER_NODE: 4 SUBNETS_PER_NODE: 4
GOSSIP_MAX_SIZE: 10485760
MIN_EPOCHS_FOR_BLOCK_REQUESTS: 33024
MAX_CHUNK_SIZE: 10485760
TTFB_TIMEOUT: 5
RESP_TIMEOUT: 10
MESSAGE_DOMAIN_INVALID_SNAPPY: 0x00000000
MESSAGE_DOMAIN_VALID_SNAPPY: 0x01000000
ATTESTATION_SUBNET_COUNT: 64
ATTESTATION_SUBNET_EXTRA_BITS: 0
ATTESTATION_SUBNET_PREFIX_BITS: 6

View File

@ -89,4 +89,14 @@ DEPOSIT_CONTRACT_ADDRESS: 0x00000000219ab540356cBB839Cbe05303d7705Fa
# Network # Network
# --------------------------------------------------------------- # ---------------------------------------------------------------
SUBNETS_PER_NODE: 2 SUBNETS_PER_NODE: 2
GOSSIP_MAX_SIZE: 10485760
MIN_EPOCHS_FOR_BLOCK_REQUESTS: 33024
MAX_CHUNK_SIZE: 10485760
TTFB_TIMEOUT: 5
RESP_TIMEOUT: 10
MESSAGE_DOMAIN_INVALID_SNAPPY: 0x00000000
MESSAGE_DOMAIN_VALID_SNAPPY: 0x01000000
ATTESTATION_SUBNET_COUNT: 64
ATTESTATION_SUBNET_EXTRA_BITS: 0
ATTESTATION_SUBNET_PREFIX_BITS: 6

View File

@ -89,4 +89,14 @@ DEPOSIT_CONTRACT_ADDRESS: 0xff50ed3d0ec03aC01D4C79aAd74928BFF48a7b2b
# Network # Network
# --------------------------------------------------------------- # ---------------------------------------------------------------
SUBNETS_PER_NODE: 2 SUBNETS_PER_NODE: 2
GOSSIP_MAX_SIZE: 10485760
MIN_EPOCHS_FOR_BLOCK_REQUESTS: 33024
MAX_CHUNK_SIZE: 10485760
TTFB_TIMEOUT: 5
RESP_TIMEOUT: 10
MESSAGE_DOMAIN_INVALID_SNAPPY: 0x00000000
MESSAGE_DOMAIN_VALID_SNAPPY: 0x01000000
ATTESTATION_SUBNET_COUNT: 64
ATTESTATION_SUBNET_EXTRA_BITS: 0
ATTESTATION_SUBNET_PREFIX_BITS: 6

View File

@ -77,4 +77,14 @@ DEPOSIT_CONTRACT_ADDRESS: 0x7f02C3E3c98b133055B8B348B2Ac625669Ed295D
# Network # Network
# --------------------------------------------------------------- # ---------------------------------------------------------------
SUBNETS_PER_NODE: 2 SUBNETS_PER_NODE: 2
GOSSIP_MAX_SIZE: 10485760
MIN_EPOCHS_FOR_BLOCK_REQUESTS: 33024
MAX_CHUNK_SIZE: 10485760
TTFB_TIMEOUT: 5
RESP_TIMEOUT: 10
MESSAGE_DOMAIN_INVALID_SNAPPY: 0x00000000
MESSAGE_DOMAIN_VALID_SNAPPY: 0x01000000
ATTESTATION_SUBNET_COUNT: 64
ATTESTATION_SUBNET_EXTRA_BITS: 0
ATTESTATION_SUBNET_PREFIX_BITS: 6

View File

@ -60,6 +60,7 @@ beacon_chain = { path = "../../beacon_node/beacon_chain" }
eth2_interop_keypairs = { path = "../../common/eth2_interop_keypairs" } eth2_interop_keypairs = { path = "../../common/eth2_interop_keypairs" }
state_processing = { path = "../state_processing" } state_processing = { path = "../state_processing" }
tokio = "1.14.0" tokio = "1.14.0"
paste = "1.0.14"
[features] [features]
default = ["sqlite", "legacy-arith"] default = ["sqlite", "legacy-arith"]

View File

@ -6,6 +6,7 @@ use serde_derive::Deserialize;
use serde_utils::quoted_u64::MaybeQuoted; use serde_utils::quoted_u64::MaybeQuoted;
use std::fs::File; use std::fs::File;
use std::path::Path; use std::path::Path;
use std::time::Duration;
use tree_hash::TreeHash; use tree_hash::TreeHash;
/// Each of the BLS signature domains. /// Each of the BLS signature domains.
@ -170,7 +171,15 @@ pub struct ChainSpec {
pub attestation_subnet_count: u64, pub attestation_subnet_count: u64,
pub subnets_per_node: u8, pub subnets_per_node: u8,
pub epochs_per_subnet_subscription: u64, pub epochs_per_subnet_subscription: u64,
pub gossip_max_size: u64,
pub min_epochs_for_block_requests: u64,
pub max_chunk_size: u64,
pub ttfb_timeout: u64,
pub resp_timeout: u64,
pub message_domain_invalid_snappy: [u8; 4],
pub message_domain_valid_snappy: [u8; 4],
pub attestation_subnet_extra_bits: u8, pub attestation_subnet_extra_bits: u8,
pub attestation_subnet_prefix_bits: u8,
/* /*
* Application params * Application params
@ -451,10 +460,16 @@ impl ChainSpec {
Hash256::from(domain) Hash256::from(domain)
} }
#[allow(clippy::arithmetic_side_effects)] pub fn maximum_gossip_clock_disparity(&self) -> Duration {
pub const fn attestation_subnet_prefix_bits(&self) -> u32 { Duration::from_millis(self.maximum_gossip_clock_disparity_millis)
let attestation_subnet_count_bits = self.attestation_subnet_count.ilog2(); }
self.attestation_subnet_extra_bits as u32 + attestation_subnet_count_bits
pub fn ttfb_timeout(&self) -> Duration {
Duration::from_secs(self.ttfb_timeout)
}
pub fn resp_timeout(&self) -> Duration {
Duration::from_secs(self.resp_timeout)
} }
/// Returns a `ChainSpec` compatible with the Ethereum Foundation specification. /// Returns a `ChainSpec` compatible with the Ethereum Foundation specification.
@ -617,8 +632,15 @@ impl ChainSpec {
maximum_gossip_clock_disparity_millis: 500, maximum_gossip_clock_disparity_millis: 500,
target_aggregators_per_committee: 16, target_aggregators_per_committee: 16,
epochs_per_subnet_subscription: 256, epochs_per_subnet_subscription: 256,
attestation_subnet_extra_bits: 0, gossip_max_size: default_gossip_max_size(),
min_epochs_for_block_requests: default_min_epochs_for_block_requests(),
max_chunk_size: default_max_chunk_size(),
ttfb_timeout: default_ttfb_timeout(),
resp_timeout: default_resp_timeout(),
message_domain_invalid_snappy: default_message_domain_invalid_snappy(),
message_domain_valid_snappy: default_message_domain_valid_snappy(),
attestation_subnet_extra_bits: default_attestation_subnet_extra_bits(),
attestation_subnet_prefix_bits: default_attestation_subnet_prefix_bits(),
/* /*
* Application specific * Application specific
*/ */
@ -842,7 +864,15 @@ impl ChainSpec {
maximum_gossip_clock_disparity_millis: 500, maximum_gossip_clock_disparity_millis: 500,
target_aggregators_per_committee: 16, target_aggregators_per_committee: 16,
epochs_per_subnet_subscription: 256, epochs_per_subnet_subscription: 256,
attestation_subnet_extra_bits: 0, gossip_max_size: default_gossip_max_size(),
min_epochs_for_block_requests: default_min_epochs_for_block_requests(),
max_chunk_size: default_max_chunk_size(),
ttfb_timeout: default_ttfb_timeout(),
resp_timeout: default_resp_timeout(),
message_domain_invalid_snappy: default_message_domain_invalid_snappy(),
message_domain_valid_snappy: default_message_domain_valid_snappy(),
attestation_subnet_extra_bits: default_attestation_subnet_extra_bits(),
attestation_subnet_prefix_bits: default_attestation_subnet_prefix_bits(),
/* /*
* Application specific * Application specific
@ -953,6 +983,34 @@ pub struct Config {
#[serde(with = "serde_utils::quoted_u64")] #[serde(with = "serde_utils::quoted_u64")]
deposit_network_id: u64, deposit_network_id: u64,
deposit_contract_address: Address, deposit_contract_address: Address,
#[serde(default = "default_gossip_max_size")]
#[serde(with = "serde_utils::quoted_u64")]
gossip_max_size: u64,
#[serde(default = "default_min_epochs_for_block_requests")]
#[serde(with = "serde_utils::quoted_u64")]
min_epochs_for_block_requests: u64,
#[serde(default = "default_max_chunk_size")]
#[serde(with = "serde_utils::quoted_u64")]
max_chunk_size: u64,
#[serde(default = "default_ttfb_timeout")]
#[serde(with = "serde_utils::quoted_u64")]
ttfb_timeout: u64,
#[serde(default = "default_resp_timeout")]
#[serde(with = "serde_utils::quoted_u64")]
resp_timeout: u64,
#[serde(default = "default_message_domain_invalid_snappy")]
#[serde(with = "serde_utils::bytes_4_hex")]
message_domain_invalid_snappy: [u8; 4],
#[serde(default = "default_message_domain_valid_snappy")]
#[serde(with = "serde_utils::bytes_4_hex")]
message_domain_valid_snappy: [u8; 4],
#[serde(default = "default_attestation_subnet_extra_bits")]
#[serde(with = "serde_utils::quoted_u8")]
attestation_subnet_extra_bits: u8,
#[serde(default = "default_attestation_subnet_prefix_bits")]
#[serde(with = "serde_utils::quoted_u8")]
attestation_subnet_prefix_bits: u8,
} }
fn default_bellatrix_fork_version() -> [u8; 4] { fn default_bellatrix_fork_version() -> [u8; 4] {
@ -993,6 +1051,42 @@ fn default_subnets_per_node() -> u8 {
2u8 2u8
} }
const fn default_gossip_max_size() -> u64 {
10485760
}
const fn default_min_epochs_for_block_requests() -> u64 {
33024
}
const fn default_max_chunk_size() -> u64 {
10485760
}
const fn default_ttfb_timeout() -> u64 {
5
}
const fn default_resp_timeout() -> u64 {
10
}
const fn default_message_domain_invalid_snappy() -> [u8; 4] {
[0, 0, 0, 0]
}
const fn default_message_domain_valid_snappy() -> [u8; 4] {
[1, 0, 0, 0]
}
const fn default_attestation_subnet_extra_bits() -> u8 {
0
}
const fn default_attestation_subnet_prefix_bits() -> u8 {
6
}
impl Default for Config { impl Default for Config {
fn default() -> Self { fn default() -> Self {
let chain_spec = MainnetEthSpec::default_spec(); let chain_spec = MainnetEthSpec::default_spec();
@ -1088,6 +1182,16 @@ impl Config {
deposit_chain_id: spec.deposit_chain_id, deposit_chain_id: spec.deposit_chain_id,
deposit_network_id: spec.deposit_network_id, deposit_network_id: spec.deposit_network_id,
deposit_contract_address: spec.deposit_contract_address, deposit_contract_address: spec.deposit_contract_address,
gossip_max_size: spec.gossip_max_size,
min_epochs_for_block_requests: spec.min_epochs_for_block_requests,
max_chunk_size: spec.max_chunk_size,
ttfb_timeout: spec.ttfb_timeout,
resp_timeout: spec.resp_timeout,
message_domain_invalid_snappy: spec.message_domain_invalid_snappy,
message_domain_valid_snappy: spec.message_domain_valid_snappy,
attestation_subnet_extra_bits: spec.attestation_subnet_extra_bits,
attestation_subnet_prefix_bits: spec.attestation_subnet_prefix_bits,
} }
} }
@ -1132,6 +1236,15 @@ impl Config {
deposit_chain_id, deposit_chain_id,
deposit_network_id, deposit_network_id,
deposit_contract_address, deposit_contract_address,
gossip_max_size,
min_epochs_for_block_requests,
max_chunk_size,
ttfb_timeout,
resp_timeout,
message_domain_invalid_snappy,
message_domain_valid_snappy,
attestation_subnet_extra_bits,
attestation_subnet_prefix_bits,
} = self; } = self;
if preset_base != T::spec_name().to_string().as_str() { if preset_base != T::spec_name().to_string().as_str() {
@ -1169,6 +1282,15 @@ impl Config {
terminal_block_hash, terminal_block_hash,
terminal_block_hash_activation_epoch, terminal_block_hash_activation_epoch,
safe_slots_to_import_optimistically, safe_slots_to_import_optimistically,
gossip_max_size,
min_epochs_for_block_requests,
max_chunk_size,
ttfb_timeout,
resp_timeout,
message_domain_invalid_snappy,
message_domain_valid_snappy,
attestation_subnet_extra_bits,
attestation_subnet_prefix_bits,
..chain_spec.clone() ..chain_spec.clone()
}) })
} }
@ -1306,6 +1428,7 @@ mod tests {
#[cfg(test)] #[cfg(test)]
mod yaml_tests { mod yaml_tests {
use super::*; use super::*;
use paste::paste;
use tempfile::NamedTempFile; use tempfile::NamedTempFile;
#[test] #[test]
@ -1410,29 +1533,35 @@ mod yaml_tests {
"#; "#;
let chain_spec: Config = serde_yaml::from_str(spec).unwrap(); let chain_spec: Config = serde_yaml::from_str(spec).unwrap();
assert_eq!(
chain_spec.terminal_total_difficulty, // Asserts that `chain_spec.$name` and `default_$name()` are equal.
default_terminal_total_difficulty() macro_rules! check_default {
); ($name: ident) => {
assert_eq!( paste! {
chain_spec.terminal_block_hash, assert_eq!(
default_terminal_block_hash() chain_spec.$name,
); [<default_ $name>](),
assert_eq!( "{} does not match default", stringify!($name));
chain_spec.terminal_block_hash_activation_epoch, }
default_terminal_block_hash_activation_epoch() };
); }
assert_eq!(
chain_spec.safe_slots_to_import_optimistically, check_default!(terminal_total_difficulty);
default_safe_slots_to_import_optimistically() check_default!(terminal_block_hash);
); check_default!(terminal_block_hash_activation_epoch);
check_default!(safe_slots_to_import_optimistically);
check_default!(bellatrix_fork_version);
check_default!(gossip_max_size);
check_default!(min_epochs_for_block_requests);
check_default!(max_chunk_size);
check_default!(ttfb_timeout);
check_default!(resp_timeout);
check_default!(message_domain_invalid_snappy);
check_default!(message_domain_valid_snappy);
check_default!(attestation_subnet_extra_bits);
check_default!(attestation_subnet_prefix_bits);
assert_eq!(chain_spec.bellatrix_fork_epoch, None); assert_eq!(chain_spec.bellatrix_fork_epoch, None);
assert_eq!(
chain_spec.bellatrix_fork_version,
default_bellatrix_fork_version()
);
} }
#[test] #[test]

View File

@ -84,7 +84,7 @@ impl SubnetId {
let subscription_duration = spec.epochs_per_subnet_subscription; let subscription_duration = spec.epochs_per_subnet_subscription;
let node_id_prefix = let node_id_prefix =
(node_id >> (256 - spec.attestation_subnet_prefix_bits() as usize)).as_usize(); (node_id >> (256 - spec.attestation_subnet_prefix_bits as usize)).as_usize();
// NOTE: The as_u64() panics if the number is larger than u64::max_value(). This cannot be // NOTE: The as_u64() panics if the number is larger than u64::max_value(). This cannot be
// true as spec.epochs_per_subnet_subscription is a u64. // true as spec.epochs_per_subnet_subscription is a u64.
@ -99,7 +99,7 @@ impl SubnetId {
let permutation_seed = let permutation_seed =
ethereum_hashing::hash(&int_to_bytes::int_to_bytes8(subscription_event_idx)); ethereum_hashing::hash(&int_to_bytes::int_to_bytes8(subscription_event_idx));
let num_subnets = 1 << spec.attestation_subnet_prefix_bits(); let num_subnets = 1 << spec.attestation_subnet_prefix_bits;
let permutated_prefix = compute_shuffled_index( let permutated_prefix = compute_shuffled_index(
node_id_prefix, node_id_prefix,
num_subnets, num_subnets,

View File

@ -81,3 +81,17 @@ PROPOSER_SCORE_BOOST: 40
DEPOSIT_CHAIN_ID: 1 DEPOSIT_CHAIN_ID: 1
DEPOSIT_NETWORK_ID: 1 DEPOSIT_NETWORK_ID: 1
DEPOSIT_CONTRACT_ADDRESS: 0x00000000219ab540356cBB839Cbe05303d7705Fa DEPOSIT_CONTRACT_ADDRESS: 0x00000000219ab540356cBB839Cbe05303d7705Fa
# Network
# ---------------------------------------------------------------
SUBNETS_PER_NODE: 2
GOSSIP_MAX_SIZE: 10485760
MIN_EPOCHS_FOR_BLOCK_REQUESTS: 33024
MAX_CHUNK_SIZE: 10485760
TTFB_TIMEOUT: 5
RESP_TIMEOUT: 10
MESSAGE_DOMAIN_INVALID_SNAPPY: 0x00000000
MESSAGE_DOMAIN_VALID_SNAPPY: 0x01000000
ATTESTATION_SUBNET_COUNT: 64
ATTESTATION_SUBNET_EXTRA_BITS: 0
ATTESTATION_SUBNET_PREFIX_BITS: 6