Merge pull request #293 from michaelsproul/operation-pool
Implement Operation Pool
This commit is contained in:
commit
206f651895
@ -3,6 +3,7 @@ members = [
|
|||||||
"eth2/attester",
|
"eth2/attester",
|
||||||
"eth2/block_proposer",
|
"eth2/block_proposer",
|
||||||
"eth2/fork_choice",
|
"eth2/fork_choice",
|
||||||
|
"eth2/operation_pool",
|
||||||
"eth2/state_processing",
|
"eth2/state_processing",
|
||||||
"eth2/state_processing/yaml_utils",
|
"eth2/state_processing/yaml_utils",
|
||||||
"eth2/types",
|
"eth2/types",
|
||||||
|
@ -15,6 +15,7 @@ hashing = { path = "../../eth2/utils/hashing" }
|
|||||||
fork_choice = { path = "../../eth2/fork_choice" }
|
fork_choice = { path = "../../eth2/fork_choice" }
|
||||||
parking_lot = "0.7"
|
parking_lot = "0.7"
|
||||||
log = "0.4"
|
log = "0.4"
|
||||||
|
operation_pool = { path = "../../eth2/operation_pool" }
|
||||||
env_logger = "0.6"
|
env_logger = "0.6"
|
||||||
serde = "1.0"
|
serde = "1.0"
|
||||||
serde_derive = "1.0"
|
serde_derive = "1.0"
|
||||||
|
@ -1,4 +1,3 @@
|
|||||||
use crate::attestation_aggregator::{AttestationAggregator, Outcome as AggregationOutcome};
|
|
||||||
use crate::checkpoint::CheckPoint;
|
use crate::checkpoint::CheckPoint;
|
||||||
use crate::errors::{BeaconChainError as Error, BlockProductionError};
|
use crate::errors::{BeaconChainError as Error, BlockProductionError};
|
||||||
use db::{
|
use db::{
|
||||||
@ -7,9 +6,15 @@ use db::{
|
|||||||
};
|
};
|
||||||
use fork_choice::{ForkChoice, ForkChoiceError};
|
use fork_choice::{ForkChoice, ForkChoiceError};
|
||||||
use log::{debug, trace};
|
use log::{debug, trace};
|
||||||
|
use operation_pool::DepositInsertStatus;
|
||||||
|
use operation_pool::OperationPool;
|
||||||
use parking_lot::{RwLock, RwLockReadGuard};
|
use parking_lot::{RwLock, RwLockReadGuard};
|
||||||
use slot_clock::SlotClock;
|
use slot_clock::SlotClock;
|
||||||
use ssz::ssz_encode;
|
use ssz::ssz_encode;
|
||||||
|
pub use state_processing::per_block_processing::errors::{
|
||||||
|
AttestationValidationError, AttesterSlashingValidationError, DepositValidationError,
|
||||||
|
ExitValidationError, ProposerSlashingValidationError, TransferValidationError,
|
||||||
|
};
|
||||||
use state_processing::{
|
use state_processing::{
|
||||||
per_block_processing, per_block_processing_without_verifying_block_signature,
|
per_block_processing, per_block_processing_without_verifying_block_signature,
|
||||||
per_slot_processing, BlockProcessingError, SlotProcessingError,
|
per_slot_processing, BlockProcessingError, SlotProcessingError,
|
||||||
@ -82,12 +87,7 @@ pub struct BeaconChain<T: ClientDB + Sized, U: SlotClock, F: ForkChoice> {
|
|||||||
pub block_store: Arc<BeaconBlockStore<T>>,
|
pub block_store: Arc<BeaconBlockStore<T>>,
|
||||||
pub state_store: Arc<BeaconStateStore<T>>,
|
pub state_store: Arc<BeaconStateStore<T>>,
|
||||||
pub slot_clock: U,
|
pub slot_clock: U,
|
||||||
pub attestation_aggregator: RwLock<AttestationAggregator>,
|
pub op_pool: OperationPool,
|
||||||
pub deposits_for_inclusion: RwLock<Vec<Deposit>>,
|
|
||||||
pub exits_for_inclusion: RwLock<Vec<VoluntaryExit>>,
|
|
||||||
pub transfers_for_inclusion: RwLock<Vec<Transfer>>,
|
|
||||||
pub proposer_slashings_for_inclusion: RwLock<Vec<ProposerSlashing>>,
|
|
||||||
pub attester_slashings_for_inclusion: RwLock<Vec<AttesterSlashing>>,
|
|
||||||
canonical_head: RwLock<CheckPoint>,
|
canonical_head: RwLock<CheckPoint>,
|
||||||
finalized_head: RwLock<CheckPoint>,
|
finalized_head: RwLock<CheckPoint>,
|
||||||
pub state: RwLock<BeaconState>,
|
pub state: RwLock<BeaconState>,
|
||||||
@ -129,7 +129,6 @@ where
|
|||||||
genesis_state.clone(),
|
genesis_state.clone(),
|
||||||
state_root,
|
state_root,
|
||||||
));
|
));
|
||||||
let attestation_aggregator = RwLock::new(AttestationAggregator::new());
|
|
||||||
|
|
||||||
genesis_state.build_epoch_cache(RelativeEpoch::Previous, &spec)?;
|
genesis_state.build_epoch_cache(RelativeEpoch::Previous, &spec)?;
|
||||||
genesis_state.build_epoch_cache(RelativeEpoch::Current, &spec)?;
|
genesis_state.build_epoch_cache(RelativeEpoch::Current, &spec)?;
|
||||||
@ -140,12 +139,7 @@ where
|
|||||||
block_store,
|
block_store,
|
||||||
state_store,
|
state_store,
|
||||||
slot_clock,
|
slot_clock,
|
||||||
attestation_aggregator,
|
op_pool: OperationPool::new(),
|
||||||
deposits_for_inclusion: RwLock::new(vec![]),
|
|
||||||
exits_for_inclusion: RwLock::new(vec![]),
|
|
||||||
transfers_for_inclusion: RwLock::new(vec![]),
|
|
||||||
proposer_slashings_for_inclusion: RwLock::new(vec![]),
|
|
||||||
attester_slashings_for_inclusion: RwLock::new(vec![]),
|
|
||||||
state: RwLock::new(genesis_state),
|
state: RwLock::new(genesis_state),
|
||||||
finalized_head,
|
finalized_head,
|
||||||
canonical_head,
|
canonical_head,
|
||||||
@ -479,13 +473,9 @@ where
|
|||||||
}
|
}
|
||||||
|
|
||||||
/// Produce an `AttestationData` that is valid for the present `slot` and given `shard`.
|
/// Produce an `AttestationData` that is valid for the present `slot` and given `shard`.
|
||||||
pub fn produce_attestation(&self, shard: u64) -> Result<AttestationData, Error> {
|
pub fn produce_attestation_data(&self, shard: u64) -> Result<AttestationData, Error> {
|
||||||
trace!("BeaconChain::produce_attestation: shard: {}", shard);
|
trace!("BeaconChain::produce_attestation: shard: {}", shard);
|
||||||
let source_epoch = self.state.read().current_justified_epoch;
|
let state = self.state.read();
|
||||||
let source_root = *self.state.read().get_block_root(
|
|
||||||
source_epoch.start_slot(self.spec.slots_per_epoch),
|
|
||||||
&self.spec,
|
|
||||||
)?;
|
|
||||||
|
|
||||||
let target_root = *self.state.read().get_block_root(
|
let target_root = *self.state.read().get_block_root(
|
||||||
self.state
|
self.state
|
||||||
@ -506,257 +496,60 @@ where
|
|||||||
epoch: self.state.read().slot.epoch(self.spec.slots_per_epoch),
|
epoch: self.state.read().slot.epoch(self.spec.slots_per_epoch),
|
||||||
crosslink_data_root: Hash256::zero(),
|
crosslink_data_root: Hash256::zero(),
|
||||||
},
|
},
|
||||||
source_epoch,
|
source_epoch: state.current_justified_epoch,
|
||||||
source_root,
|
source_root: state.current_justified_root,
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Validate a `FreeAttestation` and either:
|
/// Accept a new attestation from the network.
|
||||||
///
|
///
|
||||||
/// - Create a new `Attestation`.
|
/// If valid, the attestation is added to the `op_pool` and aggregated with another attestation
|
||||||
/// - Aggregate it to an existing `Attestation`.
|
/// if possible.
|
||||||
pub fn process_free_attestation(
|
pub fn process_attestation(
|
||||||
&self,
|
&self,
|
||||||
free_attestation: FreeAttestation,
|
attestation: Attestation,
|
||||||
) -> Result<AggregationOutcome, Error> {
|
) -> Result<(), AttestationValidationError> {
|
||||||
let aggregation_outcome = self
|
self.op_pool
|
||||||
.attestation_aggregator
|
.insert_attestation(attestation, &*self.state.read(), &self.spec)
|
||||||
.write()
|
|
||||||
.process_free_attestation(&self.state.read(), &free_attestation, &self.spec)?;
|
|
||||||
|
|
||||||
// return if the attestation is invalid
|
|
||||||
if !aggregation_outcome.valid {
|
|
||||||
return Ok(aggregation_outcome);
|
|
||||||
}
|
|
||||||
|
|
||||||
// valid attestation, proceed with fork-choice logic
|
|
||||||
self.fork_choice.write().add_attestation(
|
|
||||||
free_attestation.validator_index,
|
|
||||||
&free_attestation.data.beacon_block_root,
|
|
||||||
&self.spec,
|
|
||||||
)?;
|
|
||||||
Ok(aggregation_outcome)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Accept some deposit and queue it for inclusion in an appropriate block.
|
/// Accept some deposit and queue it for inclusion in an appropriate block.
|
||||||
pub fn receive_deposit_for_inclusion(&self, deposit: Deposit) {
|
pub fn process_deposit(
|
||||||
// TODO: deposits are not checked for validity; check them.
|
&self,
|
||||||
//
|
deposit: Deposit,
|
||||||
// https://github.com/sigp/lighthouse/issues/276
|
) -> Result<DepositInsertStatus, DepositValidationError> {
|
||||||
self.deposits_for_inclusion.write().push(deposit);
|
self.op_pool
|
||||||
}
|
.insert_deposit(deposit, &*self.state.read(), &self.spec)
|
||||||
|
|
||||||
/// Return a vec of deposits suitable for inclusion in some block.
|
|
||||||
pub fn get_deposits_for_block(&self) -> Vec<Deposit> {
|
|
||||||
// TODO: deposits are indiscriminately included; check them for validity.
|
|
||||||
//
|
|
||||||
// https://github.com/sigp/lighthouse/issues/275
|
|
||||||
self.deposits_for_inclusion.read().clone()
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Takes a list of `Deposits` that were included in recent blocks and removes them from the
|
|
||||||
/// inclusion queue.
|
|
||||||
///
|
|
||||||
/// This ensures that `Deposits` are not included twice in successive blocks.
|
|
||||||
pub fn set_deposits_as_included(&self, included_deposits: &[Deposit]) {
|
|
||||||
// TODO: method does not take forks into account; consider this.
|
|
||||||
//
|
|
||||||
// https://github.com/sigp/lighthouse/issues/275
|
|
||||||
let mut indices_to_delete = vec![];
|
|
||||||
|
|
||||||
for included in included_deposits {
|
|
||||||
for (i, for_inclusion) in self.deposits_for_inclusion.read().iter().enumerate() {
|
|
||||||
if included == for_inclusion {
|
|
||||||
indices_to_delete.push(i);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
let deposits_for_inclusion = &mut self.deposits_for_inclusion.write();
|
|
||||||
for i in indices_to_delete {
|
|
||||||
deposits_for_inclusion.remove(i);
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Accept some exit and queue it for inclusion in an appropriate block.
|
/// Accept some exit and queue it for inclusion in an appropriate block.
|
||||||
pub fn receive_exit_for_inclusion(&self, exit: VoluntaryExit) {
|
pub fn process_voluntary_exit(&self, exit: VoluntaryExit) -> Result<(), ExitValidationError> {
|
||||||
// TODO: exits are not checked for validity; check them.
|
self.op_pool
|
||||||
//
|
.insert_voluntary_exit(exit, &*self.state.read(), &self.spec)
|
||||||
// https://github.com/sigp/lighthouse/issues/276
|
|
||||||
self.exits_for_inclusion.write().push(exit);
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Return a vec of exits suitable for inclusion in some block.
|
|
||||||
pub fn get_exits_for_block(&self) -> Vec<VoluntaryExit> {
|
|
||||||
// TODO: exits are indiscriminately included; check them for validity.
|
|
||||||
//
|
|
||||||
// https://github.com/sigp/lighthouse/issues/275
|
|
||||||
self.exits_for_inclusion.read().clone()
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Takes a list of `Deposits` that were included in recent blocks and removes them from the
|
|
||||||
/// inclusion queue.
|
|
||||||
///
|
|
||||||
/// This ensures that `Deposits` are not included twice in successive blocks.
|
|
||||||
pub fn set_exits_as_included(&self, included_exits: &[VoluntaryExit]) {
|
|
||||||
// TODO: method does not take forks into account; consider this.
|
|
||||||
let mut indices_to_delete = vec![];
|
|
||||||
|
|
||||||
for included in included_exits {
|
|
||||||
for (i, for_inclusion) in self.exits_for_inclusion.read().iter().enumerate() {
|
|
||||||
if included == for_inclusion {
|
|
||||||
indices_to_delete.push(i);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
let exits_for_inclusion = &mut self.exits_for_inclusion.write();
|
|
||||||
for i in indices_to_delete {
|
|
||||||
exits_for_inclusion.remove(i);
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Accept some transfer and queue it for inclusion in an appropriate block.
|
/// Accept some transfer and queue it for inclusion in an appropriate block.
|
||||||
pub fn receive_transfer_for_inclusion(&self, transfer: Transfer) {
|
pub fn process_transfer(&self, transfer: Transfer) -> Result<(), TransferValidationError> {
|
||||||
// TODO: transfers are not checked for validity; check them.
|
self.op_pool
|
||||||
//
|
.insert_transfer(transfer, &*self.state.read(), &self.spec)
|
||||||
// https://github.com/sigp/lighthouse/issues/276
|
|
||||||
self.transfers_for_inclusion.write().push(transfer);
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Return a vec of transfers suitable for inclusion in some block.
|
|
||||||
pub fn get_transfers_for_block(&self) -> Vec<Transfer> {
|
|
||||||
// TODO: transfers are indiscriminately included; check them for validity.
|
|
||||||
//
|
|
||||||
// https://github.com/sigp/lighthouse/issues/275
|
|
||||||
self.transfers_for_inclusion.read().clone()
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Takes a list of `Deposits` that were included in recent blocks and removes them from the
|
|
||||||
/// inclusion queue.
|
|
||||||
///
|
|
||||||
/// This ensures that `Deposits` are not included twice in successive blocks.
|
|
||||||
pub fn set_transfers_as_included(&self, included_transfers: &[Transfer]) {
|
|
||||||
// TODO: method does not take forks into account; consider this.
|
|
||||||
let mut indices_to_delete = vec![];
|
|
||||||
|
|
||||||
for included in included_transfers {
|
|
||||||
for (i, for_inclusion) in self.transfers_for_inclusion.read().iter().enumerate() {
|
|
||||||
if included == for_inclusion {
|
|
||||||
indices_to_delete.push(i);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
let transfers_for_inclusion = &mut self.transfers_for_inclusion.write();
|
|
||||||
for i in indices_to_delete {
|
|
||||||
transfers_for_inclusion.remove(i);
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Accept some proposer slashing and queue it for inclusion in an appropriate block.
|
/// Accept some proposer slashing and queue it for inclusion in an appropriate block.
|
||||||
pub fn receive_proposer_slashing_for_inclusion(&self, proposer_slashing: ProposerSlashing) {
|
pub fn process_proposer_slashing(
|
||||||
// TODO: proposer_slashings are not checked for validity; check them.
|
|
||||||
//
|
|
||||||
// https://github.com/sigp/lighthouse/issues/276
|
|
||||||
self.proposer_slashings_for_inclusion
|
|
||||||
.write()
|
|
||||||
.push(proposer_slashing);
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Return a vec of proposer slashings suitable for inclusion in some block.
|
|
||||||
pub fn get_proposer_slashings_for_block(&self) -> Vec<ProposerSlashing> {
|
|
||||||
// TODO: proposer_slashings are indiscriminately included; check them for validity.
|
|
||||||
//
|
|
||||||
// https://github.com/sigp/lighthouse/issues/275
|
|
||||||
self.proposer_slashings_for_inclusion.read().clone()
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Takes a list of `ProposerSlashings` that were included in recent blocks and removes them
|
|
||||||
/// from the inclusion queue.
|
|
||||||
///
|
|
||||||
/// This ensures that `ProposerSlashings` are not included twice in successive blocks.
|
|
||||||
pub fn set_proposer_slashings_as_included(
|
|
||||||
&self,
|
&self,
|
||||||
included_proposer_slashings: &[ProposerSlashing],
|
proposer_slashing: ProposerSlashing,
|
||||||
) {
|
) -> Result<(), ProposerSlashingValidationError> {
|
||||||
// TODO: method does not take forks into account; consider this.
|
self.op_pool
|
||||||
//
|
.insert_proposer_slashing(proposer_slashing, &*self.state.read(), &self.spec)
|
||||||
// https://github.com/sigp/lighthouse/issues/275
|
|
||||||
let mut indices_to_delete = vec![];
|
|
||||||
|
|
||||||
for included in included_proposer_slashings {
|
|
||||||
for (i, for_inclusion) in self
|
|
||||||
.proposer_slashings_for_inclusion
|
|
||||||
.read()
|
|
||||||
.iter()
|
|
||||||
.enumerate()
|
|
||||||
{
|
|
||||||
if included == for_inclusion {
|
|
||||||
indices_to_delete.push(i);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
let proposer_slashings_for_inclusion = &mut self.proposer_slashings_for_inclusion.write();
|
|
||||||
for i in indices_to_delete {
|
|
||||||
proposer_slashings_for_inclusion.remove(i);
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Accept some attester slashing and queue it for inclusion in an appropriate block.
|
/// Accept some attester slashing and queue it for inclusion in an appropriate block.
|
||||||
pub fn receive_attester_slashing_for_inclusion(&self, attester_slashing: AttesterSlashing) {
|
pub fn process_attester_slashing(
|
||||||
// TODO: attester_slashings are not checked for validity; check them.
|
|
||||||
//
|
|
||||||
// https://github.com/sigp/lighthouse/issues/276
|
|
||||||
self.attester_slashings_for_inclusion
|
|
||||||
.write()
|
|
||||||
.push(attester_slashing);
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Return a vec of attester slashings suitable for inclusion in some block.
|
|
||||||
pub fn get_attester_slashings_for_block(&self) -> Vec<AttesterSlashing> {
|
|
||||||
// TODO: attester_slashings are indiscriminately included; check them for validity.
|
|
||||||
//
|
|
||||||
// https://github.com/sigp/lighthouse/issues/275
|
|
||||||
self.attester_slashings_for_inclusion.read().clone()
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Takes a list of `AttesterSlashings` that were included in recent blocks and removes them
|
|
||||||
/// from the inclusion queue.
|
|
||||||
///
|
|
||||||
/// This ensures that `AttesterSlashings` are not included twice in successive blocks.
|
|
||||||
pub fn set_attester_slashings_as_included(
|
|
||||||
&self,
|
&self,
|
||||||
included_attester_slashings: &[AttesterSlashing],
|
attester_slashing: AttesterSlashing,
|
||||||
) {
|
) -> Result<(), AttesterSlashingValidationError> {
|
||||||
// TODO: method does not take forks into account; consider this.
|
self.op_pool
|
||||||
//
|
.insert_attester_slashing(attester_slashing, &*self.state.read(), &self.spec)
|
||||||
// https://github.com/sigp/lighthouse/issues/275
|
|
||||||
let mut indices_to_delete = vec![];
|
|
||||||
|
|
||||||
for included in included_attester_slashings {
|
|
||||||
for (i, for_inclusion) in self
|
|
||||||
.attester_slashings_for_inclusion
|
|
||||||
.read()
|
|
||||||
.iter()
|
|
||||||
.enumerate()
|
|
||||||
{
|
|
||||||
if included == for_inclusion {
|
|
||||||
indices_to_delete.push(i);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
let attester_slashings_for_inclusion = &mut self.attester_slashings_for_inclusion.write();
|
|
||||||
for i in indices_to_delete {
|
|
||||||
attester_slashings_for_inclusion.remove(i);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
/// Returns `true` if the given block root has not been processed.
|
|
||||||
pub fn is_new_block_root(&self, beacon_block_root: &Hash256) -> Result<bool, Error> {
|
|
||||||
Ok(!self.block_store.exists(beacon_block_root)?)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Accept some block and attempt to add it to block DAG.
|
/// Accept some block and attempt to add it to block DAG.
|
||||||
@ -832,13 +625,6 @@ where
|
|||||||
self.block_store.put(&block_root, &ssz_encode(&block)[..])?;
|
self.block_store.put(&block_root, &ssz_encode(&block)[..])?;
|
||||||
self.state_store.put(&state_root, &ssz_encode(&state)[..])?;
|
self.state_store.put(&state_root, &ssz_encode(&state)[..])?;
|
||||||
|
|
||||||
// Update the inclusion queues so they aren't re-submitted.
|
|
||||||
self.set_deposits_as_included(&block.body.deposits[..]);
|
|
||||||
self.set_transfers_as_included(&block.body.transfers[..]);
|
|
||||||
self.set_exits_as_included(&block.body.voluntary_exits[..]);
|
|
||||||
self.set_proposer_slashings_as_included(&block.body.proposer_slashings[..]);
|
|
||||||
self.set_attester_slashings_as_included(&block.body.attester_slashings[..]);
|
|
||||||
|
|
||||||
// run the fork_choice add_block logic
|
// run the fork_choice add_block logic
|
||||||
self.fork_choice
|
self.fork_choice
|
||||||
.write()
|
.write()
|
||||||
@ -874,20 +660,13 @@ where
|
|||||||
|
|
||||||
trace!("Finding attestations for new block...");
|
trace!("Finding attestations for new block...");
|
||||||
|
|
||||||
let attestations = self
|
|
||||||
.attestation_aggregator
|
|
||||||
.read()
|
|
||||||
.get_attestations_for_state(&state, &self.spec);
|
|
||||||
|
|
||||||
trace!(
|
|
||||||
"Inserting {} attestation(s) into new block.",
|
|
||||||
attestations.len()
|
|
||||||
);
|
|
||||||
|
|
||||||
let previous_block_root = *state
|
let previous_block_root = *state
|
||||||
.get_block_root(state.slot - 1, &self.spec)
|
.get_block_root(state.slot - 1, &self.spec)
|
||||||
.map_err(|_| BlockProductionError::UnableToGetBlockRootFromState)?;
|
.map_err(|_| BlockProductionError::UnableToGetBlockRootFromState)?;
|
||||||
|
|
||||||
|
let (proposer_slashings, attester_slashings) =
|
||||||
|
self.op_pool.get_slashings(&*self.state.read(), &self.spec);
|
||||||
|
|
||||||
let mut block = BeaconBlock {
|
let mut block = BeaconBlock {
|
||||||
slot: state.slot,
|
slot: state.slot,
|
||||||
previous_block_root,
|
previous_block_root,
|
||||||
@ -900,16 +679,23 @@ where
|
|||||||
deposit_root: Hash256::zero(),
|
deposit_root: Hash256::zero(),
|
||||||
block_hash: Hash256::zero(),
|
block_hash: Hash256::zero(),
|
||||||
},
|
},
|
||||||
proposer_slashings: self.get_proposer_slashings_for_block(),
|
proposer_slashings,
|
||||||
attester_slashings: self.get_attester_slashings_for_block(),
|
attester_slashings,
|
||||||
attestations,
|
attestations: self
|
||||||
deposits: self.get_deposits_for_block(),
|
.op_pool
|
||||||
voluntary_exits: self.get_exits_for_block(),
|
.get_attestations(&*self.state.read(), &self.spec),
|
||||||
transfers: self.get_transfers_for_block(),
|
deposits: self.op_pool.get_deposits(&*self.state.read(), &self.spec),
|
||||||
|
voluntary_exits: self
|
||||||
|
.op_pool
|
||||||
|
.get_voluntary_exits(&*self.state.read(), &self.spec),
|
||||||
|
transfers: self.op_pool.get_transfers(&*self.state.read(), &self.spec),
|
||||||
},
|
},
|
||||||
};
|
};
|
||||||
|
|
||||||
trace!("BeaconChain::produce_block: updating state for new block.",);
|
debug!(
|
||||||
|
"Produced block with {} attestations, updating state.",
|
||||||
|
block.body.attestations.len()
|
||||||
|
);
|
||||||
|
|
||||||
per_block_processing_without_verifying_block_signature(&mut state, &block, &self.spec)?;
|
per_block_processing_without_verifying_block_signature(&mut state, &block, &self.spec)?;
|
||||||
|
|
||||||
@ -951,6 +737,11 @@ where
|
|||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Returns `true` if the given block root has not been processed.
|
||||||
|
pub fn is_new_block_root(&self, beacon_block_root: &Hash256) -> Result<bool, Error> {
|
||||||
|
Ok(!self.block_store.exists(beacon_block_root)?)
|
||||||
|
}
|
||||||
|
|
||||||
/// Dumps the entire canonical chain, from the head to genesis to a vector for analysis.
|
/// Dumps the entire canonical chain, from the head to genesis to a vector for analysis.
|
||||||
///
|
///
|
||||||
/// This could be a very expensive operation and should only be done in testing/analysis
|
/// This could be a very expensive operation and should only be done in testing/analysis
|
||||||
|
@ -47,6 +47,8 @@ test_cases:
|
|||||||
states:
|
states:
|
||||||
- slot: 63
|
- slot: 63
|
||||||
num_validators: 1003
|
num_validators: 1003
|
||||||
|
num_previous_epoch_attestations: 0
|
||||||
|
num_current_epoch_attestations: 10
|
||||||
slashed_validators: [11, 12, 13, 14, 42]
|
slashed_validators: [11, 12, 13, 14, 42]
|
||||||
exited_validators: []
|
exited_validators: []
|
||||||
exit_initiated_validators: [50]
|
exit_initiated_validators: [50]
|
||||||
|
@ -10,8 +10,6 @@ use log::debug;
|
|||||||
use rayon::prelude::*;
|
use rayon::prelude::*;
|
||||||
use slot_clock::TestingSlotClock;
|
use slot_clock::TestingSlotClock;
|
||||||
use ssz::TreeHash;
|
use ssz::TreeHash;
|
||||||
use std::collections::HashSet;
|
|
||||||
use std::iter::FromIterator;
|
|
||||||
use std::sync::Arc;
|
use std::sync::Arc;
|
||||||
use types::{test_utils::TestingBeaconStateBuilder, *};
|
use types::{test_utils::TestingBeaconStateBuilder, *};
|
||||||
|
|
||||||
@ -137,51 +135,64 @@ impl BeaconChainHarness {
|
|||||||
slot
|
slot
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Gather the `FreeAttestation`s from the valiators.
|
pub fn gather_attesations(&mut self) -> Vec<Attestation> {
|
||||||
///
|
|
||||||
/// Note: validators will only produce attestations _once per slot_. So, if you call this twice
|
|
||||||
/// you'll only get attestations on the first run.
|
|
||||||
pub fn gather_free_attesations(&mut self) -> Vec<FreeAttestation> {
|
|
||||||
let present_slot = self.beacon_chain.present_slot();
|
let present_slot = self.beacon_chain.present_slot();
|
||||||
|
let state = self.beacon_chain.state.read();
|
||||||
|
|
||||||
let attesting_validators = self
|
let mut attestations = vec![];
|
||||||
.beacon_chain
|
|
||||||
.state
|
for committee in state
|
||||||
.read()
|
|
||||||
.get_crosslink_committees_at_slot(present_slot, &self.spec)
|
.get_crosslink_committees_at_slot(present_slot, &self.spec)
|
||||||
.unwrap()
|
.unwrap()
|
||||||
.iter()
|
{
|
||||||
.fold(vec![], |mut acc, c| {
|
for &validator in &committee.committee {
|
||||||
acc.append(&mut c.committee.clone());
|
let duties = state
|
||||||
acc
|
.get_attestation_duties(validator, &self.spec)
|
||||||
});
|
.unwrap()
|
||||||
let attesting_validators: HashSet<usize> =
|
.expect("Attesting validators by definition have duties");
|
||||||
HashSet::from_iter(attesting_validators.iter().cloned());
|
|
||||||
|
|
||||||
let free_attestations: Vec<FreeAttestation> = self
|
// Obtain `AttestationData` from the beacon chain.
|
||||||
.validators
|
let data = self
|
||||||
.par_iter_mut()
|
.beacon_chain
|
||||||
.enumerate()
|
.produce_attestation_data(duties.shard)
|
||||||
.filter_map(|(i, validator)| {
|
.unwrap();
|
||||||
if attesting_validators.contains(&i) {
|
|
||||||
// Advance the validator slot.
|
|
||||||
validator.set_slot(present_slot);
|
|
||||||
|
|
||||||
// Prompt the validator to produce an attestation (if required).
|
// Produce an aggregate signature with a single signature.
|
||||||
validator.produce_free_attestation().ok()
|
let aggregate_signature = {
|
||||||
} else {
|
let message = AttestationDataAndCustodyBit {
|
||||||
None
|
data: data.clone(),
|
||||||
}
|
custody_bit: false,
|
||||||
})
|
}
|
||||||
.collect();
|
.hash_tree_root();
|
||||||
|
let domain = self.spec.get_domain(
|
||||||
|
state.slot.epoch(self.spec.slots_per_epoch),
|
||||||
|
Domain::Attestation,
|
||||||
|
&state.fork,
|
||||||
|
);
|
||||||
|
let sig =
|
||||||
|
Signature::new(&message, domain, &self.validators[validator].keypair.sk);
|
||||||
|
|
||||||
debug!(
|
let mut agg_sig = AggregateSignature::new();
|
||||||
"Gathered {} FreeAttestations for slot {}.",
|
agg_sig.add(&sig);
|
||||||
free_attestations.len(),
|
|
||||||
present_slot
|
|
||||||
);
|
|
||||||
|
|
||||||
free_attestations
|
agg_sig
|
||||||
|
};
|
||||||
|
|
||||||
|
let mut aggregation_bitfield = Bitfield::with_capacity(duties.committee_len);
|
||||||
|
let custody_bitfield = Bitfield::with_capacity(duties.committee_len);
|
||||||
|
|
||||||
|
aggregation_bitfield.set(duties.committee_index, true);
|
||||||
|
|
||||||
|
attestations.push(Attestation {
|
||||||
|
aggregation_bitfield,
|
||||||
|
data,
|
||||||
|
custody_bitfield,
|
||||||
|
aggregate_signature,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
attestations
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Get the block from the proposer for the slot.
|
/// Get the block from the proposer for the slot.
|
||||||
@ -200,7 +211,9 @@ impl BeaconChainHarness {
|
|||||||
|
|
||||||
// Ensure the validators slot clock is accurate.
|
// Ensure the validators slot clock is accurate.
|
||||||
self.validators[proposer].set_slot(present_slot);
|
self.validators[proposer].set_slot(present_slot);
|
||||||
self.validators[proposer].produce_block().unwrap()
|
let block = self.validators[proposer].produce_block().unwrap();
|
||||||
|
|
||||||
|
block
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Advances the chain with a BeaconBlock and attestations from all validators.
|
/// Advances the chain with a BeaconBlock and attestations from all validators.
|
||||||
@ -219,20 +232,23 @@ impl BeaconChainHarness {
|
|||||||
};
|
};
|
||||||
debug!("...block processed by BeaconChain.");
|
debug!("...block processed by BeaconChain.");
|
||||||
|
|
||||||
debug!("Producing free attestations...");
|
debug!("Producing attestations...");
|
||||||
|
|
||||||
// Produce new attestations.
|
// Produce new attestations.
|
||||||
let free_attestations = self.gather_free_attesations();
|
let attestations = self.gather_attesations();
|
||||||
|
|
||||||
debug!("Processing free attestations...");
|
debug!("Processing {} attestations...", attestations.len());
|
||||||
|
|
||||||
free_attestations.par_iter().for_each(|free_attestation| {
|
attestations
|
||||||
self.beacon_chain
|
.par_iter()
|
||||||
.process_free_attestation(free_attestation.clone())
|
.enumerate()
|
||||||
.unwrap();
|
.for_each(|(i, attestation)| {
|
||||||
});
|
self.beacon_chain
|
||||||
|
.process_attestation(attestation.clone())
|
||||||
|
.expect(&format!("Attestation {} invalid: {:?}", i, attestation));
|
||||||
|
});
|
||||||
|
|
||||||
debug!("Free attestations processed.");
|
debug!("Attestations processed.");
|
||||||
|
|
||||||
block
|
block
|
||||||
}
|
}
|
||||||
@ -285,7 +301,7 @@ impl BeaconChainHarness {
|
|||||||
/// If a new `ValidatorHarness` was created, the validator should become fully operational as
|
/// If a new `ValidatorHarness` was created, the validator should become fully operational as
|
||||||
/// if the validator were created during `BeaconChainHarness` instantiation.
|
/// if the validator were created during `BeaconChainHarness` instantiation.
|
||||||
pub fn add_deposit(&mut self, deposit: Deposit, keypair: Option<Keypair>) {
|
pub fn add_deposit(&mut self, deposit: Deposit, keypair: Option<Keypair>) {
|
||||||
self.beacon_chain.receive_deposit_for_inclusion(deposit);
|
self.beacon_chain.process_deposit(deposit).unwrap();
|
||||||
|
|
||||||
// If a keypair is present, add a new `ValidatorHarness` to the rig.
|
// If a keypair is present, add a new `ValidatorHarness` to the rig.
|
||||||
if let Some(keypair) = keypair {
|
if let Some(keypair) = keypair {
|
||||||
@ -301,24 +317,26 @@ impl BeaconChainHarness {
|
|||||||
/// will stop receiving duties from the beacon chain and just do nothing when prompted to
|
/// will stop receiving duties from the beacon chain and just do nothing when prompted to
|
||||||
/// produce/attest.
|
/// produce/attest.
|
||||||
pub fn add_exit(&mut self, exit: VoluntaryExit) {
|
pub fn add_exit(&mut self, exit: VoluntaryExit) {
|
||||||
self.beacon_chain.receive_exit_for_inclusion(exit);
|
self.beacon_chain.process_voluntary_exit(exit).unwrap();
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Submit an transfer to the `BeaconChain` for inclusion in some block.
|
/// Submit an transfer to the `BeaconChain` for inclusion in some block.
|
||||||
pub fn add_transfer(&mut self, transfer: Transfer) {
|
pub fn add_transfer(&mut self, transfer: Transfer) {
|
||||||
self.beacon_chain.receive_transfer_for_inclusion(transfer);
|
self.beacon_chain.process_transfer(transfer).unwrap();
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Submit a proposer slashing to the `BeaconChain` for inclusion in some block.
|
/// Submit a proposer slashing to the `BeaconChain` for inclusion in some block.
|
||||||
pub fn add_proposer_slashing(&mut self, proposer_slashing: ProposerSlashing) {
|
pub fn add_proposer_slashing(&mut self, proposer_slashing: ProposerSlashing) {
|
||||||
self.beacon_chain
|
self.beacon_chain
|
||||||
.receive_proposer_slashing_for_inclusion(proposer_slashing);
|
.process_proposer_slashing(proposer_slashing)
|
||||||
|
.unwrap();
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Submit an attester slashing to the `BeaconChain` for inclusion in some block.
|
/// Submit an attester slashing to the `BeaconChain` for inclusion in some block.
|
||||||
pub fn add_attester_slashing(&mut self, attester_slashing: AttesterSlashing) {
|
pub fn add_attester_slashing(&mut self, attester_slashing: AttesterSlashing) {
|
||||||
self.beacon_chain
|
self.beacon_chain
|
||||||
.receive_attester_slashing_for_inclusion(attester_slashing);
|
.process_attester_slashing(attester_slashing)
|
||||||
|
.unwrap();
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Executes the fork choice rule on the `BeaconChain`, selecting a new canonical head.
|
/// Executes the fork choice rule on the `BeaconChain`, selecting a new canonical head.
|
||||||
|
@ -16,6 +16,10 @@ pub struct StateCheck {
|
|||||||
pub slot: Slot,
|
pub slot: Slot,
|
||||||
/// Checked against `beacon_state.validator_registry.len()`.
|
/// Checked against `beacon_state.validator_registry.len()`.
|
||||||
pub num_validators: Option<usize>,
|
pub num_validators: Option<usize>,
|
||||||
|
/// The number of pending attestations from the previous epoch that should be in the state.
|
||||||
|
pub num_previous_epoch_attestations: Option<usize>,
|
||||||
|
/// The number of pending attestations from the current epoch that should be in the state.
|
||||||
|
pub num_current_epoch_attestations: Option<usize>,
|
||||||
/// A list of validator indices which have been penalized. Must be in ascending order.
|
/// A list of validator indices which have been penalized. Must be in ascending order.
|
||||||
pub slashed_validators: Option<Vec<u64>>,
|
pub slashed_validators: Option<Vec<u64>>,
|
||||||
/// A list of validator indices which have been fully exited. Must be in ascending order.
|
/// A list of validator indices which have been fully exited. Must be in ascending order.
|
||||||
@ -34,6 +38,8 @@ impl StateCheck {
|
|||||||
Self {
|
Self {
|
||||||
slot: Slot::from(as_u64(&yaml, "slot").expect("State must specify slot")),
|
slot: Slot::from(as_u64(&yaml, "slot").expect("State must specify slot")),
|
||||||
num_validators: as_usize(&yaml, "num_validators"),
|
num_validators: as_usize(&yaml, "num_validators"),
|
||||||
|
num_previous_epoch_attestations: as_usize(&yaml, "num_previous_epoch_attestations"),
|
||||||
|
num_current_epoch_attestations: as_usize(&yaml, "num_current_epoch_attestations"),
|
||||||
slashed_validators: as_vec_u64(&yaml, "slashed_validators"),
|
slashed_validators: as_vec_u64(&yaml, "slashed_validators"),
|
||||||
exited_validators: as_vec_u64(&yaml, "exited_validators"),
|
exited_validators: as_vec_u64(&yaml, "exited_validators"),
|
||||||
exit_initiated_validators: as_vec_u64(&yaml, "exit_initiated_validators"),
|
exit_initiated_validators: as_vec_u64(&yaml, "exit_initiated_validators"),
|
||||||
@ -58,6 +64,7 @@ impl StateCheck {
|
|||||||
"State slot is invalid."
|
"State slot is invalid."
|
||||||
);
|
);
|
||||||
|
|
||||||
|
// Check the validator count
|
||||||
if let Some(num_validators) = self.num_validators {
|
if let Some(num_validators) = self.num_validators {
|
||||||
assert_eq!(
|
assert_eq!(
|
||||||
state.validator_registry.len(),
|
state.validator_registry.len(),
|
||||||
@ -67,6 +74,26 @@ impl StateCheck {
|
|||||||
info!("OK: num_validators = {}.", num_validators);
|
info!("OK: num_validators = {}.", num_validators);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Check the previous epoch attestations
|
||||||
|
if let Some(n) = self.num_previous_epoch_attestations {
|
||||||
|
assert_eq!(
|
||||||
|
state.previous_epoch_attestations.len(),
|
||||||
|
n,
|
||||||
|
"previous epoch attestations count != expected."
|
||||||
|
);
|
||||||
|
info!("OK: num_previous_epoch_attestations = {}.", n);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Check the current epoch attestations
|
||||||
|
if let Some(n) = self.num_current_epoch_attestations {
|
||||||
|
assert_eq!(
|
||||||
|
state.current_epoch_attestations.len(),
|
||||||
|
n,
|
||||||
|
"current epoch attestations count != expected."
|
||||||
|
);
|
||||||
|
info!("OK: num_current_epoch_attestations = {}.", n);
|
||||||
|
}
|
||||||
|
|
||||||
// Check for slashed validators.
|
// Check for slashed validators.
|
||||||
if let Some(ref slashed_validators) = self.slashed_validators {
|
if let Some(ref slashed_validators) = self.slashed_validators {
|
||||||
let actually_slashed_validators: Vec<u64> = state
|
let actually_slashed_validators: Vec<u64> = state
|
||||||
|
@ -55,7 +55,7 @@ impl<T: ClientDB, U: SlotClock, F: ForkChoice> AttesterBeaconNode for DirectBeac
|
|||||||
_slot: Slot,
|
_slot: Slot,
|
||||||
shard: u64,
|
shard: u64,
|
||||||
) -> Result<Option<AttestationData>, NodeError> {
|
) -> Result<Option<AttestationData>, NodeError> {
|
||||||
match self.beacon_chain.produce_attestation(shard) {
|
match self.beacon_chain.produce_attestation_data(shard) {
|
||||||
Ok(attestation_data) => Ok(Some(attestation_data)),
|
Ok(attestation_data) => Ok(Some(attestation_data)),
|
||||||
Err(e) => Err(NodeError::RemoteFailure(format!("{:?}", e))),
|
Err(e) => Err(NodeError::RemoteFailure(format!("{:?}", e))),
|
||||||
}
|
}
|
||||||
|
13
eth2/operation_pool/Cargo.toml
Normal file
13
eth2/operation_pool/Cargo.toml
Normal file
@ -0,0 +1,13 @@
|
|||||||
|
[package]
|
||||||
|
name = "operation_pool"
|
||||||
|
version = "0.1.0"
|
||||||
|
authors = ["Michael Sproul <michael@sigmaprime.io>"]
|
||||||
|
edition = "2018"
|
||||||
|
|
||||||
|
[dependencies]
|
||||||
|
int_to_bytes = { path = "../utils/int_to_bytes" }
|
||||||
|
itertools = "0.8"
|
||||||
|
parking_lot = "0.7"
|
||||||
|
types = { path = "../types" }
|
||||||
|
state_processing = { path = "../state_processing" }
|
||||||
|
ssz = { path = "../utils/ssz" }
|
952
eth2/operation_pool/src/lib.rs
Normal file
952
eth2/operation_pool/src/lib.rs
Normal file
@ -0,0 +1,952 @@
|
|||||||
|
use int_to_bytes::int_to_bytes8;
|
||||||
|
use itertools::Itertools;
|
||||||
|
use parking_lot::RwLock;
|
||||||
|
use ssz::ssz_encode;
|
||||||
|
use state_processing::per_block_processing::errors::{
|
||||||
|
AttestationValidationError, AttesterSlashingValidationError, DepositValidationError,
|
||||||
|
ExitValidationError, ProposerSlashingValidationError, TransferValidationError,
|
||||||
|
};
|
||||||
|
use state_processing::per_block_processing::{
|
||||||
|
gather_attester_slashing_indices_modular, validate_attestation,
|
||||||
|
validate_attestation_time_independent_only, verify_attester_slashing, verify_deposit,
|
||||||
|
verify_exit, verify_exit_time_independent_only, verify_proposer_slashing, verify_transfer,
|
||||||
|
verify_transfer_time_independent_only,
|
||||||
|
};
|
||||||
|
use std::collections::{btree_map::Entry, hash_map, BTreeMap, HashMap, HashSet};
|
||||||
|
use types::chain_spec::Domain;
|
||||||
|
use types::{
|
||||||
|
Attestation, AttestationData, AttesterSlashing, BeaconState, ChainSpec, Deposit, Epoch,
|
||||||
|
ProposerSlashing, Transfer, Validator, VoluntaryExit,
|
||||||
|
};
|
||||||
|
|
||||||
|
#[cfg(test)]
|
||||||
|
const VERIFY_DEPOSIT_PROOFS: bool = false;
|
||||||
|
#[cfg(not(test))]
|
||||||
|
const VERIFY_DEPOSIT_PROOFS: bool = false; // TODO: enable this
|
||||||
|
|
||||||
|
#[derive(Default)]
|
||||||
|
pub struct OperationPool {
|
||||||
|
/// Map from attestation ID (see below) to vectors of attestations.
|
||||||
|
attestations: RwLock<HashMap<AttestationId, Vec<Attestation>>>,
|
||||||
|
/// Map from deposit index to deposit data.
|
||||||
|
// NOTE: We assume that there is only one deposit per index
|
||||||
|
// because the Eth1 data is updated (at most) once per epoch,
|
||||||
|
// and the spec doesn't seem to accomodate for re-orgs on a time-frame
|
||||||
|
// longer than an epoch
|
||||||
|
deposits: RwLock<BTreeMap<u64, Deposit>>,
|
||||||
|
/// Map from two attestation IDs to a slashing for those IDs.
|
||||||
|
attester_slashings: RwLock<HashMap<(AttestationId, AttestationId), AttesterSlashing>>,
|
||||||
|
/// Map from proposer index to slashing.
|
||||||
|
proposer_slashings: RwLock<HashMap<u64, ProposerSlashing>>,
|
||||||
|
/// Map from exiting validator to their exit data.
|
||||||
|
voluntary_exits: RwLock<HashMap<u64, VoluntaryExit>>,
|
||||||
|
/// Set of transfers.
|
||||||
|
transfers: RwLock<HashSet<Transfer>>,
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Serialized `AttestationData` augmented with a domain to encode the fork info.
|
||||||
|
#[derive(PartialEq, Eq, Clone, Hash, Debug)]
|
||||||
|
struct AttestationId(Vec<u8>);
|
||||||
|
|
||||||
|
/// Number of domain bytes that the end of an attestation ID is padded with.
|
||||||
|
const DOMAIN_BYTES_LEN: usize = 8;
|
||||||
|
|
||||||
|
impl AttestationId {
|
||||||
|
fn from_data(attestation: &AttestationData, state: &BeaconState, spec: &ChainSpec) -> Self {
|
||||||
|
let mut bytes = ssz_encode(attestation);
|
||||||
|
let epoch = attestation.slot.epoch(spec.slots_per_epoch);
|
||||||
|
bytes.extend_from_slice(&AttestationId::compute_domain_bytes(epoch, state, spec));
|
||||||
|
AttestationId(bytes)
|
||||||
|
}
|
||||||
|
|
||||||
|
fn compute_domain_bytes(epoch: Epoch, state: &BeaconState, spec: &ChainSpec) -> Vec<u8> {
|
||||||
|
int_to_bytes8(spec.get_domain(epoch, Domain::Attestation, &state.fork))
|
||||||
|
}
|
||||||
|
|
||||||
|
fn domain_bytes_match(&self, domain_bytes: &[u8]) -> bool {
|
||||||
|
&self.0[self.0.len() - DOMAIN_BYTES_LEN..] == domain_bytes
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Compute a fitness score for an attestation.
|
||||||
|
///
|
||||||
|
/// The score is calculated by determining the number of *new* attestations that
|
||||||
|
/// the aggregate attestation introduces, and is proportional to the size of the reward we will
|
||||||
|
/// receive for including it in a block.
|
||||||
|
// TODO: this could be optimised with a map from validator index to whether that validator has
|
||||||
|
// attested in each of the current and previous epochs. Currently quadractic in number of validators.
|
||||||
|
fn attestation_score(attestation: &Attestation, state: &BeaconState, spec: &ChainSpec) -> usize {
|
||||||
|
// Bitfield of validators whose attestations are new/fresh.
|
||||||
|
let mut new_validators = attestation.aggregation_bitfield.clone();
|
||||||
|
|
||||||
|
let attestation_epoch = attestation.data.slot.epoch(spec.slots_per_epoch);
|
||||||
|
|
||||||
|
let state_attestations = if attestation_epoch == state.current_epoch(spec) {
|
||||||
|
&state.current_epoch_attestations
|
||||||
|
} else if attestation_epoch == state.previous_epoch(spec) {
|
||||||
|
&state.previous_epoch_attestations
|
||||||
|
} else {
|
||||||
|
return 0;
|
||||||
|
};
|
||||||
|
|
||||||
|
state_attestations
|
||||||
|
.iter()
|
||||||
|
// In a single epoch, an attester should only be attesting for one shard.
|
||||||
|
// TODO: we avoid including slashable attestations in the state here,
|
||||||
|
// but maybe we should do something else with them (like construct slashings).
|
||||||
|
.filter(|current_attestation| current_attestation.data.shard == attestation.data.shard)
|
||||||
|
.for_each(|current_attestation| {
|
||||||
|
// Remove the validators who have signed the existing attestation (they are not new)
|
||||||
|
new_validators.difference_inplace(¤t_attestation.aggregation_bitfield);
|
||||||
|
});
|
||||||
|
|
||||||
|
new_validators.num_set_bits()
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Debug, PartialEq, Clone)]
|
||||||
|
pub enum DepositInsertStatus {
|
||||||
|
/// The deposit was not already in the pool.
|
||||||
|
Fresh,
|
||||||
|
/// The deposit already existed in the pool.
|
||||||
|
Duplicate,
|
||||||
|
/// The deposit conflicted with an existing deposit, which was replaced.
|
||||||
|
Replaced(Box<Deposit>),
|
||||||
|
}
|
||||||
|
|
||||||
|
impl OperationPool {
|
||||||
|
/// Create a new operation pool.
|
||||||
|
pub fn new() -> Self {
|
||||||
|
Self::default()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Insert an attestation into the pool, aggregating it with existing attestations if possible.
|
||||||
|
pub fn insert_attestation(
|
||||||
|
&self,
|
||||||
|
attestation: Attestation,
|
||||||
|
state: &BeaconState,
|
||||||
|
spec: &ChainSpec,
|
||||||
|
) -> Result<(), AttestationValidationError> {
|
||||||
|
// Check that attestation signatures are valid.
|
||||||
|
validate_attestation_time_independent_only(state, &attestation, spec)?;
|
||||||
|
|
||||||
|
let id = AttestationId::from_data(&attestation.data, state, spec);
|
||||||
|
|
||||||
|
// Take a write lock on the attestations map.
|
||||||
|
let mut attestations = self.attestations.write();
|
||||||
|
|
||||||
|
let existing_attestations = match attestations.entry(id) {
|
||||||
|
hash_map::Entry::Vacant(entry) => {
|
||||||
|
entry.insert(vec![attestation]);
|
||||||
|
return Ok(());
|
||||||
|
}
|
||||||
|
hash_map::Entry::Occupied(entry) => entry.into_mut(),
|
||||||
|
};
|
||||||
|
|
||||||
|
let mut aggregated = false;
|
||||||
|
for existing_attestation in existing_attestations.iter_mut() {
|
||||||
|
if existing_attestation.signers_disjoint_from(&attestation) {
|
||||||
|
existing_attestation.aggregate(&attestation);
|
||||||
|
aggregated = true;
|
||||||
|
} else if *existing_attestation == attestation {
|
||||||
|
aggregated = true;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if !aggregated {
|
||||||
|
existing_attestations.push(attestation);
|
||||||
|
}
|
||||||
|
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Total number of attestations in the pool, including attestations for the same data.
|
||||||
|
pub fn num_attestations(&self) -> usize {
|
||||||
|
self.attestations
|
||||||
|
.read()
|
||||||
|
.values()
|
||||||
|
.map(|atts| atts.len())
|
||||||
|
.sum()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Get a list of attestations for inclusion in a block.
|
||||||
|
pub fn get_attestations(&self, state: &BeaconState, spec: &ChainSpec) -> Vec<Attestation> {
|
||||||
|
// Attestations for the current fork, which may be from the current or previous epoch.
|
||||||
|
let prev_epoch = state.previous_epoch(spec);
|
||||||
|
let current_epoch = state.current_epoch(spec);
|
||||||
|
let prev_domain_bytes = AttestationId::compute_domain_bytes(prev_epoch, state, spec);
|
||||||
|
let curr_domain_bytes = AttestationId::compute_domain_bytes(current_epoch, state, spec);
|
||||||
|
self.attestations
|
||||||
|
.read()
|
||||||
|
.iter()
|
||||||
|
.filter(|(key, _)| {
|
||||||
|
key.domain_bytes_match(&prev_domain_bytes)
|
||||||
|
|| key.domain_bytes_match(&curr_domain_bytes)
|
||||||
|
})
|
||||||
|
.flat_map(|(_, attestations)| attestations)
|
||||||
|
// That are valid...
|
||||||
|
.filter(|attestation| validate_attestation(state, attestation, spec).is_ok())
|
||||||
|
// Scored by the number of new attestations they introduce (descending)
|
||||||
|
// TODO: need to consider attestations introduced in THIS block
|
||||||
|
.map(|att| (att, attestation_score(att, state, spec)))
|
||||||
|
// Don't include any useless attestations (score 0)
|
||||||
|
.filter(|&(_, score)| score != 0)
|
||||||
|
.sorted_by_key(|&(_, score)| std::cmp::Reverse(score))
|
||||||
|
// Limited to the maximum number of attestations per block
|
||||||
|
.take(spec.max_attestations as usize)
|
||||||
|
.map(|(att, _)| att)
|
||||||
|
.cloned()
|
||||||
|
.collect()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Remove attestations which are too old to be included in a block.
|
||||||
|
// TODO: we could probably prune other attestations here:
|
||||||
|
// - ones that are completely covered by attestations included in the state
|
||||||
|
// - maybe ones invalidated by the confirmation of one fork over another
|
||||||
|
pub fn prune_attestations(&self, finalized_state: &BeaconState, spec: &ChainSpec) {
|
||||||
|
self.attestations.write().retain(|_, attestations| {
|
||||||
|
// All the attestations in this bucket have the same data, so we only need to
|
||||||
|
// check the first one.
|
||||||
|
attestations.first().map_or(false, |att| {
|
||||||
|
finalized_state.slot < att.data.slot + spec.slots_per_epoch
|
||||||
|
})
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Add a deposit to the pool.
|
||||||
|
///
|
||||||
|
/// No two distinct deposits should be added with the same index.
|
||||||
|
pub fn insert_deposit(
|
||||||
|
&self,
|
||||||
|
deposit: Deposit,
|
||||||
|
state: &BeaconState,
|
||||||
|
spec: &ChainSpec,
|
||||||
|
) -> Result<DepositInsertStatus, DepositValidationError> {
|
||||||
|
use DepositInsertStatus::*;
|
||||||
|
|
||||||
|
match self.deposits.write().entry(deposit.index) {
|
||||||
|
Entry::Vacant(entry) => {
|
||||||
|
verify_deposit(state, &deposit, VERIFY_DEPOSIT_PROOFS, spec)?;
|
||||||
|
entry.insert(deposit);
|
||||||
|
Ok(Fresh)
|
||||||
|
}
|
||||||
|
Entry::Occupied(mut entry) => {
|
||||||
|
if entry.get() == &deposit {
|
||||||
|
Ok(Duplicate)
|
||||||
|
} else {
|
||||||
|
verify_deposit(state, &deposit, VERIFY_DEPOSIT_PROOFS, spec)?;
|
||||||
|
Ok(Replaced(Box::new(entry.insert(deposit))))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Get an ordered list of deposits for inclusion in a block.
|
||||||
|
///
|
||||||
|
/// Take at most the maximum number of deposits, beginning from the current deposit index.
|
||||||
|
pub fn get_deposits(&self, state: &BeaconState, spec: &ChainSpec) -> Vec<Deposit> {
|
||||||
|
let start_idx = state.deposit_index;
|
||||||
|
(start_idx..start_idx + spec.max_deposits)
|
||||||
|
.map(|idx| self.deposits.read().get(&idx).cloned())
|
||||||
|
.take_while(Option::is_some)
|
||||||
|
.flatten()
|
||||||
|
.collect()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Remove all deposits with index less than the deposit index of the latest finalised block.
|
||||||
|
pub fn prune_deposits(&self, state: &BeaconState) -> BTreeMap<u64, Deposit> {
|
||||||
|
let deposits_keep = self.deposits.write().split_off(&state.deposit_index);
|
||||||
|
std::mem::replace(&mut self.deposits.write(), deposits_keep)
|
||||||
|
}
|
||||||
|
|
||||||
|
/// The number of deposits stored in the pool.
|
||||||
|
pub fn num_deposits(&self) -> usize {
|
||||||
|
self.deposits.read().len()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Insert a proposer slashing into the pool.
|
||||||
|
pub fn insert_proposer_slashing(
|
||||||
|
&self,
|
||||||
|
slashing: ProposerSlashing,
|
||||||
|
state: &BeaconState,
|
||||||
|
spec: &ChainSpec,
|
||||||
|
) -> Result<(), ProposerSlashingValidationError> {
|
||||||
|
// TODO: should maybe insert anyway if the proposer is unknown in the validator index,
|
||||||
|
// because they could *become* known later
|
||||||
|
verify_proposer_slashing(&slashing, state, spec)?;
|
||||||
|
self.proposer_slashings
|
||||||
|
.write()
|
||||||
|
.insert(slashing.proposer_index, slashing);
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Compute the tuple ID that is used to identify an attester slashing.
|
||||||
|
///
|
||||||
|
/// Depends on the fork field of the state, but not on the state's epoch.
|
||||||
|
fn attester_slashing_id(
|
||||||
|
slashing: &AttesterSlashing,
|
||||||
|
state: &BeaconState,
|
||||||
|
spec: &ChainSpec,
|
||||||
|
) -> (AttestationId, AttestationId) {
|
||||||
|
(
|
||||||
|
AttestationId::from_data(&slashing.slashable_attestation_1.data, state, spec),
|
||||||
|
AttestationId::from_data(&slashing.slashable_attestation_2.data, state, spec),
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Insert an attester slashing into the pool.
|
||||||
|
pub fn insert_attester_slashing(
|
||||||
|
&self,
|
||||||
|
slashing: AttesterSlashing,
|
||||||
|
state: &BeaconState,
|
||||||
|
spec: &ChainSpec,
|
||||||
|
) -> Result<(), AttesterSlashingValidationError> {
|
||||||
|
verify_attester_slashing(state, &slashing, true, spec)?;
|
||||||
|
let id = Self::attester_slashing_id(&slashing, state, spec);
|
||||||
|
self.attester_slashings.write().insert(id, slashing);
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Get proposer and attester slashings for inclusion in a block.
|
||||||
|
///
|
||||||
|
/// This function computes both types of slashings together, because
|
||||||
|
/// attester slashings may be invalidated by proposer slashings included
|
||||||
|
/// earlier in the block.
|
||||||
|
pub fn get_slashings(
|
||||||
|
&self,
|
||||||
|
state: &BeaconState,
|
||||||
|
spec: &ChainSpec,
|
||||||
|
) -> (Vec<ProposerSlashing>, Vec<AttesterSlashing>) {
|
||||||
|
let proposer_slashings = filter_limit_operations(
|
||||||
|
self.proposer_slashings.read().values(),
|
||||||
|
|slashing| {
|
||||||
|
state
|
||||||
|
.validator_registry
|
||||||
|
.get(slashing.proposer_index as usize)
|
||||||
|
.map_or(false, |validator| !validator.slashed)
|
||||||
|
},
|
||||||
|
spec.max_proposer_slashings,
|
||||||
|
);
|
||||||
|
|
||||||
|
// Set of validators to be slashed, so we don't attempt to construct invalid attester
|
||||||
|
// slashings.
|
||||||
|
let mut to_be_slashed = proposer_slashings
|
||||||
|
.iter()
|
||||||
|
.map(|s| s.proposer_index)
|
||||||
|
.collect::<HashSet<_>>();
|
||||||
|
|
||||||
|
let attester_slashings = self
|
||||||
|
.attester_slashings
|
||||||
|
.read()
|
||||||
|
.iter()
|
||||||
|
.filter(|(id, slashing)| {
|
||||||
|
// Check the fork.
|
||||||
|
Self::attester_slashing_id(slashing, state, spec) == **id
|
||||||
|
})
|
||||||
|
.filter(|(_, slashing)| {
|
||||||
|
// Take all slashings that will slash 1 or more validators.
|
||||||
|
let slashed_validators = gather_attester_slashing_indices_modular(
|
||||||
|
state,
|
||||||
|
slashing,
|
||||||
|
|index, validator| validator.slashed || to_be_slashed.contains(&index),
|
||||||
|
spec,
|
||||||
|
);
|
||||||
|
|
||||||
|
// Extend the `to_be_slashed` set so subsequent iterations don't try to include
|
||||||
|
// useless slashings.
|
||||||
|
if let Ok(validators) = slashed_validators {
|
||||||
|
to_be_slashed.extend(validators);
|
||||||
|
true
|
||||||
|
} else {
|
||||||
|
false
|
||||||
|
}
|
||||||
|
})
|
||||||
|
.take(spec.max_attester_slashings as usize)
|
||||||
|
.map(|(_, slashing)| slashing.clone())
|
||||||
|
.collect();
|
||||||
|
|
||||||
|
(proposer_slashings, attester_slashings)
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Prune proposer slashings for all slashed or withdrawn validators.
|
||||||
|
pub fn prune_proposer_slashings(&self, finalized_state: &BeaconState, spec: &ChainSpec) {
|
||||||
|
prune_validator_hash_map(
|
||||||
|
&mut self.proposer_slashings.write(),
|
||||||
|
|validator| {
|
||||||
|
validator.slashed
|
||||||
|
|| validator.is_withdrawable_at(finalized_state.current_epoch(spec))
|
||||||
|
},
|
||||||
|
finalized_state,
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Prune attester slashings for all slashed or withdrawn validators, or attestations on another
|
||||||
|
/// fork.
|
||||||
|
pub fn prune_attester_slashings(&self, finalized_state: &BeaconState, spec: &ChainSpec) {
|
||||||
|
self.attester_slashings.write().retain(|id, slashing| {
|
||||||
|
let fork_ok = &Self::attester_slashing_id(slashing, finalized_state, spec) == id;
|
||||||
|
let curr_epoch = finalized_state.current_epoch(spec);
|
||||||
|
let slashing_ok = gather_attester_slashing_indices_modular(
|
||||||
|
finalized_state,
|
||||||
|
slashing,
|
||||||
|
|_, validator| validator.slashed || validator.is_withdrawable_at(curr_epoch),
|
||||||
|
spec,
|
||||||
|
)
|
||||||
|
.is_ok();
|
||||||
|
fork_ok && slashing_ok
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Insert a voluntary exit, validating it almost-entirely (future exits are permitted).
|
||||||
|
pub fn insert_voluntary_exit(
|
||||||
|
&self,
|
||||||
|
exit: VoluntaryExit,
|
||||||
|
state: &BeaconState,
|
||||||
|
spec: &ChainSpec,
|
||||||
|
) -> Result<(), ExitValidationError> {
|
||||||
|
verify_exit_time_independent_only(state, &exit, spec)?;
|
||||||
|
self.voluntary_exits
|
||||||
|
.write()
|
||||||
|
.insert(exit.validator_index, exit);
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Get a list of voluntary exits for inclusion in a block.
|
||||||
|
pub fn get_voluntary_exits(&self, state: &BeaconState, spec: &ChainSpec) -> Vec<VoluntaryExit> {
|
||||||
|
filter_limit_operations(
|
||||||
|
self.voluntary_exits.read().values(),
|
||||||
|
|exit| verify_exit(state, exit, spec).is_ok(),
|
||||||
|
spec.max_voluntary_exits,
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Prune if validator has already exited at the last finalized state.
|
||||||
|
pub fn prune_voluntary_exits(&self, finalized_state: &BeaconState, spec: &ChainSpec) {
|
||||||
|
prune_validator_hash_map(
|
||||||
|
&mut self.voluntary_exits.write(),
|
||||||
|
|validator| validator.is_exited_at(finalized_state.current_epoch(spec)),
|
||||||
|
finalized_state,
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Insert a transfer into the pool, checking it for validity in the process.
|
||||||
|
pub fn insert_transfer(
|
||||||
|
&self,
|
||||||
|
transfer: Transfer,
|
||||||
|
state: &BeaconState,
|
||||||
|
spec: &ChainSpec,
|
||||||
|
) -> Result<(), TransferValidationError> {
|
||||||
|
// The signature of the transfer isn't hashed, but because we check
|
||||||
|
// it before we insert into the HashSet, we can't end up with duplicate
|
||||||
|
// transactions.
|
||||||
|
verify_transfer_time_independent_only(state, &transfer, spec)?;
|
||||||
|
self.transfers.write().insert(transfer);
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Get a list of transfers for inclusion in a block.
|
||||||
|
// TODO: improve the economic optimality of this function by accounting for
|
||||||
|
// dependencies between transfers in the same block e.g. A pays B, B pays C
|
||||||
|
pub fn get_transfers(&self, state: &BeaconState, spec: &ChainSpec) -> Vec<Transfer> {
|
||||||
|
self.transfers
|
||||||
|
.read()
|
||||||
|
.iter()
|
||||||
|
.filter(|transfer| verify_transfer(state, transfer, spec).is_ok())
|
||||||
|
.sorted_by_key(|transfer| std::cmp::Reverse(transfer.fee))
|
||||||
|
.take(spec.max_transfers as usize)
|
||||||
|
.cloned()
|
||||||
|
.collect()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Prune the set of transfers by removing all those whose slot has already passed.
|
||||||
|
pub fn prune_transfers(&self, finalized_state: &BeaconState) {
|
||||||
|
self.transfers
|
||||||
|
.write()
|
||||||
|
.retain(|transfer| transfer.slot > finalized_state.slot)
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Prune all types of transactions given the latest finalized state.
|
||||||
|
pub fn prune_all(&self, finalized_state: &BeaconState, spec: &ChainSpec) {
|
||||||
|
self.prune_attestations(finalized_state, spec);
|
||||||
|
self.prune_deposits(finalized_state);
|
||||||
|
self.prune_proposer_slashings(finalized_state, spec);
|
||||||
|
self.prune_attester_slashings(finalized_state, spec);
|
||||||
|
self.prune_voluntary_exits(finalized_state, spec);
|
||||||
|
self.prune_transfers(finalized_state);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Filter up to a maximum number of operations out of an iterator.
|
||||||
|
fn filter_limit_operations<'a, T: 'a, I, F>(operations: I, filter: F, limit: u64) -> Vec<T>
|
||||||
|
where
|
||||||
|
I: IntoIterator<Item = &'a T>,
|
||||||
|
F: Fn(&T) -> bool,
|
||||||
|
T: Clone,
|
||||||
|
{
|
||||||
|
operations
|
||||||
|
.into_iter()
|
||||||
|
.filter(|x| filter(*x))
|
||||||
|
.take(limit as usize)
|
||||||
|
.cloned()
|
||||||
|
.collect()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Remove all entries from the given hash map for which `prune_if` returns true.
|
||||||
|
///
|
||||||
|
/// The keys in the map should be validator indices, which will be looked up
|
||||||
|
/// in the state's validator registry and then passed to `prune_if`.
|
||||||
|
/// Entries for unknown validators will be kept.
|
||||||
|
fn prune_validator_hash_map<T, F>(
|
||||||
|
map: &mut HashMap<u64, T>,
|
||||||
|
prune_if: F,
|
||||||
|
finalized_state: &BeaconState,
|
||||||
|
) where
|
||||||
|
F: Fn(&Validator) -> bool,
|
||||||
|
{
|
||||||
|
map.retain(|&validator_index, _| {
|
||||||
|
finalized_state
|
||||||
|
.validator_registry
|
||||||
|
.get(validator_index as usize)
|
||||||
|
.map_or(true, |validator| !prune_if(validator))
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
#[cfg(test)]
|
||||||
|
mod tests {
|
||||||
|
use super::DepositInsertStatus::*;
|
||||||
|
use super::*;
|
||||||
|
use types::test_utils::*;
|
||||||
|
use types::*;
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn insert_deposit() {
|
||||||
|
let rng = &mut XorShiftRng::from_seed([42; 16]);
|
||||||
|
let (ref spec, ref state) = test_state(rng);
|
||||||
|
let op_pool = OperationPool::new();
|
||||||
|
let deposit1 = make_deposit(rng, state, spec);
|
||||||
|
let mut deposit2 = make_deposit(rng, state, spec);
|
||||||
|
deposit2.index = deposit1.index;
|
||||||
|
|
||||||
|
assert_eq!(
|
||||||
|
op_pool.insert_deposit(deposit1.clone(), state, spec),
|
||||||
|
Ok(Fresh)
|
||||||
|
);
|
||||||
|
assert_eq!(
|
||||||
|
op_pool.insert_deposit(deposit1.clone(), state, spec),
|
||||||
|
Ok(Duplicate)
|
||||||
|
);
|
||||||
|
assert_eq!(
|
||||||
|
op_pool.insert_deposit(deposit2, state, spec),
|
||||||
|
Ok(Replaced(Box::new(deposit1)))
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn get_deposits_max() {
|
||||||
|
let rng = &mut XorShiftRng::from_seed([42; 16]);
|
||||||
|
let (spec, mut state) = test_state(rng);
|
||||||
|
let op_pool = OperationPool::new();
|
||||||
|
let start = 10000;
|
||||||
|
let max_deposits = spec.max_deposits;
|
||||||
|
let extra = 5;
|
||||||
|
let offset = 1;
|
||||||
|
assert!(offset <= extra);
|
||||||
|
|
||||||
|
let deposits = dummy_deposits(rng, &state, &spec, start, max_deposits + extra);
|
||||||
|
|
||||||
|
for deposit in &deposits {
|
||||||
|
assert_eq!(
|
||||||
|
op_pool.insert_deposit(deposit.clone(), &state, &spec),
|
||||||
|
Ok(Fresh)
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
state.deposit_index = start + offset;
|
||||||
|
let deposits_for_block = op_pool.get_deposits(&state, &spec);
|
||||||
|
|
||||||
|
assert_eq!(deposits_for_block.len() as u64, max_deposits);
|
||||||
|
assert_eq!(
|
||||||
|
deposits_for_block[..],
|
||||||
|
deposits[offset as usize..(offset + max_deposits) as usize]
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn prune_deposits() {
|
||||||
|
let rng = &mut XorShiftRng::from_seed([42; 16]);
|
||||||
|
let (spec, state) = test_state(rng);
|
||||||
|
let op_pool = OperationPool::new();
|
||||||
|
|
||||||
|
let start1 = 100;
|
||||||
|
// test is super slow in debug mode if this parameter is too high
|
||||||
|
let count = 5;
|
||||||
|
let gap = 25;
|
||||||
|
let start2 = start1 + count + gap;
|
||||||
|
|
||||||
|
let deposits1 = dummy_deposits(rng, &state, &spec, start1, count);
|
||||||
|
let deposits2 = dummy_deposits(rng, &state, &spec, start2, count);
|
||||||
|
|
||||||
|
for d in deposits1.into_iter().chain(deposits2) {
|
||||||
|
assert!(op_pool.insert_deposit(d, &state, &spec).is_ok());
|
||||||
|
}
|
||||||
|
|
||||||
|
assert_eq!(op_pool.num_deposits(), 2 * count as usize);
|
||||||
|
|
||||||
|
let mut state = BeaconState::random_for_test(rng);
|
||||||
|
state.deposit_index = start1;
|
||||||
|
|
||||||
|
// Pruning the first bunch of deposits in batches of 5 should work.
|
||||||
|
let step = 5;
|
||||||
|
let mut pool_size = step + 2 * count as usize;
|
||||||
|
for i in (start1..=(start1 + count)).step_by(step) {
|
||||||
|
state.deposit_index = i;
|
||||||
|
op_pool.prune_deposits(&state);
|
||||||
|
pool_size -= step;
|
||||||
|
assert_eq!(op_pool.num_deposits(), pool_size);
|
||||||
|
}
|
||||||
|
assert_eq!(pool_size, count as usize);
|
||||||
|
// Pruning in the gap should do nothing.
|
||||||
|
for i in (start1 + count..start2).step_by(step) {
|
||||||
|
state.deposit_index = i;
|
||||||
|
op_pool.prune_deposits(&state);
|
||||||
|
assert_eq!(op_pool.num_deposits(), count as usize);
|
||||||
|
}
|
||||||
|
// Same again for the later deposits.
|
||||||
|
pool_size += step;
|
||||||
|
for i in (start2..=(start2 + count)).step_by(step) {
|
||||||
|
state.deposit_index = i;
|
||||||
|
op_pool.prune_deposits(&state);
|
||||||
|
pool_size -= step;
|
||||||
|
assert_eq!(op_pool.num_deposits(), pool_size);
|
||||||
|
}
|
||||||
|
assert_eq!(op_pool.num_deposits(), 0);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Create a random deposit (with a valid proof of posession)
|
||||||
|
fn make_deposit(rng: &mut XorShiftRng, state: &BeaconState, spec: &ChainSpec) -> Deposit {
|
||||||
|
let keypair = Keypair::random();
|
||||||
|
let mut deposit = Deposit::random_for_test(rng);
|
||||||
|
let mut deposit_input = DepositInput {
|
||||||
|
pubkey: keypair.pk.clone(),
|
||||||
|
withdrawal_credentials: Hash256::zero(),
|
||||||
|
proof_of_possession: Signature::empty_signature(),
|
||||||
|
};
|
||||||
|
deposit_input.proof_of_possession = deposit_input.create_proof_of_possession(
|
||||||
|
&keypair.sk,
|
||||||
|
state.slot.epoch(spec.slots_per_epoch),
|
||||||
|
&state.fork,
|
||||||
|
spec,
|
||||||
|
);
|
||||||
|
deposit.deposit_data.deposit_input = deposit_input;
|
||||||
|
deposit
|
||||||
|
}
|
||||||
|
|
||||||
|
// Create `count` dummy deposits with sequential deposit IDs beginning from `start`.
|
||||||
|
fn dummy_deposits(
|
||||||
|
rng: &mut XorShiftRng,
|
||||||
|
state: &BeaconState,
|
||||||
|
spec: &ChainSpec,
|
||||||
|
start: u64,
|
||||||
|
count: u64,
|
||||||
|
) -> Vec<Deposit> {
|
||||||
|
let proto_deposit = make_deposit(rng, state, spec);
|
||||||
|
(start..start + count)
|
||||||
|
.map(|index| {
|
||||||
|
let mut deposit = proto_deposit.clone();
|
||||||
|
deposit.index = index;
|
||||||
|
deposit
|
||||||
|
})
|
||||||
|
.collect()
|
||||||
|
}
|
||||||
|
|
||||||
|
fn test_state(rng: &mut XorShiftRng) -> (ChainSpec, BeaconState) {
|
||||||
|
let spec = ChainSpec::foundation();
|
||||||
|
let mut state = BeaconState::random_for_test(rng);
|
||||||
|
state.fork = Fork::genesis(&spec);
|
||||||
|
|
||||||
|
(spec, state)
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Create a signed attestation for use in tests.
|
||||||
|
/// Signed by all validators in `committee[signing_range]` and `committee[extra_signer]`.
|
||||||
|
fn signed_attestation<R: std::slice::SliceIndex<[usize], Output = [usize]>>(
|
||||||
|
committee: &CrosslinkCommittee,
|
||||||
|
keypairs: &[Keypair],
|
||||||
|
signing_range: R,
|
||||||
|
slot: Slot,
|
||||||
|
state: &BeaconState,
|
||||||
|
spec: &ChainSpec,
|
||||||
|
extra_signer: Option<usize>,
|
||||||
|
) -> Attestation {
|
||||||
|
let mut builder = TestingAttestationBuilder::new(
|
||||||
|
state,
|
||||||
|
&committee.committee,
|
||||||
|
slot,
|
||||||
|
committee.shard,
|
||||||
|
spec,
|
||||||
|
);
|
||||||
|
let signers = &committee.committee[signing_range];
|
||||||
|
let committee_keys = signers.iter().map(|&i| &keypairs[i].sk).collect::<Vec<_>>();
|
||||||
|
builder.sign(signers, &committee_keys, &state.fork, spec);
|
||||||
|
extra_signer.map(|c_idx| {
|
||||||
|
let validator_index = committee.committee[c_idx];
|
||||||
|
builder.sign(
|
||||||
|
&[validator_index],
|
||||||
|
&[&keypairs[validator_index].sk],
|
||||||
|
&state.fork,
|
||||||
|
spec,
|
||||||
|
)
|
||||||
|
});
|
||||||
|
builder.build()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Test state for attestation-related tests.
|
||||||
|
fn attestation_test_state(
|
||||||
|
spec: &ChainSpec,
|
||||||
|
num_committees: usize,
|
||||||
|
) -> (BeaconState, Vec<Keypair>) {
|
||||||
|
let num_validators =
|
||||||
|
num_committees * (spec.slots_per_epoch * spec.target_committee_size) as usize;
|
||||||
|
let mut state_builder =
|
||||||
|
TestingBeaconStateBuilder::from_default_keypairs_file_if_exists(num_validators, spec);
|
||||||
|
let slot_offset = 1000 * spec.slots_per_epoch + spec.slots_per_epoch / 2;
|
||||||
|
let slot = spec.genesis_slot + slot_offset;
|
||||||
|
state_builder.teleport_to_slot(slot, spec);
|
||||||
|
state_builder.build_caches(spec).unwrap();
|
||||||
|
state_builder.build()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Set the latest crosslink in the state to match the attestation.
|
||||||
|
fn fake_latest_crosslink(att: &Attestation, state: &mut BeaconState, spec: &ChainSpec) {
|
||||||
|
state.latest_crosslinks[att.data.shard as usize] = Crosslink {
|
||||||
|
crosslink_data_root: att.data.crosslink_data_root,
|
||||||
|
epoch: att.data.slot.epoch(spec.slots_per_epoch),
|
||||||
|
};
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_attestation_score() {
|
||||||
|
let spec = &ChainSpec::foundation();
|
||||||
|
let (ref mut state, ref keypairs) = attestation_test_state(spec, 1);
|
||||||
|
let slot = state.slot - 1;
|
||||||
|
let committees = state
|
||||||
|
.get_crosslink_committees_at_slot(slot, spec)
|
||||||
|
.unwrap()
|
||||||
|
.clone();
|
||||||
|
|
||||||
|
for committee in committees {
|
||||||
|
let att1 = signed_attestation(&committee, keypairs, ..2, slot, state, spec, None);
|
||||||
|
let att2 = signed_attestation(&committee, keypairs, .., slot, state, spec, None);
|
||||||
|
|
||||||
|
assert_eq!(
|
||||||
|
att1.aggregation_bitfield.num_set_bits(),
|
||||||
|
attestation_score(&att1, state, spec)
|
||||||
|
);
|
||||||
|
|
||||||
|
state
|
||||||
|
.current_epoch_attestations
|
||||||
|
.push(PendingAttestation::from_attestation(&att1, state.slot));
|
||||||
|
|
||||||
|
assert_eq!(
|
||||||
|
committee.committee.len() - 2,
|
||||||
|
attestation_score(&att2, state, spec)
|
||||||
|
);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/// End-to-end test of basic attestation handling.
|
||||||
|
#[test]
|
||||||
|
fn attestation_aggregation_insert_get_prune() {
|
||||||
|
let spec = &ChainSpec::foundation();
|
||||||
|
let (ref mut state, ref keypairs) = attestation_test_state(spec, 1);
|
||||||
|
let op_pool = OperationPool::new();
|
||||||
|
|
||||||
|
let slot = state.slot - 1;
|
||||||
|
let committees = state
|
||||||
|
.get_crosslink_committees_at_slot(slot, spec)
|
||||||
|
.unwrap()
|
||||||
|
.clone();
|
||||||
|
|
||||||
|
assert_eq!(
|
||||||
|
committees.len(),
|
||||||
|
1,
|
||||||
|
"we expect just one committee with this many validators"
|
||||||
|
);
|
||||||
|
|
||||||
|
for committee in &committees {
|
||||||
|
let step_size = 2;
|
||||||
|
for i in (0..committee.committee.len()).step_by(step_size) {
|
||||||
|
let att = signed_attestation(
|
||||||
|
committee,
|
||||||
|
keypairs,
|
||||||
|
i..i + step_size,
|
||||||
|
slot,
|
||||||
|
state,
|
||||||
|
spec,
|
||||||
|
None,
|
||||||
|
);
|
||||||
|
fake_latest_crosslink(&att, state, spec);
|
||||||
|
op_pool.insert_attestation(att, state, spec).unwrap();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
assert_eq!(op_pool.attestations.read().len(), committees.len());
|
||||||
|
assert_eq!(op_pool.num_attestations(), committees.len());
|
||||||
|
|
||||||
|
// Before the min attestation inclusion delay, get_attestations shouldn't return anything.
|
||||||
|
assert_eq!(op_pool.get_attestations(state, spec).len(), 0);
|
||||||
|
|
||||||
|
// Then once the delay has elapsed, we should get a single aggregated attestation.
|
||||||
|
state.slot += spec.min_attestation_inclusion_delay;
|
||||||
|
|
||||||
|
let block_attestations = op_pool.get_attestations(state, spec);
|
||||||
|
assert_eq!(block_attestations.len(), committees.len());
|
||||||
|
|
||||||
|
let agg_att = &block_attestations[0];
|
||||||
|
assert_eq!(
|
||||||
|
agg_att.aggregation_bitfield.num_set_bits(),
|
||||||
|
spec.target_committee_size as usize
|
||||||
|
);
|
||||||
|
|
||||||
|
// Prune attestations shouldn't do anything at this point.
|
||||||
|
op_pool.prune_attestations(state, spec);
|
||||||
|
assert_eq!(op_pool.num_attestations(), committees.len());
|
||||||
|
|
||||||
|
// But once we advance to an epoch after the attestation, it should prune it out of
|
||||||
|
// existence.
|
||||||
|
state.slot = slot + spec.slots_per_epoch;
|
||||||
|
op_pool.prune_attestations(state, spec);
|
||||||
|
assert_eq!(op_pool.num_attestations(), 0);
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Adding an attestation already in the pool should not increase the size of the pool.
|
||||||
|
#[test]
|
||||||
|
fn attestation_duplicate() {
|
||||||
|
let spec = &ChainSpec::foundation();
|
||||||
|
let (ref mut state, ref keypairs) = attestation_test_state(spec, 1);
|
||||||
|
let op_pool = OperationPool::new();
|
||||||
|
|
||||||
|
let slot = state.slot - 1;
|
||||||
|
let committees = state
|
||||||
|
.get_crosslink_committees_at_slot(slot, spec)
|
||||||
|
.unwrap()
|
||||||
|
.clone();
|
||||||
|
|
||||||
|
for committee in &committees {
|
||||||
|
let att = signed_attestation(committee, keypairs, .., slot, state, spec, None);
|
||||||
|
fake_latest_crosslink(&att, state, spec);
|
||||||
|
op_pool
|
||||||
|
.insert_attestation(att.clone(), state, spec)
|
||||||
|
.unwrap();
|
||||||
|
op_pool.insert_attestation(att, state, spec).unwrap();
|
||||||
|
}
|
||||||
|
|
||||||
|
assert_eq!(op_pool.num_attestations(), committees.len());
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Adding lots of attestations that only intersect pairwise should lead to two aggregate
|
||||||
|
/// attestations.
|
||||||
|
#[test]
|
||||||
|
fn attestation_pairwise_overlapping() {
|
||||||
|
let spec = &ChainSpec::foundation();
|
||||||
|
let (ref mut state, ref keypairs) = attestation_test_state(spec, 1);
|
||||||
|
let op_pool = OperationPool::new();
|
||||||
|
|
||||||
|
let slot = state.slot - 1;
|
||||||
|
let committees = state
|
||||||
|
.get_crosslink_committees_at_slot(slot, spec)
|
||||||
|
.unwrap()
|
||||||
|
.clone();
|
||||||
|
|
||||||
|
let step_size = 2;
|
||||||
|
for committee in &committees {
|
||||||
|
// Create attestations that overlap on `step_size` validators, like:
|
||||||
|
// {0,1,2,3}, {2,3,4,5}, {4,5,6,7}, ...
|
||||||
|
for i in (0..committee.committee.len() - step_size).step_by(step_size) {
|
||||||
|
let att = signed_attestation(
|
||||||
|
committee,
|
||||||
|
keypairs,
|
||||||
|
i..i + 2 * step_size,
|
||||||
|
slot,
|
||||||
|
state,
|
||||||
|
spec,
|
||||||
|
None,
|
||||||
|
);
|
||||||
|
fake_latest_crosslink(&att, state, spec);
|
||||||
|
op_pool.insert_attestation(att, state, spec).unwrap();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// The attestations should get aggregated into two attestations that comprise all
|
||||||
|
// validators.
|
||||||
|
assert_eq!(op_pool.attestations.read().len(), committees.len());
|
||||||
|
assert_eq!(op_pool.num_attestations(), 2 * committees.len());
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Create a bunch of attestations signed by a small number of validators, and another
|
||||||
|
/// bunch signed by a larger number, such that there are at least `max_attestations`
|
||||||
|
/// signed by the larger number. Then, check that `get_attestations` only returns the
|
||||||
|
/// high-quality attestations. To ensure that no aggregation occurs, ALL attestations
|
||||||
|
/// are also signed by the 0th member of the committee.
|
||||||
|
#[test]
|
||||||
|
fn attestation_get_max() {
|
||||||
|
let spec = &ChainSpec::foundation();
|
||||||
|
let small_step_size = 2;
|
||||||
|
let big_step_size = 4;
|
||||||
|
let (ref mut state, ref keypairs) = attestation_test_state(spec, big_step_size);
|
||||||
|
let op_pool = OperationPool::new();
|
||||||
|
|
||||||
|
let slot = state.slot - 1;
|
||||||
|
let committees = state
|
||||||
|
.get_crosslink_committees_at_slot(slot, spec)
|
||||||
|
.unwrap()
|
||||||
|
.clone();
|
||||||
|
|
||||||
|
let max_attestations = spec.max_attestations as usize;
|
||||||
|
let target_committee_size = spec.target_committee_size as usize;
|
||||||
|
|
||||||
|
let mut insert_attestations = |committee, step_size| {
|
||||||
|
for i in (0..target_committee_size).step_by(step_size) {
|
||||||
|
let att = signed_attestation(
|
||||||
|
committee,
|
||||||
|
keypairs,
|
||||||
|
i..i + step_size,
|
||||||
|
slot,
|
||||||
|
state,
|
||||||
|
spec,
|
||||||
|
if i == 0 { None } else { Some(0) },
|
||||||
|
);
|
||||||
|
fake_latest_crosslink(&att, state, spec);
|
||||||
|
op_pool.insert_attestation(att, state, spec).unwrap();
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
for committee in &committees {
|
||||||
|
assert_eq!(committee.committee.len(), target_committee_size);
|
||||||
|
// Attestations signed by only 2-3 validators
|
||||||
|
insert_attestations(committee, small_step_size);
|
||||||
|
// Attestations signed by 4+ validators
|
||||||
|
insert_attestations(committee, big_step_size);
|
||||||
|
}
|
||||||
|
|
||||||
|
let num_small = target_committee_size / small_step_size;
|
||||||
|
let num_big = target_committee_size / big_step_size;
|
||||||
|
|
||||||
|
assert_eq!(op_pool.attestations.read().len(), committees.len());
|
||||||
|
assert_eq!(
|
||||||
|
op_pool.num_attestations(),
|
||||||
|
(num_small + num_big) * committees.len()
|
||||||
|
);
|
||||||
|
assert!(op_pool.num_attestations() > max_attestations);
|
||||||
|
|
||||||
|
state.slot += spec.min_attestation_inclusion_delay;
|
||||||
|
let best_attestations = op_pool.get_attestations(state, spec);
|
||||||
|
assert_eq!(best_attestations.len(), max_attestations);
|
||||||
|
|
||||||
|
// All the best attestations should be signed by at least `big_step_size` (4) validators.
|
||||||
|
for att in &best_attestations {
|
||||||
|
assert!(att.aggregation_bitfield.num_set_bits() >= big_step_size);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// TODO: more tests
|
||||||
|
}
|
@ -1,4 +1,3 @@
|
|||||||
use self::verify_proposer_slashing::verify_proposer_slashing;
|
|
||||||
use crate::common::slash_validator;
|
use crate::common::slash_validator;
|
||||||
use errors::{BlockInvalid as Invalid, BlockProcessingError as Error, IntoWithIndex};
|
use errors::{BlockInvalid as Invalid, BlockProcessingError as Error, IntoWithIndex};
|
||||||
use rayon::prelude::*;
|
use rayon::prelude::*;
|
||||||
@ -6,13 +5,20 @@ use ssz::{SignedRoot, TreeHash};
|
|||||||
use types::*;
|
use types::*;
|
||||||
|
|
||||||
pub use self::verify_attester_slashing::{
|
pub use self::verify_attester_slashing::{
|
||||||
gather_attester_slashing_indices, verify_attester_slashing,
|
gather_attester_slashing_indices, gather_attester_slashing_indices_modular,
|
||||||
|
verify_attester_slashing,
|
||||||
|
};
|
||||||
|
pub use self::verify_proposer_slashing::verify_proposer_slashing;
|
||||||
|
pub use validate_attestation::{
|
||||||
|
validate_attestation, validate_attestation_time_independent_only,
|
||||||
|
validate_attestation_without_signature,
|
||||||
};
|
};
|
||||||
pub use validate_attestation::{validate_attestation, validate_attestation_without_signature};
|
|
||||||
pub use verify_deposit::{get_existing_validator_index, verify_deposit, verify_deposit_index};
|
pub use verify_deposit::{get_existing_validator_index, verify_deposit, verify_deposit_index};
|
||||||
pub use verify_exit::verify_exit;
|
pub use verify_exit::{verify_exit, verify_exit_time_independent_only};
|
||||||
pub use verify_slashable_attestation::verify_slashable_attestation;
|
pub use verify_slashable_attestation::verify_slashable_attestation;
|
||||||
pub use verify_transfer::{execute_transfer, verify_transfer};
|
pub use verify_transfer::{
|
||||||
|
execute_transfer, verify_transfer, verify_transfer_time_independent_only,
|
||||||
|
};
|
||||||
|
|
||||||
pub mod errors;
|
pub mod errors;
|
||||||
mod validate_attestation;
|
mod validate_attestation;
|
||||||
@ -316,13 +322,7 @@ pub fn process_attestations(
|
|||||||
|
|
||||||
// Update the state in series.
|
// Update the state in series.
|
||||||
for attestation in attestations {
|
for attestation in attestations {
|
||||||
let pending_attestation = PendingAttestation {
|
let pending_attestation = PendingAttestation::from_attestation(attestation, state.slot);
|
||||||
data: attestation.data.clone(),
|
|
||||||
aggregation_bitfield: attestation.aggregation_bitfield.clone(),
|
|
||||||
custody_bitfield: attestation.custody_bitfield.clone(),
|
|
||||||
inclusion_slot: state.slot,
|
|
||||||
};
|
|
||||||
|
|
||||||
let attestation_epoch = attestation.data.slot.epoch(spec.slots_per_epoch);
|
let attestation_epoch = attestation.data.slot.epoch(spec.slots_per_epoch);
|
||||||
|
|
||||||
if attestation_epoch == state.current_epoch(spec) {
|
if attestation_epoch == state.current_epoch(spec) {
|
||||||
|
@ -390,6 +390,11 @@ pub enum TransferInvalid {
|
|||||||
///
|
///
|
||||||
/// (state_slot, transfer_slot)
|
/// (state_slot, transfer_slot)
|
||||||
StateSlotMismatch(Slot, Slot),
|
StateSlotMismatch(Slot, Slot),
|
||||||
|
/// The `transfer.slot` is in the past relative to the state slot.
|
||||||
|
///
|
||||||
|
///
|
||||||
|
/// (state_slot, transfer_slot)
|
||||||
|
TransferSlotInPast(Slot, Slot),
|
||||||
/// The `transfer.from` validator has been activated and is not withdrawable.
|
/// The `transfer.from` validator has been activated and is not withdrawable.
|
||||||
///
|
///
|
||||||
/// (from_validator)
|
/// (from_validator)
|
||||||
|
@ -14,7 +14,16 @@ pub fn validate_attestation(
|
|||||||
attestation: &Attestation,
|
attestation: &Attestation,
|
||||||
spec: &ChainSpec,
|
spec: &ChainSpec,
|
||||||
) -> Result<(), Error> {
|
) -> Result<(), Error> {
|
||||||
validate_attestation_signature_optional(state, attestation, spec, true)
|
validate_attestation_parametric(state, attestation, spec, true, false)
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Like `validate_attestation` but doesn't run checks which may become true in future states.
|
||||||
|
pub fn validate_attestation_time_independent_only(
|
||||||
|
state: &BeaconState,
|
||||||
|
attestation: &Attestation,
|
||||||
|
spec: &ChainSpec,
|
||||||
|
) -> Result<(), Error> {
|
||||||
|
validate_attestation_parametric(state, attestation, spec, true, true)
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Indicates if an `Attestation` is valid to be included in a block in the current epoch of the
|
/// Indicates if an `Attestation` is valid to be included in a block in the current epoch of the
|
||||||
@ -28,7 +37,7 @@ pub fn validate_attestation_without_signature(
|
|||||||
attestation: &Attestation,
|
attestation: &Attestation,
|
||||||
spec: &ChainSpec,
|
spec: &ChainSpec,
|
||||||
) -> Result<(), Error> {
|
) -> Result<(), Error> {
|
||||||
validate_attestation_signature_optional(state, attestation, spec, false)
|
validate_attestation_parametric(state, attestation, spec, false, false)
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Indicates if an `Attestation` is valid to be included in a block in the current epoch of the
|
/// Indicates if an `Attestation` is valid to be included in a block in the current epoch of the
|
||||||
@ -36,15 +45,13 @@ pub fn validate_attestation_without_signature(
|
|||||||
///
|
///
|
||||||
///
|
///
|
||||||
/// Spec v0.5.0
|
/// Spec v0.5.0
|
||||||
fn validate_attestation_signature_optional(
|
fn validate_attestation_parametric(
|
||||||
state: &BeaconState,
|
state: &BeaconState,
|
||||||
attestation: &Attestation,
|
attestation: &Attestation,
|
||||||
spec: &ChainSpec,
|
spec: &ChainSpec,
|
||||||
verify_signature: bool,
|
verify_signature: bool,
|
||||||
|
time_independent_only: bool,
|
||||||
) -> Result<(), Error> {
|
) -> Result<(), Error> {
|
||||||
let state_epoch = state.slot.epoch(spec.slots_per_epoch);
|
|
||||||
let attestation_epoch = attestation.data.slot.epoch(spec.slots_per_epoch);
|
|
||||||
|
|
||||||
// Can't submit pre-historic attestations.
|
// Can't submit pre-historic attestations.
|
||||||
verify!(
|
verify!(
|
||||||
attestation.data.slot >= spec.genesis_slot,
|
attestation.data.slot >= spec.genesis_slot,
|
||||||
@ -65,7 +72,8 @@ fn validate_attestation_signature_optional(
|
|||||||
|
|
||||||
// Can't submit attestation too quickly.
|
// Can't submit attestation too quickly.
|
||||||
verify!(
|
verify!(
|
||||||
attestation.data.slot + spec.min_attestation_inclusion_delay <= state.slot,
|
time_independent_only
|
||||||
|
|| attestation.data.slot + spec.min_attestation_inclusion_delay <= state.slot,
|
||||||
Invalid::IncludedTooEarly {
|
Invalid::IncludedTooEarly {
|
||||||
state: state.slot,
|
state: state.slot,
|
||||||
delay: spec.min_attestation_inclusion_delay,
|
delay: spec.min_attestation_inclusion_delay,
|
||||||
@ -74,40 +82,8 @@ fn validate_attestation_signature_optional(
|
|||||||
);
|
);
|
||||||
|
|
||||||
// Verify the justified epoch and root is correct.
|
// Verify the justified epoch and root is correct.
|
||||||
if attestation_epoch >= state_epoch {
|
if !time_independent_only {
|
||||||
verify!(
|
verify_justified_epoch_and_root(attestation, state, spec)?;
|
||||||
attestation.data.source_epoch == state.current_justified_epoch,
|
|
||||||
Invalid::WrongJustifiedEpoch {
|
|
||||||
state: state.current_justified_epoch,
|
|
||||||
attestation: attestation.data.source_epoch,
|
|
||||||
is_current: true,
|
|
||||||
}
|
|
||||||
);
|
|
||||||
verify!(
|
|
||||||
attestation.data.source_root == state.current_justified_root,
|
|
||||||
Invalid::WrongJustifiedRoot {
|
|
||||||
state: state.current_justified_root,
|
|
||||||
attestation: attestation.data.source_root,
|
|
||||||
is_current: true,
|
|
||||||
}
|
|
||||||
);
|
|
||||||
} else {
|
|
||||||
verify!(
|
|
||||||
attestation.data.source_epoch == state.previous_justified_epoch,
|
|
||||||
Invalid::WrongJustifiedEpoch {
|
|
||||||
state: state.previous_justified_epoch,
|
|
||||||
attestation: attestation.data.source_epoch,
|
|
||||||
is_current: false,
|
|
||||||
}
|
|
||||||
);
|
|
||||||
verify!(
|
|
||||||
attestation.data.source_root == state.previous_justified_root,
|
|
||||||
Invalid::WrongJustifiedRoot {
|
|
||||||
state: state.previous_justified_root,
|
|
||||||
attestation: attestation.data.source_root,
|
|
||||||
is_current: true,
|
|
||||||
}
|
|
||||||
);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// Check that the crosslink data is valid.
|
// Check that the crosslink data is valid.
|
||||||
@ -188,6 +164,56 @@ fn validate_attestation_signature_optional(
|
|||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Verify that the `source_epoch` and `source_root` of an `Attestation` correctly
|
||||||
|
/// match the current (or previous) justified epoch and root from the state.
|
||||||
|
///
|
||||||
|
/// Spec v0.5.0
|
||||||
|
fn verify_justified_epoch_and_root(
|
||||||
|
attestation: &Attestation,
|
||||||
|
state: &BeaconState,
|
||||||
|
spec: &ChainSpec,
|
||||||
|
) -> Result<(), Error> {
|
||||||
|
let state_epoch = state.slot.epoch(spec.slots_per_epoch);
|
||||||
|
let attestation_epoch = attestation.data.slot.epoch(spec.slots_per_epoch);
|
||||||
|
|
||||||
|
if attestation_epoch >= state_epoch {
|
||||||
|
verify!(
|
||||||
|
attestation.data.source_epoch == state.current_justified_epoch,
|
||||||
|
Invalid::WrongJustifiedEpoch {
|
||||||
|
state: state.current_justified_epoch,
|
||||||
|
attestation: attestation.data.source_epoch,
|
||||||
|
is_current: true,
|
||||||
|
}
|
||||||
|
);
|
||||||
|
verify!(
|
||||||
|
attestation.data.source_root == state.current_justified_root,
|
||||||
|
Invalid::WrongJustifiedRoot {
|
||||||
|
state: state.current_justified_root,
|
||||||
|
attestation: attestation.data.source_root,
|
||||||
|
is_current: true,
|
||||||
|
}
|
||||||
|
);
|
||||||
|
} else {
|
||||||
|
verify!(
|
||||||
|
attestation.data.source_epoch == state.previous_justified_epoch,
|
||||||
|
Invalid::WrongJustifiedEpoch {
|
||||||
|
state: state.previous_justified_epoch,
|
||||||
|
attestation: attestation.data.source_epoch,
|
||||||
|
is_current: false,
|
||||||
|
}
|
||||||
|
);
|
||||||
|
verify!(
|
||||||
|
attestation.data.source_root == state.previous_justified_root,
|
||||||
|
Invalid::WrongJustifiedRoot {
|
||||||
|
state: state.previous_justified_root,
|
||||||
|
attestation: attestation.data.source_root,
|
||||||
|
is_current: true,
|
||||||
|
}
|
||||||
|
);
|
||||||
|
}
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
/// Verifies an aggregate signature for some given `AttestationData`, returning `true` if the
|
/// Verifies an aggregate signature for some given `AttestationData`, returning `true` if the
|
||||||
/// `aggregate_signature` is valid.
|
/// `aggregate_signature` is valid.
|
||||||
///
|
///
|
||||||
|
@ -47,6 +47,25 @@ pub fn gather_attester_slashing_indices(
|
|||||||
attester_slashing: &AttesterSlashing,
|
attester_slashing: &AttesterSlashing,
|
||||||
spec: &ChainSpec,
|
spec: &ChainSpec,
|
||||||
) -> Result<Vec<u64>, Error> {
|
) -> Result<Vec<u64>, Error> {
|
||||||
|
gather_attester_slashing_indices_modular(
|
||||||
|
state,
|
||||||
|
attester_slashing,
|
||||||
|
|_, validator| validator.slashed,
|
||||||
|
spec,
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Same as `gather_attester_slashing_indices` but allows the caller to specify the criteria
|
||||||
|
/// for determining whether a given validator should be considered slashed.
|
||||||
|
pub fn gather_attester_slashing_indices_modular<F>(
|
||||||
|
state: &BeaconState,
|
||||||
|
attester_slashing: &AttesterSlashing,
|
||||||
|
is_slashed: F,
|
||||||
|
spec: &ChainSpec,
|
||||||
|
) -> Result<Vec<u64>, Error>
|
||||||
|
where
|
||||||
|
F: Fn(u64, &Validator) -> bool,
|
||||||
|
{
|
||||||
let slashable_attestation_1 = &attester_slashing.slashable_attestation_1;
|
let slashable_attestation_1 = &attester_slashing.slashable_attestation_1;
|
||||||
let slashable_attestation_2 = &attester_slashing.slashable_attestation_2;
|
let slashable_attestation_2 = &attester_slashing.slashable_attestation_2;
|
||||||
|
|
||||||
@ -57,7 +76,7 @@ pub fn gather_attester_slashing_indices(
|
|||||||
.get(*i as usize)
|
.get(*i as usize)
|
||||||
.ok_or_else(|| Error::Invalid(Invalid::UnknownValidator(*i)))?;
|
.ok_or_else(|| Error::Invalid(Invalid::UnknownValidator(*i)))?;
|
||||||
|
|
||||||
if slashable_attestation_2.validator_indices.contains(&i) & !validator.slashed {
|
if slashable_attestation_2.validator_indices.contains(&i) & !is_slashed(*i, validator) {
|
||||||
// TODO: verify that we should reject any slashable attestation which includes a
|
// TODO: verify that we should reject any slashable attestation which includes a
|
||||||
// withdrawn validator. PH has asked the question on gitter, awaiting response.
|
// withdrawn validator. PH has asked the question on gitter, awaiting response.
|
||||||
verify!(
|
verify!(
|
||||||
|
@ -12,6 +12,25 @@ pub fn verify_exit(
|
|||||||
state: &BeaconState,
|
state: &BeaconState,
|
||||||
exit: &VoluntaryExit,
|
exit: &VoluntaryExit,
|
||||||
spec: &ChainSpec,
|
spec: &ChainSpec,
|
||||||
|
) -> Result<(), Error> {
|
||||||
|
verify_exit_parametric(state, exit, spec, false)
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Like `verify_exit` but doesn't run checks which may become true in future states.
|
||||||
|
pub fn verify_exit_time_independent_only(
|
||||||
|
state: &BeaconState,
|
||||||
|
exit: &VoluntaryExit,
|
||||||
|
spec: &ChainSpec,
|
||||||
|
) -> Result<(), Error> {
|
||||||
|
verify_exit_parametric(state, exit, spec, true)
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Parametric version of `verify_exit` that skips some checks if `time_independent_only` is true.
|
||||||
|
fn verify_exit_parametric(
|
||||||
|
state: &BeaconState,
|
||||||
|
exit: &VoluntaryExit,
|
||||||
|
spec: &ChainSpec,
|
||||||
|
time_independent_only: bool,
|
||||||
) -> Result<(), Error> {
|
) -> Result<(), Error> {
|
||||||
let validator = state
|
let validator = state
|
||||||
.validator_registry
|
.validator_registry
|
||||||
@ -32,7 +51,7 @@ pub fn verify_exit(
|
|||||||
|
|
||||||
// Exits must specify an epoch when they become valid; they are not valid before then.
|
// Exits must specify an epoch when they become valid; they are not valid before then.
|
||||||
verify!(
|
verify!(
|
||||||
state.current_epoch(spec) >= exit.epoch,
|
time_independent_only || state.current_epoch(spec) >= exit.epoch,
|
||||||
Invalid::FutureEpoch {
|
Invalid::FutureEpoch {
|
||||||
state: state.current_epoch(spec),
|
state: state.current_epoch(spec),
|
||||||
exit: exit.epoch
|
exit: exit.epoch
|
||||||
|
@ -15,6 +15,25 @@ pub fn verify_transfer(
|
|||||||
state: &BeaconState,
|
state: &BeaconState,
|
||||||
transfer: &Transfer,
|
transfer: &Transfer,
|
||||||
spec: &ChainSpec,
|
spec: &ChainSpec,
|
||||||
|
) -> Result<(), Error> {
|
||||||
|
verify_transfer_parametric(state, transfer, spec, false)
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Like `verify_transfer` but doesn't run checks which may become true in future states.
|
||||||
|
pub fn verify_transfer_time_independent_only(
|
||||||
|
state: &BeaconState,
|
||||||
|
transfer: &Transfer,
|
||||||
|
spec: &ChainSpec,
|
||||||
|
) -> Result<(), Error> {
|
||||||
|
verify_transfer_parametric(state, transfer, spec, true)
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Parametric version of `verify_transfer` that allows some checks to be skipped.
|
||||||
|
fn verify_transfer_parametric(
|
||||||
|
state: &BeaconState,
|
||||||
|
transfer: &Transfer,
|
||||||
|
spec: &ChainSpec,
|
||||||
|
time_independent_only: bool,
|
||||||
) -> Result<(), Error> {
|
) -> Result<(), Error> {
|
||||||
let sender_balance = *state
|
let sender_balance = *state
|
||||||
.validator_balances
|
.validator_balances
|
||||||
@ -27,17 +46,18 @@ pub fn verify_transfer(
|
|||||||
.ok_or_else(|| Error::Invalid(Invalid::FeeOverflow(transfer.amount, transfer.fee)))?;
|
.ok_or_else(|| Error::Invalid(Invalid::FeeOverflow(transfer.amount, transfer.fee)))?;
|
||||||
|
|
||||||
verify!(
|
verify!(
|
||||||
sender_balance >= transfer.amount,
|
time_independent_only || sender_balance >= transfer.amount,
|
||||||
Invalid::FromBalanceInsufficient(transfer.amount, sender_balance)
|
Invalid::FromBalanceInsufficient(transfer.amount, sender_balance)
|
||||||
);
|
);
|
||||||
|
|
||||||
verify!(
|
verify!(
|
||||||
sender_balance >= transfer.fee,
|
time_independent_only || sender_balance >= transfer.fee,
|
||||||
Invalid::FromBalanceInsufficient(transfer.fee, sender_balance)
|
Invalid::FromBalanceInsufficient(transfer.fee, sender_balance)
|
||||||
);
|
);
|
||||||
|
|
||||||
verify!(
|
verify!(
|
||||||
(sender_balance == total_amount)
|
time_independent_only
|
||||||
|
|| (sender_balance == total_amount)
|
||||||
|| (sender_balance >= (total_amount + spec.min_deposit_amount)),
|
|| (sender_balance >= (total_amount + spec.min_deposit_amount)),
|
||||||
Invalid::InvalidResultingFromBalance(
|
Invalid::InvalidResultingFromBalance(
|
||||||
sender_balance - total_amount,
|
sender_balance - total_amount,
|
||||||
@ -45,10 +65,17 @@ pub fn verify_transfer(
|
|||||||
)
|
)
|
||||||
);
|
);
|
||||||
|
|
||||||
verify!(
|
if time_independent_only {
|
||||||
state.slot == transfer.slot,
|
verify!(
|
||||||
Invalid::StateSlotMismatch(state.slot, transfer.slot)
|
state.slot <= transfer.slot,
|
||||||
);
|
Invalid::TransferSlotInPast(state.slot, transfer.slot)
|
||||||
|
);
|
||||||
|
} else {
|
||||||
|
verify!(
|
||||||
|
state.slot == transfer.slot,
|
||||||
|
Invalid::StateSlotMismatch(state.slot, transfer.slot)
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
let sender_validator = state
|
let sender_validator = state
|
||||||
.validator_registry
|
.validator_registry
|
||||||
@ -57,7 +84,8 @@ pub fn verify_transfer(
|
|||||||
let epoch = state.slot.epoch(spec.slots_per_epoch);
|
let epoch = state.slot.epoch(spec.slots_per_epoch);
|
||||||
|
|
||||||
verify!(
|
verify!(
|
||||||
sender_validator.is_withdrawable_at(epoch)
|
time_independent_only
|
||||||
|
|| sender_validator.is_withdrawable_at(epoch)
|
||||||
|| sender_validator.activation_epoch == spec.far_future_epoch,
|
|| sender_validator.activation_epoch == spec.far_future_epoch,
|
||||||
Invalid::FromValidatorIneligableForTransfer(transfer.sender)
|
Invalid::FromValidatorIneligableForTransfer(transfer.sender)
|
||||||
);
|
);
|
||||||
|
@ -8,6 +8,7 @@ edition = "2018"
|
|||||||
bls = { path = "../utils/bls" }
|
bls = { path = "../utils/bls" }
|
||||||
boolean-bitfield = { path = "../utils/boolean-bitfield" }
|
boolean-bitfield = { path = "../utils/boolean-bitfield" }
|
||||||
dirs = "1.0"
|
dirs = "1.0"
|
||||||
|
derivative = "1.0"
|
||||||
ethereum-types = "0.5"
|
ethereum-types = "0.5"
|
||||||
hashing = { path = "../utils/hashing" }
|
hashing = { path = "../utils/hashing" }
|
||||||
hex = "0.3"
|
hex = "0.3"
|
||||||
|
@ -28,6 +28,29 @@ pub struct Attestation {
|
|||||||
pub aggregate_signature: AggregateSignature,
|
pub aggregate_signature: AggregateSignature,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
impl Attestation {
|
||||||
|
/// Are the aggregation bitfields of these attestations disjoint?
|
||||||
|
pub fn signers_disjoint_from(&self, other: &Attestation) -> bool {
|
||||||
|
self.aggregation_bitfield
|
||||||
|
.intersection(&other.aggregation_bitfield)
|
||||||
|
.is_zero()
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Aggregate another Attestation into this one.
|
||||||
|
///
|
||||||
|
/// The aggregation bitfields must be disjoint, and the data must be the same.
|
||||||
|
pub fn aggregate(&mut self, other: &Attestation) {
|
||||||
|
debug_assert_eq!(self.data, other.data);
|
||||||
|
debug_assert!(self.signers_disjoint_from(other));
|
||||||
|
|
||||||
|
self.aggregation_bitfield
|
||||||
|
.union_inplace(&other.aggregation_bitfield);
|
||||||
|
self.custody_bitfield.union_inplace(&other.custody_bitfield);
|
||||||
|
self.aggregate_signature
|
||||||
|
.add_aggregate(&other.aggregate_signature);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
#[cfg(test)]
|
#[cfg(test)]
|
||||||
mod tests {
|
mod tests {
|
||||||
use super::*;
|
use super::*;
|
||||||
|
@ -1,9 +1,10 @@
|
|||||||
use crate::*;
|
use crate::*;
|
||||||
use serde_derive::{Deserialize, Serialize};
|
use serde_derive::{Deserialize, Serialize};
|
||||||
|
|
||||||
#[derive(Debug, PartialEq, Clone, Default, Serialize, Deserialize)]
|
#[derive(Debug, PartialEq, Clone, Copy, Default, Serialize, Deserialize)]
|
||||||
pub struct AttestationDuty {
|
pub struct AttestationDuty {
|
||||||
pub slot: Slot,
|
pub slot: Slot,
|
||||||
pub shard: Shard,
|
pub shard: Shard,
|
||||||
pub committee_index: usize,
|
pub committee_index: usize,
|
||||||
|
pub committee_len: usize,
|
||||||
}
|
}
|
||||||
|
@ -92,6 +92,7 @@ impl EpochCache {
|
|||||||
slot,
|
slot,
|
||||||
shard,
|
shard,
|
||||||
committee_index: k,
|
committee_index: k,
|
||||||
|
committee_len: crosslink_committee.committee.len(),
|
||||||
};
|
};
|
||||||
attestation_duties[*validator_index] = Some(attestation_duty)
|
attestation_duties[*validator_index] = Some(attestation_duty)
|
||||||
}
|
}
|
||||||
|
@ -1,5 +1,5 @@
|
|||||||
use crate::test_utils::TestRandom;
|
use crate::test_utils::TestRandom;
|
||||||
use crate::{AttestationData, Bitfield, Slot};
|
use crate::{Attestation, AttestationData, Bitfield, Slot};
|
||||||
use rand::RngCore;
|
use rand::RngCore;
|
||||||
use serde_derive::{Deserialize, Serialize};
|
use serde_derive::{Deserialize, Serialize};
|
||||||
use ssz_derive::{Decode, Encode, TreeHash};
|
use ssz_derive::{Decode, Encode, TreeHash};
|
||||||
@ -16,6 +16,18 @@ pub struct PendingAttestation {
|
|||||||
pub inclusion_slot: Slot,
|
pub inclusion_slot: Slot,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
impl PendingAttestation {
|
||||||
|
/// Create a `PendingAttestation` from an `Attestation`, at the given `inclusion_slot`.
|
||||||
|
pub fn from_attestation(attestation: &Attestation, inclusion_slot: Slot) -> Self {
|
||||||
|
PendingAttestation {
|
||||||
|
data: attestation.data.clone(),
|
||||||
|
aggregation_bitfield: attestation.aggregation_bitfield.clone(),
|
||||||
|
custody_bitfield: attestation.custody_bitfield.clone(),
|
||||||
|
inclusion_slot,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
#[cfg(test)]
|
#[cfg(test)]
|
||||||
mod tests {
|
mod tests {
|
||||||
use super::*;
|
use super::*;
|
||||||
|
@ -19,7 +19,7 @@ pub fn generate_deterministic_keypairs(validator_count: usize) -> Vec<Keypair> {
|
|||||||
.collect::<Vec<usize>>()
|
.collect::<Vec<usize>>()
|
||||||
.par_iter()
|
.par_iter()
|
||||||
.map(|&i| {
|
.map(|&i| {
|
||||||
let secret = int_to_bytes48(i as u64 + 1);
|
let secret = int_to_bytes48(i as u64 + 1000);
|
||||||
let sk = SecretKey::from_bytes(&secret).unwrap();
|
let sk = SecretKey::from_bytes(&secret).unwrap();
|
||||||
let pk = PublicKey::from_secret_key(&sk);
|
let pk = PublicKey::from_secret_key(&sk);
|
||||||
Keypair { sk, pk }
|
Keypair { sk, pk }
|
||||||
|
@ -1,6 +1,7 @@
|
|||||||
use super::Slot;
|
use super::Slot;
|
||||||
use crate::test_utils::TestRandom;
|
use crate::test_utils::TestRandom;
|
||||||
use bls::{PublicKey, Signature};
|
use bls::{PublicKey, Signature};
|
||||||
|
use derivative::Derivative;
|
||||||
use rand::RngCore;
|
use rand::RngCore;
|
||||||
use serde_derive::{Deserialize, Serialize};
|
use serde_derive::{Deserialize, Serialize};
|
||||||
use ssz::TreeHash;
|
use ssz::TreeHash;
|
||||||
@ -12,7 +13,6 @@ use test_random_derive::TestRandom;
|
|||||||
/// Spec v0.5.0
|
/// Spec v0.5.0
|
||||||
#[derive(
|
#[derive(
|
||||||
Debug,
|
Debug,
|
||||||
PartialEq,
|
|
||||||
Clone,
|
Clone,
|
||||||
Serialize,
|
Serialize,
|
||||||
Deserialize,
|
Deserialize,
|
||||||
@ -21,7 +21,9 @@ use test_random_derive::TestRandom;
|
|||||||
TreeHash,
|
TreeHash,
|
||||||
TestRandom,
|
TestRandom,
|
||||||
SignedRoot,
|
SignedRoot,
|
||||||
|
Derivative,
|
||||||
)]
|
)]
|
||||||
|
#[derivative(PartialEq, Eq, Hash)]
|
||||||
pub struct Transfer {
|
pub struct Transfer {
|
||||||
pub sender: u64,
|
pub sender: u64,
|
||||||
pub recipient: u64,
|
pub recipient: u64,
|
||||||
@ -29,6 +31,7 @@ pub struct Transfer {
|
|||||||
pub fee: u64,
|
pub fee: u64,
|
||||||
pub slot: Slot,
|
pub slot: Slot,
|
||||||
pub pubkey: PublicKey,
|
pub pubkey: PublicKey,
|
||||||
|
#[derivative(Hash = "ignore")]
|
||||||
pub signature: Signature,
|
pub signature: Signature,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1,7 +1,7 @@
|
|||||||
use super::PublicKey;
|
use super::PublicKey;
|
||||||
use bls_aggregates::AggregatePublicKey as RawAggregatePublicKey;
|
use bls_aggregates::AggregatePublicKey as RawAggregatePublicKey;
|
||||||
|
|
||||||
/// A single BLS signature.
|
/// A BLS aggregate public key.
|
||||||
///
|
///
|
||||||
/// This struct is a wrapper upon a base type and provides helper functions (e.g., SSZ
|
/// This struct is a wrapper upon a base type and provides helper functions (e.g., SSZ
|
||||||
/// serialization).
|
/// serialization).
|
||||||
@ -17,7 +17,7 @@ impl AggregatePublicKey {
|
|||||||
self.0.add(public_key.as_raw())
|
self.0.add(public_key.as_raw())
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Returns the underlying signature.
|
/// Returns the underlying public key.
|
||||||
pub fn as_raw(&self) -> &RawAggregatePublicKey {
|
pub fn as_raw(&self) -> &RawAggregatePublicKey {
|
||||||
&self.0
|
&self.0
|
||||||
}
|
}
|
||||||
|
@ -36,6 +36,11 @@ impl AggregateSignature {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Add (aggregate) another `AggregateSignature`.
|
||||||
|
pub fn add_aggregate(&mut self, agg_signature: &AggregateSignature) {
|
||||||
|
self.0.add_aggregate(&agg_signature.0)
|
||||||
|
}
|
||||||
|
|
||||||
/// Verify the `AggregateSignature` against an `AggregatePublicKey`.
|
/// Verify the `AggregateSignature` against an `AggregatePublicKey`.
|
||||||
///
|
///
|
||||||
/// Only returns `true` if the set of keys in the `AggregatePublicKey` match the set of keys
|
/// Only returns `true` if the set of keys in the `AggregatePublicKey` match the set of keys
|
||||||
|
@ -31,6 +31,11 @@ impl FakeAggregateSignature {
|
|||||||
// Do nothing.
|
// Do nothing.
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Does glorious nothing.
|
||||||
|
pub fn add_aggregate(&mut self, _agg_sig: &FakeAggregateSignature) {
|
||||||
|
// Do nothing.
|
||||||
|
}
|
||||||
|
|
||||||
/// _Always_ returns `true`.
|
/// _Always_ returns `true`.
|
||||||
pub fn verify(
|
pub fn verify(
|
||||||
&self,
|
&self,
|
||||||
|
@ -33,9 +33,11 @@ impl BooleanBitfield {
|
|||||||
}
|
}
|
||||||
|
|
||||||
/// Create a new bitfield with the given length `initial_len` and all values set to `bit`.
|
/// Create a new bitfield with the given length `initial_len` and all values set to `bit`.
|
||||||
pub fn from_elem(inital_len: usize, bit: bool) -> Self {
|
pub fn from_elem(initial_len: usize, bit: bool) -> Self {
|
||||||
|
// BitVec can panic if we don't set the len to be a multiple of 8.
|
||||||
|
let len = ((initial_len + 7) / 8) * 8;
|
||||||
Self {
|
Self {
|
||||||
0: BitVec::from_elem(inital_len, bit),
|
0: BitVec::from_elem(len, bit),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -89,6 +91,11 @@ impl BooleanBitfield {
|
|||||||
self.len() == 0
|
self.len() == 0
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Returns true if all bits are set to 0.
|
||||||
|
pub fn is_zero(&self) -> bool {
|
||||||
|
self.0.none()
|
||||||
|
}
|
||||||
|
|
||||||
/// Returns the number of bytes required to represent this bitfield.
|
/// Returns the number of bytes required to represent this bitfield.
|
||||||
pub fn num_bytes(&self) -> usize {
|
pub fn num_bytes(&self) -> usize {
|
||||||
self.to_bytes().len()
|
self.to_bytes().len()
|
||||||
@ -104,6 +111,44 @@ impl BooleanBitfield {
|
|||||||
pub fn to_bytes(&self) -> Vec<u8> {
|
pub fn to_bytes(&self) -> Vec<u8> {
|
||||||
self.0.to_bytes()
|
self.0.to_bytes()
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/// Compute the intersection (binary-and) of this bitfield with another. Lengths must match.
|
||||||
|
pub fn intersection(&self, other: &Self) -> Self {
|
||||||
|
let mut res = self.clone();
|
||||||
|
res.intersection_inplace(other);
|
||||||
|
res
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Like `intersection` but in-place (updates `self`).
|
||||||
|
pub fn intersection_inplace(&mut self, other: &Self) {
|
||||||
|
self.0.intersect(&other.0);
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Compute the union (binary-or) of this bitfield with another. Lengths must match.
|
||||||
|
pub fn union(&self, other: &Self) -> Self {
|
||||||
|
let mut res = self.clone();
|
||||||
|
res.union_inplace(other);
|
||||||
|
res
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Like `union` but in-place (updates `self`).
|
||||||
|
pub fn union_inplace(&mut self, other: &Self) {
|
||||||
|
self.0.union(&other.0);
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Compute the difference (binary-minus) of this bitfield with another. Lengths must match.
|
||||||
|
///
|
||||||
|
/// Computes `self - other`.
|
||||||
|
pub fn difference(&self, other: &Self) -> Self {
|
||||||
|
let mut res = self.clone();
|
||||||
|
res.difference_inplace(other);
|
||||||
|
res
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Like `difference` but in-place (updates `self`).
|
||||||
|
pub fn difference_inplace(&mut self, other: &Self) {
|
||||||
|
self.0.difference(&other.0);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl default::Default for BooleanBitfield {
|
impl default::Default for BooleanBitfield {
|
||||||
@ -125,10 +170,11 @@ impl cmp::PartialEq for BooleanBitfield {
|
|||||||
/// Create a new bitfield that is a union of two other bitfields.
|
/// Create a new bitfield that is a union of two other bitfields.
|
||||||
///
|
///
|
||||||
/// For example `union(0101, 1000) == 1101`
|
/// For example `union(0101, 1000) == 1101`
|
||||||
impl std::ops::BitAnd for BooleanBitfield {
|
// TODO: length-independent intersection for BitAnd
|
||||||
|
impl std::ops::BitOr for BooleanBitfield {
|
||||||
type Output = Self;
|
type Output = Self;
|
||||||
|
|
||||||
fn bitand(self, other: Self) -> Self {
|
fn bitor(self, other: Self) -> Self {
|
||||||
let (biggest, smallest) = if self.len() > other.len() {
|
let (biggest, smallest) = if self.len() > other.len() {
|
||||||
(&self, &other)
|
(&self, &other)
|
||||||
} else {
|
} else {
|
||||||
@ -419,10 +465,59 @@ mod tests {
|
|||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn test_bitand() {
|
fn test_bitor() {
|
||||||
let a = BooleanBitfield::from_bytes(&vec![2, 8, 1][..]);
|
let a = BooleanBitfield::from_bytes(&vec![2, 8, 1][..]);
|
||||||
let b = BooleanBitfield::from_bytes(&vec![4, 8, 16][..]);
|
let b = BooleanBitfield::from_bytes(&vec![4, 8, 16][..]);
|
||||||
let c = BooleanBitfield::from_bytes(&vec![6, 8, 17][..]);
|
let c = BooleanBitfield::from_bytes(&vec![6, 8, 17][..]);
|
||||||
assert_eq!(c, a & b);
|
assert_eq!(c, a | b);
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_is_zero() {
|
||||||
|
let yes_data: &[&[u8]] = &[&[], &[0], &[0, 0], &[0, 0, 0]];
|
||||||
|
for bytes in yes_data {
|
||||||
|
assert!(BooleanBitfield::from_bytes(bytes).is_zero());
|
||||||
|
}
|
||||||
|
let no_data: &[&[u8]] = &[&[1], &[6], &[0, 1], &[0, 0, 1], &[0, 0, 255]];
|
||||||
|
for bytes in no_data {
|
||||||
|
assert!(!BooleanBitfield::from_bytes(bytes).is_zero());
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_intersection() {
|
||||||
|
let a = BooleanBitfield::from_bytes(&[0b1100, 0b0001]);
|
||||||
|
let b = BooleanBitfield::from_bytes(&[0b1011, 0b1001]);
|
||||||
|
let c = BooleanBitfield::from_bytes(&[0b1000, 0b0001]);
|
||||||
|
assert_eq!(a.intersection(&b), c);
|
||||||
|
assert_eq!(b.intersection(&a), c);
|
||||||
|
assert_eq!(a.intersection(&c), c);
|
||||||
|
assert_eq!(b.intersection(&c), c);
|
||||||
|
assert_eq!(a.intersection(&a), a);
|
||||||
|
assert_eq!(b.intersection(&b), b);
|
||||||
|
assert_eq!(c.intersection(&c), c);
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_union() {
|
||||||
|
let a = BooleanBitfield::from_bytes(&[0b1100, 0b0001]);
|
||||||
|
let b = BooleanBitfield::from_bytes(&[0b1011, 0b1001]);
|
||||||
|
let c = BooleanBitfield::from_bytes(&[0b1111, 0b1001]);
|
||||||
|
assert_eq!(a.union(&b), c);
|
||||||
|
assert_eq!(b.union(&a), c);
|
||||||
|
assert_eq!(a.union(&a), a);
|
||||||
|
assert_eq!(b.union(&b), b);
|
||||||
|
assert_eq!(c.union(&c), c);
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_difference() {
|
||||||
|
let a = BooleanBitfield::from_bytes(&[0b1100, 0b0001]);
|
||||||
|
let b = BooleanBitfield::from_bytes(&[0b1011, 0b1001]);
|
||||||
|
let a_b = BooleanBitfield::from_bytes(&[0b0100, 0b0000]);
|
||||||
|
let b_a = BooleanBitfield::from_bytes(&[0b0011, 0b1000]);
|
||||||
|
assert_eq!(a.difference(&b), a_b);
|
||||||
|
assert_eq!(b.difference(&a), b_a);
|
||||||
|
assert!(a.difference(&a).is_zero());
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
Loading…
Reference in New Issue
Block a user