75aea7054c
## Issue Addressed #4281 ## Proposed Changes - Change `ShufflingCache` implementation from using `LruCache` to a custom cache that removes entry with lowest epoch instead of oldest insertion time. - Protect the "enshrined" head shufflings when inserting new committee cache entries. The shuffling ids matching the head's previous, current, and future epochs will never be ejected from the cache during `Self::insert_cache_item`. ## Additional Info There is a bonus point on shuffling preferences in the issue description that hasn't been implemented yet, as I haven't figured out a good way to do this: > However I'm not convinced since there are some complexities around tie-breaking when two entries have the same epoch. Perhaps preferring entries in the canonical chain is best? We should be able to check if a block is on the canonical chain by: ```rust canonical_head .fork_choice_read_lock() .contains_block(root) ``` However we need to interleave the shuffling and fork choice locks, which may cause deadlocks if we're not careful (mentioned by @paulhauner). Alternatively, we could use the `state.block_roots` field of the `chain.canonical_head.snapshot.beacon_state`, which avoids deadlock but requires more work. I'd like to get some feedback on review & testing before I dig deeper into the preferences stuff, as having the canonical head preference may already be quite useful in preventing the issue raised. Co-authored-by: Jimmy Chen <jimmy@sigmaprime.io>
585 lines
22 KiB
Rust
585 lines
22 KiB
Rust
use std::collections::HashMap;
|
|
use std::sync::Arc;
|
|
|
|
use itertools::Itertools;
|
|
use slog::{debug, Logger};
|
|
|
|
use oneshot_broadcast::{oneshot, Receiver, Sender};
|
|
use types::{
|
|
beacon_state::CommitteeCache, AttestationShufflingId, BeaconState, Epoch, EthSpec, Hash256,
|
|
RelativeEpoch,
|
|
};
|
|
|
|
use crate::{metrics, BeaconChainError};
|
|
|
|
/// The size of the cache that stores committee caches for quicker verification.
|
|
///
|
|
/// Each entry should be `8 + 800,000 = 800,008` bytes in size with 100k validators. (8-byte hash +
|
|
/// 100k indices). Therefore, this cache should be approx `16 * 800,008 = 12.8 MB`. (Note: this
|
|
/// ignores a few extra bytes in the caches that should be insignificant compared to the indices).
|
|
pub const DEFAULT_CACHE_SIZE: usize = 16;
|
|
|
|
/// The maximum number of concurrent committee cache "promises" that can be issued. In effect, this
|
|
/// limits the number of concurrent states that can be loaded into memory for the committee cache.
|
|
/// This prevents excessive memory usage at the cost of rejecting some attestations.
|
|
///
|
|
/// We set this value to 2 since states can be quite large and have a significant impact on memory
|
|
/// usage. A healthy network cannot have more than a few committee caches and those caches should
|
|
/// always be inserted during block import. Unstable networks with a high degree of forking might
|
|
/// see some attestations dropped due to this concurrency limit, however I propose that this is
|
|
/// better than low-resource nodes going OOM.
|
|
const MAX_CONCURRENT_PROMISES: usize = 2;
|
|
|
|
#[derive(Clone)]
|
|
pub enum CacheItem {
|
|
/// A committee.
|
|
Committee(Arc<CommitteeCache>),
|
|
/// A promise for a future committee.
|
|
Promise(Receiver<Arc<CommitteeCache>>),
|
|
}
|
|
|
|
impl CacheItem {
|
|
pub fn is_promise(&self) -> bool {
|
|
matches!(self, CacheItem::Promise(_))
|
|
}
|
|
|
|
pub fn wait(self) -> Result<Arc<CommitteeCache>, BeaconChainError> {
|
|
match self {
|
|
CacheItem::Committee(cache) => Ok(cache),
|
|
CacheItem::Promise(receiver) => receiver
|
|
.recv()
|
|
.map_err(BeaconChainError::CommitteePromiseFailed),
|
|
}
|
|
}
|
|
}
|
|
|
|
/// Provides a cache for `CommitteeCache`.
|
|
///
|
|
/// It has been named `ShufflingCache` because `CommitteeCacheCache` is a bit weird and looks like
|
|
/// a find/replace error.
|
|
pub struct ShufflingCache {
|
|
cache: HashMap<AttestationShufflingId, CacheItem>,
|
|
cache_size: usize,
|
|
head_shuffling_ids: BlockShufflingIds,
|
|
logger: Logger,
|
|
}
|
|
|
|
impl ShufflingCache {
|
|
pub fn new(cache_size: usize, head_shuffling_ids: BlockShufflingIds, logger: Logger) -> Self {
|
|
Self {
|
|
cache: HashMap::new(),
|
|
cache_size,
|
|
head_shuffling_ids,
|
|
logger,
|
|
}
|
|
}
|
|
|
|
pub fn get(&mut self, key: &AttestationShufflingId) -> Option<CacheItem> {
|
|
match self.cache.get(key) {
|
|
// The cache contained the committee cache, return it.
|
|
item @ Some(CacheItem::Committee(_)) => {
|
|
metrics::inc_counter(&metrics::SHUFFLING_CACHE_HITS);
|
|
item.cloned()
|
|
}
|
|
// The cache contains a promise for the committee cache. Check to see if the promise has
|
|
// already been resolved, without waiting for it.
|
|
item @ Some(CacheItem::Promise(receiver)) => match receiver.try_recv() {
|
|
// The promise has already been resolved. Replace the entry in the cache with a
|
|
// `Committee` entry and then return the committee.
|
|
Ok(Some(committee)) => {
|
|
metrics::inc_counter(&metrics::SHUFFLING_CACHE_PROMISE_HITS);
|
|
metrics::inc_counter(&metrics::SHUFFLING_CACHE_HITS);
|
|
let ready = CacheItem::Committee(committee);
|
|
self.insert_cache_item(key.clone(), ready.clone());
|
|
Some(ready)
|
|
}
|
|
// The promise has not yet been resolved. Return the promise so the caller can await
|
|
// it.
|
|
Ok(None) => {
|
|
metrics::inc_counter(&metrics::SHUFFLING_CACHE_PROMISE_HITS);
|
|
metrics::inc_counter(&metrics::SHUFFLING_CACHE_HITS);
|
|
item.cloned()
|
|
}
|
|
// The sender has been dropped without sending a committee. There was most likely an
|
|
// error computing the committee cache. Drop the key from the cache and return
|
|
// `None` so the caller can recompute the committee.
|
|
//
|
|
// It's worth noting that this is the only place where we removed unresolved
|
|
// promises from the cache. This means unresolved promises will only be removed if
|
|
// we try to access them again. This is OK, since the promises don't consume much
|
|
// memory. We expect that *all* promises should be resolved, unless there is a
|
|
// programming or database error.
|
|
Err(oneshot_broadcast::Error::SenderDropped) => {
|
|
metrics::inc_counter(&metrics::SHUFFLING_CACHE_PROMISE_FAILS);
|
|
metrics::inc_counter(&metrics::SHUFFLING_CACHE_MISSES);
|
|
self.cache.remove(key);
|
|
None
|
|
}
|
|
},
|
|
// The cache does not have this committee and it's not already promised to be computed.
|
|
None => {
|
|
metrics::inc_counter(&metrics::SHUFFLING_CACHE_MISSES);
|
|
None
|
|
}
|
|
}
|
|
}
|
|
|
|
pub fn contains(&self, key: &AttestationShufflingId) -> bool {
|
|
self.cache.contains_key(key)
|
|
}
|
|
|
|
pub fn insert_committee_cache<C: ToArcCommitteeCache>(
|
|
&mut self,
|
|
key: AttestationShufflingId,
|
|
committee_cache: &C,
|
|
) {
|
|
if self
|
|
.cache
|
|
.get(&key)
|
|
// Replace the committee if it's not present or if it's a promise. A bird in the hand is
|
|
// worth two in the promise-bush!
|
|
.map_or(true, CacheItem::is_promise)
|
|
{
|
|
self.insert_cache_item(
|
|
key,
|
|
CacheItem::Committee(committee_cache.to_arc_committee_cache()),
|
|
);
|
|
}
|
|
}
|
|
|
|
/// Prunes the cache first before inserting a new cache item.
|
|
fn insert_cache_item(&mut self, key: AttestationShufflingId, cache_item: CacheItem) {
|
|
self.prune_cache();
|
|
self.cache.insert(key, cache_item);
|
|
}
|
|
|
|
/// Prunes the `cache` to keep the size below the `cache_size` limit, based on the following
|
|
/// preferences:
|
|
/// - Entries from more recent epochs are preferred over older ones.
|
|
/// - Entries with shuffling ids matching the head's previous, current, and future epochs must
|
|
/// not be pruned.
|
|
fn prune_cache(&mut self) {
|
|
let target_cache_size = self.cache_size.saturating_sub(1);
|
|
if let Some(prune_count) = self.cache.len().checked_sub(target_cache_size) {
|
|
let shuffling_ids_to_prune = self
|
|
.cache
|
|
.keys()
|
|
.sorted_by_key(|key| key.shuffling_epoch)
|
|
.filter(|shuffling_id| {
|
|
Some(shuffling_id)
|
|
!= self
|
|
.head_shuffling_ids
|
|
.id_for_epoch(shuffling_id.shuffling_epoch)
|
|
.as_ref()
|
|
.as_ref()
|
|
})
|
|
.take(prune_count)
|
|
.cloned()
|
|
.collect::<Vec<_>>();
|
|
|
|
for shuffling_id in shuffling_ids_to_prune.iter() {
|
|
debug!(
|
|
self.logger,
|
|
"Removing old shuffling from cache";
|
|
"shuffling_epoch" => shuffling_id.shuffling_epoch,
|
|
"shuffling_decision_block" => ?shuffling_id.shuffling_decision_block
|
|
);
|
|
self.cache.remove(shuffling_id);
|
|
}
|
|
}
|
|
}
|
|
|
|
pub fn create_promise(
|
|
&mut self,
|
|
key: AttestationShufflingId,
|
|
) -> Result<Sender<Arc<CommitteeCache>>, BeaconChainError> {
|
|
let num_active_promises = self
|
|
.cache
|
|
.iter()
|
|
.filter(|(_, item)| item.is_promise())
|
|
.count();
|
|
if num_active_promises >= MAX_CONCURRENT_PROMISES {
|
|
return Err(BeaconChainError::MaxCommitteePromises(num_active_promises));
|
|
}
|
|
|
|
let (sender, receiver) = oneshot();
|
|
self.insert_cache_item(key, CacheItem::Promise(receiver));
|
|
Ok(sender)
|
|
}
|
|
|
|
/// Inform the cache that the shuffling decision roots for the head has changed.
|
|
///
|
|
/// The shufflings for the head's previous, current, and future epochs will never be ejected from
|
|
/// the cache during `Self::insert_cache_item`.
|
|
pub fn update_head_shuffling_ids(&mut self, head_shuffling_ids: BlockShufflingIds) {
|
|
self.head_shuffling_ids = head_shuffling_ids;
|
|
}
|
|
}
|
|
|
|
/// A helper trait to allow lazy-cloning of the committee cache when inserting into the cache.
|
|
pub trait ToArcCommitteeCache {
|
|
fn to_arc_committee_cache(&self) -> Arc<CommitteeCache>;
|
|
}
|
|
|
|
impl ToArcCommitteeCache for CommitteeCache {
|
|
fn to_arc_committee_cache(&self) -> Arc<CommitteeCache> {
|
|
Arc::new(self.clone())
|
|
}
|
|
}
|
|
|
|
impl ToArcCommitteeCache for Arc<CommitteeCache> {
|
|
fn to_arc_committee_cache(&self) -> Arc<CommitteeCache> {
|
|
self.clone()
|
|
}
|
|
}
|
|
|
|
/// Contains the shuffling IDs for a beacon block.
|
|
#[derive(Clone)]
|
|
pub struct BlockShufflingIds {
|
|
pub current: AttestationShufflingId,
|
|
pub next: AttestationShufflingId,
|
|
pub previous: Option<AttestationShufflingId>,
|
|
pub block_root: Hash256,
|
|
}
|
|
|
|
impl BlockShufflingIds {
|
|
/// Returns the shuffling ID for the given epoch.
|
|
///
|
|
/// Returns `None` if `epoch` is prior to `self.previous?.shuffling_epoch` or
|
|
/// `self.current.shuffling_epoch` (if `previous` is `None`).
|
|
pub fn id_for_epoch(&self, epoch: Epoch) -> Option<AttestationShufflingId> {
|
|
if epoch == self.current.shuffling_epoch {
|
|
Some(self.current.clone())
|
|
} else if self
|
|
.previous
|
|
.as_ref()
|
|
.map_or(false, |id| id.shuffling_epoch == epoch)
|
|
{
|
|
self.previous.clone()
|
|
} else if epoch == self.next.shuffling_epoch {
|
|
Some(self.next.clone())
|
|
} else if epoch > self.next.shuffling_epoch {
|
|
Some(AttestationShufflingId::from_components(
|
|
epoch,
|
|
self.block_root,
|
|
))
|
|
} else {
|
|
None
|
|
}
|
|
}
|
|
|
|
pub fn try_from_head<T: EthSpec>(
|
|
head_block_root: Hash256,
|
|
head_state: &BeaconState<T>,
|
|
) -> Result<Self, String> {
|
|
let get_shuffling_id = |relative_epoch| {
|
|
AttestationShufflingId::new(head_block_root, head_state, relative_epoch).map_err(|e| {
|
|
format!(
|
|
"Unable to get attester shuffling decision slot for the epoch {:?}: {:?}",
|
|
relative_epoch, e
|
|
)
|
|
})
|
|
};
|
|
|
|
Ok(Self {
|
|
current: get_shuffling_id(RelativeEpoch::Current)?,
|
|
next: get_shuffling_id(RelativeEpoch::Next)?,
|
|
previous: Some(get_shuffling_id(RelativeEpoch::Previous)?),
|
|
block_root: head_block_root,
|
|
})
|
|
}
|
|
}
|
|
|
|
// Disable tests in debug since the beacon chain harness is slow unless in release.
|
|
#[cfg(not(debug_assertions))]
|
|
#[cfg(test)]
|
|
mod test {
|
|
use task_executor::test_utils::null_logger;
|
|
use types::*;
|
|
|
|
use crate::test_utils::EphemeralHarnessType;
|
|
|
|
use super::*;
|
|
|
|
type E = MinimalEthSpec;
|
|
type TestBeaconChainType = EphemeralHarnessType<E>;
|
|
type BeaconChainHarness = crate::test_utils::BeaconChainHarness<TestBeaconChainType>;
|
|
const TEST_CACHE_SIZE: usize = 5;
|
|
|
|
// Creates a new shuffling cache for testing
|
|
fn new_shuffling_cache() -> ShufflingCache {
|
|
let current_epoch = 8;
|
|
let head_shuffling_ids = BlockShufflingIds {
|
|
current: shuffling_id(current_epoch),
|
|
next: shuffling_id(current_epoch + 1),
|
|
previous: Some(shuffling_id(current_epoch - 1)),
|
|
block_root: Hash256::from_low_u64_le(0),
|
|
};
|
|
let logger = null_logger().unwrap();
|
|
ShufflingCache::new(TEST_CACHE_SIZE, head_shuffling_ids, logger)
|
|
}
|
|
|
|
/// Returns two different committee caches for testing.
|
|
fn committee_caches() -> (Arc<CommitteeCache>, Arc<CommitteeCache>) {
|
|
let harness = BeaconChainHarness::builder(MinimalEthSpec)
|
|
.default_spec()
|
|
.deterministic_keypairs(8)
|
|
.fresh_ephemeral_store()
|
|
.build();
|
|
let (mut state, _) = harness.get_current_state_and_root();
|
|
state
|
|
.build_committee_cache(RelativeEpoch::Current, &harness.chain.spec)
|
|
.unwrap();
|
|
state
|
|
.build_committee_cache(RelativeEpoch::Next, &harness.chain.spec)
|
|
.unwrap();
|
|
let committee_a = state
|
|
.committee_cache(RelativeEpoch::Current)
|
|
.unwrap()
|
|
.clone();
|
|
let committee_b = state.committee_cache(RelativeEpoch::Next).unwrap().clone();
|
|
assert!(committee_a != committee_b);
|
|
(Arc::new(committee_a), Arc::new(committee_b))
|
|
}
|
|
|
|
/// Builds a deterministic but incoherent shuffling ID from a `u64`.
|
|
fn shuffling_id(id: u64) -> AttestationShufflingId {
|
|
AttestationShufflingId {
|
|
shuffling_epoch: id.into(),
|
|
shuffling_decision_block: Hash256::from_low_u64_be(id),
|
|
}
|
|
}
|
|
|
|
#[test]
|
|
fn resolved_promise() {
|
|
let (committee_a, _) = committee_caches();
|
|
let id_a = shuffling_id(1);
|
|
let mut cache = new_shuffling_cache();
|
|
|
|
// Create a promise.
|
|
let sender = cache.create_promise(id_a.clone()).unwrap();
|
|
|
|
// Retrieve the newly created promise.
|
|
let item = cache.get(&id_a).unwrap();
|
|
assert!(
|
|
matches!(item, CacheItem::Promise(_)),
|
|
"the item should be a promise"
|
|
);
|
|
|
|
// Resolve the promise.
|
|
sender.send(committee_a.clone());
|
|
|
|
// Ensure the promise has been resolved.
|
|
let item = cache.get(&id_a).unwrap();
|
|
assert!(
|
|
matches!(item, CacheItem::Committee(committee) if committee == committee_a),
|
|
"the promise should be resolved"
|
|
);
|
|
assert_eq!(cache.cache.len(), 1, "the cache should have one entry");
|
|
}
|
|
|
|
#[test]
|
|
fn unresolved_promise() {
|
|
let id_a = shuffling_id(1);
|
|
let mut cache = new_shuffling_cache();
|
|
|
|
// Create a promise.
|
|
let sender = cache.create_promise(id_a.clone()).unwrap();
|
|
|
|
// Retrieve the newly created promise.
|
|
let item = cache.get(&id_a).unwrap();
|
|
assert!(
|
|
matches!(item, CacheItem::Promise(_)),
|
|
"the item should be a promise"
|
|
);
|
|
|
|
// Drop the sender without resolving the promise, simulating an error computing the
|
|
// committee.
|
|
drop(sender);
|
|
|
|
// Ensure the key now indicates an empty slot.
|
|
assert!(cache.get(&id_a).is_none(), "the slot should be empty");
|
|
assert!(cache.cache.is_empty(), "the cache should be empty");
|
|
}
|
|
|
|
#[test]
|
|
fn two_promises() {
|
|
let (committee_a, committee_b) = committee_caches();
|
|
let (id_a, id_b) = (shuffling_id(1), shuffling_id(2));
|
|
let mut cache = new_shuffling_cache();
|
|
|
|
// Create promise A.
|
|
let sender_a = cache.create_promise(id_a.clone()).unwrap();
|
|
|
|
// Retrieve promise A.
|
|
let item = cache.get(&id_a).unwrap();
|
|
assert!(
|
|
matches!(item, CacheItem::Promise(_)),
|
|
"item a should be a promise"
|
|
);
|
|
|
|
// Create promise B.
|
|
let sender_b = cache.create_promise(id_b.clone()).unwrap();
|
|
|
|
// Retrieve promise B.
|
|
let item = cache.get(&id_b).unwrap();
|
|
assert!(
|
|
matches!(item, CacheItem::Promise(_)),
|
|
"item b should be a promise"
|
|
);
|
|
|
|
// Resolve promise A.
|
|
sender_a.send(committee_a.clone());
|
|
// Ensure promise A has been resolved.
|
|
let item = cache.get(&id_a).unwrap();
|
|
assert!(
|
|
matches!(item, CacheItem::Committee(committee) if committee == committee_a),
|
|
"promise A should be resolved"
|
|
);
|
|
|
|
// Resolve promise B.
|
|
sender_b.send(committee_b.clone());
|
|
// Ensure promise B has been resolved.
|
|
let item = cache.get(&id_b).unwrap();
|
|
assert!(
|
|
matches!(item, CacheItem::Committee(committee) if committee == committee_b),
|
|
"promise B should be resolved"
|
|
);
|
|
|
|
// Check both entries again.
|
|
assert!(
|
|
matches!(cache.get(&id_a).unwrap(), CacheItem::Committee(committee) if committee == committee_a),
|
|
"promise A should remain resolved"
|
|
);
|
|
assert!(
|
|
matches!(cache.get(&id_b).unwrap(), CacheItem::Committee(committee) if committee == committee_b),
|
|
"promise B should remain resolved"
|
|
);
|
|
assert_eq!(cache.cache.len(), 2, "the cache should have two entries");
|
|
}
|
|
|
|
#[test]
|
|
fn too_many_promises() {
|
|
let mut cache = new_shuffling_cache();
|
|
|
|
for i in 0..MAX_CONCURRENT_PROMISES {
|
|
cache.create_promise(shuffling_id(i as u64)).unwrap();
|
|
}
|
|
|
|
// Ensure that the next promise returns an error. It is important for the application to
|
|
// dump his ass when he can't keep his promises, you're a queen and you deserve better.
|
|
assert!(matches!(
|
|
cache.create_promise(shuffling_id(MAX_CONCURRENT_PROMISES as u64)),
|
|
Err(BeaconChainError::MaxCommitteePromises(
|
|
MAX_CONCURRENT_PROMISES
|
|
))
|
|
));
|
|
assert_eq!(
|
|
cache.cache.len(),
|
|
MAX_CONCURRENT_PROMISES,
|
|
"the cache should have two entries"
|
|
);
|
|
}
|
|
|
|
#[test]
|
|
fn should_insert_committee_cache() {
|
|
let mut cache = new_shuffling_cache();
|
|
let id_a = shuffling_id(1);
|
|
let committee_cache_a = Arc::new(CommitteeCache::default());
|
|
cache.insert_committee_cache(id_a.clone(), &committee_cache_a);
|
|
assert!(
|
|
matches!(cache.get(&id_a).unwrap(), CacheItem::Committee(committee_cache) if committee_cache == committee_cache_a),
|
|
"should insert committee cache"
|
|
);
|
|
}
|
|
|
|
#[test]
|
|
fn should_prune_committee_cache_with_lowest_epoch() {
|
|
let mut cache = new_shuffling_cache();
|
|
let shuffling_id_and_committee_caches = (0..(TEST_CACHE_SIZE + 1))
|
|
.map(|i| (shuffling_id(i as u64), Arc::new(CommitteeCache::default())))
|
|
.collect::<Vec<_>>();
|
|
|
|
for (shuffling_id, committee_cache) in shuffling_id_and_committee_caches.iter() {
|
|
cache.insert_committee_cache(shuffling_id.clone(), committee_cache);
|
|
}
|
|
|
|
for i in 1..(TEST_CACHE_SIZE + 1) {
|
|
assert!(
|
|
cache.contains(&shuffling_id_and_committee_caches.get(i).unwrap().0),
|
|
"should contain recent epoch shuffling ids"
|
|
);
|
|
}
|
|
|
|
assert!(
|
|
!cache.contains(&shuffling_id_and_committee_caches.get(0).unwrap().0),
|
|
"should not contain oldest epoch shuffling id"
|
|
);
|
|
assert_eq!(
|
|
cache.cache.len(),
|
|
cache.cache_size,
|
|
"should limit cache size"
|
|
);
|
|
}
|
|
|
|
#[test]
|
|
fn should_retain_head_state_shufflings() {
|
|
let mut cache = new_shuffling_cache();
|
|
let current_epoch = 10;
|
|
let committee_cache = Arc::new(CommitteeCache::default());
|
|
|
|
// Insert a few entries for next the epoch with different decision roots.
|
|
for i in 0..TEST_CACHE_SIZE {
|
|
let shuffling_id = AttestationShufflingId {
|
|
shuffling_epoch: (current_epoch + 1).into(),
|
|
shuffling_decision_block: Hash256::from_low_u64_be(current_epoch + i as u64),
|
|
};
|
|
cache.insert_committee_cache(shuffling_id, &committee_cache);
|
|
}
|
|
|
|
// Now, update the head shuffling ids
|
|
let head_shuffling_ids = BlockShufflingIds {
|
|
current: shuffling_id(current_epoch),
|
|
next: shuffling_id(current_epoch + 1),
|
|
previous: Some(shuffling_id(current_epoch - 1)),
|
|
block_root: Hash256::from_low_u64_le(42),
|
|
};
|
|
cache.update_head_shuffling_ids(head_shuffling_ids.clone());
|
|
|
|
// Insert head state shuffling ids. Should not be overridden by other shuffling ids.
|
|
cache.insert_committee_cache(head_shuffling_ids.current.clone(), &committee_cache);
|
|
cache.insert_committee_cache(head_shuffling_ids.next.clone(), &committee_cache);
|
|
cache.insert_committee_cache(
|
|
head_shuffling_ids.previous.clone().unwrap(),
|
|
&committee_cache,
|
|
);
|
|
|
|
// Insert a few entries for older epochs.
|
|
for i in 0..TEST_CACHE_SIZE {
|
|
let shuffling_id = AttestationShufflingId {
|
|
shuffling_epoch: Epoch::from(i),
|
|
shuffling_decision_block: Hash256::from_low_u64_be(i as u64),
|
|
};
|
|
cache.insert_committee_cache(shuffling_id, &committee_cache);
|
|
}
|
|
|
|
assert!(
|
|
cache.contains(&head_shuffling_ids.current),
|
|
"should retain head shuffling id for the current epoch."
|
|
);
|
|
assert!(
|
|
cache.contains(&head_shuffling_ids.next),
|
|
"should retain head shuffling id for the next epoch."
|
|
);
|
|
assert!(
|
|
cache.contains(&head_shuffling_ids.previous.unwrap()),
|
|
"should retain head shuffling id for previous epoch."
|
|
);
|
|
assert_eq!(
|
|
cache.cache.len(),
|
|
cache.cache_size,
|
|
"should limit cache size"
|
|
);
|
|
}
|
|
}
|