Compare commits

...

28 Commits

Author SHA1 Message Date
f6a60be80e Release v5.1.3
-----BEGIN PGP SIGNATURE-----
 
 iQEzBAABCgAdFiEEyqmiaFUWoDD2agrYd7EwnS5U6RQFAmYFBh4ACgkQd7EwnS5U
 6RSEJgf/agMvtx7ZnhfQXA8D0XE9/R08WrDdymLAhdu8Mrv2Vl2eFc20mpsWwMXm
 85Yoq65u/a01063sjohttUlwIbJbr8OC7USrGupROYOZrdS4e96lIB3nTKHckoSR
 5xuBpXcwtaY83VZsetj1RCy7F3NunWgplXMAg43QTtilMI4m2uNxwitsn1mDNul6
 xE4AeZlumf4LsYJrI1BAc0BsGB0Z19pIVPtSzNz3L66rYtytt9KeW2xCfdA9GOoY
 9iOrxkCk2/e/qHfmld6X2W62KsGz0idX1DT7arbHLTRrzIA8L4rC/3cE/1bDyAFV
 cjIoXcxmE3eMpuc+GGfX2d9Os/TQOw==
 =cipa
 -----END PGP SIGNATURE-----

Merge tag 'v5.1.3' into stable
2024-07-19 00:32:27 +08:00
realbigsean
3058b96f25
Release v5.1.3 (#5497)
* Release v5.1.3
2024-03-28 05:22:30 +00:00
Pawan Dhananjay
c909941ca1
Bump duplicate cache time (#5493)
* Bump seen_ttl for gossipsub duplicate cache
2024-03-27 17:31:51 +00:00
realbigsean
9d24844f50
Lookup log improvements (#5491)
* log improvements
2024-03-27 13:24:04 +00:00
realbigsean
334aa2eabd
Single lookup improvements (#5488)
* Fix unexpected `UnrequestedBlobId` and `ExtraBlocksReturned` errors due to race conditions.

* Continue chain segment processing and skip any blocks that are already known, rather than returning an error.

* more de-dup checking

* ensure we don't reset `requested_ids` during rpc download

* better fix

* Merge branch 'unstable' of https://github.com/sigp/lighthouse into more-dup-lookup-fixes

* remove chain hash check

* Merge branch 'fix-block-lookup-race' of https://github.com/jimmygchen/lighthouse into sean-test-lookups

* remove block check

* add back tests

* Log and CI fixes

* undue extra check

* Merge branch 'sean-test-lookups' of https://github.com/realbigsean/lighthouse into sean-test-lookups

* log improvements

* Improve logging
2024-03-27 10:01:27 +00:00
Michael Sproul
250a5bdc27
Run fork choice after RPC blob import (#5475)
* Run fork choice after RPC blob import
2024-03-26 21:36:08 +00:00
João Oliveira
59ef564b1d
Move gossipsub into a separate crate (#5401)
* move gossipsub into a separate crate

* Merge branch 'unstable' of github.com:sigp/lighthouse into separate-gossipsub

* address review 2

* clippy beta

* update logging to log gossipsub logs
2024-03-26 03:10:59 +00:00
Akihito Nakano
8cec8a6793
Fix double counted metrics (#5476)
* Fix double counted metrics
2024-03-26 03:10:26 +00:00
Eitan Seri-Levi
e4d4e439cb
Add Capella & Deneb light client support (#4946)
* rebase and add comment

* conditional test

* test

* optimistic chould be working now

* finality should be working now

* try again

* try again

* clippy fix

* add lc bootstrap beacon api

* add lc optimistic/finality update to events

* fmt

* That error isn't occuring on my computer but I think this should fix it

* Merge branch 'unstable' into light_client_beacon_api_1

# Conflicts:
#	beacon_node/beacon_chain/src/events.rs
#	beacon_node/http_api/src/lib.rs
#	beacon_node/http_api/src/test_utils.rs
#	beacon_node/http_api/tests/main.rs
#	beacon_node/lighthouse_network/src/rpc/codec/ssz_snappy.rs
#	beacon_node/lighthouse_network/src/rpc/methods.rs
#	beacon_node/lighthouse_network/src/service/api_types.rs
#	beacon_node/network/src/beacon_processor/worker/rpc_methods.rs
#	beacon_node/tests/test.rs
#	common/eth2/src/types.rs
#	lighthouse/src/main.rs

* Add missing test file

* Update light client types to comply with Altair light client spec.

* Fix test compilation

* Merge branch 'unstable' into light_client_beacon_api_1

* Support deserializing light client structures for the Bellatrix fork

* Move `get_light_client_bootstrap` logic to `BeaconChain`. `LightClientBootstrap` API to return `ForkVersionedResponse`.

* Misc fixes.
- log cleanup
- move http_api config mutation to `config::get_config` for consistency
- fix light client API responses

* Add light client bootstrap API test and fix existing ones.

* Merge branch 'unstable' into light_client_beacon_api_1

* Fix test for `light-client-server` http api config.

* Appease clippy

* Add Altair light client SSZ tests

* Merge branch 'unstable' of https://github.com/sigp/lighthouse into light_client_beacon_api_1

* updates to light client header

* light client header from signed beacon block

* using options

* implement helper functions

* placeholder conversion from vec hash256 to exec branch

* add deneb

* using fixed vector

* remove unwraps

* by epoch

* compute merkle proof

* merkle proof

* update comments

* resolve merge conflicts

* linting

* Merge branch 'unstable' into light-client-ssz-tests

# Conflicts:
#	beacon_node/beacon_chain/src/beacon_chain.rs
#	consensus/types/src/light_client_bootstrap.rs
#	consensus/types/src/light_client_header.rs

* superstruct attempt

* superstruct changes

* lint

* altair

* update

* update

* changes to light_client_optimistic_ and finality

* merge unstable

* refactor

* resolved merge conflicts

* Merge branch 'unstable' of https://github.com/sigp/lighthouse into capella_deneb_light_client_types

* block_to_light_client_header fork aware

* fmt

* comment fix

* comment fix

* include merge fork, update deserialize_by_fork, refactor

* fmt

* pass by ref to prevent clone

* rename merkle proof fn

* add FIXME

* LightClientHeader TestRandom

* fix comments

* fork version deserialize

* merge unstable

* move fn arguments, fork name calc

* use task executor

* remove unneeded fns

* remove dead code

* add manual ssz decoding/encoding and add ssz_tests_by_fork macro

* merge deneb types with tests

* merge ssz tests, revert code deletion, cleanup

* move chainspec

* update ssz tests

* fmt

* light client ssz tests

* change to superstruct

* changes from feedback

* linting

* Merge branch 'unstable' of https://github.com/sigp/lighthouse into capella_deneb_light_client_types

* test fix

* cleanup

* Remove unused `derive`.

* Merge branch 'unstable' of https://github.com/sigp/lighthouse into capella_deneb_light_client_types

* beta compiler fix

* merge
2024-03-25 11:01:56 +00:00
chonghe
19b0db2cdf
Delete PRE_CAPELLA_ENGINE_CAPABILITIES (#5406)
* Adjust width

* Commit changes

* Delete PRE_CAPELLA_ENGINE_CAPABILITIES

* Revert "Adjust width"

This reverts commit 6fea81b897ce4a346c90902f1c95190875afc2d9.

* Revert "Commit changes"

This reverts commit d00859a63e526c791e0fec25acdaa2557da2fb95.

* Simplify

* Merge branch 'delete-pre-capella' of https://github.com/chong-he/lighthouse into delete-pre-capella
2024-03-23 21:30:36 +00:00
chonghe
a0e64d0652
Built-in documentation text width in Lighthouse book (#5394)
* Adjust width

* Commit changes

* Trigger Build
2024-03-23 20:52:09 +00:00
dknopik
0a6e4a11d7
Verify whether validators really are unknown during sync committee duty API request (#5174)
* Verify whether validators really are unknown during sync committee duty API request

* Merge branch 'unstable' into fix-4717

* Merge branch 'unstable' into fix-4717

* Merge branch 'unstable' of https://github.com/sigp/lighthouse into fix-4717
2024-03-23 00:01:11 +00:00
Daniel Ramírez-Chiquillo
035c378c61
Make sure all geth processes are killed when stopping a local testnet (#5383)
* Fix geth processes not being killed when stopping a local testnet

* Merge branch 'unstable' of https://github.com/sigp/lighthouse into fix_stop_testnet
2024-03-23 00:01:08 +00:00
Lion - dapplion
306d3eb0eb
Delete unused incomplete_processing_components (#5418)
* Delete unused incomplete_processing_components

* lint
2024-03-23 00:00:35 +00:00
Michael Sproul
332e1bb9e6
Fix one and hide all beacon-processor flags (#5397)
* Fix `beacon-processor-work-queue-len`

* Hide beacon-processor flags
2024-03-22 23:24:17 +00:00
zhiqiangxu
5121d655f7
chore: reduce scope of commitment (#5426)
* reduce scope of commitment

* avoid clone for last reference

* Merge branch 'unstable' of https://github.com/sigp/lighthouse into mod_merge_single_blob
2024-03-22 18:16:52 +00:00
zhiqiangxu
5bfe6a8ae3
chore: remove stale comment (#5440)
* rm stale comment

* Merge branch 'unstable' of https://github.com/sigp/lighthouse into rm_irelevant_comment
2024-03-22 18:16:49 +00:00
Afanti
003bb0ae3c
fix: tail command typo (#5456)
* fix: tail command typo

* Merge branch 'unstable' of https://github.com/sigp/lighthouse into patch-1
2024-03-22 18:16:46 +00:00
realbigsean
21cdc64bfe
Improve parent lookup logging (#5451)
* upgrade parent lookup result processing logs to debug, use display instead of debug for BlockError in case a blob parent unknown error is hit, add block root to BlockIsAlreadyKnown

* fix compile

* fix compile

* fix compile
2024-03-22 18:16:13 +00:00
realbigsean
5ce16192c7
Release v5.1.2 (#5453)
* Release v5.1.2
2024-03-21 18:04:00 +00:00
João Oliveira
6edf031490
disable libp2p upnp (#5449)
* disable libp2p upnp

when passing --disable-upnp cli flag
2024-03-21 01:13:03 +00:00
Michael Sproul
65a6118c53
Fix gossip verification of duplicate attester slashings (#5385)
* Fix gossip verification of duplicate attester slashings
2024-03-20 20:47:38 +00:00
João Oliveira
f33ce8cc34
fix NAT nat_open metrics report (#5427)
* fix nat reporting
2024-03-20 06:45:18 +00:00
Akihito Nakano
7117772fb3
Fix peer count metrics (#5404)
* Set the peers_per_client metrics directly, rather than using increment/decrement

* Move PEERS_CONNECTED related update to the same place

* Move PEERS_CONNECTED_MULTI related update to the same place

* Rename

* Remove unused variables
2024-03-20 06:45:16 +00:00
joao
aa592853df
Improvements and Fixes in Documentation, Including Corrected Command Usage (#4998)
* Fix typo: change 'periodical' to 'periodic' in progress updates description

* Fix wrong command in Usage section

* fix typo in Development Environment section

* Fix typo: change 'Explictly' to 'Explicitly'

* Fix typos in Lighthouse UI and Contributing sections

* Fix typo: replace 'confirms' with 'conforms' in Beacon Node API description

* fix minor typographical error: change 'advice' to 'advise' in SIGILL warning message

* Fix spelling error in Detailed Guide section

* Revert "Fix typo: change 'Explictly' to 'Explicitly'"

This reverts commit 6b0781682b4f9d129d9af48d428353f596b8b852.

* Revert "fix minor typographical error: change 'advice' to 'advise' in SIGILL warning message"

This reverts commit a4904a0afd2d531caf4585e1f0726a6dd65081ea.

* compiled

* Revert "compiled"

This reverts commit 425a553bd93af93340858050cc45041ae86b3733.

* Revert "Revert "compiled""

This reverts commit b1f871cb1bd41822f4c7bd696f77ba38cdd32f9d.

* Empty commit to trigger CI.
2024-03-20 06:44:44 +00:00
Eitan Seri-Levi
01ec42e75a
Fix Rust beta compiler errors 1.78.0-beta.1 (#5439)
* remove redundant imports

* fix test

* contains key

* fmt

* Merge branch 'unstable' into fix-beta-compiler
2024-03-20 05:17:02 +00:00
realbigsean
4449627c7c
Dedup parent blob requests (#5432)
* de dup parent blob requests

* add new line
2024-03-20 00:19:26 +00:00
chonghe
eab3672c6d
Add attestation simulator, blobs info and some updates to Lighthouse Book (#5364)
* Apply suggestions from code review

* Revise attestation simulator doc

* Revise blobs.md

* Summary

* Add blobs

* Simulator docs

* Revise attestation simulator

* minor formatting

* Revise vm node

* Update faq

* Update faq

* Add link to v4.6.0

* Remove minification in the docs

* Update Goerli to Holesky

* Add a note on moved vm validator monitor

* Update Rpi 4 note

* Revise attestation simulator doc

* Add docs for attestation simulator

* update database table

* Update faq on resources used

* Fix and update table
2024-03-14 06:12:25 +00:00
263 changed files with 2609 additions and 1395 deletions

365
Cargo.lock generated
View File

@ -505,96 +505,25 @@ dependencies = [
"futures-core", "futures-core",
] ]
[[package]]
name = "async-channel"
version = "2.2.0"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "f28243a43d821d11341ab73c80bed182dc015c514b951616cf79bd4af39af0c3"
dependencies = [
"concurrent-queue",
"event-listener 5.2.0",
"event-listener-strategy 0.5.0",
"futures-core",
"pin-project-lite",
]
[[package]]
name = "async-executor"
version = "1.8.0"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "17ae5ebefcc48e7452b4987947920dac9450be1110cadf34d1b8c116bdbaf97c"
dependencies = [
"async-lock 3.3.0",
"async-task",
"concurrent-queue",
"fastrand 2.0.1",
"futures-lite 2.2.0",
"slab",
]
[[package]]
name = "async-global-executor"
version = "2.4.1"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "05b1b633a2115cd122d73b955eadd9916c18c8f510ec9cd1686404c60ad1c29c"
dependencies = [
"async-channel 2.2.0",
"async-executor",
"async-io 2.3.1",
"async-lock 3.3.0",
"blocking",
"futures-lite 2.2.0",
"once_cell",
]
[[package]]
name = "async-io"
version = "1.13.0"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "0fc5b45d93ef0529756f812ca52e44c221b35341892d3dcc34132ac02f3dd2af"
dependencies = [
"async-lock 2.8.0",
"autocfg",
"cfg-if",
"concurrent-queue",
"futures-lite 1.13.0",
"log",
"parking",
"polling 2.8.0",
"rustix 0.37.27",
"slab",
"socket2 0.4.10",
"waker-fn",
]
[[package]] [[package]]
name = "async-io" name = "async-io"
version = "2.3.1" version = "2.3.1"
source = "registry+https://github.com/rust-lang/crates.io-index" source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "8f97ab0c5b00a7cdbe5a371b9a782ee7be1316095885c8a4ea1daf490eb0ef65" checksum = "8f97ab0c5b00a7cdbe5a371b9a782ee7be1316095885c8a4ea1daf490eb0ef65"
dependencies = [ dependencies = [
"async-lock 3.3.0", "async-lock",
"cfg-if", "cfg-if",
"concurrent-queue", "concurrent-queue",
"futures-io", "futures-io",
"futures-lite 2.2.0", "futures-lite",
"parking", "parking",
"polling 3.5.0", "polling",
"rustix 0.38.31", "rustix 0.38.31",
"slab", "slab",
"tracing", "tracing",
"windows-sys 0.52.0", "windows-sys 0.52.0",
] ]
[[package]]
name = "async-lock"
version = "2.8.0"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "287272293e9d8c41773cec55e365490fe034813a2f172f502d6ddcf75b2f582b"
dependencies = [
"event-listener 2.5.3",
]
[[package]] [[package]]
name = "async-lock" name = "async-lock"
version = "3.3.0" version = "3.3.0"
@ -602,78 +531,10 @@ source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "d034b430882f8381900d3fe6f0aaa3ad94f2cb4ac519b429692a1bc2dda4ae7b" checksum = "d034b430882f8381900d3fe6f0aaa3ad94f2cb4ac519b429692a1bc2dda4ae7b"
dependencies = [ dependencies = [
"event-listener 4.0.3", "event-listener 4.0.3",
"event-listener-strategy 0.4.0", "event-listener-strategy",
"pin-project-lite", "pin-project-lite",
] ]
[[package]]
name = "async-process"
version = "1.8.1"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "ea6438ba0a08d81529c69b36700fa2f95837bfe3e776ab39cde9c14d9149da88"
dependencies = [
"async-io 1.13.0",
"async-lock 2.8.0",
"async-signal",
"blocking",
"cfg-if",
"event-listener 3.1.0",
"futures-lite 1.13.0",
"rustix 0.38.31",
"windows-sys 0.48.0",
]
[[package]]
name = "async-signal"
version = "0.2.5"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "9e47d90f65a225c4527103a8d747001fc56e375203592b25ad103e1ca13124c5"
dependencies = [
"async-io 2.3.1",
"async-lock 2.8.0",
"atomic-waker",
"cfg-if",
"futures-core",
"futures-io",
"rustix 0.38.31",
"signal-hook-registry",
"slab",
"windows-sys 0.48.0",
]
[[package]]
name = "async-std"
version = "1.12.0"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "62565bb4402e926b29953c785397c6dc0391b7b446e45008b0049eb43cec6f5d"
dependencies = [
"async-channel 1.9.0",
"async-global-executor",
"async-io 1.13.0",
"async-lock 2.8.0",
"async-process",
"crossbeam-utils",
"futures-channel",
"futures-core",
"futures-io",
"futures-lite 1.13.0",
"gloo-timers",
"kv-log-macro",
"log",
"memchr",
"once_cell",
"pin-project-lite",
"pin-utils",
"slab",
"wasm-bindgen-futures",
]
[[package]]
name = "async-task"
version = "4.7.0"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "fbb36e985947064623dbd357f727af08ffd077f93d696782f3c56365fa2e2799"
[[package]] [[package]]
name = "async-trait" name = "async-trait"
version = "0.1.77" version = "0.1.77"
@ -722,12 +583,6 @@ dependencies = [
"pin-project-lite", "pin-project-lite",
] ]
[[package]]
name = "atomic-waker"
version = "1.1.2"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "1505bd5d3d116872e7271a6d4e16d81d0c8570876c8de68093a09ac269d8aac0"
[[package]] [[package]]
name = "attohttpc" name = "attohttpc"
version = "0.24.1" version = "0.24.1"
@ -935,7 +790,7 @@ dependencies = [
[[package]] [[package]]
name = "beacon_node" name = "beacon_node"
version = "5.1.1" version = "5.1.3"
dependencies = [ dependencies = [
"beacon_chain", "beacon_chain",
"clap", "clap",
@ -1130,22 +985,6 @@ version = "0.2.1"
source = "registry+https://github.com/rust-lang/crates.io-index" source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "8d696c370c750c948ada61c69a0ee2cbbb9c50b1019ddb86d9317157a99c2cae" checksum = "8d696c370c750c948ada61c69a0ee2cbbb9c50b1019ddb86d9317157a99c2cae"
[[package]]
name = "blocking"
version = "1.5.1"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "6a37913e8dc4ddcc604f0c6d3bf2887c995153af3611de9e23c352b44c1b9118"
dependencies = [
"async-channel 2.2.0",
"async-lock 3.3.0",
"async-task",
"fastrand 2.0.1",
"futures-io",
"futures-lite 2.2.0",
"piper",
"tracing",
]
[[package]] [[package]]
name = "bls" name = "bls"
version = "0.2.0" version = "0.2.0"
@ -1187,7 +1026,7 @@ dependencies = [
[[package]] [[package]]
name = "boot_node" name = "boot_node"
version = "5.1.1" version = "5.1.3"
dependencies = [ dependencies = [
"beacon_node", "beacon_node",
"clap", "clap",
@ -2419,7 +2258,7 @@ dependencies = [
name = "environment" name = "environment"
version = "0.1.2" version = "0.1.2"
dependencies = [ dependencies = [
"async-channel 1.9.0", "async-channel",
"ctrlc", "ctrlc",
"eth2_config", "eth2_config",
"eth2_network_config", "eth2_network_config",
@ -2921,17 +2760,6 @@ version = "2.5.3"
source = "registry+https://github.com/rust-lang/crates.io-index" source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "0206175f82b8d6bf6652ff7d71a1e27fd2e4efde587fd368662814d6ec1d9ce0" checksum = "0206175f82b8d6bf6652ff7d71a1e27fd2e4efde587fd368662814d6ec1d9ce0"
[[package]]
name = "event-listener"
version = "3.1.0"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "d93877bcde0eb80ca09131a08d23f0a5c18a620b01db137dba666d18cd9b30c2"
dependencies = [
"concurrent-queue",
"parking",
"pin-project-lite",
]
[[package]] [[package]]
name = "event-listener" name = "event-listener"
version = "4.0.3" version = "4.0.3"
@ -2943,17 +2771,6 @@ dependencies = [
"pin-project-lite", "pin-project-lite",
] ]
[[package]]
name = "event-listener"
version = "5.2.0"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "2b5fb89194fa3cad959b833185b3063ba881dbfc7030680b314250779fb4cc91"
dependencies = [
"concurrent-queue",
"parking",
"pin-project-lite",
]
[[package]] [[package]]
name = "event-listener-strategy" name = "event-listener-strategy"
version = "0.4.0" version = "0.4.0"
@ -2964,21 +2781,11 @@ dependencies = [
"pin-project-lite", "pin-project-lite",
] ]
[[package]]
name = "event-listener-strategy"
version = "0.5.0"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "feedafcaa9b749175d5ac357452a9d41ea2911da598fde46ce1fe02c37751291"
dependencies = [
"event-listener 5.2.0",
"pin-project-lite",
]
[[package]] [[package]]
name = "execution_engine_integration" name = "execution_engine_integration"
version = "0.1.0" version = "0.1.0"
dependencies = [ dependencies = [
"async-channel 1.9.0", "async-channel",
"deposit_contract", "deposit_contract",
"environment", "environment",
"ethers-core", "ethers-core",
@ -3072,15 +2879,6 @@ version = "0.1.9"
source = "registry+https://github.com/rust-lang/crates.io-index" source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "7360491ce676a36bf9bb3c56c1aa791658183a54d2744120f27285738d90465a" checksum = "7360491ce676a36bf9bb3c56c1aa791658183a54d2744120f27285738d90465a"
[[package]]
name = "fastrand"
version = "1.9.0"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "e51093e27b0797c359783294ca4f0a911c270184cb10f85783b118614a1501be"
dependencies = [
"instant",
]
[[package]] [[package]]
name = "fastrand" name = "fastrand"
version = "2.0.1" version = "2.0.1"
@ -3316,31 +3114,13 @@ version = "0.3.30"
source = "registry+https://github.com/rust-lang/crates.io-index" source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "a44623e20b9681a318efdd71c299b6b222ed6f231972bfe2f224ebad6311f0c1" checksum = "a44623e20b9681a318efdd71c299b6b222ed6f231972bfe2f224ebad6311f0c1"
[[package]]
name = "futures-lite"
version = "1.13.0"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "49a9d51ce47660b1e808d3c990b4709f2f415d928835a17dfd16991515c46bce"
dependencies = [
"fastrand 1.9.0",
"futures-core",
"futures-io",
"memchr",
"parking",
"pin-project-lite",
"waker-fn",
]
[[package]] [[package]]
name = "futures-lite" name = "futures-lite"
version = "2.2.0" version = "2.2.0"
source = "registry+https://github.com/rust-lang/crates.io-index" source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "445ba825b27408685aaecefd65178908c36c6e96aaf6d8599419d46e624192ba" checksum = "445ba825b27408685aaecefd65178908c36c6e96aaf6d8599419d46e624192ba"
dependencies = [ dependencies = [
"fastrand 2.0.1",
"futures-core", "futures-core",
"futures-io",
"parking",
"pin-project-lite", "pin-project-lite",
] ]
@ -3519,15 +3299,34 @@ source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "d2fabcfbdc87f4758337ca535fb41a6d701b65693ce38287d856d1674551ec9b" checksum = "d2fabcfbdc87f4758337ca535fb41a6d701b65693ce38287d856d1674551ec9b"
[[package]] [[package]]
name = "gloo-timers" name = "gossipsub"
version = "0.2.6" version = "0.5.0"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "9b995a66bb87bebce9a0f4a95aed01daca4872c050bfcb21653361c03bc35e5c"
dependencies = [ dependencies = [
"futures-channel", "async-channel",
"futures-core", "asynchronous-codec 0.7.0",
"js-sys", "base64 0.21.7",
"wasm-bindgen", "byteorder",
"bytes",
"either",
"fnv",
"futures",
"futures-ticker",
"futures-timer",
"getrandom",
"hex_fmt",
"instant",
"libp2p",
"prometheus-client",
"quick-protobuf",
"quick-protobuf-codec 0.3.1",
"quickcheck",
"rand",
"regex",
"serde",
"sha2 0.10.8",
"smallvec",
"tracing",
"void",
] ]
[[package]] [[package]]
@ -4129,7 +3928,7 @@ version = "3.2.0"
source = "registry+https://github.com/rust-lang/crates.io-index" source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "d6b0422c86d7ce0e97169cc42e04ae643caf278874a7a3c87b8150a220dc7e1e" checksum = "d6b0422c86d7ce0e97169cc42e04ae643caf278874a7a3c87b8150a220dc7e1e"
dependencies = [ dependencies = [
"async-io 2.3.1", "async-io",
"core-foundation", "core-foundation",
"fnv", "fnv",
"futures", "futures",
@ -4456,15 +4255,6 @@ dependencies = [
"tiny-keccak", "tiny-keccak",
] ]
[[package]]
name = "kv-log-macro"
version = "1.0.7"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "0de8b303297635ad57c9f5059fd9cee7a47f8e8daa09df0fcd07dd39fb22977f"
dependencies = [
"log",
]
[[package]] [[package]]
name = "kzg" name = "kzg"
version = "0.1.0" version = "0.1.0"
@ -4498,7 +4288,7 @@ checksum = "830d08ce1d1d941e6b30645f1a0eb5643013d835ce3779a5fc208261dbe10f55"
[[package]] [[package]]
name = "lcli" name = "lcli"
version = "5.1.1" version = "5.1.3"
dependencies = [ dependencies = [
"account_utils", "account_utils",
"beacon_chain", "beacon_chain",
@ -5074,7 +4864,7 @@ dependencies = [
[[package]] [[package]]
name = "lighthouse" name = "lighthouse"
version = "5.1.1" version = "5.1.3"
dependencies = [ dependencies = [
"account_manager", "account_manager",
"account_utils", "account_utils",
@ -5128,9 +4918,7 @@ dependencies = [
name = "lighthouse_network" name = "lighthouse_network"
version = "0.2.0" version = "0.2.0"
dependencies = [ dependencies = [
"async-channel 1.9.0", "async-channel",
"async-std",
"asynchronous-codec 0.7.0",
"base64 0.21.7", "base64 0.21.7",
"byteorder", "byteorder",
"bytes", "bytes",
@ -5145,8 +4933,8 @@ dependencies = [
"fnv", "fnv",
"futures", "futures",
"futures-ticker", "futures-ticker",
"futures-timer",
"getrandom", "getrandom",
"gossipsub",
"hex", "hex",
"hex_fmt", "hex_fmt",
"instant", "instant",
@ -5159,8 +4947,6 @@ dependencies = [
"lru_cache", "lru_cache",
"parking_lot 0.12.1", "parking_lot 0.12.1",
"prometheus-client", "prometheus-client",
"quick-protobuf",
"quick-protobuf-codec 0.3.1",
"quickcheck", "quickcheck",
"quickcheck_macros", "quickcheck_macros",
"rand", "rand",
@ -5211,12 +4997,6 @@ version = "0.1.4"
source = "registry+https://github.com/rust-lang/crates.io-index" source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "f051f77a7c8e6957c0696eac88f26b0117e54f52d3fc682ab19397a8812846a4" checksum = "f051f77a7c8e6957c0696eac88f26b0117e54f52d3fc682ab19397a8812846a4"
[[package]]
name = "linux-raw-sys"
version = "0.3.8"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "ef53942eb7bf7ff43a617b3e2c1c4a5ecf5944a7c1bc12d7ee39bbb15e5c1519"
[[package]] [[package]]
name = "linux-raw-sys" name = "linux-raw-sys"
version = "0.4.13" version = "0.4.13"
@ -5267,9 +5047,6 @@ name = "log"
version = "0.4.21" version = "0.4.21"
source = "registry+https://github.com/rust-lang/crates.io-index" source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "90ed8c1e510134f979dbc4f070f87d4313098b704861a105fe34231c70a3901c" checksum = "90ed8c1e510134f979dbc4f070f87d4313098b704861a105fe34231c70a3901c"
dependencies = [
"value-bag",
]
[[package]] [[package]]
name = "logging" name = "logging"
@ -5688,7 +5465,7 @@ name = "network"
version = "0.2.0" version = "0.2.0"
dependencies = [ dependencies = [
"anyhow", "anyhow",
"async-channel 1.9.0", "async-channel",
"beacon_chain", "beacon_chain",
"beacon_processor", "beacon_processor",
"delay_map", "delay_map",
@ -5702,6 +5479,7 @@ dependencies = [
"fnv", "fnv",
"futures", "futures",
"genesis", "genesis",
"gossipsub",
"hex", "hex",
"igd-next", "igd-next",
"itertools", "itertools",
@ -6299,17 +6077,6 @@ version = "0.1.0"
source = "registry+https://github.com/rust-lang/crates.io-index" source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "8b870d8c151b6f2fb93e84a13146138f05d02ed11c7e7c54f8826aaaf7c9f184" checksum = "8b870d8c151b6f2fb93e84a13146138f05d02ed11c7e7c54f8826aaaf7c9f184"
[[package]]
name = "piper"
version = "0.2.1"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "668d31b1c4eba19242f2088b2bf3316b82ca31082a8335764db4e083db7485d4"
dependencies = [
"atomic-waker",
"fastrand 2.0.1",
"futures-io",
]
[[package]] [[package]]
name = "pkcs8" name = "pkcs8"
version = "0.9.0" version = "0.9.0"
@ -6376,22 +6143,6 @@ dependencies = [
"plotters-backend", "plotters-backend",
] ]
[[package]]
name = "polling"
version = "2.8.0"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "4b2d323e8ca7996b3e23126511a523f7e62924d93ecd5ae73b333815b0eb3dce"
dependencies = [
"autocfg",
"bitflags 1.3.2",
"cfg-if",
"concurrent-queue",
"libc",
"log",
"pin-project-lite",
"windows-sys 0.48.0",
]
[[package]] [[package]]
name = "polling" name = "polling"
version = "3.5.0" version = "3.5.0"
@ -7243,20 +6994,6 @@ dependencies = [
"windows-sys 0.45.0", "windows-sys 0.45.0",
] ]
[[package]]
name = "rustix"
version = "0.37.27"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "fea8ca367a3a01fe35e6943c400addf443c0f57670e6ec51196f71a4b8762dd2"
dependencies = [
"bitflags 1.3.2",
"errno",
"io-lifetimes",
"libc",
"linux-raw-sys 0.3.8",
"windows-sys 0.48.0",
]
[[package]] [[package]]
name = "rustix" name = "rustix"
version = "0.38.31" version = "0.38.31"
@ -8348,7 +8085,7 @@ checksum = "c63f48baada5c52e65a29eef93ab4f8982681b67f9e8d29c7b05abcfec2b9ffe"
name = "task_executor" name = "task_executor"
version = "0.1.0" version = "0.1.0"
dependencies = [ dependencies = [
"async-channel 1.9.0", "async-channel",
"futures", "futures",
"lazy_static", "lazy_static",
"lighthouse_metrics", "lighthouse_metrics",
@ -8364,7 +8101,7 @@ source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "85b77fafb263dd9d05cbeac119526425676db3784113aa9295c88498cbf8bff1" checksum = "85b77fafb263dd9d05cbeac119526425676db3784113aa9295c88498cbf8bff1"
dependencies = [ dependencies = [
"cfg-if", "cfg-if",
"fastrand 2.0.1", "fastrand",
"rustix 0.38.31", "rustix 0.38.31",
"windows-sys 0.52.0", "windows-sys 0.52.0",
] ]
@ -9245,12 +8982,6 @@ version = "0.1.0"
source = "registry+https://github.com/rust-lang/crates.io-index" source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "830b7e5d4d90034032940e4ace0d9a9a057e7a45cd94e6c007832e39edb82f6d" checksum = "830b7e5d4d90034032940e4ace0d9a9a057e7a45cd94e6c007832e39edb82f6d"
[[package]]
name = "value-bag"
version = "1.7.0"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "126e423afe2dd9ac52142e7e9d5ce4135d7e13776c529d27fd6bc49f19e3280b"
[[package]] [[package]]
name = "vcpkg" name = "vcpkg"
version = "0.2.15" version = "0.2.15"
@ -9284,12 +9015,6 @@ dependencies = [
"libc", "libc",
] ]
[[package]]
name = "waker-fn"
version = "1.1.1"
source = "registry+https://github.com/rust-lang/crates.io-index"
checksum = "f3c4517f54858c779bbcbf228f4fca63d121bf85fbecb2dc578cdf4a39395690"
[[package]] [[package]]
name = "walkdir" name = "walkdir"
version = "2.5.0" version = "2.5.0"
@ -9515,7 +9240,7 @@ name = "web3signer_tests"
version = "0.1.0" version = "0.1.0"
dependencies = [ dependencies = [
"account_utils", "account_utils",
"async-channel 1.9.0", "async-channel",
"environment", "environment",
"eth2_keystore", "eth2_keystore",
"eth2_network_config", "eth2_network_config",

View File

@ -9,6 +9,7 @@ members = [
"beacon_node/client", "beacon_node/client",
"beacon_node/eth1", "beacon_node/eth1",
"beacon_node/lighthouse_network", "beacon_node/lighthouse_network",
"beacon_node/lighthouse_network/gossipsub",
"beacon_node/execution_layer", "beacon_node/execution_layer",
"beacon_node/http_api", "beacon_node/http_api",
"beacon_node/http_metrics", "beacon_node/http_metrics",
@ -200,6 +201,7 @@ execution_layer = { path = "beacon_node/execution_layer" }
filesystem = { path = "common/filesystem" } filesystem = { path = "common/filesystem" }
fork_choice = { path = "consensus/fork_choice" } fork_choice = { path = "consensus/fork_choice" }
genesis = { path = "beacon_node/genesis" } genesis = { path = "beacon_node/genesis" }
gossipsub = { path = "beacon_node/lighthouse_network/gossipsub/" }
http_api = { path = "beacon_node/http_api" } http_api = { path = "beacon_node/http_api" }
int_to_bytes = { path = "consensus/int_to_bytes" } int_to_bytes = { path = "consensus/int_to_bytes" }
kzg = { path = "crypto/kzg" } kzg = { path = "crypto/kzg" }

View File

@ -41,7 +41,7 @@ as the canonical staking deposit contract address.
The [Lighthouse Book](https://lighthouse-book.sigmaprime.io) contains information for users and The [Lighthouse Book](https://lighthouse-book.sigmaprime.io) contains information for users and
developers. developers.
The Lighthouse team maintains a blog at [lighthouse-blog.sigmaprime.io][blog] which contains periodical The Lighthouse team maintains a blog at [lighthouse-blog.sigmaprime.io][blog] which contains periodic
progress updates, roadmap insights and interesting findings. progress updates, roadmap insights and interesting findings.
## Branches ## Branches

View File

@ -29,6 +29,6 @@ Simply run `./account_manager generate` to generate a new random private key,
which will be automatically saved to the correct directory. which will be automatically saved to the correct directory.
If you prefer to use our "deterministic" keys for testing purposes, simply If you prefer to use our "deterministic" keys for testing purposes, simply
run `./accounts_manager generate_deterministic -i <index>`, where `index` is run `./account_manager generate_deterministic -i <index>`, where `index` is
the validator index for the key. This will reliably produce the same key each time the validator index for the key. This will reliably produce the same key each time
and save it to the directory. and save it to the directory.

View File

@ -1,6 +1,6 @@
[package] [package]
name = "beacon_node" name = "beacon_node"
version = "5.1.1" version = "5.1.3"
authors = [ authors = [
"Paul Hauner <paul@paulhauner.com>", "Paul Hauner <paul@paulhauner.com>",
"Age Manning <Age@AgeManning.com", "Age Manning <Age@AgeManning.com",

View File

@ -72,7 +72,7 @@ use crate::{
kzg_utils, metrics, AvailabilityPendingExecutedBlock, BeaconChainError, BeaconForkChoiceStore, kzg_utils, metrics, AvailabilityPendingExecutedBlock, BeaconChainError, BeaconForkChoiceStore,
BeaconSnapshot, CachedHead, BeaconSnapshot, CachedHead,
}; };
use eth2::types::{EventKind, SseBlobSidecar, SseBlock, SseExtendedPayloadAttributes, SyncDuty}; use eth2::types::{EventKind, SseBlobSidecar, SseBlock, SseExtendedPayloadAttributes};
use execution_layer::{ use execution_layer::{
BlockProposalContents, BlockProposalContentsType, BuilderParams, ChainHealth, ExecutionLayer, BlockProposalContents, BlockProposalContentsType, BuilderParams, ChainHealth, ExecutionLayer,
FailedCondition, PayloadAttributes, PayloadStatus, FailedCondition, PayloadAttributes, PayloadStatus,
@ -120,8 +120,7 @@ use store::{
use task_executor::{ShutdownReason, TaskExecutor}; use task_executor::{ShutdownReason, TaskExecutor};
use tokio_stream::Stream; use tokio_stream::Stream;
use tree_hash::TreeHash; use tree_hash::TreeHash;
use types::beacon_state::CloneConfig; use types::blob_sidecar::FixedBlobSidecarList;
use types::blob_sidecar::{BlobSidecarList, FixedBlobSidecarList};
use types::payload::BlockProductionVersion; use types::payload::BlockProductionVersion;
use types::*; use types::*;
@ -414,14 +413,14 @@ pub struct BeaconChain<T: BeaconChainTypes> {
/// Maintains a record of slashable message seen over the gossip network or RPC. /// Maintains a record of slashable message seen over the gossip network or RPC.
pub observed_slashable: RwLock<ObservedSlashable<T::EthSpec>>, pub observed_slashable: RwLock<ObservedSlashable<T::EthSpec>>,
/// Maintains a record of which validators have submitted voluntary exits. /// Maintains a record of which validators have submitted voluntary exits.
pub(crate) observed_voluntary_exits: Mutex<ObservedOperations<SignedVoluntaryExit, T::EthSpec>>, pub observed_voluntary_exits: Mutex<ObservedOperations<SignedVoluntaryExit, T::EthSpec>>,
/// Maintains a record of which validators we've seen proposer slashings for. /// Maintains a record of which validators we've seen proposer slashings for.
pub(crate) observed_proposer_slashings: Mutex<ObservedOperations<ProposerSlashing, T::EthSpec>>, pub observed_proposer_slashings: Mutex<ObservedOperations<ProposerSlashing, T::EthSpec>>,
/// Maintains a record of which validators we've seen attester slashings for. /// Maintains a record of which validators we've seen attester slashings for.
pub(crate) observed_attester_slashings: pub observed_attester_slashings:
Mutex<ObservedOperations<AttesterSlashing<T::EthSpec>, T::EthSpec>>, Mutex<ObservedOperations<AttesterSlashing<T::EthSpec>, T::EthSpec>>,
/// Maintains a record of which validators we've seen BLS to execution changes for. /// Maintains a record of which validators we've seen BLS to execution changes for.
pub(crate) observed_bls_to_execution_changes: pub observed_bls_to_execution_changes:
Mutex<ObservedOperations<SignedBlsToExecutionChange, T::EthSpec>>, Mutex<ObservedOperations<SignedBlsToExecutionChange, T::EthSpec>>,
/// Provides information from the Ethereum 1 (PoW) chain. /// Provides information from the Ethereum 1 (PoW) chain.
pub eth1_chain: Option<Eth1Chain<T::Eth1Chain, T::EthSpec>>, pub eth1_chain: Option<Eth1Chain<T::Eth1Chain, T::EthSpec>>,
@ -1348,11 +1347,12 @@ impl<T: BeaconChainTypes> BeaconChain<T> {
(parent_root, slot, sync_aggregate): LightClientProducerEvent<T::EthSpec>, (parent_root, slot, sync_aggregate): LightClientProducerEvent<T::EthSpec>,
) -> Result<(), Error> { ) -> Result<(), Error> {
self.light_client_server_cache.recompute_and_cache_updates( self.light_client_server_cache.recompute_and_cache_updates(
&self.log,
self.store.clone(), self.store.clone(),
&parent_root, &parent_root,
slot, slot,
&sync_aggregate, &sync_aggregate,
&self.log,
&self.spec,
) )
} }
@ -2567,7 +2567,7 @@ impl<T: BeaconChainTypes> BeaconChain<T> {
&self, &self,
epoch: Epoch, epoch: Epoch,
validator_indices: &[u64], validator_indices: &[u64],
) -> Result<Vec<Option<SyncDuty>>, Error> { ) -> Result<Vec<Result<Option<SyncDuty>, BeaconStateError>>, Error> {
self.with_head(move |head| { self.with_head(move |head| {
head.beacon_state head.beacon_state
.get_sync_committee_duties(epoch, validator_indices, &self.spec) .get_sync_committee_duties(epoch, validator_indices, &self.spec)
@ -2652,7 +2652,7 @@ impl<T: BeaconChainTypes> BeaconChain<T> {
// If the block is relevant, add it to the filtered chain segment. // If the block is relevant, add it to the filtered chain segment.
Ok(_) => filtered_chain_segment.push((block_root, block)), Ok(_) => filtered_chain_segment.push((block_root, block)),
// If the block is already known, simply ignore this block. // If the block is already known, simply ignore this block.
Err(BlockError::BlockIsAlreadyKnown) => continue, Err(BlockError::BlockIsAlreadyKnown(_)) => continue,
// If the block is the genesis block, simply ignore this block. // If the block is the genesis block, simply ignore this block.
Err(BlockError::GenesisBlock) => continue, Err(BlockError::GenesisBlock) => continue,
// If the block is is for a finalized slot, simply ignore this block. // If the block is is for a finalized slot, simply ignore this block.
@ -2796,6 +2796,12 @@ impl<T: BeaconChainTypes> BeaconChain<T> {
} }
} }
} }
Err(BlockError::BlockIsAlreadyKnown(block_root)) => {
debug!(self.log,
"Ignoring already known blocks while processing chain segment";
"block_root" => ?block_root);
continue;
}
Err(error) => { Err(error) => {
return ChainSegmentResult::Failed { return ChainSegmentResult::Failed {
imported_blocks, imported_blocks,
@ -2880,7 +2886,7 @@ impl<T: BeaconChainTypes> BeaconChain<T> {
.fork_choice_read_lock() .fork_choice_read_lock()
.contains_block(&block_root) .contains_block(&block_root)
{ {
return Err(BlockError::BlockIsAlreadyKnown); return Err(BlockError::BlockIsAlreadyKnown(blob.block_root()));
} }
if let Some(event_handler) = self.event_handler.as_ref() { if let Some(event_handler) = self.event_handler.as_ref() {
@ -2892,7 +2898,7 @@ impl<T: BeaconChainTypes> BeaconChain<T> {
} }
self.data_availability_checker self.data_availability_checker
.notify_gossip_blob(blob.slot(), block_root, &blob); .notify_gossip_blob(block_root, &blob);
let r = self.check_gossip_blob_availability_and_import(blob).await; let r = self.check_gossip_blob_availability_and_import(blob).await;
self.remove_notified(&block_root, r) self.remove_notified(&block_root, r)
} }
@ -2912,7 +2918,7 @@ impl<T: BeaconChainTypes> BeaconChain<T> {
.fork_choice_read_lock() .fork_choice_read_lock()
.contains_block(&block_root) .contains_block(&block_root)
{ {
return Err(BlockError::BlockIsAlreadyKnown); return Err(BlockError::BlockIsAlreadyKnown(block_root));
} }
if let Some(event_handler) = self.event_handler.as_ref() { if let Some(event_handler) = self.event_handler.as_ref() {
@ -2926,7 +2932,7 @@ impl<T: BeaconChainTypes> BeaconChain<T> {
} }
self.data_availability_checker self.data_availability_checker
.notify_rpc_blobs(slot, block_root, &blobs); .notify_rpc_blobs(block_root, &blobs);
let r = self let r = self
.check_rpc_blob_availability_and_import(slot, block_root, blobs) .check_rpc_blob_availability_and_import(slot, block_root, blobs)
.await; .await;
@ -3032,7 +3038,7 @@ impl<T: BeaconChainTypes> BeaconChain<T> {
match import_block.await { match import_block.await {
// The block was successfully verified and imported. Yay. // The block was successfully verified and imported. Yay.
Ok(status @ AvailabilityProcessingStatus::Imported(block_root)) => { Ok(status @ AvailabilityProcessingStatus::Imported(block_root)) => {
trace!( debug!(
self.log, self.log,
"Beacon block imported"; "Beacon block imported";
"block_root" => ?block_root, "block_root" => ?block_root,
@ -3045,7 +3051,7 @@ impl<T: BeaconChainTypes> BeaconChain<T> {
Ok(status) Ok(status)
} }
Ok(status @ AvailabilityProcessingStatus::MissingComponents(slot, block_root)) => { Ok(status @ AvailabilityProcessingStatus::MissingComponents(slot, block_root)) => {
trace!( debug!(
self.log, self.log,
"Beacon block awaiting blobs"; "Beacon block awaiting blobs";
"block_root" => ?block_root, "block_root" => ?block_root,
@ -6636,13 +6642,17 @@ impl<T: BeaconChainTypes> BeaconChain<T> {
&self, &self,
block_root: &Hash256, block_root: &Hash256,
) -> Result<Option<(LightClientBootstrap<T::EthSpec>, ForkName)>, Error> { ) -> Result<Option<(LightClientBootstrap<T::EthSpec>, ForkName)>, Error> {
let Some((state_root, slot)) = self let handle = self
.get_blinded_block(block_root)? .task_executor
.map(|block| (block.state_root(), block.slot())) .handle()
else { .ok_or(BeaconChainError::RuntimeShutdown)?;
let Some(block) = handle.block_on(async { self.get_block(block_root).await })? else {
return Ok(None); return Ok(None);
}; };
let (state_root, slot) = (block.state_root(), block.slot());
let Some(mut state) = self.get_state(&state_root, Some(slot))? else { let Some(mut state) = self.get_state(&state_root, Some(slot))? else {
return Ok(None); return Ok(None);
}; };
@ -6652,12 +6662,12 @@ impl<T: BeaconChainTypes> BeaconChain<T> {
.map_err(Error::InconsistentFork)?; .map_err(Error::InconsistentFork)?;
match fork_name { match fork_name {
ForkName::Altair | ForkName::Merge => { ForkName::Altair | ForkName::Merge | ForkName::Capella | ForkName::Deneb => {
LightClientBootstrap::from_beacon_state(&mut state) LightClientBootstrap::from_beacon_state(&mut state, &block, &self.spec)
.map(|bootstrap| Some((bootstrap, fork_name))) .map(|bootstrap| Some((bootstrap, fork_name)))
.map_err(Error::LightClientError) .map_err(Error::LightClientError)
} }
ForkName::Base | ForkName::Capella | ForkName::Deneb => Err(Error::UnsupportedFork), ForkName::Base => Err(Error::UnsupportedFork),
} }
} }
} }

View File

@ -190,7 +190,7 @@ pub enum BlockError<T: EthSpec> {
/// ## Peer scoring /// ## Peer scoring
/// ///
/// The block is valid and we have already imported a block with this hash. /// The block is valid and we have already imported a block with this hash.
BlockIsAlreadyKnown, BlockIsAlreadyKnown(Hash256),
/// The block slot exceeds the MAXIMUM_BLOCK_SLOT_NUMBER. /// The block slot exceeds the MAXIMUM_BLOCK_SLOT_NUMBER.
/// ///
/// ## Peer scoring /// ## Peer scoring
@ -832,7 +832,7 @@ impl<T: BeaconChainTypes> GossipVerifiedBlock<T> {
// already know this block. // already know this block.
let fork_choice_read_lock = chain.canonical_head.fork_choice_read_lock(); let fork_choice_read_lock = chain.canonical_head.fork_choice_read_lock();
if fork_choice_read_lock.contains_block(&block_root) { if fork_choice_read_lock.contains_block(&block_root) {
return Err(BlockError::BlockIsAlreadyKnown); return Err(BlockError::BlockIsAlreadyKnown(block_root));
} }
// Do not process a block that doesn't descend from the finalized root. // Do not process a block that doesn't descend from the finalized root.
@ -966,7 +966,7 @@ impl<T: BeaconChainTypes> GossipVerifiedBlock<T> {
SeenBlock::Slashable => { SeenBlock::Slashable => {
return Err(BlockError::Slashable); return Err(BlockError::Slashable);
} }
SeenBlock::Duplicate => return Err(BlockError::BlockIsAlreadyKnown), SeenBlock::Duplicate => return Err(BlockError::BlockIsAlreadyKnown(block_root)),
SeenBlock::UniqueNonSlashable => {} SeenBlock::UniqueNonSlashable => {}
}; };
@ -1784,7 +1784,7 @@ pub fn check_block_relevancy<T: BeaconChainTypes>(
.fork_choice_read_lock() .fork_choice_read_lock()
.contains_block(&block_root) .contains_block(&block_root)
{ {
return Err(BlockError::BlockIsAlreadyKnown); return Err(BlockError::BlockIsAlreadyKnown(block_root));
} }
Ok(block_root) Ok(block_root)

View File

@ -7,6 +7,7 @@ use crate::{get_block_root, GossipVerifiedBlock, PayloadVerificationOutcome};
use derivative::Derivative; use derivative::Derivative;
use ssz_types::VariableList; use ssz_types::VariableList;
use state_processing::ConsensusContext; use state_processing::ConsensusContext;
use std::fmt::{Debug, Formatter};
use std::sync::Arc; use std::sync::Arc;
use types::blob_sidecar::{BlobIdentifier, BlobSidecarError, FixedBlobSidecarList}; use types::blob_sidecar::{BlobIdentifier, BlobSidecarError, FixedBlobSidecarList};
use types::{ use types::{
@ -27,13 +28,19 @@ use types::{
/// Note: We make a distinction over blocks received over gossip because /// Note: We make a distinction over blocks received over gossip because
/// in a post-deneb world, the blobs corresponding to a given block that are received /// in a post-deneb world, the blobs corresponding to a given block that are received
/// over rpc do not contain the proposer signature for dos resistance. /// over rpc do not contain the proposer signature for dos resistance.
#[derive(Debug, Clone, Derivative)] #[derive(Clone, Derivative)]
#[derivative(Hash(bound = "E: EthSpec"))] #[derivative(Hash(bound = "E: EthSpec"))]
pub struct RpcBlock<E: EthSpec> { pub struct RpcBlock<E: EthSpec> {
block_root: Hash256, block_root: Hash256,
block: RpcBlockInner<E>, block: RpcBlockInner<E>,
} }
impl<E: EthSpec> Debug for RpcBlock<E> {
fn fmt(&self, f: &mut Formatter<'_>) -> std::fmt::Result {
write!(f, "RpcBlock({:?})", self.block_root)
}
}
impl<E: EthSpec> RpcBlock<E> { impl<E: EthSpec> RpcBlock<E> {
pub fn block_root(&self) -> Hash256 { pub fn block_root(&self) -> Hash256 {
self.block_root self.block_root

View File

@ -22,7 +22,7 @@ use std::sync::Arc;
use task_executor::TaskExecutor; use task_executor::TaskExecutor;
use types::beacon_block_body::KzgCommitmentOpts; use types::beacon_block_body::KzgCommitmentOpts;
use types::blob_sidecar::{BlobIdentifier, BlobSidecar, FixedBlobSidecarList}; use types::blob_sidecar::{BlobIdentifier, BlobSidecar, FixedBlobSidecarList};
use types::{BlobSidecarList, ChainSpec, Epoch, EthSpec, Hash256, SignedBeaconBlock, Slot}; use types::{BlobSidecarList, ChainSpec, Epoch, EthSpec, Hash256, SignedBeaconBlock};
mod availability_view; mod availability_view;
mod child_components; mod child_components;
@ -110,8 +110,6 @@ impl<T: BeaconChainTypes> DataAvailabilityChecker<T> {
self.processing_cache.read().get(&block_root).cloned() self.processing_cache.read().get(&block_root).cloned()
} }
/// A `None` indicates blobs are not required.
///
/// If there's no block, all possible ids will be returned that don't exist in the given blobs. /// If there's no block, all possible ids will be returned that don't exist in the given blobs.
/// If there no blobs, all possible ids will be returned. /// If there no blobs, all possible ids will be returned.
pub fn get_missing_blob_ids<V: AvailabilityView<T::EthSpec>>( pub fn get_missing_blob_ids<V: AvailabilityView<T::EthSpec>>(
@ -356,41 +354,30 @@ impl<T: BeaconChainTypes> DataAvailabilityChecker<T> {
/// them here is useful to avoid duplicate downloads of blocks, as well as understanding /// them here is useful to avoid duplicate downloads of blocks, as well as understanding
/// our blob download requirements. We will also serve this over RPC. /// our blob download requirements. We will also serve this over RPC.
pub fn notify_block(&self, block_root: Hash256, block: Arc<SignedBeaconBlock<T::EthSpec>>) { pub fn notify_block(&self, block_root: Hash256, block: Arc<SignedBeaconBlock<T::EthSpec>>) {
let slot = block.slot();
self.processing_cache self.processing_cache
.write() .write()
.entry(block_root) .entry(block_root)
.or_insert_with(|| ProcessingComponents::new(slot)) .or_default()
.merge_block(block); .merge_block(block);
} }
/// Add a single blob commitment to the processing cache. This commitment is unverified but caching /// Add a single blob commitment to the processing cache. This commitment is unverified but caching
/// them here is useful to avoid duplicate downloads of blobs, as well as understanding /// them here is useful to avoid duplicate downloads of blobs, as well as understanding
/// our block and blob download requirements. /// our block and blob download requirements.
pub fn notify_gossip_blob( pub fn notify_gossip_blob(&self, block_root: Hash256, blob: &GossipVerifiedBlob<T>) {
&self,
slot: Slot,
block_root: Hash256,
blob: &GossipVerifiedBlob<T>,
) {
let index = blob.index(); let index = blob.index();
let commitment = blob.kzg_commitment(); let commitment = blob.kzg_commitment();
self.processing_cache self.processing_cache
.write() .write()
.entry(block_root) .entry(block_root)
.or_insert_with(|| ProcessingComponents::new(slot)) .or_default()
.merge_single_blob(index as usize, commitment); .merge_single_blob(index as usize, commitment);
} }
/// Adds blob commitments to the processing cache. These commitments are unverified but caching /// Adds blob commitments to the processing cache. These commitments are unverified but caching
/// them here is useful to avoid duplicate downloads of blobs, as well as understanding /// them here is useful to avoid duplicate downloads of blobs, as well as understanding
/// our block and blob download requirements. /// our block and blob download requirements.
pub fn notify_rpc_blobs( pub fn notify_rpc_blobs(&self, block_root: Hash256, blobs: &FixedBlobSidecarList<T::EthSpec>) {
&self,
slot: Slot,
block_root: Hash256,
blobs: &FixedBlobSidecarList<T::EthSpec>,
) {
let mut commitments = KzgCommitmentOpts::<T::EthSpec>::default(); let mut commitments = KzgCommitmentOpts::<T::EthSpec>::default();
for blob in blobs.iter().flatten() { for blob in blobs.iter().flatten() {
if let Some(commitment) = commitments.get_mut(blob.index as usize) { if let Some(commitment) = commitments.get_mut(blob.index as usize) {
@ -400,7 +387,7 @@ impl<T: BeaconChainTypes> DataAvailabilityChecker<T> {
self.processing_cache self.processing_cache
.write() .write()
.entry(block_root) .entry(block_root)
.or_insert_with(|| ProcessingComponents::new(slot)) .or_default()
.merge_blobs(commitments); .merge_blobs(commitments);
} }
@ -409,14 +396,6 @@ impl<T: BeaconChainTypes> DataAvailabilityChecker<T> {
self.processing_cache.write().remove(block_root) self.processing_cache.write().remove(block_root)
} }
/// Gather all block roots for which we are not currently processing all components for the
/// given slot.
pub fn incomplete_processing_components(&self, slot: Slot) -> Vec<Hash256> {
self.processing_cache
.read()
.incomplete_processing_components(slot)
}
/// The epoch at which we require a data availability check in block processing. /// The epoch at which we require a data availability check in block processing.
/// `None` if the `Deneb` fork is disabled. /// `None` if the `Deneb` fork is disabled.
pub fn data_availability_boundary(&self) -> Option<Epoch> { pub fn data_availability_boundary(&self) -> Option<Epoch> {

View File

@ -108,11 +108,10 @@ pub trait AvailabilityView<E: EthSpec> {
/// 1. The blob entry at the index is empty and no block exists, or /// 1. The blob entry at the index is empty and no block exists, or
/// 2. The block exists and its commitment matches the blob's commitment. /// 2. The block exists and its commitment matches the blob's commitment.
fn merge_single_blob(&mut self, index: usize, blob: Self::BlobType) { fn merge_single_blob(&mut self, index: usize, blob: Self::BlobType) {
let commitment = *blob.get_commitment();
if let Some(cached_block) = self.get_cached_block() { if let Some(cached_block) = self.get_cached_block() {
let block_commitment_opt = cached_block.get_commitments().get(index).copied(); let block_commitment_opt = cached_block.get_commitments().get(index).copied();
if let Some(block_commitment) = block_commitment_opt { if let Some(block_commitment) = block_commitment_opt {
if block_commitment == commitment { if block_commitment == *blob.get_commitment() {
self.insert_blob_at_index(index, blob) self.insert_blob_at_index(index, blob)
} }
} }

View File

@ -780,7 +780,7 @@ mod test {
use store::{HotColdDB, ItemStore, LevelDB, StoreConfig}; use store::{HotColdDB, ItemStore, LevelDB, StoreConfig};
use tempfile::{tempdir, TempDir}; use tempfile::{tempdir, TempDir};
use types::non_zero_usize::new_non_zero_usize; use types::non_zero_usize::new_non_zero_usize;
use types::{ChainSpec, ExecPayload, MinimalEthSpec}; use types::{ExecPayload, MinimalEthSpec};
const LOW_VALIDATOR_COUNT: usize = 32; const LOW_VALIDATOR_COUNT: usize = 32;

View File

@ -3,7 +3,7 @@ use std::collections::hash_map::Entry;
use std::collections::HashMap; use std::collections::HashMap;
use std::sync::Arc; use std::sync::Arc;
use types::beacon_block_body::KzgCommitmentOpts; use types::beacon_block_body::KzgCommitmentOpts;
use types::{EthSpec, Hash256, SignedBeaconBlock, Slot}; use types::{EthSpec, Hash256, SignedBeaconBlock};
/// This cache is used only for gossip blocks/blobs and single block/blob lookups, to give req/resp /// This cache is used only for gossip blocks/blobs and single block/blob lookups, to give req/resp
/// a view of what we have and what we require. This cache serves a slightly different purpose than /// a view of what we have and what we require. This cache serves a slightly different purpose than
@ -29,23 +29,13 @@ impl<E: EthSpec> ProcessingCache<E> {
.get(block_root) .get(block_root)
.map_or(false, |b| b.block_exists()) .map_or(false, |b| b.block_exists())
} }
pub fn incomplete_processing_components(&self, slot: Slot) -> Vec<Hash256> {
let mut roots_missing_components = vec![];
for (&block_root, info) in self.processing_cache.iter() {
if info.slot == slot && !info.is_available() {
roots_missing_components.push(block_root);
}
}
roots_missing_components
}
pub fn len(&self) -> usize { pub fn len(&self) -> usize {
self.processing_cache.len() self.processing_cache.len()
} }
} }
#[derive(Debug, Clone)] #[derive(Default, Debug, Clone)]
pub struct ProcessingComponents<E: EthSpec> { pub struct ProcessingComponents<E: EthSpec> {
slot: Slot,
/// Blobs required for a block can only be known if we have seen the block. So `Some` here /// Blobs required for a block can only be known if we have seen the block. So `Some` here
/// means we've seen it, a `None` means we haven't. The `kzg_commitments` value helps us figure /// means we've seen it, a `None` means we haven't. The `kzg_commitments` value helps us figure
/// out whether incoming blobs actually match the block. /// out whether incoming blobs actually match the block.
@ -56,12 +46,8 @@ pub struct ProcessingComponents<E: EthSpec> {
} }
impl<E: EthSpec> ProcessingComponents<E> { impl<E: EthSpec> ProcessingComponents<E> {
pub fn new(slot: Slot) -> Self { pub fn new() -> Self {
Self { Self::default()
slot,
block: None,
blob_commitments: KzgCommitmentOpts::<E>::default(),
}
} }
} }
@ -70,7 +56,6 @@ impl<E: EthSpec> ProcessingComponents<E> {
impl<E: EthSpec> ProcessingComponents<E> { impl<E: EthSpec> ProcessingComponents<E> {
pub fn empty(_block_root: Hash256) -> Self { pub fn empty(_block_root: Hash256) -> Self {
Self { Self {
slot: Slot::new(0),
block: None, block: None,
blob_commitments: KzgCommitmentOpts::<E>::default(), blob_commitments: KzgCommitmentOpts::<E>::default(),
} }

View File

@ -6,7 +6,6 @@ use crate::{
use parking_lot::RwLock; use parking_lot::RwLock;
use proto_array::Block as ProtoBlock; use proto_array::Block as ProtoBlock;
use std::sync::Arc; use std::sync::Arc;
use types::blob_sidecar::BlobSidecarList;
use types::*; use types::*;
pub struct CacheItem<E: EthSpec> { pub struct CacheItem<E: EthSpec> {

View File

@ -250,6 +250,7 @@ easy_from_to!(StateAdvanceError, BeaconChainError);
easy_from_to!(BlockReplayError, BeaconChainError); easy_from_to!(BlockReplayError, BeaconChainError);
easy_from_to!(InconsistentFork, BeaconChainError); easy_from_to!(InconsistentFork, BeaconChainError);
easy_from_to!(AvailabilityCheckError, BeaconChainError); easy_from_to!(AvailabilityCheckError, BeaconChainError);
easy_from_to!(LightClientError, BeaconChainError);
#[derive(Debug)] #[derive(Debug)]
pub enum BlockProductionError { pub enum BlockProductionError {

View File

@ -9,7 +9,6 @@ use ssz_derive::{Decode, Encode};
use state_processing::per_block_processing::get_new_eth1_data; use state_processing::per_block_processing::get_new_eth1_data;
use std::cmp::Ordering; use std::cmp::Ordering;
use std::collections::HashMap; use std::collections::HashMap;
use std::iter::DoubleEndedIterator;
use std::marker::PhantomData; use std::marker::PhantomData;
use std::time::{SystemTime, UNIX_EPOCH}; use std::time::{SystemTime, UNIX_EPOCH};
use store::{DBColumn, Error as StoreError, StoreItem}; use store::{DBColumn, Error as StoreError, StoreItem};
@ -736,7 +735,7 @@ mod test {
mod eth1_chain_json_backend { mod eth1_chain_json_backend {
use super::*; use super::*;
use eth1::DepositLog; use eth1::DepositLog;
use types::{test_utils::generate_deterministic_keypair, EthSpec, MainnetEthSpec}; use types::{test_utils::generate_deterministic_keypair, MainnetEthSpec};
fn get_eth1_chain() -> Eth1Chain<CachingEth1Backend<E>, E> { fn get_eth1_chain() -> Eth1Chain<CachingEth1Backend<E>, E> {
let eth1_config = Eth1Config { let eth1_config = Eth1Config {

View File

@ -560,9 +560,6 @@ where
parent_beacon_block_root, parent_beacon_block_root,
); );
// Note: the suggested_fee_recipient is stored in the `execution_layer`, it will add this parameter.
//
// This future is not executed here, it's up to the caller to await it.
let block_contents = execution_layer let block_contents = execution_layer
.get_payload( .get_payload(
parent_hash, parent_hash,

View File

@ -48,7 +48,7 @@ impl<T: BeaconChainTypes> VerifiedLightClientFinalityUpdate<T> {
// verify that enough time has passed for the block to have been propagated // verify that enough time has passed for the block to have been propagated
let start_time = chain let start_time = chain
.slot_clock .slot_clock
.start_of(rcv_finality_update.signature_slot) .start_of(*rcv_finality_update.signature_slot())
.ok_or(Error::SigSlotStartIsNone)?; .ok_or(Error::SigSlotStartIsNone)?;
let one_third_slot_duration = Duration::new(chain.spec.seconds_per_slot / 3, 0); let one_third_slot_duration = Duration::new(chain.spec.seconds_per_slot / 3, 0);
if seen_timestamp + chain.spec.maximum_gossip_clock_disparity() if seen_timestamp + chain.spec.maximum_gossip_clock_disparity()

View File

@ -52,7 +52,7 @@ impl<T: BeaconChainTypes> VerifiedLightClientOptimisticUpdate<T> {
// verify that enough time has passed for the block to have been propagated // verify that enough time has passed for the block to have been propagated
let start_time = chain let start_time = chain
.slot_clock .slot_clock
.start_of(rcv_optimistic_update.signature_slot) .start_of(*rcv_optimistic_update.signature_slot())
.ok_or(Error::SigSlotStartIsNone)?; .ok_or(Error::SigSlotStartIsNone)?;
let one_third_slot_duration = Duration::new(chain.spec.seconds_per_slot / 3, 0); let one_third_slot_duration = Duration::new(chain.spec.seconds_per_slot / 3, 0);
if seen_timestamp + chain.spec.maximum_gossip_clock_disparity() if seen_timestamp + chain.spec.maximum_gossip_clock_disparity()
@ -65,10 +65,7 @@ impl<T: BeaconChainTypes> VerifiedLightClientOptimisticUpdate<T> {
let head_block = &head.snapshot.beacon_block; let head_block = &head.snapshot.beacon_block;
// check if we can process the optimistic update immediately // check if we can process the optimistic update immediately
// otherwise queue // otherwise queue
let canonical_root = rcv_optimistic_update let canonical_root = rcv_optimistic_update.get_canonical_root();
.attested_header
.beacon
.canonical_root();
if canonical_root != head_block.message().parent_root() { if canonical_root != head_block.message().parent_root() {
return Err(Error::UnknownBlockParentRoot(canonical_root)); return Err(Error::UnknownBlockParentRoot(canonical_root));
@ -84,7 +81,7 @@ impl<T: BeaconChainTypes> VerifiedLightClientOptimisticUpdate<T> {
return Err(Error::InvalidLightClientOptimisticUpdate); return Err(Error::InvalidLightClientOptimisticUpdate);
} }
let parent_root = rcv_optimistic_update.attested_header.beacon.parent_root; let parent_root = rcv_optimistic_update.get_parent_root();
Ok(Self { Ok(Self {
light_client_optimistic_update: rcv_optimistic_update, light_client_optimistic_update: rcv_optimistic_update,
parent_root, parent_root,

View File

@ -8,7 +8,7 @@ use types::light_client_update::{FinalizedRootProofLen, FINALIZED_ROOT_INDEX};
use types::non_zero_usize::new_non_zero_usize; use types::non_zero_usize::new_non_zero_usize;
use types::{ use types::{
BeaconBlockRef, BeaconState, ChainSpec, EthSpec, ForkName, Hash256, LightClientFinalityUpdate, BeaconBlockRef, BeaconState, ChainSpec, EthSpec, ForkName, Hash256, LightClientFinalityUpdate,
LightClientHeader, LightClientOptimisticUpdate, Slot, SyncAggregate, LightClientOptimisticUpdate, Slot, SyncAggregate,
}; };
/// A prev block cache miss requires to re-generate the state of the post-parent block. Items in the /// A prev block cache miss requires to re-generate the state of the post-parent block. Items in the
@ -71,11 +71,12 @@ impl<T: BeaconChainTypes> LightClientServerCache<T> {
/// results are cached either on disk or memory to be served via p2p and rest API /// results are cached either on disk or memory to be served via p2p and rest API
pub fn recompute_and_cache_updates( pub fn recompute_and_cache_updates(
&self, &self,
log: &Logger,
store: BeaconStore<T>, store: BeaconStore<T>,
block_parent_root: &Hash256, block_parent_root: &Hash256,
block_slot: Slot, block_slot: Slot,
sync_aggregate: &SyncAggregate<T::EthSpec>, sync_aggregate: &SyncAggregate<T::EthSpec>,
log: &Logger,
chain_spec: &ChainSpec,
) -> Result<(), BeaconChainError> { ) -> Result<(), BeaconChainError> {
let _timer = let _timer =
metrics::start_timer(&metrics::LIGHT_CLIENT_SERVER_CACHE_RECOMPUTE_UPDATES_TIMES); metrics::start_timer(&metrics::LIGHT_CLIENT_SERVER_CACHE_RECOMPUTE_UPDATES_TIMES);
@ -83,12 +84,13 @@ impl<T: BeaconChainTypes> LightClientServerCache<T> {
let signature_slot = block_slot; let signature_slot = block_slot;
let attested_block_root = block_parent_root; let attested_block_root = block_parent_root;
let attested_block = store.get_blinded_block(attested_block_root)?.ok_or( let attested_block =
BeaconChainError::DBInconsistent(format!( store
"Block not available {:?}", .get_full_block(attested_block_root)?
attested_block_root .ok_or(BeaconChainError::DBInconsistent(format!(
)), "Block not available {:?}",
)?; attested_block_root
)))?;
let cached_parts = self.get_or_compute_prev_block_cache( let cached_parts = self.get_or_compute_prev_block_cache(
store.clone(), store.clone(),
@ -109,11 +111,12 @@ impl<T: BeaconChainTypes> LightClientServerCache<T> {
}; };
if is_latest_optimistic { if is_latest_optimistic {
// can create an optimistic update, that is more recent // can create an optimistic update, that is more recent
*self.latest_optimistic_update.write() = Some(LightClientOptimisticUpdate { *self.latest_optimistic_update.write() = Some(LightClientOptimisticUpdate::new(
attested_header: block_to_light_client_header(attested_block.message()), &attested_block,
sync_aggregate: sync_aggregate.clone(), sync_aggregate.clone(),
signature_slot, signature_slot,
}); chain_spec,
)?);
}; };
// Spec: Full nodes SHOULD provide the LightClientFinalityUpdate with the highest // Spec: Full nodes SHOULD provide the LightClientFinalityUpdate with the highest
@ -127,17 +130,16 @@ impl<T: BeaconChainTypes> LightClientServerCache<T> {
if is_latest_finality & !cached_parts.finalized_block_root.is_zero() { if is_latest_finality & !cached_parts.finalized_block_root.is_zero() {
// Immediately after checkpoint sync the finalized block may not be available yet. // Immediately after checkpoint sync the finalized block may not be available yet.
if let Some(finalized_block) = if let Some(finalized_block) =
store.get_blinded_block(&cached_parts.finalized_block_root)? store.get_full_block(&cached_parts.finalized_block_root)?
{ {
*self.latest_finality_update.write() = Some(LightClientFinalityUpdate { *self.latest_finality_update.write() = Some(LightClientFinalityUpdate::new(
// TODO: may want to cache this result from latest_optimistic_update if producing a &attested_block,
// light_client header becomes expensive &finalized_block,
attested_header: block_to_light_client_header(attested_block.message()), cached_parts.finality_branch.clone(),
finalized_header: block_to_light_client_header(finalized_block.message()), sync_aggregate.clone(),
finality_branch: cached_parts.finality_branch.clone(),
sync_aggregate: sync_aggregate.clone(),
signature_slot, signature_slot,
}); chain_spec,
)?);
} else { } else {
debug!( debug!(
log, log,
@ -214,7 +216,7 @@ impl LightClientCachedData {
} }
} }
// Implements spec priorization rules: // Implements spec prioritization rules:
// > Full nodes SHOULD provide the LightClientFinalityUpdate with the highest attested_header.beacon.slot (if multiple, highest signature_slot) // > Full nodes SHOULD provide the LightClientFinalityUpdate with the highest attested_header.beacon.slot (if multiple, highest signature_slot)
// //
// ref: https://github.com/ethereum/consensus-specs/blob/113c58f9bf9c08867f6f5f633c4d98e0364d612a/specs/altair/light-client/full-node.md#create_light_client_finality_update // ref: https://github.com/ethereum/consensus-specs/blob/113c58f9bf9c08867f6f5f633c4d98e0364d612a/specs/altair/light-client/full-node.md#create_light_client_finality_update
@ -223,14 +225,15 @@ fn is_latest_finality_update<T: EthSpec>(
attested_slot: Slot, attested_slot: Slot,
signature_slot: Slot, signature_slot: Slot,
) -> bool { ) -> bool {
if attested_slot > prev.attested_header.beacon.slot { let prev_slot = prev.get_attested_header_slot();
if attested_slot > prev_slot {
true true
} else { } else {
attested_slot == prev.attested_header.beacon.slot && signature_slot > prev.signature_slot attested_slot == prev_slot && signature_slot > *prev.signature_slot()
} }
} }
// Implements spec priorization rules: // Implements spec prioritization rules:
// > Full nodes SHOULD provide the LightClientOptimisticUpdate with the highest attested_header.beacon.slot (if multiple, highest signature_slot) // > Full nodes SHOULD provide the LightClientOptimisticUpdate with the highest attested_header.beacon.slot (if multiple, highest signature_slot)
// //
// ref: https://github.com/ethereum/consensus-specs/blob/113c58f9bf9c08867f6f5f633c4d98e0364d612a/specs/altair/light-client/full-node.md#create_light_client_optimistic_update // ref: https://github.com/ethereum/consensus-specs/blob/113c58f9bf9c08867f6f5f633c4d98e0364d612a/specs/altair/light-client/full-node.md#create_light_client_optimistic_update
@ -239,18 +242,10 @@ fn is_latest_optimistic_update<T: EthSpec>(
attested_slot: Slot, attested_slot: Slot,
signature_slot: Slot, signature_slot: Slot,
) -> bool { ) -> bool {
if attested_slot > prev.attested_header.beacon.slot { let prev_slot = prev.get_slot();
if attested_slot > prev_slot {
true true
} else { } else {
attested_slot == prev.attested_header.beacon.slot && signature_slot > prev.signature_slot attested_slot == prev_slot && signature_slot > *prev.signature_slot()
}
}
fn block_to_light_client_header<T: EthSpec>(
block: BeaconBlockRef<T, types::BlindedPayload<T>>,
) -> LightClientHeader {
// TODO: make fork aware
LightClientHeader {
beacon: block.block_header(),
} }
} }

View File

@ -111,7 +111,7 @@ mod tests {
use super::*; use super::*;
use bls::Hash256; use bls::Hash256;
use std::sync::Arc; use std::sync::Arc;
use types::{BlobSidecar, MainnetEthSpec}; use types::MainnetEthSpec;
type E = MainnetEthSpec; type E = MainnetEthSpec;

View File

@ -153,6 +153,11 @@ impl<T: ObservableOperation<E>, E: EthSpec> ObservedOperations<T, E> {
self.current_fork = head_fork; self.current_fork = head_fork;
} }
} }
/// Reset the cache. MUST ONLY BE USED IN TESTS.
pub fn __reset_for_testing_only(&mut self) {
self.observed_validator_indices.clear();
}
} }
impl<T: ObservableOperation<E> + VerifyOperationAt<E>, E: EthSpec> ObservedOperations<T, E> { impl<T: ObservableOperation<E> + VerifyOperationAt<E>, E: EthSpec> ObservedOperations<T, E> {

View File

@ -367,10 +367,7 @@ impl<T: EthSpec> SnapshotCache<T> {
mod test { mod test {
use super::*; use super::*;
use crate::test_utils::{BeaconChainHarness, EphemeralHarnessType}; use crate::test_utils::{BeaconChainHarness, EphemeralHarnessType};
use types::{ use types::{test_utils::generate_deterministic_keypair, BeaconBlock, MainnetEthSpec};
test_utils::generate_deterministic_keypair, BeaconBlock, Epoch, MainnetEthSpec,
SignedBeaconBlock, Slot,
};
fn get_harness() -> BeaconChainHarness<EphemeralHarnessType<MainnetEthSpec>> { fn get_harness() -> BeaconChainHarness<EphemeralHarnessType<MainnetEthSpec>> {
let harness = BeaconChainHarness::builder(MainnetEthSpec) let harness = BeaconChainHarness::builder(MainnetEthSpec)

View File

@ -61,7 +61,6 @@ use task_executor::TaskExecutor;
use task_executor::{test_utils::TestRuntime, ShutdownReason}; use task_executor::{test_utils::TestRuntime, ShutdownReason};
use tree_hash::TreeHash; use tree_hash::TreeHash;
use types::payload::BlockProductionVersion; use types::payload::BlockProductionVersion;
use types::sync_selection_proof::SyncSelectionProof;
pub use types::test_utils::generate_deterministic_keypairs; pub use types::test_utils::generate_deterministic_keypairs;
use types::test_utils::TestRandom; use types::test_utils::TestRandom;
use types::{typenum::U4294967296, *}; use types::{typenum::U4294967296, *};

View File

@ -15,7 +15,6 @@ use state_processing::per_epoch_processing::{
errors::EpochProcessingError, EpochProcessingSummary, errors::EpochProcessingError, EpochProcessingSummary,
}; };
use std::collections::{HashMap, HashSet}; use std::collections::{HashMap, HashSet};
use std::convert::TryFrom;
use std::io; use std::io;
use std::marker::PhantomData; use std::marker::PhantomData;
use std::str::Utf8Error; use std::str::Utf8Error;

View File

@ -2,7 +2,6 @@ use crate::errors::BeaconChainError;
use crate::{BeaconChainTypes, BeaconStore}; use crate::{BeaconChainTypes, BeaconStore};
use ssz::{Decode, Encode}; use ssz::{Decode, Encode};
use std::collections::HashMap; use std::collections::HashMap;
use std::convert::TryInto;
use std::marker::PhantomData; use std::marker::PhantomData;
use store::{DBColumn, Error as StoreError, StoreItem, StoreOp}; use store::{DBColumn, Error as StoreError, StoreItem, StoreOp};
use types::{BeaconState, Hash256, PublicKey, PublicKeyBytes}; use types::{BeaconState, Hash256, PublicKey, PublicKeyBytes};
@ -195,7 +194,7 @@ mod test {
use logging::test_logger; use logging::test_logger;
use std::sync::Arc; use std::sync::Arc;
use store::HotColdDB; use store::HotColdDB;
use types::{BeaconState, EthSpec, Keypair, MainnetEthSpec}; use types::{EthSpec, Keypair, MainnetEthSpec};
type E = MainnetEthSpec; type E = MainnetEthSpec;
type T = EphemeralHarnessType<E>; type T = EphemeralHarnessType<E>;

View File

@ -1087,7 +1087,7 @@ async fn block_gossip_verification() {
assert!( assert!(
matches!( matches!(
unwrap_err(harness.chain.verify_block_for_gossip(Arc::new(block.clone())).await), unwrap_err(harness.chain.verify_block_for_gossip(Arc::new(block.clone())).await),
BlockError::BlockIsAlreadyKnown, BlockError::BlockIsAlreadyKnown(_),
), ),
"should register any valid signature against the proposer, even if the block failed later verification" "should register any valid signature against the proposer, even if the block failed later verification"
); );
@ -1115,7 +1115,7 @@ async fn block_gossip_verification() {
.verify_block_for_gossip(block.clone()) .verify_block_for_gossip(block.clone())
.await .await
.expect_err("should error when processing known block"), .expect_err("should error when processing known block"),
BlockError::BlockIsAlreadyKnown BlockError::BlockIsAlreadyKnown(_)
), ),
"the second proposal by this validator should be rejected" "the second proposal by this validator should be rejected"
); );

View File

@ -2,12 +2,18 @@
#![cfg(not(debug_assertions))] #![cfg(not(debug_assertions))]
use beacon_chain::observed_operations::ObservationOutcome; use beacon_chain::{
use beacon_chain::test_utils::{ observed_operations::ObservationOutcome,
test_spec, AttestationStrategy, BeaconChainHarness, BlockStrategy, DiskHarnessType, test_utils::{
test_spec, AttestationStrategy, BeaconChainHarness, BlockStrategy, DiskHarnessType,
},
BeaconChainError,
}; };
use lazy_static::lazy_static; use lazy_static::lazy_static;
use sloggers::{null::NullLoggerBuilder, Build}; use sloggers::{null::NullLoggerBuilder, Build};
use state_processing::per_block_processing::errors::{
AttesterSlashingInvalid, BlockOperationError, ExitInvalid, ProposerSlashingInvalid,
};
use std::sync::Arc; use std::sync::Arc;
use store::{LevelDB, StoreConfig}; use store::{LevelDB, StoreConfig};
use tempfile::{tempdir, TempDir}; use tempfile::{tempdir, TempDir};
@ -119,6 +125,75 @@ async fn voluntary_exit() {
)); ));
} }
#[tokio::test]
async fn voluntary_exit_duplicate_in_state() {
let db_path = tempdir().unwrap();
let store = get_store(&db_path);
let harness = get_harness(store.clone(), VALIDATOR_COUNT);
let spec = &harness.chain.spec;
harness
.extend_chain(
(E::slots_per_epoch() * (spec.shard_committee_period + 1)) as usize,
BlockStrategy::OnCanonicalHead,
AttestationStrategy::AllValidators,
)
.await;
harness.advance_slot();
// Exit a validator.
let exited_validator = 0;
let exit =
harness.make_voluntary_exit(exited_validator, Epoch::new(spec.shard_committee_period));
let ObservationOutcome::New(verified_exit) = harness
.chain
.verify_voluntary_exit_for_gossip(exit.clone())
.unwrap()
else {
panic!("exit should verify");
};
harness.chain.import_voluntary_exit(verified_exit);
// Make a new block to include the exit.
harness
.extend_chain(
1,
BlockStrategy::OnCanonicalHead,
AttestationStrategy::AllValidators,
)
.await;
// Verify validator is actually exited.
assert_ne!(
harness
.get_current_state()
.validators()
.get(exited_validator as usize)
.unwrap()
.exit_epoch,
spec.far_future_epoch
);
// Clear the in-memory gossip cache & try to verify the same exit on gossip.
// It should still fail because gossip verification should check the validator's `exit_epoch`
// field in the head state.
harness
.chain
.observed_voluntary_exits
.lock()
.__reset_for_testing_only();
assert!(matches!(
harness
.chain
.verify_voluntary_exit_for_gossip(exit)
.unwrap_err(),
BeaconChainError::ExitValidationError(BlockOperationError::Invalid(
ExitInvalid::AlreadyExited(index)
)) if index == exited_validator
));
}
#[test] #[test]
fn proposer_slashing() { fn proposer_slashing() {
let db_path = tempdir().unwrap(); let db_path = tempdir().unwrap();
@ -171,6 +246,63 @@ fn proposer_slashing() {
)); ));
} }
#[tokio::test]
async fn proposer_slashing_duplicate_in_state() {
let db_path = tempdir().unwrap();
let store = get_store(&db_path);
let harness = get_harness(store.clone(), VALIDATOR_COUNT);
// Slash a validator.
let slashed_validator = 0;
let slashing = harness.make_proposer_slashing(slashed_validator);
let ObservationOutcome::New(verified_slashing) = harness
.chain
.verify_proposer_slashing_for_gossip(slashing.clone())
.unwrap()
else {
panic!("slashing should verify");
};
harness.chain.import_proposer_slashing(verified_slashing);
// Make a new block to include the slashing.
harness
.extend_chain(
1,
BlockStrategy::OnCanonicalHead,
AttestationStrategy::AllValidators,
)
.await;
// Verify validator is actually slashed.
assert!(
harness
.get_current_state()
.validators()
.get(slashed_validator as usize)
.unwrap()
.slashed
);
// Clear the in-memory gossip cache & try to verify the same slashing on gossip.
// It should still fail because gossip verification should check the validator's `slashed` field
// in the head state.
harness
.chain
.observed_proposer_slashings
.lock()
.__reset_for_testing_only();
assert!(matches!(
harness
.chain
.verify_proposer_slashing_for_gossip(slashing)
.unwrap_err(),
BeaconChainError::ProposerSlashingValidationError(BlockOperationError::Invalid(
ProposerSlashingInvalid::ProposerNotSlashable(index)
)) if index == slashed_validator
));
}
#[test] #[test]
fn attester_slashing() { fn attester_slashing() {
let db_path = tempdir().unwrap(); let db_path = tempdir().unwrap();
@ -241,3 +373,60 @@ fn attester_slashing() {
ObservationOutcome::AlreadyKnown ObservationOutcome::AlreadyKnown
)); ));
} }
#[tokio::test]
async fn attester_slashing_duplicate_in_state() {
let db_path = tempdir().unwrap();
let store = get_store(&db_path);
let harness = get_harness(store.clone(), VALIDATOR_COUNT);
// Slash a validator.
let slashed_validator = 0;
let slashing = harness.make_attester_slashing(vec![slashed_validator]);
let ObservationOutcome::New(verified_slashing) = harness
.chain
.verify_attester_slashing_for_gossip(slashing.clone())
.unwrap()
else {
panic!("slashing should verify");
};
harness.chain.import_attester_slashing(verified_slashing);
// Make a new block to include the slashing.
harness
.extend_chain(
1,
BlockStrategy::OnCanonicalHead,
AttestationStrategy::AllValidators,
)
.await;
// Verify validator is actually slashed.
assert!(
harness
.get_current_state()
.validators()
.get(slashed_validator as usize)
.unwrap()
.slashed
);
// Clear the in-memory gossip cache & try to verify the same slashing on gossip.
// It should still fail because gossip verification should check the validator's `slashed` field
// in the head state.
harness
.chain
.observed_attester_slashings
.lock()
.__reset_for_testing_only();
assert!(matches!(
harness
.chain
.verify_attester_slashing_for_gossip(slashing)
.unwrap_err(),
BeaconChainError::AttesterSlashingValidationError(BlockOperationError::Invalid(
AttesterSlashingInvalid::NoSlashableIndices
))
));
}

View File

@ -284,7 +284,7 @@ pub struct BeaconProcessorChannels<E: EthSpec> {
impl<E: EthSpec> BeaconProcessorChannels<E> { impl<E: EthSpec> BeaconProcessorChannels<E> {
pub fn new(config: &BeaconProcessorConfig) -> Self { pub fn new(config: &BeaconProcessorConfig) -> Self {
let (beacon_processor_tx, beacon_processor_rx) = let (beacon_processor_tx, beacon_processor_rx) =
mpsc::channel(config.max_scheduled_work_queue_len); mpsc::channel(config.max_work_event_queue_len);
let (work_reprocessing_tx, work_reprocessing_rx) = let (work_reprocessing_tx, work_reprocessing_rx) =
mpsc::channel(config.max_scheduled_work_queue_len); mpsc::channel(config.max_scheduled_work_queue_len);

View File

@ -964,7 +964,6 @@ impl<S: SlotClock> ReprocessQueue<S> {
mod tests { mod tests {
use super::*; use super::*;
use slot_clock::TestingSlotClock; use slot_clock::TestingSlotClock;
use types::Slot;
#[test] #[test]
fn backfill_processing_schedule_calculation() { fn backfill_processing_schedule_calculation() {

View File

@ -118,7 +118,7 @@ impl Default for Config {
impl Config { impl Config {
/// Updates the data directory for the Client. /// Updates the data directory for the Client.
pub fn set_data_dir(&mut self, data_dir: PathBuf) { pub fn set_data_dir(&mut self, data_dir: PathBuf) {
self.data_dir = data_dir.clone(); self.data_dir.clone_from(&data_dir);
self.http_api.data_dir = data_dir; self.http_api.data_dir = data_dir;
} }

View File

@ -196,7 +196,6 @@ impl BlockCache {
#[cfg(test)] #[cfg(test)]
mod tests { mod tests {
use super::*; use super::*;
use types::Hash256;
fn get_block(i: u64, interval_secs: u64) -> Eth1Block { fn get_block(i: u64, interval_secs: u64) -> Eth1Block {
Eth1Block { Eth1Block {

View File

@ -99,7 +99,6 @@ async fn new_anvil_instance() -> Result<AnvilEth1Instance, String> {
mod eth1_cache { mod eth1_cache {
use super::*; use super::*;
use types::{EthSpec, MainnetEthSpec};
#[tokio::test] #[tokio::test]
async fn simple_scenario() { async fn simple_scenario() {

View File

@ -17,7 +17,6 @@ pub use json_structures::{JsonWithdrawal, TransitionConfigurationV1};
use pretty_reqwest_error::PrettyReqwestError; use pretty_reqwest_error::PrettyReqwestError;
use reqwest::StatusCode; use reqwest::StatusCode;
use serde::{Deserialize, Serialize}; use serde::{Deserialize, Serialize};
use std::convert::TryFrom;
use strum::IntoStaticStr; use strum::IntoStaticStr;
use superstruct::superstruct; use superstruct::superstruct;
pub use types::{ pub use types::{

View File

@ -11,7 +11,6 @@ use std::collections::HashSet;
use tokio::sync::Mutex; use tokio::sync::Mutex;
use std::time::{Duration, Instant}; use std::time::{Duration, Instant};
use types::EthSpec;
pub use deposit_log::{DepositLog, Log}; pub use deposit_log::{DepositLog, Log};
pub use reqwest::Client; pub use reqwest::Client;
@ -72,23 +71,6 @@ pub static LIGHTHOUSE_CAPABILITIES: &[&str] = &[
ENGINE_GET_PAYLOAD_BODIES_BY_RANGE_V1, ENGINE_GET_PAYLOAD_BODIES_BY_RANGE_V1,
]; ];
/// This is necessary because a user might run a capella-enabled version of
/// lighthouse before they update to a capella-enabled execution engine.
// TODO (mark): rip this out once we are post-capella on mainnet
pub static PRE_CAPELLA_ENGINE_CAPABILITIES: EngineCapabilities = EngineCapabilities {
new_payload_v1: true,
new_payload_v2: false,
new_payload_v3: false,
forkchoice_updated_v1: true,
forkchoice_updated_v2: false,
forkchoice_updated_v3: false,
get_payload_bodies_by_hash_v1: false,
get_payload_bodies_by_range_v1: false,
get_payload_v1: true,
get_payload_v2: false,
get_payload_v3: false,
};
/// Contains methods to convert arbitrary bytes to an ETH2 deposit contract object. /// Contains methods to convert arbitrary bytes to an ETH2 deposit contract object.
pub mod deposit_log { pub mod deposit_log {
use ssz::Decode; use ssz::Decode;
@ -1013,38 +995,29 @@ impl HttpJsonRpc {
pub async fn exchange_capabilities(&self) -> Result<EngineCapabilities, Error> { pub async fn exchange_capabilities(&self) -> Result<EngineCapabilities, Error> {
let params = json!([LIGHTHOUSE_CAPABILITIES]); let params = json!([LIGHTHOUSE_CAPABILITIES]);
let response: Result<HashSet<String>, _> = self let capabilities: HashSet<String> = self
.rpc_request( .rpc_request(
ENGINE_EXCHANGE_CAPABILITIES, ENGINE_EXCHANGE_CAPABILITIES,
params, params,
ENGINE_EXCHANGE_CAPABILITIES_TIMEOUT * self.execution_timeout_multiplier, ENGINE_EXCHANGE_CAPABILITIES_TIMEOUT * self.execution_timeout_multiplier,
) )
.await; .await?;
match response { Ok(EngineCapabilities {
// TODO (mark): rip this out once we are post capella on mainnet new_payload_v1: capabilities.contains(ENGINE_NEW_PAYLOAD_V1),
Err(error) => match error { new_payload_v2: capabilities.contains(ENGINE_NEW_PAYLOAD_V2),
Error::ServerMessage { code, message: _ } if code == METHOD_NOT_FOUND_CODE => { new_payload_v3: capabilities.contains(ENGINE_NEW_PAYLOAD_V3),
Ok(PRE_CAPELLA_ENGINE_CAPABILITIES) forkchoice_updated_v1: capabilities.contains(ENGINE_FORKCHOICE_UPDATED_V1),
} forkchoice_updated_v2: capabilities.contains(ENGINE_FORKCHOICE_UPDATED_V2),
_ => Err(error), forkchoice_updated_v3: capabilities.contains(ENGINE_FORKCHOICE_UPDATED_V3),
}, get_payload_bodies_by_hash_v1: capabilities
Ok(capabilities) => Ok(EngineCapabilities { .contains(ENGINE_GET_PAYLOAD_BODIES_BY_HASH_V1),
new_payload_v1: capabilities.contains(ENGINE_NEW_PAYLOAD_V1), get_payload_bodies_by_range_v1: capabilities
new_payload_v2: capabilities.contains(ENGINE_NEW_PAYLOAD_V2), .contains(ENGINE_GET_PAYLOAD_BODIES_BY_RANGE_V1),
new_payload_v3: capabilities.contains(ENGINE_NEW_PAYLOAD_V3), get_payload_v1: capabilities.contains(ENGINE_GET_PAYLOAD_V1),
forkchoice_updated_v1: capabilities.contains(ENGINE_FORKCHOICE_UPDATED_V1), get_payload_v2: capabilities.contains(ENGINE_GET_PAYLOAD_V2),
forkchoice_updated_v2: capabilities.contains(ENGINE_FORKCHOICE_UPDATED_V2), get_payload_v3: capabilities.contains(ENGINE_GET_PAYLOAD_V3),
forkchoice_updated_v3: capabilities.contains(ENGINE_FORKCHOICE_UPDATED_V3), })
get_payload_bodies_by_hash_v1: capabilities
.contains(ENGINE_GET_PAYLOAD_BODIES_BY_HASH_V1),
get_payload_bodies_by_range_v1: capabilities
.contains(ENGINE_GET_PAYLOAD_BODIES_BY_RANGE_V1),
get_payload_v1: capabilities.contains(ENGINE_GET_PAYLOAD_V1),
get_payload_v2: capabilities.contains(ENGINE_GET_PAYLOAD_V2),
get_payload_v3: capabilities.contains(ENGINE_GET_PAYLOAD_V3),
}),
}
} }
pub async fn clear_exchange_capabilties_cache(&self) { pub async fn clear_exchange_capabilties_cache(&self) {
@ -1191,7 +1164,7 @@ mod test {
use std::future::Future; use std::future::Future;
use std::str::FromStr; use std::str::FromStr;
use std::sync::Arc; use std::sync::Arc;
use types::{ExecutionPayloadMerge, MainnetEthSpec, Transactions, Unsigned, VariableList}; use types::{MainnetEthSpec, Unsigned};
struct Tester { struct Tester {
server: MockServer<MainnetEthSpec>, server: MockServer<MainnetEthSpec>,

View File

@ -4,10 +4,7 @@ use strum::EnumString;
use superstruct::superstruct; use superstruct::superstruct;
use types::beacon_block_body::KzgCommitments; use types::beacon_block_body::KzgCommitments;
use types::blob_sidecar::BlobsList; use types::blob_sidecar::BlobsList;
use types::{ use types::{FixedVector, Unsigned};
EthSpec, ExecutionBlockHash, ExecutionPayload, ExecutionPayloadCapella, ExecutionPayloadDeneb,
ExecutionPayloadMerge, FixedVector, Transactions, Unsigned, VariableList, Withdrawal,
};
#[derive(Debug, PartialEq, Serialize, Deserialize)] #[derive(Debug, PartialEq, Serialize, Deserialize)]
#[serde(rename_all = "camelCase")] #[serde(rename_all = "camelCase")]

View File

@ -5,7 +5,6 @@ use crate::test_utils::DEFAULT_MOCK_EL_PAYLOAD_VALUE_WEI;
use serde::{de::DeserializeOwned, Deserialize}; use serde::{de::DeserializeOwned, Deserialize};
use serde_json::Value as JsonValue; use serde_json::Value as JsonValue;
use std::sync::Arc; use std::sync::Arc;
use types::{EthSpec, ForkName};
pub const GENERIC_ERROR_CODE: i64 = -1234; pub const GENERIC_ERROR_CODE: i64 = -1234;
pub const BAD_PARAMS_ERROR_CODE: i64 = -32602; pub const BAD_PARAMS_ERROR_CODE: i64 = -32602;

View File

@ -71,8 +71,6 @@ pub trait BidStuff<E: EthSpec> {
fn set_withdrawals_root(&mut self, withdrawals_root: Hash256); fn set_withdrawals_root(&mut self, withdrawals_root: Hash256);
fn sign_builder_message(&mut self, sk: &SecretKey, spec: &ChainSpec) -> Signature; fn sign_builder_message(&mut self, sk: &SecretKey, spec: &ChainSpec) -> Signature;
fn to_signed_bid(self, signature: Signature) -> SignedBuilderBid<E>;
} }
impl<E: EthSpec> BidStuff<E> for BuilderBid<E> { impl<E: EthSpec> BidStuff<E> for BuilderBid<E> {
@ -183,13 +181,6 @@ impl<E: EthSpec> BidStuff<E> for BuilderBid<E> {
let message = self.signing_root(domain); let message = self.signing_root(domain);
sk.sign(message) sk.sign(message)
} }
fn to_signed_bid(self, signature: Signature) -> SignedBuilderBid<E> {
SignedBuilderBid {
message: self,
signature,
}
}
} }
#[derive(Clone)] #[derive(Clone)]

View File

@ -2,14 +2,12 @@ use crate::{
test_utils::{ test_utils::{
MockServer, DEFAULT_JWT_SECRET, DEFAULT_TERMINAL_BLOCK, DEFAULT_TERMINAL_DIFFICULTY, MockServer, DEFAULT_JWT_SECRET, DEFAULT_TERMINAL_BLOCK, DEFAULT_TERMINAL_DIFFICULTY,
}, },
Config, *, *,
}; };
use keccak_hash::H256; use keccak_hash::H256;
use kzg::Kzg; use kzg::Kzg;
use sensitive_url::SensitiveUrl;
use task_executor::TaskExecutor;
use tempfile::NamedTempFile; use tempfile::NamedTempFile;
use types::{Address, ChainSpec, Epoch, EthSpec, Hash256, MainnetEthSpec}; use types::MainnetEthSpec;
pub struct MockExecutionLayer<T: EthSpec> { pub struct MockExecutionLayer<T: EthSpec> {
pub server: MockServer<T>, pub server: MockServer<T>,

View File

@ -137,7 +137,7 @@ pub fn interop_genesis_state_with_withdrawal_credentials<T: EthSpec>(
#[cfg(test)] #[cfg(test)]
mod test { mod test {
use super::*; use super::*;
use types::{test_utils::generate_deterministic_keypairs, EthSpec, MinimalEthSpec}; use types::{test_utils::generate_deterministic_keypairs, MinimalEthSpec};
type TestEthSpec = MinimalEthSpec; type TestEthSpec = MinimalEthSpec;

View File

@ -132,7 +132,7 @@ impl<T: EthSpec> PackingEfficiencyHandler<T> {
} }
// Remove duplicate attestations as these yield no reward. // Remove duplicate attestations as these yield no reward.
attestations_in_block.retain(|x, _| self.included_attestations.get(x).is_none()); attestations_in_block.retain(|x, _| !self.included_attestations.contains_key(x));
self.included_attestations self.included_attestations
.extend(attestations_in_block.clone()); .extend(attestations_in_block.clone());
@ -179,8 +179,9 @@ impl<T: EthSpec> PackingEfficiencyHandler<T> {
.collect::<Vec<_>>() .collect::<Vec<_>>()
}; };
self.committee_store.previous_epoch_committees = self.committee_store
self.committee_store.current_epoch_committees.clone(); .previous_epoch_committees
.clone_from(&self.committee_store.current_epoch_committees);
self.committee_store.current_epoch_committees = new_committees; self.committee_store.current_epoch_committees = new_committees;

View File

@ -2337,7 +2337,7 @@ pub fn serve<T: BeaconChainTypes>(
let fork_name = chain let fork_name = chain
.spec .spec
.fork_name_at_slot::<T::EthSpec>(update.signature_slot); .fork_name_at_slot::<T::EthSpec>(*update.signature_slot());
match accept_header { match accept_header {
Some(api_types::Accept::Ssz) => Response::builder() Some(api_types::Accept::Ssz) => Response::builder()
.status(200) .status(200)
@ -2384,7 +2384,7 @@ pub fn serve<T: BeaconChainTypes>(
let fork_name = chain let fork_name = chain
.spec .spec
.fork_name_at_slot::<T::EthSpec>(update.signature_slot); .fork_name_at_slot::<T::EthSpec>(*update.signature_slot());
match accept_header { match accept_header {
Some(api_types::Accept::Ssz) => Response::builder() Some(api_types::Accept::Ssz) => Response::builder()
.status(200) .status(200)

View File

@ -82,11 +82,11 @@ pub async fn publish_block<T: BeaconChainTypes, B: IntoGossipVerifiedBlockConten
| SignedBeaconBlock::Altair(_) | SignedBeaconBlock::Altair(_)
| SignedBeaconBlock::Merge(_) | SignedBeaconBlock::Merge(_)
| SignedBeaconBlock::Capella(_) => { | SignedBeaconBlock::Capella(_) => {
crate::publish_pubsub_message(&sender, PubsubMessage::BeaconBlock(block.clone())) crate::publish_pubsub_message(&sender, PubsubMessage::BeaconBlock(block))
.map_err(|_| BlockError::BeaconChainError(BeaconChainError::UnableToPublish))?; .map_err(|_| BlockError::BeaconChainError(BeaconChainError::UnableToPublish))?;
} }
SignedBeaconBlock::Deneb(_) => { SignedBeaconBlock::Deneb(_) => {
let mut pubsub_messages = vec![PubsubMessage::BeaconBlock(block.clone())]; let mut pubsub_messages = vec![PubsubMessage::BeaconBlock(block)];
if let Some(blob_sidecars) = blobs_opt { if let Some(blob_sidecars) = blobs_opt {
for (blob_index, blob) in blob_sidecars.into_iter().enumerate() { for (blob_index, blob) in blob_sidecars.into_iter().enumerate() {
pubsub_messages.push(PubsubMessage::BlobSidecar(Box::new(( pubsub_messages.push(PubsubMessage::BlobSidecar(Box::new((
@ -113,7 +113,7 @@ pub async fn publish_block<T: BeaconChainTypes, B: IntoGossipVerifiedBlockConten
let (gossip_verified_block, gossip_verified_blobs) = let (gossip_verified_block, gossip_verified_blobs) =
match block_contents.into_gossip_verified_block(&chain) { match block_contents.into_gossip_verified_block(&chain) {
Ok(b) => b, Ok(b) => b,
Err(BlockContentsError::BlockError(BlockError::BlockIsAlreadyKnown)) Err(BlockContentsError::BlockError(BlockError::BlockIsAlreadyKnown(_)))
| Err(BlockContentsError::BlobError( | Err(BlockContentsError::BlobError(
beacon_chain::blob_verification::GossipBlobError::RepeatBlob { .. }, beacon_chain::blob_verification::GossipBlobError::RepeatBlob { .. },
)) => { )) => {
@ -133,7 +133,7 @@ pub async fn publish_block<T: BeaconChainTypes, B: IntoGossipVerifiedBlockConten
log, log,
"Not publishing block - not gossip verified"; "Not publishing block - not gossip verified";
"slot" => slot, "slot" => slot,
"error" => ?e "error" => %e
); );
return Err(warp_utils::reject::custom_bad_request(e.to_string())); return Err(warp_utils::reject::custom_bad_request(e.to_string()));
} }

View File

@ -45,7 +45,12 @@ pub fn sync_committee_duties<T: BeaconChainTypes>(
// the vast majority of requests. Rather than checking if we think the request will succeed in a // the vast majority of requests. Rather than checking if we think the request will succeed in a
// way prone to data races, we attempt the request immediately and check the error code. // way prone to data races, we attempt the request immediately and check the error code.
match chain.sync_committee_duties_from_head(request_epoch, request_indices) { match chain.sync_committee_duties_from_head(request_epoch, request_indices) {
Ok(duties) => return Ok(convert_to_response(duties, execution_optimistic)), Ok(duties) => {
return Ok(convert_to_response(
verify_unknown_validators(duties, request_epoch, chain)?,
execution_optimistic,
))
}
Err(BeaconChainError::SyncDutiesError(BeaconStateError::SyncCommitteeNotKnown { Err(BeaconChainError::SyncDutiesError(BeaconStateError::SyncCommitteeNotKnown {
.. ..
})) }))
@ -64,7 +69,10 @@ pub fn sync_committee_duties<T: BeaconChainTypes>(
)), )),
e => warp_utils::reject::beacon_chain_error(e), e => warp_utils::reject::beacon_chain_error(e),
})?; })?;
Ok(convert_to_response(duties, execution_optimistic)) Ok(convert_to_response(
verify_unknown_validators(duties, request_epoch, chain)?,
execution_optimistic,
))
} }
/// Slow path for duties: load a state and use it to compute the duties. /// Slow path for duties: load a state and use it to compute the duties.
@ -73,7 +81,7 @@ fn duties_from_state_load<T: BeaconChainTypes>(
request_indices: &[u64], request_indices: &[u64],
altair_fork_epoch: Epoch, altair_fork_epoch: Epoch,
chain: &BeaconChain<T>, chain: &BeaconChain<T>,
) -> Result<Vec<Option<SyncDuty>>, BeaconChainError> { ) -> Result<Vec<Result<Option<SyncDuty>, BeaconStateError>>, BeaconChainError> {
// Determine what the current epoch would be if we fast-forward our system clock by // Determine what the current epoch would be if we fast-forward our system clock by
// `MAXIMUM_GOSSIP_CLOCK_DISPARITY`. // `MAXIMUM_GOSSIP_CLOCK_DISPARITY`.
// //
@ -121,6 +129,45 @@ fn duties_from_state_load<T: BeaconChainTypes>(
} }
} }
fn verify_unknown_validators<T: BeaconChainTypes>(
duties: Vec<Result<Option<SyncDuty>, BeaconStateError>>,
request_epoch: Epoch,
chain: &BeaconChain<T>,
) -> Result<Vec<Option<SyncDuty>>, warp::reject::Rejection> {
// Lazily load the request_epoch_state, as it is only needed if there are any UnknownValidator
let mut request_epoch_state = None;
duties
.into_iter()
.map(|res| {
res.or_else(|err| {
// Make sure the validator is really unknown w.r.t. the request_epoch
if let BeaconStateError::UnknownValidator(idx) = err {
let request_epoch_state = match &mut request_epoch_state {
Some(state) => state,
None => request_epoch_state.insert(chain.state_at_slot(
request_epoch.start_slot(T::EthSpec::slots_per_epoch()),
StateSkipConfig::WithoutStateRoots,
)?),
};
request_epoch_state
.get_validator(idx)
.map_err(BeaconChainError::SyncDutiesError)
.map(|_| None)
} else {
Err(BeaconChainError::SyncDutiesError(err))
}
})
})
.collect::<Result<Vec<_>, _>>()
.map_err(|err| match err {
BeaconChainError::SyncDutiesError(BeaconStateError::UnknownValidator(idx)) => {
warp_utils::reject::custom_bad_request(format!("invalid validator index: {idx}"))
}
e => warp_utils::reject::beacon_chain_error(e),
})
}
fn convert_to_response(duties: Vec<Option<SyncDuty>>, execution_optimistic: bool) -> SyncDuties { fn convert_to_response(duties: Vec<Option<SyncDuty>>, execution_optimistic: bool) -> SyncDuties {
api_types::GenericResponse::from(duties.into_iter().flatten().collect::<Vec<_>>()) api_types::GenericResponse::from(duties.into_iter().flatten().collect::<Vec<_>>())
.add_execution_optimistic(execution_optimistic) .add_execution_optimistic(execution_optimistic)

View File

@ -1717,7 +1717,7 @@ impl ApiTester {
}; };
let expected = block.slot(); let expected = block.slot();
assert_eq!(result.header.beacon.slot, expected); assert_eq!(result.get_slot(), expected);
self self
} }

View File

@ -5,8 +5,8 @@ authors = ["Sigma Prime <contact@sigmaprime.io>"]
edition = { workspace = true } edition = { workspace = true }
[dependencies] [dependencies]
async-channel = { workspace = true }
discv5 = { workspace = true } discv5 = { workspace = true }
gossipsub = { workspace = true }
unsigned-varint = { version = "0.6", features = ["codec"] } unsigned-varint = { version = "0.6", features = ["codec"] }
ssz_types = { workspace = true } ssz_types = { workspace = true }
types = { workspace = true } types = { workspace = true }
@ -50,16 +50,12 @@ either = { workspace = true }
# Local dependencies # Local dependencies
futures-ticker = "0.0.3" futures-ticker = "0.0.3"
futures-timer = "3.0.2"
getrandom = "0.2.11" getrandom = "0.2.11"
hex_fmt = "0.3.0" hex_fmt = "0.3.0"
instant = "0.1.12" instant = "0.1.12"
quick-protobuf = "0.8"
void = "1.0.2" void = "1.0.2"
asynchronous-codec = "0.7.0"
base64 = "0.21.5" base64 = "0.21.5"
libp2p-mplex = "0.41" libp2p-mplex = "0.41"
quick-protobuf-codec = "0.3"
[dependencies.libp2p] [dependencies.libp2p]
version = "0.53" version = "0.53"
@ -72,7 +68,7 @@ slog-async = { workspace = true }
tempfile = { workspace = true } tempfile = { workspace = true }
quickcheck = { workspace = true } quickcheck = { workspace = true }
quickcheck_macros = { workspace = true } quickcheck_macros = { workspace = true }
async-std = { version = "1.6.3", features = ["unstable"] } async-channel = { workspace = true }
[features] [features]
libp2p-websocket = [] libp2p-websocket = []

View File

@ -0,0 +1,378 @@
## 0.5 Sigma Prime fork
- Attempt to publish to at least mesh_n peers when publishing a message when flood publish is disabled.
See [PR 5357](https://github.com/sigp/lighthouse/pull/5357).
- Drop `Publish` and `Forward` gossipsub stale messages when polling ConnectionHandler.
See [PR 5175](https://github.com/sigp/lighthouse/pull/5175).
- Apply back pressure by setting a limit in the ConnectionHandler message queue.
See [PR 5066](https://github.com/sigp/lighthouse/pull/5066).
## 0.46.1
- Deprecate `Rpc` in preparation for removing it from the public API because it is an internal type.
See [PR 4833](https://github.com/libp2p/rust-libp2p/pull/4833).
## 0.46.0
- Remove `fast_message_id_fn` mechanism from `Config`.
See [PR 4285](https://github.com/libp2p/rust-libp2p/pull/4285).
- Remove deprecated `gossipsub::Config::idle_timeout` in favor of `SwarmBuilder::idle_connection_timeout`.
See [PR 4642](https://github.com/libp2p/rust-libp2p/pull/4642).
- Return typed error from config builder.
See [PR 4445](https://github.com/libp2p/rust-libp2p/pull/4445).
- Process outbound stream before inbound stream in `EnabledHandler::poll(..)`.
See [PR 4778](https://github.com/libp2p/rust-libp2p/pull/4778).
## 0.45.2
- Deprecate `gossipsub::Config::idle_timeout` in favor of `SwarmBuilder::idle_connection_timeout`.
See [PR 4648].
<!-- Interal changes:
- Allow new clippy lint.
-->
[PR 4648]: (https://github.com/libp2p/rust-libp2p/pull/4648)
<!-- Internal changes
- Allow deprecated usage of `KeepAlive::Until`
-->
## 0.45.1
- Add getter function to o btain `TopicScoreParams`.
See [PR 4231].
[PR 4231]: https://github.com/libp2p/rust-libp2p/pull/4231
## 0.45.0
- Raise MSRV to 1.65.
See [PR 3715].
- Remove deprecated items. See [PR 3862].
[PR 3715]: https://github.com/libp2p/rust-libp2p/pull/3715
[PR 3862]: https://github.com/libp2p/rust-libp2p/pull/3862
## 0.44.4
- Deprecate `metrics`, `protocol`, `subscription_filter`, `time_cache` modules to make them private. See [PR 3777].
- Honor the `gossipsub::Config::support_floodsub` in all cases.
Previously, it was ignored when a custom protocol id was set via `gossipsub::Config::protocol_id`.
See [PR 3837].
[PR 3777]: https://github.com/libp2p/rust-libp2p/pull/3777
[PR 3837]: https://github.com/libp2p/rust-libp2p/pull/3837
## 0.44.3
- Fix erroneously duplicate message IDs. See [PR 3716].
- Gracefully disable handler on stream errors. Deprecate a few variants of `HandlerError`.
See [PR 3625].
[PR 3716]: https://github.com/libp2p/rust-libp2p/pull/3716
[PR 3625]: https://github.com/libp2p/rust-libp2p/pull/3325
## 0.44.2
- Signed messages now use sequential integers in the sequence number field.
See [PR 3551].
[PR 3551]: https://github.com/libp2p/rust-libp2p/pull/3551
## 0.44.1
- Migrate from `prost` to `quick-protobuf`. This removes `protoc` dependency. See [PR 3312].
[PR 3312]: https://github.com/libp2p/rust-libp2p/pull/3312
## 0.44.0
- Update to `prometheus-client` `v0.19.0`. See [PR 3207].
- Update to `libp2p-core` `v0.39.0`.
- Update to `libp2p-swarm` `v0.42.0`.
- Initialize `ProtocolConfig` via `GossipsubConfig`. See [PR 3381].
- Rename types as per [discussion 2174].
`Gossipsub` has been renamed to `Behaviour`.
The `Gossipsub` prefix has been removed from various types like `GossipsubConfig` or `GossipsubMessage`.
It is preferred to import the gossipsub protocol as a module (`use libp2p::gossipsub;`), and refer to its types via `gossipsub::`.
For example: `gossipsub::Behaviour` or `gossipsub::RawMessage`. See [PR 3303].
[PR 3207]: https://github.com/libp2p/rust-libp2p/pull/3207/
[PR 3303]: https://github.com/libp2p/rust-libp2p/pull/3303/
[PR 3381]: https://github.com/libp2p/rust-libp2p/pull/3381/
[discussion 2174]: https://github.com/libp2p/rust-libp2p/discussions/2174
## 0.43.0
- Update to `libp2p-core` `v0.38.0`.
- Update to `libp2p-swarm` `v0.41.0`.
- Update to `prost-codec` `v0.3.0`.
- Refactoring GossipsubCodec to use common protobuf Codec. See [PR 3070].
- Replace `Gossipsub`'s `NetworkBehaviour` implementation `inject_*` methods with the new `on_*` methods.
See [PR 3011].
- Replace `GossipsubHandler`'s `ConnectionHandler` implementation `inject_*` methods with the new `on_*` methods.
See [PR 3085].
- Update `rust-version` to reflect the actual MSRV: 1.62.0. See [PR 3090].
[PR 3085]: https://github.com/libp2p/rust-libp2p/pull/3085
[PR 3070]: https://github.com/libp2p/rust-libp2p/pull/3070
[PR 3011]: https://github.com/libp2p/rust-libp2p/pull/3011
[PR 3090]: https://github.com/libp2p/rust-libp2p/pull/3090
## 0.42.0
- Bump rand to 0.8 and quickcheck to 1. See [PR 2857].
- Update to `libp2p-core` `v0.37.0`.
- Update to `libp2p-swarm` `v0.40.0`.
[PR 2857]: https://github.com/libp2p/rust-libp2p/pull/2857
## 0.41.0
- Update to `libp2p-swarm` `v0.39.0`.
- Update to `libp2p-core` `v0.36.0`.
- Allow publishing with any `impl Into<TopicHash>` as a topic. See [PR 2862].
[PR 2862]: https://github.com/libp2p/rust-libp2p/pull/2862
## 0.40.0
- Update prost requirement from 0.10 to 0.11 which no longer installs the protoc Protobuf compiler.
Thus you will need protoc installed locally. See [PR 2788].
- Update to `libp2p-swarm` `v0.38.0`.
- Update to `libp2p-core` `v0.35.0`.
- Update to `prometheus-client` `v0.18.0`. See [PR 2822].
[PR 2822]: https://github.com/libp2p/rust-libp2p/pull/2761/
[PR 2788]: https://github.com/libp2p/rust-libp2p/pull/2788
## 0.39.0
- Update to `libp2p-core` `v0.34.0`.
- Update to `libp2p-swarm` `v0.37.0`.
- Allow for custom protocol ID via `GossipsubConfigBuilder::protocol_id()`. See [PR 2718].
[PR 2718]: https://github.com/libp2p/rust-libp2p/pull/2718/
## 0.38.1
- Fix duplicate connection id. See [PR 2702].
[PR 2702]: https://github.com/libp2p/rust-libp2p/pull/2702
## 0.38.0
- Update to `libp2p-core` `v0.33.0`.
- Update to `libp2p-swarm` `v0.36.0`.
- changed `TimeCache::contains_key` and `DuplicateCache::contains` to immutable methods. See [PR 2620].
- Update to `prometheus-client` `v0.16.0`. See [PR 2631].
[PR 2620]: https://github.com/libp2p/rust-libp2p/pull/2620
[PR 2631]: https://github.com/libp2p/rust-libp2p/pull/2631
## 0.37.0
- Update to `libp2p-swarm` `v0.35.0`.
- Fix gossipsub metric (see [PR 2558]).
- Allow the user to set the buckets for the score histogram, and to adjust them from the score thresholds. See [PR 2595].
[PR 2558]: https://github.com/libp2p/rust-libp2p/pull/2558
[PR 2595]: https://github.com/libp2p/rust-libp2p/pull/2595
## 0.36.0 [2022-02-22]
- Update to `libp2p-core` `v0.32.0`.
- Update to `libp2p-swarm` `v0.34.0`.
- Move from `open-metrics-client` to `prometheus-client` (see [PR 2442]).
- Emit gossip of all non empty topics (see [PR 2481]).
- Merge NetworkBehaviour's inject_\* paired methods (see [PR 2445]).
- Revert to wasm-timer (see [PR 2506]).
- Do not overwrite msg's peers if put again into mcache (see [PR 2493]).
[PR 2442]: https://github.com/libp2p/rust-libp2p/pull/2442
[PR 2481]: https://github.com/libp2p/rust-libp2p/pull/2481
[PR 2445]: https://github.com/libp2p/rust-libp2p/pull/2445
[PR 2506]: https://github.com/libp2p/rust-libp2p/pull/2506
[PR 2493]: https://github.com/libp2p/rust-libp2p/pull/2493
## 0.35.0 [2022-01-27]
- Update dependencies.
- Migrate to Rust edition 2021 (see [PR 2339]).
- Add metrics for network and configuration performance analysis (see [PR 2346]).
- Improve bandwidth performance by tracking IWANTs and reducing duplicate sends
(see [PR 2327]).
- Implement `Serialize` and `Deserialize` for `MessageId` and `FastMessageId` (see [PR 2408])
- Fix `GossipsubConfigBuilder::build()` requiring `&self` to live for `'static` (see [PR 2409])
- Implement Unsubscribe backoff as per [libp2p specs PR 383] (see [PR 2403]).
[PR 2346]: https://github.com/libp2p/rust-libp2p/pull/2346
[PR 2339]: https://github.com/libp2p/rust-libp2p/pull/2339
[PR 2327]: https://github.com/libp2p/rust-libp2p/pull/2327
[PR 2408]: https://github.com/libp2p/rust-libp2p/pull/2408
[PR 2409]: https://github.com/libp2p/rust-libp2p/pull/2409
[PR 2403]: https://github.com/libp2p/rust-libp2p/pull/2403
[libp2p specs PR 383]: https://github.com/libp2p/specs/pull/383
## 0.34.0 [2021-11-16]
- Add topic and mesh metrics (see [PR 2316]).
- Fix bug in internal peer's topics tracking (see [PR 2325]).
- Use `instant` and `futures-timer` instead of `wasm-timer` (see [PR 2245]).
- Update dependencies.
[PR 2245]: https://github.com/libp2p/rust-libp2p/pull/2245
[PR 2325]: https://github.com/libp2p/rust-libp2p/pull/2325
[PR 2316]: https://github.com/libp2p/rust-libp2p/pull/2316
## 0.33.0 [2021-11-01]
- Add an event to register peers that do not support the gossipsub protocol
[PR 2241](https://github.com/libp2p/rust-libp2p/pull/2241)
- Make default features of `libp2p-core` optional.
[PR 2181](https://github.com/libp2p/rust-libp2p/pull/2181)
- Improve internal peer tracking.
[PR 2175](https://github.com/libp2p/rust-libp2p/pull/2175)
- Update dependencies.
- Allow `message_id_fn`s to accept closures that capture variables.
[PR 2103](https://github.com/libp2p/rust-libp2p/pull/2103)
- Implement std::error::Error for error types.
[PR 2254](https://github.com/libp2p/rust-libp2p/pull/2254)
## 0.32.0 [2021-07-12]
- Update dependencies.
- Reduce log levels across the crate to lessen noisiness of libp2p-gossipsub (see [PR 2101]).
[PR 2101]: https://github.com/libp2p/rust-libp2p/pull/2101
## 0.31.0 [2021-05-17]
- Keep connections to peers in a mesh alive. Allow closing idle connections to peers not in a mesh
[PR-2043].
[PR-2043]: https://github.com/libp2p/rust-libp2p/pull/2043https://github.com/libp2p/rust-libp2p/pull/2043
## 0.30.1 [2021-04-27]
- Remove `regex-filter` feature flag thus always enabling `regex::RegexSubscriptionFilter` [PR
2056](https://github.com/libp2p/rust-libp2p/pull/2056).
## 0.30.0 [2021-04-13]
- Update `libp2p-swarm`.
- Update dependencies.
## 0.29.0 [2021-03-17]
- Update `libp2p-swarm`.
- Update dependencies.
## 0.28.0 [2021-02-15]
- Prevent non-published messages being added to caches.
[PR 1930](https://github.com/libp2p/rust-libp2p/pull/1930)
- Update dependencies.
## 0.27.0 [2021-01-12]
- Update dependencies.
- Implement Gossipsub v1.1 specification.
[PR 1720](https://github.com/libp2p/rust-libp2p/pull/1720)
## 0.26.0 [2020-12-17]
- Update `libp2p-swarm` and `libp2p-core`.
## 0.25.0 [2020-11-25]
- Update `libp2p-swarm` and `libp2p-core`.
## 0.24.0 [2020-11-09]
- Update dependencies.
## 0.23.0 [2020-10-16]
- Update dependencies.
## 0.22.0 [2020-09-09]
- Update `libp2p-swarm` and `libp2p-core`.
## 0.21.0 [2020-08-18]
- Add public API to list topics and peers. [PR 1677](https://github.com/libp2p/rust-libp2p/pull/1677).
- Add message signing and extended privacy/validation configurations. [PR 1583](https://github.com/libp2p/rust-libp2p/pull/1583).
- `Debug` instance for `Gossipsub`. [PR 1673](https://github.com/libp2p/rust-libp2p/pull/1673).
- Bump `libp2p-core` and `libp2p-swarm` dependency.
## 0.20.0 [2020-07-01]
- Updated dependencies.
## 0.19.3 [2020-06-23]
- Maintenance release fixing linter warnings.
## 0.19.2 [2020-06-22]
- Updated dependencies.

View File

@ -0,0 +1,50 @@
[package]
name = "gossipsub"
edition = "2021"
description = "Sigma prime's version of Gossipsub protocol for libp2p"
version = "0.5.0"
authors = ["Age Manning <Age@AgeManning.com>"]
license = "MIT"
repository = "https://github.com/sigp/lighthouse/"
keywords = ["peer-to-peer", "libp2p", "networking"]
categories = ["network-programming", "asynchronous"]
[features]
wasm-bindgen = ["getrandom/js", "instant/wasm-bindgen"]
[dependencies]
async-channel = { workspace = true }
asynchronous-codec = "0.7.0"
base64 = "0.21.7"
byteorder = "1.5.0"
bytes = "1.5"
either = "1.9"
fnv = "1.0.7"
futures = "0.3.30"
futures-ticker = "0.0.3"
futures-timer = "3.0.2"
getrandom = "0.2.12"
hex_fmt = "0.3.0"
instant = "0.1.12"
libp2p = { version = "0.53", default-features = false }
quick-protobuf = "0.8"
quick-protobuf-codec = "0.3"
rand = "0.8"
regex = "1.10.3"
serde = { version = "1", optional = true, features = ["derive"] }
sha2 = "0.10.8"
smallvec = "1.13.1"
tracing = "0.1.37"
void = "1.0.2"
prometheus-client = "0.22.0"
[dev-dependencies]
quickcheck = { workspace = true }
# Passing arguments to the docsrs builder in order to properly document cfg's.
# More information: https://docs.rs/about/builds#cross-compiling
[package.metadata.docs.rs]
all-features = true
rustdoc-args = ["--cfg", "docsrs"]
rustc-args = ["--cfg", "docsrs"]

View File

@ -19,7 +19,7 @@
// DEALINGS IN THE SOFTWARE. // DEALINGS IN THE SOFTWARE.
//! Data structure for efficiently storing known back-off's when pruning peers. //! Data structure for efficiently storing known back-off's when pruning peers.
use crate::gossipsub::topic::TopicHash; use crate::topic::TopicHash;
use instant::Instant; use instant::Instant;
use libp2p::identity::PeerId; use libp2p::identity::PeerId;
use std::collections::{ use std::collections::{

View File

@ -57,8 +57,8 @@ use super::time_cache::DuplicateCache;
use super::topic::{Hasher, Topic, TopicHash}; use super::topic::{Hasher, Topic, TopicHash};
use super::transform::{DataTransform, IdentityTransform}; use super::transform::{DataTransform, IdentityTransform};
use super::types::{ use super::types::{
ControlAction, Message, MessageAcceptance, MessageId, PeerInfo, RawMessage, Subscription, ControlAction, FailedMessages, Message, MessageAcceptance, MessageId, PeerInfo, RawMessage,
SubscriptionAction, Subscription, SubscriptionAction,
}; };
use super::types::{Graft, IHave, IWant, PeerConnections, PeerKind, Prune}; use super::types::{Graft, IHave, IWant, PeerConnections, PeerKind, Prune};
use super::{backoff::BackoffStorage, types::RpcSender}; use super::{backoff::BackoffStorage, types::RpcSender};
@ -66,7 +66,7 @@ use super::{
config::{Config, ValidationMode}, config::{Config, ValidationMode},
types::RpcOut, types::RpcOut,
}; };
use super::{FailedMessages, PublishError, SubscriptionError, TopicScoreParams, ValidationError}; use super::{PublishError, SubscriptionError, TopicScoreParams, ValidationError};
use instant::SystemTime; use instant::SystemTime;
use quick_protobuf::{MessageWrite, Writer}; use quick_protobuf::{MessageWrite, Writer};
use std::{cmp::Ordering::Equal, fmt::Debug}; use std::{cmp::Ordering::Equal, fmt::Debug};
@ -525,7 +525,7 @@ where
return Err(SubscriptionError::NotAllowed); return Err(SubscriptionError::NotAllowed);
} }
if self.mesh.get(&topic_hash).is_some() { if self.mesh.contains_key(&topic_hash) {
tracing::debug!(%topic, "Topic is already in the mesh"); tracing::debug!(%topic, "Topic is already in the mesh");
return Ok(false); return Ok(false);
} }
@ -551,7 +551,7 @@ where
tracing::debug!(%topic, "Unsubscribing from topic"); tracing::debug!(%topic, "Unsubscribing from topic");
let topic_hash = topic.hash(); let topic_hash = topic.hash();
if self.mesh.get(&topic_hash).is_none() { if !self.mesh.contains_key(&topic_hash) {
tracing::debug!(topic=%topic_hash, "Already unsubscribed from topic"); tracing::debug!(topic=%topic_hash, "Already unsubscribed from topic");
// we are not subscribed // we are not subscribed
return Ok(false); return Ok(false);

View File

@ -21,19 +21,18 @@
// Collection of tests for the gossipsub network behaviour // Collection of tests for the gossipsub network behaviour
use super::*; use super::*;
use crate::gossipsub::subscription_filter::WhitelistSubscriptionFilter; use crate::subscription_filter::WhitelistSubscriptionFilter;
use crate::gossipsub::transform::{DataTransform, IdentityTransform}; use crate::transform::{DataTransform, IdentityTransform};
use crate::gossipsub::types::{RpcOut, RpcReceiver}; use crate::types::{RpcOut, RpcReceiver};
use crate::gossipsub::ValidationError; use crate::ValidationError;
use crate::gossipsub::{ use crate::{
config::Config, config::ConfigBuilder, types::Rpc, IdentTopic as Topic, TopicScoreParams, config::Config, config::ConfigBuilder, types::Rpc, IdentTopic as Topic, TopicScoreParams,
}; };
use async_std::net::Ipv4Addr;
use byteorder::{BigEndian, ByteOrder}; use byteorder::{BigEndian, ByteOrder};
use libp2p::core::{ConnectedPoint, Endpoint}; use libp2p::core::ConnectedPoint;
use rand::Rng; use rand::Rng;
use std::net::Ipv4Addr;
use std::thread::sleep; use std::thread::sleep;
use std::time::Duration;
#[derive(Default, Debug)] #[derive(Default, Debug)]
struct InjectNodes<D, F> struct InjectNodes<D, F>
@ -427,7 +426,7 @@ fn test_subscribe() {
.create_network(); .create_network();
assert!( assert!(
gs.mesh.get(&topic_hashes[0]).is_some(), gs.mesh.contains_key(&topic_hashes[0]),
"Subscribe should add a new entry to the mesh[topic] hashmap" "Subscribe should add a new entry to the mesh[topic] hashmap"
); );
@ -477,7 +476,7 @@ fn test_unsubscribe() {
"Topic_peers contain a topic entry" "Topic_peers contain a topic entry"
); );
assert!( assert!(
gs.mesh.get(topic_hash).is_some(), gs.mesh.contains_key(topic_hash),
"mesh should contain a topic entry" "mesh should contain a topic entry"
); );
} }
@ -511,7 +510,7 @@ fn test_unsubscribe() {
// check we clean up internal structures // check we clean up internal structures
for topic_hash in &topic_hashes { for topic_hash in &topic_hashes {
assert!( assert!(
gs.mesh.get(topic_hash).is_none(), !gs.mesh.contains_key(topic_hash),
"All topics should have been removed from the mesh" "All topics should have been removed from the mesh"
); );
} }
@ -694,7 +693,7 @@ fn test_publish_without_flood_publishing() {
.create_network(); .create_network();
assert!( assert!(
gs.mesh.get(&topic_hashes[0]).is_some(), gs.mesh.contains_key(&topic_hashes[0]),
"Subscribe should add a new entry to the mesh[topic] hashmap" "Subscribe should add a new entry to the mesh[topic] hashmap"
); );
@ -774,7 +773,7 @@ fn test_fanout() {
.create_network(); .create_network();
assert!( assert!(
gs.mesh.get(&topic_hashes[0]).is_some(), gs.mesh.contains_key(&topic_hashes[0]),
"Subscribe should add a new entry to the mesh[topic] hashmap" "Subscribe should add a new entry to the mesh[topic] hashmap"
); );
// Unsubscribe from topic // Unsubscribe from topic
@ -946,7 +945,7 @@ fn test_handle_received_subscriptions() {
); );
assert!( assert!(
gs.connected_peers.get(&unknown_peer).is_none(), !gs.connected_peers.contains_key(&unknown_peer),
"Unknown peer should not have been added" "Unknown peer should not have been added"
); );
@ -1347,7 +1346,7 @@ fn test_handle_graft_multiple_topics() {
} }
assert!( assert!(
gs.mesh.get(&topic_hashes[2]).is_none(), !gs.mesh.contains_key(&topic_hashes[2]),
"Expected the second topic to not be in the mesh" "Expected the second topic to not be in the mesh"
); );
} }
@ -5228,7 +5227,7 @@ fn test_graft_without_subscribe() {
.create_network(); .create_network();
assert!( assert!(
gs.mesh.get(&topic_hashes[0]).is_some(), gs.mesh.contains_key(&topic_hashes[0]),
"Subscribe should add a new entry to the mesh[topic] hashmap" "Subscribe should add a new entry to the mesh[topic] hashmap"
); );

View File

@ -36,7 +36,7 @@ pub enum ValidationMode {
/// be present as well as the sequence number. All messages must have valid signatures. /// be present as well as the sequence number. All messages must have valid signatures.
/// ///
/// NOTE: This setting will reject messages from nodes using /// NOTE: This setting will reject messages from nodes using
/// [`crate::gossipsub::behaviour::MessageAuthenticity::Anonymous`] and all messages that do not have /// [`crate::behaviour::MessageAuthenticity::Anonymous`] and all messages that do not have
/// signatures. /// signatures.
Strict, Strict,
/// This setting permits messages that have no author, sequence number or signature. If any of /// This setting permits messages that have no author, sequence number or signature. If any of
@ -195,7 +195,7 @@ impl Config {
/// When set to `true`, prevents automatic forwarding of all received messages. This setting /// When set to `true`, prevents automatic forwarding of all received messages. This setting
/// allows a user to validate the messages before propagating them to their peers. If set to /// allows a user to validate the messages before propagating them to their peers. If set to
/// true, the user must manually call [`crate::gossipsub::Behaviour::report_message_validation_result()`] /// true, the user must manually call [`crate::Behaviour::report_message_validation_result()`]
/// on the behaviour to forward message once validated (default is `false`). /// on the behaviour to forward message once validated (default is `false`).
/// The default is `false`. /// The default is `false`.
pub fn validate_messages(&self) -> bool { pub fn validate_messages(&self) -> bool {
@ -611,7 +611,7 @@ impl ConfigBuilder {
/// When set, prevents automatic forwarding of all received messages. This setting /// When set, prevents automatic forwarding of all received messages. This setting
/// allows a user to validate the messages before propagating them to their peers. If set, /// allows a user to validate the messages before propagating them to their peers. If set,
/// the user must manually call [`crate::gossipsub::Behaviour::report_message_validation_result()`] on the /// the user must manually call [`crate::Behaviour::report_message_validation_result()`] on the
/// behaviour to forward a message once validated. /// behaviour to forward a message once validated.
pub fn validate_messages(&mut self) -> &mut Self { pub fn validate_messages(&mut self) -> &mut Self {
self.config.validate_messages = true; self.config.validate_messages = true;
@ -902,11 +902,10 @@ impl std::fmt::Debug for Config {
#[cfg(test)] #[cfg(test)]
mod test { mod test {
use super::*; use super::*;
use crate::gossipsub::topic::IdentityHash; use crate::topic::IdentityHash;
use crate::gossipsub::types::PeerKind; use crate::types::PeerKind;
use crate::gossipsub::Topic; use crate::Topic;
use libp2p::core::UpgradeInfo; use libp2p::core::UpgradeInfo;
use libp2p::swarm::StreamProtocol;
use std::collections::hash_map::DefaultHasher; use std::collections::hash_map::DefaultHasher;
use std::hash::{Hash, Hasher}; use std::hash::{Hash, Hasher};

View File

@ -0,0 +1,134 @@
// Copyright 2020 Sigma Prime Pty Ltd.
//
// Permission is hereby granted, free of charge, to any person obtaining a
// copy of this software and associated documentation files (the "Software"),
// to deal in the Software without restriction, including without limitation
// the rights to use, copy, modify, merge, publish, distribute, sublicense,
// and/or sell copies of the Software, and to permit persons to whom the
// Software is furnished to do so, subject to the following conditions:
//
// The above copyright notice and this permission notice shall be included in
// all copies or substantial portions of the Software.
//
// THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
// OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
// FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
// AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
// LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
// FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
// DEALINGS IN THE SOFTWARE.
//! Implementation of the [Gossipsub](https://github.com/libp2p/specs/blob/master/pubsub/gossipsub/README.md) protocol.
//!
//! Gossipsub is a P2P pubsub (publish/subscription) routing layer designed to extend upon
//! floodsub and meshsub routing protocols.
//!
//! # Overview
//!
//! *Note: The gossipsub protocol specifications
//! (<https://github.com/libp2p/specs/tree/master/pubsub/gossipsub>) provide an outline for the
//! routing protocol. They should be consulted for further detail.*
//!
//! Gossipsub is a blend of meshsub for data and randomsub for mesh metadata. It provides bounded
//! degree and amplification factor with the meshsub construction and augments it using gossip
//! propagation of metadata with the randomsub technique.
//!
//! The router maintains an overlay mesh network of peers on which to efficiently send messages and
//! metadata. Peers use control messages to broadcast and request known messages and
//! subscribe/unsubscribe from topics in the mesh network.
//!
//! # Important Discrepancies
//!
//! This section outlines the current implementation's potential discrepancies from that of other
//! implementations, due to undefined elements in the current specification.
//!
//! - **Topics** - In gossipsub, topics configurable by the `hash_topics` configuration parameter.
//! Topics are of type [`TopicHash`]. The current go implementation uses raw utf-8 strings, and this
//! is default configuration in rust-libp2p. Topics can be hashed (SHA256 hashed then base64
//! encoded) by setting the `hash_topics` configuration parameter to true.
//!
//! - **Sequence Numbers** - A message on the gossipsub network is identified by the source
//! [`PeerId`](libp2p_identity::PeerId) and a nonce (sequence number) of the message. The sequence numbers in
//! this implementation are sent as raw bytes across the wire. They are 64-bit big-endian unsigned
//! integers. When messages are signed, they are monotonically increasing integers starting from a
//! random value and wrapping around u64::MAX. When messages are unsigned, they are chosen at random.
//! NOTE: These numbers are sequential in the current go implementation.
//!
//! # Peer Discovery
//!
//! Gossipsub does not provide peer discovery by itself. Peer discovery is the process by which
//! peers in a p2p network exchange information about each other among other reasons to become resistant
//! against the failure or replacement of the
//! [boot nodes](https://docs.libp2p.io/reference/glossary/#boot-node) of the network.
//!
//! Peer
//! discovery can e.g. be implemented with the help of the [Kademlia](https://github.com/libp2p/specs/blob/master/kad-dht/README.md) protocol
//! in combination with the [Identify](https://github.com/libp2p/specs/tree/master/identify) protocol. See the
//! Kademlia implementation documentation for more information.
//!
//! # Using Gossipsub
//!
//! ## Gossipsub Config
//!
//! The [`Config`] struct specifies various network performance/tuning configuration
//! parameters. Specifically it specifies:
//!
//! [`Config`]: struct.Config.html
//!
//! This struct implements the [`Default`] trait and can be initialised via
//! [`Config::default()`].
//!
//!
//! ## Behaviour
//!
//! The [`Behaviour`] struct implements the [`libp2p_swarm::NetworkBehaviour`] trait allowing it to
//! act as the routing behaviour in a [`libp2p_swarm::Swarm`]. This struct requires an instance of
//! [`PeerId`](libp2p_identity::PeerId) and [`Config`].
//!
//! [`Behaviour`]: struct.Behaviour.html
//! ## Example
//!
//! For an example on how to use gossipsub, see the [chat-example](https://github.com/libp2p/rust-libp2p/tree/master/examples/chat).
#![cfg_attr(docsrs, feature(doc_cfg, doc_auto_cfg))]
mod backoff;
mod behaviour;
mod config;
mod error;
mod gossip_promises;
mod handler;
mod mcache;
mod metrics;
mod peer_score;
mod protocol;
mod rpc_proto;
mod subscription_filter;
mod time_cache;
mod topic;
mod transform;
mod types;
pub use self::behaviour::{Behaviour, Event, MessageAuthenticity};
pub use self::config::{Config, ConfigBuilder, ValidationMode, Version};
pub use self::error::{ConfigBuilderError, PublishError, SubscriptionError, ValidationError};
pub use self::metrics::Config as MetricsConfig;
pub use self::peer_score::{
score_parameter_decay, score_parameter_decay_with_base, PeerScoreParams, PeerScoreThresholds,
TopicScoreParams,
};
pub use self::subscription_filter::{
AllowAllSubscriptionFilter, CallbackSubscriptionFilter, CombinedSubscriptionFilters,
MaxCountSubscriptionFilter, RegexSubscriptionFilter, TopicSubscriptionFilter,
WhitelistSubscriptionFilter,
};
pub use self::topic::{Hasher, Topic, TopicHash};
pub use self::transform::{DataTransform, IdentityTransform};
pub use self::types::{FailedMessages, Message, MessageAcceptance, MessageId, RawMessage};
#[deprecated(note = "Will be removed from the public API.")]
pub type Rpc = self::types::Rpc;
pub type IdentTopic = Topic<self::topic::IdentityHash>;
pub type Sha256Topic = Topic<self::topic::Sha256Hash>;

View File

@ -221,7 +221,7 @@ impl MessageCache {
#[cfg(test)] #[cfg(test)]
mod tests { mod tests {
use super::*; use super::*;
use crate::gossipsub::types::RawMessage; use crate::types::RawMessage;
use crate::{IdentTopic as Topic, TopicHash}; use crate::{IdentTopic as Topic, TopicHash};
use libp2p::identity::PeerId; use libp2p::identity::PeerId;

View File

@ -102,7 +102,7 @@ impl PeerStats {
topic_hash: TopicHash, topic_hash: TopicHash,
params: &PeerScoreParams, params: &PeerScoreParams,
) -> Option<&mut TopicStats> { ) -> Option<&mut TopicStats> {
if params.topics.get(&topic_hash).is_some() { if params.topics.contains_key(&topic_hash) {
Some(self.topics.entry(topic_hash).or_default()) Some(self.topics.entry(topic_hash).or_default())
} else { } else {
self.topics.get_mut(&topic_hash) self.topics.get_mut(&topic_hash)
@ -310,7 +310,7 @@ impl PeerScore {
// P6: IP collocation factor // P6: IP collocation factor
for ip in peer_stats.known_ips.iter() { for ip in peer_stats.known_ips.iter() {
if self.params.ip_colocation_factor_whitelist.get(ip).is_some() { if self.params.ip_colocation_factor_whitelist.contains(ip) {
continue; continue;
} }
@ -705,7 +705,7 @@ impl PeerScore {
) { ) {
let record = self.deliveries.entry(msg_id.clone()).or_default(); let record = self.deliveries.entry(msg_id.clone()).or_default();
if record.peers.get(from).is_some() { if record.peers.contains(from) {
// we have already seen this duplicate! // we have already seen this duplicate!
return; return;
} }

View File

@ -18,7 +18,7 @@
// FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER // FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
// DEALINGS IN THE SOFTWARE. // DEALINGS IN THE SOFTWARE.
use crate::gossipsub::TopicHash; use crate::TopicHash;
use std::collections::{HashMap, HashSet}; use std::collections::{HashMap, HashSet};
use std::net::IpAddr; use std::net::IpAddr;
use std::time::Duration; use std::time::Duration;

View File

@ -21,8 +21,8 @@
/// A collection of unit tests mostly ported from the go implementation. /// A collection of unit tests mostly ported from the go implementation.
use super::*; use super::*;
use crate::gossipsub::types::RawMessage; use crate::types::RawMessage;
use crate::gossipsub::{IdentTopic as Topic, Message}; use crate::{IdentTopic as Topic, Message};
// estimates a value within variance // estimates a value within variance
fn within_variance(value: f64, expected: f64, variance: f64) -> bool { fn within_variance(value: f64, expected: f64, variance: f64) -> bool {

View File

@ -30,7 +30,6 @@ use super::ValidationError;
use asynchronous_codec::{Decoder, Encoder, Framed}; use asynchronous_codec::{Decoder, Encoder, Framed};
use byteorder::{BigEndian, ByteOrder}; use byteorder::{BigEndian, ByteOrder};
use bytes::BytesMut; use bytes::BytesMut;
use futures::future;
use futures::prelude::*; use futures::prelude::*;
use libp2p::core::{InboundUpgrade, OutboundUpgrade, UpgradeInfo}; use libp2p::core::{InboundUpgrade, OutboundUpgrade, UpgradeInfo};
use libp2p::identity::{PeerId, PublicKey}; use libp2p::identity::{PeerId, PublicKey};
@ -508,10 +507,10 @@ impl Decoder for GossipsubCodec {
#[cfg(test)] #[cfg(test)]
mod tests { mod tests {
use super::*; use super::*;
use crate::gossipsub::config::Config; use crate::config::Config;
use crate::gossipsub::protocol::{BytesMut, GossipsubCodec, HandlerEvent}; use crate::protocol::{BytesMut, GossipsubCodec, HandlerEvent};
use crate::gossipsub::*; use crate::{Behaviour, ConfigBuilder, MessageAuthenticity};
use crate::gossipsub::{IdentTopic as Topic, Version}; use crate::{IdentTopic as Topic, Version};
use libp2p::identity::Keypair; use libp2p::identity::Keypair;
use quickcheck::*; use quickcheck::*;
@ -586,7 +585,7 @@ mod tests {
fn prop(message: Message) { fn prop(message: Message) {
let message = message.0; let message = message.0;
let rpc = crate::gossipsub::types::Rpc { let rpc = crate::types::Rpc {
messages: vec![message.clone()], messages: vec![message.clone()],
subscriptions: vec![], subscriptions: vec![],
control_msgs: vec![], control_msgs: vec![],

View File

@ -26,8 +26,8 @@ pub(crate) mod proto {
#[cfg(test)] #[cfg(test)]
mod test { mod test {
use crate::gossipsub::rpc_proto::proto::compat; use crate::rpc_proto::proto::compat;
use crate::gossipsub::IdentTopic as Topic; use crate::IdentTopic as Topic;
use libp2p::identity::PeerId; use libp2p::identity::PeerId;
use quick_protobuf::{BytesReader, MessageRead, MessageWrite, Writer}; use quick_protobuf::{BytesReader, MessageRead, MessageWrite, Writer};
use rand::Rng; use rand::Rng;

View File

@ -18,8 +18,8 @@
// FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER // FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
// DEALINGS IN THE SOFTWARE. // DEALINGS IN THE SOFTWARE.
use crate::gossipsub::types::Subscription; use crate::types::Subscription;
use crate::gossipsub::TopicHash; use crate::TopicHash;
use std::collections::{BTreeSet, HashMap, HashSet}; use std::collections::{BTreeSet, HashMap, HashSet};
pub trait TopicSubscriptionFilter { pub trait TopicSubscriptionFilter {
@ -128,7 +128,7 @@ impl<T: TopicSubscriptionFilter> TopicSubscriptionFilter for MaxCountSubscriptio
.filter .filter
.filter_incoming_subscriptions(subscriptions, currently_subscribed_topics)?; .filter_incoming_subscriptions(subscriptions, currently_subscribed_topics)?;
use crate::gossipsub::types::SubscriptionAction::*; use crate::types::SubscriptionAction::*;
let mut unsubscribed = 0; let mut unsubscribed = 0;
let mut new_subscribed = 0; let mut new_subscribed = 0;
@ -211,7 +211,7 @@ impl TopicSubscriptionFilter for RegexSubscriptionFilter {
#[cfg(test)] #[cfg(test)]
mod test { mod test {
use super::*; use super::*;
use crate::gossipsub::types::SubscriptionAction::*; use crate::types::SubscriptionAction::*;
use std::iter::FromIterator; use std::iter::FromIterator;
#[test] #[test]

View File

@ -18,7 +18,7 @@
// FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER // FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
// DEALINGS IN THE SOFTWARE. // DEALINGS IN THE SOFTWARE.
use crate::gossipsub::rpc_proto::proto; use crate::rpc_proto::proto;
use base64::prelude::*; use base64::prelude::*;
use prometheus_client::encoding::EncodeLabelSet; use prometheus_client::encoding::EncodeLabelSet;
use quick_protobuf::Writer; use quick_protobuf::Writer;

View File

@ -25,11 +25,11 @@
//! algorithms that can be topic-specific. Once the raw data is transformed the message-id is then //! algorithms that can be topic-specific. Once the raw data is transformed the message-id is then
//! calculated, allowing for applications to employ message-id functions post compression. //! calculated, allowing for applications to employ message-id functions post compression.
use crate::gossipsub::{Message, RawMessage, TopicHash}; use crate::{Message, RawMessage, TopicHash};
/// A general trait of transforming a [`RawMessage`] into a [`Message`]. The /// A general trait of transforming a [`RawMessage`] into a [`Message`]. The
/// [`RawMessage`] is obtained from the wire and the [`Message`] is used to /// [`RawMessage`] is obtained from the wire and the [`Message`] is used to
/// calculate the [`crate::gossipsub::MessageId`] of the message and is what is sent to the application. /// calculate the [`crate::MessageId`] of the message and is what is sent to the application.
/// ///
/// The inbound/outbound transforms must be inverses. Applying the inbound transform and then the /// The inbound/outbound transforms must be inverses. Applying the inbound transform and then the
/// outbound transform MUST leave the underlying data un-modified. /// outbound transform MUST leave the underlying data un-modified.
@ -40,7 +40,7 @@ pub trait DataTransform {
fn inbound_transform(&self, raw_message: RawMessage) -> Result<Message, std::io::Error>; fn inbound_transform(&self, raw_message: RawMessage) -> Result<Message, std::io::Error>;
/// Takes the data to be published (a topic and associated data) transforms the data. The /// Takes the data to be published (a topic and associated data) transforms the data. The
/// transformed data will then be used to create a [`crate::gossipsub::RawMessage`] to be sent to peers. /// transformed data will then be used to create a [`crate::RawMessage`] to be sent to peers.
fn outbound_transform( fn outbound_transform(
&self, &self,
topic: &TopicHash, topic: &TopicHash,

View File

@ -19,8 +19,8 @@
// DEALINGS IN THE SOFTWARE. // DEALINGS IN THE SOFTWARE.
//! A collection of types using the Gossipsub system. //! A collection of types using the Gossipsub system.
use crate::gossipsub::metrics::Metrics; use crate::metrics::Metrics;
use crate::gossipsub::TopicHash; use crate::TopicHash;
use async_channel::{Receiver, Sender}; use async_channel::{Receiver, Sender};
use futures::stream::Peekable; use futures::stream::Peekable;
use futures::{Future, Stream, StreamExt}; use futures::{Future, Stream, StreamExt};
@ -37,7 +37,7 @@ use std::sync::Arc;
use std::task::{Context, Poll}; use std::task::{Context, Poll};
use std::{fmt, pin::Pin}; use std::{fmt, pin::Pin};
use crate::gossipsub::rpc_proto::proto; use crate::rpc_proto::proto;
#[cfg(feature = "serde")] #[cfg(feature = "serde")]
use serde::{Deserialize, Serialize}; use serde::{Deserialize, Serialize};
@ -190,7 +190,7 @@ impl From<RawMessage> for proto::Message {
} }
/// The message sent to the user after a [`RawMessage`] has been transformed by a /// The message sent to the user after a [`RawMessage`] has been transformed by a
/// [`crate::gossipsub::DataTransform`]. /// [`crate::DataTransform`].
#[derive(Clone, PartialEq, Eq, Hash)] #[derive(Clone, PartialEq, Eq, Hash)]
pub struct Message { pub struct Message {
/// Id of the peer that published this message. /// Id of the peer that published this message.

View File

@ -1,4 +1,3 @@
use crate::gossipsub;
use crate::listen_addr::{ListenAddr, ListenAddress}; use crate::listen_addr::{ListenAddr, ListenAddress};
use crate::rpc::config::{InboundRateLimiterConfig, OutboundRateLimiterConfig}; use crate::rpc::config::{InboundRateLimiterConfig, OutboundRateLimiterConfig};
use crate::types::GossipKind; use crate::types::GossipKind;
@ -21,20 +20,6 @@ pub const DEFAULT_TCP_PORT: u16 = 9000u16;
pub const DEFAULT_DISC_PORT: u16 = 9000u16; pub const DEFAULT_DISC_PORT: u16 = 9000u16;
pub const DEFAULT_QUIC_PORT: u16 = 9001u16; pub const DEFAULT_QUIC_PORT: u16 = 9001u16;
/// The cache time is set to accommodate the circulation time of an attestation.
///
/// The p2p spec declares that we accept attestations within the following range:
///
/// ```ignore
/// ATTESTATION_PROPAGATION_SLOT_RANGE = 32
/// attestation.data.slot + ATTESTATION_PROPAGATION_SLOT_RANGE >= current_slot >= attestation.data.slot
/// ```
///
/// Therefore, we must accept attestations across a span of 33 slots (where each slot is 12
/// seconds). We add an additional second to account for the 500ms gossip clock disparity, and
/// another 500ms for "fudge factor".
pub const DUPLICATE_CACHE_TIME: Duration = Duration::from_secs(33 * 12 + 1);
/// The maximum size of gossip messages. /// The maximum size of gossip messages.
pub fn gossip_max_size(is_merge_enabled: bool, gossip_max_size: usize) -> usize { pub fn gossip_max_size(is_merge_enabled: bool, gossip_max_size: usize) -> usize {
if is_merge_enabled { if is_merge_enabled {
@ -453,6 +438,8 @@ pub fn gossipsub_config(
network_load: u8, network_load: u8,
fork_context: Arc<ForkContext>, fork_context: Arc<ForkContext>,
gossipsub_config_params: GossipsubConfigParams, gossipsub_config_params: GossipsubConfigParams,
seconds_per_slot: u64,
slots_per_epoch: u64,
) -> gossipsub::Config { ) -> gossipsub::Config {
fn prefix( fn prefix(
prefix: [u8; 4], prefix: [u8; 4],
@ -492,6 +479,13 @@ pub fn gossipsub_config(
let load = NetworkLoad::from(network_load); let load = NetworkLoad::from(network_load);
// Since EIP 7045 (activated at the deneb fork), we allow attestations that are
// 2 epochs old to be circulated around the p2p network.
// To accommodate the increase, we should increase the duplicate cache time to filter older seen messages.
// 2 epochs is quite sane for pre-deneb network parameters as well.
// Hence we keep the same parameters for pre-deneb networks as well to avoid switching at the fork.
let duplicate_cache_time = Duration::from_secs(slots_per_epoch * seconds_per_slot * 2);
gossipsub::ConfigBuilder::default() gossipsub::ConfigBuilder::default()
.max_transmit_size(gossip_max_size( .max_transmit_size(gossip_max_size(
is_merge_enabled, is_merge_enabled,
@ -510,7 +504,7 @@ pub fn gossipsub_config(
.history_gossip(load.history_gossip) .history_gossip(load.history_gossip)
.validate_messages() // require validation before propagation .validate_messages() // require validation before propagation
.validation_mode(gossipsub::ValidationMode::Anonymous) .validation_mode(gossipsub::ValidationMode::Anonymous)
.duplicate_cache_time(DUPLICATE_CACHE_TIME) .duplicate_cache_time(duplicate_cache_time)
.message_id_fn(gossip_message_id) .message_id_fn(gossip_message_id)
.allow_self_origin(true) .allow_self_origin(true)
.build() .build()

View File

@ -10,7 +10,6 @@ pub mod service;
#[allow(clippy::mutable_key_type)] // PeerId in hashmaps are no longer permitted by clippy #[allow(clippy::mutable_key_type)] // PeerId in hashmaps are no longer permitted by clippy
pub mod discovery; pub mod discovery;
pub mod gossipsub;
pub mod listen_addr; pub mod listen_addr;
pub mod metrics; pub mod metrics;
pub mod peer_manager; pub mod peer_manager;

View File

@ -26,6 +26,8 @@ pub use libp2p::identity::Keypair;
#[allow(clippy::mutable_key_type)] // PeerId in hashmaps are no longer permitted by clippy #[allow(clippy::mutable_key_type)] // PeerId in hashmaps are no longer permitted by clippy
pub mod peerdb; pub mod peerdb;
use crate::peer_manager::peerdb::client::ClientKind;
use libp2p::multiaddr;
pub use peerdb::peer_info::{ pub use peerdb::peer_info::{
ConnectionDirection, PeerConnectionStatus, PeerConnectionStatus::*, PeerInfo, ConnectionDirection, PeerConnectionStatus, PeerConnectionStatus::*, PeerInfo,
}; };
@ -33,6 +35,8 @@ use peerdb::score::{PeerAction, ReportSource};
pub use peerdb::sync_status::{SyncInfo, SyncStatus}; pub use peerdb::sync_status::{SyncInfo, SyncStatus};
use std::collections::{hash_map::Entry, HashMap}; use std::collections::{hash_map::Entry, HashMap};
use std::net::IpAddr; use std::net::IpAddr;
use strum::IntoEnumIterator;
pub mod config; pub mod config;
mod network_behaviour; mod network_behaviour;
@ -464,19 +468,6 @@ impl<TSpec: EthSpec> PeerManager<TSpec> {
"observed_address" => ?info.observed_addr, "observed_address" => ?info.observed_addr,
"protocols" => ?info.protocols "protocols" => ?info.protocols
); );
// update the peer client kind metric if the peer is connected
if matches!(
peer_info.connection_status(),
PeerConnectionStatus::Connected { .. }
| PeerConnectionStatus::Disconnecting { .. }
) {
metrics::inc_gauge_vec(
&metrics::PEERS_PER_CLIENT,
&[peer_info.client().kind.as_ref()],
);
metrics::dec_gauge_vec(&metrics::PEERS_PER_CLIENT, &[previous_kind.as_ref()]);
}
} }
} else { } else {
error!(self.log, "Received an Identify response from an unknown peer"; "peer_id" => peer_id.to_string()); error!(self.log, "Received an Identify response from an unknown peer"; "peer_id" => peer_id.to_string());
@ -812,12 +803,6 @@ impl<TSpec: EthSpec> PeerManager<TSpec> {
// start a ping and status timer for the peer // start a ping and status timer for the peer
self.status_peers.insert(*peer_id); self.status_peers.insert(*peer_id);
let connected_peers = self.network_globals.connected_peers() as i64;
// increment prometheus metrics
metrics::inc_counter(&metrics::PEER_CONNECT_EVENT_COUNT);
metrics::set_gauge(&metrics::PEERS_CONNECTED, connected_peers);
true true
} }
@ -1267,6 +1252,70 @@ impl<TSpec: EthSpec> PeerManager<TSpec> {
); );
} }
} }
// Update peer count related metrics.
fn update_peer_count_metrics(&self) {
let mut peers_connected = 0;
let mut clients_per_peer = HashMap::new();
let mut peers_connected_mutli: HashMap<(&str, &str), i32> = HashMap::new();
for (_, peer_info) in self.network_globals.peers.read().connected_peers() {
peers_connected += 1;
*clients_per_peer
.entry(peer_info.client().kind.to_string())
.or_default() += 1;
let direction = match peer_info.connection_direction() {
Some(ConnectionDirection::Incoming) => "inbound",
Some(ConnectionDirection::Outgoing) => "outbound",
None => "none",
};
// Note: the `transport` is set to `unknown` if the `listening_addresses` list is empty.
// This situation occurs when the peer is initially registered in PeerDB, but the peer
// info has not yet been updated at `PeerManager::identify`.
let transport = peer_info
.listening_addresses()
.iter()
.find_map(|addr| {
addr.iter().find_map(|proto| match proto {
multiaddr::Protocol::QuicV1 => Some("quic"),
multiaddr::Protocol::Tcp(_) => Some("tcp"),
_ => None,
})
})
.unwrap_or("unknown");
*peers_connected_mutli
.entry((direction, transport))
.or_default() += 1;
}
// PEERS_CONNECTED
metrics::set_gauge(&metrics::PEERS_CONNECTED, peers_connected);
// PEERS_PER_CLIENT
for client_kind in ClientKind::iter() {
let value = clients_per_peer.get(&client_kind.to_string()).unwrap_or(&0);
metrics::set_gauge_vec(
&metrics::PEERS_PER_CLIENT,
&[client_kind.as_ref()],
*value as i64,
);
}
// PEERS_CONNECTED_MULTI
for direction in ["inbound", "outbound", "none"] {
for transport in ["quic", "tcp", "unknown"] {
metrics::set_gauge_vec(
&metrics::PEERS_CONNECTED_MULTI,
&[direction, transport],
*peers_connected_mutli
.get(&(direction, transport))
.unwrap_or(&0) as i64,
);
}
}
}
} }
enum ConnectingType { enum ConnectingType {

View File

@ -4,7 +4,7 @@ use std::net::IpAddr;
use std::task::{Context, Poll}; use std::task::{Context, Poll};
use futures::StreamExt; use futures::StreamExt;
use libp2p::core::{multiaddr, ConnectedPoint}; use libp2p::core::ConnectedPoint;
use libp2p::identity::PeerId; use libp2p::identity::PeerId;
use libp2p::swarm::behaviour::{ConnectionClosed, ConnectionEstablished, DialFailure, FromSwarm}; use libp2p::swarm::behaviour::{ConnectionClosed, ConnectionEstablished, DialFailure, FromSwarm};
use libp2p::swarm::dial_opts::{DialOpts, PeerCondition}; use libp2p::swarm::dial_opts::{DialOpts, PeerCondition};
@ -243,35 +243,11 @@ impl<TSpec: EthSpec> PeerManager<TSpec> {
self.events.push(PeerManagerEvent::MetaData(peer_id)); self.events.push(PeerManagerEvent::MetaData(peer_id));
} }
// increment prometheus metrics // Update the prometheus metrics
if self.metrics_enabled { if self.metrics_enabled {
let remote_addr = endpoint.get_remote_address();
let direction = if endpoint.is_dialer() {
"outbound"
} else {
"inbound"
};
match remote_addr.iter().find(|proto| {
matches!(
proto,
multiaddr::Protocol::QuicV1 | multiaddr::Protocol::Tcp(_)
)
}) {
Some(multiaddr::Protocol::QuicV1) => {
metrics::inc_gauge_vec(&metrics::PEERS_CONNECTED_MULTI, &[direction, "quic"]);
}
Some(multiaddr::Protocol::Tcp(_)) => {
metrics::inc_gauge_vec(&metrics::PEERS_CONNECTED_MULTI, &[direction, "tcp"]);
}
Some(_) => unreachable!(),
None => {
error!(self.log, "Connection established via unknown transport"; "addr" => %remote_addr)
}
};
metrics::inc_gauge(&metrics::PEERS_CONNECTED);
metrics::inc_counter(&metrics::PEER_CONNECT_EVENT_COUNT); metrics::inc_counter(&metrics::PEER_CONNECT_EVENT_COUNT);
self.update_peer_count_metrics();
} }
// Count dialing peers in the limit if the peer dialed us. // Count dialing peers in the limit if the peer dialed us.
@ -309,7 +285,7 @@ impl<TSpec: EthSpec> PeerManager<TSpec> {
fn on_connection_closed( fn on_connection_closed(
&mut self, &mut self,
peer_id: PeerId, peer_id: PeerId,
endpoint: &ConnectedPoint, _endpoint: &ConnectedPoint,
remaining_established: usize, remaining_established: usize,
) { ) {
if remaining_established > 0 { if remaining_established > 0 {
@ -337,33 +313,12 @@ impl<TSpec: EthSpec> PeerManager<TSpec> {
// reference so that peer manager can track this peer. // reference so that peer manager can track this peer.
self.inject_disconnect(&peer_id); self.inject_disconnect(&peer_id);
let remote_addr = endpoint.get_remote_address();
// Update the prometheus metrics // Update the prometheus metrics
if self.metrics_enabled { if self.metrics_enabled {
let direction = if endpoint.is_dialer() {
"outbound"
} else {
"inbound"
};
match remote_addr.iter().find(|proto| {
matches!(
proto,
multiaddr::Protocol::QuicV1 | multiaddr::Protocol::Tcp(_)
)
}) {
Some(multiaddr::Protocol::QuicV1) => {
metrics::dec_gauge_vec(&metrics::PEERS_CONNECTED_MULTI, &[direction, "quic"]);
}
Some(multiaddr::Protocol::Tcp(_)) => {
metrics::dec_gauge_vec(&metrics::PEERS_CONNECTED_MULTI, &[direction, "tcp"]);
}
// If it's an unknown protocol we already logged when connection was established.
_ => {}
};
// Legacy standard metrics. // Legacy standard metrics.
metrics::dec_gauge(&metrics::PEERS_CONNECTED);
metrics::inc_counter(&metrics::PEER_DISCONNECT_EVENT_COUNT); metrics::inc_counter(&metrics::PEER_DISCONNECT_EVENT_COUNT);
self.update_peer_count_metrics();
} }
} }

View File

@ -1251,7 +1251,6 @@ impl BannedPeersCount {
mod tests { mod tests {
use super::*; use super::*;
use libp2p::core::multiaddr::Protocol; use libp2p::core::multiaddr::Protocol;
use libp2p::core::Multiaddr;
use slog::{o, Drain}; use slog::{o, Drain};
use std::net::{Ipv4Addr, Ipv6Addr}; use std::net::{Ipv4Addr, Ipv6Addr};
use types::MinimalEthSpec; use types::MinimalEthSpec;

View File

@ -3,7 +3,7 @@ use crate::rpc::{
codec::base::OutboundCodec, codec::base::OutboundCodec,
protocol::{Encoding, ProtocolId, RPCError, SupportedProtocol, ERROR_TYPE_MAX, ERROR_TYPE_MIN}, protocol::{Encoding, ProtocolId, RPCError, SupportedProtocol, ERROR_TYPE_MAX, ERROR_TYPE_MIN},
}; };
use crate::rpc::{InboundRequest, OutboundRequest, RPCCodedResponse, RPCResponse}; use crate::rpc::{InboundRequest, OutboundRequest};
use libp2p::bytes::BytesMut; use libp2p::bytes::BytesMut;
use snap::read::FrameDecoder; use snap::read::FrameDecoder;
use snap::write::FrameEncoder; use snap::write::FrameEncoder;
@ -590,9 +590,18 @@ fn handle_rpc_response<T: EthSpec>(
SupportedProtocol::MetaDataV1 => Ok(Some(RPCResponse::MetaData(MetaData::V1( SupportedProtocol::MetaDataV1 => Ok(Some(RPCResponse::MetaData(MetaData::V1(
MetaDataV1::from_ssz_bytes(decoded_buffer)?, MetaDataV1::from_ssz_bytes(decoded_buffer)?,
)))), )))),
SupportedProtocol::LightClientBootstrapV1 => Ok(Some(RPCResponse::LightClientBootstrap( SupportedProtocol::LightClientBootstrapV1 => match fork_name {
LightClientBootstrap::from_ssz_bytes(decoded_buffer)?, Some(fork_name) => Ok(Some(RPCResponse::LightClientBootstrap(Arc::new(
))), LightClientBootstrap::from_ssz_bytes(decoded_buffer, fork_name)?,
)))),
None => Err(RPCError::ErrorResponse(
RPCResponseErrorCode::InvalidRequest,
format!(
"No context bytes provided for {:?} response",
versioned_protocol
),
)),
},
// MetaData V2 responses have no context bytes, so behave similarly to V1 responses // MetaData V2 responses have no context bytes, so behave similarly to V1 responses
SupportedProtocol::MetaDataV2 => Ok(Some(RPCResponse::MetaData(MetaData::V2( SupportedProtocol::MetaDataV2 => Ok(Some(RPCResponse::MetaData(MetaData::V2(
MetaDataV2::from_ssz_bytes(decoded_buffer)?, MetaDataV2::from_ssz_bytes(decoded_buffer)?,
@ -676,22 +685,13 @@ fn context_bytes_to_fork_name(
mod tests { mod tests {
use super::*; use super::*;
use crate::rpc::{protocol::*, MetaData}; use crate::rpc::protocol::*;
use crate::{ use crate::types::{EnrAttestationBitfield, EnrSyncCommitteeBitfield};
rpc::{methods::StatusMessage, Ping, RPCResponseErrorCode},
types::{EnrAttestationBitfield, EnrSyncCommitteeBitfield},
};
use std::sync::Arc;
use types::{ use types::{
blob_sidecar::BlobIdentifier, BeaconBlock, BeaconBlockAltair, BeaconBlockBase, blob_sidecar::BlobIdentifier, BeaconBlock, BeaconBlockAltair, BeaconBlockBase,
BeaconBlockMerge, ChainSpec, EmptyBlock, Epoch, ForkContext, FullPayload, Hash256, BeaconBlockMerge, EmptyBlock, Epoch, FullPayload, Signature, Slot,
Signature, SignedBeaconBlock, Slot,
}; };
use snap::write::FrameEncoder;
use ssz::Encode;
use std::io::Write;
type Spec = types::MainnetEthSpec; type Spec = types::MainnetEthSpec;
fn fork_context(fork_name: ForkName) -> ForkContext { fn fork_context(fork_name: ForkName) -> ForkContext {

View File

@ -9,7 +9,7 @@ use crate::rpc::outbound::{OutboundFramed, OutboundRequest};
use crate::rpc::protocol::InboundFramed; use crate::rpc::protocol::InboundFramed;
use fnv::FnvHashMap; use fnv::FnvHashMap;
use futures::prelude::*; use futures::prelude::*;
use futures::{Sink, SinkExt}; use futures::SinkExt;
use libp2p::swarm::handler::{ use libp2p::swarm::handler::{
ConnectionEvent, ConnectionHandler, ConnectionHandlerEvent, DialUpgradeError, ConnectionEvent, ConnectionHandler, ConnectionHandlerEvent, DialUpgradeError,
FullyNegotiatedInbound, FullyNegotiatedOutbound, StreamUpgradeError, SubstreamProtocol, FullyNegotiatedInbound, FullyNegotiatedOutbound, StreamUpgradeError, SubstreamProtocol,

View File

@ -6,6 +6,7 @@ use serde::Serialize;
use ssz::Encode; use ssz::Encode;
use ssz_derive::{Decode, Encode}; use ssz_derive::{Decode, Encode};
use ssz_types::{typenum::U256, VariableList}; use ssz_types::{typenum::U256, VariableList};
use std::fmt::Display;
use std::marker::PhantomData; use std::marker::PhantomData;
use std::ops::Deref; use std::ops::Deref;
use std::sync::Arc; use std::sync::Arc;
@ -44,11 +45,13 @@ impl Deref for ErrorType {
} }
} }
impl ToString for ErrorType { impl Display for ErrorType {
fn to_string(&self) -> String { fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
#[allow(clippy::invalid_regex)] #[allow(clippy::invalid_regex)]
let re = Regex::new("\\p{C}").expect("Regex is valid"); let re = Regex::new("\\p{C}").expect("Regex is valid");
String::from_utf8_lossy(&re.replace_all(self.0.deref(), &b""[..])).to_string() let error_type_str =
String::from_utf8_lossy(&re.replace_all(self.0.deref(), &b""[..])).to_string();
write!(f, "{}", error_type_str)
} }
} }
@ -385,7 +388,7 @@ pub enum RPCResponse<T: EthSpec> {
BlobsByRange(Arc<BlobSidecar<T>>), BlobsByRange(Arc<BlobSidecar<T>>),
/// A response to a get LIGHT_CLIENT_BOOTSTRAP request. /// A response to a get LIGHT_CLIENT_BOOTSTRAP request.
LightClientBootstrap(LightClientBootstrap<T>), LightClientBootstrap(Arc<LightClientBootstrap<T>>),
/// A response to a get BLOBS_BY_ROOT request. /// A response to a get BLOBS_BY_ROOT request.
BlobsByRoot(Arc<BlobSidecar<T>>), BlobsByRoot(Arc<BlobSidecar<T>>),
@ -566,11 +569,7 @@ impl<T: EthSpec> std::fmt::Display for RPCResponse<T> {
RPCResponse::Pong(ping) => write!(f, "Pong: {}", ping.data), RPCResponse::Pong(ping) => write!(f, "Pong: {}", ping.data),
RPCResponse::MetaData(metadata) => write!(f, "Metadata: {}", metadata.seq_number()), RPCResponse::MetaData(metadata) => write!(f, "Metadata: {}", metadata.seq_number()),
RPCResponse::LightClientBootstrap(bootstrap) => { RPCResponse::LightClientBootstrap(bootstrap) => {
write!( write!(f, "LightClientBootstrap Slot: {}", bootstrap.get_slot())
f,
"LightClientBootstrap Slot: {}",
bootstrap.header.beacon.slot
)
} }
} }
} }
@ -580,7 +579,7 @@ impl<T: EthSpec> std::fmt::Display for RPCCodedResponse<T> {
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result { fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
match self { match self {
RPCCodedResponse::Success(res) => write!(f, "{}", res), RPCCodedResponse::Success(res) => write!(f, "{}", res),
RPCCodedResponse::Error(code, err) => write!(f, "{}: {}", code, err.to_string()), RPCCodedResponse::Error(code, err) => write!(f, "{}: {}", code, err),
RPCCodedResponse::StreamTermination(_) => write!(f, "Stream Termination"), RPCCodedResponse::StreamTermination(_) => write!(f, "Stream Termination"),
} }
} }

View File

@ -2,11 +2,10 @@ use super::methods::*;
use super::protocol::ProtocolId; use super::protocol::ProtocolId;
use super::protocol::SupportedProtocol; use super::protocol::SupportedProtocol;
use super::RPCError; use super::RPCError;
use crate::rpc::protocol::Encoding; use crate::rpc::codec::{
use crate::rpc::{ base::BaseOutboundCodec, ssz_snappy::SSZSnappyOutboundCodec, OutboundCodec,
codec::{base::BaseOutboundCodec, ssz_snappy::SSZSnappyOutboundCodec, OutboundCodec},
methods::ResponseTermination,
}; };
use crate::rpc::protocol::Encoding;
use futures::future::BoxFuture; use futures::future::BoxFuture;
use futures::prelude::{AsyncRead, AsyncWrite}; use futures::prelude::{AsyncRead, AsyncWrite};
use futures::{FutureExt, SinkExt}; use futures::{FutureExt, SinkExt};

View File

@ -1,8 +1,5 @@
use super::methods::*; use super::methods::*;
use crate::rpc::{ use crate::rpc::codec::{base::BaseInboundCodec, ssz_snappy::SSZSnappyInboundCodec, InboundCodec};
codec::{base::BaseInboundCodec, ssz_snappy::SSZSnappyInboundCodec, InboundCodec},
methods::{MaxErrorLen, ResponseTermination, MAX_ERROR_LEN},
};
use futures::future::BoxFuture; use futures::future::BoxFuture;
use futures::prelude::{AsyncRead, AsyncWrite}; use futures::prelude::{AsyncRead, AsyncWrite};
use futures::{FutureExt, StreamExt}; use futures::{FutureExt, StreamExt};

View File

@ -3,7 +3,6 @@ use crate::rpc::Protocol;
use fnv::FnvHashMap; use fnv::FnvHashMap;
use libp2p::PeerId; use libp2p::PeerId;
use serde::{Deserialize, Serialize}; use serde::{Deserialize, Serialize};
use std::convert::TryInto;
use std::future::Future; use std::future::Future;
use std::hash::Hash; use std::hash::Hash;
use std::pin::Pin; use std::pin::Pin;

View File

@ -93,7 +93,7 @@ pub enum Response<TSpec: EthSpec> {
/// A response to a get BLOBS_BY_ROOT request. /// A response to a get BLOBS_BY_ROOT request.
BlobsByRoot(Option<Arc<BlobSidecar<TSpec>>>), BlobsByRoot(Option<Arc<BlobSidecar<TSpec>>>),
/// A response to a LightClientUpdate request. /// A response to a LightClientUpdate request.
LightClientBootstrap(LightClientBootstrap<TSpec>), LightClientBootstrap(Arc<LightClientBootstrap<TSpec>>),
} }
impl<TSpec: EthSpec> std::convert::From<Response<TSpec>> for RPCCodedResponse<TSpec> { impl<TSpec: EthSpec> std::convert::From<Response<TSpec>> for RPCCodedResponse<TSpec> {

View File

@ -3,8 +3,8 @@ use crate::peer_manager::PeerManager;
use crate::rpc::{ReqId, RPC}; use crate::rpc::{ReqId, RPC};
use crate::types::SnappyTransform; use crate::types::SnappyTransform;
use crate::gossipsub;
use libp2p::identify; use libp2p::identify;
use libp2p::swarm::behaviour::toggle::Toggle;
use libp2p::swarm::NetworkBehaviour; use libp2p::swarm::NetworkBehaviour;
use libp2p::upnp::tokio::Behaviour as Upnp; use libp2p::upnp::tokio::Behaviour as Upnp;
use types::EthSpec; use types::EthSpec;
@ -34,7 +34,7 @@ where
/// Provides IP addresses and peer information. /// Provides IP addresses and peer information.
pub identify: identify::Behaviour, pub identify: identify::Behaviour,
/// Libp2p UPnP port mapping. /// Libp2p UPnP port mapping.
pub upnp: Upnp, pub upnp: Toggle<Upnp>,
/// The routing pub-sub mechanism for eth2. /// The routing pub-sub mechanism for eth2.
pub gossipsub: Gossipsub, pub gossipsub: Gossipsub,
} }

View File

@ -268,8 +268,6 @@ impl futures::stream::Stream for GossipCache {
#[cfg(test)] #[cfg(test)]
mod tests { mod tests {
use crate::types::GossipKind;
use super::*; use super::*;
use futures::stream::StreamExt; use futures::stream::StreamExt;

View File

@ -1,9 +1,9 @@
use crate::gossipsub::{ use crate::types::{GossipEncoding, GossipKind, GossipTopic};
use crate::{error, TopicHash};
use gossipsub::{
Config as GossipsubConfig, IdentTopic as Topic, PeerScoreParams, PeerScoreThresholds, Config as GossipsubConfig, IdentTopic as Topic, PeerScoreParams, PeerScoreThresholds,
TopicScoreParams, TopicScoreParams,
}; };
use crate::types::{GossipEncoding, GossipKind, GossipTopic};
use crate::{error, TopicHash};
use std::cmp::max; use std::cmp::max;
use std::collections::HashMap; use std::collections::HashMap;
use std::marker::PhantomData; use std::marker::PhantomData;

View File

@ -4,10 +4,6 @@ use crate::config::{gossipsub_config, GossipsubConfigParams, NetworkLoad};
use crate::discovery::{ use crate::discovery::{
subnet_predicate, DiscoveredPeers, Discovery, FIND_NODE_QUERY_CLOSEST_PEERS, subnet_predicate, DiscoveredPeers, Discovery, FIND_NODE_QUERY_CLOSEST_PEERS,
}; };
use crate::gossipsub::{
self, IdentTopic as Topic, MessageAcceptance, MessageAuthenticity, MessageId, PublishError,
TopicScoreParams,
};
use crate::peer_manager::{ use crate::peer_manager::{
config::Config as PeerManagerCfg, peerdb::score::PeerAction, peerdb::score::ReportSource, config::Config as PeerManagerCfg, peerdb::score::PeerAction, peerdb::score::ReportSource,
ConnectionDirection, PeerManager, PeerManagerEvent, ConnectionDirection, PeerManager, PeerManagerEvent,
@ -27,8 +23,13 @@ use crate::Eth2Enr;
use crate::{error, metrics, Enr, NetworkGlobals, PubsubMessage, TopicHash}; use crate::{error, metrics, Enr, NetworkGlobals, PubsubMessage, TopicHash};
use api_types::{PeerRequestId, Request, RequestId, Response}; use api_types::{PeerRequestId, Request, RequestId, Response};
use futures::stream::StreamExt; use futures::stream::StreamExt;
use gossipsub::{
IdentTopic as Topic, MessageAcceptance, MessageAuthenticity, MessageId, PublishError,
TopicScoreParams,
};
use gossipsub_scoring_parameters::{lighthouse_gossip_thresholds, PeerScoreSettings}; use gossipsub_scoring_parameters::{lighthouse_gossip_thresholds, PeerScoreSettings};
use libp2p::multiaddr::{self, Multiaddr, Protocol as MProtocol}; use libp2p::multiaddr::{self, Multiaddr, Protocol as MProtocol};
use libp2p::swarm::behaviour::toggle::Toggle;
use libp2p::swarm::{Swarm, SwarmEvent}; use libp2p::swarm::{Swarm, SwarmEvent};
use libp2p::{identify, PeerId, SwarmBuilder}; use libp2p::{identify, PeerId, SwarmBuilder};
use slog::{crit, debug, info, o, trace, warn}; use slog::{crit, debug, info, o, trace, warn};
@ -255,6 +256,8 @@ impl<AppReqId: ReqId, TSpec: EthSpec> Network<AppReqId, TSpec> {
config.network_load, config.network_load,
ctx.fork_context.clone(), ctx.fork_context.clone(),
gossipsub_config_params, gossipsub_config_params,
ctx.chain_spec.seconds_per_slot,
TSpec::slots_per_epoch(),
); );
// If metrics are enabled for libp2p build the configuration // If metrics are enabled for libp2p build the configuration
@ -379,6 +382,11 @@ impl<AppReqId: ReqId, TSpec: EthSpec> Network<AppReqId, TSpec> {
libp2p::connection_limits::Behaviour::new(limits) libp2p::connection_limits::Behaviour::new(limits)
}; };
let upnp = Toggle::from(
config
.upnp_enabled
.then_some(libp2p::upnp::tokio::Behaviour::default()),
);
let behaviour = { let behaviour = {
Behaviour { Behaviour {
gossipsub, gossipsub,
@ -387,7 +395,7 @@ impl<AppReqId: ReqId, TSpec: EthSpec> Network<AppReqId, TSpec> {
identify, identify,
peer_manager, peer_manager,
connection_limits, connection_limits,
upnp: Default::default(), upnp,
} }
}; };

View File

@ -1,4 +1,3 @@
use crate::gossipsub;
use crate::multiaddr::Protocol; use crate::multiaddr::Protocol;
use crate::rpc::{MetaData, MetaDataV1, MetaDataV2}; use crate::rpc::{MetaData, MetaDataV1, MetaDataV2};
use crate::types::{ use crate::types::{

View File

@ -1,11 +1,9 @@
//! Handles the encoding and decoding of pubsub messages. //! Handles the encoding and decoding of pubsub messages.
use crate::gossipsub;
use crate::types::{GossipEncoding, GossipKind, GossipTopic}; use crate::types::{GossipEncoding, GossipKind, GossipTopic};
use crate::TopicHash; use crate::TopicHash;
use snap::raw::{decompress_len, Decoder, Encoder}; use snap::raw::{decompress_len, Decoder, Encoder};
use ssz::{Decode, Encode}; use ssz::{Decode, Encode};
use std::boxed::Box;
use std::io::{Error, ErrorKind}; use std::io::{Error, ErrorKind};
use std::sync::Arc; use std::sync::Arc;
use types::{ use types::{
@ -265,17 +263,31 @@ impl<T: EthSpec> PubsubMessage<T> {
))) )))
} }
GossipKind::LightClientFinalityUpdate => { GossipKind::LightClientFinalityUpdate => {
let light_client_finality_update = let light_client_finality_update = match fork_context.from_context_bytes(gossip_topic.fork_digest) {
LightClientFinalityUpdate::from_ssz_bytes(data) Some(&fork_name) => {
.map_err(|e| format!("{:?}", e))?; LightClientFinalityUpdate::from_ssz_bytes(data, fork_name)
.map_err(|e| format!("{:?}", e))?
},
None => return Err(format!(
"light_client_finality_update topic invalid for given fork digest {:?}",
gossip_topic.fork_digest
)),
};
Ok(PubsubMessage::LightClientFinalityUpdate(Box::new( Ok(PubsubMessage::LightClientFinalityUpdate(Box::new(
light_client_finality_update, light_client_finality_update,
))) )))
} }
GossipKind::LightClientOptimisticUpdate => { GossipKind::LightClientOptimisticUpdate => {
let light_client_optimistic_update = let light_client_optimistic_update = match fork_context.from_context_bytes(gossip_topic.fork_digest) {
LightClientOptimisticUpdate::from_ssz_bytes(data) Some(&fork_name) => {
.map_err(|e| format!("{:?}", e))?; LightClientOptimisticUpdate::from_ssz_bytes(data, fork_name)
.map_err(|e| format!("{:?}", e))?
},
None => return Err(format!(
"light_client_optimistic_update topic invalid for given fork digest {:?}",
gossip_topic.fork_digest
)),
};
Ok(PubsubMessage::LightClientOptimisticUpdate(Box::new( Ok(PubsubMessage::LightClientOptimisticUpdate(Box::new(
light_client_optimistic_update, light_client_optimistic_update,
))) )))

View File

@ -1,4 +1,4 @@
use crate::gossipsub::{IdentTopic as Topic, TopicHash}; use gossipsub::{IdentTopic as Topic, TopicHash};
use serde::{Deserialize, Serialize}; use serde::{Deserialize, Serialize};
use strum::AsRefStr; use strum::AsRefStr;
use types::{ChainSpec, EthSpec, ForkName, SubnetId, SyncSubnetId, Unsigned}; use types::{ChainSpec, EthSpec, ForkName, SubnetId, SyncSubnetId, Unsigned};

View File

@ -1,5 +1,4 @@
#![cfg(test)] #![cfg(test)]
use lighthouse_network::gossipsub;
use lighthouse_network::service::Network as LibP2PService; use lighthouse_network::service::Network as LibP2PService;
use lighthouse_network::Enr; use lighthouse_network::Enr;
use lighthouse_network::EnrExt; use lighthouse_network::EnrExt;

View File

@ -11,6 +11,7 @@ matches = "0.1.8"
slog-term = { workspace = true } slog-term = { workspace = true }
slog-async = { workspace = true } slog-async = { workspace = true }
eth2 = { workspace = true } eth2 = { workspace = true }
gossipsub = { workspace = true }
[dependencies] [dependencies]
async-channel = { workspace = true } async-channel = { workspace = true }

View File

@ -963,7 +963,7 @@ impl<T: BeaconChainTypes> NetworkBeaconProcessor<T> {
self.propagate_validation_result(message_id, peer_id, MessageAcceptance::Ignore); self.propagate_validation_result(message_id, peer_id, MessageAcceptance::Ignore);
return None; return None;
} }
Err(BlockError::BlockIsAlreadyKnown) => { Err(BlockError::BlockIsAlreadyKnown(_)) => {
debug!( debug!(
self.log, self.log,
"Gossip block is already known"; "Gossip block is already known";

View File

@ -6,7 +6,6 @@ use beacon_chain::{BeaconChainError, BeaconChainTypes, HistoricalBlockError, Whe
use beacon_processor::SendOnDrop; use beacon_processor::SendOnDrop;
use itertools::process_results; use itertools::process_results;
use lighthouse_network::rpc::methods::{BlobsByRangeRequest, BlobsByRootRequest}; use lighthouse_network::rpc::methods::{BlobsByRangeRequest, BlobsByRootRequest};
use lighthouse_network::rpc::StatusMessage;
use lighthouse_network::rpc::*; use lighthouse_network::rpc::*;
use lighthouse_network::{PeerId, PeerRequestId, ReportSource, Response, SyncInfo}; use lighthouse_network::{PeerId, PeerRequestId, ReportSource, Response, SyncInfo};
use slog::{debug, error, warn}; use slog::{debug, error, warn};
@ -305,7 +304,7 @@ impl<T: BeaconChainTypes> NetworkBeaconProcessor<T> {
match self.chain.get_light_client_bootstrap(&block_root) { match self.chain.get_light_client_bootstrap(&block_root) {
Ok(Some((bootstrap, _))) => self.send_response( Ok(Some((bootstrap, _))) => self.send_response(
peer_id, peer_id,
Response::LightClientBootstrap(bootstrap), Response::LightClientBootstrap(Arc::new(bootstrap)),
request_id, request_id,
), ),
Ok(None) => self.send_error_response( Ok(None) => self.send_error_response(

View File

@ -117,6 +117,7 @@ impl<T: BeaconChainTypes> NetworkBeaconProcessor<T> {
"Gossip block is being processed"; "Gossip block is being processed";
"action" => "sending rpc block to reprocessing queue", "action" => "sending rpc block to reprocessing queue",
"block_root" => %block_root, "block_root" => %block_root,
"process_type" => ?process_type,
); );
// Send message to work reprocess queue to retry the block // Send message to work reprocess queue to retry the block
@ -149,6 +150,7 @@ impl<T: BeaconChainTypes> NetworkBeaconProcessor<T> {
"proposer" => block.message().proposer_index(), "proposer" => block.message().proposer_index(),
"slot" => block.slot(), "slot" => block.slot(),
"commitments" => commitments_formatted, "commitments" => commitments_formatted,
"process_type" => ?process_type,
); );
let result = self let result = self
@ -267,6 +269,7 @@ impl<T: BeaconChainTypes> NetworkBeaconProcessor<T> {
"slot" => %slot, "slot" => %slot,
"block_hash" => %hash, "block_hash" => %hash,
); );
self.chain.recompute_head_at_current_slot().await;
} }
Ok(AvailabilityProcessingStatus::MissingComponents(_, _)) => { Ok(AvailabilityProcessingStatus::MissingComponents(_, _)) => {
debug!( debug!(
@ -276,7 +279,7 @@ impl<T: BeaconChainTypes> NetworkBeaconProcessor<T> {
"slot" => %slot, "slot" => %slot,
); );
} }
Err(BlockError::BlockIsAlreadyKnown) => { Err(BlockError::BlockIsAlreadyKnown(_)) => {
debug!( debug!(
self.log, self.log,
"Blobs have already been imported"; "Blobs have already been imported";
@ -417,7 +420,11 @@ impl<T: BeaconChainTypes> NetworkBeaconProcessor<T> {
} }
} }
(imported_blocks, Ok(_)) => { (imported_blocks, Ok(_)) => {
debug!(self.log, "Parent lookup processed successfully"); debug!(
self.log, "Parent lookup processed successfully";
"chain_hash" => %chain_head,
"imported_blocks" => imported_blocks
);
BatchProcessResult::Success { BatchProcessResult::Success {
was_non_empty: imported_blocks > 0, was_non_empty: imported_blocks > 0,
} }
@ -639,7 +646,7 @@ impl<T: BeaconChainTypes> NetworkBeaconProcessor<T> {
peer_action: Some(PeerAction::LowToleranceError), peer_action: Some(PeerAction::LowToleranceError),
}) })
} }
BlockError::BlockIsAlreadyKnown => { BlockError::BlockIsAlreadyKnown(_) => {
// This can happen for many reasons. Head sync's can download multiples and parent // This can happen for many reasons. Head sync's can download multiples and parent
// lookups can download blocks before range sync // lookups can download blocks before range sync
Ok(()) Ok(())

View File

@ -60,11 +60,10 @@ impl StoreItem for PersistedDht {
#[cfg(test)] #[cfg(test)]
mod tests { mod tests {
use super::*; use super::*;
use lighthouse_network::Enr;
use sloggers::{null::NullLoggerBuilder, Build}; use sloggers::{null::NullLoggerBuilder, Build};
use std::str::FromStr; use std::str::FromStr;
use store::config::StoreConfig; use store::config::StoreConfig;
use store::{HotColdDB, MemoryStore}; use store::MemoryStore;
use types::{ChainSpec, MinimalEthSpec}; use types::{ChainSpec, MinimalEthSpec};
#[test] #[test]
fn test_persisted_dht() { fn test_persisted_dht() {

Some files were not shown because too many files have changed in this diff Show More