* F3-370: integrate F3 dynamic manifest * F3-370: make linter happy * Set manifest sender identities * Decode manifest sender peer ID from string before using it Peer ID is of type string internally but the internal string representation is not the same as the encoded string representation. Therefore, the latter needs to be decoded and cannot be casted to the former. Otherwise, it will represent a different ID. * `make gen` the pain of my life --------- Co-authored-by: adlrocha <6717133+adlrocha@users.noreply.github.com> Co-authored-by: Masih H. Derkani <m@derkani.org>
700 lines
21 KiB
Go
700 lines
21 KiB
Go
package lp2p
|
|
|
|
import (
|
|
"context"
|
|
"encoding/json"
|
|
"fmt"
|
|
"net"
|
|
"time"
|
|
|
|
pubsub "github.com/libp2p/go-libp2p-pubsub"
|
|
pubsub_pb "github.com/libp2p/go-libp2p-pubsub/pb"
|
|
"github.com/libp2p/go-libp2p/core/host"
|
|
"github.com/libp2p/go-libp2p/core/peer"
|
|
"github.com/minio/blake2b-simd"
|
|
ma "github.com/multiformats/go-multiaddr"
|
|
"go.opencensus.io/stats"
|
|
"go.uber.org/fx"
|
|
"golang.org/x/xerrors"
|
|
|
|
"github.com/filecoin-project/go-f3/gpbft"
|
|
"github.com/filecoin-project/go-f3/manifest"
|
|
|
|
"github.com/filecoin-project/lotus/build"
|
|
"github.com/filecoin-project/lotus/metrics"
|
|
"github.com/filecoin-project/lotus/node/config"
|
|
"github.com/filecoin-project/lotus/node/modules/dtypes"
|
|
"github.com/filecoin-project/lotus/node/modules/helpers"
|
|
"github.com/filecoin-project/lotus/node/modules/tracer"
|
|
)
|
|
|
|
func init() {
|
|
// configure larger overlay parameters
|
|
pubsub.GossipSubD = 8
|
|
pubsub.GossipSubDscore = 6
|
|
pubsub.GossipSubDout = 3
|
|
pubsub.GossipSubDlo = 6
|
|
pubsub.GossipSubDhi = 12
|
|
pubsub.GossipSubDlazy = 12
|
|
pubsub.GossipSubDirectConnectInitialDelay = 30 * time.Second
|
|
pubsub.GossipSubIWantFollowupTime = 5 * time.Second
|
|
pubsub.GossipSubHistoryLength = 10
|
|
pubsub.GossipSubGossipFactor = 0.1
|
|
}
|
|
|
|
const (
|
|
GossipScoreThreshold = -500
|
|
PublishScoreThreshold = -1000
|
|
GraylistScoreThreshold = -2500
|
|
AcceptPXScoreThreshold = 1000
|
|
OpportunisticGraftScoreThreshold = 3.5
|
|
|
|
// Determines the max. number of configuration changes
|
|
// that are allowed for the dynamic manifest.
|
|
// If the manifest changes more than this number, the F3
|
|
// message topic will be filtered
|
|
MaxDynamicManifestChangesAllowed = 1000
|
|
)
|
|
|
|
func ScoreKeeper() *dtypes.ScoreKeeper {
|
|
return new(dtypes.ScoreKeeper)
|
|
}
|
|
|
|
type PeerScoreTracker interface {
|
|
UpdatePeerScore(scores map[peer.ID]*pubsub.PeerScoreSnapshot)
|
|
}
|
|
|
|
type peerScoreTracker struct {
|
|
sk *dtypes.ScoreKeeper
|
|
lt tracer.LotusTracer
|
|
}
|
|
|
|
func newPeerScoreTracker(lt tracer.LotusTracer, sk *dtypes.ScoreKeeper) PeerScoreTracker {
|
|
return &peerScoreTracker{
|
|
sk: sk,
|
|
lt: lt,
|
|
}
|
|
}
|
|
|
|
func (pst *peerScoreTracker) UpdatePeerScore(scores map[peer.ID]*pubsub.PeerScoreSnapshot) {
|
|
if pst.lt != nil {
|
|
pst.lt.PeerScores(scores)
|
|
}
|
|
|
|
pst.sk.Update(scores)
|
|
}
|
|
|
|
type GossipIn struct {
|
|
fx.In
|
|
Mctx helpers.MetricsCtx
|
|
Lc fx.Lifecycle
|
|
Host host.Host
|
|
Nn dtypes.NetworkName
|
|
Bp dtypes.BootstrapPeers
|
|
Db dtypes.DrandBootstrap
|
|
Cfg *config.Pubsub
|
|
Sk *dtypes.ScoreKeeper
|
|
Dr dtypes.DrandSchedule
|
|
}
|
|
|
|
func getDrandTopic(chainInfoJSON string) (string, error) {
|
|
var drandInfo = struct {
|
|
Hash string `json:"hash"`
|
|
}{}
|
|
err := json.Unmarshal([]byte(chainInfoJSON), &drandInfo)
|
|
if err != nil {
|
|
return "", xerrors.Errorf("could not unmarshal drand chain info: %w", err)
|
|
}
|
|
return "/drand/pubsub/v0.0.0/" + drandInfo.Hash, nil
|
|
}
|
|
|
|
func GossipSub(in GossipIn) (service *pubsub.PubSub, err error) {
|
|
bootstrappers := make(map[peer.ID]struct{})
|
|
for _, pi := range in.Bp {
|
|
bootstrappers[pi.ID] = struct{}{}
|
|
}
|
|
drandBootstrappers := make(map[peer.ID]struct{})
|
|
for _, pi := range in.Db {
|
|
drandBootstrappers[pi.ID] = struct{}{}
|
|
}
|
|
|
|
isBootstrapNode := in.Cfg.Bootstrapper
|
|
|
|
drandTopicParams := &pubsub.TopicScoreParams{
|
|
// expected 2 beaconsn/min
|
|
TopicWeight: 0.5, // 5x block topic; max cap is 62.5
|
|
|
|
// 1 tick per second, maxes at 1 after 1 hour
|
|
TimeInMeshWeight: 0.00027, // ~1/3600
|
|
TimeInMeshQuantum: time.Second,
|
|
TimeInMeshCap: 1,
|
|
|
|
// deliveries decay after 1 hour, cap at 25 beacons
|
|
FirstMessageDeliveriesWeight: 5, // max value is 125
|
|
FirstMessageDeliveriesDecay: pubsub.ScoreParameterDecay(time.Hour),
|
|
FirstMessageDeliveriesCap: 25, // the maximum expected in an hour is ~26, including the decay
|
|
|
|
// Mesh Delivery Failure is currently turned off for beacons
|
|
// This is on purpose as
|
|
// - the traffic is very low for meaningful distribution of incoming edges.
|
|
// - the reaction time needs to be very slow -- in the order of 10 min at least
|
|
// so we might as well let opportunistic grafting repair the mesh on its own
|
|
// pace.
|
|
// - the network is too small, so large asymmetries can be expected between mesh
|
|
// edges.
|
|
// We should revisit this once the network grows.
|
|
|
|
// invalid messages decay after 1 hour
|
|
InvalidMessageDeliveriesWeight: -1000,
|
|
InvalidMessageDeliveriesDecay: pubsub.ScoreParameterDecay(time.Hour),
|
|
}
|
|
|
|
ingestTopicParams := &pubsub.TopicScoreParams{
|
|
// expected ~0.5 confirmed deals / min. sampled
|
|
TopicWeight: 0.1,
|
|
|
|
TimeInMeshWeight: 0.00027, // ~1/3600
|
|
TimeInMeshQuantum: time.Second,
|
|
TimeInMeshCap: 1,
|
|
|
|
FirstMessageDeliveriesWeight: 0.5,
|
|
FirstMessageDeliveriesDecay: pubsub.ScoreParameterDecay(time.Hour),
|
|
FirstMessageDeliveriesCap: 100, // allowing for burstiness
|
|
|
|
InvalidMessageDeliveriesWeight: -1000,
|
|
InvalidMessageDeliveriesDecay: pubsub.ScoreParameterDecay(time.Hour),
|
|
}
|
|
|
|
topicParams := map[string]*pubsub.TopicScoreParams{
|
|
build.BlocksTopic(in.Nn): {
|
|
// expected 10 blocks/min
|
|
TopicWeight: 0.1, // max cap is 50, max mesh penalty is -10, single invalid message is -100
|
|
|
|
// 1 tick per second, maxes at 1 after 1 hour
|
|
TimeInMeshWeight: 0.00027, // ~1/3600
|
|
TimeInMeshQuantum: time.Second,
|
|
TimeInMeshCap: 1,
|
|
|
|
// deliveries decay after 1 hour, cap at 100 blocks
|
|
FirstMessageDeliveriesWeight: 5, // max value is 500
|
|
FirstMessageDeliveriesDecay: pubsub.ScoreParameterDecay(time.Hour),
|
|
FirstMessageDeliveriesCap: 100, // 100 blocks in an hour
|
|
|
|
// Mesh Delivery Failure is currently turned off for blocks
|
|
// This is on purpose as
|
|
// - the traffic is very low for meaningful distribution of incoming edges.
|
|
// - the reaction time needs to be very slow -- in the order of 10 min at least
|
|
// so we might as well let opportunistic grafting repair the mesh on its own
|
|
// pace.
|
|
// - the network is too small, so large asymmetries can be expected between mesh
|
|
// edges.
|
|
// We should revisit this once the network grows.
|
|
//
|
|
// // tracks deliveries in the last minute
|
|
// // penalty activates at 1 minute and expects ~0.4 blocks
|
|
// MeshMessageDeliveriesWeight: -576, // max penalty is -100
|
|
// MeshMessageDeliveriesDecay: pubsub.ScoreParameterDecay(time.Minute),
|
|
// MeshMessageDeliveriesCap: 10, // 10 blocks in a minute
|
|
// MeshMessageDeliveriesThreshold: 0.41666, // 10/12/2 blocks/min
|
|
// MeshMessageDeliveriesWindow: 10 * time.Millisecond,
|
|
// MeshMessageDeliveriesActivation: time.Minute,
|
|
//
|
|
// // decays after 15 min
|
|
// MeshFailurePenaltyWeight: -576,
|
|
// MeshFailurePenaltyDecay: pubsub.ScoreParameterDecay(15 * time.Minute),
|
|
|
|
// invalid messages decay after 1 hour
|
|
InvalidMessageDeliveriesWeight: -1000,
|
|
InvalidMessageDeliveriesDecay: pubsub.ScoreParameterDecay(time.Hour),
|
|
},
|
|
build.MessagesTopic(in.Nn): {
|
|
// expected > 1 tx/second
|
|
TopicWeight: 0.1, // max cap is 5, single invalid message is -100
|
|
|
|
// 1 tick per second, maxes at 1 hour
|
|
TimeInMeshWeight: 0.0002778, // ~1/3600
|
|
TimeInMeshQuantum: time.Second,
|
|
TimeInMeshCap: 1,
|
|
|
|
// deliveries decay after 10min, cap at 100 tx
|
|
FirstMessageDeliveriesWeight: 0.5, // max value is 50
|
|
FirstMessageDeliveriesDecay: pubsub.ScoreParameterDecay(10 * time.Minute),
|
|
FirstMessageDeliveriesCap: 100, // 100 messages in 10 minutes
|
|
|
|
// Mesh Delivery Failure is currently turned off for messages
|
|
// This is on purpose as the network is still too small, which results in
|
|
// asymmetries and potential unmeshing from negative scores.
|
|
// // tracks deliveries in the last minute
|
|
// // penalty activates at 1 min and expects 2.5 txs
|
|
// MeshMessageDeliveriesWeight: -16, // max penalty is -100
|
|
// MeshMessageDeliveriesDecay: pubsub.ScoreParameterDecay(time.Minute),
|
|
// MeshMessageDeliveriesCap: 100, // 100 txs in a minute
|
|
// MeshMessageDeliveriesThreshold: 2.5, // 60/12/2 txs/minute
|
|
// MeshMessageDeliveriesWindow: 10 * time.Millisecond,
|
|
// MeshMessageDeliveriesActivation: time.Minute,
|
|
|
|
// // decays after 5min
|
|
// MeshFailurePenaltyWeight: -16,
|
|
// MeshFailurePenaltyDecay: pubsub.ScoreParameterDecay(5 * time.Minute),
|
|
|
|
// invalid messages decay after 1 hour
|
|
InvalidMessageDeliveriesWeight: -1000,
|
|
InvalidMessageDeliveriesDecay: pubsub.ScoreParameterDecay(time.Hour),
|
|
},
|
|
}
|
|
|
|
pgTopicWeights := map[string]float64{
|
|
build.BlocksTopic(in.Nn): 10,
|
|
build.MessagesTopic(in.Nn): 1,
|
|
}
|
|
|
|
var drandTopics []string
|
|
for _, d := range in.Dr {
|
|
topic, err := getDrandTopic(d.Config.ChainInfoJSON)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
topicParams[topic] = drandTopicParams
|
|
pgTopicWeights[topic] = 5
|
|
drandTopics = append(drandTopics, topic)
|
|
}
|
|
|
|
// Index ingestion whitelist
|
|
topicParams[build.IndexerIngestTopic(in.Nn)] = ingestTopicParams
|
|
|
|
// IP colocation whitelist
|
|
var ipcoloWhitelist []*net.IPNet
|
|
for _, cidr := range in.Cfg.IPColocationWhitelist {
|
|
_, ipnet, err := net.ParseCIDR(cidr)
|
|
if err != nil {
|
|
return nil, xerrors.Errorf("error parsing IPColocation subnet %s: %w", cidr, err)
|
|
}
|
|
ipcoloWhitelist = append(ipcoloWhitelist, ipnet)
|
|
}
|
|
|
|
options := []pubsub.Option{
|
|
// Gossipsubv1.1 configuration
|
|
pubsub.WithFloodPublish(true),
|
|
pubsub.WithMessageIdFn(HashMsgId),
|
|
pubsub.WithPeerScore(
|
|
&pubsub.PeerScoreParams{
|
|
AppSpecificScore: func(p peer.ID) float64 {
|
|
// return a heavy positive score for bootstrappers so that we don't unilaterally prune
|
|
// them and accept PX from them.
|
|
// we don't do that in the bootstrappers themselves to avoid creating a closed mesh
|
|
// between them (however we might want to consider doing just that)
|
|
_, ok := bootstrappers[p]
|
|
if ok && !isBootstrapNode {
|
|
return 2500
|
|
}
|
|
|
|
_, ok = drandBootstrappers[p]
|
|
if ok && !isBootstrapNode {
|
|
return 1500
|
|
}
|
|
|
|
// TODO: we want to plug the application specific score to the node itself in order
|
|
// to provide feedback to the pubsub system based on observed behaviour
|
|
return 0
|
|
},
|
|
AppSpecificWeight: 1,
|
|
|
|
// This sets the IP colocation threshold to 5 peers before we apply penalties
|
|
IPColocationFactorThreshold: 5,
|
|
IPColocationFactorWeight: -100,
|
|
IPColocationFactorWhitelist: ipcoloWhitelist,
|
|
|
|
// P7: behavioural penalties, decay after 1hr
|
|
BehaviourPenaltyThreshold: 6,
|
|
BehaviourPenaltyWeight: -10,
|
|
BehaviourPenaltyDecay: pubsub.ScoreParameterDecay(time.Hour),
|
|
|
|
DecayInterval: pubsub.DefaultDecayInterval,
|
|
DecayToZero: pubsub.DefaultDecayToZero,
|
|
|
|
// this retains non-positive scores for 6 hours
|
|
RetainScore: 6 * time.Hour,
|
|
|
|
// topic parameters
|
|
Topics: topicParams,
|
|
},
|
|
&pubsub.PeerScoreThresholds{
|
|
GossipThreshold: GossipScoreThreshold,
|
|
PublishThreshold: PublishScoreThreshold,
|
|
GraylistThreshold: GraylistScoreThreshold,
|
|
AcceptPXThreshold: AcceptPXScoreThreshold,
|
|
OpportunisticGraftThreshold: OpportunisticGraftScoreThreshold,
|
|
},
|
|
),
|
|
}
|
|
|
|
// enable Peer eXchange on bootstrappers
|
|
if isBootstrapNode {
|
|
// turn off the mesh in bootstrappers -- only do gossip and PX
|
|
pubsub.GossipSubD = 0
|
|
pubsub.GossipSubDscore = 0
|
|
pubsub.GossipSubDlo = 0
|
|
pubsub.GossipSubDhi = 0
|
|
pubsub.GossipSubDout = 0
|
|
pubsub.GossipSubDlazy = 64
|
|
pubsub.GossipSubGossipFactor = 0.25
|
|
pubsub.GossipSubPruneBackoff = 5 * time.Minute
|
|
// turn on PX
|
|
options = append(options, pubsub.WithPeerExchange(true))
|
|
}
|
|
|
|
// direct peers
|
|
if in.Cfg.DirectPeers != nil {
|
|
var directPeerInfo []peer.AddrInfo
|
|
|
|
for _, addr := range in.Cfg.DirectPeers {
|
|
a, err := ma.NewMultiaddr(addr)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
pi, err := peer.AddrInfoFromP2pAddr(a)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
directPeerInfo = append(directPeerInfo, *pi)
|
|
}
|
|
|
|
options = append(options, pubsub.WithDirectPeers(directPeerInfo))
|
|
}
|
|
|
|
// validation queue RED
|
|
var pgParams *pubsub.PeerGaterParams
|
|
|
|
if isBootstrapNode {
|
|
pgParams = pubsub.NewPeerGaterParams(
|
|
0.33,
|
|
pubsub.ScoreParameterDecay(2*time.Minute),
|
|
pubsub.ScoreParameterDecay(10*time.Minute),
|
|
).WithTopicDeliveryWeights(pgTopicWeights)
|
|
} else {
|
|
pgParams = pubsub.NewPeerGaterParams(
|
|
0.33,
|
|
pubsub.ScoreParameterDecay(2*time.Minute),
|
|
pubsub.ScoreParameterDecay(time.Hour),
|
|
).WithTopicDeliveryWeights(pgTopicWeights)
|
|
}
|
|
|
|
options = append(options, pubsub.WithPeerGater(pgParams))
|
|
|
|
allowTopics := []string{
|
|
build.BlocksTopic(in.Nn),
|
|
build.MessagesTopic(in.Nn),
|
|
build.IndexerIngestTopic(in.Nn),
|
|
}
|
|
|
|
if build.IsF3Enabled() {
|
|
f3TopicName := manifest.PubSubTopicFromNetworkName(gpbft.NetworkName(in.Nn))
|
|
allowTopics = append(allowTopics, f3TopicName)
|
|
|
|
// allow dynamic manifest topic and the new topic names after a reconfiguration.
|
|
// Note: This is pretty ugly, but I tried to use a regex subscription filter
|
|
// as the commented code below, but unfortunately it overwrites previous filters. A simple fix would
|
|
// be to allow combining several topic filters, but for now this works.
|
|
//
|
|
// pattern := fmt.Sprintf(`^\/f3\/%s\/0\.0\.1\/?[0-9]*$`, in.Nn)
|
|
// rx, err := regexp.Compile(pattern)
|
|
// if err != nil {
|
|
// return nil, xerrors.Errorf("failed to compile manifest topic regex: %w", err)
|
|
// }
|
|
// options = append(options,
|
|
// pubsub.WithSubscriptionFilter(
|
|
// pubsub.WrapLimitSubscriptionFilter(
|
|
// pubsub.NewRegexpSubscriptionFilter(rx),
|
|
// 100)))
|
|
allowTopics = append(allowTopics, manifest.ManifestPubSubTopicName)
|
|
for i := 0; i < MaxDynamicManifestChangesAllowed; i++ {
|
|
allowTopics = append(allowTopics, f3TopicName+"/"+fmt.Sprintf("%d", i))
|
|
}
|
|
}
|
|
|
|
allowTopics = append(allowTopics, drandTopics...)
|
|
options = append(options,
|
|
pubsub.WithSubscriptionFilter(
|
|
pubsub.WrapLimitSubscriptionFilter(
|
|
pubsub.NewAllowlistSubscriptionFilter(allowTopics...),
|
|
100)))
|
|
|
|
var transports []tracer.TracerTransport
|
|
if in.Cfg.JsonTracer != "" {
|
|
jsonTransport, err := tracer.NewJsonTracerTransport(in.Cfg.JsonTracer)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
transports = append(transports, jsonTransport)
|
|
}
|
|
|
|
tps := make([]string, 0) // range of topics that will be submited to the traces
|
|
addTopicToList := func(topicList []string, newTopic string) []string {
|
|
// check if the topic is already in the list
|
|
for _, tp := range topicList {
|
|
if tp == newTopic {
|
|
return topicList
|
|
}
|
|
}
|
|
// add it otherwise
|
|
return append(topicList, newTopic)
|
|
}
|
|
|
|
// tracer
|
|
if in.Cfg.ElasticSearchTracer != "" {
|
|
elasticSearchTransport, err := tracer.NewElasticSearchTransport(
|
|
in.Cfg.ElasticSearchTracer,
|
|
in.Cfg.ElasticSearchIndex,
|
|
)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
transports = append(transports, elasticSearchTransport)
|
|
tps = addTopicToList(tps, build.BlocksTopic(in.Nn))
|
|
}
|
|
|
|
lt := tracer.NewLotusTracer(transports, in.Host.ID(), in.Cfg.TracerSourceAuth)
|
|
|
|
// tracer
|
|
if in.Cfg.RemoteTracer != "" {
|
|
a, err := ma.NewMultiaddr(in.Cfg.RemoteTracer)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
pi, err := peer.AddrInfoFromP2pAddr(a)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
tr, err := pubsub.NewRemoteTracer(context.TODO(), in.Host, *pi)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
pst := newPeerScoreTracker(lt, in.Sk)
|
|
trw := newTracerWrapper(tr, lt, tps...)
|
|
|
|
options = append(options, pubsub.WithEventTracer(trw))
|
|
options = append(options, pubsub.WithPeerScoreInspect(pst.UpdatePeerScore, 10*time.Second))
|
|
} else {
|
|
// still instantiate a tracer for collecting metrics
|
|
pst := newPeerScoreTracker(lt, in.Sk)
|
|
trw := newTracerWrapper(nil, lt, tps...)
|
|
|
|
options = append(options, pubsub.WithEventTracer(trw))
|
|
options = append(options, pubsub.WithPeerScoreInspect(pst.UpdatePeerScore, 10*time.Second))
|
|
}
|
|
|
|
return pubsub.NewGossipSub(helpers.LifecycleCtx(in.Mctx, in.Lc), in.Host, options...)
|
|
}
|
|
|
|
func HashMsgId(m *pubsub_pb.Message) string {
|
|
hash := blake2b.Sum256(m.Data)
|
|
return string(hash[:])
|
|
}
|
|
|
|
func newTracerWrapper(
|
|
lp2pTracer pubsub.EventTracer,
|
|
lotusTracer pubsub.EventTracer,
|
|
topics ...string,
|
|
) pubsub.EventTracer {
|
|
var topicsMap map[string]struct{}
|
|
if len(topics) > 0 {
|
|
topicsMap = make(map[string]struct{})
|
|
for _, topic := range topics {
|
|
topicsMap[topic] = struct{}{}
|
|
}
|
|
}
|
|
return &tracerWrapper{lp2pTracer: lp2pTracer, lotusTracer: lotusTracer, topics: topicsMap}
|
|
}
|
|
|
|
type tracerWrapper struct {
|
|
lp2pTracer pubsub.EventTracer
|
|
lotusTracer pubsub.EventTracer
|
|
topics map[string]struct{}
|
|
}
|
|
|
|
func (trw *tracerWrapper) traceMessage(topic string) bool {
|
|
_, ok := trw.topics[topic]
|
|
return ok
|
|
}
|
|
|
|
func (trw *tracerWrapper) Trace(evt *pubsub_pb.TraceEvent) {
|
|
// this filters the trace events reported to the remote tracer to include only
|
|
// JOIN/LEAVE/GRAFT/PRUNE/PUBLISH/DELIVER. This significantly reduces bandwidth usage and still
|
|
// collects enough data to recover the state of the mesh and compute message delivery latency
|
|
// distributions.
|
|
// Furthermore, we only trace message publication and deliveries for specified topics
|
|
// (here just the blocks topic).
|
|
switch evt.GetType() {
|
|
case pubsub_pb.TraceEvent_PUBLISH_MESSAGE:
|
|
stats.Record(context.TODO(), metrics.PubsubPublishMessage.M(1))
|
|
if trw.traceMessage(evt.GetPublishMessage().GetTopic()) {
|
|
if trw.lp2pTracer != nil {
|
|
trw.lp2pTracer.Trace(evt)
|
|
}
|
|
if trw.lotusTracer != nil {
|
|
trw.lotusTracer.Trace(evt)
|
|
}
|
|
}
|
|
|
|
case pubsub_pb.TraceEvent_DELIVER_MESSAGE:
|
|
stats.Record(context.TODO(), metrics.PubsubDeliverMessage.M(1))
|
|
|
|
case pubsub_pb.TraceEvent_REJECT_MESSAGE:
|
|
stats.Record(context.TODO(), metrics.PubsubRejectMessage.M(1))
|
|
if trw.traceMessage(evt.GetRejectMessage().GetTopic()) {
|
|
if trw.lp2pTracer != nil {
|
|
trw.lp2pTracer.Trace(evt)
|
|
}
|
|
if trw.lotusTracer != nil {
|
|
trw.lotusTracer.Trace(evt)
|
|
}
|
|
}
|
|
|
|
case pubsub_pb.TraceEvent_DUPLICATE_MESSAGE:
|
|
stats.Record(context.TODO(), metrics.PubsubDuplicateMessage.M(1))
|
|
if trw.traceMessage(evt.GetDuplicateMessage().GetTopic()) {
|
|
if trw.lp2pTracer != nil {
|
|
trw.lp2pTracer.Trace(evt)
|
|
}
|
|
if trw.lotusTracer != nil {
|
|
trw.lotusTracer.Trace(evt)
|
|
}
|
|
}
|
|
|
|
case pubsub_pb.TraceEvent_JOIN:
|
|
if trw.traceMessage(evt.GetJoin().GetTopic()) {
|
|
if trw.lp2pTracer != nil {
|
|
trw.lp2pTracer.Trace(evt)
|
|
}
|
|
if trw.lotusTracer != nil {
|
|
trw.lotusTracer.Trace(evt)
|
|
}
|
|
}
|
|
case pubsub_pb.TraceEvent_LEAVE:
|
|
if trw.traceMessage(evt.GetLeave().GetTopic()) {
|
|
if trw.lp2pTracer != nil {
|
|
trw.lp2pTracer.Trace(evt)
|
|
}
|
|
if trw.lotusTracer != nil {
|
|
trw.lotusTracer.Trace(evt)
|
|
}
|
|
}
|
|
|
|
case pubsub_pb.TraceEvent_GRAFT:
|
|
if trw.traceMessage(evt.GetGraft().GetTopic()) {
|
|
if trw.lp2pTracer != nil {
|
|
trw.lp2pTracer.Trace(evt)
|
|
}
|
|
|
|
if trw.lotusTracer != nil {
|
|
trw.lotusTracer.Trace(evt)
|
|
}
|
|
}
|
|
|
|
case pubsub_pb.TraceEvent_PRUNE:
|
|
stats.Record(context.TODO(), metrics.PubsubPruneMessage.M(1))
|
|
if trw.traceMessage(evt.GetPrune().GetTopic()) {
|
|
if trw.lp2pTracer != nil {
|
|
trw.lp2pTracer.Trace(evt)
|
|
}
|
|
if trw.lotusTracer != nil {
|
|
trw.lotusTracer.Trace(evt)
|
|
}
|
|
}
|
|
|
|
case pubsub_pb.TraceEvent_ADD_PEER:
|
|
if trw.lp2pTracer != nil {
|
|
trw.lp2pTracer.Trace(evt)
|
|
}
|
|
if trw.lotusTracer != nil {
|
|
trw.lotusTracer.Trace(evt)
|
|
}
|
|
|
|
case pubsub_pb.TraceEvent_REMOVE_PEER:
|
|
if trw.lp2pTracer != nil {
|
|
trw.lp2pTracer.Trace(evt)
|
|
}
|
|
if trw.lotusTracer != nil {
|
|
trw.lotusTracer.Trace(evt)
|
|
}
|
|
|
|
case pubsub_pb.TraceEvent_RECV_RPC:
|
|
stats.Record(context.TODO(), metrics.PubsubRecvRPC.M(1))
|
|
// only track the RPC Calls from IWANT / IHAVE / BLOCK topic
|
|
controlRPC := evt.GetRecvRPC().GetMeta().GetControl()
|
|
ihave := controlRPC.GetIhave()
|
|
iwant := controlRPC.GetIwant()
|
|
msgsRPC := evt.GetRecvRPC().GetMeta().GetMessages()
|
|
|
|
// check if any of the messages we are sending belong to a trackable topic
|
|
var validTopic = false
|
|
for _, topic := range msgsRPC {
|
|
if trw.traceMessage(topic.GetTopic()) {
|
|
validTopic = true
|
|
break
|
|
}
|
|
}
|
|
// track if the Iwant / Ihave messages are from a valid Topic
|
|
var validIhave = false
|
|
for _, msgs := range ihave {
|
|
if trw.traceMessage(msgs.GetTopic()) {
|
|
validIhave = true
|
|
break
|
|
}
|
|
}
|
|
// check if we have any of iwant msgs (it doesn't classify per topic - just msg.ID)
|
|
validIwant := len(iwant) > 0
|
|
|
|
// trace the event if any of the flags was triggered
|
|
if validIhave || validIwant || validTopic {
|
|
if trw.lp2pTracer != nil {
|
|
trw.lp2pTracer.Trace(evt)
|
|
}
|
|
if trw.lotusTracer != nil {
|
|
trw.lotusTracer.Trace(evt)
|
|
}
|
|
}
|
|
case pubsub_pb.TraceEvent_SEND_RPC:
|
|
stats.Record(context.TODO(), metrics.PubsubSendRPC.M(1))
|
|
// only track the RPC Calls from IWANT / IHAVE / BLOCK topic
|
|
controlRPC := evt.GetSendRPC().GetMeta().GetControl()
|
|
ihave := controlRPC.GetIhave()
|
|
iwant := controlRPC.GetIwant()
|
|
msgsRPC := evt.GetSendRPC().GetMeta().GetMessages()
|
|
|
|
// check if any of the messages we are sending belong to a trackable topic
|
|
var validTopic = false
|
|
for _, topic := range msgsRPC {
|
|
if trw.traceMessage(topic.GetTopic()) {
|
|
validTopic = true
|
|
break
|
|
}
|
|
}
|
|
// track if the Iwant / Ihave messages are from a valid Topic
|
|
var validIhave = false
|
|
for _, msgs := range ihave {
|
|
if trw.traceMessage(msgs.GetTopic()) {
|
|
validIhave = true
|
|
break
|
|
}
|
|
}
|
|
// check if there was any of the Iwant msgs
|
|
validIwant := len(iwant) > 0
|
|
|
|
// trace the msgs if any of the flags was triggered
|
|
if validIhave || validIwant || validTopic {
|
|
if trw.lp2pTracer != nil {
|
|
trw.lp2pTracer.Trace(evt)
|
|
}
|
|
if trw.lotusTracer != nil {
|
|
trw.lotusTracer.Trace(evt)
|
|
}
|
|
}
|
|
case pubsub_pb.TraceEvent_DROP_RPC:
|
|
stats.Record(context.TODO(), metrics.PubsubDropRPC.M(1))
|
|
}
|
|
}
|