mirror of
https://gitlab.com/pulsechaincom/prysm-pulse.git
synced 2024-12-25 12:57:18 +00:00
7896e087e5
* add in better logging * gaz * last one * mess about * check strings * fix
148 lines
4.8 KiB
Go
148 lines
4.8 KiB
Go
package sync
|
|
|
|
import (
|
|
"context"
|
|
"reflect"
|
|
"strings"
|
|
|
|
libp2pcore "github.com/libp2p/go-libp2p-core"
|
|
"github.com/libp2p/go-libp2p-core/helpers"
|
|
"github.com/libp2p/go-libp2p-core/mux"
|
|
"github.com/libp2p/go-libp2p-core/network"
|
|
"github.com/prysmaticlabs/prysm/beacon-chain/p2p"
|
|
pb "github.com/prysmaticlabs/prysm/proto/beacon/p2p/v1"
|
|
"github.com/prysmaticlabs/prysm/shared/params"
|
|
"github.com/prysmaticlabs/prysm/shared/roughtime"
|
|
"github.com/prysmaticlabs/prysm/shared/traceutil"
|
|
"go.opencensus.io/trace"
|
|
)
|
|
|
|
// Time to first byte timeout. The maximum time to wait for first byte of
|
|
// request response (time-to-first-byte). The client is expected to give up if
|
|
// they don't receive the first byte within 5 seconds.
|
|
var ttfbTimeout = params.BeaconNetworkConfig().TtfbTimeout
|
|
|
|
// respTimeout is the maximum time for complete response transfer.
|
|
var respTimeout = params.BeaconNetworkConfig().RespTimeout
|
|
|
|
// rpcHandler is responsible for handling and responding to any incoming message.
|
|
// This method may return an error to internal monitoring, but the error will
|
|
// not be relayed to the peer.
|
|
type rpcHandler func(context.Context, interface{}, libp2pcore.Stream) error
|
|
|
|
// registerRPCHandlers for p2p RPC.
|
|
func (s *Service) registerRPCHandlers() {
|
|
s.registerRPC(
|
|
p2p.RPCStatusTopic,
|
|
&pb.Status{},
|
|
s.statusRPCHandler,
|
|
)
|
|
s.registerRPC(
|
|
p2p.RPCGoodByeTopic,
|
|
new(uint64),
|
|
s.goodbyeRPCHandler,
|
|
)
|
|
s.registerRPC(
|
|
p2p.RPCBlocksByRangeTopic,
|
|
&pb.BeaconBlocksByRangeRequest{},
|
|
s.beaconBlocksByRangeRPCHandler,
|
|
)
|
|
s.registerRPC(
|
|
p2p.RPCBlocksByRootTopic,
|
|
[][32]byte{},
|
|
s.beaconBlocksRootRPCHandler,
|
|
)
|
|
s.registerRPC(
|
|
p2p.RPCPingTopic,
|
|
new(uint64),
|
|
s.pingHandler,
|
|
)
|
|
s.registerRPC(
|
|
p2p.RPCMetaDataTopic,
|
|
new(interface{}),
|
|
s.metaDataHandler,
|
|
)
|
|
}
|
|
|
|
// registerRPC for a given topic with an expected protobuf message type.
|
|
func (s *Service) registerRPC(topic string, base interface{}, handle rpcHandler) {
|
|
topic += s.p2p.Encoding().ProtocolSuffix()
|
|
log := log.WithField("topic", topic)
|
|
s.p2p.SetStreamHandler(topic, func(stream network.Stream) {
|
|
ctx, cancel := context.WithTimeout(context.Background(), ttfbTimeout)
|
|
defer cancel()
|
|
defer func() {
|
|
if err := helpers.FullClose(stream); err != nil && err.Error() != mux.ErrReset.Error() {
|
|
log.WithError(err).Debug("Failed to reset stream")
|
|
}
|
|
}()
|
|
ctx, span := trace.StartSpan(ctx, "sync.rpc")
|
|
defer span.End()
|
|
span.AddAttributes(trace.StringAttribute("topic", topic))
|
|
span.AddAttributes(trace.StringAttribute("peer", stream.Conn().RemotePeer().Pretty()))
|
|
log := log.WithField("peer", stream.Conn().RemotePeer().Pretty())
|
|
|
|
if err := stream.SetReadDeadline(roughtime.Now().Add(ttfbTimeout)); err != nil {
|
|
log.WithError(err).Error("Could not set stream read deadline")
|
|
return
|
|
}
|
|
|
|
// Increment message received counter.
|
|
messageReceivedCounter.WithLabelValues(topic).Inc()
|
|
|
|
// since metadata requests do not have any data in the payload, we
|
|
// do not decode anything.
|
|
if strings.Contains(topic, p2p.RPCMetaDataTopic) {
|
|
if err := handle(ctx, new(interface{}), stream); err != nil {
|
|
messageFailedProcessingCounter.WithLabelValues(topic).Inc()
|
|
if err != errWrongForkDigestVersion {
|
|
log.WithError(err).Warn("Failed to handle p2p RPC")
|
|
}
|
|
traceutil.AnnotateError(span, err)
|
|
}
|
|
return
|
|
}
|
|
|
|
// Given we have an input argument that can be pointer or [][32]byte, this gives us
|
|
// a way to check for its reflect.Kind and based on the result, we can decode
|
|
// accordingly.
|
|
t := reflect.TypeOf(base)
|
|
if t.Kind() == reflect.Ptr {
|
|
msg := reflect.New(t.Elem())
|
|
if err := s.p2p.Encoding().DecodeWithMaxLength(stream, msg.Interface()); err != nil {
|
|
// Debug logs for goodbye/status errors
|
|
if strings.Contains(topic, p2p.RPCGoodByeTopic) || strings.Contains(topic, p2p.RPCStatusTopic) {
|
|
log.WithError(err).Debug("Failed to decode goodbye stream message")
|
|
traceutil.AnnotateError(span, err)
|
|
return
|
|
}
|
|
log.WithError(err).Warn("Failed to decode stream message")
|
|
traceutil.AnnotateError(span, err)
|
|
return
|
|
}
|
|
if err := handle(ctx, msg.Interface(), stream); err != nil {
|
|
messageFailedProcessingCounter.WithLabelValues(topic).Inc()
|
|
if err != errWrongForkDigestVersion {
|
|
log.WithError(err).Warn("Failed to handle p2p RPC")
|
|
}
|
|
traceutil.AnnotateError(span, err)
|
|
}
|
|
} else {
|
|
msg := reflect.New(t)
|
|
if err := s.p2p.Encoding().DecodeWithMaxLength(stream, msg.Interface()); err != nil {
|
|
log.WithError(err).Warn("Failed to decode stream message")
|
|
traceutil.AnnotateError(span, err)
|
|
return
|
|
}
|
|
if err := handle(ctx, msg.Elem().Interface(), stream); err != nil {
|
|
messageFailedProcessingCounter.WithLabelValues(topic).Inc()
|
|
if err != errWrongForkDigestVersion {
|
|
log.WithError(err).Warn("Failed to handle p2p RPC")
|
|
}
|
|
traceutil.AnnotateError(span, err)
|
|
}
|
|
}
|
|
|
|
})
|
|
}
|