mirror of
https://gitlab.com/pulsechaincom/erigon-pulse.git
synced 2024-12-21 19:20:39 +00:00
509a7af26a
This fixes an issue where the mumbai testnet node struggle to find peers. Before this fix in general test peer numbers are typically around 20 in total between eth66, eth67 and eth68. For new peers some can struggle to find even a single peer after days of operation. These are the numbers after 12 hours or running on a node which previously could not find any peers: eth66=13, eth67=76, eth68=91. The root cause of this issue is the following: - A significant number of mumbai peers around the boot node return network ids which are different from those currently available in the DHT - The available nodes are all consequently busy and return 'too many peers' for long periods These issues case a significant number of discovery timeouts, some of the queries will never receive a response. This causes the discovery read loop to enter a channel deadlock - which means that no responses are processed, nor timeouts fired. This causes the discovery process in the node to stop. From then on it just re-requests handshakes from a relatively small number of peers. This check in fixes this situation with the following changes: - Remove the deadlock by running the timer in a separate go-routine so it can run independently of the main request processing. - Allow the discovery process matcher to match on port if no id match can be established on initial ping. This allows subsequent node validation to proceed and if the node proves to be valid via the remainder of the look-up and handshake process it us used as a valid peer. - Completely unsolicited responses, i.e. those which come from a completely unknown ip:port combination continue to be ignored. -
365 lines
11 KiB
Go
365 lines
11 KiB
Go
// Copyright 2014 The go-ethereum Authors
|
|
// This file is part of the go-ethereum library.
|
|
//
|
|
// The go-ethereum library is free software: you can redistribute it and/or modify
|
|
// it under the terms of the GNU Lesser General Public License as published by
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
// (at your option) any later version.
|
|
//
|
|
// The go-ethereum library is distributed in the hope that it will be useful,
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
// GNU Lesser General Public License for more details.
|
|
//
|
|
// You should have received a copy of the GNU Lesser General Public License
|
|
// along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
package p2p
|
|
|
|
import (
|
|
"encoding/binary"
|
|
"errors"
|
|
"fmt"
|
|
"math/rand"
|
|
"net"
|
|
"reflect"
|
|
"strconv"
|
|
"strings"
|
|
"testing"
|
|
"time"
|
|
|
|
"github.com/ledgerwatch/log/v3"
|
|
|
|
"github.com/ledgerwatch/erigon/p2p/enode"
|
|
"github.com/ledgerwatch/erigon/p2p/enr"
|
|
)
|
|
|
|
var discard = Protocol{
|
|
Name: "discard",
|
|
Length: 1,
|
|
Run: func(p *Peer, rw MsgReadWriter) *PeerError {
|
|
for {
|
|
msg, err := rw.ReadMsg()
|
|
if err != nil {
|
|
return NewPeerError(PeerErrorTest, DiscProtocolError, err, "peer_test: 'discard' protocol ReadMsg error")
|
|
}
|
|
fmt.Printf("discarding %d\n", msg.Code)
|
|
msg.Discard()
|
|
}
|
|
},
|
|
}
|
|
|
|
// uintID encodes i into a node ID.
|
|
func uintID(i uint16) enode.ID {
|
|
var id enode.ID
|
|
binary.BigEndian.PutUint16(id[:], i)
|
|
return id
|
|
}
|
|
|
|
// newNode creates a node record with the given address.
|
|
func newNode(id enode.ID, addr string) *enode.Node {
|
|
var r enr.Record
|
|
if addr != "" {
|
|
// Set the port if present.
|
|
if strings.Contains(addr, ":") {
|
|
hs, ps, err := net.SplitHostPort(addr)
|
|
if err != nil {
|
|
panic(fmt.Errorf("invalid address %q", addr))
|
|
}
|
|
port, err := strconv.Atoi(ps)
|
|
if err != nil {
|
|
panic(fmt.Errorf("invalid port in %q", addr))
|
|
}
|
|
r.Set(enr.TCP(port))
|
|
r.Set(enr.UDP(port))
|
|
addr = hs
|
|
}
|
|
// Set the IP.
|
|
ip := net.ParseIP(addr)
|
|
if ip == nil {
|
|
panic(fmt.Errorf("invalid IP %q", addr))
|
|
}
|
|
r.Set(enr.IP(ip))
|
|
}
|
|
return enode.SignNull(&r, id)
|
|
}
|
|
|
|
func testPeer(protos []Protocol) (func(), *conn, *Peer, <-chan *PeerError) {
|
|
var (
|
|
fd1, fd2 = net.Pipe()
|
|
key1, key2 = newkey(), newkey()
|
|
t1 = newTestTransport(&key2.PublicKey, fd1, nil)
|
|
t2 = newTestTransport(&key1.PublicKey, fd2, &key1.PublicKey)
|
|
)
|
|
|
|
c1 := &conn{fd: fd1, node: newNode(uintID(1), ""), transport: t1}
|
|
c2 := &conn{fd: fd2, node: newNode(uintID(2), ""), transport: t2}
|
|
for _, p := range protos {
|
|
c1.caps = append(c1.caps, p.cap())
|
|
c2.caps = append(c2.caps, p.cap())
|
|
}
|
|
|
|
peer := newPeer(log.Root(), c1, protos, [64]byte{1}, true)
|
|
errc := make(chan *PeerError, 1)
|
|
go func() {
|
|
err := peer.run()
|
|
errc <- err
|
|
}()
|
|
|
|
closer := func() { c2.close(errors.New("close func called")) }
|
|
return closer, c2, peer, errc
|
|
}
|
|
|
|
func TestPeerProtoReadMsg(t *testing.T) {
|
|
proto := Protocol{
|
|
Name: "a",
|
|
Length: 5,
|
|
Run: func(peer *Peer, rw MsgReadWriter) *PeerError {
|
|
if err := ExpectMsg(rw, 2, []uint{1}); err != nil {
|
|
t.Error(err)
|
|
}
|
|
if err := ExpectMsg(rw, 3, []uint{2}); err != nil {
|
|
t.Error(err)
|
|
}
|
|
if err := ExpectMsg(rw, 4, []uint{3}); err != nil {
|
|
t.Error(err)
|
|
}
|
|
return nil
|
|
},
|
|
}
|
|
|
|
closer, rw, _, errc := testPeer([]Protocol{proto})
|
|
defer closer()
|
|
|
|
Send(rw, baseProtocolLength+2, []uint{1})
|
|
Send(rw, baseProtocolLength+3, []uint{2})
|
|
Send(rw, baseProtocolLength+4, []uint{3})
|
|
|
|
select {
|
|
case err := <-errc:
|
|
if (err != nil) && (err.Reason != DiscQuitting) {
|
|
t.Errorf("peer returned error: %v", err)
|
|
}
|
|
case <-time.After(2 * time.Second):
|
|
t.Errorf("receive timeout")
|
|
}
|
|
}
|
|
|
|
func TestPeerProtoEncodeMsg(t *testing.T) {
|
|
proto := Protocol{
|
|
Name: "a",
|
|
Length: 2,
|
|
Run: func(peer *Peer, rw MsgReadWriter) *PeerError {
|
|
if err := SendItems(rw, 2); err == nil {
|
|
t.Error("expected error for out-of-range msg code, got nil")
|
|
}
|
|
if err := SendItems(rw, 1, "foo", "bar"); err != nil {
|
|
t.Errorf("write error: %v", err)
|
|
}
|
|
return nil
|
|
},
|
|
}
|
|
closer, rw, _, _ := testPeer([]Protocol{proto})
|
|
defer closer()
|
|
|
|
if err := ExpectMsg(rw, 17, []string{"foo", "bar"}); err != nil {
|
|
t.Error(err)
|
|
}
|
|
}
|
|
|
|
func TestPeerPing(t *testing.T) {
|
|
closer, rw, _, _ := testPeer(nil)
|
|
defer closer()
|
|
if err := SendItems(rw, pingMsg); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
if err := ExpectMsg(rw, pongMsg, nil); err != nil {
|
|
t.Error(err)
|
|
}
|
|
}
|
|
|
|
// This test checks that a disconnect message sent by a peer is returned
|
|
// as the error from Peer.run.
|
|
func TestPeerDisconnect(t *testing.T) {
|
|
closer, rw, _, disc := testPeer(nil)
|
|
defer closer()
|
|
|
|
if err := SendItems(rw, discMsg, DiscQuitting); err != nil {
|
|
t.Fatal(err)
|
|
}
|
|
select {
|
|
case reason := <-disc:
|
|
if !strings.Contains(reason.Error(), DiscQuitting.Error()) {
|
|
t.Errorf("run returned wrong reason: got %v, want %v", reason.Error(), DiscQuitting.Error())
|
|
}
|
|
case <-time.After(500 * time.Millisecond):
|
|
t.Error("peer did not return")
|
|
}
|
|
}
|
|
|
|
// This test is supposed to verify that Peer can reliably handle
|
|
// multiple causes of disconnection occurring at the same time.
|
|
func TestPeerDisconnectRace(t *testing.T) {
|
|
maybe := func() bool { return rand.Intn(2) == 1 }
|
|
|
|
for i := 0; i < 1000; i++ {
|
|
protoclose := make(chan *PeerError)
|
|
protodisc := make(chan *PeerError)
|
|
closer, rw, p, disc := testPeer([]Protocol{
|
|
{
|
|
Name: "closereq",
|
|
Run: func(p *Peer, rw MsgReadWriter) *PeerError { return <-protoclose },
|
|
Length: 1,
|
|
},
|
|
{
|
|
Name: "disconnect",
|
|
Run: func(p *Peer, rw MsgReadWriter) *PeerError {
|
|
p.Disconnect(<-protodisc)
|
|
return nil
|
|
},
|
|
Length: 1,
|
|
},
|
|
})
|
|
|
|
// Simulate incoming messages.
|
|
go SendItems(rw, baseProtocolLength+1)
|
|
go SendItems(rw, baseProtocolLength+2)
|
|
// Close the network connection.
|
|
go closer()
|
|
// Make protocol "closereq" return.
|
|
protoclose <- NewPeerError(PeerErrorTest, DiscRequested, nil, "peer_test.TestPeerDisconnectRace: protocol closed")
|
|
// Make protocol "disconnect" call peer.Disconnect
|
|
protodisc <- NewPeerError(PeerErrorTest, DiscAlreadyConnected, nil, "peer_test.TestPeerDisconnectRace: protocol called peer.Disconnect()")
|
|
// In some cases, simulate something else calling peer.Disconnect.
|
|
if maybe() {
|
|
go p.Disconnect(NewPeerError(PeerErrorTest, DiscInvalidIdentity, nil, "peer_test.TestPeerDisconnectRace: something else called peer.Disconnect()"))
|
|
}
|
|
// In some cases, simulate remote requesting a disconnect.
|
|
if maybe() {
|
|
go SendItems(rw, discMsg, DiscQuitting)
|
|
}
|
|
|
|
select {
|
|
case <-disc:
|
|
case <-time.After(2 * time.Second):
|
|
// Peer.run should return quickly. If it doesn't the Peer
|
|
// goroutines are probably deadlocked. Call panic in order to
|
|
// show the stacks.
|
|
panic("Peer.run took to long to return.")
|
|
}
|
|
}
|
|
}
|
|
|
|
func TestNewPeer(t *testing.T) {
|
|
name := "nodename"
|
|
caps := []Cap{{"foo", 2}, {"bar", 3}}
|
|
id := randomID()
|
|
p := NewPeer(id, [64]byte{1}, name, caps, true)
|
|
if p.ID() != id {
|
|
t.Errorf("ID mismatch: got %v, expected %v", p.ID(), id)
|
|
}
|
|
if p.Name() != name {
|
|
t.Errorf("Name mismatch: got %v, expected %v", p.Name(), name)
|
|
}
|
|
if !reflect.DeepEqual(p.Caps(), caps) {
|
|
t.Errorf("Caps mismatch: got %v, expected %v", p.Caps(), caps)
|
|
}
|
|
|
|
p.Disconnect(NewPeerError(PeerErrorTest, DiscAlreadyConnected, nil, "TestNewPeer Disconnect")) // Should not hang
|
|
}
|
|
|
|
func TestMatchProtocols(t *testing.T) {
|
|
tests := []struct {
|
|
Remote []Cap
|
|
Local []Protocol
|
|
Match map[string]protoRW
|
|
}{
|
|
{
|
|
// No remote capabilities
|
|
Local: []Protocol{{Name: "a"}},
|
|
},
|
|
{
|
|
// No local protocols
|
|
Remote: []Cap{{Name: "a"}},
|
|
},
|
|
{
|
|
// No mutual protocols
|
|
Remote: []Cap{{Name: "a"}},
|
|
Local: []Protocol{{Name: "b"}},
|
|
},
|
|
{
|
|
// Some matches, some differences
|
|
Remote: []Cap{{Name: "local"}, {Name: "match1"}, {Name: "match2"}},
|
|
Local: []Protocol{{Name: "match1"}, {Name: "match2"}, {Name: "remote"}},
|
|
Match: map[string]protoRW{"match1": {Protocol: Protocol{Name: "match1"}}, "match2": {Protocol: Protocol{Name: "match2"}}},
|
|
},
|
|
{
|
|
// Various alphabetical ordering
|
|
Remote: []Cap{{Name: "aa"}, {Name: "ab"}, {Name: "bb"}, {Name: "ba"}},
|
|
Local: []Protocol{{Name: "ba"}, {Name: "bb"}, {Name: "ab"}, {Name: "aa"}},
|
|
Match: map[string]protoRW{"aa": {Protocol: Protocol{Name: "aa"}}, "ab": {Protocol: Protocol{Name: "ab"}}, "ba": {Protocol: Protocol{Name: "ba"}}, "bb": {Protocol: Protocol{Name: "bb"}}},
|
|
},
|
|
{
|
|
// No mutual versions
|
|
Remote: []Cap{{Version: 1}},
|
|
Local: []Protocol{{Version: 2}},
|
|
},
|
|
{
|
|
// Multiple versions, single common
|
|
Remote: []Cap{{Version: 1}, {Version: 2}},
|
|
Local: []Protocol{{Version: 2}, {Version: 3}},
|
|
Match: map[string]protoRW{"": {Protocol: Protocol{Version: 2}}},
|
|
},
|
|
{
|
|
// Multiple versions, multiple common
|
|
Remote: []Cap{{Version: 1}, {Version: 2}, {Version: 3}, {Version: 4}},
|
|
Local: []Protocol{{Version: 2}, {Version: 3}},
|
|
Match: map[string]protoRW{"": {Protocol: Protocol{Version: 3}}},
|
|
},
|
|
{
|
|
// Various version orderings
|
|
Remote: []Cap{{Version: 4}, {Version: 1}, {Version: 3}, {Version: 2}},
|
|
Local: []Protocol{{Version: 2}, {Version: 3}, {Version: 1}},
|
|
Match: map[string]protoRW{"": {Protocol: Protocol{Version: 3}}},
|
|
},
|
|
{
|
|
// Versions overriding sub-protocol lengths
|
|
Remote: []Cap{{Version: 1}, {Version: 2}, {Version: 3}, {Name: "a"}},
|
|
Local: []Protocol{{Version: 1, Length: 1}, {Version: 2, Length: 2}, {Version: 3, Length: 3}, {Name: "a"}},
|
|
Match: map[string]protoRW{"": {Protocol: Protocol{Version: 3}}, "a": {Protocol: Protocol{Name: "a"}, offset: 3}},
|
|
},
|
|
}
|
|
|
|
for i, tt := range tests {
|
|
result := matchProtocols(tt.Local, tt.Remote, nil, log.Root())
|
|
if len(result) != len(tt.Match) {
|
|
t.Errorf("test %d: negotiation mismatch: have %v, want %v", i, len(result), len(tt.Match))
|
|
continue
|
|
}
|
|
// Make sure all negotiated protocols are needed and correct
|
|
for name, proto := range result {
|
|
match, ok := tt.Match[name]
|
|
if !ok {
|
|
t.Errorf("test %d, proto '%s': negotiated but shouldn't have", i, name)
|
|
continue
|
|
}
|
|
if proto.Name != match.Name {
|
|
t.Errorf("test %d, proto '%s': name mismatch: have %v, want %v", i, name, proto.Name, match.Name)
|
|
}
|
|
if proto.Version != match.Version {
|
|
t.Errorf("test %d, proto '%s': version mismatch: have %v, want %v", i, name, proto.Version, match.Version)
|
|
}
|
|
if proto.offset-baseProtocolLength != match.offset {
|
|
t.Errorf("test %d, proto '%s': offset mismatch: have %v, want %v", i, name, proto.offset-baseProtocolLength, match.offset)
|
|
}
|
|
}
|
|
// Make sure no protocols missed negotiation
|
|
for name := range tt.Match {
|
|
if _, ok := result[name]; !ok {
|
|
t.Errorf("test %d, proto '%s': not negotiated, should have", i, name)
|
|
continue
|
|
}
|
|
}
|
|
}
|
|
}
|