forked from cerc-io/plugeth
590c99a98f
If the timeout fired (even just nanoseconds) before the deadline of the next pending reply, the timer was not rescheduled. The timer would've been rescheduled anyway once the next packet was sent, but there were cases where no next packet could ever be sent due to the locking issue fixed in the previous commit. As timing-related bugs go, this issue had been present for a long time and I could never reproduce it. The test added in this commit did reproduce the issue on about one out of 15 runs.
452 lines
14 KiB
Go
452 lines
14 KiB
Go
// Copyright 2015 The go-ethereum Authors
|
|
// This file is part of the go-ethereum library.
|
|
//
|
|
// The go-ethereum library is free software: you can redistribute it and/or modify
|
|
// it under the terms of the GNU Lesser General Public License as published by
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
// (at your option) any later version.
|
|
//
|
|
// The go-ethereum library is distributed in the hope that it will be useful,
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
// GNU Lesser General Public License for more details.
|
|
//
|
|
// You should have received a copy of the GNU Lesser General Public License
|
|
// along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
package discover
|
|
|
|
import (
|
|
"bytes"
|
|
"crypto/ecdsa"
|
|
"encoding/binary"
|
|
"errors"
|
|
"fmt"
|
|
"io"
|
|
logpkg "log"
|
|
"math/rand"
|
|
"net"
|
|
"os"
|
|
"path/filepath"
|
|
"reflect"
|
|
"runtime"
|
|
"sync"
|
|
"testing"
|
|
"time"
|
|
|
|
"github.com/ethereum/go-ethereum/crypto"
|
|
"github.com/ethereum/go-ethereum/logger"
|
|
)
|
|
|
|
func init() {
|
|
logger.AddLogSystem(logger.NewStdLogSystem(os.Stdout, logpkg.LstdFlags, logger.ErrorLevel))
|
|
}
|
|
|
|
// shared test variables
|
|
var (
|
|
futureExp = uint64(time.Now().Add(10 * time.Hour).Unix())
|
|
testTarget = NodeID{0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1, 0, 1}
|
|
testRemote = rpcEndpoint{IP: net.ParseIP("1.1.1.1").To4(), UDP: 1, TCP: 2}
|
|
testLocalAnnounced = rpcEndpoint{IP: net.ParseIP("2.2.2.2").To4(), UDP: 3, TCP: 4}
|
|
testLocal = rpcEndpoint{IP: net.ParseIP("3.3.3.3").To4(), UDP: 5, TCP: 6}
|
|
)
|
|
|
|
type udpTest struct {
|
|
t *testing.T
|
|
pipe *dgramPipe
|
|
table *Table
|
|
udp *udp
|
|
sent [][]byte
|
|
localkey, remotekey *ecdsa.PrivateKey
|
|
remoteaddr *net.UDPAddr
|
|
}
|
|
|
|
func newUDPTest(t *testing.T) *udpTest {
|
|
test := &udpTest{
|
|
t: t,
|
|
pipe: newpipe(),
|
|
localkey: newkey(),
|
|
remotekey: newkey(),
|
|
remoteaddr: &net.UDPAddr{IP: net.IP{1, 2, 3, 4}, Port: 30303},
|
|
}
|
|
test.table, test.udp = newUDP(test.localkey, test.pipe, nil, "")
|
|
return test
|
|
}
|
|
|
|
// handles a packet as if it had been sent to the transport.
|
|
func (test *udpTest) packetIn(wantError error, ptype byte, data packet) error {
|
|
enc, err := encodePacket(test.remotekey, ptype, data)
|
|
if err != nil {
|
|
return test.errorf("packet (%d) encode error: %v", err)
|
|
}
|
|
test.sent = append(test.sent, enc)
|
|
if err = test.udp.handlePacket(test.remoteaddr, enc); err != wantError {
|
|
return test.errorf("error mismatch: got %q, want %q", err, wantError)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// waits for a packet to be sent by the transport.
|
|
// validate should have type func(*udpTest, X) error, where X is a packet type.
|
|
func (test *udpTest) waitPacketOut(validate interface{}) error {
|
|
dgram := test.pipe.waitPacketOut()
|
|
p, _, _, err := decodePacket(dgram)
|
|
if err != nil {
|
|
return test.errorf("sent packet decode error: %v", err)
|
|
}
|
|
fn := reflect.ValueOf(validate)
|
|
exptype := fn.Type().In(0)
|
|
if reflect.TypeOf(p) != exptype {
|
|
return test.errorf("sent packet type mismatch, got: %v, want: %v", reflect.TypeOf(p), exptype)
|
|
}
|
|
fn.Call([]reflect.Value{reflect.ValueOf(p)})
|
|
return nil
|
|
}
|
|
|
|
func (test *udpTest) errorf(format string, args ...interface{}) error {
|
|
_, file, line, ok := runtime.Caller(2) // errorf + waitPacketOut
|
|
if ok {
|
|
file = filepath.Base(file)
|
|
} else {
|
|
file = "???"
|
|
line = 1
|
|
}
|
|
err := fmt.Errorf(format, args...)
|
|
fmt.Printf("\t%s:%d: %v\n", file, line, err)
|
|
test.t.Fail()
|
|
return err
|
|
}
|
|
|
|
func TestUDP_packetErrors(t *testing.T) {
|
|
test := newUDPTest(t)
|
|
defer test.table.Close()
|
|
|
|
test.packetIn(errExpired, pingPacket, &ping{From: testRemote, To: testLocalAnnounced, Version: Version})
|
|
test.packetIn(errBadVersion, pingPacket, &ping{From: testRemote, To: testLocalAnnounced, Version: 99, Expiration: futureExp})
|
|
test.packetIn(errUnsolicitedReply, pongPacket, &pong{ReplyTok: []byte{}, Expiration: futureExp})
|
|
test.packetIn(errUnknownNode, findnodePacket, &findnode{Expiration: futureExp})
|
|
test.packetIn(errUnsolicitedReply, neighborsPacket, &neighbors{Expiration: futureExp})
|
|
}
|
|
|
|
func TestUDP_pingTimeout(t *testing.T) {
|
|
t.Parallel()
|
|
test := newUDPTest(t)
|
|
defer test.table.Close()
|
|
|
|
toaddr := &net.UDPAddr{IP: net.ParseIP("1.2.3.4"), Port: 2222}
|
|
toid := NodeID{1, 2, 3, 4}
|
|
if err := test.udp.ping(toid, toaddr); err != errTimeout {
|
|
t.Error("expected timeout error, got", err)
|
|
}
|
|
}
|
|
|
|
func TestUDP_responseTimeouts(t *testing.T) {
|
|
t.Parallel()
|
|
test := newUDPTest(t)
|
|
defer test.table.Close()
|
|
|
|
rand.Seed(time.Now().UnixNano())
|
|
randomDuration := func(max time.Duration) time.Duration {
|
|
return time.Duration(rand.Int63n(int64(max)))
|
|
}
|
|
|
|
var (
|
|
nReqs = 200
|
|
nTimeouts = 0 // number of requests with ptype > 128
|
|
nilErr = make(chan error, nReqs) // for requests that get a reply
|
|
timeoutErr = make(chan error, nReqs) // for requests that time out
|
|
)
|
|
for i := 0; i < nReqs; i++ {
|
|
// Create a matcher for a random request in udp.loop. Requests
|
|
// with ptype <= 128 will not get a reply and should time out.
|
|
// For all other requests, a reply is scheduled to arrive
|
|
// within the timeout window.
|
|
p := &pending{
|
|
ptype: byte(rand.Intn(255)),
|
|
callback: func(interface{}) bool { return true },
|
|
}
|
|
binary.BigEndian.PutUint64(p.from[:], uint64(i))
|
|
if p.ptype <= 128 {
|
|
p.errc = timeoutErr
|
|
nTimeouts++
|
|
} else {
|
|
p.errc = nilErr
|
|
time.AfterFunc(randomDuration(60*time.Millisecond), func() {
|
|
if !test.udp.handleReply(p.from, p.ptype, nil) {
|
|
t.Logf("not matched: %v", p)
|
|
}
|
|
})
|
|
}
|
|
test.udp.addpending <- p
|
|
time.Sleep(randomDuration(30 * time.Millisecond))
|
|
}
|
|
|
|
// Check that all timeouts were delivered and that the rest got nil errors.
|
|
// The replies must be delivered.
|
|
var (
|
|
recvDeadline = time.After(20 * time.Second)
|
|
nTimeoutsRecv, nNil = 0, 0
|
|
)
|
|
for i := 0; i < nReqs; i++ {
|
|
select {
|
|
case err := <-timeoutErr:
|
|
if err != errTimeout {
|
|
t.Fatalf("got non-timeout error on timeoutErr %d: %v", i, err)
|
|
}
|
|
nTimeoutsRecv++
|
|
case err := <-nilErr:
|
|
if err != nil {
|
|
t.Fatalf("got non-nil error on nilErr %d: %v", i, err)
|
|
}
|
|
nNil++
|
|
case <-recvDeadline:
|
|
t.Fatalf("exceeded recv deadline")
|
|
}
|
|
}
|
|
if nTimeoutsRecv != nTimeouts {
|
|
t.Errorf("wrong number of timeout errors received: got %d, want %d", nTimeoutsRecv, nTimeouts)
|
|
}
|
|
if nNil != nReqs-nTimeouts {
|
|
t.Errorf("wrong number of successful replies: got %d, want %d", nNil, nReqs-nTimeouts)
|
|
}
|
|
}
|
|
|
|
func TestUDP_findnodeTimeout(t *testing.T) {
|
|
t.Parallel()
|
|
test := newUDPTest(t)
|
|
defer test.table.Close()
|
|
|
|
toaddr := &net.UDPAddr{IP: net.ParseIP("1.2.3.4"), Port: 2222}
|
|
toid := NodeID{1, 2, 3, 4}
|
|
target := NodeID{4, 5, 6, 7}
|
|
result, err := test.udp.findnode(toid, toaddr, target)
|
|
if err != errTimeout {
|
|
t.Error("expected timeout error, got", err)
|
|
}
|
|
if len(result) > 0 {
|
|
t.Error("expected empty result, got", result)
|
|
}
|
|
}
|
|
|
|
func TestUDP_findnode(t *testing.T) {
|
|
test := newUDPTest(t)
|
|
defer test.table.Close()
|
|
|
|
// put a few nodes into the table. their exact
|
|
// distribution shouldn't matter much, altough we need to
|
|
// take care not to overflow any bucket.
|
|
targetHash := crypto.Sha3Hash(testTarget[:])
|
|
nodes := &nodesByDistance{target: targetHash}
|
|
for i := 0; i < bucketSize; i++ {
|
|
nodes.push(nodeAtDistance(test.table.self.sha, i+2), bucketSize)
|
|
}
|
|
test.table.stuff(nodes.entries)
|
|
|
|
// ensure there's a bond with the test node,
|
|
// findnode won't be accepted otherwise.
|
|
test.table.db.updateNode(newNode(
|
|
PubkeyID(&test.remotekey.PublicKey),
|
|
test.remoteaddr.IP,
|
|
uint16(test.remoteaddr.Port),
|
|
99,
|
|
))
|
|
// check that closest neighbors are returned.
|
|
test.packetIn(nil, findnodePacket, &findnode{Target: testTarget, Expiration: futureExp})
|
|
expected := test.table.closest(targetHash, bucketSize)
|
|
|
|
waitNeighbors := func(want []*Node) {
|
|
test.waitPacketOut(func(p *neighbors) {
|
|
if len(p.Nodes) != len(want) {
|
|
t.Errorf("wrong number of results: got %d, want %d", len(p.Nodes), bucketSize)
|
|
}
|
|
for i := range p.Nodes {
|
|
if p.Nodes[i].ID != want[i].ID {
|
|
t.Errorf("result mismatch at %d:\n got: %v\n want: %v", i, p.Nodes[i], expected.entries[i])
|
|
}
|
|
}
|
|
})
|
|
}
|
|
waitNeighbors(expected.entries[:maxNeighbors])
|
|
waitNeighbors(expected.entries[maxNeighbors:])
|
|
}
|
|
|
|
func TestUDP_findnodeMultiReply(t *testing.T) {
|
|
test := newUDPTest(t)
|
|
defer test.table.Close()
|
|
|
|
// queue a pending findnode request
|
|
resultc, errc := make(chan []*Node), make(chan error)
|
|
go func() {
|
|
rid := PubkeyID(&test.remotekey.PublicKey)
|
|
ns, err := test.udp.findnode(rid, test.remoteaddr, testTarget)
|
|
if err != nil && len(ns) == 0 {
|
|
errc <- err
|
|
} else {
|
|
resultc <- ns
|
|
}
|
|
}()
|
|
|
|
// wait for the findnode to be sent.
|
|
// after it is sent, the transport is waiting for a reply
|
|
test.waitPacketOut(func(p *findnode) {
|
|
if p.Target != testTarget {
|
|
t.Errorf("wrong target: got %v, want %v", p.Target, testTarget)
|
|
}
|
|
})
|
|
|
|
// send the reply as two packets.
|
|
list := []*Node{
|
|
MustParseNode("enode://ba85011c70bcc5c04d8607d3a0ed29aa6179c092cbdda10d5d32684fb33ed01bd94f588ca8f91ac48318087dcb02eaf36773a7a453f0eedd6742af668097b29c@10.0.1.16:30303?discport=30304"),
|
|
MustParseNode("enode://81fa361d25f157cd421c60dcc28d8dac5ef6a89476633339c5df30287474520caca09627da18543d9079b5b288698b542d56167aa5c09111e55acdbbdf2ef799@10.0.1.16:30303"),
|
|
MustParseNode("enode://9bffefd833d53fac8e652415f4973bee289e8b1a5c6c4cbe70abf817ce8a64cee11b823b66a987f51aaa9fba0d6a91b3e6bf0d5a5d1042de8e9eeea057b217f8@10.0.1.36:30301?discport=17"),
|
|
MustParseNode("enode://1b5b4aa662d7cb44a7221bfba67302590b643028197a7d5214790f3bac7aaa4a3241be9e83c09cf1f6c69d007c634faae3dc1b1221793e8446c0b3a09de65960@10.0.1.16:30303"),
|
|
}
|
|
rpclist := make([]rpcNode, len(list))
|
|
for i := range list {
|
|
rpclist[i] = nodeToRPC(list[i])
|
|
}
|
|
test.packetIn(nil, neighborsPacket, &neighbors{Expiration: futureExp, Nodes: rpclist[:2]})
|
|
test.packetIn(nil, neighborsPacket, &neighbors{Expiration: futureExp, Nodes: rpclist[2:]})
|
|
|
|
// check that the sent neighbors are all returned by findnode
|
|
select {
|
|
case result := <-resultc:
|
|
if !reflect.DeepEqual(result, list) {
|
|
t.Errorf("neighbors mismatch:\n got: %v\n want: %v", result, list)
|
|
}
|
|
case err := <-errc:
|
|
t.Errorf("findnode error: %v", err)
|
|
case <-time.After(5 * time.Second):
|
|
t.Error("findnode did not return within 5 seconds")
|
|
}
|
|
}
|
|
|
|
func TestUDP_successfulPing(t *testing.T) {
|
|
test := newUDPTest(t)
|
|
added := make(chan *Node, 1)
|
|
test.table.nodeAddedHook = func(n *Node) { added <- n }
|
|
defer test.table.Close()
|
|
|
|
// The remote side sends a ping packet to initiate the exchange.
|
|
go test.packetIn(nil, pingPacket, &ping{From: testRemote, To: testLocalAnnounced, Version: Version, Expiration: futureExp})
|
|
|
|
// the ping is replied to.
|
|
test.waitPacketOut(func(p *pong) {
|
|
pinghash := test.sent[0][:macSize]
|
|
if !bytes.Equal(p.ReplyTok, pinghash) {
|
|
t.Errorf("got pong.ReplyTok %x, want %x", p.ReplyTok, pinghash)
|
|
}
|
|
wantTo := rpcEndpoint{
|
|
// The mirrored UDP address is the UDP packet sender
|
|
IP: test.remoteaddr.IP, UDP: uint16(test.remoteaddr.Port),
|
|
// The mirrored TCP port is the one from the ping packet
|
|
TCP: testRemote.TCP,
|
|
}
|
|
if !reflect.DeepEqual(p.To, wantTo) {
|
|
t.Errorf("got pong.To %v, want %v", p.To, wantTo)
|
|
}
|
|
})
|
|
|
|
// remote is unknown, the table pings back.
|
|
test.waitPacketOut(func(p *ping) error {
|
|
if !reflect.DeepEqual(p.From, test.udp.ourEndpoint) {
|
|
t.Errorf("got ping.From %v, want %v", p.From, test.udp.ourEndpoint)
|
|
}
|
|
wantTo := rpcEndpoint{
|
|
// The mirrored UDP address is the UDP packet sender.
|
|
IP: test.remoteaddr.IP, UDP: uint16(test.remoteaddr.Port),
|
|
TCP: 0,
|
|
}
|
|
if !reflect.DeepEqual(p.To, wantTo) {
|
|
t.Errorf("got ping.To %v, want %v", p.To, wantTo)
|
|
}
|
|
return nil
|
|
})
|
|
test.packetIn(nil, pongPacket, &pong{Expiration: futureExp})
|
|
|
|
// the node should be added to the table shortly after getting the
|
|
// pong packet.
|
|
select {
|
|
case n := <-added:
|
|
rid := PubkeyID(&test.remotekey.PublicKey)
|
|
if n.ID != rid {
|
|
t.Errorf("node has wrong ID: got %v, want %v", n.ID, rid)
|
|
}
|
|
if !bytes.Equal(n.IP, test.remoteaddr.IP) {
|
|
t.Errorf("node has wrong IP: got %v, want: %v", n.IP, test.remoteaddr.IP)
|
|
}
|
|
if int(n.UDP) != test.remoteaddr.Port {
|
|
t.Errorf("node has wrong UDP port: got %v, want: %v", n.UDP, test.remoteaddr.Port)
|
|
}
|
|
if n.TCP != testRemote.TCP {
|
|
t.Errorf("node has wrong TCP port: got %v, want: %v", n.TCP, testRemote.TCP)
|
|
}
|
|
case <-time.After(2 * time.Second):
|
|
t.Errorf("node was not added within 2 seconds")
|
|
}
|
|
}
|
|
|
|
// dgramPipe is a fake UDP socket. It queues all sent datagrams.
|
|
type dgramPipe struct {
|
|
mu *sync.Mutex
|
|
cond *sync.Cond
|
|
closing chan struct{}
|
|
closed bool
|
|
queue [][]byte
|
|
}
|
|
|
|
func newpipe() *dgramPipe {
|
|
mu := new(sync.Mutex)
|
|
return &dgramPipe{
|
|
closing: make(chan struct{}),
|
|
cond: &sync.Cond{L: mu},
|
|
mu: mu,
|
|
}
|
|
}
|
|
|
|
// WriteToUDP queues a datagram.
|
|
func (c *dgramPipe) WriteToUDP(b []byte, to *net.UDPAddr) (n int, err error) {
|
|
msg := make([]byte, len(b))
|
|
copy(msg, b)
|
|
c.mu.Lock()
|
|
defer c.mu.Unlock()
|
|
if c.closed {
|
|
return 0, errors.New("closed")
|
|
}
|
|
c.queue = append(c.queue, msg)
|
|
c.cond.Signal()
|
|
return len(b), nil
|
|
}
|
|
|
|
// ReadFromUDP just hangs until the pipe is closed.
|
|
func (c *dgramPipe) ReadFromUDP(b []byte) (n int, addr *net.UDPAddr, err error) {
|
|
<-c.closing
|
|
return 0, nil, io.EOF
|
|
}
|
|
|
|
func (c *dgramPipe) Close() error {
|
|
c.mu.Lock()
|
|
defer c.mu.Unlock()
|
|
if !c.closed {
|
|
close(c.closing)
|
|
c.closed = true
|
|
}
|
|
return nil
|
|
}
|
|
|
|
func (c *dgramPipe) LocalAddr() net.Addr {
|
|
return &net.UDPAddr{IP: testLocal.IP, Port: int(testLocal.UDP)}
|
|
}
|
|
|
|
func (c *dgramPipe) waitPacketOut() []byte {
|
|
c.mu.Lock()
|
|
defer c.mu.Unlock()
|
|
for len(c.queue) == 0 {
|
|
c.cond.Wait()
|
|
}
|
|
p := c.queue[0]
|
|
copy(c.queue, c.queue[1:])
|
|
c.queue = c.queue[:len(c.queue)-1]
|
|
return p
|
|
}
|