forked from cerc-io/plugeth
b4a2681120
This PR reimplements the light client server pool. It is also a first step to move certain logic into a new lespay package. This package will contain the implementation of the lespay token sale functions, the token buying and selling logic and other components related to peer selection/prioritization and service quality evaluation. Over the long term this package will be reusable for incentivizing future protocols. Since the LES peer logic is now based on enode.Iterator, it can now use DNS-based fallback discovery to find servers. This document describes the function of the new components: https://gist.github.com/zsfelfoldi/3c7ace895234b7b345ab4f71dab102d4
353 lines
9.4 KiB
Go
353 lines
9.4 KiB
Go
// Copyright 2020 The go-ethereum Authors
|
|
// This file is part of the go-ethereum library.
|
|
//
|
|
// The go-ethereum library is free software: you can redistribute it and/or modify
|
|
// it under the terms of the GNU Lesser General Public License as published by
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
// (at your option) any later version.
|
|
//
|
|
// The go-ethereum library is distributed in the hope that it will be useful,
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
// GNU Lesser General Public License for more details.
|
|
//
|
|
// You should have received a copy of the GNU Lesser General Public License
|
|
// along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
package les
|
|
|
|
import (
|
|
"math/rand"
|
|
"sync/atomic"
|
|
"testing"
|
|
"time"
|
|
|
|
"github.com/ethereum/go-ethereum/common/mclock"
|
|
"github.com/ethereum/go-ethereum/ethdb"
|
|
"github.com/ethereum/go-ethereum/ethdb/memorydb"
|
|
lpc "github.com/ethereum/go-ethereum/les/lespay/client"
|
|
"github.com/ethereum/go-ethereum/p2p"
|
|
"github.com/ethereum/go-ethereum/p2p/enode"
|
|
"github.com/ethereum/go-ethereum/p2p/enr"
|
|
)
|
|
|
|
const (
|
|
spTestNodes = 1000
|
|
spTestTarget = 5
|
|
spTestLength = 10000
|
|
spMinTotal = 40000
|
|
spMaxTotal = 50000
|
|
)
|
|
|
|
func testNodeID(i int) enode.ID {
|
|
return enode.ID{42, byte(i % 256), byte(i / 256)}
|
|
}
|
|
|
|
func testNodeIndex(id enode.ID) int {
|
|
if id[0] != 42 {
|
|
return -1
|
|
}
|
|
return int(id[1]) + int(id[2])*256
|
|
}
|
|
|
|
type serverPoolTest struct {
|
|
db ethdb.KeyValueStore
|
|
clock *mclock.Simulated
|
|
quit chan struct{}
|
|
preNeg, preNegFail bool
|
|
vt *lpc.ValueTracker
|
|
sp *serverPool
|
|
input enode.Iterator
|
|
testNodes []spTestNode
|
|
trusted []string
|
|
waitCount, waitEnded int32
|
|
|
|
cycle, conn, servedConn int
|
|
serviceCycles, dialCount int
|
|
disconnect map[int][]int
|
|
}
|
|
|
|
type spTestNode struct {
|
|
connectCycles, waitCycles int
|
|
nextConnCycle, totalConn int
|
|
connected, service bool
|
|
peer *serverPeer
|
|
}
|
|
|
|
func newServerPoolTest(preNeg, preNegFail bool) *serverPoolTest {
|
|
nodes := make([]*enode.Node, spTestNodes)
|
|
for i := range nodes {
|
|
nodes[i] = enode.SignNull(&enr.Record{}, testNodeID(i))
|
|
}
|
|
return &serverPoolTest{
|
|
clock: &mclock.Simulated{},
|
|
db: memorydb.New(),
|
|
input: enode.CycleNodes(nodes),
|
|
testNodes: make([]spTestNode, spTestNodes),
|
|
preNeg: preNeg,
|
|
preNegFail: preNegFail,
|
|
}
|
|
}
|
|
|
|
func (s *serverPoolTest) beginWait() {
|
|
// ensure that dialIterator and the maximal number of pre-neg queries are not all stuck in a waiting state
|
|
for atomic.AddInt32(&s.waitCount, 1) > preNegLimit {
|
|
atomic.AddInt32(&s.waitCount, -1)
|
|
s.clock.Run(time.Second)
|
|
}
|
|
}
|
|
|
|
func (s *serverPoolTest) endWait() {
|
|
atomic.AddInt32(&s.waitCount, -1)
|
|
atomic.AddInt32(&s.waitEnded, 1)
|
|
}
|
|
|
|
func (s *serverPoolTest) addTrusted(i int) {
|
|
s.trusted = append(s.trusted, enode.SignNull(&enr.Record{}, testNodeID(i)).String())
|
|
}
|
|
|
|
func (s *serverPoolTest) start() {
|
|
var testQuery queryFunc
|
|
if s.preNeg {
|
|
testQuery = func(node *enode.Node) int {
|
|
idx := testNodeIndex(node.ID())
|
|
n := &s.testNodes[idx]
|
|
canConnect := !n.connected && n.connectCycles != 0 && s.cycle >= n.nextConnCycle
|
|
if s.preNegFail {
|
|
// simulate a scenario where UDP queries never work
|
|
s.beginWait()
|
|
s.clock.Sleep(time.Second * 5)
|
|
s.endWait()
|
|
return -1
|
|
} else {
|
|
switch idx % 3 {
|
|
case 0:
|
|
// pre-neg returns true only if connection is possible
|
|
if canConnect {
|
|
return 1
|
|
} else {
|
|
return 0
|
|
}
|
|
case 1:
|
|
// pre-neg returns true but connection might still fail
|
|
return 1
|
|
case 2:
|
|
// pre-neg returns true if connection is possible, otherwise timeout (node unresponsive)
|
|
if canConnect {
|
|
return 1
|
|
} else {
|
|
s.beginWait()
|
|
s.clock.Sleep(time.Second * 5)
|
|
s.endWait()
|
|
return -1
|
|
}
|
|
}
|
|
return -1
|
|
}
|
|
}
|
|
}
|
|
|
|
s.vt = lpc.NewValueTracker(s.db, s.clock, requestList, time.Minute, 1/float64(time.Hour), 1/float64(time.Hour*100), 1/float64(time.Hour*1000))
|
|
s.sp = newServerPool(s.db, []byte("serverpool:"), s.vt, s.input, 0, testQuery, s.clock, s.trusted)
|
|
s.sp.validSchemes = enode.ValidSchemesForTesting
|
|
s.sp.unixTime = func() int64 { return int64(s.clock.Now()) / int64(time.Second) }
|
|
s.disconnect = make(map[int][]int)
|
|
s.sp.start()
|
|
s.quit = make(chan struct{})
|
|
go func() {
|
|
last := int32(-1)
|
|
for {
|
|
select {
|
|
case <-time.After(time.Millisecond * 100):
|
|
c := atomic.LoadInt32(&s.waitEnded)
|
|
if c == last {
|
|
// advance clock if test is stuck (might happen in rare cases)
|
|
s.clock.Run(time.Second)
|
|
}
|
|
last = c
|
|
case <-s.quit:
|
|
return
|
|
}
|
|
}
|
|
}()
|
|
}
|
|
|
|
func (s *serverPoolTest) stop() {
|
|
close(s.quit)
|
|
s.sp.stop()
|
|
s.vt.Stop()
|
|
for i := range s.testNodes {
|
|
n := &s.testNodes[i]
|
|
if n.connected {
|
|
n.totalConn += s.cycle
|
|
}
|
|
n.connected = false
|
|
n.peer = nil
|
|
n.nextConnCycle = 0
|
|
}
|
|
s.conn, s.servedConn = 0, 0
|
|
}
|
|
|
|
func (s *serverPoolTest) run() {
|
|
for count := spTestLength; count > 0; count-- {
|
|
if dcList := s.disconnect[s.cycle]; dcList != nil {
|
|
for _, idx := range dcList {
|
|
n := &s.testNodes[idx]
|
|
s.sp.unregisterPeer(n.peer)
|
|
n.totalConn += s.cycle
|
|
n.connected = false
|
|
n.peer = nil
|
|
s.conn--
|
|
if n.service {
|
|
s.servedConn--
|
|
}
|
|
n.nextConnCycle = s.cycle + n.waitCycles
|
|
}
|
|
delete(s.disconnect, s.cycle)
|
|
}
|
|
if s.conn < spTestTarget {
|
|
s.dialCount++
|
|
s.beginWait()
|
|
s.sp.dialIterator.Next()
|
|
s.endWait()
|
|
dial := s.sp.dialIterator.Node()
|
|
id := dial.ID()
|
|
idx := testNodeIndex(id)
|
|
n := &s.testNodes[idx]
|
|
if !n.connected && n.connectCycles != 0 && s.cycle >= n.nextConnCycle {
|
|
s.conn++
|
|
if n.service {
|
|
s.servedConn++
|
|
}
|
|
n.totalConn -= s.cycle
|
|
n.connected = true
|
|
dc := s.cycle + n.connectCycles
|
|
s.disconnect[dc] = append(s.disconnect[dc], idx)
|
|
n.peer = &serverPeer{peerCommons: peerCommons{Peer: p2p.NewPeer(id, "", nil)}}
|
|
s.sp.registerPeer(n.peer)
|
|
if n.service {
|
|
s.vt.Served(s.vt.GetNode(id), []lpc.ServedRequest{{ReqType: 0, Amount: 100}}, 0)
|
|
}
|
|
}
|
|
}
|
|
s.serviceCycles += s.servedConn
|
|
s.clock.Run(time.Second)
|
|
s.cycle++
|
|
}
|
|
}
|
|
|
|
func (s *serverPoolTest) setNodes(count, conn, wait int, service, trusted bool) (res []int) {
|
|
for ; count > 0; count-- {
|
|
idx := rand.Intn(spTestNodes)
|
|
for s.testNodes[idx].connectCycles != 0 || s.testNodes[idx].connected {
|
|
idx = rand.Intn(spTestNodes)
|
|
}
|
|
res = append(res, idx)
|
|
s.testNodes[idx] = spTestNode{
|
|
connectCycles: conn,
|
|
waitCycles: wait,
|
|
service: service,
|
|
}
|
|
if trusted {
|
|
s.addTrusted(idx)
|
|
}
|
|
}
|
|
return
|
|
}
|
|
|
|
func (s *serverPoolTest) resetNodes() {
|
|
for i, n := range s.testNodes {
|
|
if n.connected {
|
|
n.totalConn += s.cycle
|
|
s.sp.unregisterPeer(n.peer)
|
|
}
|
|
s.testNodes[i] = spTestNode{totalConn: n.totalConn}
|
|
}
|
|
s.conn, s.servedConn = 0, 0
|
|
s.disconnect = make(map[int][]int)
|
|
s.trusted = nil
|
|
}
|
|
|
|
func (s *serverPoolTest) checkNodes(t *testing.T, nodes []int) {
|
|
var sum int
|
|
for _, idx := range nodes {
|
|
n := &s.testNodes[idx]
|
|
if n.connected {
|
|
n.totalConn += s.cycle
|
|
}
|
|
sum += n.totalConn
|
|
n.totalConn = 0
|
|
if n.connected {
|
|
n.totalConn -= s.cycle
|
|
}
|
|
}
|
|
if sum < spMinTotal || sum > spMaxTotal {
|
|
t.Errorf("Total connection amount %d outside expected range %d to %d", sum, spMinTotal, spMaxTotal)
|
|
}
|
|
}
|
|
|
|
func TestServerPool(t *testing.T) { testServerPool(t, false, false) }
|
|
func TestServerPoolWithPreNeg(t *testing.T) { testServerPool(t, true, false) }
|
|
func TestServerPoolWithPreNegFail(t *testing.T) { testServerPool(t, true, true) }
|
|
func testServerPool(t *testing.T, preNeg, fail bool) {
|
|
s := newServerPoolTest(preNeg, fail)
|
|
nodes := s.setNodes(100, 200, 200, true, false)
|
|
s.setNodes(100, 20, 20, false, false)
|
|
s.start()
|
|
s.run()
|
|
s.stop()
|
|
s.checkNodes(t, nodes)
|
|
}
|
|
|
|
func TestServerPoolChangedNodes(t *testing.T) { testServerPoolChangedNodes(t, false) }
|
|
func TestServerPoolChangedNodesWithPreNeg(t *testing.T) { testServerPoolChangedNodes(t, true) }
|
|
func testServerPoolChangedNodes(t *testing.T, preNeg bool) {
|
|
s := newServerPoolTest(preNeg, false)
|
|
nodes := s.setNodes(100, 200, 200, true, false)
|
|
s.setNodes(100, 20, 20, false, false)
|
|
s.start()
|
|
s.run()
|
|
s.checkNodes(t, nodes)
|
|
for i := 0; i < 3; i++ {
|
|
s.resetNodes()
|
|
nodes := s.setNodes(100, 200, 200, true, false)
|
|
s.setNodes(100, 20, 20, false, false)
|
|
s.run()
|
|
s.checkNodes(t, nodes)
|
|
}
|
|
s.stop()
|
|
}
|
|
|
|
func TestServerPoolRestartNoDiscovery(t *testing.T) { testServerPoolRestartNoDiscovery(t, false) }
|
|
func TestServerPoolRestartNoDiscoveryWithPreNeg(t *testing.T) {
|
|
testServerPoolRestartNoDiscovery(t, true)
|
|
}
|
|
func testServerPoolRestartNoDiscovery(t *testing.T, preNeg bool) {
|
|
s := newServerPoolTest(preNeg, false)
|
|
nodes := s.setNodes(100, 200, 200, true, false)
|
|
s.setNodes(100, 20, 20, false, false)
|
|
s.start()
|
|
s.run()
|
|
s.stop()
|
|
s.checkNodes(t, nodes)
|
|
s.input = nil
|
|
s.start()
|
|
s.run()
|
|
s.stop()
|
|
s.checkNodes(t, nodes)
|
|
}
|
|
|
|
func TestServerPoolTrustedNoDiscovery(t *testing.T) { testServerPoolTrustedNoDiscovery(t, false) }
|
|
func TestServerPoolTrustedNoDiscoveryWithPreNeg(t *testing.T) {
|
|
testServerPoolTrustedNoDiscovery(t, true)
|
|
}
|
|
func testServerPoolTrustedNoDiscovery(t *testing.T, preNeg bool) {
|
|
s := newServerPoolTest(preNeg, false)
|
|
trusted := s.setNodes(200, 200, 200, true, true)
|
|
s.input = nil
|
|
s.start()
|
|
s.run()
|
|
s.stop()
|
|
s.checkNodes(t, trusted)
|
|
}
|