* les: move client pool to les/vflux/server * les/vflux/server: un-expose NodeBalance, remove unused fn, fix bugs * tests/fuzzers/vflux: add ClientPool fuzzer * les/vflux/server: fixed balance tests * les: rebase fix * les/vflux/server: fixed more bugs * les/vflux/server: unexported NodeStateMachine fields and flags * les/vflux/server: unexport all internal components and functions * les/vflux/server: fixed priorityPool test * les/vflux/server: polish balance * les/vflux/server: fixed mutex locking error * les/vflux/server: priorityPool bug fixed * common/prque: make Prque wrap-around priority handling optional * les/vflux/server: rename funcs, small optimizations * les/vflux/server: fixed timeUntil * les/vflux/server: separated balance.posValue and negValue * les/vflux/server: polish setup * les/vflux/server: enforce capacity curve monotonicity * les/vflux/server: simplified requestCapacity * les/vflux/server: requestCapacity with target range, no iterations in SetCapacity * les/vflux/server: minor changes * les/vflux/server: moved default factors to balanceTracker * les/vflux/server: set inactiveFlag in priorityPool * les/vflux/server: moved related metrics to vfs package * les/vflux/client: make priorityPool temp state logic cleaner * les/vflux/server: changed log.Crit to log.Error * add vflux fuzzer to oss-fuzz Co-authored-by: rjl493456442 <garyrong0905@gmail.com>
		
			
				
	
	
		
			234 lines
		
	
	
		
			6.4 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
			
		
		
	
	
			234 lines
		
	
	
		
			6.4 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
// Copyright 2020 The go-ethereum Authors
 | 
						|
// This file is part of the go-ethereum library.
 | 
						|
//
 | 
						|
// The go-ethereum library is free software: you can redistribute it and/or modify
 | 
						|
// it under the terms of the GNU Lesser General Public License as published by
 | 
						|
// the Free Software Foundation, either version 3 of the License, or
 | 
						|
// (at your option) any later version.
 | 
						|
//
 | 
						|
// The go-ethereum library is distributed in the hope that it will be useful,
 | 
						|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
 | 
						|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
 | 
						|
// GNU Lesser General Public License for more details.
 | 
						|
//
 | 
						|
// You should have received a copy of the GNU Lesser General Public License
 | 
						|
// along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
 | 
						|
 | 
						|
package server
 | 
						|
 | 
						|
import (
 | 
						|
	"math/rand"
 | 
						|
	"reflect"
 | 
						|
	"testing"
 | 
						|
	"time"
 | 
						|
 | 
						|
	"github.com/ethereum/go-ethereum/common/mclock"
 | 
						|
	"github.com/ethereum/go-ethereum/p2p/enode"
 | 
						|
	"github.com/ethereum/go-ethereum/p2p/enr"
 | 
						|
	"github.com/ethereum/go-ethereum/p2p/nodestate"
 | 
						|
)
 | 
						|
 | 
						|
const (
 | 
						|
	testCapacityStepDiv      = 100
 | 
						|
	testCapacityToleranceDiv = 10
 | 
						|
	testMinCap               = 100
 | 
						|
)
 | 
						|
 | 
						|
type ppTestClient struct {
 | 
						|
	node         *enode.Node
 | 
						|
	balance, cap uint64
 | 
						|
}
 | 
						|
 | 
						|
func (c *ppTestClient) priority(cap uint64) int64 {
 | 
						|
	return int64(c.balance / cap)
 | 
						|
}
 | 
						|
 | 
						|
func (c *ppTestClient) estimatePriority(cap uint64, addBalance int64, future, bias time.Duration, update bool) int64 {
 | 
						|
	return int64(c.balance / cap)
 | 
						|
}
 | 
						|
 | 
						|
func TestPriorityPool(t *testing.T) {
 | 
						|
	clock := &mclock.Simulated{}
 | 
						|
	setup := newServerSetup()
 | 
						|
	setup.balanceField = setup.setup.NewField("ppTestClient", reflect.TypeOf(&ppTestClient{}))
 | 
						|
	ns := nodestate.NewNodeStateMachine(nil, nil, clock, setup.setup)
 | 
						|
 | 
						|
	ns.SubscribeField(setup.capacityField, func(node *enode.Node, state nodestate.Flags, oldValue, newValue interface{}) {
 | 
						|
		if n := ns.GetField(node, setup.balanceField); n != nil {
 | 
						|
			c := n.(*ppTestClient)
 | 
						|
			c.cap = newValue.(uint64)
 | 
						|
		}
 | 
						|
	})
 | 
						|
	pp := newPriorityPool(ns, setup, clock, testMinCap, 0, testCapacityStepDiv, testCapacityStepDiv)
 | 
						|
	ns.Start()
 | 
						|
	pp.SetLimits(100, 1000000)
 | 
						|
	clients := make([]*ppTestClient, 100)
 | 
						|
	raise := func(c *ppTestClient) {
 | 
						|
		for {
 | 
						|
			var ok bool
 | 
						|
			ns.Operation(func() {
 | 
						|
				newCap := c.cap + c.cap/testCapacityStepDiv
 | 
						|
				ok = pp.requestCapacity(c.node, newCap, newCap, 0) == newCap
 | 
						|
			})
 | 
						|
			if !ok {
 | 
						|
				return
 | 
						|
			}
 | 
						|
		}
 | 
						|
	}
 | 
						|
	var sumBalance uint64
 | 
						|
	check := func(c *ppTestClient) {
 | 
						|
		expCap := 1000000 * c.balance / sumBalance
 | 
						|
		capTol := expCap / testCapacityToleranceDiv
 | 
						|
		if c.cap < expCap-capTol || c.cap > expCap+capTol {
 | 
						|
			t.Errorf("Wrong node capacity (expected %d, got %d)", expCap, c.cap)
 | 
						|
		}
 | 
						|
	}
 | 
						|
 | 
						|
	for i := range clients {
 | 
						|
		c := &ppTestClient{
 | 
						|
			node:    enode.SignNull(&enr.Record{}, enode.ID{byte(i)}),
 | 
						|
			balance: 100000000000,
 | 
						|
			cap:     1000,
 | 
						|
		}
 | 
						|
		sumBalance += c.balance
 | 
						|
		clients[i] = c
 | 
						|
		ns.SetField(c.node, setup.balanceField, c)
 | 
						|
		ns.SetState(c.node, setup.inactiveFlag, nodestate.Flags{}, 0)
 | 
						|
		raise(c)
 | 
						|
		check(c)
 | 
						|
	}
 | 
						|
 | 
						|
	for count := 0; count < 100; count++ {
 | 
						|
		c := clients[rand.Intn(len(clients))]
 | 
						|
		oldBalance := c.balance
 | 
						|
		c.balance = uint64(rand.Int63n(100000000000) + 100000000000)
 | 
						|
		sumBalance += c.balance - oldBalance
 | 
						|
		pp.ns.SetState(c.node, setup.updateFlag, nodestate.Flags{}, 0)
 | 
						|
		pp.ns.SetState(c.node, nodestate.Flags{}, setup.updateFlag, 0)
 | 
						|
		if c.balance > oldBalance {
 | 
						|
			raise(c)
 | 
						|
		} else {
 | 
						|
			for _, c := range clients {
 | 
						|
				raise(c)
 | 
						|
			}
 | 
						|
		}
 | 
						|
		// check whether capacities are proportional to balances
 | 
						|
		for _, c := range clients {
 | 
						|
			check(c)
 | 
						|
		}
 | 
						|
		if count%10 == 0 {
 | 
						|
			// test available capacity calculation with capacity curve
 | 
						|
			c = clients[rand.Intn(len(clients))]
 | 
						|
			curve := pp.getCapacityCurve().exclude(c.node.ID())
 | 
						|
 | 
						|
			add := uint64(rand.Int63n(10000000000000))
 | 
						|
			c.balance += add
 | 
						|
			sumBalance += add
 | 
						|
			expCap := curve.maxCapacity(func(cap uint64) int64 {
 | 
						|
				return int64(c.balance / cap)
 | 
						|
			})
 | 
						|
			var ok bool
 | 
						|
			expFail := expCap + 10
 | 
						|
			if expFail < testMinCap {
 | 
						|
				expFail = testMinCap
 | 
						|
			}
 | 
						|
			ns.Operation(func() {
 | 
						|
				ok = pp.requestCapacity(c.node, expFail, expFail, 0) == expFail
 | 
						|
			})
 | 
						|
			if ok {
 | 
						|
				t.Errorf("Request for more than expected available capacity succeeded")
 | 
						|
			}
 | 
						|
			if expCap >= testMinCap {
 | 
						|
				ns.Operation(func() {
 | 
						|
					ok = pp.requestCapacity(c.node, expCap, expCap, 0) == expCap
 | 
						|
				})
 | 
						|
				if !ok {
 | 
						|
					t.Errorf("Request for expected available capacity failed")
 | 
						|
				}
 | 
						|
			}
 | 
						|
			c.balance -= add
 | 
						|
			sumBalance -= add
 | 
						|
			pp.ns.SetState(c.node, setup.updateFlag, nodestate.Flags{}, 0)
 | 
						|
			pp.ns.SetState(c.node, nodestate.Flags{}, setup.updateFlag, 0)
 | 
						|
			for _, c := range clients {
 | 
						|
				raise(c)
 | 
						|
			}
 | 
						|
		}
 | 
						|
	}
 | 
						|
 | 
						|
	ns.Stop()
 | 
						|
}
 | 
						|
 | 
						|
func TestCapacityCurve(t *testing.T) {
 | 
						|
	clock := &mclock.Simulated{}
 | 
						|
	setup := newServerSetup()
 | 
						|
	setup.balanceField = setup.setup.NewField("ppTestClient", reflect.TypeOf(&ppTestClient{}))
 | 
						|
	ns := nodestate.NewNodeStateMachine(nil, nil, clock, setup.setup)
 | 
						|
 | 
						|
	pp := newPriorityPool(ns, setup, clock, 400000, 0, 2, 2)
 | 
						|
	ns.Start()
 | 
						|
	pp.SetLimits(10, 10000000)
 | 
						|
	clients := make([]*ppTestClient, 10)
 | 
						|
 | 
						|
	for i := range clients {
 | 
						|
		c := &ppTestClient{
 | 
						|
			node:    enode.SignNull(&enr.Record{}, enode.ID{byte(i)}),
 | 
						|
			balance: 100000000000 * uint64(i+1),
 | 
						|
			cap:     1000000,
 | 
						|
		}
 | 
						|
		clients[i] = c
 | 
						|
		ns.SetField(c.node, setup.balanceField, c)
 | 
						|
		ns.SetState(c.node, setup.inactiveFlag, nodestate.Flags{}, 0)
 | 
						|
		ns.Operation(func() {
 | 
						|
			pp.requestCapacity(c.node, c.cap, c.cap, 0)
 | 
						|
		})
 | 
						|
	}
 | 
						|
 | 
						|
	curve := pp.getCapacityCurve()
 | 
						|
	check := func(balance, expCap uint64) {
 | 
						|
		cap := curve.maxCapacity(func(cap uint64) int64 {
 | 
						|
			return int64(balance / cap)
 | 
						|
		})
 | 
						|
		var fail bool
 | 
						|
		if cap == 0 || expCap == 0 {
 | 
						|
			fail = cap != expCap
 | 
						|
		} else {
 | 
						|
			pri := balance / cap
 | 
						|
			expPri := balance / expCap
 | 
						|
			fail = pri != expPri && pri != expPri+1
 | 
						|
		}
 | 
						|
		if fail {
 | 
						|
			t.Errorf("Incorrect capacity for %d balance (got %d, expected %d)", balance, cap, expCap)
 | 
						|
		}
 | 
						|
	}
 | 
						|
 | 
						|
	check(0, 0)
 | 
						|
	check(10000000000, 100000)
 | 
						|
	check(50000000000, 500000)
 | 
						|
	check(100000000000, 1000000)
 | 
						|
	check(200000000000, 1000000)
 | 
						|
	check(300000000000, 1500000)
 | 
						|
	check(450000000000, 1500000)
 | 
						|
	check(600000000000, 2000000)
 | 
						|
	check(800000000000, 2000000)
 | 
						|
	check(1000000000000, 2500000)
 | 
						|
 | 
						|
	pp.SetLimits(11, 10000000)
 | 
						|
	curve = pp.getCapacityCurve()
 | 
						|
 | 
						|
	check(0, 0)
 | 
						|
	check(10000000000, 100000)
 | 
						|
	check(50000000000, 500000)
 | 
						|
	check(150000000000, 750000)
 | 
						|
	check(200000000000, 1000000)
 | 
						|
	check(220000000000, 1100000)
 | 
						|
	check(275000000000, 1100000)
 | 
						|
	check(375000000000, 1500000)
 | 
						|
	check(450000000000, 1500000)
 | 
						|
	check(600000000000, 2000000)
 | 
						|
	check(800000000000, 2000000)
 | 
						|
	check(1000000000000, 2500000)
 | 
						|
 | 
						|
	ns.Stop()
 | 
						|
}
 |