2019-07-22 09:17:27 +00:00
|
|
|
// Copyright 2019 The go-ethereum Authors
|
2019-02-26 11:32:48 +00:00
|
|
|
// This file is part of the go-ethereum library.
|
|
|
|
//
|
|
|
|
// The go-ethereum library is free software: you can redistribute it and/or modify
|
|
|
|
// it under the terms of the GNU Lesser General Public License as published by
|
|
|
|
// the Free Software Foundation, either version 3 of the License, or
|
|
|
|
// (at your option) any later version.
|
|
|
|
//
|
|
|
|
// The go-ethereum library is distributed in the hope that it will be useful,
|
|
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
|
|
// GNU Lesser General Public License for more details.
|
|
|
|
//
|
|
|
|
// You should have received a copy of the GNU Lesser General Public License
|
|
|
|
// along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
|
|
|
|
package les
|
|
|
|
|
|
|
|
import (
|
|
|
|
"sync"
|
2019-05-30 18:51:13 +00:00
|
|
|
"sync/atomic"
|
2019-02-26 11:32:48 +00:00
|
|
|
|
|
|
|
"github.com/ethereum/go-ethereum/common/mclock"
|
|
|
|
"github.com/ethereum/go-ethereum/common/prque"
|
2023-06-19 09:33:48 +00:00
|
|
|
"golang.org/x/exp/slices"
|
2019-02-26 11:32:48 +00:00
|
|
|
)
|
|
|
|
|
|
|
|
// servingQueue allows running tasks in a limited number of threads and puts the
|
|
|
|
// waiting tasks in a priority queue
|
|
|
|
type servingQueue struct {
|
2019-05-30 18:51:13 +00:00
|
|
|
recentTime, queuedTime, servingTimeDiff uint64
|
|
|
|
burstLimit, burstDropLimit uint64
|
|
|
|
burstDecRate float64
|
|
|
|
lastUpdate mclock.AbsTime
|
|
|
|
|
2019-02-26 11:32:48 +00:00
|
|
|
queueAddCh, queueBestCh chan *servingTask
|
|
|
|
stopThreadCh, quit chan struct{}
|
|
|
|
setThreadsCh chan int
|
|
|
|
|
|
|
|
wg sync.WaitGroup
|
2023-02-09 11:03:54 +00:00
|
|
|
threadCount int // number of currently running threads
|
|
|
|
queue *prque.Prque[int64, *servingTask] // priority queue for waiting or suspended tasks
|
|
|
|
best *servingTask // the highest priority task (not included in the queue)
|
|
|
|
suspendBias int64 // priority bias against suspending an already running task
|
2019-02-26 11:32:48 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// servingTask represents a request serving task. Tasks can be implemented to
|
|
|
|
// run in multiple steps, allowing the serving queue to suspend execution between
|
|
|
|
// steps if higher priority tasks are entered. The creator of the task should
|
|
|
|
// set the following fields:
|
|
|
|
//
|
|
|
|
// - priority: greater value means higher priority; values can wrap around the int64 range
|
|
|
|
// - run: execute a single step; return true if finished
|
|
|
|
// - after: executed after run finishes or returns an error, receives the total serving time
|
|
|
|
type servingTask struct {
|
2019-05-30 18:51:13 +00:00
|
|
|
sq *servingQueue
|
|
|
|
servingTime, timeAdded, maxTime, expTime uint64
|
2020-02-26 09:41:24 +00:00
|
|
|
peer *clientPeer
|
2019-05-30 18:51:13 +00:00
|
|
|
priority int64
|
|
|
|
biasAdded bool
|
|
|
|
token runToken
|
|
|
|
tokenCh chan runToken
|
2019-02-26 11:32:48 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// runToken received by servingTask.start allows the task to run. Closing the
|
|
|
|
// channel by servingTask.stop signals the thread controller to allow a new task
|
|
|
|
// to start running.
|
|
|
|
type runToken chan struct{}
|
|
|
|
|
|
|
|
// start blocks until the task can start and returns true if it is allowed to run.
|
|
|
|
// Returning false means that the task should be cancelled.
|
|
|
|
func (t *servingTask) start() bool {
|
2019-05-30 18:51:13 +00:00
|
|
|
if t.peer.isFrozen() {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
t.tokenCh = make(chan runToken, 1)
|
2019-02-26 11:32:48 +00:00
|
|
|
select {
|
2019-05-30 18:51:13 +00:00
|
|
|
case t.sq.queueAddCh <- t:
|
|
|
|
case <-t.sq.quit:
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
select {
|
|
|
|
case t.token = <-t.tokenCh:
|
|
|
|
case <-t.sq.quit:
|
|
|
|
return false
|
2019-02-26 11:32:48 +00:00
|
|
|
}
|
|
|
|
if t.token == nil {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
t.servingTime -= uint64(mclock.Now())
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
|
|
|
|
// done signals the thread controller about the task being finished and returns
|
|
|
|
// the total serving time of the task in nanoseconds.
|
|
|
|
func (t *servingTask) done() uint64 {
|
|
|
|
t.servingTime += uint64(mclock.Now())
|
|
|
|
close(t.token)
|
2019-05-30 18:51:13 +00:00
|
|
|
diff := t.servingTime - t.timeAdded
|
|
|
|
t.timeAdded = t.servingTime
|
|
|
|
if t.expTime > diff {
|
|
|
|
t.expTime -= diff
|
|
|
|
atomic.AddUint64(&t.sq.servingTimeDiff, t.expTime)
|
|
|
|
} else {
|
|
|
|
t.expTime = 0
|
|
|
|
}
|
2019-02-26 11:32:48 +00:00
|
|
|
return t.servingTime
|
|
|
|
}
|
|
|
|
|
|
|
|
// waitOrStop can be called during the execution of the task. It blocks if there
|
|
|
|
// is a higher priority task waiting (a bias is applied in favor of the currently
|
|
|
|
// running task). Returning true means that the execution can be resumed. False
|
|
|
|
// means the task should be cancelled.
|
|
|
|
func (t *servingTask) waitOrStop() bool {
|
|
|
|
t.done()
|
|
|
|
if !t.biasAdded {
|
|
|
|
t.priority += t.sq.suspendBias
|
|
|
|
t.biasAdded = true
|
|
|
|
}
|
|
|
|
return t.start()
|
|
|
|
}
|
|
|
|
|
|
|
|
// newServingQueue returns a new servingQueue
|
2019-07-03 18:23:06 +00:00
|
|
|
func newServingQueue(suspendBias int64, utilTarget float64) *servingQueue {
|
2019-02-26 11:32:48 +00:00
|
|
|
sq := &servingQueue{
|
2023-02-09 11:03:54 +00:00
|
|
|
queue: prque.New[int64, *servingTask](nil),
|
2019-05-30 18:51:13 +00:00
|
|
|
suspendBias: suspendBias,
|
|
|
|
queueAddCh: make(chan *servingTask, 100),
|
|
|
|
queueBestCh: make(chan *servingTask),
|
|
|
|
stopThreadCh: make(chan struct{}),
|
|
|
|
quit: make(chan struct{}),
|
|
|
|
setThreadsCh: make(chan int, 10),
|
|
|
|
burstLimit: uint64(utilTarget * bufLimitRatio * 1200000),
|
|
|
|
burstDropLimit: uint64(utilTarget * bufLimitRatio * 1000000),
|
|
|
|
burstDecRate: utilTarget,
|
|
|
|
lastUpdate: mclock.Now(),
|
2019-02-26 11:32:48 +00:00
|
|
|
}
|
|
|
|
sq.wg.Add(2)
|
|
|
|
go sq.queueLoop()
|
|
|
|
go sq.threadCountLoop()
|
|
|
|
return sq
|
|
|
|
}
|
|
|
|
|
|
|
|
// newTask creates a new task with the given priority
|
2020-02-26 09:41:24 +00:00
|
|
|
func (sq *servingQueue) newTask(peer *clientPeer, maxTime uint64, priority int64) *servingTask {
|
2019-02-26 11:32:48 +00:00
|
|
|
return &servingTask{
|
|
|
|
sq: sq,
|
2019-05-30 18:51:13 +00:00
|
|
|
peer: peer,
|
|
|
|
maxTime: maxTime,
|
|
|
|
expTime: maxTime,
|
2019-02-26 11:32:48 +00:00
|
|
|
priority: priority,
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// threadController is started in multiple goroutines and controls the execution
|
|
|
|
// of tasks. The number of active thread controllers equals the allowed number of
|
|
|
|
// concurrently running threads. It tries to fetch the highest priority queued
|
|
|
|
// task first. If there are no queued tasks waiting then it can directly catch
|
|
|
|
// run tokens from the token channel and allow the corresponding tasks to run
|
|
|
|
// without entering the priority queue.
|
|
|
|
func (sq *servingQueue) threadController() {
|
2021-07-01 12:01:19 +00:00
|
|
|
defer sq.wg.Done()
|
2019-02-26 11:32:48 +00:00
|
|
|
for {
|
|
|
|
token := make(runToken)
|
|
|
|
select {
|
|
|
|
case best := <-sq.queueBestCh:
|
|
|
|
best.tokenCh <- token
|
2019-05-30 18:51:13 +00:00
|
|
|
case <-sq.stopThreadCh:
|
|
|
|
return
|
|
|
|
case <-sq.quit:
|
|
|
|
return
|
2019-02-26 11:32:48 +00:00
|
|
|
}
|
|
|
|
select {
|
|
|
|
case <-sq.stopThreadCh:
|
|
|
|
return
|
|
|
|
case <-sq.quit:
|
|
|
|
return
|
2021-07-01 12:01:19 +00:00
|
|
|
case <-token:
|
2019-02-26 11:32:48 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-06-19 09:33:48 +00:00
|
|
|
// peerTasks lists the tasks received from a given peer when selecting peers to freeze
|
|
|
|
type peerTasks struct {
|
|
|
|
peer *clientPeer
|
|
|
|
list []*servingTask
|
|
|
|
sumTime uint64
|
|
|
|
priority float64
|
2019-05-30 18:51:13 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// freezePeers selects the peers with the worst priority queued tasks and freezes
|
|
|
|
// them until burstTime goes under burstDropLimit or all peers are frozen
|
|
|
|
func (sq *servingQueue) freezePeers() {
|
2020-02-26 09:41:24 +00:00
|
|
|
peerMap := make(map[*clientPeer]*peerTasks)
|
2023-06-19 09:33:48 +00:00
|
|
|
var peerList []*peerTasks
|
2019-05-30 18:51:13 +00:00
|
|
|
if sq.best != nil {
|
|
|
|
sq.queue.Push(sq.best, sq.best.priority)
|
|
|
|
}
|
|
|
|
sq.best = nil
|
|
|
|
for sq.queue.Size() > 0 {
|
2023-02-09 11:03:54 +00:00
|
|
|
task := sq.queue.PopItem()
|
2019-05-30 18:51:13 +00:00
|
|
|
tasks := peerMap[task.peer]
|
|
|
|
if tasks == nil {
|
|
|
|
bufValue, bufLimit := task.peer.fcClient.BufferStatus()
|
|
|
|
if bufLimit < 1 {
|
|
|
|
bufLimit = 1
|
|
|
|
}
|
|
|
|
tasks = &peerTasks{
|
|
|
|
peer: task.peer,
|
|
|
|
priority: float64(bufValue) / float64(bufLimit), // lower value comes first
|
|
|
|
}
|
|
|
|
peerMap[task.peer] = tasks
|
|
|
|
peerList = append(peerList, tasks)
|
|
|
|
}
|
|
|
|
tasks.list = append(tasks.list, task)
|
|
|
|
tasks.sumTime += task.expTime
|
|
|
|
}
|
2023-06-19 09:33:48 +00:00
|
|
|
slices.SortFunc(peerList, func(a, b *peerTasks) bool {
|
|
|
|
return a.priority < b.priority
|
|
|
|
})
|
2019-05-30 18:51:13 +00:00
|
|
|
drop := true
|
|
|
|
for _, tasks := range peerList {
|
|
|
|
if drop {
|
2020-02-26 09:41:24 +00:00
|
|
|
tasks.peer.freeze()
|
2019-05-30 18:51:13 +00:00
|
|
|
tasks.peer.fcClient.Freeze()
|
|
|
|
sq.queuedTime -= tasks.sumTime
|
2019-07-03 18:23:06 +00:00
|
|
|
sqQueuedGauge.Update(int64(sq.queuedTime))
|
|
|
|
clientFreezeMeter.Mark(1)
|
2019-05-30 18:51:13 +00:00
|
|
|
drop = sq.recentTime+sq.queuedTime > sq.burstDropLimit
|
|
|
|
for _, task := range tasks.list {
|
|
|
|
task.tokenCh <- nil
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
for _, task := range tasks.list {
|
|
|
|
sq.queue.Push(task, task.priority)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if sq.queue.Size() > 0 {
|
2023-02-09 11:03:54 +00:00
|
|
|
sq.best = sq.queue.PopItem()
|
2019-05-30 18:51:13 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// updateRecentTime recalculates the recent serving time value
|
|
|
|
func (sq *servingQueue) updateRecentTime() {
|
|
|
|
subTime := atomic.SwapUint64(&sq.servingTimeDiff, 0)
|
|
|
|
now := mclock.Now()
|
|
|
|
dt := now - sq.lastUpdate
|
|
|
|
sq.lastUpdate = now
|
|
|
|
if dt > 0 {
|
|
|
|
subTime += uint64(float64(dt) * sq.burstDecRate)
|
|
|
|
}
|
|
|
|
if sq.recentTime > subTime {
|
|
|
|
sq.recentTime -= subTime
|
|
|
|
} else {
|
|
|
|
sq.recentTime = 0
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2019-02-26 11:32:48 +00:00
|
|
|
// addTask inserts a task into the priority queue
|
|
|
|
func (sq *servingQueue) addTask(task *servingTask) {
|
|
|
|
if sq.best == nil {
|
|
|
|
sq.best = task
|
2021-04-06 18:42:50 +00:00
|
|
|
} else if task.priority-sq.best.priority > 0 {
|
2019-02-26 11:32:48 +00:00
|
|
|
sq.queue.Push(sq.best, sq.best.priority)
|
|
|
|
sq.best = task
|
|
|
|
} else {
|
|
|
|
sq.queue.Push(task, task.priority)
|
|
|
|
}
|
2019-05-30 18:51:13 +00:00
|
|
|
sq.updateRecentTime()
|
|
|
|
sq.queuedTime += task.expTime
|
2019-07-03 18:23:06 +00:00
|
|
|
sqServedGauge.Update(int64(sq.recentTime))
|
|
|
|
sqQueuedGauge.Update(int64(sq.queuedTime))
|
2019-05-30 18:51:13 +00:00
|
|
|
if sq.recentTime+sq.queuedTime > sq.burstLimit {
|
|
|
|
sq.freezePeers()
|
|
|
|
}
|
2019-02-26 11:32:48 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
// queueLoop is an event loop running in a goroutine. It receives tasks from queueAddCh
|
|
|
|
// and always tries to send the highest priority task to queueBestCh. Successfully sent
|
|
|
|
// tasks are removed from the queue.
|
|
|
|
func (sq *servingQueue) queueLoop() {
|
2021-07-01 12:01:19 +00:00
|
|
|
defer sq.wg.Done()
|
2019-02-26 11:32:48 +00:00
|
|
|
for {
|
|
|
|
if sq.best != nil {
|
2019-05-30 18:51:13 +00:00
|
|
|
expTime := sq.best.expTime
|
2019-02-26 11:32:48 +00:00
|
|
|
select {
|
|
|
|
case task := <-sq.queueAddCh:
|
|
|
|
sq.addTask(task)
|
|
|
|
case sq.queueBestCh <- sq.best:
|
2019-05-30 18:51:13 +00:00
|
|
|
sq.updateRecentTime()
|
|
|
|
sq.queuedTime -= expTime
|
|
|
|
sq.recentTime += expTime
|
2019-07-03 18:23:06 +00:00
|
|
|
sqServedGauge.Update(int64(sq.recentTime))
|
|
|
|
sqQueuedGauge.Update(int64(sq.queuedTime))
|
2019-02-26 11:32:48 +00:00
|
|
|
if sq.queue.Size() == 0 {
|
|
|
|
sq.best = nil
|
|
|
|
} else {
|
2023-02-09 11:03:54 +00:00
|
|
|
sq.best = sq.queue.PopItem()
|
2019-02-26 11:32:48 +00:00
|
|
|
}
|
|
|
|
case <-sq.quit:
|
|
|
|
return
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
select {
|
|
|
|
case task := <-sq.queueAddCh:
|
|
|
|
sq.addTask(task)
|
|
|
|
case <-sq.quit:
|
|
|
|
return
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// threadCountLoop is an event loop running in a goroutine. It adjusts the number
|
|
|
|
// of active thread controller goroutines.
|
|
|
|
func (sq *servingQueue) threadCountLoop() {
|
|
|
|
var threadCountTarget int
|
2021-07-01 12:01:19 +00:00
|
|
|
defer sq.wg.Done()
|
2019-02-26 11:32:48 +00:00
|
|
|
for {
|
|
|
|
for threadCountTarget > sq.threadCount {
|
|
|
|
sq.wg.Add(1)
|
|
|
|
go sq.threadController()
|
|
|
|
sq.threadCount++
|
|
|
|
}
|
|
|
|
if threadCountTarget < sq.threadCount {
|
|
|
|
select {
|
|
|
|
case threadCountTarget = <-sq.setThreadsCh:
|
|
|
|
case sq.stopThreadCh <- struct{}{}:
|
|
|
|
sq.threadCount--
|
|
|
|
case <-sq.quit:
|
|
|
|
return
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
select {
|
|
|
|
case threadCountTarget = <-sq.setThreadsCh:
|
|
|
|
case <-sq.quit:
|
|
|
|
return
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// setThreads sets the allowed processing thread count, suspending tasks as soon as
|
|
|
|
// possible if necessary.
|
|
|
|
func (sq *servingQueue) setThreads(threadCount int) {
|
|
|
|
select {
|
|
|
|
case sq.setThreadsCh <- threadCount:
|
|
|
|
case <-sq.quit:
|
|
|
|
return
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// stop stops task processing as soon as possible and shuts down the serving queue.
|
|
|
|
func (sq *servingQueue) stop() {
|
|
|
|
close(sq.quit)
|
|
|
|
sq.wg.Wait()
|
|
|
|
}
|