6fd93ed170
* Fix 1.21 regression: GET_32G_MAX_CONCURRENT + mixed prepared/executing leads to stuck scheduler If you have 12 GET tasks and GET_32G_MAX_CONCURRENT=1, sealing jobs will only show assigned tasks for GET of the miner and is stuck. I believe this to be a regression of 1.21 unifying the counters, in the case of GETs where PrepType and TaskType both being seal/v0/fetch leading to a state where tasks are blocked since already counted towards the limit. * itests: Repro issue from PR #10633 * make counters int (non-working) * fix: worker sched: Send taskDone notifs after tasks are done * itests: Make TestPledgeMaxConcurrentGet actually reproduce the issue * make the linter happy --------- Co-authored-by: Steffen Butzer <steffen.butzer@outlook.com>
100 lines
2.4 KiB
Go
100 lines
2.4 KiB
Go
package sealer
|
|
|
|
import (
|
|
"math"
|
|
|
|
"github.com/filecoin-project/lotus/storage/sealer/storiface"
|
|
)
|
|
|
|
func NewLowestUtilizationAssigner() Assigner {
|
|
return &AssignerCommon{
|
|
WindowSel: LowestUtilizationWS,
|
|
}
|
|
}
|
|
|
|
func LowestUtilizationWS(sh *Scheduler, queueLen int, acceptableWindows [][]int, windows []SchedWindow) int {
|
|
scheduled := 0
|
|
rmQueue := make([]int, 0, queueLen)
|
|
workerUtil := map[storiface.WorkerID]float64{}
|
|
|
|
for sqi := 0; sqi < queueLen; sqi++ {
|
|
task := (*sh.SchedQueue)[sqi]
|
|
|
|
selectedWindow := -1
|
|
var needRes storiface.Resources
|
|
var info storiface.WorkerInfo
|
|
var bestWid storiface.WorkerID
|
|
bestUtilization := math.MaxFloat64 // smaller = better
|
|
|
|
for i, wnd := range acceptableWindows[task.IndexHeap] {
|
|
wid := sh.OpenWindows[wnd].Worker
|
|
w := sh.Workers[wid]
|
|
|
|
res := w.Info.Resources.ResourceSpec(task.Sector.ProofType, task.TaskType)
|
|
|
|
log.Debugf("SCHED try assign sqi:%d sector %d to window %d (awi:%d)", sqi, task.Sector.ID.Number, wnd, i)
|
|
|
|
// TODO: allow bigger windows
|
|
if !windows[wnd].Allocated.CanHandleRequest(task.SchedId, task.SealTask(), res, wid, "schedAssign", w.Info) {
|
|
continue
|
|
}
|
|
|
|
wu, found := workerUtil[wid]
|
|
if !found {
|
|
wu = w.Utilization()
|
|
workerUtil[wid] = wu
|
|
}
|
|
if wu >= bestUtilization {
|
|
// acceptable worker list is initially sorted by utilization, and the initially-best workers
|
|
// will be assigned tasks first. This means that if we find a worker which isn't better, it
|
|
// probably means that the other workers aren't better either.
|
|
//
|
|
// utilization
|
|
// ^
|
|
// | /
|
|
// | \ /
|
|
// | \ /
|
|
// | *
|
|
// #--------> acceptableWindow index
|
|
//
|
|
// * -> we're here
|
|
break
|
|
}
|
|
|
|
info = w.Info
|
|
needRes = res
|
|
bestWid = wid
|
|
selectedWindow = wnd
|
|
bestUtilization = wu
|
|
}
|
|
|
|
if selectedWindow < 0 {
|
|
// all windows full
|
|
continue
|
|
}
|
|
|
|
log.Debugw("SCHED ASSIGNED",
|
|
"assigner", "util",
|
|
"sqi", sqi,
|
|
"sector", task.Sector.ID.Number,
|
|
"task", task.TaskType,
|
|
"window", selectedWindow,
|
|
"worker", bestWid,
|
|
"utilization", bestUtilization)
|
|
|
|
workerUtil[bestWid] += windows[selectedWindow].Allocated.Add(task.SchedId, task.SealTask(), info.Resources, needRes)
|
|
windows[selectedWindow].Todo = append(windows[selectedWindow].Todo, task)
|
|
|
|
rmQueue = append(rmQueue, sqi)
|
|
scheduled++
|
|
}
|
|
|
|
if len(rmQueue) > 0 {
|
|
for i := len(rmQueue) - 1; i >= 0; i-- {
|
|
sh.SchedQueue.Remove(rmQueue[i])
|
|
}
|
|
}
|
|
|
|
return scheduled
|
|
}
|