forked from cerc-io/plugeth
ethdb/pebble: add level file metrics (#28271)
This commit is contained in:
parent
8976a0c97a
commit
7776a3214a
@ -71,6 +71,8 @@ type Database struct {
|
||||
seekCompGauge metrics.Gauge // Gauge for tracking the number of table compaction caused by read opt
|
||||
manualMemAllocGauge metrics.Gauge // Gauge for tracking amount of non-managed memory currently allocated
|
||||
|
||||
levelsGauge []metrics.Gauge // Gauge for tracking the number of tables in levels
|
||||
|
||||
quitLock sync.RWMutex // Mutex protecting the quit channel and the closed flag
|
||||
quitChan chan chan error // Quit channel to stop the metrics collection before closing the database
|
||||
closed bool // keep track of whether we're Closed
|
||||
@ -230,7 +232,7 @@ func New(file string, cache int, handles int, namespace string, readonly bool, e
|
||||
db.manualMemAllocGauge = metrics.NewRegisteredGauge(namespace+"memory/manualalloc", nil)
|
||||
|
||||
// Start up the metrics gathering and return
|
||||
go db.meter(metricsGatheringInterval)
|
||||
go db.meter(metricsGatheringInterval, namespace)
|
||||
return db, nil
|
||||
}
|
||||
|
||||
@ -427,7 +429,7 @@ func (d *Database) Path() string {
|
||||
|
||||
// meter periodically retrieves internal pebble counters and reports them to
|
||||
// the metrics subsystem.
|
||||
func (d *Database) meter(refresh time.Duration) {
|
||||
func (d *Database) meter(refresh time.Duration, namespace string) {
|
||||
var errc chan error
|
||||
timer := time.NewTimer(refresh)
|
||||
defer timer.Stop()
|
||||
@ -450,7 +452,7 @@ func (d *Database) meter(refresh time.Duration) {
|
||||
compRead int64
|
||||
nWrite int64
|
||||
|
||||
metrics = d.db.Metrics()
|
||||
stats = d.db.Metrics()
|
||||
compTime = d.compTime.Load()
|
||||
writeDelayCount = d.writeDelayCount.Load()
|
||||
writeDelayTime = d.writeDelayTime.Load()
|
||||
@ -461,14 +463,14 @@ func (d *Database) meter(refresh time.Duration) {
|
||||
writeDelayCounts[i%2] = writeDelayCount
|
||||
compTimes[i%2] = compTime
|
||||
|
||||
for _, levelMetrics := range metrics.Levels {
|
||||
for _, levelMetrics := range stats.Levels {
|
||||
nWrite += int64(levelMetrics.BytesCompacted)
|
||||
nWrite += int64(levelMetrics.BytesFlushed)
|
||||
compWrite += int64(levelMetrics.BytesCompacted)
|
||||
compRead += int64(levelMetrics.BytesRead)
|
||||
}
|
||||
|
||||
nWrite += int64(metrics.WAL.BytesWritten)
|
||||
nWrite += int64(stats.WAL.BytesWritten)
|
||||
|
||||
compWrites[i%2] = compWrite
|
||||
compReads[i%2] = compRead
|
||||
@ -490,7 +492,7 @@ func (d *Database) meter(refresh time.Duration) {
|
||||
d.compWriteMeter.Mark(compWrites[i%2] - compWrites[(i-1)%2])
|
||||
}
|
||||
if d.diskSizeGauge != nil {
|
||||
d.diskSizeGauge.Update(int64(metrics.DiskSpaceUsage()))
|
||||
d.diskSizeGauge.Update(int64(stats.DiskSpaceUsage()))
|
||||
}
|
||||
if d.diskReadMeter != nil {
|
||||
d.diskReadMeter.Mark(0) // pebble doesn't track non-compaction reads
|
||||
@ -499,12 +501,20 @@ func (d *Database) meter(refresh time.Duration) {
|
||||
d.diskWriteMeter.Mark(nWrites[i%2] - nWrites[(i-1)%2])
|
||||
}
|
||||
// See https://github.com/cockroachdb/pebble/pull/1628#pullrequestreview-1026664054
|
||||
manuallyAllocated := metrics.BlockCache.Size + int64(metrics.MemTable.Size) + int64(metrics.MemTable.ZombieSize)
|
||||
manuallyAllocated := stats.BlockCache.Size + int64(stats.MemTable.Size) + int64(stats.MemTable.ZombieSize)
|
||||
d.manualMemAllocGauge.Update(manuallyAllocated)
|
||||
d.memCompGauge.Update(metrics.Flush.Count)
|
||||
d.memCompGauge.Update(stats.Flush.Count)
|
||||
d.nonlevel0CompGauge.Update(nonLevel0CompCount)
|
||||
d.level0CompGauge.Update(level0CompCount)
|
||||
d.seekCompGauge.Update(metrics.Compact.ReadCount)
|
||||
d.seekCompGauge.Update(stats.Compact.ReadCount)
|
||||
|
||||
for i, level := range stats.Levels {
|
||||
// Append metrics for additional layers
|
||||
if i >= len(d.levelsGauge) {
|
||||
d.levelsGauge = append(d.levelsGauge, metrics.NewRegisteredGauge(namespace+fmt.Sprintf("tables/level%v", i), nil))
|
||||
}
|
||||
d.levelsGauge[i].Update(level.NumFiles)
|
||||
}
|
||||
|
||||
// Sleep a bit, then repeat the stats collection
|
||||
select {
|
||||
|
Loading…
Reference in New Issue
Block a user