feat: Add small cache to execution traces

This PR adds a small cache to calls to ExecutionTrace which helps
 improve performance for node operators like exchanges and block
explorers.

If items is in cache calls to this function will be 2-3x faster.

Fixes: https://github.com/filecoin-project/lotus/issues/10504
This commit is contained in:
Fridrik Asmundsson 2023-03-20 15:17:27 +00:00
parent 5aa6dd100c
commit 1ea7e05cde
2 changed files with 27 additions and 2 deletions

View File

@ -128,10 +128,17 @@ func (sm *StateManager) ExecutionTraceWithMonitor(ctx context.Context, ts *types
} }
func (sm *StateManager) ExecutionTrace(ctx context.Context, ts *types.TipSet) (cid.Cid, []*api.InvocResult, error) { func (sm *StateManager) ExecutionTrace(ctx context.Context, ts *types.TipSet) (cid.Cid, []*api.InvocResult, error) {
if entry, ok := sm.execTraceCache.Get(ts); ok {
return entry.cid, entry.invocTrace, nil
}
var invocTrace []*api.InvocResult var invocTrace []*api.InvocResult
st, err := sm.ExecutionTraceWithMonitor(ctx, ts, &InvocationTracer{trace: &invocTrace}) st, err := sm.ExecutionTraceWithMonitor(ctx, ts, &InvocationTracer{trace: &invocTrace})
if err != nil { if err != nil {
return cid.Undef, nil, err return cid.Undef, nil, err
} }
sm.execTraceCache.Add(ts, tipSetCacheEntry{st, invocTrace})
return st, invocTrace, nil return st, invocTrace, nil
} }

View File

@ -5,6 +5,7 @@ import (
"fmt" "fmt"
"sync" "sync"
lru "github.com/hashicorp/golang-lru/v2"
"github.com/ipfs/go-cid" "github.com/ipfs/go-cid"
dstore "github.com/ipfs/go-datastore" dstore "github.com/ipfs/go-datastore"
cbor "github.com/ipfs/go-ipld-cbor" cbor "github.com/ipfs/go-ipld-cbor"
@ -39,6 +40,8 @@ import (
const LookbackNoLimit = api.LookbackNoLimit const LookbackNoLimit = api.LookbackNoLimit
const ReceiptAmtBitwidth = 3 const ReceiptAmtBitwidth = 3
const execTraceCacheSize = 16
var log = logging.Logger("statemgr") var log = logging.Logger("statemgr")
type StateManagerAPI interface { type StateManagerAPI interface {
@ -138,6 +141,10 @@ type StateManager struct {
beacon beacon.Schedule beacon beacon.Schedule
msgIndex index.MsgIndex msgIndex index.MsgIndex
// We keep a small cache for calls to ExecutionTrace which helps improve
// performance for node operators like exchanges and block explorers
execTraceCache *lru.ARCCache[*types.TipSet, tipSetCacheEntry]
} }
// Caches a single state tree // Caches a single state tree
@ -146,6 +153,11 @@ type treeCache struct {
tree *state.StateTree tree *state.StateTree
} }
type tipSetCacheEntry struct {
cid cid.Cid
invocTrace []*api.InvocResult
}
func NewStateManager(cs *store.ChainStore, exec Executor, sys vm.SyscallBuilder, us UpgradeSchedule, beacon beacon.Schedule, metadataDs dstore.Batching, msgIndex index.MsgIndex) (*StateManager, error) { func NewStateManager(cs *store.ChainStore, exec Executor, sys vm.SyscallBuilder, us UpgradeSchedule, beacon beacon.Schedule, metadataDs dstore.Batching, msgIndex index.MsgIndex) (*StateManager, error) {
// If we have upgrades, make sure they're in-order and make sense. // If we have upgrades, make sure they're in-order and make sense.
if err := us.Validate(); err != nil { if err := us.Validate(); err != nil {
@ -185,6 +197,11 @@ func NewStateManager(cs *store.ChainStore, exec Executor, sys vm.SyscallBuilder,
} }
} }
execTraceCache, err := lru.NewARC[*types.TipSet, tipSetCacheEntry](execTraceCacheSize)
if err != nil {
return nil, err
}
return &StateManager{ return &StateManager{
networkVersions: networkVersions, networkVersions: networkVersions,
latestVersion: lastVersion, latestVersion: lastVersion,
@ -202,6 +219,7 @@ func NewStateManager(cs *store.ChainStore, exec Executor, sys vm.SyscallBuilder,
}, },
compWait: make(map[string]chan struct{}), compWait: make(map[string]chan struct{}),
msgIndex: msgIndex, msgIndex: msgIndex,
execTraceCache: execTraceCache,
}, nil }, nil
} }