core/state, eth, trie: stabilize memory use, fix memory leak

This commit is contained in:
Péter Szilágyi 2020-08-26 13:05:06 +03:00
parent 63a9d4b2ae
commit d8da0b3d81
No known key found for this signature in database
GPG Key ID: E9AE538CEDF8293D
7 changed files with 58 additions and 29 deletions

View File

@ -847,7 +847,7 @@ func (s *StateDB) Commit(deleteEmptyObjects bool) (common.Hash, error) {
// The onleaf func is called _serially_, so we can reuse the same account // The onleaf func is called _serially_, so we can reuse the same account
// for unmarshalling every time. // for unmarshalling every time.
var account Account var account Account
root, err := s.trie.Commit(func(leaf []byte, parent common.Hash) error { root, err := s.trie.Commit(func(path []byte, leaf []byte, parent common.Hash) error {
if err := rlp.DecodeBytes(leaf, &account); err != nil { if err := rlp.DecodeBytes(leaf, &account); err != nil {
return nil return nil
} }

View File

@ -28,13 +28,13 @@ import (
// NewStateSync create a new state trie download scheduler. // NewStateSync create a new state trie download scheduler.
func NewStateSync(root common.Hash, database ethdb.KeyValueReader, bloom *trie.SyncBloom) *trie.Sync { func NewStateSync(root common.Hash, database ethdb.KeyValueReader, bloom *trie.SyncBloom) *trie.Sync {
var syncer *trie.Sync var syncer *trie.Sync
callback := func(leaf []byte, parent common.Hash) error { callback := func(path []byte, leaf []byte, parent common.Hash) error {
var obj Account var obj Account
if err := rlp.Decode(bytes.NewReader(leaf), &obj); err != nil { if err := rlp.Decode(bytes.NewReader(leaf), &obj); err != nil {
return err return err
} }
syncer.AddSubTrie(obj.Root, 64, parent, nil) syncer.AddSubTrie(obj.Root, path, parent, nil)
syncer.AddCodeEntry(common.BytesToHash(obj.CodeHash), 64, parent) syncer.AddCodeEntry(common.BytesToHash(obj.CodeHash), path, parent)
return nil return nil
} }
syncer = trie.NewSync(root, database, callback, bloom) syncer = trie.NewSync(root, database, callback, bloom)

View File

@ -1611,7 +1611,13 @@ func (d *Downloader) processFastSyncContent(latest *types.Header) error {
// Start syncing state of the reported head block. This should get us most of // Start syncing state of the reported head block. This should get us most of
// the state of the pivot block. // the state of the pivot block.
sync := d.syncState(latest.Root) sync := d.syncState(latest.Root)
defer sync.Cancel() defer func() {
// The `sync` object is replaced every time the pivot moves. We need to
// defer close the very last active one, hence the lazy evaluation vs.
// calling defer sync.Cancel() !!!
sync.Cancel()
}()
closeOnErr := func(s *stateSync) { closeOnErr := func(s *stateSync) {
if err := s.Wait(); err != nil && err != errCancelStateFetch && err != errCanceled { if err := s.Wait(); err != nil && err != errCancelStateFetch && err != errCanceled {
d.queue.Close() // wake up Results d.queue.Close() // wake up Results
@ -1674,9 +1680,8 @@ func (d *Downloader) processFastSyncContent(latest *types.Header) error {
// If new pivot block found, cancel old state retrieval and restart // If new pivot block found, cancel old state retrieval and restart
if oldPivot != P { if oldPivot != P {
sync.Cancel() sync.Cancel()
sync = d.syncState(P.Header.Root) sync = d.syncState(P.Header.Root)
defer sync.Cancel()
go closeOnErr(sync) go closeOnErr(sync)
oldPivot = P oldPivot = P
} }

View File

@ -226,12 +226,12 @@ func (c *committer) commitLoop(db *Database) {
switch n := n.(type) { switch n := n.(type) {
case *shortNode: case *shortNode:
if child, ok := n.Val.(valueNode); ok { if child, ok := n.Val.(valueNode); ok {
c.onleaf(child, hash) c.onleaf(nil, child, hash)
} }
case *fullNode: case *fullNode:
for i := 0; i < 16; i++ { for i := 0; i < 16; i++ {
if child, ok := n.Children[i].(valueNode); ok { if child, ok := n.Children[i].(valueNode); ok {
c.onleaf(child, hash) c.onleaf(nil, child, hash)
} }
} }
} }

View File

@ -34,14 +34,19 @@ var ErrNotRequested = errors.New("not requested")
// node it already processed previously. // node it already processed previously.
var ErrAlreadyProcessed = errors.New("already processed") var ErrAlreadyProcessed = errors.New("already processed")
// maxFetchesPerDepth is the maximum number of pending trie nodes per depth. The
// role of this value is to limit the number of trie nodes that get expanded in
// memory if the node was configured with a significant number of peers.
const maxFetchesPerDepth = 16384
// request represents a scheduled or already in-flight state retrieval request. // request represents a scheduled or already in-flight state retrieval request.
type request struct { type request struct {
path []byte // Merkle path leading to this node for prioritization
hash common.Hash // Hash of the node data content to retrieve hash common.Hash // Hash of the node data content to retrieve
data []byte // Data content of the node, cached until all subtrees complete data []byte // Data content of the node, cached until all subtrees complete
code bool // Whether this is a code entry code bool // Whether this is a code entry
parents []*request // Parent state nodes referencing this entry (notify all upon completion) parents []*request // Parent state nodes referencing this entry (notify all upon completion)
depth int // Depth level within the trie the node is located to prioritise DFS
deps int // Number of dependencies before allowed to commit this node deps int // Number of dependencies before allowed to commit this node
callback LeafCallback // Callback to invoke if a leaf node it reached on this branch callback LeafCallback // Callback to invoke if a leaf node it reached on this branch
@ -89,6 +94,7 @@ type Sync struct {
nodeReqs map[common.Hash]*request // Pending requests pertaining to a trie node hash nodeReqs map[common.Hash]*request // Pending requests pertaining to a trie node hash
codeReqs map[common.Hash]*request // Pending requests pertaining to a code hash codeReqs map[common.Hash]*request // Pending requests pertaining to a code hash
queue *prque.Prque // Priority queue with the pending requests queue *prque.Prque // Priority queue with the pending requests
fetches map[int]int // Number of active fetches per trie node depth
bloom *SyncBloom // Bloom filter for fast state existence checks bloom *SyncBloom // Bloom filter for fast state existence checks
} }
@ -100,14 +106,15 @@ func NewSync(root common.Hash, database ethdb.KeyValueReader, callback LeafCallb
nodeReqs: make(map[common.Hash]*request), nodeReqs: make(map[common.Hash]*request),
codeReqs: make(map[common.Hash]*request), codeReqs: make(map[common.Hash]*request),
queue: prque.New(nil), queue: prque.New(nil),
fetches: make(map[int]int),
bloom: bloom, bloom: bloom,
} }
ts.AddSubTrie(root, 0, common.Hash{}, callback) ts.AddSubTrie(root, nil, common.Hash{}, callback)
return ts return ts
} }
// AddSubTrie registers a new trie to the sync code, rooted at the designated parent. // AddSubTrie registers a new trie to the sync code, rooted at the designated parent.
func (s *Sync) AddSubTrie(root common.Hash, depth int, parent common.Hash, callback LeafCallback) { func (s *Sync) AddSubTrie(root common.Hash, path []byte, parent common.Hash, callback LeafCallback) {
// Short circuit if the trie is empty or already known // Short circuit if the trie is empty or already known
if root == emptyRoot { if root == emptyRoot {
return return
@ -128,8 +135,8 @@ func (s *Sync) AddSubTrie(root common.Hash, depth int, parent common.Hash, callb
} }
// Assemble the new sub-trie sync request // Assemble the new sub-trie sync request
req := &request{ req := &request{
path: path,
hash: root, hash: root,
depth: depth,
callback: callback, callback: callback,
} }
// If this sub-trie has a designated parent, link them together // If this sub-trie has a designated parent, link them together
@ -147,7 +154,7 @@ func (s *Sync) AddSubTrie(root common.Hash, depth int, parent common.Hash, callb
// AddCodeEntry schedules the direct retrieval of a contract code that should not // AddCodeEntry schedules the direct retrieval of a contract code that should not
// be interpreted as a trie node, but rather accepted and stored into the database // be interpreted as a trie node, but rather accepted and stored into the database
// as is. // as is.
func (s *Sync) AddCodeEntry(hash common.Hash, depth int, parent common.Hash) { func (s *Sync) AddCodeEntry(hash common.Hash, path []byte, parent common.Hash) {
// Short circuit if the entry is empty or already known // Short circuit if the entry is empty or already known
if hash == emptyState { if hash == emptyState {
return return
@ -170,9 +177,9 @@ func (s *Sync) AddCodeEntry(hash common.Hash, depth int, parent common.Hash) {
} }
// Assemble the new sub-trie sync request // Assemble the new sub-trie sync request
req := &request{ req := &request{
hash: hash, path: path,
code: true, hash: hash,
depth: depth, code: true,
} }
// If this sub-trie has a designated parent, link them together // If this sub-trie has a designated parent, link them together
if parent != (common.Hash{}) { if parent != (common.Hash{}) {
@ -190,7 +197,18 @@ func (s *Sync) AddCodeEntry(hash common.Hash, depth int, parent common.Hash) {
func (s *Sync) Missing(max int) []common.Hash { func (s *Sync) Missing(max int) []common.Hash {
var requests []common.Hash var requests []common.Hash
for !s.queue.Empty() && (max == 0 || len(requests) < max) { for !s.queue.Empty() && (max == 0 || len(requests) < max) {
requests = append(requests, s.queue.PopItem().(common.Hash)) // Retrieve th enext item in line
item, prio := s.queue.Peek()
// If we have too many already-pending tasks for this depth, throttle
depth := int(prio >> 56)
if s.fetches[depth] > maxFetchesPerDepth {
break
}
// Item is allowed to be scheduled, add it to the task list
s.queue.Pop()
s.fetches[depth]++
requests = append(requests, item.(common.Hash))
} }
return requests return requests
} }
@ -285,7 +303,11 @@ func (s *Sync) schedule(req *request) {
// is a trie node and code has same hash. In this case two elements // is a trie node and code has same hash. In this case two elements
// with same hash and same or different depth will be pushed. But it's // with same hash and same or different depth will be pushed. But it's
// ok the worst case is the second response will be treated as duplicated. // ok the worst case is the second response will be treated as duplicated.
s.queue.Push(req.hash, int64(req.depth)) prio := int64(len(req.path)) << 56 // depth >= 128 will never happen, storage leaves will be included in their parents
for i := 0; i < 14 && i < len(req.path); i++ {
prio |= int64(15-req.path[i]) << (52 - i*4) // 15-nibble => lexicographic order
}
s.queue.Push(req.hash, prio)
} }
// children retrieves all the missing children of a state trie entry for future // children retrieves all the missing children of a state trie entry for future
@ -293,23 +315,23 @@ func (s *Sync) schedule(req *request) {
func (s *Sync) children(req *request, object node) ([]*request, error) { func (s *Sync) children(req *request, object node) ([]*request, error) {
// Gather all the children of the node, irrelevant whether known or not // Gather all the children of the node, irrelevant whether known or not
type child struct { type child struct {
node node path []byte
depth int node node
} }
var children []child var children []child
switch node := (object).(type) { switch node := (object).(type) {
case *shortNode: case *shortNode:
children = []child{{ children = []child{{
node: node.Val, node: node.Val,
depth: req.depth + len(node.Key), path: append(append([]byte(nil), req.path...), node.Key...),
}} }}
case *fullNode: case *fullNode:
for i := 0; i < 17; i++ { for i := 0; i < 17; i++ {
if node.Children[i] != nil { if node.Children[i] != nil {
children = append(children, child{ children = append(children, child{
node: node.Children[i], node: node.Children[i],
depth: req.depth + 1, path: append(append([]byte(nil), req.path...), byte(i)),
}) })
} }
} }
@ -322,7 +344,7 @@ func (s *Sync) children(req *request, object node) ([]*request, error) {
// Notify any external watcher of a new key/value node // Notify any external watcher of a new key/value node
if req.callback != nil { if req.callback != nil {
if node, ok := (child.node).(valueNode); ok { if node, ok := (child.node).(valueNode); ok {
if err := req.callback(node, req.hash); err != nil { if err := req.callback(req.path, node, req.hash); err != nil {
return nil, err return nil, err
} }
} }
@ -346,9 +368,9 @@ func (s *Sync) children(req *request, object node) ([]*request, error) {
} }
// Locally unknown node, schedule for retrieval // Locally unknown node, schedule for retrieval
requests = append(requests, &request{ requests = append(requests, &request{
path: child.path,
hash: hash, hash: hash,
parents: []*request{req}, parents: []*request{req},
depth: child.depth,
callback: req.callback, callback: req.callback,
}) })
} }
@ -364,9 +386,11 @@ func (s *Sync) commit(req *request) (err error) {
if req.code { if req.code {
s.membatch.codes[req.hash] = req.data s.membatch.codes[req.hash] = req.data
delete(s.codeReqs, req.hash) delete(s.codeReqs, req.hash)
s.fetches[len(req.path)]--
} else { } else {
s.membatch.nodes[req.hash] = req.data s.membatch.nodes[req.hash] = req.data
delete(s.nodeReqs, req.hash) delete(s.nodeReqs, req.hash)
s.fetches[len(req.path)]--
} }
// Check all parents for completion // Check all parents for completion
for _, parent := range req.parents { for _, parent := range req.parents {

View File

@ -38,7 +38,7 @@ var (
// LeafCallback is a callback type invoked when a trie operation reaches a leaf // LeafCallback is a callback type invoked when a trie operation reaches a leaf
// node. It's used by state sync and commit to allow handling external references // node. It's used by state sync and commit to allow handling external references
// between account and storage tries. // between account and storage tries.
type LeafCallback func(leaf []byte, parent common.Hash) error type LeafCallback func(path []byte, leaf []byte, parent common.Hash) error
// Trie is a Merkle Patricia Trie. // Trie is a Merkle Patricia Trie.
// The zero value is an empty trie with no database. // The zero value is an empty trie with no database.

View File

@ -565,7 +565,7 @@ func BenchmarkCommitAfterHash(b *testing.B) {
benchmarkCommitAfterHash(b, nil) benchmarkCommitAfterHash(b, nil)
}) })
var a account var a account
onleaf := func(leaf []byte, parent common.Hash) error { onleaf := func(path []byte, leaf []byte, parent common.Hash) error {
rlp.DecodeBytes(leaf, &a) rlp.DecodeBytes(leaf, &a)
return nil return nil
} }