* ethdb: use pebble Co-authored-by: Gary Rong <garyrong0905@gmail.com> foo update * apply suggested changes * flags: go format node: fix ddir lookup mistake accounts/abi/bind: fix go.mod replacement for generated binding deps: update pebble + with fix 32-bit build * ethdb/pebble: respect max memtable size * core/rawdb, ethdb: enable pebble on non-32bit platforms only * core/rawdb: fix build tags, fix some review concerns * core/rawdb: refactor methods for database opening * core/rawdb: remove erroneous build tag * cmd/geth: fix the flag default handling + testcase * cmd/geth: improve testing regarding custom backends * ethdb/pebble, deps: update pebble dependency * core/rawdb: replace method with Open * ethdb/pebble: several updates for pebble (#49) * ethdb/pebble: fix size count in batch * ethdb/pebble: disable seek compaction * ethdb/pebble: more fixes * ethdb, core, cmd: polish and fixes (#50) * cmd/utils, core/rawdb, ethdb/pebble: address some review concerns * Update flags.go * ethdb/pebble: minor refactors * ethdb/pebble: avoid copy on batch replay * ethdb: fix compilation flaw * cmd: fix test fail due to mismatching error message * cmd/geth, node: rename backingdb to db.engine --------- Co-authored-by: Jared Wasinger <j-wasinger@hotmail.com> Co-authored-by: rjl493456442 <garyrong0905@gmail.com> Co-authored-by: Péter Szilágyi <peterke@gmail.com>
		
			
				
	
	
		
			584 lines
		
	
	
		
			22 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
			
		
		
	
	
			584 lines
		
	
	
		
			22 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
| // Copyright 2018 The go-ethereum Authors
 | |
| // This file is part of the go-ethereum library.
 | |
| //
 | |
| // The go-ethereum library is free software: you can redistribute it and/or modify
 | |
| // it under the terms of the GNU Lesser General Public License as published by
 | |
| // the Free Software Foundation, either version 3 of the License, or
 | |
| // (at your option) any later version.
 | |
| //
 | |
| // The go-ethereum library is distributed in the hope that it will be useful,
 | |
| // but WITHOUT ANY WARRANTY; without even the implied warranty of
 | |
| // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
 | |
| // GNU Lesser General Public License for more details.
 | |
| //
 | |
| // You should have received a copy of the GNU Lesser General Public License
 | |
| // along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
 | |
| 
 | |
| package rawdb
 | |
| 
 | |
| import (
 | |
| 	"bytes"
 | |
| 	"errors"
 | |
| 	"fmt"
 | |
| 	"os"
 | |
| 	"path"
 | |
| 	"path/filepath"
 | |
| 	"strings"
 | |
| 	"sync/atomic"
 | |
| 	"time"
 | |
| 
 | |
| 	"github.com/ethereum/go-ethereum/common"
 | |
| 	"github.com/ethereum/go-ethereum/ethdb"
 | |
| 	"github.com/ethereum/go-ethereum/ethdb/leveldb"
 | |
| 	"github.com/ethereum/go-ethereum/ethdb/memorydb"
 | |
| 	"github.com/ethereum/go-ethereum/log"
 | |
| 	"github.com/olekukonko/tablewriter"
 | |
| )
 | |
| 
 | |
| // freezerdb is a database wrapper that enabled freezer data retrievals.
 | |
| type freezerdb struct {
 | |
| 	ancientRoot string
 | |
| 	ethdb.KeyValueStore
 | |
| 	ethdb.AncientStore
 | |
| }
 | |
| 
 | |
| // AncientDatadir returns the path of root ancient directory.
 | |
| func (frdb *freezerdb) AncientDatadir() (string, error) {
 | |
| 	return frdb.ancientRoot, nil
 | |
| }
 | |
| 
 | |
| // Close implements io.Closer, closing both the fast key-value store as well as
 | |
| // the slow ancient tables.
 | |
| func (frdb *freezerdb) Close() error {
 | |
| 	var errs []error
 | |
| 	if err := frdb.AncientStore.Close(); err != nil {
 | |
| 		errs = append(errs, err)
 | |
| 	}
 | |
| 	if err := frdb.KeyValueStore.Close(); err != nil {
 | |
| 		errs = append(errs, err)
 | |
| 	}
 | |
| 	if len(errs) != 0 {
 | |
| 		return fmt.Errorf("%v", errs)
 | |
| 	}
 | |
| 	return nil
 | |
| }
 | |
| 
 | |
| // Freeze is a helper method used for external testing to trigger and block until
 | |
| // a freeze cycle completes, without having to sleep for a minute to trigger the
 | |
| // automatic background run.
 | |
| func (frdb *freezerdb) Freeze(threshold uint64) error {
 | |
| 	if frdb.AncientStore.(*chainFreezer).readonly {
 | |
| 		return errReadOnly
 | |
| 	}
 | |
| 	// Set the freezer threshold to a temporary value
 | |
| 	defer func(old uint64) {
 | |
| 		atomic.StoreUint64(&frdb.AncientStore.(*chainFreezer).threshold, old)
 | |
| 	}(atomic.LoadUint64(&frdb.AncientStore.(*chainFreezer).threshold))
 | |
| 	atomic.StoreUint64(&frdb.AncientStore.(*chainFreezer).threshold, threshold)
 | |
| 
 | |
| 	// Trigger a freeze cycle and block until it's done
 | |
| 	trigger := make(chan struct{}, 1)
 | |
| 	frdb.AncientStore.(*chainFreezer).trigger <- trigger
 | |
| 	<-trigger
 | |
| 	return nil
 | |
| }
 | |
| 
 | |
| // nofreezedb is a database wrapper that disables freezer data retrievals.
 | |
| type nofreezedb struct {
 | |
| 	ethdb.KeyValueStore
 | |
| }
 | |
| 
 | |
| // HasAncient returns an error as we don't have a backing chain freezer.
 | |
| func (db *nofreezedb) HasAncient(kind string, number uint64) (bool, error) {
 | |
| 	return false, errNotSupported
 | |
| }
 | |
| 
 | |
| // Ancient returns an error as we don't have a backing chain freezer.
 | |
| func (db *nofreezedb) Ancient(kind string, number uint64) ([]byte, error) {
 | |
| 	return nil, errNotSupported
 | |
| }
 | |
| 
 | |
| // AncientRange returns an error as we don't have a backing chain freezer.
 | |
| func (db *nofreezedb) AncientRange(kind string, start, max, maxByteSize uint64) ([][]byte, error) {
 | |
| 	return nil, errNotSupported
 | |
| }
 | |
| 
 | |
| // Ancients returns an error as we don't have a backing chain freezer.
 | |
| func (db *nofreezedb) Ancients() (uint64, error) {
 | |
| 	return 0, errNotSupported
 | |
| }
 | |
| 
 | |
| // Tail returns an error as we don't have a backing chain freezer.
 | |
| func (db *nofreezedb) Tail() (uint64, error) {
 | |
| 	return 0, errNotSupported
 | |
| }
 | |
| 
 | |
| // AncientSize returns an error as we don't have a backing chain freezer.
 | |
| func (db *nofreezedb) AncientSize(kind string) (uint64, error) {
 | |
| 	return 0, errNotSupported
 | |
| }
 | |
| 
 | |
| // ModifyAncients is not supported.
 | |
| func (db *nofreezedb) ModifyAncients(func(ethdb.AncientWriteOp) error) (int64, error) {
 | |
| 	return 0, errNotSupported
 | |
| }
 | |
| 
 | |
| // TruncateHead returns an error as we don't have a backing chain freezer.
 | |
| func (db *nofreezedb) TruncateHead(items uint64) error {
 | |
| 	return errNotSupported
 | |
| }
 | |
| 
 | |
| // TruncateTail returns an error as we don't have a backing chain freezer.
 | |
| func (db *nofreezedb) TruncateTail(items uint64) error {
 | |
| 	return errNotSupported
 | |
| }
 | |
| 
 | |
| // Sync returns an error as we don't have a backing chain freezer.
 | |
| func (db *nofreezedb) Sync() error {
 | |
| 	return errNotSupported
 | |
| }
 | |
| 
 | |
| func (db *nofreezedb) ReadAncients(fn func(reader ethdb.AncientReaderOp) error) (err error) {
 | |
| 	// Unlike other ancient-related methods, this method does not return
 | |
| 	// errNotSupported when invoked.
 | |
| 	// The reason for this is that the caller might want to do several things:
 | |
| 	// 1. Check if something is in freezer,
 | |
| 	// 2. If not, check leveldb.
 | |
| 	//
 | |
| 	// This will work, since the ancient-checks inside 'fn' will return errors,
 | |
| 	// and the leveldb work will continue.
 | |
| 	//
 | |
| 	// If we instead were to return errNotSupported here, then the caller would
 | |
| 	// have to explicitly check for that, having an extra clause to do the
 | |
| 	// non-ancient operations.
 | |
| 	return fn(db)
 | |
| }
 | |
| 
 | |
| // MigrateTable processes the entries in a given table in sequence
 | |
| // converting them to a new format if they're of an old format.
 | |
| func (db *nofreezedb) MigrateTable(kind string, convert convertLegacyFn) error {
 | |
| 	return errNotSupported
 | |
| }
 | |
| 
 | |
| // AncientDatadir returns an error as we don't have a backing chain freezer.
 | |
| func (db *nofreezedb) AncientDatadir() (string, error) {
 | |
| 	return "", errNotSupported
 | |
| }
 | |
| 
 | |
| // NewDatabase creates a high level database on top of a given key-value data
 | |
| // store without a freezer moving immutable chain segments into cold storage.
 | |
| func NewDatabase(db ethdb.KeyValueStore) ethdb.Database {
 | |
| 	return &nofreezedb{KeyValueStore: db}
 | |
| }
 | |
| 
 | |
| // resolveChainFreezerDir is a helper function which resolves the absolute path
 | |
| // of chain freezer by considering backward compatibility.
 | |
| func resolveChainFreezerDir(ancient string) string {
 | |
| 	// Check if the chain freezer is already present in the specified
 | |
| 	// sub folder, if not then two possibilities:
 | |
| 	// - chain freezer is not initialized
 | |
| 	// - chain freezer exists in legacy location (root ancient folder)
 | |
| 	freezer := path.Join(ancient, chainFreezerName)
 | |
| 	if !common.FileExist(freezer) {
 | |
| 		if !common.FileExist(ancient) {
 | |
| 			// The entire ancient store is not initialized, still use the sub
 | |
| 			// folder for initialization.
 | |
| 		} else {
 | |
| 			// Ancient root is already initialized, then we hold the assumption
 | |
| 			// that chain freezer is also initialized and located in root folder.
 | |
| 			// In this case fallback to legacy location.
 | |
| 			freezer = ancient
 | |
| 			log.Info("Found legacy ancient chain path", "location", ancient)
 | |
| 		}
 | |
| 	}
 | |
| 	return freezer
 | |
| }
 | |
| 
 | |
| // NewDatabaseWithFreezer creates a high level database on top of a given key-
 | |
| // value data store with a freezer moving immutable chain segments into cold
 | |
| // storage. The passed ancient indicates the path of root ancient directory
 | |
| // where the chain freezer can be opened.
 | |
| func NewDatabaseWithFreezer(db ethdb.KeyValueStore, ancient string, namespace string, readonly bool) (ethdb.Database, error) {
 | |
| 	// Create the idle freezer instance
 | |
| 	frdb, err := newChainFreezer(resolveChainFreezerDir(ancient), namespace, readonly, freezerTableSize, chainFreezerNoSnappy)
 | |
| 	if err != nil {
 | |
| 		return nil, err
 | |
| 	}
 | |
| 	// Since the freezer can be stored separately from the user's key-value database,
 | |
| 	// there's a fairly high probability that the user requests invalid combinations
 | |
| 	// of the freezer and database. Ensure that we don't shoot ourselves in the foot
 | |
| 	// by serving up conflicting data, leading to both datastores getting corrupted.
 | |
| 	//
 | |
| 	//   - If both the freezer and key-value store is empty (no genesis), we just
 | |
| 	//     initialized a new empty freezer, so everything's fine.
 | |
| 	//   - If the key-value store is empty, but the freezer is not, we need to make
 | |
| 	//     sure the user's genesis matches the freezer. That will be checked in the
 | |
| 	//     blockchain, since we don't have the genesis block here (nor should we at
 | |
| 	//     this point care, the key-value/freezer combo is valid).
 | |
| 	//   - If neither the key-value store nor the freezer is empty, cross validate
 | |
| 	//     the genesis hashes to make sure they are compatible. If they are, also
 | |
| 	//     ensure that there's no gap between the freezer and subsequently leveldb.
 | |
| 	//   - If the key-value store is not empty, but the freezer is we might just be
 | |
| 	//     upgrading to the freezer release, or we might have had a small chain and
 | |
| 	//     not frozen anything yet. Ensure that no blocks are missing yet from the
 | |
| 	//     key-value store, since that would mean we already had an old freezer.
 | |
| 
 | |
| 	// If the genesis hash is empty, we have a new key-value store, so nothing to
 | |
| 	// validate in this method. If, however, the genesis hash is not nil, compare
 | |
| 	// it to the freezer content.
 | |
| 	if kvgenesis, _ := db.Get(headerHashKey(0)); len(kvgenesis) > 0 {
 | |
| 		if frozen, _ := frdb.Ancients(); frozen > 0 {
 | |
| 			// If the freezer already contains something, ensure that the genesis blocks
 | |
| 			// match, otherwise we might mix up freezers across chains and destroy both
 | |
| 			// the freezer and the key-value store.
 | |
| 			frgenesis, err := frdb.Ancient(chainFreezerHashTable, 0)
 | |
| 			if err != nil {
 | |
| 				return nil, fmt.Errorf("failed to retrieve genesis from ancient %v", err)
 | |
| 			} else if !bytes.Equal(kvgenesis, frgenesis) {
 | |
| 				return nil, fmt.Errorf("genesis mismatch: %#x (leveldb) != %#x (ancients)", kvgenesis, frgenesis)
 | |
| 			}
 | |
| 			// Key-value store and freezer belong to the same network. Ensure that they
 | |
| 			// are contiguous, otherwise we might end up with a non-functional freezer.
 | |
| 			if kvhash, _ := db.Get(headerHashKey(frozen)); len(kvhash) == 0 {
 | |
| 				// Subsequent header after the freezer limit is missing from the database.
 | |
| 				// Reject startup if the database has a more recent head.
 | |
| 				if ldbNum := *ReadHeaderNumber(db, ReadHeadHeaderHash(db)); ldbNum > frozen-1 {
 | |
| 					return nil, fmt.Errorf("gap in the chain between ancients (#%d) and leveldb (#%d) ", frozen, ldbNum)
 | |
| 				}
 | |
| 				// Database contains only older data than the freezer, this happens if the
 | |
| 				// state was wiped and reinited from an existing freezer.
 | |
| 			}
 | |
| 			// Otherwise, key-value store continues where the freezer left off, all is fine.
 | |
| 			// We might have duplicate blocks (crash after freezer write but before key-value
 | |
| 			// store deletion, but that's fine).
 | |
| 		} else {
 | |
| 			// If the freezer is empty, ensure nothing was moved yet from the key-value
 | |
| 			// store, otherwise we'll end up missing data. We check block #1 to decide
 | |
| 			// if we froze anything previously or not, but do take care of databases with
 | |
| 			// only the genesis block.
 | |
| 			if ReadHeadHeaderHash(db) != common.BytesToHash(kvgenesis) {
 | |
| 				// Key-value store contains more data than the genesis block, make sure we
 | |
| 				// didn't freeze anything yet.
 | |
| 				if kvblob, _ := db.Get(headerHashKey(1)); len(kvblob) == 0 {
 | |
| 					return nil, errors.New("ancient chain segments already extracted, please set --datadir.ancient to the correct path")
 | |
| 				}
 | |
| 				// Block #1 is still in the database, we're allowed to init a new freezer
 | |
| 			}
 | |
| 			// Otherwise, the head header is still the genesis, we're allowed to init a new
 | |
| 			// freezer.
 | |
| 		}
 | |
| 	}
 | |
| 	// Freezer is consistent with the key-value database, permit combining the two
 | |
| 	if !frdb.readonly {
 | |
| 		frdb.wg.Add(1)
 | |
| 		go func() {
 | |
| 			frdb.freeze(db)
 | |
| 			frdb.wg.Done()
 | |
| 		}()
 | |
| 	}
 | |
| 	return &freezerdb{
 | |
| 		ancientRoot:   ancient,
 | |
| 		KeyValueStore: db,
 | |
| 		AncientStore:  frdb,
 | |
| 	}, nil
 | |
| }
 | |
| 
 | |
| // NewMemoryDatabase creates an ephemeral in-memory key-value database without a
 | |
| // freezer moving immutable chain segments into cold storage.
 | |
| func NewMemoryDatabase() ethdb.Database {
 | |
| 	return NewDatabase(memorydb.New())
 | |
| }
 | |
| 
 | |
| // NewMemoryDatabaseWithCap creates an ephemeral in-memory key-value database
 | |
| // with an initial starting capacity, but without a freezer moving immutable
 | |
| // chain segments into cold storage.
 | |
| func NewMemoryDatabaseWithCap(size int) ethdb.Database {
 | |
| 	return NewDatabase(memorydb.NewWithCap(size))
 | |
| }
 | |
| 
 | |
| // NewLevelDBDatabase creates a persistent key-value database without a freezer
 | |
| // moving immutable chain segments into cold storage.
 | |
| func NewLevelDBDatabase(file string, cache int, handles int, namespace string, readonly bool) (ethdb.Database, error) {
 | |
| 	db, err := leveldb.New(file, cache, handles, namespace, readonly)
 | |
| 	if err != nil {
 | |
| 		return nil, err
 | |
| 	}
 | |
| 	log.Info("Using LevelDB as the backing database")
 | |
| 	return NewDatabase(db), nil
 | |
| }
 | |
| 
 | |
| const (
 | |
| 	dbPebble  = "pebble"
 | |
| 	dbLeveldb = "leveldb"
 | |
| )
 | |
| 
 | |
| // hasPreexistingDb checks the given data directory whether a database is already
 | |
| // instantiated at that location, and if so, returns the type of database (or the
 | |
| // empty string).
 | |
| func hasPreexistingDb(path string) string {
 | |
| 	if _, err := os.Stat(filepath.Join(path, "CURRENT")); err != nil {
 | |
| 		return "" // No pre-existing db
 | |
| 	}
 | |
| 	if matches, err := filepath.Glob(filepath.Join(path, "OPTIONS*")); len(matches) > 0 || err != nil {
 | |
| 		if err != nil {
 | |
| 			panic(err) // only possible if the pattern is malformed
 | |
| 		}
 | |
| 		return dbPebble
 | |
| 	}
 | |
| 	return dbLeveldb
 | |
| }
 | |
| 
 | |
| // OpenOptions contains the options to apply when opening a database.
 | |
| // OBS: If AncientsDirectory is empty, it indicates that no freezer is to be used.
 | |
| type OpenOptions struct {
 | |
| 	Type              string // "leveldb" | "pebble"
 | |
| 	Directory         string // the datadir
 | |
| 	AncientsDirectory string // the ancients-dir
 | |
| 	Namespace         string // the namespace for database relevant metrics
 | |
| 	Cache             int    // the capacity(in megabytes) of the data caching
 | |
| 	Handles           int    // number of files to be open simultaneously
 | |
| 	ReadOnly          bool
 | |
| }
 | |
| 
 | |
| // openKeyValueDatabase opens a disk-based key-value database, e.g. leveldb or pebble.
 | |
| //
 | |
| //	                      type == null          type != null
 | |
| //	                   +----------------------------------------
 | |
| //	db is non-existent |  leveldb default  |  specified type
 | |
| //	db is existent     |  from db          |  specified type (if compatible)
 | |
| func openKeyValueDatabase(o OpenOptions) (ethdb.Database, error) {
 | |
| 	existingDb := hasPreexistingDb(o.Directory)
 | |
| 	if len(existingDb) != 0 && len(o.Type) != 0 && o.Type != existingDb {
 | |
| 		return nil, fmt.Errorf("db.engine choice was %v but found pre-existing %v database in specified data directory", o.Type, existingDb)
 | |
| 	}
 | |
| 	if o.Type == dbPebble || existingDb == dbPebble {
 | |
| 		if PebbleEnabled {
 | |
| 			log.Info("Using pebble as the backing database")
 | |
| 			return NewPebbleDBDatabase(o.Directory, o.Cache, o.Handles, o.Namespace, o.ReadOnly)
 | |
| 		} else {
 | |
| 			return nil, errors.New("db.engine 'pebble' not supported on this platform")
 | |
| 		}
 | |
| 	}
 | |
| 	if len(o.Type) != 0 && o.Type != dbLeveldb {
 | |
| 		return nil, fmt.Errorf("unknown db.engine %v", o.Type)
 | |
| 	}
 | |
| 	log.Info("Using leveldb as the backing database")
 | |
| 	// Use leveldb, either as default (no explicit choice), or pre-existing, or chosen explicitly
 | |
| 	return NewLevelDBDatabase(o.Directory, o.Cache, o.Handles, o.Namespace, o.ReadOnly)
 | |
| }
 | |
| 
 | |
| // Open opens both a disk-based key-value database such as leveldb or pebble, but also
 | |
| // integrates it with a freezer database -- if the AncientDir option has been
 | |
| // set on the provided OpenOptions.
 | |
| // The passed o.AncientDir indicates the path of root ancient directory where
 | |
| // the chain freezer can be opened.
 | |
| func Open(o OpenOptions) (ethdb.Database, error) {
 | |
| 	kvdb, err := openKeyValueDatabase(o)
 | |
| 	if err != nil {
 | |
| 		return nil, err
 | |
| 	}
 | |
| 	if len(o.AncientsDirectory) == 0 {
 | |
| 		return kvdb, nil
 | |
| 	}
 | |
| 	frdb, err := NewDatabaseWithFreezer(kvdb, o.AncientsDirectory, o.Namespace, o.ReadOnly)
 | |
| 	if err != nil {
 | |
| 		kvdb.Close()
 | |
| 		return nil, err
 | |
| 	}
 | |
| 	return frdb, nil
 | |
| }
 | |
| 
 | |
| type counter uint64
 | |
| 
 | |
| func (c counter) String() string {
 | |
| 	return fmt.Sprintf("%d", c)
 | |
| }
 | |
| 
 | |
| func (c counter) Percentage(current uint64) string {
 | |
| 	return fmt.Sprintf("%d", current*100/uint64(c))
 | |
| }
 | |
| 
 | |
| // stat stores sizes and count for a parameter
 | |
| type stat struct {
 | |
| 	size  common.StorageSize
 | |
| 	count counter
 | |
| }
 | |
| 
 | |
| // Add size to the stat and increase the counter by 1
 | |
| func (s *stat) Add(size common.StorageSize) {
 | |
| 	s.size += size
 | |
| 	s.count++
 | |
| }
 | |
| 
 | |
| func (s *stat) Size() string {
 | |
| 	return s.size.String()
 | |
| }
 | |
| 
 | |
| func (s *stat) Count() string {
 | |
| 	return s.count.String()
 | |
| }
 | |
| 
 | |
| // InspectDatabase traverses the entire database and checks the size
 | |
| // of all different categories of data.
 | |
| func InspectDatabase(db ethdb.Database, keyPrefix, keyStart []byte) error {
 | |
| 	it := db.NewIterator(keyPrefix, keyStart)
 | |
| 	defer it.Release()
 | |
| 
 | |
| 	var (
 | |
| 		count  int64
 | |
| 		start  = time.Now()
 | |
| 		logged = time.Now()
 | |
| 
 | |
| 		// Key-value store statistics
 | |
| 		headers         stat
 | |
| 		bodies          stat
 | |
| 		receipts        stat
 | |
| 		tds             stat
 | |
| 		numHashPairings stat
 | |
| 		hashNumPairings stat
 | |
| 		tries           stat
 | |
| 		codes           stat
 | |
| 		txLookups       stat
 | |
| 		accountSnaps    stat
 | |
| 		storageSnaps    stat
 | |
| 		preimages       stat
 | |
| 		bloomBits       stat
 | |
| 		beaconHeaders   stat
 | |
| 		cliqueSnaps     stat
 | |
| 
 | |
| 		// Les statistic
 | |
| 		chtTrieNodes   stat
 | |
| 		bloomTrieNodes stat
 | |
| 
 | |
| 		// Meta- and unaccounted data
 | |
| 		metadata    stat
 | |
| 		unaccounted stat
 | |
| 
 | |
| 		// Totals
 | |
| 		total common.StorageSize
 | |
| 	)
 | |
| 	// Inspect key-value database first.
 | |
| 	for it.Next() {
 | |
| 		var (
 | |
| 			key  = it.Key()
 | |
| 			size = common.StorageSize(len(key) + len(it.Value()))
 | |
| 		)
 | |
| 		total += size
 | |
| 		switch {
 | |
| 		case bytes.HasPrefix(key, headerPrefix) && len(key) == (len(headerPrefix)+8+common.HashLength):
 | |
| 			headers.Add(size)
 | |
| 		case bytes.HasPrefix(key, blockBodyPrefix) && len(key) == (len(blockBodyPrefix)+8+common.HashLength):
 | |
| 			bodies.Add(size)
 | |
| 		case bytes.HasPrefix(key, blockReceiptsPrefix) && len(key) == (len(blockReceiptsPrefix)+8+common.HashLength):
 | |
| 			receipts.Add(size)
 | |
| 		case bytes.HasPrefix(key, headerPrefix) && bytes.HasSuffix(key, headerTDSuffix):
 | |
| 			tds.Add(size)
 | |
| 		case bytes.HasPrefix(key, headerPrefix) && bytes.HasSuffix(key, headerHashSuffix):
 | |
| 			numHashPairings.Add(size)
 | |
| 		case bytes.HasPrefix(key, headerNumberPrefix) && len(key) == (len(headerNumberPrefix)+common.HashLength):
 | |
| 			hashNumPairings.Add(size)
 | |
| 		case len(key) == common.HashLength:
 | |
| 			tries.Add(size)
 | |
| 		case bytes.HasPrefix(key, CodePrefix) && len(key) == len(CodePrefix)+common.HashLength:
 | |
| 			codes.Add(size)
 | |
| 		case bytes.HasPrefix(key, txLookupPrefix) && len(key) == (len(txLookupPrefix)+common.HashLength):
 | |
| 			txLookups.Add(size)
 | |
| 		case bytes.HasPrefix(key, SnapshotAccountPrefix) && len(key) == (len(SnapshotAccountPrefix)+common.HashLength):
 | |
| 			accountSnaps.Add(size)
 | |
| 		case bytes.HasPrefix(key, SnapshotStoragePrefix) && len(key) == (len(SnapshotStoragePrefix)+2*common.HashLength):
 | |
| 			storageSnaps.Add(size)
 | |
| 		case bytes.HasPrefix(key, PreimagePrefix) && len(key) == (len(PreimagePrefix)+common.HashLength):
 | |
| 			preimages.Add(size)
 | |
| 		case bytes.HasPrefix(key, configPrefix) && len(key) == (len(configPrefix)+common.HashLength):
 | |
| 			metadata.Add(size)
 | |
| 		case bytes.HasPrefix(key, genesisPrefix) && len(key) == (len(genesisPrefix)+common.HashLength):
 | |
| 			metadata.Add(size)
 | |
| 		case bytes.HasPrefix(key, bloomBitsPrefix) && len(key) == (len(bloomBitsPrefix)+10+common.HashLength):
 | |
| 			bloomBits.Add(size)
 | |
| 		case bytes.HasPrefix(key, BloomBitsIndexPrefix):
 | |
| 			bloomBits.Add(size)
 | |
| 		case bytes.HasPrefix(key, skeletonHeaderPrefix) && len(key) == (len(skeletonHeaderPrefix)+8):
 | |
| 			beaconHeaders.Add(size)
 | |
| 		case bytes.HasPrefix(key, CliqueSnapshotPrefix) && len(key) == 7+common.HashLength:
 | |
| 			cliqueSnaps.Add(size)
 | |
| 		case bytes.HasPrefix(key, ChtTablePrefix) ||
 | |
| 			bytes.HasPrefix(key, ChtIndexTablePrefix) ||
 | |
| 			bytes.HasPrefix(key, ChtPrefix): // Canonical hash trie
 | |
| 			chtTrieNodes.Add(size)
 | |
| 		case bytes.HasPrefix(key, BloomTrieTablePrefix) ||
 | |
| 			bytes.HasPrefix(key, BloomTrieIndexPrefix) ||
 | |
| 			bytes.HasPrefix(key, BloomTriePrefix): // Bloomtrie sub
 | |
| 			bloomTrieNodes.Add(size)
 | |
| 		default:
 | |
| 			var accounted bool
 | |
| 			for _, meta := range [][]byte{
 | |
| 				databaseVersionKey, headHeaderKey, headBlockKey, headFastBlockKey, headFinalizedBlockKey,
 | |
| 				lastPivotKey, fastTrieProgressKey, snapshotDisabledKey, SnapshotRootKey, snapshotJournalKey,
 | |
| 				snapshotGeneratorKey, snapshotRecoveryKey, txIndexTailKey, fastTxLookupLimitKey,
 | |
| 				uncleanShutdownKey, badBlockKey, transitionStatusKey, skeletonSyncStatusKey,
 | |
| 			} {
 | |
| 				if bytes.Equal(key, meta) {
 | |
| 					metadata.Add(size)
 | |
| 					accounted = true
 | |
| 					break
 | |
| 				}
 | |
| 			}
 | |
| 			if !accounted {
 | |
| 				unaccounted.Add(size)
 | |
| 			}
 | |
| 		}
 | |
| 		count++
 | |
| 		if count%1000 == 0 && time.Since(logged) > 8*time.Second {
 | |
| 			log.Info("Inspecting database", "count", count, "elapsed", common.PrettyDuration(time.Since(start)))
 | |
| 			logged = time.Now()
 | |
| 		}
 | |
| 	}
 | |
| 	// Display the database statistic of key-value store.
 | |
| 	stats := [][]string{
 | |
| 		{"Key-Value store", "Headers", headers.Size(), headers.Count()},
 | |
| 		{"Key-Value store", "Bodies", bodies.Size(), bodies.Count()},
 | |
| 		{"Key-Value store", "Receipt lists", receipts.Size(), receipts.Count()},
 | |
| 		{"Key-Value store", "Difficulties", tds.Size(), tds.Count()},
 | |
| 		{"Key-Value store", "Block number->hash", numHashPairings.Size(), numHashPairings.Count()},
 | |
| 		{"Key-Value store", "Block hash->number", hashNumPairings.Size(), hashNumPairings.Count()},
 | |
| 		{"Key-Value store", "Transaction index", txLookups.Size(), txLookups.Count()},
 | |
| 		{"Key-Value store", "Bloombit index", bloomBits.Size(), bloomBits.Count()},
 | |
| 		{"Key-Value store", "Contract codes", codes.Size(), codes.Count()},
 | |
| 		{"Key-Value store", "Trie nodes", tries.Size(), tries.Count()},
 | |
| 		{"Key-Value store", "Trie preimages", preimages.Size(), preimages.Count()},
 | |
| 		{"Key-Value store", "Account snapshot", accountSnaps.Size(), accountSnaps.Count()},
 | |
| 		{"Key-Value store", "Storage snapshot", storageSnaps.Size(), storageSnaps.Count()},
 | |
| 		{"Key-Value store", "Beacon sync headers", beaconHeaders.Size(), beaconHeaders.Count()},
 | |
| 		{"Key-Value store", "Clique snapshots", cliqueSnaps.Size(), cliqueSnaps.Count()},
 | |
| 		{"Key-Value store", "Singleton metadata", metadata.Size(), metadata.Count()},
 | |
| 		{"Light client", "CHT trie nodes", chtTrieNodes.Size(), chtTrieNodes.Count()},
 | |
| 		{"Light client", "Bloom trie nodes", bloomTrieNodes.Size(), bloomTrieNodes.Count()},
 | |
| 	}
 | |
| 	// Inspect all registered append-only file store then.
 | |
| 	ancients, err := inspectFreezers(db)
 | |
| 	if err != nil {
 | |
| 		return err
 | |
| 	}
 | |
| 	for _, ancient := range ancients {
 | |
| 		for _, table := range ancient.sizes {
 | |
| 			stats = append(stats, []string{
 | |
| 				fmt.Sprintf("Ancient store (%s)", strings.Title(ancient.name)),
 | |
| 				strings.Title(table.name),
 | |
| 				table.size.String(),
 | |
| 				fmt.Sprintf("%d", ancient.count()),
 | |
| 			})
 | |
| 		}
 | |
| 		total += ancient.size()
 | |
| 	}
 | |
| 	table := tablewriter.NewWriter(os.Stdout)
 | |
| 	table.SetHeader([]string{"Database", "Category", "Size", "Items"})
 | |
| 	table.SetFooter([]string{"", "Total", total.String(), " "})
 | |
| 	table.AppendBulk(stats)
 | |
| 	table.Render()
 | |
| 
 | |
| 	if unaccounted.size > 0 {
 | |
| 		log.Error("Database contains unaccounted data", "size", unaccounted.size, "count", unaccounted.count)
 | |
| 	}
 | |
| 	return nil
 | |
| }
 |