// Copyright © 2020 Vulcanize, Inc // // This program is free software: you can redistribute it and/or modify // it under the terms of the GNU Affero General Public License as published by // the Free Software Foundation, either version 3 of the License, or // (at your option) any later version. // // This program is distributed in the hope that it will be useful, // but WITHOUT ANY WARRANTY; without even the implied warranty of // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the // GNU Affero General Public License for more details. // // You should have received a copy of the GNU Affero General Public License // along with this program. If not, see . package snapshot import ( "context" "fmt" "math/big" "sync" "github.com/cerc-io/eth-iterator-utils/tracker" statediff "github.com/cerc-io/plugeth-statediff" "github.com/cerc-io/plugeth-statediff/adapt" "github.com/cerc-io/plugeth-statediff/indexer" "github.com/cerc-io/plugeth-statediff/types" "github.com/ethereum/go-ethereum/common" "github.com/ethereum/go-ethereum/core/rawdb" "github.com/ethereum/go-ethereum/core/state" "github.com/ethereum/go-ethereum/crypto" "github.com/ethereum/go-ethereum/ethdb" "github.com/ethereum/go-ethereum/rlp" log "github.com/sirupsen/logrus" ) var ( emptyNode, _ = rlp.EncodeToBytes(&[]byte{}) emptyCodeHash = crypto.Keccak256([]byte{}) emptyContractRoot = crypto.Keccak256Hash(emptyNode) defaultBatchSize = uint(100) ) // Service holds ethDB and stateDB to read data from lvldb and Publisher // to publish trie in postgres DB. type Service struct { ethDB ethdb.Database stateDB state.Database // ipfsPublisher Publisher indexer indexer.Indexer maxBatchSize uint // tracker *tracker.Tracker recoveryFile string } func NewLevelDB(con *EthConfig) (ethdb.Database, error) { kvdb, _ := rawdb.NewLevelDBDatabase(con.LevelDBPath, 1024, 256, "ipld-eth-state-snapshot", true) edb, err := rawdb.NewDatabaseWithFreezer(kvdb, con.AncientDBPath, "ipld-eth-state-snapshot", true) if err != nil { return nil, fmt.Errorf("unable to create NewLevelDBDatabaseWithFreezer: %s", err) } return edb, nil } // NewSnapshotService creates Service. func NewSnapshotService(edb ethdb.Database, indexer indexer.Indexer, recoveryFile string) (*Service, error) { return &Service{ ethDB: edb, stateDB: state.NewDatabase(edb), // ipfsPublisher: pub, indexer: indexer, maxBatchSize: defaultBatchSize, recoveryFile: recoveryFile, }, nil } type SnapshotParams struct { WatchedAddresses []common.Address Height uint64 Workers uint } func (s *Service) CreateSnapshot(params SnapshotParams) error { // extract header from lvldb and publish to PG-IPFS // hold onto the headerID so that we can link the state nodes to this header log.Infof("Creating snapshot at height %d", params.Height) hash := rawdb.ReadCanonicalHash(s.ethDB, params.Height) header := rawdb.ReadHeader(s.ethDB, hash, params.Height) if header == nil { return fmt.Errorf("unable to read canonical header at height %d", params.Height) } log.Infof("head hash: %s head height: %d", hash.Hex(), params.Height) // Context for snapshot work ctx, cancelCtx := context.WithCancel(context.Background()) defer cancelCtx() var err error tx := s.indexer.BeginTx(header.Number, ctx) defer tx.RollbackOnFailure(err) var headerid string headerid, err = s.indexer.PushHeader(tx, header, big.NewInt(0), big.NewInt(0)) if err != nil { return err } tracker := tracker.New(s.recoveryFile, params.Workers) tracker.CaptureSignal(cancelCtx) defer func() { err := tracker.HaltAndDump() if err != nil { log.Errorf("failed to write recovery file: %v", err) } }() var nodeMtx, ipldMtx sync.Mutex nodeSink := func(node types.StateLeafNode) error { nodeMtx.Lock() defer nodeMtx.Unlock() return s.indexer.PushStateNode(tx, node, headerid) } ipldSink := func(c types.IPLD) error { ipldMtx.Lock() defer ipldMtx.Unlock() return s.indexer.PushIPLD(tx, c) } // Build a diff compared against the zero hash to get a full snapshot sdargs := statediff.Args{ NewStateRoot: header.Root, BlockHash: header.Hash(), BlockNumber: header.Number, } sdparams := statediff.Params{ WatchedAddresses: params.WatchedAddresses, } sdparams.ComputeWatchedAddressesLeafPaths() builder := statediff.NewBuilder(adapt.GethStateView(s.stateDB)) builder.SetSubtrieWorkers(params.Workers) if err = builder.WriteStateDiffTracked(sdargs, sdparams, nodeSink, ipldSink, &tracker); err != nil { return err } if err = tx.Submit(); err != nil { return fmt.Errorf("batch transaction submission failed: %w", err) } return err } // CreateLatestSnapshot snapshot at head (ignores height param) func (s *Service) CreateLatestSnapshot(workers uint, watchedAddresses []common.Address) error { log.Info("Creating snapshot at head") hash := rawdb.ReadHeadHeaderHash(s.ethDB) height := rawdb.ReadHeaderNumber(s.ethDB, hash) if height == nil { return fmt.Errorf("unable to read header height for header hash %s", hash.String()) } return s.CreateSnapshot(SnapshotParams{Height: *height, Workers: workers, WatchedAddresses: watchedAddresses}) }