lotus/blockstore/splitstore/splitstore.go

1577 lines
36 KiB
Go
Raw Normal View History

package splitstore
import (
2021-03-13 10:00:28 +00:00
"bytes"
"context"
2020-11-24 17:41:07 +00:00
"encoding/binary"
"errors"
"sort"
"sync"
2020-11-29 13:10:30 +00:00
"sync/atomic"
2020-11-25 08:11:42 +00:00
"time"
2021-03-03 08:56:41 +00:00
"go.uber.org/multierr"
"golang.org/x/xerrors"
blocks "github.com/ipfs/go-block-format"
cid "github.com/ipfs/go-cid"
2020-11-24 17:41:07 +00:00
dstore "github.com/ipfs/go-datastore"
logging "github.com/ipfs/go-log/v2"
2021-03-13 10:00:28 +00:00
cbg "github.com/whyrusleeping/cbor-gen"
"github.com/filecoin-project/go-state-types/abi"
2021-03-01 07:25:52 +00:00
bstore "github.com/filecoin-project/lotus/blockstore"
2020-11-24 17:15:38 +00:00
"github.com/filecoin-project/lotus/build"
"github.com/filecoin-project/lotus/chain/types"
2021-03-05 09:54:06 +00:00
"github.com/filecoin-project/lotus/metrics"
"github.com/filecoin-project/specs-actors/v2/actors/builtin"
2021-03-05 09:54:06 +00:00
"go.opencensus.io/stats"
)
2021-02-28 11:51:42 +00:00
var (
// CompactionThreshold is the number of epochs that need to have elapsed
// from the previously compacted epoch to trigger a new compaction.
//
// |················· CompactionThreshold ··················|
2021-07-04 07:06:55 +00:00
// | |
// =======‖≡≡≡≡≡≡≡≡≡≡≡≡≡≡≡≡≡≡≡≡‖------------------------»
// | | chain --> ↑__ current epoch
// | archived epochs ___↑
// ↑________ CompactionBoundary
//
// === :: cold (already archived)
// ≡≡≡ :: to be archived in this compaction
// --- :: hot
CompactionThreshold = 5 * build.Finality
// CompactionBoundary is the number of epochs from the current epoch at which
2021-03-13 10:00:28 +00:00
// we will walk the chain for live objects.
2021-06-21 11:50:03 +00:00
CompactionBoundary = 4 * build.Finality
2021-07-04 16:21:00 +00:00
// CompactionLookback is the number of epochs from the current epoch at which
// we will consider marking an old block reference.
CompactionLookback = 2 * build.Finality
// SyncGapTime is the time delay from a tipset's min timestamp before we decide
// there is a sync gap
SyncGapTime = time.Minute
)
2020-11-24 17:26:28 +00:00
var (
// baseEpochKey stores the base epoch (last compaction epoch) in the
// metadata store.
baseEpochKey = dstore.NewKey("/splitstore/baseEpoch")
// warmupEpochKey stores whether a hot store warmup has been performed.
// On first start, the splitstore will walk the state tree and will copy
// all active blocks into the hotstore.
warmupEpochKey = dstore.NewKey("/splitstore/warmupEpoch")
2021-03-02 08:04:02 +00:00
2021-03-05 08:00:17 +00:00
// markSetSizeKey stores the current estimate for the mark set size.
// this is first computed at warmup and updated in every compaction
markSetSizeKey = dstore.NewKey("/splitstore/markSetSize")
2021-03-02 08:04:02 +00:00
log = logging.Logger("splitstore")
2021-06-22 07:10:24 +00:00
// used to signal end of walk
errStopWalk = errors.New("stop walk")
2021-06-22 07:10:24 +00:00
// set this to true if you are debugging the splitstore to enable debug logging
enableDebugLog = false
// set this to true if you want to track origin stack traces in the write log
enableDebugLogWriteTraces = false
)
2020-11-24 17:41:07 +00:00
2021-03-02 16:59:00 +00:00
const (
batchSize = 16384
defaultColdPurgeSize = 7_000_000
)
2021-03-01 18:30:15 +00:00
type Config struct {
// MarkSetType is the type of mark set to use.
//
// Sane values are: "mapts", "bolt" (if you are memory constrained).
MarkSetType string
// DiscardColdBlocks indicates whether to skip moving cold blocks to the coldstore.
// If the splitstore is running with a noop coldstore then this option is set to true
// which skips moving (as it is a noop, but still takes time to read all the cold objects)
// and directly purges cold blocks.
2021-07-04 06:53:58 +00:00
DiscardColdBlocks bool
}
// ChainAccessor allows the Splitstore to access the chain. It will most likely
// be a ChainStore at runtime.
type ChainAccessor interface {
2021-03-19 10:17:32 +00:00
GetGenesis() (*types.BlockHeader, error)
GetTipsetByHeight(context.Context, abi.ChainEpoch, *types.TipSet, bool) (*types.TipSet, error)
GetHeaviestTipSet() *types.TipSet
SubscribeHeadChanges(change func(revert []*types.TipSet, apply []*types.TipSet) error)
}
type SplitStore struct {
compacting int32 // compaction (or warmp up) in progress
critsection int32 // compaction critical section
closing int32 // the split store is closing
2020-11-29 13:10:30 +00:00
cfg *Config
baseEpoch abi.ChainEpoch
warmupEpoch abi.ChainEpoch
writeEpoch abi.ChainEpoch
2020-11-29 13:10:30 +00:00
2021-03-02 16:59:00 +00:00
coldPurgeSize int
2020-11-29 13:10:30 +00:00
mx sync.Mutex
curTs *types.TipSet
2021-07-04 06:53:58 +00:00
chain ChainAccessor
ds dstore.Datastore
hot bstore.Blockstore
cold bstore.Blockstore
2020-11-24 17:26:28 +00:00
markSetEnv MarkSetEnv
markSetSize int64
ctx context.Context
cancel func()
2021-06-21 12:17:00 +00:00
debug *debugLog
// protection for concurrent read/writes during compaction
txnLk sync.RWMutex
txnActive bool
2021-07-04 16:21:00 +00:00
txnLookbackEpoch abi.ChainEpoch
txnProtect MarkSet
txnRefsMx sync.Mutex
txnRefs map[cid.Cid]struct{}
txnMissing map[cid.Cid]struct{}
}
var _ bstore.Blockstore = (*SplitStore)(nil)
2021-03-01 17:39:00 +00:00
// Open opens an existing splistore, or creates a new splitstore. The splitstore
// is backed by the provided hot and cold stores. The returned SplitStore MUST be
// attached to the ChainStore with Start in order to trigger compaction.
func Open(path string, ds dstore.Datastore, hot, cold bstore.Blockstore, cfg *Config) (*SplitStore, error) {
2021-07-04 06:53:58 +00:00
// hot blockstore must support BlockstoreIterator
if _, ok := hot.(bstore.BlockstoreIterator); !ok {
return nil, xerrors.Errorf("hot blockstore does not support efficient iteration: %T", hot)
2020-12-01 15:17:34 +00:00
}
// the markset env
markSetEnv, err := OpenMarkSetEnv(path, cfg.MarkSetType)
if err != nil {
return nil, err
2020-12-01 15:17:34 +00:00
}
// and now we can make a SplitStore
ss := &SplitStore{
cfg: cfg,
ds: ds,
hot: hot,
cold: cold,
markSetEnv: markSetEnv,
2021-03-02 16:59:00 +00:00
coldPurgeSize: defaultColdPurgeSize,
}
ss.ctx, ss.cancel = context.WithCancel(context.Background())
2021-06-22 07:10:24 +00:00
if enableDebugLog {
ss.debug, err = openDebugLog(path)
if err != nil {
return nil, err
}
}
2020-12-01 15:17:34 +00:00
return ss, nil
}
// Blockstore interface
2021-03-01 17:39:00 +00:00
func (s *SplitStore) DeleteBlock(_ cid.Cid) error {
// afaict we don't seem to be using this method, so it's not implemented
2020-11-24 22:01:10 +00:00
return errors.New("DeleteBlock not implemented on SplitStore; don't do this Luke!") //nolint
}
2021-03-02 14:45:45 +00:00
func (s *SplitStore) DeleteMany(_ []cid.Cid) error {
// afaict we don't seem to be using this method, so it's not implemented
return errors.New("DeleteMany not implemented on SplitStore; don't do this Luke!") //nolint
}
func (s *SplitStore) Has(c cid.Cid) (bool, error) {
s.txnLk.RLock()
defer s.txnLk.RUnlock()
has, err := s.hot.Has(c)
if err != nil {
return has, err
}
if has {
err = s.trackTxnRef(c)
return true, err
}
return s.cold.Has(c)
}
func (s *SplitStore) Get(cid cid.Cid) (blocks.Block, error) {
s.txnLk.RLock()
defer s.txnLk.RUnlock()
blk, err := s.hot.Get(cid)
switch err {
case nil:
err = s.trackTxnRef(cid)
return blk, err
case bstore.ErrNotFound:
2021-07-04 06:53:58 +00:00
if s.debug != nil {
s.mx.Lock()
warm := s.warmupEpoch > 0
curTs := s.curTs
s.mx.Unlock()
if warm {
s.debug.LogReadMiss(curTs, cid)
}
2021-06-22 07:56:35 +00:00
}
2021-06-21 12:17:00 +00:00
2021-03-05 09:54:06 +00:00
blk, err = s.cold.Get(cid)
if err == nil {
2021-03-05 09:54:06 +00:00
stats.Record(context.Background(), metrics.SplitstoreMiss.M(1))
2021-06-21 12:17:00 +00:00
2021-03-05 09:54:06 +00:00
}
return blk, err
default:
return nil, err
}
}
func (s *SplitStore) GetSize(cid cid.Cid) (int, error) {
s.txnLk.RLock()
defer s.txnLk.RUnlock()
size, err := s.hot.GetSize(cid)
switch err {
case nil:
err = s.trackTxnRef(cid)
return size, err
case bstore.ErrNotFound:
2021-07-04 06:53:58 +00:00
if s.debug != nil {
s.mx.Lock()
warm := s.warmupEpoch > 0
curTs := s.curTs
s.mx.Unlock()
if warm {
s.debug.LogReadMiss(curTs, cid)
}
2021-06-22 07:56:35 +00:00
}
2021-06-21 12:17:00 +00:00
2021-03-05 09:54:06 +00:00
size, err = s.cold.GetSize(cid)
if err == nil {
2021-03-05 09:54:06 +00:00
stats.Record(context.Background(), metrics.SplitstoreMiss.M(1))
}
return size, err
default:
return 0, err
}
}
func (s *SplitStore) Put(blk blocks.Block) error {
s.txnLk.RLock()
defer s.txnLk.RUnlock()
err := s.hot.Put(blk)
if err == nil {
2021-07-04 06:53:58 +00:00
if s.debug != nil {
s.mx.Lock()
curTs := s.curTs
writeEpoch := s.writeEpoch
s.mx.Unlock()
s.debug.LogWrite(curTs, blk, writeEpoch)
}
err = s.trackTxnRef(blk.Cid())
2021-06-28 12:21:51 +00:00
}
return err
}
func (s *SplitStore) PutMany(blks []blocks.Block) error {
batch := make([]cid.Cid, 0, len(blks))
for _, blk := range blks {
batch = append(batch, blk.Cid())
}
s.txnLk.RLock()
defer s.txnLk.RUnlock()
err := s.hot.PutMany(blks)
if err == nil {
2021-07-04 06:53:58 +00:00
if s.debug != nil {
s.mx.Lock()
curTs := s.curTs
writeEpoch := s.writeEpoch
s.mx.Unlock()
s.debug.LogWriteMany(curTs, blks, writeEpoch)
}
err = s.trackTxnRefMany(batch)
2021-06-28 12:21:51 +00:00
}
return err
}
func (s *SplitStore) AllKeysChan(ctx context.Context) (<-chan cid.Cid, error) {
ctx, cancel := context.WithCancel(ctx)
chHot, err := s.hot.AllKeysChan(ctx)
if err != nil {
2020-11-24 22:01:10 +00:00
cancel()
return nil, err
}
chCold, err := s.cold.AllKeysChan(ctx)
if err != nil {
cancel()
return nil, err
}
ch := make(chan cid.Cid)
go func() {
defer cancel()
defer close(ch)
for _, in := range []<-chan cid.Cid{chHot, chCold} {
for cid := range in {
select {
case ch <- cid:
case <-ctx.Done():
return
}
}
}
}()
return ch, nil
}
func (s *SplitStore) HashOnRead(enabled bool) {
s.hot.HashOnRead(enabled)
s.cold.HashOnRead(enabled)
}
func (s *SplitStore) View(cid cid.Cid, cb func([]byte) error) error {
s.txnLk.RLock()
defer s.txnLk.RUnlock()
err := s.hot.View(cid, cb)
switch err {
case nil:
err = s.trackTxnRef(cid)
return err
case bstore.ErrNotFound:
2021-07-04 06:53:58 +00:00
if s.debug != nil {
s.mx.Lock()
warm := s.warmupEpoch > 0
curTs := s.curTs
s.mx.Unlock()
if warm {
s.debug.LogReadMiss(curTs, cid)
}
2021-06-22 07:56:35 +00:00
}
2021-06-21 12:17:00 +00:00
2021-03-16 20:05:22 +00:00
err = s.cold.View(cid, cb)
if err == nil {
stats.Record(context.Background(), metrics.SplitstoreMiss.M(1))
}
return err
default:
return err
}
}
2020-11-24 17:15:38 +00:00
2020-11-24 17:26:28 +00:00
// State tracking
func (s *SplitStore) Start(chain ChainAccessor) error {
s.chain = chain
s.curTs = chain.GetHeaviestTipSet()
2020-11-24 17:41:07 +00:00
// load base epoch from metadata ds
// if none, then use current epoch because it's a fresh start
bs, err := s.ds.Get(baseEpochKey)
switch err {
case nil:
2020-11-26 18:37:02 +00:00
s.baseEpoch = bytesToEpoch(bs)
2020-11-24 17:41:07 +00:00
case dstore.ErrNotFound:
if s.curTs == nil {
// this can happen in some tests
break
}
2020-11-24 17:41:07 +00:00
err = s.setBaseEpoch(s.curTs.Height())
if err != nil {
return xerrors.Errorf("error saving base epoch: %w", err)
2020-11-24 17:41:07 +00:00
}
default:
return xerrors.Errorf("error loading base epoch: %w", err)
}
// load warmup epoch from metadata ds
// if none, then the splitstore will warm up the hotstore at first head change notif
// by walking the current tipset
bs, err = s.ds.Get(warmupEpochKey)
switch err {
case nil:
s.warmupEpoch = bytesToEpoch(bs)
case dstore.ErrNotFound:
// the hotstore hasn't warmed up, load the genesis into the hotstore
err = s.warmup(s.curTs)
if err != nil {
return xerrors.Errorf("error warming up: %w", err)
}
default:
return xerrors.Errorf("error loading warmup epoch: %w", err)
2020-11-24 17:41:07 +00:00
}
2021-03-05 08:00:17 +00:00
// load markSetSize from metadata ds
// if none, the splitstore will compute it during warmup and update in every compaction
bs, err = s.ds.Get(markSetSizeKey)
switch err {
case nil:
s.markSetSize = bytesToInt64(bs)
case dstore.ErrNotFound:
default:
return xerrors.Errorf("error loading mark set size: %w", err)
}
2021-07-04 06:53:58 +00:00
log.Infow("starting splitstore", "baseEpoch", s.baseEpoch, "warmupEpoch", s.warmupEpoch)
2021-07-04 06:53:58 +00:00
if s.debug != nil {
go s.background()
}
2021-03-02 17:07:54 +00:00
2020-11-24 17:41:07 +00:00
// watch the chain
chain.SubscribeHeadChanges(s.HeadChange)
2020-11-24 17:41:07 +00:00
return nil
2020-11-24 17:26:28 +00:00
}
2020-11-26 15:49:47 +00:00
func (s *SplitStore) Close() error {
atomic.StoreInt32(&s.closing, 1)
if atomic.LoadInt32(&s.critsection) == 1 {
log.Warn("ongoing compaction in critical section; waiting for it to finish...")
for atomic.LoadInt32(&s.critsection) == 1 {
2020-11-26 15:49:47 +00:00
time.Sleep(time.Second)
}
}
s.cancel()
2021-07-04 06:53:58 +00:00
return multierr.Combine(s.markSetEnv.Close(), s.debug.Close())
2020-11-26 15:49:47 +00:00
}
func (s *SplitStore) HeadChange(_, apply []*types.TipSet) error {
// Revert only.
if len(apply) == 0 {
return nil
}
s.mx.Lock()
curTs := apply[len(apply)-1]
epoch := curTs.Height()
s.curTs = curTs
s.mx.Unlock()
timestamp := time.Unix(int64(curTs.MinTimestamp()), 0)
if time.Since(timestamp) > SyncGapTime {
// don't attempt compaction before we have caught up syncing
return nil
}
2020-11-29 13:10:30 +00:00
if !atomic.CompareAndSwapInt32(&s.compacting, 0, 1) {
// we are currently compacting, do nothing and wait for the next head change
return nil
}
if epoch-s.baseEpoch > CompactionThreshold {
// it's time to compact
2020-11-24 17:26:28 +00:00
go func() {
2020-11-29 13:10:30 +00:00
defer atomic.StoreInt32(&s.compacting, 0)
2020-11-25 08:11:42 +00:00
log.Info("compacting splitstore")
start := time.Now()
s.compact(curTs)
2020-11-25 08:11:42 +00:00
log.Infow("compaction done", "took", time.Since(start))
2020-11-24 17:26:28 +00:00
}()
2020-11-29 13:10:30 +00:00
} else {
// no compaction necessary
atomic.StoreInt32(&s.compacting, 0)
2020-11-24 17:26:28 +00:00
}
return nil
}
2021-07-04 06:53:58 +00:00
func (s *SplitStore) background() {
ticker := time.NewTicker(time.Second)
defer ticker.Stop()
for {
select {
case <-s.ctx.Done():
return
case <-ticker.C:
s.updateWriteEpoch()
}
}
}
func (s *SplitStore) updateWriteEpoch() {
s.mx.Lock()
defer s.mx.Unlock()
curTs := s.curTs
timestamp := time.Unix(int64(curTs.MinTimestamp()), 0)
dt := time.Since(timestamp)
if dt < 0 {
writeEpoch := curTs.Height() + 1
if writeEpoch > s.writeEpoch {
s.writeEpoch = writeEpoch
}
return
}
writeEpoch := curTs.Height() + abi.ChainEpoch(dt.Seconds())/builtin.EpochDurationSeconds + 1
if writeEpoch > s.writeEpoch {
s.writeEpoch = writeEpoch
}
}
func (s *SplitStore) trackTxnRef(c cid.Cid) error {
2021-07-04 06:53:58 +00:00
if !s.txnActive {
// not compacting
return nil
}
2021-07-04 06:53:58 +00:00
if s.txnRefs != nil {
// we haven't finished marking yet, so track the reference
s.txnRefsMx.Lock()
2021-07-04 06:53:58 +00:00
s.txnRefs[c] = struct{}{}
s.txnRefsMx.Unlock()
2021-07-04 06:53:58 +00:00
return nil
}
// we have finished marking, protect the reference
return s.doTxnProtect(c, nil)
2021-07-04 16:21:00 +00:00
}
2021-07-04 16:33:49 +00:00
func (s *SplitStore) trackTxnRefMany(cids []cid.Cid) error {
if !s.txnActive {
// not compacting
return nil
}
if s.txnRefs != nil {
// we haven't finished marking yet, so track the reference
s.txnRefsMx.Lock()
for _, c := range cids {
s.txnRefs[c] = struct{}{}
}
s.txnRefsMx.Unlock()
return nil
}
// we have finished marking, protect the refs
batch := make(map[cid.Cid]struct{}, len(cids))
for _, c := range cids {
batch[c] = struct{}{}
}
for _, c := range cids {
err := s.doTxnProtect(c, batch)
if err != nil {
return err
}
}
return nil
}
func (s *SplitStore) doTxnProtect(root cid.Cid, batch map[cid.Cid]struct{}) error {
// Note: cold objects are deleted heaviest first, so the consituents of an object
// cannot be deleted before the object itself.
err := s.walkObjectIncomplete(root, cid.NewSet(),
2021-07-04 16:21:00 +00:00
func(c cid.Cid) error {
if isFilCommitment(c) {
return errStopWalk
}
2021-07-04 16:33:49 +00:00
if c != root {
_, ok := batch[c]
if ok {
// it's on the same batch, stop walk
return errStopWalk
}
}
mark, err := s.txnProtect.Has(c)
2021-07-04 16:21:00 +00:00
if err != nil {
return xerrors.Errorf("error checking mark set for %s: %w", c, err)
}
// it's marked, nothing to do
if mark {
return errStopWalk
}
// old block reference -- see comment in doCompact about the necessity of this
isOldBlock, err := s.isOldBlockHeader(c, s.txnLookbackEpoch)
if err != nil {
return xerrors.Errorf("error checking object type for %s: %w", c, err)
}
if isOldBlock {
return errStopWalk
}
return s.txnProtect.Mark(c)
},
func(c cid.Cid) error {
log.Warnf("missing object reference %s in %s", c, root)
if s.txnMissing != nil {
s.txnRefsMx.Lock()
s.txnMissing[c] = struct{}{}
s.txnRefsMx.Unlock()
}
return errStopWalk
2021-07-04 16:21:00 +00:00
})
if err != nil {
log.Warnf("error protecting object (cid: %s): %s", root, err)
}
return err
}
func (s *SplitStore) warmup(curTs *types.TipSet) error {
if !atomic.CompareAndSwapInt32(&s.compacting, 0, 1) {
return xerrors.Errorf("error locking compaction")
}
go func() {
defer atomic.StoreInt32(&s.compacting, 0)
log.Info("warming up hotstore")
start := time.Now()
2021-07-04 06:53:58 +00:00
err := s.doWarmup(curTs)
if err != nil {
log.Errorf("error warming up hotstore: %s", err)
return
}
log.Infow("warm up done", "took", time.Since(start))
}()
return nil
}
func (s *SplitStore) doWarmup(curTs *types.TipSet) error {
epoch := curTs.Height()
2021-03-01 18:11:35 +00:00
batchHot := make([]blocks.Block, 0, batchSize)
count := int64(0)
2021-03-13 10:00:28 +00:00
xcount := int64(0)
missing := int64(0)
2021-07-04 06:53:58 +00:00
err := s.walkChain(curTs, epoch, false,
func(c cid.Cid) error {
if isFilCommitment(c) {
return errStopWalk
}
count++
has, err := s.hot.Has(c)
if err != nil {
return err
}
if has {
return nil
}
blk, err := s.cold.Get(c)
if err != nil {
2021-03-13 10:00:28 +00:00
if err == bstore.ErrNotFound {
missing++
return nil
}
return err
}
2021-03-13 10:00:28 +00:00
xcount++
2021-03-01 18:11:35 +00:00
batchHot = append(batchHot, blk)
if len(batchHot) == batchSize {
err = s.hot.PutMany(batchHot)
if err != nil {
return err
}
batchHot = batchHot[:0]
}
2021-03-01 18:11:35 +00:00
return nil
})
if err != nil {
2021-03-13 10:00:28 +00:00
return err
}
2021-03-01 18:11:35 +00:00
if len(batchHot) > 0 {
err = s.hot.PutMany(batchHot)
if err != nil {
2021-03-13 10:00:28 +00:00
return err
2021-03-01 18:11:35 +00:00
}
}
log.Infow("warmup stats", "visited", count, "warm", xcount, "missing", missing)
2021-03-13 10:00:28 +00:00
s.markSetSize = count + count>>2 // overestimate a bit
2021-03-05 08:00:17 +00:00
err = s.ds.Put(markSetSizeKey, int64ToBytes(s.markSetSize))
if err != nil {
2021-03-13 10:00:28 +00:00
log.Warnf("error saving mark set size: %s", err)
2021-03-05 08:00:17 +00:00
}
2021-03-13 10:00:28 +00:00
// save the warmup epoch
err = s.ds.Put(warmupEpochKey, epochToBytes(epoch))
if err != nil {
return xerrors.Errorf("error saving warm up epoch: %w", err)
}
s.warmupEpoch = epoch
2021-03-13 10:00:28 +00:00
return nil
}
2020-11-24 17:15:38 +00:00
// Compaction/GC Algorithm
func (s *SplitStore) compact(curTs *types.TipSet) {
2021-03-05 09:54:06 +00:00
start := time.Now()
2021-07-04 06:53:58 +00:00
err := s.doCompact(curTs)
2021-03-05 09:54:06 +00:00
took := time.Since(start).Milliseconds()
stats.Record(context.Background(), metrics.SplitstoreCompactionTimeSeconds.M(float64(took)/1e3))
2021-03-05 08:29:49 +00:00
if err != nil {
log.Errorf("COMPACTION ERROR: %s", err)
}
}
func (s *SplitStore) doCompact(curTs *types.TipSet) error {
currentEpoch := curTs.Height()
boundaryEpoch := currentEpoch - CompactionBoundary
2021-07-04 16:21:00 +00:00
lookbackEpoch := currentEpoch - CompactionLookback
2021-07-04 16:21:00 +00:00
log.Infow("running compaction", "currentEpoch", currentEpoch, "baseEpoch", s.baseEpoch, "boundaryEpoch", boundaryEpoch, "lookbackEpoch", lookbackEpoch)
markSet, err := s.markSetEnv.Create("live", s.markSetSize)
if err != nil {
2021-03-05 08:29:49 +00:00
return xerrors.Errorf("error creating mark set: %w", err)
}
defer markSet.Close() //nolint:errcheck
defer s.debug.Flush()
2021-07-04 06:53:58 +00:00
// 0. Prepare the transaction
s.prepareTxnProtect(lookbackEpoch)
2021-07-04 06:53:58 +00:00
// 1. mark reachable objects by walking the chain from the current epoch to the boundary epoch
2021-07-05 08:41:09 +00:00
log.Info("marking reachable objects")
startMark := time.Now()
var count int64
2021-07-04 06:53:58 +00:00
err = s.walkChain(curTs, boundaryEpoch, true,
func(c cid.Cid) error {
if isFilCommitment(c) {
return errStopWalk
}
2021-02-28 19:35:18 +00:00
count++
return markSet.Mark(c)
})
if err != nil {
2021-07-05 08:41:09 +00:00
return xerrors.Errorf("error marking: %w", err)
}
s.markSetSize = count + count>>2 // overestimate a bit
2021-02-28 19:35:18 +00:00
2021-03-13 10:00:28 +00:00
log.Infow("marking done", "took", time.Since(startMark), "marked", count)
// begin transactional protection and fetch references created while marking
txnRefs := s.beginTxnProtect(markSet)
defer s.endTxnProtect()
2021-07-04 06:53:58 +00:00
// 1.1 Update markset for references created during marking
if len(txnRefs) > 0 {
2021-07-05 10:30:31 +00:00
log.Infow("updating mark set for live references", "refs", len(txnRefs))
startMark = time.Now()
walked := cid.NewSet()
count = 0
for c := range txnRefs {
if isFilCommitment(c) {
continue
}
mark, err := markSet.Has(c)
if err != nil {
2021-07-04 06:53:58 +00:00
return xerrors.Errorf("error checking markset for %s: %w", c, err)
}
if mark {
continue
}
err = s.walkObjectIncomplete(c, walked,
func(c cid.Cid) error {
if isFilCommitment(c) {
return errStopWalk
}
mark, err := markSet.Has(c)
if err != nil {
return xerrors.Errorf("error checking markset for %s: %w", c, err)
}
if mark {
return errStopWalk
}
// we also short-circuit on old blocks, as these can come from a network request
// and cause us to fail because we have purged its consistituents (or missing from
// the beginning in case of snapshot sync, e.g. parent message receipts or old messages)
// if these blocks are on our chain, they would have been marked but they might be
// from a fork.
//
// Ideally, we would have API options to preclude us from trcking references to such
// objects, but we don't so we have to do this check
2021-07-04 16:21:00 +00:00
isOldBlock, err := s.isOldBlockHeader(c, lookbackEpoch)
if err != nil {
return xerrors.Errorf("error checking object type for %s: %w", c, err)
}
if isOldBlock {
return errStopWalk
}
count++
return markSet.Mark(c)
},
func(cm cid.Cid) error {
log.Warnf("missing object reference %s in %s", cm, c) //nolint
s.txnRefsMx.Lock()
s.txnMissing[cm] = struct{}{}
s.txnRefsMx.Unlock()
return errStopWalk
})
if err != nil {
return xerrors.Errorf("error walking %s for marking: %w", c, err)
2021-07-04 09:14:29 +00:00
}
}
2021-07-04 09:14:29 +00:00
log.Infow("update mark set done", "took", time.Since(startMark), "marked", count)
}
2021-07-04 06:53:58 +00:00
// 2. iterate through the hotstore to collect cold objects
log.Info("collecting cold objects")
startCollect := time.Now()
// some stats for logging
var hotCnt, coldCnt int
2021-07-04 06:53:58 +00:00
cold := make([]cid.Cid, 0, s.coldPurgeSize)
err = s.hot.(bstore.BlockstoreIterator).ForEachKey(func(c cid.Cid) error {
// was it marked?
mark, err := markSet.Has(c)
if err != nil {
2021-07-04 06:53:58 +00:00
return xerrors.Errorf("error checkiing mark set for %s: %w", c, err)
}
if mark {
2021-07-04 06:53:58 +00:00
hotCnt++
return nil
}
2021-07-04 06:53:58 +00:00
// it's cold, mark it as candidate for move
cold = append(cold, c)
2021-07-04 06:53:58 +00:00
coldCnt++
return nil
})
if err != nil {
return xerrors.Errorf("error collecting candidate cold objects: %w", err)
}
log.Infow("cold collection done", "took", time.Since(startCollect))
2021-07-04 06:53:58 +00:00
if coldCnt > 0 {
s.coldPurgeSize = coldCnt + coldCnt>>2 // overestimate a bit
}
log.Infow("compaction stats", "hot", hotCnt, "cold", coldCnt)
2021-03-05 09:54:06 +00:00
stats.Record(context.Background(), metrics.SplitstoreCompactionHot.M(int64(hotCnt)))
stats.Record(context.Background(), metrics.SplitstoreCompactionCold.M(int64(coldCnt)))
// now that we have collected cold objects, check for missing references from transactional i/o
// and disable further collection of such references (they will not be acted upon as we can't
// possibly delete objects we didn't have when we were collecting cold objects)
s.waitForMissingRefs()
2021-07-04 06:53:58 +00:00
// 3. copy the cold objects to the coldstore -- if we have one
if !s.cfg.DiscardColdBlocks {
2021-07-05 08:41:09 +00:00
log.Info("moving cold objects to the coldstore")
startMove := time.Now()
err = s.moveColdBlocks(cold)
if err != nil {
2021-07-05 08:41:09 +00:00
return xerrors.Errorf("error moving cold objects: %w", err)
}
log.Infow("moving done", "took", time.Since(startMove))
2021-03-02 09:20:39 +00:00
}
2021-07-03 05:13:26 +00:00
// 4. sort cold objects so that the dags with most references are deleted first
// this ensures that we can't refer to a dag with its consituents already deleted
log.Info("sorting cold objects")
startSort := time.Now()
err = s.sortObjects(cold)
if err != nil {
return xerrors.Errorf("error sorting objects: %w", err)
}
log.Infow("sorting done", "took", time.Since(startSort))
// Enter critical section
log.Info("entering critical section")
atomic.StoreInt32(&s.critsection, 1)
defer atomic.StoreInt32(&s.critsection, 0)
// check to see if we are closing first; if that's the case just return
if atomic.LoadInt32(&s.closing) == 1 {
log.Info("splitstore is closing; aborting compaction")
return xerrors.Errorf("compaction aborted")
}
// 5. purge cold objects from the hotstore, taking protected references into account
2021-03-02 09:20:39 +00:00
log.Info("purging cold objects from the hotstore")
startPurge := time.Now()
err = s.purge(curTs, cold)
2021-03-02 09:20:39 +00:00
if err != nil {
2021-03-05 08:29:49 +00:00
return xerrors.Errorf("error purging cold blocks: %w", err)
2021-03-02 09:20:39 +00:00
}
2021-07-05 08:41:09 +00:00
log.Infow("purging cold objects from hotstore done", "took", time.Since(startPurge))
2021-03-02 09:20:39 +00:00
// we are done; do some housekeeping
s.endTxnProtect()
2021-03-08 17:46:21 +00:00
s.gcHotstore()
2021-07-04 07:06:55 +00:00
err = s.setBaseEpoch(boundaryEpoch)
2021-03-02 09:20:39 +00:00
if err != nil {
2021-03-05 08:29:49 +00:00
return xerrors.Errorf("error saving base epoch: %w", err)
2021-03-02 09:20:39 +00:00
}
2021-03-05 08:00:17 +00:00
err = s.ds.Put(markSetSizeKey, int64ToBytes(s.markSetSize))
if err != nil {
2021-03-05 08:29:49 +00:00
return xerrors.Errorf("error saving mark set size: %w", err)
2021-03-05 08:00:17 +00:00
}
2021-03-05 08:29:49 +00:00
return nil
2021-03-02 09:20:39 +00:00
}
func (s *SplitStore) prepareTxnProtect(lookbackEpoch abi.ChainEpoch) {
s.txnLk.Lock()
defer s.txnLk.Unlock()
s.txnRefs = make(map[cid.Cid]struct{})
s.txnActive = true
s.txnLookbackEpoch = lookbackEpoch
}
func (s *SplitStore) beginTxnProtect(markSet MarkSet) map[cid.Cid]struct{} {
s.txnLk.Lock()
defer s.txnLk.Unlock()
txnRefs := s.txnRefs
s.txnRefs = nil
s.txnMissing = make(map[cid.Cid]struct{})
s.txnProtect = markSet
return txnRefs
}
func (s *SplitStore) endTxnProtect() {
s.txnLk.Lock()
defer s.txnLk.Unlock()
2021-07-05 20:51:10 +00:00
if s.txnProtect != nil {
_ = s.txnProtect.Close()
}
s.txnActive = false
s.txnProtect = nil
s.txnRefs = nil
s.txnMissing = nil
}
2021-07-04 06:53:58 +00:00
func (s *SplitStore) walkChain(ts *types.TipSet, boundary abi.ChainEpoch, inclMsgs bool,
2021-07-01 11:06:28 +00:00
f func(cid.Cid) error) error {
visited := cid.NewSet()
2021-03-13 10:00:28 +00:00
walked := cid.NewSet()
toWalk := ts.Cids()
2021-07-01 11:10:57 +00:00
walkCnt := 0
scanCnt := 0
2021-03-13 10:00:28 +00:00
walkBlock := func(c cid.Cid) error {
if !visited.Visit(c) {
2021-03-13 10:00:28 +00:00
return nil
}
2021-07-01 11:10:57 +00:00
walkCnt++
2021-07-01 11:06:28 +00:00
if err := f(c); err != nil {
return err
}
2021-03-13 10:00:28 +00:00
var hdr types.BlockHeader
err := s.view(c, func(data []byte) error {
return hdr.UnmarshalCBOR(bytes.NewBuffer(data))
})
if err != nil {
2021-03-13 10:00:28 +00:00
return xerrors.Errorf("error unmarshaling block header (cid: %s): %w", c, err)
}
2021-07-04 06:53:58 +00:00
// we only scan the block if it is at or above the boundary
if hdr.Height >= boundary || hdr.Height == 0 {
2021-07-01 11:10:57 +00:00
scanCnt++
if inclMsgs && hdr.Height > 0 {
2021-07-04 06:53:58 +00:00
if err := s.walkObject(hdr.Messages, walked, f); err != nil {
return xerrors.Errorf("error walking messages (cid: %s): %w", hdr.Messages, err)
}
2021-07-04 06:53:58 +00:00
if err := s.walkObject(hdr.ParentMessageReceipts, walked, f); err != nil {
return xerrors.Errorf("error walking message receipts (cid: %s): %w", hdr.ParentMessageReceipts, err)
}
}
2021-03-13 10:00:28 +00:00
2021-07-04 06:53:58 +00:00
if err := s.walkObject(hdr.ParentStateRoot, walked, f); err != nil {
return xerrors.Errorf("error walking state root (cid: %s): %w", hdr.ParentStateRoot, err)
}
2021-03-13 10:00:28 +00:00
}
if hdr.Height > 0 {
toWalk = append(toWalk, hdr.Parents...)
}
2021-07-01 11:06:28 +00:00
2021-03-13 10:00:28 +00:00
return nil
}
for len(toWalk) > 0 {
walking := toWalk
toWalk = nil
for _, c := range walking {
if err := walkBlock(c); err != nil {
return xerrors.Errorf("error walking block (cid: %s): %w", c, err)
}
}
}
2021-07-01 11:10:57 +00:00
log.Infow("chain walk done", "walked", walkCnt, "scanned", scanCnt)
2021-07-01 11:06:28 +00:00
2021-03-13 10:00:28 +00:00
return nil
}
func (s *SplitStore) walkObject(c cid.Cid, walked *cid.Set, f func(cid.Cid) error) error {
if !walked.Visit(c) {
return nil
}
if err := f(c); err != nil {
if err == errStopWalk {
return nil
}
return err
}
if c.Prefix().Codec != cid.DagCBOR {
return nil
}
var links []cid.Cid
err := s.view(c, func(data []byte) error {
return cbg.ScanForLinks(bytes.NewReader(data), func(c cid.Cid) {
links = append(links, c)
})
})
if err != nil {
return xerrors.Errorf("error scanning linked block (cid: %s): %w", c, err)
}
for _, c := range links {
err := s.walkObject(c, walked, f)
if err != nil {
return xerrors.Errorf("error walking link (cid: %s): %w", c, err)
}
}
return nil
}
// like walkObject, but the object may be potentially incomplete (references missing from the hotstore)
func (s *SplitStore) walkObjectIncomplete(c cid.Cid, walked *cid.Set, f, missing func(cid.Cid) error) error {
2021-03-13 10:00:28 +00:00
if !walked.Visit(c) {
return nil
}
2021-07-05 08:11:08 +00:00
// occurs check -- only for DAGs
if c.Prefix().Codec == cid.DagCBOR {
has, err := s.has(c)
2021-07-05 08:11:08 +00:00
if err != nil {
return xerrors.Errorf("error occur checking %s: %w", c, err)
}
2021-07-05 08:11:08 +00:00
if !has {
err = missing(c)
if err == errStopWalk {
return nil
}
return err
}
}
if err := f(c); err != nil {
if err == errStopWalk {
return nil
}
2021-03-13 10:00:28 +00:00
return err
}
if c.Prefix().Codec != cid.DagCBOR {
return nil
}
var links []cid.Cid
2021-07-05 08:11:08 +00:00
err := s.view(c, func(data []byte) error {
return cbg.ScanForLinks(bytes.NewReader(data), func(c cid.Cid) {
links = append(links, c)
})
})
2021-03-13 10:00:28 +00:00
if err != nil {
return xerrors.Errorf("error scanning linked block (cid: %s): %w", c, err)
2021-03-13 10:00:28 +00:00
}
for _, c := range links {
err := s.walkObjectIncomplete(c, walked, f, missing)
2021-07-01 11:06:28 +00:00
if err != nil {
return xerrors.Errorf("error walking link (cid: %s): %w", c, err)
2021-07-01 11:06:28 +00:00
}
2021-03-13 10:00:28 +00:00
}
return nil
2021-03-13 10:00:28 +00:00
}
// internal version used by walk
func (s *SplitStore) view(cid cid.Cid, cb func([]byte) error) error {
err := s.hot.View(cid, cb)
switch err {
case bstore.ErrNotFound:
return s.cold.View(cid, cb)
default:
return err
}
}
func (s *SplitStore) has(c cid.Cid) (bool, error) {
has, err := s.hot.Has(c)
if has || err != nil {
return has, err
}
return s.cold.Has(c)
}
func (s *SplitStore) isOldBlockHeader(c cid.Cid, epoch abi.ChainEpoch) (isOldBlock bool, err error) {
if c.Prefix().Codec != cid.DagCBOR {
return false, nil
}
err = s.view(c, func(data []byte) error {
var hdr types.BlockHeader
if hdr.UnmarshalCBOR(bytes.NewBuffer(data)) == nil {
isOldBlock = hdr.Height < epoch
}
return nil
})
return isOldBlock, err
}
func (s *SplitStore) isBlockHeader(c cid.Cid) (isBlock bool, err error) {
if c.Prefix().Codec != cid.DagCBOR {
return false, nil
}
err = s.view(c, func(data []byte) error {
var hdr types.BlockHeader
isBlock = hdr.UnmarshalCBOR(bytes.NewBuffer(data)) == nil
return nil
})
return isBlock, err
}
2021-03-02 16:59:00 +00:00
func (s *SplitStore) moveColdBlocks(cold []cid.Cid) error {
batch := make([]blocks.Block, 0, batchSize)
2021-07-04 06:53:58 +00:00
for _, c := range cold {
blk, err := s.hot.Get(c)
if err != nil {
2021-06-28 10:35:06 +00:00
if err == bstore.ErrNotFound {
2021-07-04 06:53:58 +00:00
log.Warnf("hotstore missing block %s", c)
continue
}
2021-07-04 06:53:58 +00:00
return xerrors.Errorf("error retrieving block %s from hotstore: %w", c, err)
}
batch = append(batch, blk)
if len(batch) == batchSize {
err = s.cold.PutMany(batch)
if err != nil {
2021-03-02 09:20:39 +00:00
return xerrors.Errorf("error putting batch to coldstore: %w", err)
}
batch = batch[:0]
}
}
if len(batch) > 0 {
2021-03-02 09:20:39 +00:00
err := s.cold.PutMany(batch)
if err != nil {
2021-03-02 09:20:39 +00:00
return xerrors.Errorf("error putting cold to coldstore: %w", err)
}
}
2021-03-02 09:20:39 +00:00
return nil
}
func (s *SplitStore) sortObjects(cids []cid.Cid) error {
// we cache the keys to avoid making a gazillion of strings
keys := make(map[cid.Cid]string)
key := func(c cid.Cid) string {
s, ok := keys[c]
if !ok {
s = string(c.Hash())
keys[c] = s
}
return s
}
// compute sorting weights as the cumulative number of DAG links
weights := make(map[string]int)
for _, c := range cids {
w := s.getObjectWeight(c, weights, key)
weights[key(c)] = w
}
// sort!
sort.Slice(cids, func(i, j int) bool {
wi := weights[key(cids[i])]
wj := weights[key(cids[j])]
if wi == wj {
return bytes.Compare(cids[i].Hash(), cids[j].Hash()) > 0
}
return wi > wj
})
return nil
}
func (s *SplitStore) getObjectWeight(c cid.Cid, weights map[string]int, key func(cid.Cid) string) int {
w, ok := weights[key(c)]
if ok {
return w
}
// we treat block headers specially to avoid walking the entire chain
var hdr types.BlockHeader
err := s.view(c, func(data []byte) error {
return hdr.UnmarshalCBOR(bytes.NewBuffer(data))
})
if err == nil {
w1 := s.getObjectWeight(hdr.ParentStateRoot, weights, key)
weights[key(hdr.ParentStateRoot)] = w1
w2 := s.getObjectWeight(hdr.Messages, weights, key)
weights[key(hdr.Messages)] = w2
return 1 + w1 + w2
}
var links []cid.Cid
err = s.view(c, func(data []byte) error {
return cbg.ScanForLinks(bytes.NewReader(data), func(c cid.Cid) {
links = append(links, c)
})
})
if err != nil {
return 1
}
for _, c := range links {
// these are internal refs, so dags will be dags
if c.Prefix().Codec != cid.DagCBOR {
w++
continue
}
wc := s.getObjectWeight(c, weights, key)
weights[key(c)] = wc
w += wc
}
return w
}
2021-03-03 18:04:29 +00:00
func (s *SplitStore) purgeBatch(cids []cid.Cid, deleteBatch func([]cid.Cid) error) error {
if len(cids) == 0 {
return nil
}
// we don't delete one giant batch of millions of objects, but rather do smaller batches
// so that we don't stop the world for an extended period of time
done := false
2021-03-03 17:36:13 +00:00
for i := 0; !done; i++ {
start := i * batchSize
end := start + batchSize
if end >= len(cids) {
end = len(cids)
done = true
}
2021-03-03 18:04:29 +00:00
err := deleteBatch(cids[start:end])
if err != nil {
2021-03-03 18:04:29 +00:00
return xerrors.Errorf("error deleting batch: %w", err)
}
}
2021-03-02 09:20:39 +00:00
return nil
}
func (s *SplitStore) purge(curTs *types.TipSet, cids []cid.Cid) error {
deadCids := make([]cid.Cid, 0, batchSize)
var purgeCnt, liveCnt int
2021-06-30 17:35:00 +00:00
defer func() {
2021-07-03 05:13:26 +00:00
log.Infow("purged cold objects", "purged", purgeCnt, "live", liveCnt)
2021-06-30 17:35:00 +00:00
}()
return s.purgeBatch(cids,
func(cids []cid.Cid) error {
deadCids := deadCids[:0]
s.txnLk.Lock()
defer s.txnLk.Unlock()
for _, c := range cids {
live, err := s.txnProtect.Has(c)
if err != nil {
return xerrors.Errorf("error checking for liveness: %w", err)
}
if live {
liveCnt++
continue
}
deadCids = append(deadCids, c)
s.debug.LogMove(curTs, c)
}
2021-07-04 06:53:58 +00:00
err := s.hot.DeleteMany(deadCids)
if err != nil {
return xerrors.Errorf("error purging cold objects: %w", err)
}
purgeCnt += len(deadCids)
return nil
})
}
// I really don't like having this code, but we seem to have some DAG references with missing
// constituents. During testing in mainnet *some* of these references *sometimes* appeared after a
// little bit.
// We need to figure out where they are coming from and eliminate that vector, but until then we
// have this gem[TM].
// My best guess is that they are parent message receipts or yet to be computed state roots.
func (s *SplitStore) waitForMissingRefs() {
s.txnLk.Lock()
missing := s.txnMissing
s.txnMissing = nil
s.txnLk.Unlock()
if len(missing) == 0 {
return
}
log.Info("waiting for missing references")
start := time.Now()
count := 0
defer func() {
log.Infow("waiting for missing references done", "took", time.Since(start), "marked", count)
}()
for i := 0; i < 3 && len(missing) > 0; i++ {
wait := time.Duration(i) * time.Minute
log.Infof("retrying for %d missing references in %s (attempt: %d)", len(missing), wait, i+1)
if wait > 0 {
time.Sleep(wait)
}
towalk := missing
walked := cid.NewSet()
missing = make(map[cid.Cid]struct{})
for c := range towalk {
err := s.walkObjectIncomplete(c, walked,
func(c cid.Cid) error {
if isFilCommitment(c) {
return errStopWalk
}
mark, err := s.txnProtect.Has(c)
if err != nil {
return xerrors.Errorf("error checking markset for %s: %w", c, err)
}
if mark {
return errStopWalk
}
isOldBlock, err := s.isOldBlockHeader(c, s.txnLookbackEpoch)
if err != nil {
return xerrors.Errorf("error checking object type for %s: %w", c, err)
}
if isOldBlock {
return errStopWalk
}
count++
return s.txnProtect.Mark(c)
},
func(c cid.Cid) error {
missing[c] = struct{}{}
return errStopWalk
})
if err != nil {
log.Warnf("error marking: %s", err)
}
}
}
if len(missing) > 0 {
log.Warnf("still missing %d references", len(missing))
for c := range missing {
log.Warnf("unresolved missing reference: %s", c)
}
}
}
2021-03-08 17:46:21 +00:00
func (s *SplitStore) gcHotstore() {
if compact, ok := s.hot.(interface{ Compact() error }); ok {
log.Infof("compacting hotstore")
startCompact := time.Now()
err := compact.Compact()
if err != nil {
log.Warnf("error compacting hotstore: %s", err)
return
}
2021-03-11 11:10:44 +00:00
log.Infow("hotstore compaction done", "took", time.Since(startCompact))
}
2021-03-08 17:46:21 +00:00
if gc, ok := s.hot.(interface{ CollectGarbage() error }); ok {
log.Infof("garbage collecting hotstore")
startGC := time.Now()
err := gc.CollectGarbage()
if err != nil {
log.Warnf("error garbage collecting hotstore: %s", err)
2021-03-11 11:10:44 +00:00
return
2021-03-08 17:46:21 +00:00
}
2021-03-11 11:10:44 +00:00
log.Infow("hotstore garbage collection done", "took", time.Since(startGC))
2021-03-08 17:46:21 +00:00
}
}
2020-11-24 17:41:07 +00:00
func (s *SplitStore) setBaseEpoch(epoch abi.ChainEpoch) error {
s.baseEpoch = epoch
2020-11-26 18:37:02 +00:00
return s.ds.Put(baseEpochKey, epochToBytes(epoch))
}
func epochToBytes(epoch abi.ChainEpoch) []byte {
2021-03-05 08:00:17 +00:00
return uint64ToBytes(uint64(epoch))
}
func bytesToEpoch(buf []byte) abi.ChainEpoch {
return abi.ChainEpoch(bytesToUint64(buf))
}
func int64ToBytes(i int64) []byte {
return uint64ToBytes(uint64(i))
}
func bytesToInt64(buf []byte) int64 {
return int64(bytesToUint64(buf))
}
func uint64ToBytes(i uint64) []byte {
2020-11-26 18:37:02 +00:00
buf := make([]byte, 16)
2021-03-05 08:00:17 +00:00
n := binary.PutUvarint(buf, i)
2020-11-26 18:37:02 +00:00
return buf[:n]
}
2021-03-05 08:00:17 +00:00
func bytesToUint64(buf []byte) uint64 {
i, _ := binary.Uvarint(buf)
return i
2020-11-24 17:15:38 +00:00
}
func isFilCommitment(c cid.Cid) bool {
switch c.Prefix().Codec {
case cid.FilCommitmentSealed, cid.FilCommitmentUnsealed:
return true
default:
return false
}
}