241 lines
7.3 KiB
Go
241 lines
7.3 KiB
Go
package dagstore
|
|
|
|
import (
|
|
"context"
|
|
"errors"
|
|
"io"
|
|
"sync"
|
|
"time"
|
|
|
|
"github.com/ipfs/go-cid"
|
|
ds "github.com/ipfs/go-datastore"
|
|
bstore "github.com/ipfs/go-ipfs-blockstore"
|
|
logging "github.com/ipfs/go-log/v2"
|
|
"golang.org/x/xerrors"
|
|
|
|
"github.com/filecoin-project/dagstore"
|
|
"github.com/filecoin-project/dagstore/mount"
|
|
"github.com/filecoin-project/dagstore/shard"
|
|
"github.com/filecoin-project/go-fil-markets/carstore"
|
|
"github.com/filecoin-project/go-fil-markets/shared"
|
|
)
|
|
|
|
var log = logging.Logger("dagstore-wrapper")
|
|
|
|
// MarketDAGStoreConfig is the config the market needs to then construct a DAG Store.
|
|
type MarketDAGStoreConfig struct {
|
|
TransientsDir string
|
|
IndexDir string
|
|
Datastore ds.Datastore
|
|
MaxConcurrentFetch int
|
|
MaxConcurrentIndex int
|
|
GCInterval time.Duration
|
|
}
|
|
|
|
// DAGStore provides an interface for the DAG store that can be mocked out
|
|
// by tests
|
|
type DAGStore interface {
|
|
RegisterShard(ctx context.Context, key shard.Key, mnt mount.Mount, out chan dagstore.ShardResult, opts dagstore.RegisterOpts) error
|
|
AcquireShard(ctx context.Context, key shard.Key, out chan dagstore.ShardResult, _ dagstore.AcquireOpts) error
|
|
RecoverShard(ctx context.Context, key shard.Key, out chan dagstore.ShardResult, _ dagstore.RecoverOpts) error
|
|
GC(ctx context.Context) (map[shard.Key]error, error)
|
|
Close() error
|
|
}
|
|
|
|
type closableBlockstore struct {
|
|
bstore.Blockstore
|
|
io.Closer
|
|
}
|
|
|
|
type Wrapper struct {
|
|
ctx context.Context
|
|
cancel context.CancelFunc
|
|
backgroundWg sync.WaitGroup
|
|
|
|
dagStore DAGStore
|
|
mountApi LotusAccessor
|
|
failureCh chan dagstore.ShardResult
|
|
traceCh chan dagstore.Trace
|
|
gcInterval time.Duration
|
|
}
|
|
|
|
var _ shared.DagStoreWrapper = (*Wrapper)(nil)
|
|
|
|
func NewDagStoreWrapper(cfg MarketDAGStoreConfig, mountApi LotusAccessor) (*Wrapper, error) {
|
|
// construct the DAG Store.
|
|
registry := mount.NewRegistry()
|
|
if err := registry.Register(lotusScheme, NewLotusMountTemplate(mountApi)); err != nil {
|
|
return nil, xerrors.Errorf("failed to create registry: %w", err)
|
|
}
|
|
|
|
// The dagstore will write Shard failures to the `failureCh` here.
|
|
failureCh := make(chan dagstore.ShardResult, 1)
|
|
// The dagstore will write Trace events to the `traceCh` here.
|
|
traceCh := make(chan dagstore.Trace, 32)
|
|
dcfg := dagstore.Config{
|
|
TransientsDir: cfg.TransientsDir,
|
|
IndexDir: cfg.IndexDir,
|
|
Datastore: cfg.Datastore,
|
|
MountRegistry: registry,
|
|
FailureCh: failureCh,
|
|
TraceCh: traceCh,
|
|
MaxConcurrentFetch: cfg.MaxConcurrentFetch,
|
|
MaxConcurrentIndex: cfg.MaxConcurrentIndex,
|
|
}
|
|
dagStore, err := dagstore.NewDAGStore(dcfg)
|
|
if err != nil {
|
|
return nil, xerrors.Errorf("failed to create DAG store: %w", err)
|
|
}
|
|
|
|
return &Wrapper{
|
|
dagStore: dagStore,
|
|
mountApi: mountApi,
|
|
failureCh: failureCh,
|
|
traceCh: traceCh,
|
|
gcInterval: cfg.GCInterval,
|
|
}, nil
|
|
}
|
|
|
|
func (ds *Wrapper) Start(ctx context.Context) error {
|
|
ds.ctx, ds.cancel = context.WithCancel(ctx)
|
|
|
|
err := ds.mountApi.Start(ctx)
|
|
if err != nil {
|
|
return xerrors.Errorf("failed to start mount API: %w", err)
|
|
}
|
|
|
|
ds.backgroundWg.Add(1)
|
|
|
|
// Run a go-routine to handle failures, traces and GC
|
|
go ds.background()
|
|
|
|
return nil
|
|
}
|
|
|
|
func (ds *Wrapper) background() {
|
|
defer ds.backgroundWg.Done()
|
|
|
|
gcTicker := time.NewTicker(ds.gcInterval)
|
|
defer gcTicker.Stop()
|
|
|
|
recoverShardResults := make(chan dagstore.ShardResult, 32)
|
|
for ds.ctx.Err() == nil {
|
|
select {
|
|
|
|
// GC the DAG store on every tick
|
|
case <-gcTicker.C:
|
|
_, _ = ds.dagStore.GC(ds.ctx)
|
|
|
|
// Log trace events from the DAG store
|
|
case tr := <-ds.traceCh:
|
|
log.Debugw("trace",
|
|
"shard-key", tr.Key.String(),
|
|
"op-type", tr.Op.String(),
|
|
"after", tr.After.String())
|
|
|
|
// Handle shard failures by attempting to recover the shard
|
|
case f := <-ds.failureCh:
|
|
log.Warnw("shard failed", "shard-key", f.Key.String(), "error", f.Error)
|
|
if err := ds.dagStore.RecoverShard(ds.ctx, f.Key, recoverShardResults, dagstore.RecoverOpts{}); err != nil {
|
|
log.Warnw("shard recovery failed", "shard-key", f.Key.String(), "error", err)
|
|
}
|
|
|
|
// Consume recover shard results
|
|
case res := <-recoverShardResults:
|
|
if res.Error != nil {
|
|
log.Warnw("shard recovery failed", "shard-key", res.Key.String(), "error", res.Error)
|
|
}
|
|
|
|
// Exit when the DAG store wrapper is shutdown
|
|
case <-ds.ctx.Done():
|
|
return
|
|
}
|
|
}
|
|
}
|
|
|
|
func (ds *Wrapper) LoadShard(ctx context.Context, pieceCid cid.Cid) (carstore.ClosableBlockstore, error) {
|
|
key := shard.KeyFromCID(pieceCid)
|
|
resch := make(chan dagstore.ShardResult, 1)
|
|
err := ds.dagStore.AcquireShard(ctx, key, resch, dagstore.AcquireOpts{})
|
|
|
|
if err != nil {
|
|
if !errors.Is(err, dagstore.ErrShardUnknown) {
|
|
return nil, xerrors.Errorf("failed to schedule acquire shard for piece CID %s: %w", pieceCid, err)
|
|
}
|
|
|
|
// if the DAGStore does not know about the Shard -> register it and then try to acquire it again.
|
|
log.Warnw("failed to load shard as shard is not registered, will re-register", "pieceCID", pieceCid)
|
|
// The path of a transient file that we can ask the DAG Store to use
|
|
// to perform the Indexing rather than fetching it via the Mount if
|
|
// we already have a transient file. However, we don't have it here
|
|
// and therefore we pass an empty file path.
|
|
carPath := ""
|
|
if err := shared.RegisterShardSync(ctx, ds, pieceCid, carPath, false); err != nil {
|
|
return nil, xerrors.Errorf("failed to re-register shard during loading piece CID %s: %w", pieceCid, err)
|
|
}
|
|
log.Warnw("successfully re-registered shard", "pieceCID", pieceCid)
|
|
|
|
resch = make(chan dagstore.ShardResult, 1)
|
|
if err := ds.dagStore.AcquireShard(ctx, key, resch, dagstore.AcquireOpts{}); err != nil {
|
|
return nil, xerrors.Errorf("failed to acquire Shard for piece CID %s after re-registering: %w", pieceCid, err)
|
|
}
|
|
}
|
|
|
|
// TODO: The context is not yet being actively monitored by the DAG store,
|
|
// so we need to select against ctx.Done() until the following issue is
|
|
// implemented:
|
|
// https://github.com/filecoin-project/dagstore/issues/39
|
|
var res dagstore.ShardResult
|
|
select {
|
|
case <-ctx.Done():
|
|
return nil, ctx.Err()
|
|
case res = <-resch:
|
|
if res.Error != nil {
|
|
return nil, xerrors.Errorf("failed to acquire shard for piece CID %s: %w", pieceCid, res.Error)
|
|
}
|
|
}
|
|
|
|
bs, err := res.Accessor.Blockstore()
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
return &closableBlockstore{Blockstore: NewReadOnlyBlockstore(bs), Closer: res.Accessor}, nil
|
|
}
|
|
|
|
func (ds *Wrapper) RegisterShard(ctx context.Context, pieceCid cid.Cid, carPath string, eagerInit bool, resch chan dagstore.ShardResult) error {
|
|
// Create a lotus mount with the piece CID
|
|
key := shard.KeyFromCID(pieceCid)
|
|
mt, err := NewLotusMount(pieceCid, ds.mountApi)
|
|
if err != nil {
|
|
return xerrors.Errorf("failed to create lotus mount for piece CID %s: %w", pieceCid, err)
|
|
}
|
|
|
|
// Register the shard
|
|
opts := dagstore.RegisterOpts{
|
|
ExistingTransient: carPath,
|
|
LazyInitialization: !eagerInit,
|
|
}
|
|
err = ds.dagStore.RegisterShard(ctx, key, mt, resch, opts)
|
|
if err != nil {
|
|
return xerrors.Errorf("failed to schedule register shard for piece CID %s: %w", pieceCid, err)
|
|
}
|
|
|
|
return nil
|
|
}
|
|
|
|
func (ds *Wrapper) Close() error {
|
|
// Cancel the context
|
|
ds.cancel()
|
|
|
|
// Close the DAG store
|
|
if err := ds.dagStore.Close(); err != nil {
|
|
return xerrors.Errorf("failed to close DAG store: %w", err)
|
|
}
|
|
|
|
// Wait for the background go routine to exit
|
|
ds.backgroundWg.Wait()
|
|
|
|
return nil
|
|
}
|