Address review

This commit is contained in:
Aayush Rajasekaran 2022-01-12 12:57:34 -05:00 committed by Jennifer Wang
parent dafd1f77fa
commit 5b38419c5f
2 changed files with 73 additions and 51 deletions

View File

@ -15,19 +15,25 @@ import (
// blockstore logger. // blockstore logger.
var autolog = log.Named("auto") var autolog = log.Named("auto")
// contains the same set of blocks twice, once as an ordered list for flushing, and as a map for fast access
type blockBatch struct {
blockList []block.Block
blockMap map[cid.Cid]block.Block
}
type AutobatchBlockstore struct { type AutobatchBlockstore struct {
// TODO: drop if memory consumption is too high // TODO: drop if memory consumption is too high
addedCids map[cid.Cid]struct{} addedCids map[cid.Cid]struct{}
bufferedLk sync.Mutex lock sync.Mutex
bufferedBlksOrdered []block.Block bufferedBatch blockBatch
bufferedBlksMap map[cid.Cid]block.Block
flushingLk sync.Mutex // the flush worker has sole control (including read) over the flushingBatch.blockList and flushErr until shutdown
flushingBlksMap map[cid.Cid]block.Block flushingBatch blockBatch
flushErr error
flushCh chan struct{}
flushCh chan struct{}
flushErr error
flushRetryDelay time.Duration flushRetryDelay time.Duration
flushCtx context.Context flushCtx context.Context
shutdownCh chan struct{} shutdownCh chan struct{}
@ -40,29 +46,32 @@ type AutobatchBlockstore struct {
func NewAutobatch(ctx context.Context, backingBs Blockstore, bufferCapacity int) *AutobatchBlockstore { func NewAutobatch(ctx context.Context, backingBs Blockstore, bufferCapacity int) *AutobatchBlockstore {
bs := &AutobatchBlockstore{ bs := &AutobatchBlockstore{
addedCids: make(map[cid.Cid]struct{}), addedCids: make(map[cid.Cid]struct{}),
backingBs: backingBs, backingBs: backingBs,
bufferCapacity: bufferCapacity, bufferCapacity: bufferCapacity,
bufferedBlksMap: make(map[cid.Cid]block.Block), flushCtx: ctx,
flushingBlksMap: make(map[cid.Cid]block.Block), flushCh: make(chan struct{}, 1),
flushCtx: ctx,
flushCh: make(chan struct{}, 1),
// could be made configable // could be made configable
flushRetryDelay: time.Second * 5, flushRetryDelay: time.Millisecond * 100,
} }
bs.bufferedBatch.blockMap = make(map[cid.Cid]block.Block)
bs.flushingBatch.blockMap = make(map[cid.Cid]block.Block)
go bs.flushWorker() go bs.flushWorker()
return bs return bs
} }
func (bs *AutobatchBlockstore) Put(ctx context.Context, blk block.Block) error { func (bs *AutobatchBlockstore) Put(ctx context.Context, blk block.Block) error {
bs.bufferedLk.Lock() bs.lock.Lock()
defer bs.lock.Unlock()
_, ok := bs.addedCids[blk.Cid()] _, ok := bs.addedCids[blk.Cid()]
if !ok { if !ok {
bs.addedCids[blk.Cid()] = struct{}{} bs.addedCids[blk.Cid()] = struct{}{}
bs.bufferedBlksOrdered = append(bs.bufferedBlksOrdered, blk) bs.bufferedBatch.blockList = append(bs.bufferedBatch.blockList, blk)
bs.bufferedBlksMap[blk.Cid()] = blk bs.bufferedBatch.blockMap[blk.Cid()] = blk
bs.bufferSize += len(blk.RawData()) bs.bufferSize += len(blk.RawData())
if bs.bufferSize >= bs.bufferCapacity { if bs.bufferSize >= bs.bufferCapacity {
// signal that a flush is appropriate, may be ignored // signal that a flush is appropriate, may be ignored
@ -73,7 +82,7 @@ func (bs *AutobatchBlockstore) Put(ctx context.Context, blk block.Block) error {
} }
} }
} }
bs.bufferedLk.Unlock()
return nil return nil
} }
@ -85,11 +94,9 @@ func (bs *AutobatchBlockstore) flushWorker() {
for putErr != nil { for putErr != nil {
select { select {
case <-bs.shutdownCh: case <-bs.shutdownCh:
bs.flushErr = putErr
return return
default: case <-time.After(bs.flushRetryDelay):
autolog.Errorf("FLUSH ERRORED: %w, retrying in %v", putErr, bs.flushRetryDelay) autolog.Errorf("FLUSH ERRORED: %w, retrying after %v", putErr, bs.flushRetryDelay)
time.Sleep(bs.flushRetryDelay)
putErr = bs.doFlush(bs.flushCtx) putErr = bs.doFlush(bs.flushCtx)
} }
} }
@ -99,31 +106,31 @@ func (bs *AutobatchBlockstore) flushWorker() {
} }
} }
// caller must NOT hold lock
func (bs *AutobatchBlockstore) doFlush(ctx context.Context) error { func (bs *AutobatchBlockstore) doFlush(ctx context.Context) error {
bs.bufferedLk.Lock() if bs.flushErr == nil {
bs.flushingLk.Lock() bs.lock.Lock()
// We do NOT clear addedCids here, because its purpose is to expedite Puts // We do NOT clear addedCids here, because its purpose is to expedite Puts
flushingBlksOrdered := bs.bufferedBlksOrdered bs.flushingBatch = bs.bufferedBatch
bs.flushingBlksMap = bs.bufferedBlksMap bs.bufferedBatch.blockList = make([]block.Block, 0, len(bs.flushingBatch.blockList))
bs.bufferedBlksOrdered = []block.Block{} bs.bufferedBatch.blockMap = make(map[cid.Cid]block.Block, len(bs.flushingBatch.blockMap))
bs.bufferedBlksMap = make(map[cid.Cid]block.Block) bs.lock.Unlock()
bs.bufferedLk.Unlock() }
bs.flushingLk.Unlock()
return bs.backingBs.PutMany(ctx, flushingBlksOrdered) bs.flushErr = bs.backingBs.PutMany(ctx, bs.flushingBatch.blockList)
return bs.flushErr
}
// caller must NOT hold lock
func (bs *AutobatchBlockstore) Flush(ctx context.Context) error {
return bs.doFlush(ctx)
} }
func (bs *AutobatchBlockstore) Shutdown(ctx context.Context) error { func (bs *AutobatchBlockstore) Shutdown(ctx context.Context) error {
// request one last flush of the worker
bs.flushCh <- struct{}{}
// shutdown the flush worker // shutdown the flush worker
bs.shutdownCh <- struct{}{} bs.shutdownCh <- struct{}{}
// if it ever errored, this method fails
if bs.flushErr != nil {
return xerrors.Errorf("flushWorker errored: %w", bs.flushErr)
}
// one last flush in case it's needed return bs.flushErr
return bs.doFlush(ctx)
} }
func (bs *AutobatchBlockstore) Get(ctx context.Context, c cid.Cid) (block.Block, error) { func (bs *AutobatchBlockstore) Get(ctx context.Context, c cid.Cid) (block.Block, error) {
@ -137,24 +144,28 @@ func (bs *AutobatchBlockstore) Get(ctx context.Context, c cid.Cid) (block.Block,
return blk, err return blk, err
} }
bs.flushingLk.Lock() bs.lock.Lock()
v, ok := bs.flushingBlksMap[c] defer bs.lock.Unlock()
bs.flushingLk.Unlock() v, ok := bs.flushingBatch.blockMap[c]
if ok { if ok {
return v, nil return v, nil
} }
bs.bufferedLk.Lock() v, ok = bs.flushingBatch.blockMap[c]
v, ok = bs.bufferedBlksMap[c]
bs.bufferedLk.Unlock()
if ok { if ok {
return v, nil return v, nil
} }
return nil, ErrNotFound // check the backingBs in case it just got put in the backingBs (and removed from the batch maps) while we were here
return bs.backingBs.Get(ctx, c)
} }
func (bs *AutobatchBlockstore) DeleteBlock(context.Context, cid.Cid) error { func (bs *AutobatchBlockstore) DeleteBlock(context.Context, cid.Cid) error {
// if we wanted to support this, we would have to:
// - flush
// - delete from the backingBs (if present)
// - remove from addedCids (if present)
// - if present in addedCids, also walk the ordered lists and remove if present
return xerrors.New("deletion is unsupported") return xerrors.New("deletion is unsupported")
} }
@ -195,8 +206,11 @@ func (bs *AutobatchBlockstore) PutMany(ctx context.Context, blks []block.Block)
} }
func (bs *AutobatchBlockstore) AllKeysChan(ctx context.Context) (<-chan cid.Cid, error) { func (bs *AutobatchBlockstore) AllKeysChan(ctx context.Context) (<-chan cid.Cid, error) {
return nil, xerrors.New("unsupported") if err := bs.Flush(ctx); err != nil {
return nil, err
}
return bs.backingBs.AllKeysChan(ctx)
} }
func (bs *AutobatchBlockstore) HashOnRead(enabled bool) { func (bs *AutobatchBlockstore) HashOnRead(enabled bool) {
@ -204,5 +218,9 @@ func (bs *AutobatchBlockstore) HashOnRead(enabled bool) {
} }
func (bs *AutobatchBlockstore) View(ctx context.Context, cid cid.Cid, callback func([]byte) error) error { func (bs *AutobatchBlockstore) View(ctx context.Context, cid cid.Cid, callback func([]byte) error) error {
return xerrors.New("unsupported") if err := bs.Flush(ctx); err != nil {
return err
}
return bs.backingBs.View(ctx, cid, callback)
} }

View File

@ -1298,8 +1298,12 @@ func upgradeActorsV7Common(
} }
// Persists the new tree and shuts down the flush worker // Persists the new tree and shuts down the flush worker
if err := writeStore.Flush(ctx); err != nil {
return cid.Undef, xerrors.Errorf("writeStore flush failed: %w", err)
}
if err := writeStore.Shutdown(ctx); err != nil { if err := writeStore.Shutdown(ctx); err != nil {
return cid.Undef, xerrors.Errorf("writeStore failed: %w", err) return cid.Undef, xerrors.Errorf("writeStore shutdown failed: %w", err)
} }
return newRoot, nil return newRoot, nil