diff --git a/extern/sector-storage/manager.go b/extern/sector-storage/manager.go index 2b7e85e3c..51558aaad 100644 --- a/extern/sector-storage/manager.go +++ b/extern/sector-storage/manager.go @@ -235,14 +235,14 @@ func (m *Manager) SectorsUnsealPiece(ctx context.Context, sector storage.SectorR // one of it's sealing scratch spaces to store them after fetching them from another worker. selector := newExistingSelector(m.index, sector.ID, storiface.FTSealed|storiface.FTCache, true) - log.Debugf("schedule unseal for sector %d", sector.ID) + log.Debugf("will schedule unseal for sector %d", sector.ID) err = m.sched.Schedule(ctx, sector, sealtasks.TTUnseal, selector, sealFetch, func(ctx context.Context, w Worker) error { // TODO: make restartable // NOTE: we're unsealing the whole sector here as with SDR we can't really // unseal the sector partially. Requesting the whole sector here can // save us some work in case another piece is requested from here - log.Debugf("unseal sector %d", sector.ID) + log.Debugf("calling unseal sector on worker, sectoID=%d", sector.ID) // Note: This unseal piece call will essentially become a no-op if the worker already has an Unsealed sector file for the given sector. _, err := m.waitSimpleCall(ctx)(w.UnsealPiece(ctx, sector, 0, abi.PaddedPieceSize(ssize).Unpadded(), ticket, *unsealed)) diff --git a/extern/sector-storage/piece_provider.go b/extern/sector-storage/piece_provider.go index 34ef44df5..7d46ed92a 100644 --- a/extern/sector-storage/piece_provider.go +++ b/extern/sector-storage/piece_provider.go @@ -57,7 +57,7 @@ func (p *pieceProvider) tryReadUnsealedPiece(ctx context.Context, sector storage // the unsealed piece. r, err := p.storage.Reader(ctx, sector, abi.PaddedPieceSize(offset.Padded()), size.Padded()) if err != nil { - log.Debugf("failed storage reader;sector=%+v, err:%s", sector.ID, err) + log.Debugf("did not get storage reader;sector=%+v, err:%s", sector.ID, err) cancel() return nil, nil, err } @@ -81,13 +81,14 @@ func (p *pieceProvider) ReadPiece(ctx context.Context, sector storage.SectorRef, r, unlock, err := p.tryReadUnsealedPiece(ctx, sector, offset, size) - log.Debugf("result of tryReadUnsealedPiece: r=%+v, err=%s", r, err) + log.Debugf("result of first tryReadUnsealedPiece: r=%+v, err=%s", r, err) if xerrors.Is(err, storiface.ErrSectorNotFound) { log.Debugf("no unsealed sector file with unsealed piece, sector=%+v, offset=%d, size=%d", sector, offset, size) err = nil } if err != nil { + log.Errorf("returning error from ReadPiece:%s", err) return nil, false, err } @@ -103,6 +104,7 @@ func (p *pieceProvider) ReadPiece(ctx context.Context, sector storage.SectorRef, commd = nil } if err := p.uns.SectorsUnsealPiece(ctx, sector, offset, size, ticket, commd); err != nil { + log.Errorf("failed to SectorsUnsealPiece: %s", err) return nil, false, xerrors.Errorf("unsealing piece: %w", err) } @@ -110,9 +112,11 @@ func (p *pieceProvider) ReadPiece(ctx context.Context, sector storage.SectorRef, r, unlock, err = p.tryReadUnsealedPiece(ctx, sector, offset, size) if err != nil { + log.Errorf("failed to tryReadUnsealedPiece after SectorsUnsealPiece: %s", err) return nil, true, xerrors.Errorf("read after unsealing: %w", err) } if r == nil { + log.Errorf("got no reader after unsealing piece") return nil, true, xerrors.Errorf("got no reader after unsealing piece") } log.Debugf("got a reader to read unsealed piece, sector=%+v, offset=%d, size=%d", sector, offset, size) diff --git a/extern/sector-storage/stores/remote.go b/extern/sector-storage/stores/remote.go index 1bb6b041b..2906756de 100644 --- a/extern/sector-storage/stores/remote.go +++ b/extern/sector-storage/stores/remote.go @@ -548,6 +548,7 @@ func (r *Remote) Reader(ctx context.Context, s storage.SectorRef, offset, size a // if they have the unsealed piece in the unsealed sector file. si, err := r.index.StorageFindSector(ctx, s.ID, ft, 0, false) if err != nil { + log.Debugf("Reader, did not find unsealed file on any of the workers %s (+%d,%d)", path, offset, size) return nil, err }