// SPDX-License-Identifier: GPL-2.0 /* * Copyright (C) 2010 Red Hat, Inc. * Copyright (C) 2016-2023 Christoph Hellwig. */ #include #include #include #include "internal.h" #include "trace.h" static DEFINE_SPINLOCK(failed_read_lock); static struct bio_list failed_read_list = BIO_EMPTY_LIST; static u32 __iomap_read_end_io(struct bio *bio, int error) { struct folio_iter fi; u32 folio_count = 0; bio_for_each_folio_all(fi, bio) { iomap_finish_folio_read(fi.folio, fi.offset, fi.length, error); folio_count++; } if (bio_integrity(bio)) fs_bio_integrity_free(bio); bio_put(bio); return folio_count; } static void iomap_fail_reads( struct work_struct *work) { struct bio *bio; struct bio_list tmp = BIO_EMPTY_LIST; unsigned long flags; spin_lock_irqsave(&failed_read_lock, flags); bio_list_merge_init(&tmp, &failed_read_list); spin_unlock_irqrestore(&failed_read_lock, flags); while ((bio = bio_list_pop(&tmp)) != NULL) { __iomap_read_end_io(bio, blk_status_to_errno(bio->bi_status)); cond_resched(); } } static DECLARE_WORK(failed_read_work, iomap_fail_reads); static void iomap_fail_buffered_read(struct bio *bio) { unsigned long flags; /* * Bounce I/O errors to a workqueue to avoid nested i_lock acquisitions * in the fserror code. The caller no longer owns the bio reference * after the spinlock drops. */ spin_lock_irqsave(&failed_read_lock, flags); if (bio_list_empty(&failed_read_list)) WARN_ON_ONCE(!schedule_work(&failed_read_work)); bio_list_add(&failed_read_list, bio); spin_unlock_irqrestore(&failed_read_lock, flags); } static void iomap_read_end_io(struct bio *bio) { if (bio->bi_status) { iomap_fail_buffered_read(bio); return; } __iomap_read_end_io(bio, 0); } u32 iomap_finish_ioend_buffered_read(struct iomap_ioend *ioend) { return __iomap_read_end_io(&ioend->io_bio, ioend->io_error); } static void iomap_bio_submit_read(const struct iomap_iter *iter, struct iomap_read_folio_ctx *ctx) { struct bio *bio = ctx->read_ctx; if (iter->iomap.flags & IOMAP_F_INTEGRITY) fs_bio_integrity_alloc(bio); submit_bio(bio); } static struct bio_set *iomap_read_bio_set(struct iomap_read_folio_ctx *ctx) { if (ctx->ops && ctx->ops->bio_set) return ctx->ops->bio_set; return &fs_bio_set; } static void iomap_read_alloc_bio(const struct iomap_iter *iter, struct iomap_read_folio_ctx *ctx, size_t plen) { const struct iomap *iomap = &iter->iomap; unsigned int nr_vecs = DIV_ROUND_UP(iomap_length(iter), PAGE_SIZE); struct bio_set *bio_set = iomap_read_bio_set(ctx); struct folio *folio = ctx->cur_folio; gfp_t gfp = mapping_gfp_constraint(folio->mapping, GFP_KERNEL); gfp_t orig_gfp = gfp; struct bio *bio; /* Submit the existing range if there was one. */ if (ctx->read_ctx) ctx->ops->submit_read(iter, ctx); /* Same as readahead_gfp_mask: */ if (ctx->rac) gfp |= __GFP_NORETRY | __GFP_NOWARN; /* * If the bio_alloc fails, try it again for a single page to avoid * having to deal with partial page reads. This emulates what * do_mpage_read_folio does. */ bio = bio_alloc_bioset(iomap->bdev, bio_max_segs(nr_vecs), REQ_OP_READ, gfp, bio_set); if (!bio) bio = bio_alloc_bioset(iomap->bdev, 1, REQ_OP_READ, orig_gfp, bio_set); if (ctx->rac) bio->bi_opf |= REQ_RAHEAD; bio->bi_iter.bi_sector = iomap_sector(iomap, iter->pos); bio->bi_end_io = iomap_read_end_io; bio_add_folio_nofail(bio, folio, plen, offset_in_folio(folio, iter->pos)); ctx->read_ctx = bio; ctx->read_ctx_file_offset = iter->pos; } int iomap_bio_read_folio_range(const struct iomap_iter *iter, struct iomap_read_folio_ctx *ctx, size_t plen) { struct folio *folio = ctx->cur_folio; struct bio *bio = ctx->read_ctx; if (!bio || bio_end_sector(bio) != iomap_sector(&iter->iomap, iter->pos) || bio->bi_iter.bi_size > iomap_max_bio_size(&iter->iomap) - plen || !bio_add_folio(bio, folio, plen, offset_in_folio(folio, iter->pos))) iomap_read_alloc_bio(iter, ctx, plen); return 0; } EXPORT_SYMBOL_GPL(iomap_bio_read_folio_range); const struct iomap_read_ops iomap_bio_read_ops = { .read_folio_range = iomap_bio_read_folio_range, .submit_read = iomap_bio_submit_read, }; EXPORT_SYMBOL_GPL(iomap_bio_read_ops); int iomap_bio_read_folio_range_sync(const struct iomap_iter *iter, struct folio *folio, loff_t pos, size_t len) { const struct iomap *srcmap = iomap_iter_srcmap(iter); sector_t sector = iomap_sector(srcmap, pos); struct bio_vec bvec; struct bio bio; int error; bio_init(&bio, srcmap->bdev, &bvec, 1, REQ_OP_READ); bio.bi_iter.bi_sector = sector; bio_add_folio_nofail(&bio, folio, len, offset_in_folio(folio, pos)); if (srcmap->flags & IOMAP_F_INTEGRITY) fs_bio_integrity_alloc(&bio); error = submit_bio_wait(&bio); if (srcmap->flags & IOMAP_F_INTEGRITY) { if (!error) error = fs_bio_integrity_verify(&bio, sector, len); fs_bio_integrity_free(&bio); } return error; }