Received: by 2002:a05:6358:3188:b0:123:57c1:9b43 with SMTP id q8csp6452261rwd; Mon, 5 Jun 2023 19:09:02 -0700 (PDT) X-Google-Smtp-Source: ACHHUZ4kFQ0H3xRKp+wi4OokqGn8f08ijkakGTgRXOI9YZ9dsL8/tj2PpiYBVvlxbsv2ISjvJveS X-Received: by 2002:a05:620a:3997:b0:75b:23a1:3683 with SMTP id ro23-20020a05620a399700b0075b23a13683mr418624qkn.68.1686017342626; Mon, 05 Jun 2023 19:09:02 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1686017342; cv=none; d=google.com; s=arc-20160816; b=eMcXY5sTp6l3fYR1LI965V7aogxmLC2n9Eu4heZw1yZ29vLDsa08CBSKZfRGPAb6/j JYjA9b628JfBFHwQRPdreQ6TBaZ/xKb4p8wzr8+zGHCYTWaofbjkuz1L0Q8NRSeJkqj2 pnvN+ylfHQtk/WoZ47NE4quI5cJVuvyD7lkSf+NsXkXvFOLPbL6IcAOdjcF3tPN8h4Vr F+kD9HbrUSmxFx40fnKyfXZnCEuQHDGSWYw7zEY9Lf/Q6VwftkJjoE45zBHDcwXIwGs7 ksUEckNQQut/FhHAb2fgEV9nMq05NThCEqREqXk25K0vRwSQrEvvieXtHHFL5J7cG9+z fl7A== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:in-reply-to:content-disposition:mime-version :references:message-id:subject:cc:to:from:date:dkim-signature; bh=zNuATo7mlaZdnzuS+aQ5PhELIIcPe1r+5lraojHcL3g=; b=0av7GO3TnEoRezByETJdAgvooMZR5/lYEIwc+4s66AI9grO9aj6xZ/NiOc4F+vY5Qr EOsRdl0+2sC99UYEnTTVP6AVRc5eLErZhZCGmuYiqrUS335s7To3vpHzSIpHs7K65g2r hBSfIQgdPdKX1/SQPLQjjg/6ADKPDpbiCnKJgtREWJKDsSlB31NTg5sQaRRdUKoBef5G s/mLY1bSwddWjSKPVEfd1F6zquFaTZH2AnWgtzwj1hTtVim3rIQ9ntrGoXNJ8qJVUhgK RcmKiuStxp3wNC8bLW05HQUdfO00OyUcxMCfQNAHAZ5t/3EfUETKU9bkURQoLSUXxle+ H1Bw== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@redhat.com header.s=mimecast20190719 header.b="K6eQk/bB"; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=redhat.com Return-Path: Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id f16-20020ae9ea10000000b0075cef656c50si5149252qkg.726.2023.06.05.19.08.48; Mon, 05 Jun 2023 19:09:02 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; dkim=pass header.i=@redhat.com header.s=mimecast20190719 header.b="K6eQk/bB"; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=redhat.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S232528AbjFFCE5 (ORCPT + 99 others); Mon, 5 Jun 2023 22:04:57 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:43356 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S233561AbjFFCEx (ORCPT ); Mon, 5 Jun 2023 22:04:53 -0400 Received: from us-smtp-delivery-124.mimecast.com (us-smtp-delivery-124.mimecast.com [170.10.133.124]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 6469211A for ; Mon, 5 Jun 2023 19:03:58 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1686017037; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version:content-type:content-type: in-reply-to:in-reply-to:references:references; bh=zNuATo7mlaZdnzuS+aQ5PhELIIcPe1r+5lraojHcL3g=; b=K6eQk/bBBLQMFUOHWa1FeSRY/CSm/UQmryTjrKhxrS+W1HXPXv2wVsxHJpl081PDCfcfyc BHkE8Xz8Lr6SA5nHGkf8ZHfF3oD9xoRs4Zbnqu9GDnTFTKqaz78mA0RuFRT99Gh8SF5iMS yXsBxDZe612mO+93My/596iQ+fhXc9E= Received: from mimecast-mx02.redhat.com (mimecast-mx02.redhat.com [66.187.233.88]) by relay.mimecast.com with ESMTP with STARTTLS (version=TLSv1.2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id us-mta-646-1T1R0cs7Mja3ABaaQnDAdQ-1; Mon, 05 Jun 2023 22:03:56 -0400 X-MC-Unique: 1T1R0cs7Mja3ABaaQnDAdQ-1 Received: from smtp.corp.redhat.com (int-mx06.intmail.prod.int.rdu2.redhat.com [10.11.54.6]) (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits)) (No client certificate requested) by mimecast-mx02.redhat.com (Postfix) with ESMTPS id A02F080120A; Tue, 6 Jun 2023 02:03:55 +0000 (UTC) Received: from ovpn-8-17.pek2.redhat.com (ovpn-8-17.pek2.redhat.com [10.72.8.17]) by smtp.corp.redhat.com (Postfix) with ESMTPS id 60CDE2166B25; Tue, 6 Jun 2023 02:03:49 +0000 (UTC) Date: Tue, 6 Jun 2023 10:03:45 +0800 From: Ming Lei To: Kent Overstreet Cc: axboe@kernel.dk, linux-kernel@vger.kernel.org, linux-block@vger.kernel.org, Phillip Lougher , ming.lei@redhat.com Subject: Re: [PATCH v2 1/5] block: Rework bio_for_each_segment_all() Message-ID: References: <20230605212717.2570570-1-kent.overstreet@linux.dev> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20230605212717.2570570-1-kent.overstreet@linux.dev> X-Scanned-By: MIMEDefang 3.1 on 10.11.54.6 X-Spam-Status: No, score=-2.1 required=5.0 tests=BAYES_00,DKIMWL_WL_HIGH, DKIM_SIGNED,DKIM_VALID,DKIM_VALID_AU,DKIM_VALID_EF,RCVD_IN_DNSWL_NONE, SPF_HELO_NONE,SPF_NONE,T_SCC_BODY_TEXT_LINE,URIBL_BLOCKED autolearn=ham autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Mon, Jun 05, 2023 at 05:27:13PM -0400, Kent Overstreet wrote: > This patch reworks bio_for_each_segment_all() to be more inline with how > the other bio iterators work: > > - bio_iter_all_peek() now returns a synthesized bio_vec; we don't stash > one in the iterator and pass a pointer to it - bad. This way makes it > clearer what's a constructed value vs. a reference to something > pre-existing, and it also will help with cleaning up and > consolidating code with bio_for_each_folio_all(). by-value is more readable & clean, the current by-reference is just for aligning with original interface. > > - We now provide bio_for_each_segment_all_continue(), for squashfs: > this makes their code clearer. Probably it is better to move the new helper into new patch, but I guess it could be hard for this patch. > > Signed-off-by: Kent Overstreet > Cc: Jens Axboe > Cc: linux-block@vger.kernel.org > Cc: Ming Lei > Cc: Phillip Lougher > --- > block/bio.c | 38 ++++++++++++------------ > block/blk-map.c | 38 ++++++++++++------------ > block/bounce.c | 12 ++++---- > drivers/md/bcache/btree.c | 8 ++--- > drivers/md/dm-crypt.c | 10 +++---- > drivers/md/raid1.c | 4 +-- > fs/btrfs/disk-io.c | 4 +-- > fs/btrfs/extent_io.c | 50 +++++++++++++++---------------- > fs/btrfs/raid56.c | 14 ++++----- > fs/erofs/zdata.c | 4 +-- > fs/f2fs/data.c | 20 ++++++------- > fs/gfs2/lops.c | 10 +++---- > fs/gfs2/meta_io.c | 8 ++--- > fs/squashfs/block.c | 48 +++++++++++++++++------------- > fs/squashfs/lz4_wrapper.c | 17 ++++++----- > fs/squashfs/lzo_wrapper.c | 17 ++++++----- > fs/squashfs/xz_wrapper.c | 19 ++++++------ > fs/squashfs/zlib_wrapper.c | 18 ++++++----- > fs/squashfs/zstd_wrapper.c | 19 ++++++------ > include/linux/bio.h | 28 ++++++++++++----- > include/linux/bvec.h | 61 ++++++++++++++++++++++---------------- > 21 files changed, 242 insertions(+), 205 deletions(-) > > diff --git a/block/bio.c b/block/bio.c > index 798cc4cf3b..2af9f50643 100644 > --- a/block/bio.c > +++ b/block/bio.c > @@ -1163,13 +1163,13 @@ EXPORT_SYMBOL(bio_add_folio); > > void __bio_release_pages(struct bio *bio, bool mark_dirty) > { > - struct bvec_iter_all iter_all; > - struct bio_vec *bvec; > + struct bvec_iter_all iter; > + struct bio_vec bvec; > > - bio_for_each_segment_all(bvec, bio, iter_all) { > - if (mark_dirty && !PageCompound(bvec->bv_page)) > - set_page_dirty_lock(bvec->bv_page); > - bio_release_page(bio, bvec->bv_page); > + bio_for_each_segment_all(bvec, bio, iter) { > + if (mark_dirty && !PageCompound(bvec.bv_page)) > + set_page_dirty_lock(bvec.bv_page); > + bio_release_page(bio, bvec.bv_page); > } > } > EXPORT_SYMBOL_GPL(__bio_release_pages); > @@ -1435,11 +1435,11 @@ EXPORT_SYMBOL(bio_copy_data); > > void bio_free_pages(struct bio *bio) > { > - struct bio_vec *bvec; > - struct bvec_iter_all iter_all; > + struct bvec_iter_all iter; > + struct bio_vec bvec; > > - bio_for_each_segment_all(bvec, bio, iter_all) > - __free_page(bvec->bv_page); > + bio_for_each_segment_all(bvec, bio, iter) > + __free_page(bvec.bv_page); > } > EXPORT_SYMBOL(bio_free_pages); > > @@ -1474,12 +1474,12 @@ EXPORT_SYMBOL(bio_free_pages); > */ > void bio_set_pages_dirty(struct bio *bio) > { > - struct bio_vec *bvec; > - struct bvec_iter_all iter_all; > + struct bvec_iter_all iter; > + struct bio_vec bvec; > > - bio_for_each_segment_all(bvec, bio, iter_all) { > - if (!PageCompound(bvec->bv_page)) > - set_page_dirty_lock(bvec->bv_page); > + bio_for_each_segment_all(bvec, bio, iter) { > + if (!PageCompound(bvec.bv_page)) > + set_page_dirty_lock(bvec.bv_page); > } > } > > @@ -1522,12 +1522,12 @@ static void bio_dirty_fn(struct work_struct *work) > > void bio_check_pages_dirty(struct bio *bio) > { > - struct bio_vec *bvec; > + struct bvec_iter_all iter; > + struct bio_vec bvec; > unsigned long flags; > - struct bvec_iter_all iter_all; > > - bio_for_each_segment_all(bvec, bio, iter_all) { > - if (!PageDirty(bvec->bv_page) && !PageCompound(bvec->bv_page)) > + bio_for_each_segment_all(bvec, bio, iter) { > + if (!PageDirty(bvec.bv_page) && !PageCompound(bvec.bv_page)) > goto defer; > } > > diff --git a/block/blk-map.c b/block/blk-map.c > index 3551c3ff17..524d84011f 100644 > --- a/block/blk-map.c > +++ b/block/blk-map.c > @@ -47,21 +47,21 @@ static struct bio_map_data *bio_alloc_map_data(struct iov_iter *data, > */ > static int bio_copy_from_iter(struct bio *bio, struct iov_iter *iter) > { > - struct bio_vec *bvec; > - struct bvec_iter_all iter_all; > + struct bvec_iter_all bv_iter; > + struct bio_vec bvec; > > - bio_for_each_segment_all(bvec, bio, iter_all) { > + bio_for_each_segment_all(bvec, bio, bv_iter) { > ssize_t ret; > > - ret = copy_page_from_iter(bvec->bv_page, > - bvec->bv_offset, > - bvec->bv_len, > + ret = copy_page_from_iter(bvec.bv_page, > + bvec.bv_offset, > + bvec.bv_len, > iter); > > if (!iov_iter_count(iter)) > break; > > - if (ret < bvec->bv_len) > + if (ret < bvec.bv_len) > return -EFAULT; > } > > @@ -78,21 +78,21 @@ static int bio_copy_from_iter(struct bio *bio, struct iov_iter *iter) > */ > static int bio_copy_to_iter(struct bio *bio, struct iov_iter iter) > { > - struct bio_vec *bvec; > - struct bvec_iter_all iter_all; > + struct bvec_iter_all bv_iter; > + struct bio_vec bvec; > > - bio_for_each_segment_all(bvec, bio, iter_all) { > + bio_for_each_segment_all(bvec, bio, bv_iter) { > ssize_t ret; > > - ret = copy_page_to_iter(bvec->bv_page, > - bvec->bv_offset, > - bvec->bv_len, > + ret = copy_page_to_iter(bvec.bv_page, > + bvec.bv_offset, > + bvec.bv_len, > &iter); > > if (!iov_iter_count(&iter)) > break; > > - if (ret < bvec->bv_len) > + if (ret < bvec.bv_len) > return -EFAULT; > } > > @@ -443,12 +443,12 @@ static void bio_copy_kern_endio(struct bio *bio) > static void bio_copy_kern_endio_read(struct bio *bio) > { > char *p = bio->bi_private; > - struct bio_vec *bvec; > - struct bvec_iter_all iter_all; > + struct bvec_iter_all iter; > + struct bio_vec bvec; > > - bio_for_each_segment_all(bvec, bio, iter_all) { > - memcpy_from_bvec(p, bvec); > - p += bvec->bv_len; > + bio_for_each_segment_all(bvec, bio, iter) { > + memcpy_from_bvec(p, &bvec); > + p += bvec.bv_len; > } > > bio_copy_kern_endio(bio); > diff --git a/block/bounce.c b/block/bounce.c > index 7cfcb242f9..e701832d76 100644 > --- a/block/bounce.c > +++ b/block/bounce.c > @@ -102,18 +102,18 @@ static void copy_to_high_bio_irq(struct bio *to, struct bio *from) > static void bounce_end_io(struct bio *bio) > { > struct bio *bio_orig = bio->bi_private; > - struct bio_vec *bvec, orig_vec; > + struct bio_vec bvec, orig_vec; > struct bvec_iter orig_iter = bio_orig->bi_iter; > - struct bvec_iter_all iter_all; > + struct bvec_iter_all iter; > > /* > * free up bounce indirect pages used > */ > - bio_for_each_segment_all(bvec, bio, iter_all) { > + bio_for_each_segment_all(bvec, bio, iter) { > orig_vec = bio_iter_iovec(bio_orig, orig_iter); > - if (bvec->bv_page != orig_vec.bv_page) { > - dec_zone_page_state(bvec->bv_page, NR_BOUNCE); > - mempool_free(bvec->bv_page, &page_pool); > + if (bvec.bv_page != orig_vec.bv_page) { > + dec_zone_page_state(bvec.bv_page, NR_BOUNCE); > + mempool_free(bvec.bv_page, &page_pool); > } > bio_advance_iter(bio_orig, &orig_iter, orig_vec.bv_len); > } > diff --git a/drivers/md/bcache/btree.c b/drivers/md/bcache/btree.c > index 147c493a98..98ce12b239 100644 > --- a/drivers/md/bcache/btree.c > +++ b/drivers/md/bcache/btree.c > @@ -373,12 +373,12 @@ static void do_btree_node_write(struct btree *b) > bset_sector_offset(&b->keys, i)); > > if (!bch_bio_alloc_pages(b->bio, __GFP_NOWARN|GFP_NOWAIT)) { > - struct bio_vec *bv; > + struct bio_vec bv; > void *addr = (void *) ((unsigned long) i & ~(PAGE_SIZE - 1)); > - struct bvec_iter_all iter_all; > + struct bvec_iter_all iter; > > - bio_for_each_segment_all(bv, b->bio, iter_all) { > - memcpy(page_address(bv->bv_page), addr, PAGE_SIZE); > + bio_for_each_segment_all(bv, b->bio, iter) { > + memcpy(page_address(bv.bv_page), addr, PAGE_SIZE); > addr += PAGE_SIZE; > } > > diff --git a/drivers/md/dm-crypt.c b/drivers/md/dm-crypt.c > index 8b47b913ee..8ffc225f7b 100644 > --- a/drivers/md/dm-crypt.c > +++ b/drivers/md/dm-crypt.c > @@ -1713,12 +1713,12 @@ static struct bio *crypt_alloc_buffer(struct dm_crypt_io *io, unsigned int size) > > static void crypt_free_buffer_pages(struct crypt_config *cc, struct bio *clone) > { > - struct bio_vec *bv; > - struct bvec_iter_all iter_all; > + struct bvec_iter_all iter; > + struct bio_vec bv; > > - bio_for_each_segment_all(bv, clone, iter_all) { > - BUG_ON(!bv->bv_page); > - mempool_free(bv->bv_page, &cc->page_pool); > + bio_for_each_segment_all(bv, clone, iter) { > + BUG_ON(!bv.bv_page); > + mempool_free(bv.bv_page, &cc->page_pool); > } > } > > diff --git a/drivers/md/raid1.c b/drivers/md/raid1.c > index 68a9e2d998..4f58cae37e 100644 > --- a/drivers/md/raid1.c > +++ b/drivers/md/raid1.c > @@ -2188,7 +2188,7 @@ static void process_checks(struct r1bio *r1_bio) > blk_status_t status = sbio->bi_status; > struct page **ppages = get_resync_pages(pbio)->pages; > struct page **spages = get_resync_pages(sbio)->pages; > - struct bio_vec *bi; > + struct bio_vec bi; > int page_len[RESYNC_PAGES] = { 0 }; > struct bvec_iter_all iter_all; > > @@ -2198,7 +2198,7 @@ static void process_checks(struct r1bio *r1_bio) > sbio->bi_status = 0; > > bio_for_each_segment_all(bi, sbio, iter_all) > - page_len[j++] = bi->bv_len; > + page_len[j++] = bi.bv_len; > > if (!status) { > for (j = vcnt; j-- ; ) { > diff --git a/fs/btrfs/disk-io.c b/fs/btrfs/disk-io.c > index fbf9006c62..10326f978a 100644 > --- a/fs/btrfs/disk-io.c > +++ b/fs/btrfs/disk-io.c > @@ -3807,12 +3807,12 @@ ALLOW_ERROR_INJECTION(open_ctree, ERRNO); > static void btrfs_end_super_write(struct bio *bio) > { > struct btrfs_device *device = bio->bi_private; > - struct bio_vec *bvec; > + struct bio_vec bvec; > struct bvec_iter_all iter_all; > struct page *page; > > bio_for_each_segment_all(bvec, bio, iter_all) { > - page = bvec->bv_page; > + page = bvec.bv_page; > > if (bio->bi_status) { > btrfs_warn_rl_in_rcu(device->fs_info, > diff --git a/fs/btrfs/extent_io.c b/fs/btrfs/extent_io.c > index a1adadd5d2..e3b7fbbad7 100644 > --- a/fs/btrfs/extent_io.c > +++ b/fs/btrfs/extent_io.c > @@ -573,34 +573,34 @@ static void end_bio_extent_writepage(struct btrfs_bio *bbio) > { > struct bio *bio = &bbio->bio; > int error = blk_status_to_errno(bio->bi_status); > - struct bio_vec *bvec; > + struct bio_vec bvec; > u64 start; > u64 end; > struct bvec_iter_all iter_all; > > ASSERT(!bio_flagged(bio, BIO_CLONED)); > bio_for_each_segment_all(bvec, bio, iter_all) { > - struct page *page = bvec->bv_page; > + struct page *page = bvec.bv_page; > struct inode *inode = page->mapping->host; > struct btrfs_fs_info *fs_info = btrfs_sb(inode->i_sb); > const u32 sectorsize = fs_info->sectorsize; > > /* Our read/write should always be sector aligned. */ > - if (!IS_ALIGNED(bvec->bv_offset, sectorsize)) > + if (!IS_ALIGNED(bvec.bv_offset, sectorsize)) > btrfs_err(fs_info, > "partial page write in btrfs with offset %u and length %u", > - bvec->bv_offset, bvec->bv_len); > - else if (!IS_ALIGNED(bvec->bv_len, sectorsize)) > + bvec.bv_offset, bvec.bv_len); > + else if (!IS_ALIGNED(bvec.bv_len, sectorsize)) > btrfs_info(fs_info, > "incomplete page write with offset %u and length %u", > - bvec->bv_offset, bvec->bv_len); > + bvec.bv_offset, bvec.bv_len); > > - start = page_offset(page) + bvec->bv_offset; > - end = start + bvec->bv_len - 1; > + start = page_offset(page) + bvec.bv_offset; > + end = start + bvec.bv_len - 1; > > end_extent_writepage(page, error, start, end); > > - btrfs_page_clear_writeback(fs_info, page, start, bvec->bv_len); > + btrfs_page_clear_writeback(fs_info, page, start, bvec.bv_len); > } > > bio_put(bio); > @@ -728,7 +728,7 @@ static struct extent_buffer *find_extent_buffer_readpage( > static void end_bio_extent_readpage(struct btrfs_bio *bbio) > { > struct bio *bio = &bbio->bio; > - struct bio_vec *bvec; > + struct bio_vec bvec; > struct processed_extent processed = { 0 }; > /* > * The offset to the beginning of a bio, since one bio can never be > @@ -741,7 +741,7 @@ static void end_bio_extent_readpage(struct btrfs_bio *bbio) > ASSERT(!bio_flagged(bio, BIO_CLONED)); > bio_for_each_segment_all(bvec, bio, iter_all) { > bool uptodate = !bio->bi_status; > - struct page *page = bvec->bv_page; > + struct page *page = bvec.bv_page; > struct inode *inode = page->mapping->host; > struct btrfs_fs_info *fs_info = btrfs_sb(inode->i_sb); > const u32 sectorsize = fs_info->sectorsize; > @@ -761,19 +761,19 @@ static void end_bio_extent_readpage(struct btrfs_bio *bbio) > * for unaligned offsets, and an error if they don't add up to > * a full sector. > */ > - if (!IS_ALIGNED(bvec->bv_offset, sectorsize)) > + if (!IS_ALIGNED(bvec.bv_offset, sectorsize)) > btrfs_err(fs_info, > "partial page read in btrfs with offset %u and length %u", > - bvec->bv_offset, bvec->bv_len); > - else if (!IS_ALIGNED(bvec->bv_offset + bvec->bv_len, > + bvec.bv_offset, bvec.bv_len); > + else if (!IS_ALIGNED(bvec.bv_offset + bvec.bv_len, > sectorsize)) > btrfs_info(fs_info, > "incomplete page read with offset %u and length %u", > - bvec->bv_offset, bvec->bv_len); > + bvec.bv_offset, bvec.bv_len); > > - start = page_offset(page) + bvec->bv_offset; > - end = start + bvec->bv_len - 1; > - len = bvec->bv_len; > + start = page_offset(page) + bvec.bv_offset; > + end = start + bvec.bv_len - 1; > + len = bvec.bv_len; > > mirror = bbio->mirror_num; > if (uptodate && !is_data_inode(inode) && > @@ -1879,7 +1879,7 @@ static void end_bio_subpage_eb_writepage(struct btrfs_bio *bbio) > { > struct bio *bio = &bbio->bio; > struct btrfs_fs_info *fs_info; > - struct bio_vec *bvec; > + struct bio_vec bvec; > struct bvec_iter_all iter_all; > > fs_info = btrfs_sb(bio_first_page_all(bio)->mapping->host->i_sb); > @@ -1887,12 +1887,12 @@ static void end_bio_subpage_eb_writepage(struct btrfs_bio *bbio) > > ASSERT(!bio_flagged(bio, BIO_CLONED)); > bio_for_each_segment_all(bvec, bio, iter_all) { > - struct page *page = bvec->bv_page; > - u64 bvec_start = page_offset(page) + bvec->bv_offset; > - u64 bvec_end = bvec_start + bvec->bv_len - 1; > + struct page *page = bvec.bv_page; > + u64 bvec_start = page_offset(page) + bvec.bv_offset; > + u64 bvec_end = bvec_start + bvec.bv_len - 1; > u64 cur_bytenr = bvec_start; > > - ASSERT(IS_ALIGNED(bvec->bv_len, fs_info->nodesize)); > + ASSERT(IS_ALIGNED(bvec.bv_len, fs_info->nodesize)); > > /* Iterate through all extent buffers in the range */ > while (cur_bytenr <= bvec_end) { > @@ -1936,14 +1936,14 @@ static void end_bio_subpage_eb_writepage(struct btrfs_bio *bbio) > static void end_bio_extent_buffer_writepage(struct btrfs_bio *bbio) > { > struct bio *bio = &bbio->bio; > - struct bio_vec *bvec; > + struct bio_vec bvec; > struct extent_buffer *eb; > int done; > struct bvec_iter_all iter_all; > > ASSERT(!bio_flagged(bio, BIO_CLONED)); > bio_for_each_segment_all(bvec, bio, iter_all) { > - struct page *page = bvec->bv_page; > + struct page *page = bvec.bv_page; > > eb = (struct extent_buffer *)page->private; > BUG_ON(!eb); > diff --git a/fs/btrfs/raid56.c b/fs/btrfs/raid56.c > index 2fab37f062..8da6097c42 100644 > --- a/fs/btrfs/raid56.c > +++ b/fs/btrfs/raid56.c > @@ -1393,7 +1393,7 @@ static struct sector_ptr *find_stripe_sector(struct btrfs_raid_bio *rbio, > static void set_bio_pages_uptodate(struct btrfs_raid_bio *rbio, struct bio *bio) > { > const u32 sectorsize = rbio->bioc->fs_info->sectorsize; > - struct bio_vec *bvec; > + struct bio_vec bvec; > struct bvec_iter_all iter_all; > > ASSERT(!bio_flagged(bio, BIO_CLONED)); > @@ -1402,9 +1402,9 @@ static void set_bio_pages_uptodate(struct btrfs_raid_bio *rbio, struct bio *bio) > struct sector_ptr *sector; > int pgoff; > > - for (pgoff = bvec->bv_offset; pgoff - bvec->bv_offset < bvec->bv_len; > + for (pgoff = bvec.bv_offset; pgoff - bvec.bv_offset < bvec.bv_len; > pgoff += sectorsize) { > - sector = find_stripe_sector(rbio, bvec->bv_page, pgoff); > + sector = find_stripe_sector(rbio, bvec.bv_page, pgoff); > ASSERT(sector); > if (sector) > sector->uptodate = 1; > @@ -1458,7 +1458,7 @@ static void verify_bio_data_sectors(struct btrfs_raid_bio *rbio, > { > struct btrfs_fs_info *fs_info = rbio->bioc->fs_info; > int total_sector_nr = get_bio_sector_nr(rbio, bio); > - struct bio_vec *bvec; > + struct bio_vec bvec; > struct bvec_iter_all iter_all; > > /* No data csum for the whole stripe, no need to verify. */ > @@ -1472,8 +1472,8 @@ static void verify_bio_data_sectors(struct btrfs_raid_bio *rbio, > bio_for_each_segment_all(bvec, bio, iter_all) { > int bv_offset; > > - for (bv_offset = bvec->bv_offset; > - bv_offset < bvec->bv_offset + bvec->bv_len; > + for (bv_offset = bvec.bv_offset; > + bv_offset < bvec.bv_offset + bvec.bv_len; > bv_offset += fs_info->sectorsize, total_sector_nr++) { > u8 csum_buf[BTRFS_CSUM_SIZE]; > u8 *expected_csum = rbio->csum_buf + > @@ -1484,7 +1484,7 @@ static void verify_bio_data_sectors(struct btrfs_raid_bio *rbio, > if (!test_bit(total_sector_nr, rbio->csum_bitmap)) > continue; > > - ret = btrfs_check_sector_csum(fs_info, bvec->bv_page, > + ret = btrfs_check_sector_csum(fs_info, bvec.bv_page, > bv_offset, csum_buf, expected_csum); > if (ret < 0) > set_bit(total_sector_nr, rbio->error_bitmap); > diff --git a/fs/erofs/zdata.c b/fs/erofs/zdata.c > index 45f21db230..fc0ce6d685 100644 > --- a/fs/erofs/zdata.c > +++ b/fs/erofs/zdata.c > @@ -1648,11 +1648,11 @@ static void z_erofs_decompressqueue_endio(struct bio *bio) > { > struct z_erofs_decompressqueue *q = bio->bi_private; > blk_status_t err = bio->bi_status; > - struct bio_vec *bvec; > + struct bio_vec bvec; > struct bvec_iter_all iter_all; > > bio_for_each_segment_all(bvec, bio, iter_all) { > - struct page *page = bvec->bv_page; > + struct page *page = bvec.bv_page; > > DBG_BUGON(PageUptodate(page)); > DBG_BUGON(z_erofs_page_is_invalidated(page)); > diff --git a/fs/f2fs/data.c b/fs/f2fs/data.c > index 7165b1202f..d687e11bdf 100644 > --- a/fs/f2fs/data.c > +++ b/fs/f2fs/data.c > @@ -139,12 +139,12 @@ struct bio_post_read_ctx { > */ > static void f2fs_finish_read_bio(struct bio *bio, bool in_task) > { > - struct bio_vec *bv; > + struct bio_vec bv; > struct bvec_iter_all iter_all; > struct bio_post_read_ctx *ctx = bio->bi_private; > > bio_for_each_segment_all(bv, bio, iter_all) { > - struct page *page = bv->bv_page; > + struct page *page = bv.bv_page; > > if (f2fs_is_compressed_page(page)) { > if (ctx && !ctx->decompression_attempted) > @@ -189,11 +189,11 @@ static void f2fs_verify_bio(struct work_struct *work) > * as those were handled separately by f2fs_end_read_compressed_page(). > */ > if (may_have_compressed_pages) { > - struct bio_vec *bv; > + struct bio_vec bv; > struct bvec_iter_all iter_all; > > bio_for_each_segment_all(bv, bio, iter_all) { > - struct page *page = bv->bv_page; > + struct page *page = bv.bv_page; > > if (!f2fs_is_compressed_page(page) && > !fsverity_verify_page(page)) { > @@ -241,13 +241,13 @@ static void f2fs_verify_and_finish_bio(struct bio *bio, bool in_task) > static void f2fs_handle_step_decompress(struct bio_post_read_ctx *ctx, > bool in_task) > { > - struct bio_vec *bv; > + struct bio_vec bv; > struct bvec_iter_all iter_all; > bool all_compressed = true; > block_t blkaddr = ctx->fs_blkaddr; > > bio_for_each_segment_all(bv, ctx->bio, iter_all) { > - struct page *page = bv->bv_page; > + struct page *page = bv.bv_page; > > if (f2fs_is_compressed_page(page)) > f2fs_end_read_compressed_page(page, false, blkaddr, > @@ -327,7 +327,7 @@ static void f2fs_read_end_io(struct bio *bio) > static void f2fs_write_end_io(struct bio *bio) > { > struct f2fs_sb_info *sbi; > - struct bio_vec *bvec; > + struct bio_vec bvec; > struct bvec_iter_all iter_all; > > iostat_update_and_unbind_ctx(bio); > @@ -337,7 +337,7 @@ static void f2fs_write_end_io(struct bio *bio) > bio->bi_status = BLK_STS_IOERR; > > bio_for_each_segment_all(bvec, bio, iter_all) { > - struct page *page = bvec->bv_page; > + struct page *page = bvec.bv_page; > enum count_type type = WB_DATA_TYPE(page); > > if (page_private_dummy(page)) { > @@ -583,7 +583,7 @@ static void __submit_merged_bio(struct f2fs_bio_info *io) > static bool __has_merged_page(struct bio *bio, struct inode *inode, > struct page *page, nid_t ino) > { > - struct bio_vec *bvec; > + struct bio_vec bvec; > struct bvec_iter_all iter_all; > > if (!bio) > @@ -593,7 +593,7 @@ static bool __has_merged_page(struct bio *bio, struct inode *inode, > return true; > > bio_for_each_segment_all(bvec, bio, iter_all) { > - struct page *target = bvec->bv_page; > + struct page *target = bvec.bv_page; > > if (fscrypt_is_bounce_page(target)) { > target = fscrypt_pagecache_page(target); > diff --git a/fs/gfs2/lops.c b/fs/gfs2/lops.c > index 1902413d5d..7f62fe8eb7 100644 > --- a/fs/gfs2/lops.c > +++ b/fs/gfs2/lops.c > @@ -202,7 +202,7 @@ static void gfs2_end_log_write_bh(struct gfs2_sbd *sdp, > static void gfs2_end_log_write(struct bio *bio) > { > struct gfs2_sbd *sdp = bio->bi_private; > - struct bio_vec *bvec; > + struct bio_vec bvec; > struct page *page; > struct bvec_iter_all iter_all; > > @@ -217,9 +217,9 @@ static void gfs2_end_log_write(struct bio *bio) > } > > bio_for_each_segment_all(bvec, bio, iter_all) { > - page = bvec->bv_page; > + page = bvec.bv_page; > if (page_has_buffers(page)) > - gfs2_end_log_write_bh(sdp, bvec, bio->bi_status); > + gfs2_end_log_write_bh(sdp, &bvec, bio->bi_status); > else > mempool_free(page, gfs2_page_pool); > } > @@ -395,11 +395,11 @@ static void gfs2_log_write_page(struct gfs2_sbd *sdp, struct page *page) > static void gfs2_end_log_read(struct bio *bio) > { > struct page *page; > - struct bio_vec *bvec; > + struct bio_vec bvec; > struct bvec_iter_all iter_all; > > bio_for_each_segment_all(bvec, bio, iter_all) { > - page = bvec->bv_page; > + page = bvec.bv_page; > if (bio->bi_status) { > int err = blk_status_to_errno(bio->bi_status); > > diff --git a/fs/gfs2/meta_io.c b/fs/gfs2/meta_io.c > index 924361fa51..832572784e 100644 > --- a/fs/gfs2/meta_io.c > +++ b/fs/gfs2/meta_io.c > @@ -193,15 +193,15 @@ struct buffer_head *gfs2_meta_new(struct gfs2_glock *gl, u64 blkno) > > static void gfs2_meta_read_endio(struct bio *bio) > { > - struct bio_vec *bvec; > + struct bio_vec bvec; > struct bvec_iter_all iter_all; > > bio_for_each_segment_all(bvec, bio, iter_all) { > - struct page *page = bvec->bv_page; > + struct page *page = bvec.bv_page; > struct buffer_head *bh = page_buffers(page); > - unsigned int len = bvec->bv_len; > + unsigned int len = bvec.bv_len; > > - while (bh_offset(bh) < bvec->bv_offset) > + while (bh_offset(bh) < bvec.bv_offset) > bh = bh->b_this_page; > do { > struct buffer_head *next = bh->b_this_page; > diff --git a/fs/squashfs/block.c b/fs/squashfs/block.c > index bed3bb8b27..83e8b44518 100644 > --- a/fs/squashfs/block.c > +++ b/fs/squashfs/block.c > @@ -35,30 +35,33 @@ static int copy_bio_to_actor(struct bio *bio, > int offset, int req_length) > { > void *actor_addr; > - struct bvec_iter_all iter_all = {}; > - struct bio_vec *bvec = bvec_init_iter_all(&iter_all); > + struct bvec_iter_all iter; > + struct bio_vec bvec; > int copied_bytes = 0; > int actor_offset = 0; > + int bytes_to_copy; > > squashfs_actor_nobuff(actor); > actor_addr = squashfs_first_page(actor); > > - if (WARN_ON_ONCE(!bio_next_segment(bio, &iter_all))) > - return 0; > + bvec_iter_all_init(&iter); > + bio_iter_all_advance(bio, &iter, offset); > > - while (copied_bytes < req_length) { > - int bytes_to_copy = min_t(int, bvec->bv_len - offset, > + while (copied_bytes < req_length && > + iter.idx < bio->bi_vcnt) { > + bvec = bio_iter_all_peek(bio, &iter); > + > + bytes_to_copy = min_t(int, bvec.bv_len, > PAGE_SIZE - actor_offset); > > bytes_to_copy = min_t(int, bytes_to_copy, > req_length - copied_bytes); > if (!IS_ERR(actor_addr)) > - memcpy(actor_addr + actor_offset, bvec_virt(bvec) + > - offset, bytes_to_copy); > + memcpy(actor_addr + actor_offset, bvec_virt(&bvec), > + bytes_to_copy); > > actor_offset += bytes_to_copy; > copied_bytes += bytes_to_copy; > - offset += bytes_to_copy; > > if (actor_offset >= PAGE_SIZE) { > actor_addr = squashfs_next_page(actor); > @@ -66,11 +69,8 @@ static int copy_bio_to_actor(struct bio *bio, > break; > actor_offset = 0; > } > - if (offset >= bvec->bv_len) { > - if (!bio_next_segment(bio, &iter_all)) > - break; > - offset = 0; > - } > + > + bio_iter_all_advance(bio, &iter, bytes_to_copy); > } > squashfs_finish_page(actor); > return copied_bytes; > @@ -159,8 +159,10 @@ int squashfs_read_data(struct super_block *sb, u64 index, int length, > * Metadata block. > */ > const u8 *data; > - struct bvec_iter_all iter_all = {}; > - struct bio_vec *bvec = bvec_init_iter_all(&iter_all); > + struct bvec_iter_all iter; > + struct bio_vec bvec; > + > + bvec_iter_all_init(&iter); > > if (index + 2 > msblk->bytes_used) { > res = -EIO; > @@ -170,21 +172,25 @@ int squashfs_read_data(struct super_block *sb, u64 index, int length, > if (res) > goto out; > > - if (WARN_ON_ONCE(!bio_next_segment(bio, &iter_all))) { > + bvec = bio_iter_all_peek(bio, &iter); > + > + if (WARN_ON_ONCE(!bvec.bv_len)) { > res = -EIO; > goto out_free_bio; > } > /* Extract the length of the metadata block */ > - data = bvec_virt(bvec); > + data = bvec_virt(&bvec); > length = data[offset]; > - if (offset < bvec->bv_len - 1) { > + if (offset < bvec.bv_len - 1) { > length |= data[offset + 1] << 8; > } else { > - if (WARN_ON_ONCE(!bio_next_segment(bio, &iter_all))) { > + bio_iter_all_advance(bio, &iter, bvec.bv_len); > + > + if (WARN_ON_ONCE(!bvec.bv_len)) { > res = -EIO; > goto out_free_bio; > } > - data = bvec_virt(bvec); > + data = bvec_virt(&bvec); > length |= data[0] << 8; > } > bio_free_pages(bio); > diff --git a/fs/squashfs/lz4_wrapper.c b/fs/squashfs/lz4_wrapper.c > index 49797729f1..bd0dd787d2 100644 > --- a/fs/squashfs/lz4_wrapper.c > +++ b/fs/squashfs/lz4_wrapper.c > @@ -92,20 +92,23 @@ static int lz4_uncompress(struct squashfs_sb_info *msblk, void *strm, > struct bio *bio, int offset, int length, > struct squashfs_page_actor *output) > { > - struct bvec_iter_all iter_all = {}; > - struct bio_vec *bvec = bvec_init_iter_all(&iter_all); > + struct bvec_iter_all iter; > + struct bio_vec bvec; > struct squashfs_lz4 *stream = strm; > void *buff = stream->input, *data; > int bytes = length, res; > > - while (bio_next_segment(bio, &iter_all)) { > - int avail = min(bytes, ((int)bvec->bv_len) - offset); > + bvec_iter_all_init(&iter); > + bio_iter_all_advance(bio, &iter, offset); > > - data = bvec_virt(bvec); > - memcpy(buff, data + offset, avail); > + bio_for_each_segment_all_continue(bvec, bio, iter) { > + unsigned avail = min_t(unsigned, bytes, bvec.bv_len); > + > + memcpy(buff, bvec_virt(&bvec), avail); > buff += avail; > bytes -= avail; > - offset = 0; > + if (!bytes) > + break; > } > > res = LZ4_decompress_safe(stream->input, stream->output, > diff --git a/fs/squashfs/lzo_wrapper.c b/fs/squashfs/lzo_wrapper.c > index d216aeefa8..bccfcfa12e 100644 > --- a/fs/squashfs/lzo_wrapper.c > +++ b/fs/squashfs/lzo_wrapper.c > @@ -66,21 +66,24 @@ static int lzo_uncompress(struct squashfs_sb_info *msblk, void *strm, > struct bio *bio, int offset, int length, > struct squashfs_page_actor *output) > { > - struct bvec_iter_all iter_all = {}; > - struct bio_vec *bvec = bvec_init_iter_all(&iter_all); > + struct bvec_iter_all iter; > + struct bio_vec bvec; > struct squashfs_lzo *stream = strm; > void *buff = stream->input, *data; > int bytes = length, res; > size_t out_len = output->length; > > - while (bio_next_segment(bio, &iter_all)) { > - int avail = min(bytes, ((int)bvec->bv_len) - offset); > + bvec_iter_all_init(&iter); > + bio_iter_all_advance(bio, &iter, offset); > > - data = bvec_virt(bvec); > - memcpy(buff, data + offset, avail); > + bio_for_each_segment_all_continue(bvec, bio, iter) { > + unsigned avail = min_t(unsigned, bytes, bvec.bv_len); > + > + memcpy(buff, bvec_virt(&bvec), avail); > buff += avail; > bytes -= avail; > - offset = 0; > + if (!bytes) > + break; > } > > res = lzo1x_decompress_safe(stream->input, (size_t)length, > diff --git a/fs/squashfs/xz_wrapper.c b/fs/squashfs/xz_wrapper.c > index 6c49481a2f..6cf0e11e3b 100644 > --- a/fs/squashfs/xz_wrapper.c > +++ b/fs/squashfs/xz_wrapper.c > @@ -120,8 +120,7 @@ static int squashfs_xz_uncompress(struct squashfs_sb_info *msblk, void *strm, > struct bio *bio, int offset, int length, > struct squashfs_page_actor *output) > { > - struct bvec_iter_all iter_all = {}; > - struct bio_vec *bvec = bvec_init_iter_all(&iter_all); > + struct bvec_iter_all iter; > int total = 0, error = 0; > struct squashfs_xz *stream = strm; > > @@ -136,26 +135,28 @@ static int squashfs_xz_uncompress(struct squashfs_sb_info *msblk, void *strm, > goto finish; > } > > + bvec_iter_all_init(&iter); > + bio_iter_all_advance(bio, &iter, offset); > + > for (;;) { > enum xz_ret xz_err; > > if (stream->buf.in_pos == stream->buf.in_size) { > - const void *data; > - int avail; > + struct bio_vec bvec = bio_iter_all_peek(bio, &iter); > + unsigned avail = min_t(unsigned, length, bvec.bv_len); > > - if (!bio_next_segment(bio, &iter_all)) { > + if (iter.idx >= bio->bi_vcnt) { > /* XZ_STREAM_END must be reached. */ > error = -EIO; > break; > } > > - avail = min(length, ((int)bvec->bv_len) - offset); > - data = bvec_virt(bvec); > length -= avail; > - stream->buf.in = data + offset; > + stream->buf.in = bvec_virt(&bvec); > stream->buf.in_size = avail; > stream->buf.in_pos = 0; > - offset = 0; > + > + bio_iter_all_advance(bio, &iter, avail); > } > > if (stream->buf.out_pos == stream->buf.out_size) { > diff --git a/fs/squashfs/zlib_wrapper.c b/fs/squashfs/zlib_wrapper.c > index cbb7afe7bc..981ca5e410 100644 > --- a/fs/squashfs/zlib_wrapper.c > +++ b/fs/squashfs/zlib_wrapper.c > @@ -53,8 +53,7 @@ static int zlib_uncompress(struct squashfs_sb_info *msblk, void *strm, > struct bio *bio, int offset, int length, > struct squashfs_page_actor *output) > { > - struct bvec_iter_all iter_all = {}; > - struct bio_vec *bvec = bvec_init_iter_all(&iter_all); > + struct bvec_iter_all iter; > int zlib_init = 0, error = 0; > z_stream *stream = strm; > > @@ -67,25 +66,28 @@ static int zlib_uncompress(struct squashfs_sb_info *msblk, void *strm, > goto finish; > } > > + bvec_iter_all_init(&iter); > + bio_iter_all_advance(bio, &iter, offset); > + > for (;;) { > int zlib_err; > > if (stream->avail_in == 0) { > - const void *data; > + struct bio_vec bvec = bio_iter_all_peek(bio, &iter); > int avail; > > - if (!bio_next_segment(bio, &iter_all)) { > + if (iter.idx >= bio->bi_vcnt) { It is bad to expose iterator/bio internal here, probably you can return bool from bio_iter_all_peek() by moveing the check into helper, meantime passing bvec reference to bio_iter_all_peek(). > /* Z_STREAM_END must be reached. */ > error = -EIO; > break; > } > > - avail = min(length, ((int)bvec->bv_len) - offset); > - data = bvec_virt(bvec); > + avail = min_t(unsigned, length, bvec.bv_len); > length -= avail; > - stream->next_in = data + offset; > + stream->next_in = bvec_virt(&bvec); > stream->avail_in = avail; > - offset = 0; > + > + bio_iter_all_advance(bio, &iter, avail); > } > > if (stream->avail_out == 0) { > diff --git a/fs/squashfs/zstd_wrapper.c b/fs/squashfs/zstd_wrapper.c > index 0e407c4d8b..658e5d462a 100644 > --- a/fs/squashfs/zstd_wrapper.c > +++ b/fs/squashfs/zstd_wrapper.c > @@ -68,8 +68,7 @@ static int zstd_uncompress(struct squashfs_sb_info *msblk, void *strm, > int error = 0; > zstd_in_buffer in_buf = { NULL, 0, 0 }; > zstd_out_buffer out_buf = { NULL, 0, 0 }; > - struct bvec_iter_all iter_all = {}; > - struct bio_vec *bvec = bvec_init_iter_all(&iter_all); > + struct bvec_iter_all iter; > > stream = zstd_init_dstream(wksp->window_size, wksp->mem, wksp->mem_size); > > @@ -85,25 +84,27 @@ static int zstd_uncompress(struct squashfs_sb_info *msblk, void *strm, > goto finish; > } > > + bvec_iter_all_init(&iter); > + bio_iter_all_advance(bio, &iter, offset); > + > for (;;) { > size_t zstd_err; > > if (in_buf.pos == in_buf.size) { > - const void *data; > - int avail; > + struct bio_vec bvec = bio_iter_all_peek(bio, &iter); > + unsigned avail = min_t(unsigned, length, bvec.bv_len); > > - if (!bio_next_segment(bio, &iter_all)) { > + if (iter.idx >= bio->bi_vcnt) { Same with above. > error = -EIO; > break; > } > > - avail = min(length, ((int)bvec->bv_len) - offset); > - data = bvec_virt(bvec); > length -= avail; > - in_buf.src = data + offset; > + in_buf.src = bvec_virt(&bvec); > in_buf.size = avail; > in_buf.pos = 0; > - offset = 0; > + > + bio_iter_all_advance(bio, &iter, avail); > } > > if (out_buf.pos == out_buf.size) { > diff --git a/include/linux/bio.h b/include/linux/bio.h > index 8588bcfbc6..39d8a1cc6e 100644 > --- a/include/linux/bio.h > +++ b/include/linux/bio.h > @@ -78,22 +78,34 @@ static inline void *bio_data(struct bio *bio) > return NULL; > } > > -static inline bool bio_next_segment(const struct bio *bio, > - struct bvec_iter_all *iter) > +static inline struct bio_vec bio_iter_all_peek(const struct bio *bio, > + struct bvec_iter_all *iter) > { > - if (iter->idx >= bio->bi_vcnt) > - return false; > + return bvec_iter_all_peek(bio->bi_io_vec, iter); > +} > > - bvec_advance(&bio->bi_io_vec[iter->idx], iter); > - return true; > +static inline void bio_iter_all_advance(const struct bio *bio, > + struct bvec_iter_all *iter, > + unsigned bytes) > +{ > + bvec_iter_all_advance(bio->bi_io_vec, iter, bytes); > } > > +#define bio_for_each_segment_all_continue(bvl, bio, iter) \ > + for (; \ > + iter.idx < bio->bi_vcnt && \ > + ((bvl = bio_iter_all_peek(bio, &iter)), true); \ > + bio_iter_all_advance((bio), &iter, bvl.bv_len)) > + > /* > * drivers should _never_ use the all version - the bio may have been split > * before it got to the driver and the driver won't own all of it > */ > -#define bio_for_each_segment_all(bvl, bio, iter) \ > - for (bvl = bvec_init_iter_all(&iter); bio_next_segment((bio), &iter); ) > +#define bio_for_each_segment_all(bvl, bio, iter) \ > + for (bvec_iter_all_init(&iter); \ > + iter.idx < (bio)->bi_vcnt && \ > + ((bvl = bio_iter_all_peek((bio), &iter)), true); \ The check of 'iter.idx < (bio)->bi_vcnt' could be moved to bio_iter_all_peek(), meantime passing &bvl to bio_iter_all_peek(). > + bio_iter_all_advance((bio), &iter, bvl.bv_len)) > > static inline void bio_advance_iter(const struct bio *bio, > struct bvec_iter *iter, unsigned int bytes) > diff --git a/include/linux/bvec.h b/include/linux/bvec.h > index 555aae5448..635fb54143 100644 > --- a/include/linux/bvec.h > +++ b/include/linux/bvec.h > @@ -85,12 +85,6 @@ struct bvec_iter { > current bvec */ > } __packed; > > -struct bvec_iter_all { > - struct bio_vec bv; > - int idx; > - unsigned done; > -}; > - > /* > * various member access, note that bio_data should of course not be used > * on highmem page vectors > @@ -184,7 +178,10 @@ static inline void bvec_iter_advance_single(const struct bio_vec *bv, > ((bvl = bvec_iter_bvec((bio_vec), (iter))), 1); \ > bvec_iter_advance_single((bio_vec), &(iter), (bvl).bv_len)) > > -/* for iterating one bio from start to end */ > +/* > + * bvec_iter_all: for advancing over a bio as it was originally created, but > + * with the usual bio_for_each_segment interface - nonstandard, do not use: > + */ > #define BVEC_ITER_ALL_INIT (struct bvec_iter) \ > { \ > .bi_sector = 0, \ > @@ -193,33 +190,45 @@ static inline void bvec_iter_advance_single(const struct bio_vec *bv, > .bi_bvec_done = 0, \ > } > > -static inline struct bio_vec *bvec_init_iter_all(struct bvec_iter_all *iter_all) > +/* > + * bvec_iter_all: for advancing over individual pages in a bio, as it was when > + * it was first created: > + */ > +struct bvec_iter_all { > + int idx; > + unsigned done; > +}; > + > +static inline void bvec_iter_all_init(struct bvec_iter_all *iter_all) > { > iter_all->done = 0; > iter_all->idx = 0; > +} > > - return &iter_all->bv; > +static inline struct bio_vec bvec_iter_all_peek(const struct bio_vec *bvec, > + struct bvec_iter_all *iter) > +{ > + struct bio_vec bv = bvec[iter->idx]; > + > + bv.bv_offset += iter->done; > + bv.bv_len -= iter->done; > + > + bv.bv_page += bv.bv_offset >> PAGE_SHIFT; > + bv.bv_offset &= ~PAGE_MASK; > + bv.bv_len = min_t(unsigned, PAGE_SIZE - bv.bv_offset, bv.bv_len); > + > + return bv; > } Here I think compiler optimizes the last bvec copy by using external bvec directly, and I run fio/t/io_uring and not find IOPS degrade. As mentioned above, it can be done explicitly by passing &bvl to bvec_iter_all_peek(), then user of bvec_iter_all_peek() can know if one real bvec is peeked, and check of 'if (iter.idx >= bio->bi_vcnt)' can be avoided in external callers. Otherwise, this patch looks fine for me. thanks, Ming