Received: by 2002:ad5:474a:0:0:0:0:0 with SMTP id i10csp2430395imu; Sun, 18 Nov 2018 23:53:41 -0800 (PST) X-Google-Smtp-Source: AJdET5dhCgf7d6jUbT2nl8QP3RDLSL0Yfc0DY31rcxPzb80d6mSV9+wh8smvsrExF/DCm9pc94VG X-Received: by 2002:a63:4926:: with SMTP id w38mr18526903pga.353.1542614021595; Sun, 18 Nov 2018 23:53:41 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1542614021; cv=none; d=google.com; s=arc-20160816; b=DqkJrzgy4+iNVGTLgE51Y6uvICWkMPjEqsSbI4fyX1qMNR1lpp0OfnYY0nqtP3JNMk vzsi4Nf03qoFBp5HUz4jpHWjKDGZ8DGYKyZD9sVn9LcluTgHYrhhYpbT2o/qUyqlsarn g048F4hG4mXQbW9qrvKeN6MKwJ77jP+vSWthIhBWMsP84vEX3sstxEnnjKg/8WVjRs54 vE9qc656WEywZwtK6ftGVDqJqcdvr719Z8V5CpsowkbeTT4NKpBWNiW6XWio89HZ0JZH 6okFJ3QsZLR4X9Z9LmBIh4jGE/t/A43JO6iHz0HeExzFGOXGbxMAj6TtaPLfCOIPzsbs 0gGA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:user-agent:in-reply-to :content-disposition:mime-version:references:message-id:subject:cc :to:from:date; bh=g3z/FM8+EQ3T3qe8WAxD50T9rEfyHf5ZcQJVGmM8rnE=; b=jFtn4A4t6DNVl0o+GKVNHzNBCO1TZtwGz9lYvfDOCzGSXJqMLwdYN54iZLPgr3Qud6 gxWU0s9dfP/bSWCycnTB0yMubhwn5QcqFjqaD3N6yupJJ5na3wv3zof8gzNoRSvElqO/ REG2z6MN737T5VP1wILfqv6eQYffrh1fD9qwOkfuOfqX9ny6lw4a+tlhEcwQG1m5HKfh By09gRDroECGb7KbS5nUncH3IdnCwmIWAJ165MsFaFdtTik3zuz0yw6Gq7FnjGHP6IAv beC1uXzk7+Ym/QRsE9G4SfY0b0kMSaoPxhQ9coyUglAGXSsabcdfl0DS2kX9sFs7zY1a Xxmg== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=redhat.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id q8si38872466pgk.40.2018.11.18.23.53.27; Sun, 18 Nov 2018 23:53:41 -0800 (PST) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=redhat.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1726529AbeKSSNp (ORCPT + 99 others); Mon, 19 Nov 2018 13:13:45 -0500 Received: from mx1.redhat.com ([209.132.183.28]:51456 "EHLO mx1.redhat.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726348AbeKSSNp (ORCPT ); Mon, 19 Nov 2018 13:13:45 -0500 Received: from smtp.corp.redhat.com (int-mx03.intmail.prod.int.phx2.redhat.com [10.5.11.13]) (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits)) (No client certificate requested) by mx1.redhat.com (Postfix) with ESMTPS id 0E3D13002061; Mon, 19 Nov 2018 07:50:55 +0000 (UTC) Received: from ming.t460p (ovpn-8-26.pek2.redhat.com [10.72.8.26]) by smtp.corp.redhat.com (Postfix) with ESMTPS id C00CE608E7; Mon, 19 Nov 2018 07:50:30 +0000 (UTC) Date: Mon, 19 Nov 2018 15:50:25 +0800 From: Ming Lei To: Omar Sandoval Cc: Jens Axboe , linux-block@vger.kernel.org, linux-kernel@vger.kernel.org, linux-mm@kvack.org, Dave Chinner , Kent Overstreet , Mike Snitzer , dm-devel@redhat.com, Alexander Viro , linux-fsdevel@vger.kernel.org, Shaohua Li , linux-raid@vger.kernel.org, linux-erofs@lists.ozlabs.org, David Sterba , linux-btrfs@vger.kernel.org, "Darrick J . Wong" , linux-xfs@vger.kernel.org, Gao Xiang , Christoph Hellwig , Theodore Ts'o , linux-ext4@vger.kernel.org, Coly Li , linux-bcache@vger.kernel.org, Boaz Harrosh , Bob Peterson , cluster-devel@redhat.com Subject: Re: [PATCH V10 03/19] block: use bio_for_each_bvec() to compute multi-page bvec count Message-ID: <20181119075024.GA16519@ming.t460p> References: <20181115085306.9910-1-ming.lei@redhat.com> <20181115085306.9910-4-ming.lei@redhat.com> <20181115202028.GC9348@vader> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20181115202028.GC9348@vader> User-Agent: Mutt/1.9.1 (2017-09-22) X-Scanned-By: MIMEDefang 2.79 on 10.5.11.13 X-Greylist: Sender IP whitelisted, not delayed by milter-greylist-4.5.16 (mx1.redhat.com [10.5.110.47]); Mon, 19 Nov 2018 07:50:55 +0000 (UTC) Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Thu, Nov 15, 2018 at 12:20:28PM -0800, Omar Sandoval wrote: > On Thu, Nov 15, 2018 at 04:52:50PM +0800, Ming Lei wrote: > > First it is more efficient to use bio_for_each_bvec() in both > > blk_bio_segment_split() and __blk_recalc_rq_segments() to compute how > > many multi-page bvecs there are in the bio. > > > > Secondly once bio_for_each_bvec() is used, the bvec may need to be > > splitted because its length can be very longer than max segment size, > > so we have to split the big bvec into several segments. > > > > Thirdly when splitting multi-page bvec into segments, the max segment > > limit may be reached, so the bio split need to be considered under > > this situation too. > > > > Cc: Dave Chinner > > Cc: Kent Overstreet > > Cc: Mike Snitzer > > Cc: dm-devel@redhat.com > > Cc: Alexander Viro > > Cc: linux-fsdevel@vger.kernel.org > > Cc: Shaohua Li > > Cc: linux-raid@vger.kernel.org > > Cc: linux-erofs@lists.ozlabs.org > > Cc: David Sterba > > Cc: linux-btrfs@vger.kernel.org > > Cc: Darrick J. Wong > > Cc: linux-xfs@vger.kernel.org > > Cc: Gao Xiang > > Cc: Christoph Hellwig > > Cc: Theodore Ts'o > > Cc: linux-ext4@vger.kernel.org > > Cc: Coly Li > > Cc: linux-bcache@vger.kernel.org > > Cc: Boaz Harrosh > > Cc: Bob Peterson > > Cc: cluster-devel@redhat.com > > Signed-off-by: Ming Lei > > --- > > block/blk-merge.c | 90 ++++++++++++++++++++++++++++++++++++++++++++++--------- > > 1 file changed, 76 insertions(+), 14 deletions(-) > > > > diff --git a/block/blk-merge.c b/block/blk-merge.c > > index 91b2af332a84..6f7deb94a23f 100644 > > --- a/block/blk-merge.c > > +++ b/block/blk-merge.c > > @@ -160,6 +160,62 @@ static inline unsigned get_max_io_size(struct request_queue *q, > > return sectors; > > } > > > > +/* > > + * Split the bvec @bv into segments, and update all kinds of > > + * variables. > > + */ > > +static bool bvec_split_segs(struct request_queue *q, struct bio_vec *bv, > > + unsigned *nsegs, unsigned *last_seg_size, > > + unsigned *front_seg_size, unsigned *sectors) > > +{ > > + bool need_split = false; > > + unsigned len = bv->bv_len; > > + unsigned total_len = 0; > > + unsigned new_nsegs = 0, seg_size = 0; > > "unsigned int" here and everywhere else. > > > + if ((*nsegs >= queue_max_segments(q)) || !len) > > + return need_split; > > + > > + /* > > + * Multipage bvec may be too big to hold in one segment, > > + * so the current bvec has to be splitted as multiple > > + * segments. > > + */ > > + while (new_nsegs + *nsegs < queue_max_segments(q)) { > > + seg_size = min(queue_max_segment_size(q), len); > > + > > + new_nsegs++; > > + total_len += seg_size; > > + len -= seg_size; > > + > > + if ((queue_virt_boundary(q) && ((bv->bv_offset + > > + total_len) & queue_virt_boundary(q))) || !len) > > + break; > > Checking queue_virt_boundary(q) != 0 is superfluous, and the len check > could just control the loop, i.e., > > while (len && new_nsegs + *nsegs < queue_max_segments(q)) { > seg_size = min(queue_max_segment_size(q), len); > > new_nsegs++; > total_len += seg_size; > len -= seg_size; > > if ((bv->bv_offset + total_len) & queue_virt_boundary(q)) > break; > } > > And if you rewrite it this way, I _think_ you can get rid of this > special case: > > if ((*nsegs >= queue_max_segments(q)) || !len) > return need_split; > > above. Good point, will do in next version. > > > + } > > + > > + /* split in the middle of the bvec */ > > + if (len) > > + need_split = true; > > need_split is unnecessary, just return len != 0. OK. > > > + > > + /* update front segment size */ > > + if (!*nsegs) { > > + unsigned first_seg_size = seg_size; > > + > > + if (new_nsegs > 1) > > + first_seg_size = queue_max_segment_size(q); > > + if (*front_seg_size < first_seg_size) > > + *front_seg_size = first_seg_size; > > + } > > + > > + /* update other varibles */ > > + *last_seg_size = seg_size; > > + *nsegs += new_nsegs; > > + if (sectors) > > + *sectors += total_len >> 9; > > + > > + return need_split; > > +} > > + > > static struct bio *blk_bio_segment_split(struct request_queue *q, > > struct bio *bio, > > struct bio_set *bs, > > @@ -173,7 +229,7 @@ static struct bio *blk_bio_segment_split(struct request_queue *q, > > struct bio *new = NULL; > > const unsigned max_sectors = get_max_io_size(q, bio); > > > > - bio_for_each_segment(bv, bio, iter) { > > + bio_for_each_bvec(bv, bio, iter) { > > /* > > * If the queue doesn't support SG gaps and adding this > > * offset would create a gap, disallow it. > > @@ -188,8 +244,12 @@ static struct bio *blk_bio_segment_split(struct request_queue *q, > > */ > > if (nsegs < queue_max_segments(q) && > > sectors < max_sectors) { > > - nsegs++; > > - sectors = max_sectors; > > + /* split in the middle of bvec */ > > + bv.bv_len = (max_sectors - sectors) << 9; > > + bvec_split_segs(q, &bv, &nsegs, > > + &seg_size, > > + &front_seg_size, > > + §ors); > > } > > goto split; > > } > > @@ -214,11 +274,12 @@ static struct bio *blk_bio_segment_split(struct request_queue *q, > > if (nsegs == 1 && seg_size > front_seg_size) > > front_seg_size = seg_size; > > Hm, do we still need to check this here now that we're updating > front_seg_size inside of bvec_split_segs()? Right, the check & update can be removed. > > > > > - nsegs++; > > bvprv = bv; > > bvprvp = &bvprv; > > - seg_size = bv.bv_len; > > - sectors += bv.bv_len >> 9; > > + > > + if (bvec_split_segs(q, &bv, &nsegs, &seg_size, > > + &front_seg_size, §ors)) > > What happened to the indent alignment here? Will fix it in next version. > > > + goto split; > > > > } > > > > @@ -296,6 +357,7 @@ static unsigned int __blk_recalc_rq_segments(struct request_queue *q, > > struct bio_vec bv, bvprv = { NULL }; > > int cluster, prev = 0; > > unsigned int seg_size, nr_phys_segs; > > + unsigned front_seg_size = bio->bi_seg_front_size; > > struct bio *fbio, *bbio; > > struct bvec_iter iter; > > > > @@ -316,7 +378,7 @@ static unsigned int __blk_recalc_rq_segments(struct request_queue *q, > > seg_size = 0; > > nr_phys_segs = 0; > > for_each_bio(bio) { > > - bio_for_each_segment(bv, bio, iter) { > > + bio_for_each_bvec(bv, bio, iter) { > > /* > > * If SG merging is disabled, each bio vector is > > * a segment > > @@ -336,20 +398,20 @@ static unsigned int __blk_recalc_rq_segments(struct request_queue *q, > > continue; > > } > > new_segment: > > - if (nr_phys_segs == 1 && seg_size > > > - fbio->bi_seg_front_size) > > - fbio->bi_seg_front_size = seg_size; > > + if (nr_phys_segs == 1 && seg_size > front_seg_size) > > + front_seg_size = seg_size; > > Same comment as in blk_bio_segment_split(), do we still need to check > this if we're updating front_seg_size in bvec_split_segs()? I think we can remove it too. Thanks, Ming