Received: by 2002:a05:6a10:5bc5:0:0:0:0 with SMTP id os5csp885873pxb; Tue, 19 Oct 2021 15:23:14 -0700 (PDT) X-Google-Smtp-Source: ABdhPJxzPPj/Nza8eRWqhlxLlbiaJMQjwxQMZyhz/KXMPcZntrTA3uonVatPe0w3IbvIsw6vF5UE X-Received: by 2002:a05:6a00:1a46:b0:44d:7a3f:2043 with SMTP id h6-20020a056a001a4600b0044d7a3f2043mr2531422pfv.56.1634682193987; Tue, 19 Oct 2021 15:23:13 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1634682193; cv=none; d=google.com; s=arc-20160816; b=rkZ1pUaE8SpoIYsbBOsg42qWxlsZrCtOWyGozrv4AJhUTJzzUOR1zw0jjBLBa6onMO 4vIgtcbnwtiCfqfiDnmaZznWFli9g81jc3bXZYy8v5uE8kT2QkVufxWwb/UaWUBMKmHX AbRidsJBx7sO/Et/T4bB55JfMvCzL3W7tQIMbjEewLg6VGUykpEjNcgG9ekvw//cK1JC 0p6XLFuBlKWpuaEWoehwUab1Zu9cwXUbSYtRbU3Izl0YQ/llvrKUxns3Uub0v6iXGULc lSKzx+rOp4S4itjZ89ihVpruQGmn3klb1AbEwtbomB4aPw5G9JJi5TguVZaKZsagqSVf aBIA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:content-language :in-reply-to:mime-version:user-agent:date:message-id:from:cc :references:to:subject; bh=6TwTP43YabzsIQbDofBw8F623M2Eokqd0NU5JFCt3i8=; b=lCG5TDO9mp8BG/Q3HER07eVB5g7wHUTIfY1ilPyOgXuw/ITWxE8GA4ehS6grQ6o4QV NQsEac2bAgvkFinn/DCt5KTH2n/SQcl+v4WhCEO9e3zs+7dRd2JuQQJKUntM+7LHuIlg yMjpthZqsMWCMYqTwWgvAmiL8wiHnxg2qsZuWZlWmshbVjS2lm8XCCmN/Kc5BXIK3vkN 21DHsnF//BYdsqE84XyxiFwHj/86ucMlgN58lOjBnjw6QNCg72+CZefOXNFLZZCZuDTn aWbukT9c+j+U/pn9YwiCvkD/7AyLagG8B+E88ivjVy/gdwcBOIjzH3QBNZxg9jwgvwva 1F6w== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=collabora.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id a12si460433pfh.148.2021.10.19.15.22.58; Tue, 19 Oct 2021 15:23:13 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=collabora.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S229757AbhJSWXZ (ORCPT + 99 others); Tue, 19 Oct 2021 18:23:25 -0400 Received: from bhuna.collabora.co.uk ([46.235.227.227]:52932 "EHLO bhuna.collabora.co.uk" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S229548AbhJSWXY (ORCPT ); Tue, 19 Oct 2021 18:23:24 -0400 Received: from [127.0.0.1] (localhost [127.0.0.1]) (Authenticated sender: gtucker) with ESMTPSA id 6A7951F43E76 Subject: Re: [PATCH 06/14] block: store elevator state in request To: Jens Axboe References: <20211017013748.76461-1-axboe@kernel.dk> <20211017013748.76461-7-axboe@kernel.dk> Cc: linux-block@vger.kernel.org, Mark Brown , "linux-kernel@vger.kernel.org" , "kernelci@groups.io" From: Guillaume Tucker Message-ID: <1a08141a-407f-0fda-4b27-e9dd350c6e79@collabora.com> Date: Tue, 19 Oct 2021 23:21:07 +0100 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:78.0) Gecko/20100101 Thunderbird/78.14.0 MIME-Version: 1.0 In-Reply-To: <20211017013748.76461-7-axboe@kernel.dk> Content-Type: text/plain; charset=utf-8 Content-Language: en-US Content-Transfer-Encoding: 7bit Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Hi Jens, On 17/10/2021 02:37, Jens Axboe wrote: > Add an rq private RQF_ELV flag, which tells the block layer that this > request was initialized on a queue that has an IO scheduler attached. > This allows for faster checking in the fast path, rather than having to > deference rq->q later on. > > Elevator switching does full quiesce of the queue before detaching an > IO scheduler, so it's safe to cache this in the request itself. A kernelci.org automated bisection found that this patch introduced a regression in next-20211019 with a NULL pointer dereference, which only seems to be affecting QEMU but across all architectures. More details about the regression can be found here: https://linux.kernelci.org/test/job/next/branch/master/kernel/next-20211019/plan/baseline/ https://linux.kernelci.org/test/case/id/616ea20eb7104071c43358ea/ See also all the test jobs involved in the automated bisection: https://lava.collabora.co.uk/scheduler/device_type/qemu?dt_search=bisection-287 If you do send a fix, please include this trailer: Reported-by: "kernelci.org bot" Please let us know if this seems like a valid bisection result and if you need any help to debug the issue or try a fix. Best wishes, Guillaume GitHub: https://github.com/kernelci/kernelci-project/issues/68 > Signed-off-by: Jens Axboe > --- > block/blk-mq-sched.h | 27 ++++++++++++++++----------- > block/blk-mq.c | 20 +++++++++++--------- > include/linux/blk-mq.h | 2 ++ > 3 files changed, 29 insertions(+), 20 deletions(-) > > diff --git a/block/blk-mq-sched.h b/block/blk-mq-sched.h > index fe252278ed9a..98836106b25f 100644 > --- a/block/blk-mq-sched.h > +++ b/block/blk-mq-sched.h > @@ -56,29 +56,34 @@ static inline bool > blk_mq_sched_allow_merge(struct request_queue *q, struct request *rq, > struct bio *bio) > { > - struct elevator_queue *e = q->elevator; > - > - if (e && e->type->ops.allow_merge) > - return e->type->ops.allow_merge(q, rq, bio); > + if (rq->rq_flags & RQF_ELV) { > + struct elevator_queue *e = q->elevator; > > + if (e->type->ops.allow_merge) > + return e->type->ops.allow_merge(q, rq, bio); > + } > return true; > } > > static inline void blk_mq_sched_completed_request(struct request *rq, u64 now) > { > - struct elevator_queue *e = rq->q->elevator; > + if (rq->rq_flags & RQF_ELV) { > + struct elevator_queue *e = rq->q->elevator; > > - if (e && e->type->ops.completed_request) > - e->type->ops.completed_request(rq, now); > + if (e->type->ops.completed_request) > + e->type->ops.completed_request(rq, now); > + } > } > > static inline void blk_mq_sched_requeue_request(struct request *rq) > { > - struct request_queue *q = rq->q; > - struct elevator_queue *e = q->elevator; > + if (rq->rq_flags & RQF_ELV) { > + struct request_queue *q = rq->q; > + struct elevator_queue *e = q->elevator; > > - if ((rq->rq_flags & RQF_ELVPRIV) && e && e->type->ops.requeue_request) > - e->type->ops.requeue_request(rq); > + if ((rq->rq_flags & RQF_ELVPRIV) && e->type->ops.requeue_request) > + e->type->ops.requeue_request(rq); > + } > } > > static inline bool blk_mq_sched_has_work(struct blk_mq_hw_ctx *hctx) > diff --git a/block/blk-mq.c b/block/blk-mq.c > index fa5b12200404..5d22c228f6df 100644 > --- a/block/blk-mq.c > +++ b/block/blk-mq.c > @@ -299,7 +299,7 @@ void blk_mq_wake_waiters(struct request_queue *q) > */ > static inline bool blk_mq_need_time_stamp(struct request *rq) > { > - return (rq->rq_flags & (RQF_IO_STAT | RQF_STATS)) || rq->q->elevator; > + return (rq->rq_flags & (RQF_IO_STAT | RQF_STATS | RQF_ELV)); > } > > static struct request *blk_mq_rq_ctx_init(struct blk_mq_alloc_data *data, > @@ -309,9 +309,11 @@ static struct request *blk_mq_rq_ctx_init(struct blk_mq_alloc_data *data, > struct request *rq = tags->static_rqs[tag]; > > if (data->q->elevator) { > + rq->rq_flags = RQF_ELV; > rq->tag = BLK_MQ_NO_TAG; > rq->internal_tag = tag; > } else { > + rq->rq_flags = 0; > rq->tag = tag; > rq->internal_tag = BLK_MQ_NO_TAG; > } > @@ -320,7 +322,6 @@ static struct request *blk_mq_rq_ctx_init(struct blk_mq_alloc_data *data, > rq->q = data->q; > rq->mq_ctx = data->ctx; > rq->mq_hctx = data->hctx; > - rq->rq_flags = 0; > rq->cmd_flags = data->cmd_flags; > if (data->flags & BLK_MQ_REQ_PM) > rq->rq_flags |= RQF_PM; > @@ -356,11 +357,11 @@ static struct request *blk_mq_rq_ctx_init(struct blk_mq_alloc_data *data, > data->ctx->rq_dispatched[op_is_sync(data->cmd_flags)]++; > refcount_set(&rq->ref, 1); > > - if (!op_is_flush(data->cmd_flags)) { > + if (!op_is_flush(data->cmd_flags) && (rq->rq_flags & RQF_ELV)) { > struct elevator_queue *e = data->q->elevator; > > rq->elv.icq = NULL; > - if (e && e->type->ops.prepare_request) { > + if (e->type->ops.prepare_request) { > if (e->type->icq_cache) > blk_mq_sched_assign_ioc(rq); > > @@ -575,12 +576,13 @@ static void __blk_mq_free_request(struct request *rq) > void blk_mq_free_request(struct request *rq) > { > struct request_queue *q = rq->q; > - struct elevator_queue *e = q->elevator; > struct blk_mq_ctx *ctx = rq->mq_ctx; > struct blk_mq_hw_ctx *hctx = rq->mq_hctx; > > - if (rq->rq_flags & RQF_ELVPRIV) { > - if (e && e->type->ops.finish_request) > + if (rq->rq_flags & (RQF_ELVPRIV | RQF_ELV)) { > + struct elevator_queue *e = q->elevator; > + > + if (e->type->ops.finish_request) > e->type->ops.finish_request(rq); > if (rq->elv.icq) { > put_io_context(rq->elv.icq->ioc); > @@ -2239,7 +2241,7 @@ static blk_status_t __blk_mq_try_issue_directly(struct blk_mq_hw_ctx *hctx, > goto insert; > } > > - if (q->elevator && !bypass_insert) > + if ((rq->rq_flags & RQF_ELV) && !bypass_insert) > goto insert; > > budget_token = blk_mq_get_dispatch_budget(q); > @@ -2475,7 +2477,7 @@ void blk_mq_submit_bio(struct bio *bio) > } > > blk_add_rq_to_plug(plug, rq); > - } else if (q->elevator) { > + } else if (rq->rq_flags & RQF_ELV) { > /* Insert the request at the IO scheduler queue */ > blk_mq_sched_insert_request(rq, false, true, true); > } else if (plug && !blk_queue_nomerges(q)) { > diff --git a/include/linux/blk-mq.h b/include/linux/blk-mq.h > index a9c1d0882550..3a399aa372b5 100644 > --- a/include/linux/blk-mq.h > +++ b/include/linux/blk-mq.h > @@ -55,6 +55,8 @@ typedef __u32 __bitwise req_flags_t; > #define RQF_MQ_POLL_SLEPT ((__force req_flags_t)(1 << 20)) > /* ->timeout has been called, don't expire again */ > #define RQF_TIMED_OUT ((__force req_flags_t)(1 << 21)) > +/* queue has elevator attached */ > +#define RQF_ELV ((__force req_flags_t)(1 << 22)) > > /* flags that prevent us from merging requests: */ > #define RQF_NOMERGE_FLAGS \ >