Received: by 2002:a25:683:0:0:0:0:0 with SMTP id 125csp1094153ybg; Tue, 2 Jun 2020 00:54:18 -0700 (PDT) X-Google-Smtp-Source: ABdhPJx2+D3AAb/YUNxl6QnpBYHfd5FWmU/kRky8h5ikkq6q6XrdQJrD+WD2yQcEdg/DXcFkRFyp X-Received: by 2002:a17:906:4b50:: with SMTP id j16mr10650425ejv.415.1591084458618; Tue, 02 Jun 2020 00:54:18 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1591084458; cv=none; d=google.com; s=arc-20160816; b=biP2p7MzlX4qqIdksvf9RDBfYYCzQvdeLBtGqG1+l/QrRkgejWk+Y2VW6fDQaysi73 3OGuQzYGoGWdFpNeqwu/xdubsOzj+5RqIQ0GXhq6UsFB0prr6nATxwUskTqknSBGIJsv jl7m2YA+mx254gG10w/n/jok9ryHYTmFgW0QhLlUMAXy4RWBgu3pwehkep50NoXeNHl7 yCtkPkt1sh2cvUjap40dF+WGCgrb5vfe03KX4RXYZxyy0k6r4F2fEe3mhF3K4EhRz84m JnGcHwF+9nBu//C315SzBwpdVwY1HWvVwYOT3pMLLnkt7cRoIJnyD9aZ80vqcaK6HHtB so2A== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:in-reply-to:content-disposition :mime-version:references:message-id:subject:cc:to:from:date; bh=R0y5itrZTxOYcFZpx6szXtUpYX5WvCpqiFTWBtLVZ7s=; b=helS0ky3zP6Lph5630bANbL9B/Ejn0dDmxUNKzyodu6xtQ6DIywnsophheSNsMGEeY PcyYcR53nob+BTZF9SPTon96UTsutYF1UvAOlCVGJxwsX14u3BlRlQSOiEoUTU0F9VhC GXsGqk4QUlY4CRB2d0j1E+/XEX6pfso47wf1yn6h/uQSn37lsVGv6OQnMzNgc/7q7Vj9 a7myQlLEoEGFdjRUvnmLZ6d9ivOIN79MP3Lmca5j0v4T8EtyRQYQF+dR/J1gHRJVhyyf M4hi+ZV+6L0wR2NqC8DpaOiLzBLw+8jlBbywinXF3CSlKY0S+7V1+Ze6FMlTLaLg3yFh 7IJA== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id q19si991329ejt.484.2020.06.02.00.53.55; Tue, 02 Jun 2020 00:54:18 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1726177AbgFBHv2 (ORCPT + 99 others); Tue, 2 Jun 2020 03:51:28 -0400 Received: from mx2.suse.de ([195.135.220.15]:52438 "EHLO mx2.suse.de" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1725819AbgFBHv2 (ORCPT ); Tue, 2 Jun 2020 03:51:28 -0400 X-Virus-Scanned: by amavisd-new at test-mx.suse.de Received: from relay2.suse.de (unknown [195.135.220.254]) by mx2.suse.de (Postfix) with ESMTP id B4E5BAB64; Tue, 2 Jun 2020 07:51:28 +0000 (UTC) Date: Tue, 2 Jun 2020 09:51:24 +0200 From: Daniel Wagner To: Dongli Zhang Cc: linux-block@vger.kernel.org, axboe@kernel.dk, hare@suse.de, ming.lei@redhat.com, linux-kernel@vger.kernel.org Subject: Re: [PATCH 1/1] blk-mq: get ctx in order to handle BLK_MQ_S_INACTIVE in blk_mq_get_tag() Message-ID: <20200602075124.3igsbsgbzf3varib@beryllium.lan> References: <20200602061749.32029-1-dongli.zhang@oracle.com> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20200602061749.32029-1-dongli.zhang@oracle.com> Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Mon, Jun 01, 2020 at 11:17:49PM -0700, Dongli Zhang wrote: > When scheduler is set, we hit below page fault when we offline cpu. > > [ 1061.007725] BUG: kernel NULL pointer dereference, address: 0000000000000040 > [ 1061.008710] #PF: supervisor read access in kernel mode > [ 1061.009492] #PF: error_code(0x0000) - not-present page > [ 1061.010241] PGD 0 P4D 0 > [ 1061.010614] Oops: 0000 [#1] SMP PTI > [ 1061.011130] CPU: 0 PID: 122 Comm: kworker/0:1H Not tainted 5.7.0-rc7+ #2' > ... ... > [ 1061.013760] Workqueue: kblockd blk_mq_run_work_fn > [ 1061.014446] RIP: 0010:blk_mq_put_tag+0xf/0x30 > ... ... > [ 1061.017726] RSP: 0018:ffffa5c18037fc70 EFLAGS: 00010287 > [ 1061.018475] RAX: 0000000000000000 RBX: ffffa5c18037fcf0 RCX: 0000000000000004 > [ 1061.019507] RDX: 0000000000000000 RSI: 0000000000000000 RDI: ffff911535dc1180 > ... ... > [ 1061.028454] Call Trace: > [ 1061.029307] blk_mq_get_tag+0x26e/0x280 > [ 1061.029866] ? wait_woken+0x80/0x80 > [ 1061.030378] blk_mq_get_driver_tag+0x99/0x110 > [ 1061.031009] blk_mq_dispatch_rq_list+0x107/0x5e0 > [ 1061.031672] ? elv_rb_del+0x1a/0x30 > [ 1061.032178] blk_mq_do_dispatch_sched+0xe2/0x130 > [ 1061.032844] __blk_mq_sched_dispatch_requests+0xcc/0x150 > [ 1061.033638] blk_mq_sched_dispatch_requests+0x2b/0x50 > [ 1061.034239] __blk_mq_run_hw_queue+0x75/0x110 > [ 1061.034867] process_one_work+0x15c/0x370 > [ 1061.035450] worker_thread+0x44/0x3d0 > [ 1061.035980] kthread+0xf3/0x130 > [ 1061.036440] ? max_active_store+0x80/0x80 > [ 1061.037018] ? kthread_bind+0x10/0x10 > [ 1061.037554] ret_from_fork+0x35/0x40 > [ 1061.038073] Modules linked in: > [ 1061.038543] CR2: 0000000000000040 > [ 1061.038962] ---[ end trace d20e1df7d028e69f ]--- > > This is because blk_mq_get_driver_tag() would be used to allocate tag once > scheduler (e.g., mq-deadline) is set. However, in order to handle > BLK_MQ_S_INACTIVE in blk_mq_get_tag(), we need to set data->ctx for > blk_mq_put_tag(). > > Fixes: bf0beec0607db3c6 ("blk-mq: drain I/O when all CPUs in a hctx are offline") > Signed-off-by: Dongli Zhang Reviewed-by: Daniel Wagner