Received: by 2002:a25:1506:0:0:0:0:0 with SMTP id 6csp5532963ybv; Tue, 11 Feb 2020 18:01:13 -0800 (PST) X-Google-Smtp-Source: APXvYqxsecfqrSl7Nj4/xgXJJuHXYxCnwIjqawFYR5taisHOYH9gMNUMdeR3Lbse05HFvwr3sPFZ X-Received: by 2002:aca:3f54:: with SMTP id m81mr4604058oia.73.1581472873656; Tue, 11 Feb 2020 18:01:13 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1581472873; cv=none; d=google.com; s=arc-20160816; b=F6jjHgzVzZFon6pANRfcCzp7VF9rYIE70dYZi7fMmgYUZ/j5SAo08OzBhg1byFE7ZC vpEdoSIOHT87bMzfh6rWlA/OwLhiCwp3cVSxA2X6ilUwUvvB8m5n+3TihwChhmAh1XOy r+EgXGHF7+pMjBJdv68JVrpP3O+td9RWKvzz6RnqczGCl5RtXw1CK6Iq9wbg+alhI3Tn Y/VDp3tt6QdAfJdTddxpkEvugEAB4aomkAEq++/XbuDR4qyUrQTbLnwZo+kM87Vmh4vY Ktk9Y9F9t64cFRN5EfGCrjCzxMdMbh+3C5wNlISIZ4H1bpgjTwMj6bB1DFWvFmpJwoCB lRhw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:content-transfer-encoding:in-reply-to :mime-version:user-agent:date:message-id:from:references:cc:to :subject; bh=D2UqRjCBa5Jv6ScYGhWG28/bPTss91KF0aLpn+sXhWc=; b=zoBhiruIvWV6m+H6fasbZXxDLAxBlVzPOY1Ok8TmHl5K7cMiGTjnwBkZ/KHJ6lignj 3LM9AyyZ/HBPKNu6ct/4kdU2/Ml5zoUmsAsjL/BY2xtYQvNT5eyo1ANsxD08mTFlhqzJ tcLlwhTJctu72OkcOVpfbZdMdkEh2g7yUcNdA3LKiCXopVK3qj9W6WN22PFiYJti/GWb 07XRiKrqm/mbiI7nijtqMk5O/qhGMf6nR7AXe6dWM1ssnNKwZgI4PkybvNlMlDKA3yme BMAs0Nr9AIavVTuBlQjH9wooNP8AScy/EHf+Eb4tOSSspFC14gCZzUtar+VhKulzSv57 wM3Q== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id t10si2889709otp.310.2020.02.11.18.01.01; Tue, 11 Feb 2020 18:01:13 -0800 (PST) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1727602AbgBLCAm (ORCPT + 99 others); Tue, 11 Feb 2020 21:00:42 -0500 Received: from szxga05-in.huawei.com ([45.249.212.191]:9724 "EHLO huawei.com" rhost-flags-OK-OK-OK-FAIL) by vger.kernel.org with ESMTP id S1726968AbgBLCAm (ORCPT ); Tue, 11 Feb 2020 21:00:42 -0500 Received: from DGGEMS413-HUB.china.huawei.com (unknown [172.30.72.59]) by Forcepoint Email with ESMTP id CC7B9611207F6164D34C; Wed, 12 Feb 2020 10:00:39 +0800 (CST) Received: from [127.0.0.1] (10.173.222.66) by DGGEMS413-HUB.china.huawei.com (10.3.19.213) with Microsoft SMTP Server id 14.3.439.0; Wed, 12 Feb 2020 10:00:38 +0800 Subject: Re: [v3] nbd: fix potential NULL pointer fault in nbd_genl_disconnect To: Mike Christie , , CC: , , References: <20200210073241.41813-1-sunke32@huawei.com> <5E418D62.8090102@redhat.com> <5E42D8B1.406@redhat.com> From: "sunke (E)" Message-ID: <1b1110b2-1db6-9781-89cf-82b1403b1641@huawei.com> Date: Wed, 12 Feb 2020 10:00:37 +0800 User-Agent: Mozilla/5.0 (Windows NT 10.0; WOW64; rv:68.0) Gecko/20100101 Thunderbird/68.2.2 MIME-Version: 1.0 In-Reply-To: <5E42D8B1.406@redhat.com> Content-Type: text/plain; charset="utf-8"; format=flowed Content-Transfer-Encoding: 8bit X-Originating-IP: [10.173.222.66] X-CFilter-Loop: Reflected Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org 在 2020/2/12 0:39, Mike Christie 写道: > On 02/10/2020 10:12 PM, sunke (E) wrote: >> >> >> 在 2020/2/11 1:05, Mike Christie 写道: >>> On 02/10/2020 01:32 AM, Sun Ke wrote: >>>> Open /dev/nbdX first, the config_refs will be 1 and >>>> the pointers in nbd_device are still null. Disconnect >>>> /dev/nbdX, then reference a null recv_workq. The >>>> protection by config_refs in nbd_genl_disconnect is useless. >>>> >>>> To fix it, just add a check for a non null task_recv in >>>> nbd_genl_disconnect. >>>> >>>> Signed-off-by: Sun Ke >>>> --- >>>> v1 -> v2: >>>> Add an omitted mutex_unlock. >>>> >>>> v2 -> v3: >>>> Add nbd->config_lock, suggested by Josef. >>>> --- >>>> drivers/block/nbd.c | 8 ++++++++ >>>> 1 file changed, 8 insertions(+) >>>> >>>> diff --git a/drivers/block/nbd.c b/drivers/block/nbd.c >>>> index b4607dd96185..870b3fd0c101 100644 >>>> --- a/drivers/block/nbd.c >>>> +++ b/drivers/block/nbd.c >>>> @@ -2008,12 +2008,20 @@ static int nbd_genl_disconnect(struct sk_buff >>>> *skb, struct genl_info *info) >>>> index); >>>> return -EINVAL; >>>> } >>>> + mutex_lock(&nbd->config_lock); >>>> if (!refcount_inc_not_zero(&nbd->refs)) { >>>> + mutex_unlock(&nbd->config_lock); >>>> mutex_unlock(&nbd_index_mutex); >>>> printk(KERN_ERR "nbd: device at index %d is going down\n", >>>> index); >>>> return -EINVAL; >>>> } >>>> + if (!nbd->recv_workq) { >>>> + mutex_unlock(&nbd->config_lock); >>>> + mutex_unlock(&nbd_index_mutex); >>>> + return -EINVAL; >>>> + } >>>> + mutex_unlock(&nbd->config_lock); >>>> mutex_unlock(&nbd_index_mutex); >>>> if (!refcount_inc_not_zero(&nbd->config_refs)) { >>>> nbd_put(nbd); >>>> >>> >>> With my other patch then we will not need this right? It handles your >>> case by just being integrated with the existing checks in: >>> >>> nbd_disconnect_and_put->nbd_clear_sock->sock_shutdown >>> >>> ... >>> >>> static void sock_shutdown(struct nbd_device *nbd) >>> { >>> >>> .... >>> >>> if (config->num_connections == 0) >>> return; >>> >>> >>> num_connections is zero for your case since we never did a >>> nbd_genl_disconnect so we would return here. >>> >>> >>> . >>> >> Hi Mike >> >> Your point is not right totally. >> >> Yes, config->num_connections is 0 and will return in sock_shutdown. Then >> it will back to nbd_disconnect_and_put and do flush_workqueue >> (nbd->recv_workq). >> >> nbd_disconnect_and_put >> ->nbd_clear_sock >> ->sock_shutdown >> ->flush_workqueue >> > > My patch removed that extra flush_workqueue in nbd_disconnect_and_put. > > The idea of the patch was to move the flush calls to when we do > sock_shutdown in the config (connect, disconnect, clear sock) code > paths, because that is the time we know we will need to kill the recv > workers and wait for them to complete so we know they are not still > running when userspace does a new config operation. > Yes, I see.