Received: by 2002:a05:6a10:22f:0:0:0:0 with SMTP id 15csp4114755pxk; Tue, 22 Sep 2020 10:34:34 -0700 (PDT) X-Google-Smtp-Source: ABdhPJxaEk73fDj9eemnsY/FRXwRXLdDU5nwMAqTxyMzTOF6KS20PiWfkwWwq9eh9MAOWiWVksyM X-Received: by 2002:a05:6402:1353:: with SMTP id y19mr5112517edw.71.1600796074656; Tue, 22 Sep 2020 10:34:34 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1600796074; cv=none; d=google.com; s=arc-20160816; b=glDK9vi//FyjF2rPEkyxgsOo7afmRdDACxhxQKx8LOHkv9re8jJ6KaruL6JA0bBNDX c44sNfoSBXtiSldGspc2CHTo42WiRbwQO/T3zMdW/Zg5eMRf7HwXaGjIPJE/KKRNSazA F6G57GR6K53yuczwq0N/UgLg1LJZ8yvnn905CVp5KVFgcogl0hyOK66LcI1iQivjgFBC 3mBXLuUVr2AkbRZ32hdRgjrbubIpaYGQaDgTs573uImTGtAZ2eEo0Rrlj62x3ZG7UEN7 TDnnwcC+MzewhKGIOSXZrQEoe0rqf/moGjnRsLzOprjcpbWWatUBpOU83S7EbzVNYxEW qKsA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:dkim-signature:content-transfer-encoding :content-language:in-reply-to:mime-version:user-agent:date :message-id:from:references:cc:to:subject; bh=FlTvSkg2VeAuml+QolVZZc+aE27r29315SlOmjxh39E=; b=sXYhQ2nGo2P+E+uVtbwRVVAmRzpt9dNQfvDFaLhI3cWVvDr7R+ovM5nYWyj+BAQuAJ fCJLanPE1ky+uzScKij5sVG2JYYz/ifBQ/4lRhJCbGMdp+A3yccEhZbbNGRWOoqV1aCY nMZIzWFC7uebs3rLe2yVhi0KXj4f2QOkcSF1yBYXes4EM+B4rFN/RfDvXNNq1Epj282S lEkynU1VLVCUAB7z/o1cxrvJTg1rSEsi4daYpNtcyQ8aX0IWVr3qaNzX2qatN0MpcRfK 5ZZ3veepfe4DvhEIz0rYSoOyiF1tiiB+AUG3wn3YHgC3DUB6ku5ZvuMBKgGGF5skxD2o mRnw== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@nvidia.com header.s=n1 header.b=n2rX4xtV; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=nvidia.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id yd22si14273898ejb.546.2020.09.22.10.34.10; Tue, 22 Sep 2020 10:34:34 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=pass header.i=@nvidia.com header.s=n1 header.b=n2rX4xtV; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=nvidia.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1726650AbgIVRbG (ORCPT + 99 others); Tue, 22 Sep 2020 13:31:06 -0400 Received: from hqnvemgate26.nvidia.com ([216.228.121.65]:11730 "EHLO hqnvemgate26.nvidia.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726526AbgIVRbG (ORCPT ); Tue, 22 Sep 2020 13:31:06 -0400 Received: from hqmail.nvidia.com (Not Verified[216.228.121.13]) by hqnvemgate26.nvidia.com (using TLS: TLSv1.2, AES256-SHA) id ; Tue, 22 Sep 2020 10:30:53 -0700 Received: from rcampbell-dev.nvidia.com (172.20.13.39) by HQMAIL107.nvidia.com (172.20.187.13) with Microsoft SMTP Server (TLS) id 15.0.1473.3; Tue, 22 Sep 2020 17:31:02 +0000 Subject: Re: [PATCH] mm/hmm/test: use after free in dmirror_allocate_chunk() To: Dan Carpenter , =?UTF-8?B?SsOpcsO0bWUgR2xpc3Nl?= CC: Jason Gunthorpe , Wei Yongjun , , , References: <20200922081234.GA1274646@mwanda> X-Nvconfidentiality: public From: Ralph Campbell Message-ID: Date: Tue, 22 Sep 2020 10:31:01 -0700 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:68.0) Gecko/20100101 Thunderbird/68.2.2 MIME-Version: 1.0 In-Reply-To: <20200922081234.GA1274646@mwanda> Content-Type: text/plain; charset="utf-8"; format=flowed Content-Language: en-US Content-Transfer-Encoding: 7bit X-Originating-IP: [172.20.13.39] X-ClientProxiedBy: HQMAIL107.nvidia.com (172.20.187.13) To HQMAIL107.nvidia.com (172.20.187.13) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=nvidia.com; s=n1; t=1600795853; bh=FlTvSkg2VeAuml+QolVZZc+aE27r29315SlOmjxh39E=; h=Subject:To:CC:References:X-Nvconfidentiality:From:Message-ID:Date: User-Agent:MIME-Version:In-Reply-To:Content-Type:Content-Language: Content-Transfer-Encoding:X-Originating-IP:X-ClientProxiedBy; b=n2rX4xtVEiYQ+FBTCrawg4dcPx1HKTKZZHFu0wsmNM8BaYERwT533PghchwBmk3j8 s5yMw/N9YH0s1iYauljogWzTK3X8jTUXNAPy33WzDSN5K1btEOE4hNX5k53kx58ECM ClVIdF7hiCenQf2kBsArGICJ/IlK45S6P8oh719RJxsPpoYc3+kuLDJnsztru8Ln64 ievHxAsZuu3L1QDQjmjxs4HymhO5JABI6r7NJcW4L6BSXOqBnR3yr6Mk5xaMHZlfls NppkhtfMRJXOM5ZSAZ5JN/3HUGAAf9ZMeXSf1fTzM1YLnmaCiveLzz+YAwb3FK04q+ fpbNBkpGmJlgQ== Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 9/22/20 1:12 AM, Dan Carpenter wrote: > The error handling code does this: > > err_free: > kfree(devmem); > ^^^^^^^^^^^^^ > err_release: > release_mem_region(devmem->pagemap.range.start, range_len(&devmem->pagemap.range)); > ^^^^^^^^ > The problem is that when we use "devmem->pagemap.range.start" the > "devmem" pointer is either NULL or freed. > > Neither the allocation nor the call to request_free_mem_region() has to > be done under the lock so I moved those to the start of the function. > > Fixes: b2ef9f5a5cb3 ("mm/hmm/test: add selftest driver for HMM") > Signed-off-by: Dan Carpenter > --- > It's weird that I didn't catch the use after free when this code was > merged in May... My bad. Not sure what happened there. How I found > this was that I have been reviewing release_mem_region() leaks and the > NULL dereference path is a leak. > Thanks for fixing this. I missed it too. :-) > lib/test_hmm.c | 47 ++++++++++++++++++++++++----------------------- > 1 file changed, 24 insertions(+), 23 deletions(-) > > diff --git a/lib/test_hmm.c b/lib/test_hmm.c > index c8133f50160b..0503c78cb322 100644 > --- a/lib/test_hmm.c > +++ b/lib/test_hmm.c > @@ -459,6 +459,22 @@ static bool dmirror_allocate_chunk(struct dmirror_device *mdevice, > unsigned long pfn_last; > void *ptr; > > + devmem = kzalloc(sizeof(*devmem), GFP_KERNEL); > + if (!devmem) > + return -ENOMEM; > + > + res = request_free_mem_region(&iomem_resource, DEVMEM_CHUNK_SIZE, > + "hmm_dmirror"); > + if (IS_ERR(res)) > + goto err_devmem; > + > + devmem->pagemap.type = MEMORY_DEVICE_PRIVATE; > + devmem->pagemap.range.start = res->start; > + devmem->pagemap.range.end = res->end; > + devmem->pagemap.nr_range = 1; > + devmem->pagemap.ops = &dmirror_devmem_ops; > + devmem->pagemap.owner = mdevice; > + > mutex_lock(&mdevice->devmem_lock); > > if (mdevice->devmem_count == mdevice->devmem_capacity) { > @@ -471,30 +487,16 @@ static bool dmirror_allocate_chunk(struct dmirror_device *mdevice, > sizeof(new_chunks[0]) * new_capacity, > GFP_KERNEL); > if (!new_chunks) Need to call mutex_unlock(&mdevice->devmem_lock). In fact, why not make this goto err_unlock and add err_unlock: mutex_unlock() before the err_release:. > - goto err; > + goto err_release;> mdevice->devmem_capacity = new_capacity; > mdevice->devmem_chunks = new_chunks; > } > > - res = request_free_mem_region(&iomem_resource, DEVMEM_CHUNK_SIZE, > - "hmm_dmirror"); > - if (IS_ERR(res)) > - goto err; > - > - devmem = kzalloc(sizeof(*devmem), GFP_KERNEL); > - if (!devmem) > - goto err_release; > - > - devmem->pagemap.type = MEMORY_DEVICE_PRIVATE; > - devmem->pagemap.range.start = res->start; > - devmem->pagemap.range.end = res->end; > - devmem->pagemap.nr_range = 1; > - devmem->pagemap.ops = &dmirror_devmem_ops; > - devmem->pagemap.owner = mdevice; > - > ptr = memremap_pages(&devmem->pagemap, numa_node_id()); > - if (IS_ERR(ptr)) > - goto err_free; > + if (IS_ERR(ptr)) { > + mutex_unlock(&mdevice->devmem_lock); > + goto err_release; > + } This could then be just goto err_unlock. > devmem->mdevice = mdevice; > pfn_first = devmem->pagemap.range.start >> PAGE_SHIFT; > @@ -525,12 +527,11 @@ static bool dmirror_allocate_chunk(struct dmirror_device *mdevice, > > return true; > > -err_free: > - kfree(devmem); > err_release: > release_mem_region(devmem->pagemap.range.start, range_len(&devmem->pagemap.range)); > -err: > - mutex_unlock(&mdevice->devmem_lock); > +err_devmem: > + kfree(devmem); > + > return false; > } > With the suggested change, you can add Reviewed-by: Ralph Campbell