Received: by 2002:ad5:474a:0:0:0:0:0 with SMTP id i10csp472992imu; Tue, 20 Nov 2018 02:05:39 -0800 (PST) X-Google-Smtp-Source: AJdET5cvb9rF7R5atqRIgl3ZrRxg3vPOnT7kgnih8AgW4CukETXT+rSF3xPXn9iMrckZtunbZm4F X-Received: by 2002:a62:3a04:: with SMTP id h4mr1483266pfa.119.1542708339859; Tue, 20 Nov 2018 02:05:39 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1542708339; cv=none; d=google.com; s=arc-20160816; b=H3GNqT6jlxQ3R5bozntYq5lgExYVeIYPgJ+n1YhsfCVRccK2CvQYlHbIdOFe4eJRtt 3hLDvwL2DE51RgvDHKBQFxQ0s4uh13zWpPr49cUzuy/l0MJbaHQVOZJ2aUInG9p9vpoV pv+53eg375ax3Jcvmxyu8DaLXuB/+fNbCnj/kmjRweANtVqwF3HNGWH50zgQ/FzuSMjs InW8IkZjPN3Sg9pxV7yN+n3hhjh9lMXkvPlM9/FTYspIDJ1zA9+c3caC2PUReSzKZ931 mlUEkd9d4nCa+srqnk+6E4LeP+2nCWt1OlHRXd6H6EKpHunRP8xkHk/sFx28bA/oK6Sj Wmzw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:content-transfer-encoding:in-reply-to :mime-version:user-agent:date:message-id:from:cc:references:to :subject; bh=oyUH9Ymirwu/ySwc7l3zzF1FEiYwywD1OHnBBJHwcwc=; b=rXZfDl8duk4kN0bXZIctCgGTMMOZbe44bpZzEPWpN5BA6b4t0kznoAujAO0UJWNhSA YUdHgrH4x+6KBOUf7bX1kLo1kbBuLjdb6/+MkyLjOGPQ2YbiB+icDGubs8W3vTsqvVGm 4UkCtXf74q7bumJ/M4S9i8qOE6zrGHxJZLP9/lQRNaKpSoqq39mhfEYoJY9HFOXNUQtL bJcCeHHqW+1sz44FsbKxAJI0XZkxPqrtoPIQi4EEmAcUlY0RyaPMRVkpZejzyssoOGPB m8g+5KdDr7UhtjtTHREd1gTxiEpNy4e2Ql4WXijpFHubTMID72Kqe6zROME6lReebthn xs8A== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id m3si9500207pld.331.2018.11.20.02.05.25; Tue, 20 Nov 2018 02:05:39 -0800 (PST) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1728115AbeKTU0J (ORCPT + 99 others); Tue, 20 Nov 2018 15:26:09 -0500 Received: from szxga07-in.huawei.com ([45.249.212.35]:46001 "EHLO huawei.com" rhost-flags-OK-OK-OK-FAIL) by vger.kernel.org with ESMTP id S1725999AbeKTU0J (ORCPT ); Tue, 20 Nov 2018 15:26:09 -0500 Received: from DGGEMS405-HUB.china.huawei.com (unknown [172.30.72.60]) by Forcepoint Email with ESMTP id 7C4D22808A7E0; Tue, 20 Nov 2018 17:57:49 +0800 (CST) Received: from [127.0.0.1] (10.202.226.41) by DGGEMS405-HUB.china.huawei.com (10.3.19.205) with Microsoft SMTP Server id 14.3.408.0; Tue, 20 Nov 2018 17:57:42 +0800 Subject: Re: [PATCH] iommu/dma: Use NUMA aware memory allocations in __iommu_dma_alloc_pages() To: References: <1541699709-25474-1-git-send-email-john.garry@huawei.com> CC: , , , , , , , From: John Garry Message-ID: <1e2b8689-dc1f-3ede-4eeb-4a91d3a3f63f@huawei.com> Date: Tue, 20 Nov 2018 09:57:36 +0000 User-Agent: Mozilla/5.0 (Windows NT 10.0; WOW64; rv:45.0) Gecko/20100101 Thunderbird/45.3.0 MIME-Version: 1.0 In-Reply-To: <1541699709-25474-1-git-send-email-john.garry@huawei.com> Content-Type: text/plain; charset="windows-1252"; format=flowed Content-Transfer-Encoding: 7bit X-Originating-IP: [10.202.226.41] X-CFilter-Loop: Reflected Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 08/11/2018 17:55, John Garry wrote: > Change function __iommu_dma_alloc_pages() to allocate memory/pages > for DMA from respective device NUMA node. > Ping.... a friendly reminder on this patch. Thanks > Originally-from: Ganapatrao Kulkarni > Signed-off-by: John Garry > --- > > This patch was originally posted by Ganapatrao in [1] *. > > However, after initial review, it was never reposted (due to lack of > cycles, I think). In addition, the functionality in its sibling patches > were merged through patches, as mentioned in [2]; this also refers to a > discussion on device local allocations vs CPU local allocations for DMA > pool, and which is better [3]. > > However, as mentioned in [3], dma_alloc_coherent() uses the locality > information from the device - as in direct DMA - so this patch is just > applying this same policy. > > [1] https://lore.kernel.org/patchwork/patch/833004/ > [2] https://lkml.org/lkml/2018/8/22/391 > [3] https://www.mail-archive.com/linux-kernel@vger.kernel.org/msg1692998.html > > * Authorship on this updated patch may need to be fixed - I add not want to > add Ganapatrao's SOB without permission. > > diff --git a/drivers/iommu/dma-iommu.c b/drivers/iommu/dma-iommu.c > index d1b0475..ada00bc 100644 > --- a/drivers/iommu/dma-iommu.c > +++ b/drivers/iommu/dma-iommu.c > @@ -449,20 +449,17 @@ static void __iommu_dma_free_pages(struct page **pages, int count) > kvfree(pages); > } > > -static struct page **__iommu_dma_alloc_pages(unsigned int count, > - unsigned long order_mask, gfp_t gfp) > +static struct page **__iommu_dma_alloc_pages(struct device *dev, > + unsigned int count, unsigned long order_mask, gfp_t gfp) > { > struct page **pages; > - unsigned int i = 0, array_size = count * sizeof(*pages); > + unsigned int i = 0, nid = dev_to_node(dev); > > order_mask &= (2U << MAX_ORDER) - 1; > if (!order_mask) > return NULL; > > - if (array_size <= PAGE_SIZE) > - pages = kzalloc(array_size, GFP_KERNEL); > - else > - pages = vzalloc(array_size); > + pages = kvzalloc_node(count * sizeof(*pages), GFP_KERNEL, nid); > if (!pages) > return NULL; > > @@ -483,8 +480,10 @@ static struct page **__iommu_dma_alloc_pages(unsigned int count, > unsigned int order = __fls(order_mask); > > order_size = 1U << order; > - page = alloc_pages((order_mask - order_size) ? > - gfp | __GFP_NORETRY : gfp, order); > + page = alloc_pages_node(nid, > + (order_mask - order_size) ? > + gfp | __GFP_NORETRY : gfp, > + order); > if (!page) > continue; > if (!order) > @@ -569,7 +568,8 @@ struct page **iommu_dma_alloc(struct device *dev, size_t size, gfp_t gfp, > alloc_sizes = min_size; > > count = PAGE_ALIGN(size) >> PAGE_SHIFT; > - pages = __iommu_dma_alloc_pages(count, alloc_sizes >> PAGE_SHIFT, gfp); > + pages = __iommu_dma_alloc_pages(dev, count, alloc_sizes >> PAGE_SHIFT, > + gfp); > if (!pages) > return NULL; > >