Received: by 2002:a25:4158:0:0:0:0:0 with SMTP id o85csp4095062yba; Tue, 9 Apr 2019 11:01:56 -0700 (PDT) X-Google-Smtp-Source: APXvYqw8LE3x4nfcmb5yDcmMNrRquiLe1bAnbf4p8o3bbsPXbC1Mipb1uefmTseotRpdlSJxYsba X-Received: by 2002:a63:79c3:: with SMTP id u186mr35014807pgc.20.1554832916875; Tue, 09 Apr 2019 11:01:56 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1554832916; cv=none; d=google.com; s=arc-20160816; b=aQp+3imhCyR1n9m2oHWvZ37DNsyWgfhIZQgUevLtjVsckuG3l4V/Z+w/QvpweS4DIZ cuv+tld7dmxmXyD5OrlzpOH4L6uDafMhNokIbXnjqP77gOVumPV25UikzwGMPGIFUHNO VneODJ20+xmxOP56tRzs5kizTFtxn1q/stdhJTK13ET+7zkLM1NVAViB5eXBA6bCknGW SHCxV33GEfHdGlKrQZ97LOPqrvdCU8D6RNWVfSV6tSSEd+AfAk5ACL8zf5MHdPSBetLy 7jPChjCtzuBg2qPvNmM2eqPpHpavndTTM18lWkyrNUJTbSLmHdU0ytPkcvooqsl/cWLS gq6g== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:content-transfer-encoding :content-language:in-reply-to:mime-version:user-agent:date :message-id:from:references:cc:to:subject; bh=KscWdtf7bI5S4pN2qVmFMMzm3Iz0SAbwXU1/SdPk8V4=; b=0spTdCxzLYjsx3q6uEHZW8H9fQf8Dp9lFERpmyb8VHnPifkafMlrFJJfg6gRjiUHXI Oc1LVyouBgKGTZiN3abqfgr6tFpEuHd/EfFiEoJQSCF5X+2vdJr2NOmWVIxinIOxTfgn WYlqpTD5PRdMz4ooIALxjXFqMH/mP4tyZADRtzYH4OX6EI8vdUJUDJHzF35Pbphum7aK mpwDJKktLsZTzX1hv70hDAmQv1UsIbmsK/eObcvMSW2uIuH8nKfXF0imPTLWfytV1oFd 0W0cp/VWbgpgVZ6yeNfUzrpRBRW3H8NVUw0WfRVkTz6W7oM2Nuai5hieQ+DvOvWmXU9B J1JQ== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id a67si22597249pla.350.2019.04.09.11.01.39; Tue, 09 Apr 2019 11:01:56 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1726532AbfDIR7g (ORCPT + 99 others); Tue, 9 Apr 2019 13:59:36 -0400 Received: from usa-sjc-mx-foss1.foss.arm.com ([217.140.101.70]:42460 "EHLO foss.arm.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726415AbfDIR7g (ORCPT ); Tue, 9 Apr 2019 13:59:36 -0400 Received: from usa-sjc-imap-foss1.foss.arm.com (unknown [10.72.51.249]) by usa-sjc-mx-foss1.foss.arm.com (Postfix) with ESMTP id 7446615AB; Tue, 9 Apr 2019 10:59:35 -0700 (PDT) Received: from [10.1.196.75] (e110467-lin.cambridge.arm.com [10.1.196.75]) by usa-sjc-imap-foss1.foss.arm.com (Postfix) with ESMTPSA id 1B9993F59C; Tue, 9 Apr 2019 10:59:33 -0700 (PDT) Subject: Re: [PATCH 12/21] dma-iommu: factor atomic pool allocations into helpers To: Christoph Hellwig Cc: Joerg Roedel , Catalin Marinas , Will Deacon , Tom Lendacky , iommu@lists.linux-foundation.org, linux-arm-kernel@lists.infradead.org, linux-kernel@vger.kernel.org References: <20190327080448.5500-1-hch@lst.de> <20190327080448.5500-13-hch@lst.de> From: Robin Murphy Message-ID: Date: Tue, 9 Apr 2019 18:59:32 +0100 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:60.0) Gecko/20100101 Thunderbird/60.6.1 MIME-Version: 1.0 In-Reply-To: <20190327080448.5500-13-hch@lst.de> Content-Type: text/plain; charset=utf-8; format=flowed Content-Language: en-GB Content-Transfer-Encoding: 7bit Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 27/03/2019 08:04, Christoph Hellwig wrote: > This keeps the code together and will simplify compiling the code > out on architectures that are always dma coherent. And this is where things take a turn in the direction I just can't get on with - I'm looking at the final result and the twisty maze of little disjoint helpers all overlapping each other in functionality is really difficult to follow. And I would *much* rather have things rely on compile-time constant optimisation than spend the future having to fix the #ifdefed parts for arm64 whenever x86-centric changes fail to test them. Conceptually, everything except the iommu_dma_alloc_remap() case is more or less just dma_direct_alloc() with an IOMMU mapping on top - if we could pass that an internal flag to say "don't fail or bounce because of masks" it seems like that approach could end up being quite a bit simpler (I did once have lofty plans to refactor the old arm64 code in such a way...) Robin. > > Signed-off-by: Christoph Hellwig > --- > drivers/iommu/dma-iommu.c | 51 +++++++++++++++++++++++++++++---------- > 1 file changed, 38 insertions(+), 13 deletions(-) > > diff --git a/drivers/iommu/dma-iommu.c b/drivers/iommu/dma-iommu.c > index 2013c650718a..8ec69176673d 100644 > --- a/drivers/iommu/dma-iommu.c > +++ b/drivers/iommu/dma-iommu.c > @@ -673,6 +673,35 @@ static int iommu_dma_get_sgtable_remap(struct sg_table *sgt, void *cpu_addr, > GFP_KERNEL); > } > > +static void iommu_dma_free_pool(struct device *dev, size_t size, > + void *vaddr, dma_addr_t dma_handle) > +{ > + __iommu_dma_unmap(iommu_get_domain_for_dev(dev), dma_handle, size); > + dma_free_from_pool(vaddr, PAGE_ALIGN(size)); > +} > + > +static void *iommu_dma_alloc_pool(struct device *dev, size_t size, > + dma_addr_t *dma_handle, gfp_t gfp, unsigned long attrs) > +{ > + bool coherent = dev_is_dma_coherent(dev); > + struct page *page; > + void *vaddr; > + > + vaddr = dma_alloc_from_pool(PAGE_ALIGN(size), &page, gfp); > + if (!vaddr) > + return NULL; > + > + *dma_handle = __iommu_dma_map(dev, page_to_phys(page), size, > + dma_info_to_prot(DMA_BIDIRECTIONAL, coherent, attrs), > + iommu_get_domain_for_dev(dev)); > + if (*dma_handle == DMA_MAPPING_ERROR) { > + dma_free_from_pool(vaddr, PAGE_ALIGN(size)); > + return NULL; > + } > + > + return vaddr; > +} > + > static void iommu_dma_sync_single_for_cpu(struct device *dev, > dma_addr_t dma_handle, size_t size, enum dma_data_direction dir) > { > @@ -981,21 +1010,18 @@ static void *iommu_dma_alloc(struct device *dev, size_t size, > * get the virtually contiguous buffer we need by way of a > * physically contiguous allocation. > */ > - if (coherent) { > - page = alloc_pages(gfp, get_order(size)); > - addr = page ? page_address(page) : NULL; > - } else { > - addr = dma_alloc_from_pool(size, &page, gfp); > - } > - if (!addr) > + if (!coherent) > + return iommu_dma_alloc_pool(dev, iosize, handle, gfp, > + attrs); > + > + page = alloc_pages(gfp, get_order(size)); > + if (!page) > return NULL; > > + addr = page_address(page); > *handle = __iommu_dma_map_page(dev, page, 0, iosize, ioprot); > if (*handle == DMA_MAPPING_ERROR) { > - if (coherent) > - __free_pages(page, get_order(size)); > - else > - dma_free_from_pool(addr, size); > + __free_pages(page, get_order(size)); > addr = NULL; > } > } else if (attrs & DMA_ATTR_FORCE_CONTIGUOUS) { > @@ -1049,8 +1075,7 @@ static void iommu_dma_free(struct device *dev, size_t size, void *cpu_addr, > * Hence how dodgy the below logic looks... > */ > if (dma_in_atomic_pool(cpu_addr, size)) { > - __iommu_dma_unmap_page(dev, handle, iosize, 0, 0); > - dma_free_from_pool(cpu_addr, size); > + iommu_dma_free_pool(dev, size, cpu_addr, handle); > } else if (attrs & DMA_ATTR_FORCE_CONTIGUOUS) { > struct page *page = vmalloc_to_page(cpu_addr); > >