2021-08-06 10:36:29

by David Stevens

[permalink] [raw]
Subject: [PATCH v2 3/9] dma-iommu: bounce buffers for untrusted devices

From: David Stevens <[email protected]>

Add support for dynamic bounce buffers to the dma-api for use with
subgranule IOMMU mappings with untrusted devices. Bounce buffer
management is split into two parts. First, there is a buffer manager
that is responsible for allocating and tracking buffers. Second, there
is a layer that uses the managed buffers as bounce buffers. It is
responsible for managing the IOMMU mapping and for syncing between the
original and bounce buffers.

For now, buffer management is very simple - every mapping allocates a
new bounce buffer.

Signed-off-by: David Stevens <[email protected]>
---
drivers/iommu/Makefile | 2 +-
drivers/iommu/dma-iommu.c | 70 +++++-
drivers/iommu/io-bounce-buffers.c | 358 ++++++++++++++++++++++++++++++
drivers/iommu/io-bounce-buffers.h | 46 ++++
drivers/iommu/io-buffer-manager.c | 212 ++++++++++++++++++
drivers/iommu/io-buffer-manager.h | 43 ++++
6 files changed, 728 insertions(+), 3 deletions(-)
create mode 100644 drivers/iommu/io-bounce-buffers.c
create mode 100644 drivers/iommu/io-bounce-buffers.h
create mode 100644 drivers/iommu/io-buffer-manager.c
create mode 100644 drivers/iommu/io-buffer-manager.h

diff --git a/drivers/iommu/Makefile b/drivers/iommu/Makefile
index c0fb0ba88143..4edaf7adc082 100644
--- a/drivers/iommu/Makefile
+++ b/drivers/iommu/Makefile
@@ -4,7 +4,7 @@ obj-$(CONFIG_IOMMU_API) += iommu.o
obj-$(CONFIG_IOMMU_API) += iommu-traces.o
obj-$(CONFIG_IOMMU_API) += iommu-sysfs.o
obj-$(CONFIG_IOMMU_DEBUGFS) += iommu-debugfs.o
-obj-$(CONFIG_IOMMU_DMA) += dma-iommu.o
+obj-$(CONFIG_IOMMU_DMA) += dma-iommu.o io-bounce-buffers.o io-buffer-manager.o
obj-$(CONFIG_IOMMU_IO_PGTABLE) += io-pgtable.o
obj-$(CONFIG_IOMMU_IO_PGTABLE_ARMV7S) += io-pgtable-arm-v7s.o
obj-$(CONFIG_IOMMU_IO_PGTABLE_LPAE) += io-pgtable-arm.o
diff --git a/drivers/iommu/dma-iommu.c b/drivers/iommu/dma-iommu.c
index 055ccda5eba1..908eb6fb7dc3 100644
--- a/drivers/iommu/dma-iommu.c
+++ b/drivers/iommu/dma-iommu.c
@@ -24,6 +24,8 @@
#include <linux/vmalloc.h>
#include <linux/crash_dump.h>

+#include "io-bounce-buffers.h"
+
struct iommu_dma_msi_page {
struct list_head list;
dma_addr_t iova;
@@ -44,6 +46,7 @@ struct iommu_dma_cookie {
dma_addr_t msi_iova;
};
struct list_head msi_page_list;
+ struct io_bounce_buffers *bounce_buffers;

/* Domain for flush queue callback; NULL if flush queue not in use */
struct iommu_domain *fq_domain;
@@ -81,6 +84,14 @@ static inline size_t cookie_msi_granule(struct iommu_dma_cookie *cookie)
return PAGE_SIZE;
}

+static struct io_bounce_buffers *dev_to_io_bounce_buffers(struct device *dev)
+{
+ struct iommu_domain *domain = iommu_get_dma_domain(dev);
+ struct iommu_dma_cookie *cookie = domain->iova_cookie;
+
+ return cookie->bounce_buffers;
+}
+
static struct iommu_dma_cookie *cookie_alloc(enum iommu_dma_cookie_type type)
{
struct iommu_dma_cookie *cookie;
@@ -160,6 +171,9 @@ void iommu_put_dma_cookie(struct iommu_domain *domain)
if (!cookie)
return;

+ if (cookie->bounce_buffers)
+ io_bounce_buffers_destroy(cookie->bounce_buffers);
+
if (cookie->type == IOMMU_DMA_IOVA_COOKIE && cookie->iovad.granule)
put_iova_domain(&cookie->iovad);

@@ -333,6 +347,7 @@ static int iommu_dma_init_domain(struct iommu_domain *domain, dma_addr_t base,
struct iommu_dma_cookie *cookie = domain->iova_cookie;
unsigned long order, base_pfn;
struct iova_domain *iovad;
+ int ret;

if (!cookie || cookie->type != IOMMU_DMA_IOVA_COOKIE)
return -EINVAL;
@@ -380,7 +395,16 @@ static int iommu_dma_init_domain(struct iommu_domain *domain, dma_addr_t base,
if (!dev)
return 0;

- return iova_reserve_iommu_regions(dev, domain);
+ ret = iova_reserve_iommu_regions(dev, domain);
+
+ if (ret == 0 && dev_is_untrusted(dev)) {
+ cookie->bounce_buffers =
+ io_bounce_buffers_init(dev, domain, iovad);
+ if (IS_ERR(cookie->bounce_buffers))
+ ret = PTR_ERR(cookie->bounce_buffers);
+ }
+
+ return ret;
}

/**
@@ -710,8 +734,13 @@ static void iommu_dma_free_noncontiguous(struct device *dev, size_t size,
static void iommu_dma_sync_single_for_cpu(struct device *dev,
dma_addr_t dma_handle, size_t size, enum dma_data_direction dir)
{
+ struct io_bounce_buffers *bounce = dev_to_io_bounce_buffers(dev);
phys_addr_t phys;

+ if (bounce && io_bounce_buffers_sync_single(bounce, dma_handle,
+ size, dir, true))
+ return;
+
if (dev_is_dma_coherent(dev))
return;

@@ -722,8 +751,13 @@ static void iommu_dma_sync_single_for_cpu(struct device *dev,
static void iommu_dma_sync_single_for_device(struct device *dev,
dma_addr_t dma_handle, size_t size, enum dma_data_direction dir)
{
+ struct io_bounce_buffers *bounce = dev_to_io_bounce_buffers(dev);
phys_addr_t phys;

+ if (bounce && io_bounce_buffers_sync_single(bounce, dma_handle,
+ size, dir, false))
+ return;
+
if (dev_is_dma_coherent(dev))
return;

@@ -735,9 +769,13 @@ static void iommu_dma_sync_sg_for_cpu(struct device *dev,
struct scatterlist *sgl, int nelems,
enum dma_data_direction dir)
{
+ struct io_bounce_buffers *bounce = dev_to_io_bounce_buffers(dev);
struct scatterlist *sg;
int i;

+ if (bounce && io_bounce_buffers_sync_sg(bounce, sgl, nelems, dir, true))
+ return;
+
if (dev_is_dma_coherent(dev))
return;

@@ -749,9 +787,14 @@ static void iommu_dma_sync_sg_for_device(struct device *dev,
struct scatterlist *sgl, int nelems,
enum dma_data_direction dir)
{
+ struct io_bounce_buffers *bounce = dev_to_io_bounce_buffers(dev);
struct scatterlist *sg;
int i;

+ if (bounce && io_bounce_buffers_sync_sg(bounce, sgl,
+ nelems, dir, false))
+ return;
+
if (dev_is_dma_coherent(dev))
return;

@@ -763,11 +806,19 @@ static dma_addr_t iommu_dma_map_page(struct device *dev, struct page *page,
unsigned long offset, size_t size, enum dma_data_direction dir,
unsigned long attrs)
{
+ struct iommu_domain *domain = iommu_get_dma_domain(dev);
+ struct iommu_dma_cookie *cookie = domain->iova_cookie;
phys_addr_t phys = page_to_phys(page) + offset;
bool coherent = dev_is_dma_coherent(dev);
int prot = dma_info_to_prot(dir, coherent, attrs);
dma_addr_t dma_handle;

+ if (cookie->bounce_buffers &&
+ io_bounce_buffers_map_page(cookie->bounce_buffers, dev, page,
+ offset, size, prot, dir, attrs,
+ &dma_handle))
+ return dma_handle;
+
dma_handle = __iommu_dma_map(dev, phys, size, prot, dma_get_mask(dev));
if (!coherent && !(attrs & DMA_ATTR_SKIP_CPU_SYNC) &&
dma_handle != DMA_MAPPING_ERROR)
@@ -778,6 +829,12 @@ static dma_addr_t iommu_dma_map_page(struct device *dev, struct page *page,
static void iommu_dma_unmap_page(struct device *dev, dma_addr_t dma_handle,
size_t size, enum dma_data_direction dir, unsigned long attrs)
{
+ struct io_bounce_buffers *bounce = dev_to_io_bounce_buffers(dev);
+
+ if (bounce &&
+ io_bounce_buffers_unmap_page(bounce, dma_handle, size, dir, attrs))
+ return;
+
if (!(attrs & DMA_ATTR_SKIP_CPU_SYNC))
iommu_dma_sync_single_for_cpu(dev, dma_handle, size, dir);
__iommu_dma_unmap(dev, dma_handle, size);
@@ -876,12 +933,17 @@ static int iommu_dma_map_sg(struct device *dev, struct scatterlist *sg,
dma_addr_t iova;
size_t iova_len = 0;
unsigned long mask = dma_get_seg_boundary(dev);
- int i;
+ int i, ret;

if (static_branch_unlikely(&iommu_deferred_attach_enabled) &&
iommu_deferred_attach(dev, domain))
return 0;

+ if (cookie->bounce_buffers &&
+ io_bounce_buffers_map_sg(cookie->bounce_buffers, dev, sg, nents,
+ prot, dir, attrs, &ret))
+ return ret;
+
if (!(attrs & DMA_ATTR_SKIP_CPU_SYNC))
iommu_dma_sync_sg_for_device(dev, sg, nents, dir);

@@ -947,10 +1009,14 @@ static int iommu_dma_map_sg(struct device *dev, struct scatterlist *sg,
static void iommu_dma_unmap_sg(struct device *dev, struct scatterlist *sg,
int nents, enum dma_data_direction dir, unsigned long attrs)
{
+ struct io_bounce_buffers *bounce = dev_to_io_bounce_buffers(dev);
dma_addr_t start, end;
struct scatterlist *tmp;
int i;

+ if (bounce && io_bounce_buffers_unmap_sg(bounce, sg, nents, dir, attrs))
+ return;
+
if (!(attrs & DMA_ATTR_SKIP_CPU_SYNC))
iommu_dma_sync_sg_for_cpu(dev, sg, nents, dir);

diff --git a/drivers/iommu/io-bounce-buffers.c b/drivers/iommu/io-bounce-buffers.c
new file mode 100644
index 000000000000..78b4440b58c8
--- /dev/null
+++ b/drivers/iommu/io-bounce-buffers.c
@@ -0,0 +1,358 @@
+// SPDX-License-Identifier: GPL-2.0-only
+/*
+ * Dynamic IOMMU mapped bounce buffers.
+ *
+ * Copyright (C) 2021 Google, Inc.
+ */
+
+#include <linux/dma-iommu.h>
+#include <linux/dma-map-ops.h>
+#include <linux/highmem.h>
+#include <linux/list.h>
+#include <linux/slab.h>
+
+#include "io-buffer-manager.h"
+#include "io-bounce-buffers.h"
+
+struct io_bounce_buffers {
+ struct iommu_domain *domain;
+ struct iova_domain *iovad;
+ unsigned int nid;
+ struct io_buffer_manager manager;
+};
+
+bool io_bounce_buffers_release_buffer_cb(struct io_buffer_manager *manager,
+ dma_addr_t iova, size_t size)
+{
+ struct io_bounce_buffers *buffers =
+ container_of(manager, struct io_bounce_buffers, manager);
+ return iommu_unmap(buffers->domain, iova, size) >= size;
+}
+
+struct io_bounce_buffers *io_bounce_buffers_init(struct device *dev,
+ struct iommu_domain *domain,
+ struct iova_domain *iovad)
+{
+ int ret;
+ struct io_bounce_buffers *buffers;
+
+ buffers = kzalloc(sizeof(*buffers), GFP_KERNEL);
+ if (!buffers)
+ return ERR_PTR(-ENOMEM);
+
+ ret = io_buffer_manager_init(&buffers->manager);
+ if (ret) {
+ kfree(buffers);
+ return ERR_PTR(ret);
+ }
+
+ buffers->domain = domain;
+ buffers->iovad = iovad;
+ buffers->nid = dev_to_node(dev);
+
+ return buffers;
+}
+
+void io_bounce_buffers_destroy(struct io_bounce_buffers *buffers)
+{
+ kfree(buffers);
+}
+
+static bool should_sync_buffer(enum dma_data_direction dir, bool sync_for_cpu)
+{
+ return dir == DMA_BIDIRECTIONAL ||
+ (dir == DMA_FROM_DEVICE && sync_for_cpu) ||
+ (dir == DMA_TO_DEVICE && !sync_for_cpu);
+}
+
+static void io_bounce_buffers_do_sync(struct io_bounce_buffers *buffers,
+ struct page **bounce_buffer,
+ size_t bounce_offset, struct page *orig,
+ size_t orig_offset, size_t size,
+ enum dma_data_direction dir, int prot,
+ bool sync_for_cpu)
+{
+ bool needs_bounce_sync = should_sync_buffer(dir, sync_for_cpu);
+ char *orig_lowmem_ptr;
+ bool dma_is_coherent = prot & IOMMU_CACHE;
+
+ if (dma_is_coherent && !needs_bounce_sync)
+ return;
+
+ orig_lowmem_ptr = PageHighMem(orig) ? NULL : page_to_virt(orig);
+
+ while (size) {
+ size_t copy_len, bounce_page_offset;
+ struct page *bounce_page;
+
+ bounce_page = bounce_buffer[bounce_offset / PAGE_SIZE];
+ bounce_page_offset = bounce_offset % PAGE_SIZE;
+
+ copy_len = size;
+ if (copy_len + bounce_page_offset > PAGE_SIZE)
+ copy_len = PAGE_SIZE - bounce_page_offset;
+
+ if (!dma_is_coherent && sync_for_cpu) {
+ phys_addr_t paddr = page_to_phys(bounce_page);
+
+ arch_sync_dma_for_cpu(paddr + bounce_page_offset,
+ copy_len, dir);
+ }
+
+ if (needs_bounce_sync) {
+ char *bounce_page_ptr = kmap_local_page(bounce_page);
+ char *bounce_ptr = bounce_page_ptr + bounce_page_offset;
+
+ if (!orig_lowmem_ptr) {
+ size_t remaining = copy_len;
+ size_t offset = orig_offset % PAGE_SIZE;
+ size_t orig_page_idx = orig_offset / PAGE_SIZE;
+
+ while (remaining) {
+ char *orig_ptr;
+ size_t sz = min(remaining,
+ PAGE_SIZE - offset);
+
+ orig_ptr = kmap_local_page(
+ nth_page(orig, orig_page_idx));
+ if (sync_for_cpu) {
+ memcpy(orig_ptr + offset,
+ bounce_ptr, sz);
+ } else {
+ memcpy(bounce_ptr,
+ orig_ptr + offset, sz);
+ }
+ kunmap_local(orig_ptr);
+
+ remaining -= sz;
+ orig_page_idx += 1;
+ bounce_ptr += sz;
+ offset = 0;
+ }
+ } else if (sync_for_cpu) {
+ memcpy(orig_lowmem_ptr + orig_offset,
+ bounce_ptr, copy_len);
+ } else {
+ memcpy(bounce_ptr,
+ orig_lowmem_ptr + orig_offset, copy_len);
+ }
+
+ kunmap_local(bounce_page_ptr);
+ }
+
+ if (!dma_is_coherent && !sync_for_cpu) {
+ phys_addr_t paddr = page_to_phys(bounce_page);
+
+ arch_sync_dma_for_device(paddr + bounce_page_offset,
+ copy_len, dir);
+ }
+
+ bounce_offset += copy_len;
+ orig_offset += copy_len;
+ size -= copy_len;
+ }
+}
+
+bool io_bounce_buffers_sync_single(struct io_bounce_buffers *buffers,
+ dma_addr_t dma_handle, size_t size,
+ enum dma_data_direction dir,
+ bool sync_for_cpu)
+{
+ struct io_bounce_buffer_info info;
+ void *orig_buffer;
+ size_t offset;
+ int prot;
+
+ if (!io_buffer_manager_find_buffer(&buffers->manager, dma_handle, &info,
+ &orig_buffer, &prot))
+ return false;
+
+ offset = dma_handle - info.iova;
+ io_bounce_buffers_do_sync(buffers, info.bounce_buffer, offset,
+ orig_buffer, offset, size, dir, prot,
+ sync_for_cpu);
+ return true;
+}
+
+static void __io_bounce_buffers_sync_sg(struct io_bounce_buffers *buffers,
+ struct scatterlist *sgl, int nents,
+ struct page **bounce_buffer,
+ enum dma_data_direction dir, int prot,
+ bool sync_for_cpu)
+{
+ size_t bounce_offset = 0;
+ struct scatterlist *iter;
+ int i;
+
+ for_each_sg(sgl, iter, nents, i) {
+ io_bounce_buffers_do_sync(buffers, bounce_buffer, bounce_offset,
+ sg_page(iter), iter->offset,
+ iter->length, dir, prot,
+ sync_for_cpu);
+ bounce_offset += iter->length;
+ }
+}
+
+bool io_bounce_buffers_sync_sg(struct io_bounce_buffers *buffers,
+ struct scatterlist *sgl, int nents,
+ enum dma_data_direction dir, bool sync_for_cpu)
+{
+ struct io_bounce_buffer_info info;
+ void *orig_buffer;
+ int prot;
+
+ if (!io_buffer_manager_find_buffer(&buffers->manager,
+ sg_dma_address(sgl), &info,
+ &orig_buffer, &prot))
+ return false;
+
+ // In the non bounce buffer case, iommu_dma_map_sg syncs before setting
+ // up the new mapping's dma address. This check handles false positives
+ // in find_buffer caused by sgl being reused for a non bounce buffer
+ // case after being used with a bounce buffer.
+ if (orig_buffer != sgl)
+ return false;
+
+ __io_bounce_buffers_sync_sg(buffers, sgl, nents, info.bounce_buffer,
+ dir, prot, sync_for_cpu);
+
+ return true;
+}
+
+bool io_bounce_buffers_unmap_page(struct io_bounce_buffers *buffers,
+ dma_addr_t handle, size_t size,
+ enum dma_data_direction dir,
+ unsigned long attrs)
+{
+ if (!(attrs & DMA_ATTR_SKIP_CPU_SYNC))
+ io_bounce_buffers_sync_single(buffers, handle, size, dir, true);
+
+ return io_buffer_manager_release_buffer(&buffers->manager,
+ buffers->domain, handle, true);
+}
+
+bool io_bounce_buffers_unmap_sg(struct io_bounce_buffers *buffers,
+ struct scatterlist *sgl, int nents,
+ enum dma_data_direction dir,
+ unsigned long attrs)
+{
+ if (!(attrs & DMA_ATTR_SKIP_CPU_SYNC))
+ io_bounce_buffers_sync_sg(buffers, sgl, nents, dir, true);
+
+ return io_buffer_manager_release_buffer(
+ &buffers->manager, buffers->domain, sg_dma_address(sgl), true);
+}
+
+static bool io_bounce_buffers_map_buffer(struct io_bounce_buffers *buffers,
+ struct io_bounce_buffer_info *info,
+ int prot)
+{
+ unsigned int count = info->size >> PAGE_SHIFT;
+ struct sg_table sgt;
+ size_t mapped;
+
+ if (sg_alloc_table_from_pages(&sgt, info->bounce_buffer, count, 0,
+ info->size, GFP_ATOMIC))
+ return false;
+
+ mapped = iommu_map_sg_atomic(buffers->domain, info->iova, sgt.sgl,
+ sgt.orig_nents, prot);
+
+ sg_free_table(&sgt);
+ return mapped >= info->size;
+}
+
+bool io_bounce_buffers_map_page(struct io_bounce_buffers *buffers,
+ struct device *dev, struct page *page,
+ unsigned long offset, size_t size, int prot,
+ enum dma_data_direction dir,
+ unsigned long attrs, dma_addr_t *handle)
+{
+ bool skip_cpu_sync = attrs & DMA_ATTR_SKIP_CPU_SYNC;
+ struct io_bounce_buffer_info info;
+ bool force_bounce = iova_offset(buffers->iovad, offset | size);
+
+ if (!force_bounce)
+ return false;
+
+ *handle = DMA_MAPPING_ERROR;
+ if (!io_buffer_manager_alloc_buffer(&buffers->manager, dev, page,
+ offset + size, prot, buffers->nid,
+ &info))
+ return true;
+
+ if (!skip_cpu_sync)
+ io_bounce_buffers_do_sync(buffers, info.bounce_buffer, offset,
+ page, offset, size, dir, prot, false);
+
+ if (!io_bounce_buffers_map_buffer(buffers, &info, prot)) {
+ io_buffer_manager_release_buffer(
+ &buffers->manager, buffers->domain, info.iova, false);
+ return true;
+ }
+
+ *handle = info.iova + offset;
+ return true;
+}
+
+bool io_bounce_buffers_map_sg(struct io_bounce_buffers *buffers,
+ struct device *dev, struct scatterlist *sgl,
+ int nents, int prot, enum dma_data_direction dir,
+ unsigned long attrs, int *out_nents)
+{
+ struct io_bounce_buffer_info info;
+ struct scatterlist *iter;
+ size_t size = 0;
+ bool skip_cpu_sync = attrs & DMA_ATTR_SKIP_CPU_SYNC;
+ dma_addr_t seg_iova;
+ int i;
+ bool force_bounce = false;
+
+ for_each_sg(sgl, iter, nents, i) {
+ size += iter->length;
+ force_bounce |= iova_offset(buffers->iovad,
+ iter->offset | iter->length);
+ }
+
+ if (!force_bounce)
+ return false;
+
+ *out_nents = 0;
+ if (!io_buffer_manager_alloc_buffer(&buffers->manager, dev, sgl, size,
+ prot, buffers->nid, &info))
+ return true;
+
+ if (!skip_cpu_sync)
+ __io_bounce_buffers_sync_sg(buffers, sgl, nents,
+ info.bounce_buffer, dir, prot,
+ false);
+
+ if (!io_bounce_buffers_map_buffer(buffers, &info, prot)) {
+ io_buffer_manager_release_buffer(
+ &buffers->manager, buffers->domain, info.iova, false);
+ return true;
+ }
+
+ i = 0;
+ seg_iova = info.iova;
+ while (size > 0) {
+ size_t seg_size = min_t(size_t, size,
+ dma_get_max_seg_size(dev));
+
+ sg_dma_len(sgl) = seg_size;
+ sg_dma_address(sgl) = seg_iova;
+
+ sgl = sg_next(sgl);
+ size -= seg_size;
+ seg_iova += seg_size;
+ i++;
+ }
+
+ if (sgl) {
+ sg_dma_address(sgl) = DMA_MAPPING_ERROR;
+ sg_dma_len(sgl) = 0;
+ }
+
+ *out_nents = i;
+ return true;
+}
diff --git a/drivers/iommu/io-bounce-buffers.h b/drivers/iommu/io-bounce-buffers.h
new file mode 100644
index 000000000000..6d132a27646c
--- /dev/null
+++ b/drivers/iommu/io-bounce-buffers.h
@@ -0,0 +1,46 @@
+/* SPDX-License-Identifier: GPL-2.0-only */
+/*
+ * Copyright (C) 2021 Google, Inc.
+ */
+
+#ifndef _LINUX_IO_BOUNCE_BUFFERS_H
+#define _LINUX_IO_BOUNCE_BUFFERS_H
+
+#include <linux/dma-iommu.h>
+#include <linux/iova.h>
+
+struct io_bounce_buffers;
+
+struct io_bounce_buffers *io_bounce_buffers_init(struct device *dev,
+ struct iommu_domain *domain,
+ struct iova_domain *iovad);
+void io_bounce_buffers_destroy(struct io_bounce_buffers *buffers);
+
+bool io_bounce_buffers_sync_single(struct io_bounce_buffers *buffers,
+ dma_addr_t dma_handle, size_t size,
+ enum dma_data_direction dir,
+ bool sync_for_cpu);
+bool io_bounce_buffers_sync_sg(struct io_bounce_buffers *buffers,
+ struct scatterlist *sgl, int nents,
+ enum dma_data_direction dir, bool sync_for_cpu);
+
+bool io_bounce_buffers_map_page(struct io_bounce_buffers *buffers,
+ struct device *dev, struct page *page,
+ unsigned long offset, size_t size, int prot,
+ enum dma_data_direction dir,
+ unsigned long attrs, dma_addr_t *handle);
+bool io_bounce_buffers_map_sg(struct io_bounce_buffers *buffers,
+ struct device *dev, struct scatterlist *sgl,
+ int nents, int prot, enum dma_data_direction dir,
+ unsigned long attrs, int *out_nents);
+
+bool io_bounce_buffers_unmap_page(struct io_bounce_buffers *buffers,
+ dma_addr_t handle, size_t size,
+ enum dma_data_direction dir,
+ unsigned long attrs);
+bool io_bounce_buffers_unmap_sg(struct io_bounce_buffers *buffers,
+ struct scatterlist *sgl, int nents,
+ enum dma_data_direction dir,
+ unsigned long attrs);
+
+#endif /* _LINUX_IO_BOUNCE_BUFFERS_H */
diff --git a/drivers/iommu/io-buffer-manager.c b/drivers/iommu/io-buffer-manager.c
new file mode 100644
index 000000000000..24e95a2faa37
--- /dev/null
+++ b/drivers/iommu/io-buffer-manager.c
@@ -0,0 +1,212 @@
+// SPDX-License-Identifier: GPL-2.0-only
+/*
+ * Manager which allocates and tracks bounce buffers and their IOVAs. Does
+ * not actually manage the IOMMU mapping nor do the bounce copies.
+ *
+ * Copyright (C) 2021 Google, Inc.
+ */
+
+#include "io-buffer-manager.h"
+
+#include <linux/slab.h>
+
+struct io_buffer_node {
+ struct rb_node node;
+ struct io_bounce_buffer_info info;
+ void *orig_buffer;
+ int prot;
+};
+
+static void io_buffer_manager_free_pages(struct page **pages, int count)
+{
+ while (count--)
+ __free_page(pages[count]);
+ kfree(pages);
+}
+
+static struct page **io_buffer_manager_alloc_pages(int count, unsigned int nid)
+{
+ struct page **pages;
+ unsigned int i;
+
+ pages = kmalloc_array(count, sizeof(*pages), GFP_ATOMIC);
+ if (!pages)
+ return NULL;
+
+ // The IOMMU can map highmem pages, but try to allocate non-highmem
+ // pages first to make accessing the buffer cheaper.
+ for (i = 0; i < count; i++) {
+ pages[i] = alloc_pages_node(
+ nid,
+ GFP_ATOMIC | __GFP_ZERO | __GFP_NORETRY | __GFP_NOWARN,
+ 0);
+ if (!pages[i]) {
+ pages[i] = alloc_pages_node(
+ nid, GFP_ATOMIC | __GFP_ZERO | __GFP_HIGHMEM,
+ 0);
+ if (!pages[i]) {
+ io_buffer_manager_free_pages(pages, i);
+ return NULL;
+ }
+ }
+ }
+
+ return pages;
+}
+
+struct io_buffer_node *find_fallback_node(struct rb_root *root, dma_addr_t iova)
+{
+ struct rb_node *node = root->rb_node;
+
+ while (node) {
+ struct io_buffer_node *cur =
+ container_of(node, struct io_buffer_node, node);
+
+ if (iova < cur->info.iova)
+ node = node->rb_left;
+ else if (iova >= cur->info.iova + cur->info.size)
+ node = node->rb_right;
+ else
+ return cur;
+ }
+ return NULL;
+}
+
+bool insert_fallback_node(struct rb_root *root, struct io_buffer_node *node)
+{
+ struct rb_node **new = &(root->rb_node), *parent = NULL;
+ dma_addr_t node_end = node->info.iova + node->info.size;
+
+ while (*new) {
+ struct io_buffer_node *cur =
+ container_of(*new, struct io_buffer_node, node);
+ dma_addr_t cur_end = cur->info.iova + cur->info.size;
+
+ parent = *new;
+ if (node_end <= cur->info.iova)
+ new = &((*new)->rb_left);
+ else if (node->info.iova >= cur_end)
+ new = &((*new)->rb_right);
+ else {
+ pr_crit("IOVA collision new=[%llx,%llx) old=[%llx,%llx)\n",
+ node->info.iova, node_end, cur->info.iova,
+ cur_end);
+ return false;
+ }
+ }
+
+ rb_link_node(&node->node, parent, new);
+ rb_insert_color(&node->node, root);
+ return true;
+}
+
+bool io_buffer_manager_alloc_buffer(struct io_buffer_manager *manager,
+ struct device *dev, void *orig_buffer,
+ size_t size, int prot, unsigned int nid,
+ struct io_bounce_buffer_info *info)
+{
+ struct iommu_domain *domain = iommu_get_dma_domain(dev);
+ struct io_buffer_node *node;
+ unsigned long flags;
+
+ node = kzalloc(sizeof(*node), GFP_ATOMIC);
+ if (!node)
+ return false;
+
+ size = PAGE_ALIGN(size);
+ node->info.iova =
+ __iommu_dma_alloc_iova(domain, size, dma_get_mask(dev), dev);
+ if (!node->info.iova)
+ goto free_node;
+
+ node->info.bounce_buffer =
+ io_buffer_manager_alloc_pages(size >> PAGE_SHIFT, nid);
+ if (!node->info.bounce_buffer)
+ goto free_iova;
+
+ spin_lock_irqsave(&manager->fallback_lock, flags);
+ if (!insert_fallback_node(&manager->fallback_buffers, node))
+ goto fallback_lock_unlock;
+ spin_unlock_irqrestore(&manager->fallback_lock, flags);
+
+ node->orig_buffer = orig_buffer;
+ node->prot = prot;
+ node->info.size = size;
+
+ *info = node->info;
+
+ return true;
+
+fallback_lock_unlock:
+ spin_unlock_irqrestore(&manager->fallback_lock, flags);
+free_iova:
+ __iommu_dma_free_iova(domain->iova_cookie, node->info.iova, size, NULL);
+free_node:
+ kfree(node);
+ return false;
+}
+
+bool io_buffer_manager_find_buffer(struct io_buffer_manager *manager,
+ dma_addr_t handle,
+ struct io_bounce_buffer_info *info,
+ void **orig_buffer, int *prot)
+{
+ struct io_buffer_node *node;
+ unsigned long flags;
+
+ spin_lock_irqsave(&manager->fallback_lock, flags);
+ node = find_fallback_node(&manager->fallback_buffers, handle);
+ spin_unlock_irqrestore(&manager->fallback_lock, flags);
+
+ if (!node)
+ return false;
+
+ *info = node->info;
+ *orig_buffer = node->orig_buffer;
+ *prot = node->prot;
+ return true;
+}
+
+bool io_buffer_manager_release_buffer(struct io_buffer_manager *manager,
+ struct iommu_domain *domain,
+ dma_addr_t handle, bool inited)
+{
+ struct io_buffer_node *node;
+ unsigned long flags;
+ bool free_buffer;
+
+ spin_lock_irqsave(&manager->fallback_lock, flags);
+ node = find_fallback_node(&manager->fallback_buffers, handle);
+ if (node)
+ rb_erase(&node->node, &manager->fallback_buffers);
+ spin_unlock_irqrestore(&manager->fallback_lock, flags);
+
+ if (!node)
+ return false;
+
+ if (inited)
+ free_buffer = io_bounce_buffers_release_buffer_cb(
+ manager, node->info.iova, node->info.size);
+ else
+ free_buffer = true;
+
+ if (free_buffer) {
+ io_buffer_manager_free_pages(node->info.bounce_buffer,
+ node->info.size >> PAGE_SHIFT);
+ __iommu_dma_free_iova(domain->iova_cookie, node->info.iova,
+ node->info.size, NULL);
+ } else {
+ pr_warn("Bounce buffer release failed; leaking buffer\n");
+ }
+
+ kfree(node);
+ return true;
+}
+
+int io_buffer_manager_init(struct io_buffer_manager *manager)
+{
+ manager->fallback_buffers = RB_ROOT;
+ spin_lock_init(&manager->fallback_lock);
+
+ return 0;
+}
diff --git a/drivers/iommu/io-buffer-manager.h b/drivers/iommu/io-buffer-manager.h
new file mode 100644
index 000000000000..aae560cc8512
--- /dev/null
+++ b/drivers/iommu/io-buffer-manager.h
@@ -0,0 +1,43 @@
+/* SPDX-License-Identifier: GPL-2.0-only */
+/*
+ * Copyright (C) 2021 Google, Inc.
+ */
+
+#ifndef _LINUX_IO_BUFFER_MANAGER_H
+#define _LINUX_IO_BUFFER_MANAGER_H
+
+#include <linux/dma-iommu.h>
+#include <linux/iova.h>
+#include <linux/spinlock.h>
+
+struct io_buffer_manager {
+ spinlock_t fallback_lock;
+ struct rb_root fallback_buffers;
+};
+
+struct io_bounce_buffer_info {
+ struct page **bounce_buffer;
+ dma_addr_t iova;
+ size_t size;
+};
+
+bool io_buffer_manager_alloc_buffer(struct io_buffer_manager *manager,
+ struct device *dev, void *orig_buffer,
+ size_t size, int prot, unsigned int nid,
+ struct io_bounce_buffer_info *info);
+
+bool io_buffer_manager_find_buffer(struct io_buffer_manager *manager,
+ dma_addr_t handle,
+ struct io_bounce_buffer_info *info,
+ void **orig_buffer, int *prot);
+
+bool io_buffer_manager_release_buffer(struct io_buffer_manager *manager,
+ struct iommu_domain *domain,
+ dma_addr_t handle, bool inited);
+
+int io_buffer_manager_init(struct io_buffer_manager *manager);
+
+bool io_bounce_buffers_release_buffer_cb(struct io_buffer_manager *manager,
+ dma_addr_t iova, size_t size);
+
+#endif /* _LINUX_IO_BUFFER_MANAGER_H */
--
2.32.0.605.g8dce9f2422-goog


2021-08-06 23:18:54

by kernel test robot

[permalink] [raw]
Subject: Re: [PATCH v2 3/9] dma-iommu: bounce buffers for untrusted devices

Hi David,

Thank you for the patch! Perhaps something to improve:

[auto build test WARNING on iommu/next]
[also build test WARNING on drm-intel/for-linux-next hch-configfs/for-next linus/master v5.14-rc4]
[cannot apply to next-20210805]
[If your patch is applied to the wrong git tree, kindly drop us a note.
And when submitting patch, we suggest to use '--base' as documented in
https://git-scm.com/docs/git-format-patch]

url: https://github.com/0day-ci/linux/commits/David-Stevens/Add-dynamic-iommu-backed-bounce-buffers/20210806-183631
base: https://git.kernel.org/pub/scm/linux/kernel/git/joro/iommu.git next
config: ia64-randconfig-r025-20210804 (attached as .config)
compiler: ia64-linux-gcc (GCC) 10.3.0
reproduce (this is a W=1 build):
wget https://raw.githubusercontent.com/intel/lkp-tests/master/sbin/make.cross -O ~/bin/make.cross
chmod +x ~/bin/make.cross
# https://github.com/0day-ci/linux/commit/c5f1f9fa88a7062c1ded50fa165f6b01ed73f161
git remote add linux-review https://github.com/0day-ci/linux
git fetch --no-tags linux-review David-Stevens/Add-dynamic-iommu-backed-bounce-buffers/20210806-183631
git checkout c5f1f9fa88a7062c1ded50fa165f6b01ed73f161
# save the attached .config to linux build tree
COMPILER_INSTALL_PATH=$HOME/0day COMPILER=gcc-10.3.0 make.cross ARCH=ia64

If you fix the issue, kindly add following tag as appropriate
Reported-by: kernel test robot <[email protected]>

All warnings (new ones prefixed by >>):

>> drivers/iommu/io-buffer-manager.c:57:24: warning: no previous prototype for 'find_fallback_node' [-Wmissing-prototypes]
57 | struct io_buffer_node *find_fallback_node(struct rb_root *root, dma_addr_t iova)
| ^~~~~~~~~~~~~~~~~~
>> drivers/iommu/io-buffer-manager.c:75:6: warning: no previous prototype for 'insert_fallback_node' [-Wmissing-prototypes]
75 | bool insert_fallback_node(struct rb_root *root, struct io_buffer_node *node)
| ^~~~~~~~~~~~~~~~~~~~


vim +/find_fallback_node +57 drivers/iommu/io-buffer-manager.c

56
> 57 struct io_buffer_node *find_fallback_node(struct rb_root *root, dma_addr_t iova)
58 {
59 struct rb_node *node = root->rb_node;
60
61 while (node) {
62 struct io_buffer_node *cur =
63 container_of(node, struct io_buffer_node, node);
64
65 if (iova < cur->info.iova)
66 node = node->rb_left;
67 else if (iova >= cur->info.iova + cur->info.size)
68 node = node->rb_right;
69 else
70 return cur;
71 }
72 return NULL;
73 }
74
> 75 bool insert_fallback_node(struct rb_root *root, struct io_buffer_node *node)
76 {
77 struct rb_node **new = &(root->rb_node), *parent = NULL;
78 dma_addr_t node_end = node->info.iova + node->info.size;
79
80 while (*new) {
81 struct io_buffer_node *cur =
82 container_of(*new, struct io_buffer_node, node);
83 dma_addr_t cur_end = cur->info.iova + cur->info.size;
84
85 parent = *new;
86 if (node_end <= cur->info.iova)
87 new = &((*new)->rb_left);
88 else if (node->info.iova >= cur_end)
89 new = &((*new)->rb_right);
90 else {
91 pr_crit("IOVA collision new=[%llx,%llx) old=[%llx,%llx)\n",
92 node->info.iova, node_end, cur->info.iova,
93 cur_end);
94 return false;
95 }
96 }
97
98 rb_link_node(&node->node, parent, new);
99 rb_insert_color(&node->node, root);
100 return true;
101 }
102

---
0-DAY CI Kernel Test Service, Intel Corporation
https://lists.01.org/hyperkitty/list/[email protected]


Attachments:
(No filename) (3.70 kB)
.config.gz (35.99 kB)
Download all attachments

2021-08-10 04:28:32

by Mi, Dapeng1

[permalink] [raw]
Subject: RE: [PATCH v2 3/9] dma-iommu: bounce buffers for untrusted devices

Hi David,

I like this patch set and this is crucial for reducing the significant vIOMMU performance. It looks you totally rewrite the IOMMU mapping/unmapping part and use the dynamically allocated memory from buddy system as bounce buffer instead of using the legacy SWIOTLB bounce buffer. As I know, some legacy devices' DMA could not access the memory larger than 32-bit memory space and the dynamically allocated memory address could exceed the 32-bit memory space. Is it a problem?

Thx,
Dapeng Mi

-----Original Message-----
From: iommu <[email protected]> On Behalf Of David Stevens
Sent: Friday, August 6, 2021 6:34 PM
To: Robin Murphy <[email protected]>
Cc: [email protected]; Sergey Senozhatsky <[email protected]>; [email protected]; David Stevens <[email protected]>; Will Deacon <[email protected]>; Christoph Hellwig <[email protected]>
Subject: [PATCH v2 3/9] dma-iommu: bounce buffers for untrusted devices

From: David Stevens <[email protected]>

Add support for dynamic bounce buffers to the dma-api for use with subgranule IOMMU mappings with untrusted devices. Bounce buffer management is split into two parts. First, there is a buffer manager that is responsible for allocating and tracking buffers. Second, there is a layer that uses the managed buffers as bounce buffers. It is responsible for managing the IOMMU mapping and for syncing between the original and bounce buffers.

For now, buffer management is very simple - every mapping allocates a new bounce buffer.

Signed-off-by: David Stevens <[email protected]>
---
drivers/iommu/Makefile | 2 +-
drivers/iommu/dma-iommu.c | 70 +++++-
drivers/iommu/io-bounce-buffers.c | 358 ++++++++++++++++++++++++++++++ drivers/iommu/io-bounce-buffers.h | 46 ++++ drivers/iommu/io-buffer-manager.c | 212 ++++++++++++++++++ drivers/iommu/io-buffer-manager.h | 43 ++++
6 files changed, 728 insertions(+), 3 deletions(-) create mode 100644 drivers/iommu/io-bounce-buffers.c create mode 100644 drivers/iommu/io-bounce-buffers.h create mode 100644 drivers/iommu/io-buffer-manager.c create mode 100644 drivers/iommu/io-buffer-manager.h

diff --git a/drivers/iommu/Makefile b/drivers/iommu/Makefile index c0fb0ba88143..4edaf7adc082 100644
--- a/drivers/iommu/Makefile
+++ b/drivers/iommu/Makefile
@@ -4,7 +4,7 @@ obj-$(CONFIG_IOMMU_API) += iommu.o
obj-$(CONFIG_IOMMU_API) += iommu-traces.o
obj-$(CONFIG_IOMMU_API) += iommu-sysfs.o
obj-$(CONFIG_IOMMU_DEBUGFS) += iommu-debugfs.o
-obj-$(CONFIG_IOMMU_DMA) += dma-iommu.o
+obj-$(CONFIG_IOMMU_DMA) += dma-iommu.o io-bounce-buffers.o
+io-buffer-manager.o
obj-$(CONFIG_IOMMU_IO_PGTABLE) += io-pgtable.o
obj-$(CONFIG_IOMMU_IO_PGTABLE_ARMV7S) += io-pgtable-arm-v7s.o
obj-$(CONFIG_IOMMU_IO_PGTABLE_LPAE) += io-pgtable-arm.o diff --git a/drivers/iommu/dma-iommu.c b/drivers/iommu/dma-iommu.c index 055ccda5eba1..908eb6fb7dc3 100644
--- a/drivers/iommu/dma-iommu.c
+++ b/drivers/iommu/dma-iommu.c
@@ -24,6 +24,8 @@
#include <linux/vmalloc.h>
#include <linux/crash_dump.h>

+#include "io-bounce-buffers.h"
+
struct iommu_dma_msi_page {
struct list_head list;
dma_addr_t iova;
@@ -44,6 +46,7 @@ struct iommu_dma_cookie {
dma_addr_t msi_iova;
};
struct list_head msi_page_list;
+ struct io_bounce_buffers *bounce_buffers;

/* Domain for flush queue callback; NULL if flush queue not in use */
struct iommu_domain *fq_domain;
@@ -81,6 +84,14 @@ static inline size_t cookie_msi_granule(struct iommu_dma_cookie *cookie)
return PAGE_SIZE;
}

+static struct io_bounce_buffers *dev_to_io_bounce_buffers(struct device
+*dev) {
+ struct iommu_domain *domain = iommu_get_dma_domain(dev);
+ struct iommu_dma_cookie *cookie = domain->iova_cookie;
+
+ return cookie->bounce_buffers;
+}
+
static struct iommu_dma_cookie *cookie_alloc(enum iommu_dma_cookie_type type) {
struct iommu_dma_cookie *cookie;
@@ -160,6 +171,9 @@ void iommu_put_dma_cookie(struct iommu_domain *domain)
if (!cookie)
return;

+ if (cookie->bounce_buffers)
+ io_bounce_buffers_destroy(cookie->bounce_buffers);
+
if (cookie->type == IOMMU_DMA_IOVA_COOKIE && cookie->iovad.granule)
put_iova_domain(&cookie->iovad);

@@ -333,6 +347,7 @@ static int iommu_dma_init_domain(struct iommu_domain *domain, dma_addr_t base,
struct iommu_dma_cookie *cookie = domain->iova_cookie;
unsigned long order, base_pfn;
struct iova_domain *iovad;
+ int ret;

if (!cookie || cookie->type != IOMMU_DMA_IOVA_COOKIE)
return -EINVAL;
@@ -380,7 +395,16 @@ static int iommu_dma_init_domain(struct iommu_domain *domain, dma_addr_t base,
if (!dev)
return 0;

- return iova_reserve_iommu_regions(dev, domain);
+ ret = iova_reserve_iommu_regions(dev, domain);
+
+ if (ret == 0 && dev_is_untrusted(dev)) {
+ cookie->bounce_buffers =
+ io_bounce_buffers_init(dev, domain, iovad);
+ if (IS_ERR(cookie->bounce_buffers))
+ ret = PTR_ERR(cookie->bounce_buffers);
+ }
+
+ return ret;
}

/**
@@ -710,8 +734,13 @@ static void iommu_dma_free_noncontiguous(struct device *dev, size_t size, static void iommu_dma_sync_single_for_cpu(struct device *dev,
dma_addr_t dma_handle, size_t size, enum dma_data_direction dir) {
+ struct io_bounce_buffers *bounce = dev_to_io_bounce_buffers(dev);
phys_addr_t phys;

+ if (bounce && io_bounce_buffers_sync_single(bounce, dma_handle,
+ size, dir, true))
+ return;
+
if (dev_is_dma_coherent(dev))
return;

@@ -722,8 +751,13 @@ static void iommu_dma_sync_single_for_cpu(struct device *dev, static void iommu_dma_sync_single_for_device(struct device *dev,
dma_addr_t dma_handle, size_t size, enum dma_data_direction dir) {
+ struct io_bounce_buffers *bounce = dev_to_io_bounce_buffers(dev);
phys_addr_t phys;

+ if (bounce && io_bounce_buffers_sync_single(bounce, dma_handle,
+ size, dir, false))
+ return;
+
if (dev_is_dma_coherent(dev))
return;

@@ -735,9 +769,13 @@ static void iommu_dma_sync_sg_for_cpu(struct device *dev,
struct scatterlist *sgl, int nelems,
enum dma_data_direction dir)
{
+ struct io_bounce_buffers *bounce = dev_to_io_bounce_buffers(dev);
struct scatterlist *sg;
int i;

+ if (bounce && io_bounce_buffers_sync_sg(bounce, sgl, nelems, dir, true))
+ return;
+
if (dev_is_dma_coherent(dev))
return;

@@ -749,9 +787,14 @@ static void iommu_dma_sync_sg_for_device(struct device *dev,
struct scatterlist *sgl, int nelems,
enum dma_data_direction dir)
{
+ struct io_bounce_buffers *bounce = dev_to_io_bounce_buffers(dev);
struct scatterlist *sg;
int i;

+ if (bounce && io_bounce_buffers_sync_sg(bounce, sgl,
+ nelems, dir, false))
+ return;
+
if (dev_is_dma_coherent(dev))
return;

@@ -763,11 +806,19 @@ static dma_addr_t iommu_dma_map_page(struct device *dev, struct page *page,
unsigned long offset, size_t size, enum dma_data_direction dir,
unsigned long attrs)
{
+ struct iommu_domain *domain = iommu_get_dma_domain(dev);
+ struct iommu_dma_cookie *cookie = domain->iova_cookie;
phys_addr_t phys = page_to_phys(page) + offset;
bool coherent = dev_is_dma_coherent(dev);
int prot = dma_info_to_prot(dir, coherent, attrs);
dma_addr_t dma_handle;

+ if (cookie->bounce_buffers &&
+ io_bounce_buffers_map_page(cookie->bounce_buffers, dev, page,
+ offset, size, prot, dir, attrs,
+ &dma_handle))
+ return dma_handle;
+
dma_handle = __iommu_dma_map(dev, phys, size, prot, dma_get_mask(dev));
if (!coherent && !(attrs & DMA_ATTR_SKIP_CPU_SYNC) &&
dma_handle != DMA_MAPPING_ERROR)
@@ -778,6 +829,12 @@ static dma_addr_t iommu_dma_map_page(struct device *dev, struct page *page, static void iommu_dma_unmap_page(struct device *dev, dma_addr_t dma_handle,
size_t size, enum dma_data_direction dir, unsigned long attrs) {
+ struct io_bounce_buffers *bounce = dev_to_io_bounce_buffers(dev);
+
+ if (bounce &&
+ io_bounce_buffers_unmap_page(bounce, dma_handle, size, dir, attrs))
+ return;
+
if (!(attrs & DMA_ATTR_SKIP_CPU_SYNC))
iommu_dma_sync_single_for_cpu(dev, dma_handle, size, dir);
__iommu_dma_unmap(dev, dma_handle, size); @@ -876,12 +933,17 @@ static int iommu_dma_map_sg(struct device *dev, struct scatterlist *sg,
dma_addr_t iova;
size_t iova_len = 0;
unsigned long mask = dma_get_seg_boundary(dev);
- int i;
+ int i, ret;

if (static_branch_unlikely(&iommu_deferred_attach_enabled) &&
iommu_deferred_attach(dev, domain))
return 0;

+ if (cookie->bounce_buffers &&
+ io_bounce_buffers_map_sg(cookie->bounce_buffers, dev, sg, nents,
+ prot, dir, attrs, &ret))
+ return ret;
+
if (!(attrs & DMA_ATTR_SKIP_CPU_SYNC))
iommu_dma_sync_sg_for_device(dev, sg, nents, dir);

@@ -947,10 +1009,14 @@ static int iommu_dma_map_sg(struct device *dev, struct scatterlist *sg, static void iommu_dma_unmap_sg(struct device *dev, struct scatterlist *sg,
int nents, enum dma_data_direction dir, unsigned long attrs) {
+ struct io_bounce_buffers *bounce = dev_to_io_bounce_buffers(dev);
dma_addr_t start, end;
struct scatterlist *tmp;
int i;

+ if (bounce && io_bounce_buffers_unmap_sg(bounce, sg, nents, dir, attrs))
+ return;
+
if (!(attrs & DMA_ATTR_SKIP_CPU_SYNC))
iommu_dma_sync_sg_for_cpu(dev, sg, nents, dir);

diff --git a/drivers/iommu/io-bounce-buffers.c b/drivers/iommu/io-bounce-buffers.c
new file mode 100644
index 000000000000..78b4440b58c8
--- /dev/null
+++ b/drivers/iommu/io-bounce-buffers.c
@@ -0,0 +1,358 @@
+// SPDX-License-Identifier: GPL-2.0-only
+/*
+ * Dynamic IOMMU mapped bounce buffers.
+ *
+ * Copyright (C) 2021 Google, Inc.
+ */
+
+#include <linux/dma-iommu.h>
+#include <linux/dma-map-ops.h>
+#include <linux/highmem.h>
+#include <linux/list.h>
+#include <linux/slab.h>
+
+#include "io-buffer-manager.h"
+#include "io-bounce-buffers.h"
+
+struct io_bounce_buffers {
+ struct iommu_domain *domain;
+ struct iova_domain *iovad;
+ unsigned int nid;
+ struct io_buffer_manager manager;
+};
+
+bool io_bounce_buffers_release_buffer_cb(struct io_buffer_manager *manager,
+ dma_addr_t iova, size_t size)
+{
+ struct io_bounce_buffers *buffers =
+ container_of(manager, struct io_bounce_buffers, manager);
+ return iommu_unmap(buffers->domain, iova, size) >= size; }
+
+struct io_bounce_buffers *io_bounce_buffers_init(struct device *dev,
+ struct iommu_domain *domain,
+ struct iova_domain *iovad)
+{
+ int ret;
+ struct io_bounce_buffers *buffers;
+
+ buffers = kzalloc(sizeof(*buffers), GFP_KERNEL);
+ if (!buffers)
+ return ERR_PTR(-ENOMEM);
+
+ ret = io_buffer_manager_init(&buffers->manager);
+ if (ret) {
+ kfree(buffers);
+ return ERR_PTR(ret);
+ }
+
+ buffers->domain = domain;
+ buffers->iovad = iovad;
+ buffers->nid = dev_to_node(dev);
+
+ return buffers;
+}
+
+void io_bounce_buffers_destroy(struct io_bounce_buffers *buffers) {
+ kfree(buffers);
+}
+
+static bool should_sync_buffer(enum dma_data_direction dir, bool
+sync_for_cpu) {
+ return dir == DMA_BIDIRECTIONAL ||
+ (dir == DMA_FROM_DEVICE && sync_for_cpu) ||
+ (dir == DMA_TO_DEVICE && !sync_for_cpu); }
+
+static void io_bounce_buffers_do_sync(struct io_bounce_buffers *buffers,
+ struct page **bounce_buffer,
+ size_t bounce_offset, struct page *orig,
+ size_t orig_offset, size_t size,
+ enum dma_data_direction dir, int prot,
+ bool sync_for_cpu)
+{
+ bool needs_bounce_sync = should_sync_buffer(dir, sync_for_cpu);
+ char *orig_lowmem_ptr;
+ bool dma_is_coherent = prot & IOMMU_CACHE;
+
+ if (dma_is_coherent && !needs_bounce_sync)
+ return;
+
+ orig_lowmem_ptr = PageHighMem(orig) ? NULL : page_to_virt(orig);
+
+ while (size) {
+ size_t copy_len, bounce_page_offset;
+ struct page *bounce_page;
+
+ bounce_page = bounce_buffer[bounce_offset / PAGE_SIZE];
+ bounce_page_offset = bounce_offset % PAGE_SIZE;
+
+ copy_len = size;
+ if (copy_len + bounce_page_offset > PAGE_SIZE)
+ copy_len = PAGE_SIZE - bounce_page_offset;
+
+ if (!dma_is_coherent && sync_for_cpu) {
+ phys_addr_t paddr = page_to_phys(bounce_page);
+
+ arch_sync_dma_for_cpu(paddr + bounce_page_offset,
+ copy_len, dir);
+ }
+
+ if (needs_bounce_sync) {
+ char *bounce_page_ptr = kmap_local_page(bounce_page);
+ char *bounce_ptr = bounce_page_ptr + bounce_page_offset;
+
+ if (!orig_lowmem_ptr) {
+ size_t remaining = copy_len;
+ size_t offset = orig_offset % PAGE_SIZE;
+ size_t orig_page_idx = orig_offset / PAGE_SIZE;
+
+ while (remaining) {
+ char *orig_ptr;
+ size_t sz = min(remaining,
+ PAGE_SIZE - offset);
+
+ orig_ptr = kmap_local_page(
+ nth_page(orig, orig_page_idx));
+ if (sync_for_cpu) {
+ memcpy(orig_ptr + offset,
+ bounce_ptr, sz);
+ } else {
+ memcpy(bounce_ptr,
+ orig_ptr + offset, sz);
+ }
+ kunmap_local(orig_ptr);
+
+ remaining -= sz;
+ orig_page_idx += 1;
+ bounce_ptr += sz;
+ offset = 0;
+ }
+ } else if (sync_for_cpu) {
+ memcpy(orig_lowmem_ptr + orig_offset,
+ bounce_ptr, copy_len);
+ } else {
+ memcpy(bounce_ptr,
+ orig_lowmem_ptr + orig_offset, copy_len);
+ }
+
+ kunmap_local(bounce_page_ptr);
+ }
+
+ if (!dma_is_coherent && !sync_for_cpu) {
+ phys_addr_t paddr = page_to_phys(bounce_page);
+
+ arch_sync_dma_for_device(paddr + bounce_page_offset,
+ copy_len, dir);
+ }
+
+ bounce_offset += copy_len;
+ orig_offset += copy_len;
+ size -= copy_len;
+ }
+}
+
+bool io_bounce_buffers_sync_single(struct io_bounce_buffers *buffers,
+ dma_addr_t dma_handle, size_t size,
+ enum dma_data_direction dir,
+ bool sync_for_cpu)
+{
+ struct io_bounce_buffer_info info;
+ void *orig_buffer;
+ size_t offset;
+ int prot;
+
+ if (!io_buffer_manager_find_buffer(&buffers->manager, dma_handle, &info,
+ &orig_buffer, &prot))
+ return false;
+
+ offset = dma_handle - info.iova;
+ io_bounce_buffers_do_sync(buffers, info.bounce_buffer, offset,
+ orig_buffer, offset, size, dir, prot,
+ sync_for_cpu);
+ return true;
+}
+
+static void __io_bounce_buffers_sync_sg(struct io_bounce_buffers *buffers,
+ struct scatterlist *sgl, int nents,
+ struct page **bounce_buffer,
+ enum dma_data_direction dir, int prot,
+ bool sync_for_cpu)
+{
+ size_t bounce_offset = 0;
+ struct scatterlist *iter;
+ int i;
+
+ for_each_sg(sgl, iter, nents, i) {
+ io_bounce_buffers_do_sync(buffers, bounce_buffer, bounce_offset,
+ sg_page(iter), iter->offset,
+ iter->length, dir, prot,
+ sync_for_cpu);
+ bounce_offset += iter->length;
+ }
+}
+
+bool io_bounce_buffers_sync_sg(struct io_bounce_buffers *buffers,
+ struct scatterlist *sgl, int nents,
+ enum dma_data_direction dir, bool sync_for_cpu) {
+ struct io_bounce_buffer_info info;
+ void *orig_buffer;
+ int prot;
+
+ if (!io_buffer_manager_find_buffer(&buffers->manager,
+ sg_dma_address(sgl), &info,
+ &orig_buffer, &prot))
+ return false;
+
+ // In the non bounce buffer case, iommu_dma_map_sg syncs before setting
+ // up the new mapping's dma address. This check handles false positives
+ // in find_buffer caused by sgl being reused for a non bounce buffer
+ // case after being used with a bounce buffer.
+ if (orig_buffer != sgl)
+ return false;
+
+ __io_bounce_buffers_sync_sg(buffers, sgl, nents, info.bounce_buffer,
+ dir, prot, sync_for_cpu);
+
+ return true;
+}
+
+bool io_bounce_buffers_unmap_page(struct io_bounce_buffers *buffers,
+ dma_addr_t handle, size_t size,
+ enum dma_data_direction dir,
+ unsigned long attrs)
+{
+ if (!(attrs & DMA_ATTR_SKIP_CPU_SYNC))
+ io_bounce_buffers_sync_single(buffers, handle, size, dir, true);
+
+ return io_buffer_manager_release_buffer(&buffers->manager,
+ buffers->domain, handle, true);
+}
+
+bool io_bounce_buffers_unmap_sg(struct io_bounce_buffers *buffers,
+ struct scatterlist *sgl, int nents,
+ enum dma_data_direction dir,
+ unsigned long attrs)
+{
+ if (!(attrs & DMA_ATTR_SKIP_CPU_SYNC))
+ io_bounce_buffers_sync_sg(buffers, sgl, nents, dir, true);
+
+ return io_buffer_manager_release_buffer(
+ &buffers->manager, buffers->domain, sg_dma_address(sgl), true); }
+
+static bool io_bounce_buffers_map_buffer(struct io_bounce_buffers *buffers,
+ struct io_bounce_buffer_info *info,
+ int prot)
+{
+ unsigned int count = info->size >> PAGE_SHIFT;
+ struct sg_table sgt;
+ size_t mapped;
+
+ if (sg_alloc_table_from_pages(&sgt, info->bounce_buffer, count, 0,
+ info->size, GFP_ATOMIC))
+ return false;
+
+ mapped = iommu_map_sg_atomic(buffers->domain, info->iova, sgt.sgl,
+ sgt.orig_nents, prot);
+
+ sg_free_table(&sgt);
+ return mapped >= info->size;
+}
+
+bool io_bounce_buffers_map_page(struct io_bounce_buffers *buffers,
+ struct device *dev, struct page *page,
+ unsigned long offset, size_t size, int prot,
+ enum dma_data_direction dir,
+ unsigned long attrs, dma_addr_t *handle) {
+ bool skip_cpu_sync = attrs & DMA_ATTR_SKIP_CPU_SYNC;
+ struct io_bounce_buffer_info info;
+ bool force_bounce = iova_offset(buffers->iovad, offset | size);
+
+ if (!force_bounce)
+ return false;
+
+ *handle = DMA_MAPPING_ERROR;
+ if (!io_buffer_manager_alloc_buffer(&buffers->manager, dev, page,
+ offset + size, prot, buffers->nid,
+ &info))
+ return true;
+
+ if (!skip_cpu_sync)
+ io_bounce_buffers_do_sync(buffers, info.bounce_buffer, offset,
+ page, offset, size, dir, prot, false);
+
+ if (!io_bounce_buffers_map_buffer(buffers, &info, prot)) {
+ io_buffer_manager_release_buffer(
+ &buffers->manager, buffers->domain, info.iova, false);
+ return true;
+ }
+
+ *handle = info.iova + offset;
+ return true;
+}
+
+bool io_bounce_buffers_map_sg(struct io_bounce_buffers *buffers,
+ struct device *dev, struct scatterlist *sgl,
+ int nents, int prot, enum dma_data_direction dir,
+ unsigned long attrs, int *out_nents) {
+ struct io_bounce_buffer_info info;
+ struct scatterlist *iter;
+ size_t size = 0;
+ bool skip_cpu_sync = attrs & DMA_ATTR_SKIP_CPU_SYNC;
+ dma_addr_t seg_iova;
+ int i;
+ bool force_bounce = false;
+
+ for_each_sg(sgl, iter, nents, i) {
+ size += iter->length;
+ force_bounce |= iova_offset(buffers->iovad,
+ iter->offset | iter->length);
+ }
+
+ if (!force_bounce)
+ return false;
+
+ *out_nents = 0;
+ if (!io_buffer_manager_alloc_buffer(&buffers->manager, dev, sgl, size,
+ prot, buffers->nid, &info))
+ return true;
+
+ if (!skip_cpu_sync)
+ __io_bounce_buffers_sync_sg(buffers, sgl, nents,
+ info.bounce_buffer, dir, prot,
+ false);
+
+ if (!io_bounce_buffers_map_buffer(buffers, &info, prot)) {
+ io_buffer_manager_release_buffer(
+ &buffers->manager, buffers->domain, info.iova, false);
+ return true;
+ }
+
+ i = 0;
+ seg_iova = info.iova;
+ while (size > 0) {
+ size_t seg_size = min_t(size_t, size,
+ dma_get_max_seg_size(dev));
+
+ sg_dma_len(sgl) = seg_size;
+ sg_dma_address(sgl) = seg_iova;
+
+ sgl = sg_next(sgl);
+ size -= seg_size;
+ seg_iova += seg_size;
+ i++;
+ }
+
+ if (sgl) {
+ sg_dma_address(sgl) = DMA_MAPPING_ERROR;
+ sg_dma_len(sgl) = 0;
+ }
+
+ *out_nents = i;
+ return true;
+}
diff --git a/drivers/iommu/io-bounce-buffers.h b/drivers/iommu/io-bounce-buffers.h
new file mode 100644
index 000000000000..6d132a27646c
--- /dev/null
+++ b/drivers/iommu/io-bounce-buffers.h
@@ -0,0 +1,46 @@
+/* SPDX-License-Identifier: GPL-2.0-only */
+/*
+ * Copyright (C) 2021 Google, Inc.
+ */
+
+#ifndef _LINUX_IO_BOUNCE_BUFFERS_H
+#define _LINUX_IO_BOUNCE_BUFFERS_H
+
+#include <linux/dma-iommu.h>
+#include <linux/iova.h>
+
+struct io_bounce_buffers;
+
+struct io_bounce_buffers *io_bounce_buffers_init(struct device *dev,
+ struct iommu_domain *domain,
+ struct iova_domain *iovad);
+void io_bounce_buffers_destroy(struct io_bounce_buffers *buffers);
+
+bool io_bounce_buffers_sync_single(struct io_bounce_buffers *buffers,
+ dma_addr_t dma_handle, size_t size,
+ enum dma_data_direction dir,
+ bool sync_for_cpu);
+bool io_bounce_buffers_sync_sg(struct io_bounce_buffers *buffers,
+ struct scatterlist *sgl, int nents,
+ enum dma_data_direction dir, bool sync_for_cpu);
+
+bool io_bounce_buffers_map_page(struct io_bounce_buffers *buffers,
+ struct device *dev, struct page *page,
+ unsigned long offset, size_t size, int prot,
+ enum dma_data_direction dir,
+ unsigned long attrs, dma_addr_t *handle); bool
+io_bounce_buffers_map_sg(struct io_bounce_buffers *buffers,
+ struct device *dev, struct scatterlist *sgl,
+ int nents, int prot, enum dma_data_direction dir,
+ unsigned long attrs, int *out_nents);
+
+bool io_bounce_buffers_unmap_page(struct io_bounce_buffers *buffers,
+ dma_addr_t handle, size_t size,
+ enum dma_data_direction dir,
+ unsigned long attrs);
+bool io_bounce_buffers_unmap_sg(struct io_bounce_buffers *buffers,
+ struct scatterlist *sgl, int nents,
+ enum dma_data_direction dir,
+ unsigned long attrs);
+
+#endif /* _LINUX_IO_BOUNCE_BUFFERS_H */
diff --git a/drivers/iommu/io-buffer-manager.c b/drivers/iommu/io-buffer-manager.c
new file mode 100644
index 000000000000..24e95a2faa37
--- /dev/null
+++ b/drivers/iommu/io-buffer-manager.c
@@ -0,0 +1,212 @@
+// SPDX-License-Identifier: GPL-2.0-only
+/*
+ * Manager which allocates and tracks bounce buffers and their IOVAs.
+Does
+ * not actually manage the IOMMU mapping nor do the bounce copies.
+ *
+ * Copyright (C) 2021 Google, Inc.
+ */
+
+#include "io-buffer-manager.h"
+
+#include <linux/slab.h>
+
+struct io_buffer_node {
+ struct rb_node node;
+ struct io_bounce_buffer_info info;
+ void *orig_buffer;
+ int prot;
+};
+
+static void io_buffer_manager_free_pages(struct page **pages, int
+count) {
+ while (count--)
+ __free_page(pages[count]);
+ kfree(pages);
+}
+
+static struct page **io_buffer_manager_alloc_pages(int count, unsigned
+int nid) {
+ struct page **pages;
+ unsigned int i;
+
+ pages = kmalloc_array(count, sizeof(*pages), GFP_ATOMIC);
+ if (!pages)
+ return NULL;
+
+ // The IOMMU can map highmem pages, but try to allocate non-highmem
+ // pages first to make accessing the buffer cheaper.
+ for (i = 0; i < count; i++) {
+ pages[i] = alloc_pages_node(
+ nid,
+ GFP_ATOMIC | __GFP_ZERO | __GFP_NORETRY | __GFP_NOWARN,
+ 0);
+ if (!pages[i]) {
+ pages[i] = alloc_pages_node(
+ nid, GFP_ATOMIC | __GFP_ZERO | __GFP_HIGHMEM,
+ 0);
+ if (!pages[i]) {
+ io_buffer_manager_free_pages(pages, i);
+ return NULL;
+ }
+ }
+ }
+
+ return pages;
+}
+
+struct io_buffer_node *find_fallback_node(struct rb_root *root,
+dma_addr_t iova) {
+ struct rb_node *node = root->rb_node;
+
+ while (node) {
+ struct io_buffer_node *cur =
+ container_of(node, struct io_buffer_node, node);
+
+ if (iova < cur->info.iova)
+ node = node->rb_left;
+ else if (iova >= cur->info.iova + cur->info.size)
+ node = node->rb_right;
+ else
+ return cur;
+ }
+ return NULL;
+}
+
+bool insert_fallback_node(struct rb_root *root, struct io_buffer_node
+*node) {
+ struct rb_node **new = &(root->rb_node), *parent = NULL;
+ dma_addr_t node_end = node->info.iova + node->info.size;
+
+ while (*new) {
+ struct io_buffer_node *cur =
+ container_of(*new, struct io_buffer_node, node);
+ dma_addr_t cur_end = cur->info.iova + cur->info.size;
+
+ parent = *new;
+ if (node_end <= cur->info.iova)
+ new = &((*new)->rb_left);
+ else if (node->info.iova >= cur_end)
+ new = &((*new)->rb_right);
+ else {
+ pr_crit("IOVA collision new=[%llx,%llx) old=[%llx,%llx)\n",
+ node->info.iova, node_end, cur->info.iova,
+ cur_end);
+ return false;
+ }
+ }
+
+ rb_link_node(&node->node, parent, new);
+ rb_insert_color(&node->node, root);
+ return true;
+}
+
+bool io_buffer_manager_alloc_buffer(struct io_buffer_manager *manager,
+ struct device *dev, void *orig_buffer,
+ size_t size, int prot, unsigned int nid,
+ struct io_bounce_buffer_info *info) {
+ struct iommu_domain *domain = iommu_get_dma_domain(dev);
+ struct io_buffer_node *node;
+ unsigned long flags;
+
+ node = kzalloc(sizeof(*node), GFP_ATOMIC);
+ if (!node)
+ return false;
+
+ size = PAGE_ALIGN(size);
+ node->info.iova =
+ __iommu_dma_alloc_iova(domain, size, dma_get_mask(dev), dev);
+ if (!node->info.iova)
+ goto free_node;
+
+ node->info.bounce_buffer =
+ io_buffer_manager_alloc_pages(size >> PAGE_SHIFT, nid);
+ if (!node->info.bounce_buffer)
+ goto free_iova;
+
+ spin_lock_irqsave(&manager->fallback_lock, flags);
+ if (!insert_fallback_node(&manager->fallback_buffers, node))
+ goto fallback_lock_unlock;
+ spin_unlock_irqrestore(&manager->fallback_lock, flags);
+
+ node->orig_buffer = orig_buffer;
+ node->prot = prot;
+ node->info.size = size;
+
+ *info = node->info;
+
+ return true;
+
+fallback_lock_unlock:
+ spin_unlock_irqrestore(&manager->fallback_lock, flags);
+free_iova:
+ __iommu_dma_free_iova(domain->iova_cookie, node->info.iova, size,
+NULL);
+free_node:
+ kfree(node);
+ return false;
+}
+
+bool io_buffer_manager_find_buffer(struct io_buffer_manager *manager,
+ dma_addr_t handle,
+ struct io_bounce_buffer_info *info,
+ void **orig_buffer, int *prot)
+{
+ struct io_buffer_node *node;
+ unsigned long flags;
+
+ spin_lock_irqsave(&manager->fallback_lock, flags);
+ node = find_fallback_node(&manager->fallback_buffers, handle);
+ spin_unlock_irqrestore(&manager->fallback_lock, flags);
+
+ if (!node)
+ return false;
+
+ *info = node->info;
+ *orig_buffer = node->orig_buffer;
+ *prot = node->prot;
+ return true;
+}
+
+bool io_buffer_manager_release_buffer(struct io_buffer_manager *manager,
+ struct iommu_domain *domain,
+ dma_addr_t handle, bool inited) {
+ struct io_buffer_node *node;
+ unsigned long flags;
+ bool free_buffer;
+
+ spin_lock_irqsave(&manager->fallback_lock, flags);
+ node = find_fallback_node(&manager->fallback_buffers, handle);
+ if (node)
+ rb_erase(&node->node, &manager->fallback_buffers);
+ spin_unlock_irqrestore(&manager->fallback_lock, flags);
+
+ if (!node)
+ return false;
+
+ if (inited)
+ free_buffer = io_bounce_buffers_release_buffer_cb(
+ manager, node->info.iova, node->info.size);
+ else
+ free_buffer = true;
+
+ if (free_buffer) {
+ io_buffer_manager_free_pages(node->info.bounce_buffer,
+ node->info.size >> PAGE_SHIFT);
+ __iommu_dma_free_iova(domain->iova_cookie, node->info.iova,
+ node->info.size, NULL);
+ } else {
+ pr_warn("Bounce buffer release failed; leaking buffer\n");
+ }
+
+ kfree(node);
+ return true;
+}
+
+int io_buffer_manager_init(struct io_buffer_manager *manager) {
+ manager->fallback_buffers = RB_ROOT;
+ spin_lock_init(&manager->fallback_lock);
+
+ return 0;
+}
diff --git a/drivers/iommu/io-buffer-manager.h b/drivers/iommu/io-buffer-manager.h
new file mode 100644
index 000000000000..aae560cc8512
--- /dev/null
+++ b/drivers/iommu/io-buffer-manager.h
@@ -0,0 +1,43 @@
+/* SPDX-License-Identifier: GPL-2.0-only */
+/*
+ * Copyright (C) 2021 Google, Inc.
+ */
+
+#ifndef _LINUX_IO_BUFFER_MANAGER_H
+#define _LINUX_IO_BUFFER_MANAGER_H
+
+#include <linux/dma-iommu.h>
+#include <linux/iova.h>
+#include <linux/spinlock.h>
+
+struct io_buffer_manager {
+ spinlock_t fallback_lock;
+ struct rb_root fallback_buffers;
+};
+
+struct io_bounce_buffer_info {
+ struct page **bounce_buffer;
+ dma_addr_t iova;
+ size_t size;
+};
+
+bool io_buffer_manager_alloc_buffer(struct io_buffer_manager *manager,
+ struct device *dev, void *orig_buffer,
+ size_t size, int prot, unsigned int nid,
+ struct io_bounce_buffer_info *info);
+
+bool io_buffer_manager_find_buffer(struct io_buffer_manager *manager,
+ dma_addr_t handle,
+ struct io_bounce_buffer_info *info,
+ void **orig_buffer, int *prot);
+
+bool io_buffer_manager_release_buffer(struct io_buffer_manager *manager,
+ struct iommu_domain *domain,
+ dma_addr_t handle, bool inited);
+
+int io_buffer_manager_init(struct io_buffer_manager *manager);
+
+bool io_bounce_buffers_release_buffer_cb(struct io_buffer_manager *manager,
+ dma_addr_t iova, size_t size);
+
+#endif /* _LINUX_IO_BUFFER_MANAGER_H */
--
2.32.0.605.g8dce9f2422-goog

_______________________________________________
iommu mailing list
[email protected]
https://lists.linuxfoundation.org/mailman/listinfo/iommu

2021-08-10 05:18:15

by David Stevens

[permalink] [raw]
Subject: Re: [PATCH v2 3/9] dma-iommu: bounce buffers for untrusted devices

On Tue, Aug 10, 2021 at 10:19 AM Mi, Dapeng1 <[email protected]> wrote:
>
> Hi David,
>
> I like this patch set and this is crucial for reducing the significant vIOMMU performance. It looks you totally rewrite the IOMMU mapping/unmapping part and use the dynamically allocated memory from buddy system as bounce buffer instead of using the legacy SWIOTLB bounce buffer. As I know, some legacy devices' DMA could not access the memory larger than 32-bit memory space and the dynamically allocated memory address could exceed the 32-bit memory space. Is it a problem?

My understanding is that when devices with that sort of limitation sit
behind an IOMMU, the IOVA is what matters, not the physical address.
The bounce bounce buffers use the same limits for IOVA allocation as
the regular dma-iommu path, so compatible IOVAs will be allocated for
the bounce buffers.

-David

> Thx,
> Dapeng Mi