This patch adds two new regions aiming to handle nested mode
translation faults.
The first region (two host kernel pages) is read-only from the
user-space perspective. The first page contains an header
that provides information about the circular buffer located in the
second page. The circular buffer is put in a different page in
the prospect to be mmappable.
The max user API version supported by the kernel is returned
through a dedicated fault region capability.
The prod header contains
- the user API version in use (potentially inferior to the one
returned in the capability),
- the offset of the queue within the region,
- the producer index relative to the start of the queue
- the max number of fault records,
- the size of each record.
The second region is write-only from the user perspective. It
contains the version of the requested fault ABI and the consumer
index that is updated by the userspace each time this latter has
consumed fault records.
The natural order of operation for the userspace is:
- retrieve the highest supported fault ABI version
- set the requested fault ABI version in the consumer region
Until the ABI version is not set by the userspace, the kernel
cannot return a comprehensive set of information inside the
prod header (entry size and number of entries in the fault queue).
Signed-off-by: Eric Auger <[email protected]>
---
v4 -> v5
- check cons is not null in vfio_pci_check_cons_fault
v3 -> v4:
- use 2 separate regions, respectively in read and write modes
- add the version capability
---
drivers/vfio/pci/vfio_pci.c | 105 ++++++++++++++++++++++++++++
drivers/vfio/pci/vfio_pci_private.h | 17 +++++
drivers/vfio/pci/vfio_pci_rdwr.c | 73 +++++++++++++++++++
include/uapi/linux/vfio.h | 42 +++++++++++
4 files changed, 237 insertions(+)
diff --git a/drivers/vfio/pci/vfio_pci.c b/drivers/vfio/pci/vfio_pci.c
index cab71da46f4a..f75f61127277 100644
--- a/drivers/vfio/pci/vfio_pci.c
+++ b/drivers/vfio/pci/vfio_pci.c
@@ -261,6 +261,106 @@ int vfio_pci_set_power_state(struct vfio_pci_device *vdev, pci_power_t state)
return ret;
}
+void vfio_pci_fault_release(struct vfio_pci_device *vdev,
+ struct vfio_pci_region *region)
+{
+}
+
+static const struct vfio_pci_fault_abi fault_abi_versions[] = {
+ [0] = {
+ .entry_size = sizeof(struct iommu_fault),
+ },
+};
+
+#define NR_FAULT_ABIS ARRAY_SIZE(fault_abi_versions)
+
+static int vfio_pci_fault_prod_add_capability(struct vfio_pci_device *vdev,
+ struct vfio_pci_region *region, struct vfio_info_cap *caps)
+{
+ struct vfio_region_info_cap_fault cap = {
+ .header.id = VFIO_REGION_INFO_CAP_PRODUCER_FAULT,
+ .header.version = 1,
+ .version = NR_FAULT_ABIS,
+ };
+ return vfio_info_add_capability(caps, &cap.header, sizeof(cap));
+}
+
+static const struct vfio_pci_regops vfio_pci_fault_cons_regops = {
+ .rw = vfio_pci_fault_cons_rw,
+ .release = vfio_pci_fault_release,
+};
+
+static const struct vfio_pci_regops vfio_pci_fault_prod_regops = {
+ .rw = vfio_pci_fault_prod_rw,
+ .release = vfio_pci_fault_release,
+ .add_capability = vfio_pci_fault_prod_add_capability,
+};
+
+static int vfio_pci_init_fault_region(struct vfio_pci_device *vdev)
+{
+ struct vfio_region_fault_prod *header;
+ int ret;
+
+ mutex_init(&vdev->fault_queue_lock);
+
+ vdev->fault_pages = kzalloc(3 * PAGE_SIZE, GFP_KERNEL);
+ if (!vdev->fault_pages)
+ return -ENOMEM;
+
+ ret = vfio_pci_register_dev_region(vdev,
+ VFIO_REGION_TYPE_NESTED,
+ VFIO_REGION_SUBTYPE_NESTED_FAULT_PROD,
+ &vfio_pci_fault_prod_regops, 2 * PAGE_SIZE,
+ VFIO_REGION_INFO_FLAG_READ, vdev->fault_pages);
+ if (ret)
+ goto out;
+
+ ret = vfio_pci_register_dev_region(vdev,
+ VFIO_REGION_TYPE_NESTED,
+ VFIO_REGION_SUBTYPE_NESTED_FAULT_CONS,
+ &vfio_pci_fault_cons_regops,
+ sizeof(struct vfio_region_fault_cons),
+ VFIO_REGION_INFO_FLAG_WRITE,
+ vdev->fault_pages + 2 * PAGE_SIZE);
+ if (ret)
+ goto out;
+
+ header = (struct vfio_region_fault_prod *)vdev->fault_pages;
+ header->version = -1;
+ header->offset = PAGE_SIZE;
+ return 0;
+out:
+ kfree(vdev->fault_pages);
+ return ret;
+}
+
+int vfio_pci_check_cons_fault(struct vfio_pci_device *vdev,
+ struct vfio_region_fault_cons *cons_header)
+{
+ struct vfio_region_fault_prod *prod_header =
+ (struct vfio_region_fault_prod *)vdev->fault_pages;
+
+ if (cons_header->version > NR_FAULT_ABIS)
+ return -EINVAL;
+
+ if (!vdev->fault_abi) {
+ vdev->fault_abi = cons_header->version;
+ prod_header->entry_size =
+ fault_abi_versions[vdev->fault_abi - 1].entry_size;
+ prod_header->nb_entries = PAGE_SIZE / prod_header->entry_size;
+ return 0;
+ }
+
+ /* Fault ABI is set */
+ if (cons_header->version != vdev->fault_abi)
+ return -EINVAL;
+
+ if (cons_header->cons && cons_header->cons >= prod_header->nb_entries)
+ return -EINVAL;
+
+ return 0;
+}
+
static int vfio_pci_enable(struct vfio_pci_device *vdev)
{
struct pci_dev *pdev = vdev->pdev;
@@ -359,6 +459,10 @@ static int vfio_pci_enable(struct vfio_pci_device *vdev)
}
}
+ ret = vfio_pci_init_fault_region(vdev);
+ if (ret)
+ goto disable_exit;
+
vfio_pci_probe_mmaps(vdev);
return 0;
@@ -1374,6 +1478,7 @@ static void vfio_pci_remove(struct pci_dev *pdev)
vfio_iommu_group_put(pdev->dev.iommu_group, &pdev->dev);
kfree(vdev->region);
+ kfree(vdev->fault_pages);
mutex_destroy(&vdev->ioeventfds_lock);
if (!disable_idle_d3)
diff --git a/drivers/vfio/pci/vfio_pci_private.h b/drivers/vfio/pci/vfio_pci_private.h
index 1812cf22fc4f..8e0a55682d3f 100644
--- a/drivers/vfio/pci/vfio_pci_private.h
+++ b/drivers/vfio/pci/vfio_pci_private.h
@@ -122,9 +122,12 @@ struct vfio_pci_device {
int ioeventfds_nr;
struct eventfd_ctx *err_trigger;
struct eventfd_ctx *req_trigger;
+ struct mutex fault_queue_lock;
+ int fault_abi;
struct list_head dummy_resources_list;
struct mutex ioeventfds_lock;
struct list_head ioeventfds_list;
+ u8 *fault_pages;
};
#define is_intx(vdev) (vdev->irq_type == VFIO_PCI_INTX_IRQ_INDEX)
@@ -153,6 +156,18 @@ extern ssize_t vfio_pci_vga_rw(struct vfio_pci_device *vdev, char __user *buf,
extern long vfio_pci_ioeventfd(struct vfio_pci_device *vdev, loff_t offset,
uint64_t data, int count, int fd);
+struct vfio_pci_fault_abi {
+ u32 entry_size;
+};
+
+extern size_t vfio_pci_fault_cons_rw(struct vfio_pci_device *vdev,
+ char __user *buf, size_t count,
+ loff_t *ppos, bool iswrite);
+
+extern size_t vfio_pci_fault_prod_rw(struct vfio_pci_device *vdev,
+ char __user *buf, size_t count,
+ loff_t *ppos, bool iswrite);
+
extern int vfio_pci_init_perm_bits(void);
extern void vfio_pci_uninit_perm_bits(void);
@@ -166,6 +181,8 @@ extern int vfio_pci_register_dev_region(struct vfio_pci_device *vdev,
extern int vfio_pci_set_power_state(struct vfio_pci_device *vdev,
pci_power_t state);
+extern int vfio_pci_check_cons_fault(struct vfio_pci_device *vdev,
+ struct vfio_region_fault_cons *header);
#ifdef CONFIG_VFIO_PCI_IGD
extern int vfio_pci_igd_init(struct vfio_pci_device *vdev);
diff --git a/drivers/vfio/pci/vfio_pci_rdwr.c b/drivers/vfio/pci/vfio_pci_rdwr.c
index a6029d0a5524..67cd9363f4e7 100644
--- a/drivers/vfio/pci/vfio_pci_rdwr.c
+++ b/drivers/vfio/pci/vfio_pci_rdwr.c
@@ -277,6 +277,79 @@ ssize_t vfio_pci_vga_rw(struct vfio_pci_device *vdev, char __user *buf,
return done;
}
+/* Read-only region */
+size_t vfio_pci_fault_prod_rw(struct vfio_pci_device *vdev, char __user *buf,
+ size_t count, loff_t *ppos, bool iswrite)
+{
+ unsigned int i = VFIO_PCI_OFFSET_TO_INDEX(*ppos) - VFIO_PCI_NUM_REGIONS;
+ void *base = vdev->region[i].data;
+ loff_t pos = *ppos & VFIO_PCI_OFFSET_MASK;
+ int ret = 0;
+
+ if (iswrite)
+ return 0;
+
+ if (!vdev->fault_abi)
+ return -EINVAL;
+
+ if (pos >= vdev->region[i].size)
+ return -EINVAL;
+
+ count = min(count, (size_t)(vdev->region[i].size - pos));
+
+ mutex_lock(&vdev->fault_queue_lock);
+
+ if (copy_to_user(buf, base + pos, count)) {
+ ret = -EFAULT;
+ goto unlock;
+ }
+ *ppos += count;
+ ret = count;
+unlock:
+ mutex_unlock(&vdev->fault_queue_lock);
+ return ret;
+}
+
+
+/* write only */
+size_t vfio_pci_fault_cons_rw(struct vfio_pci_device *vdev, char __user *buf,
+ size_t count, loff_t *ppos, bool iswrite)
+{
+ unsigned int i = VFIO_PCI_OFFSET_TO_INDEX(*ppos) - VFIO_PCI_NUM_REGIONS;
+ void *base = vdev->region[i].data;
+ loff_t pos = *ppos & VFIO_PCI_OFFSET_MASK;
+ struct vfio_region_fault_cons *header;
+ struct vfio_region_fault_cons orig_header =
+ *(struct vfio_region_fault_cons *)base;
+ int ret = 0;
+
+ if (!iswrite)
+ return 0;
+
+ if (pos >= vdev->region[i].size)
+ return -EINVAL;
+
+ count = min(count, (size_t)(vdev->region[i].size - pos));
+
+ mutex_lock(&vdev->fault_queue_lock);
+
+ if (copy_from_user(base + pos, buf, count)) {
+ ret = -EFAULT;
+ goto unlock;
+ }
+ header = (struct vfio_region_fault_cons *)base;
+ ret = vfio_pci_check_cons_fault(vdev, header);
+ if (ret) {
+ *header = orig_header;
+ goto unlock;
+ }
+ *ppos += count;
+ ret = count;
+unlock:
+ mutex_unlock(&vdev->fault_queue_lock);
+ return ret;
+}
+
static int vfio_pci_ioeventfd_handler(void *opaque, void *unused)
{
struct vfio_pci_ioeventfd *ioeventfd = opaque;
diff --git a/include/uapi/linux/vfio.h b/include/uapi/linux/vfio.h
index 2774a1ab37ae..13e041b84d48 100644
--- a/include/uapi/linux/vfio.h
+++ b/include/uapi/linux/vfio.h
@@ -307,6 +307,10 @@ struct vfio_region_info_cap_type {
#define VFIO_REGION_TYPE_GFX (1)
#define VFIO_REGION_SUBTYPE_GFX_EDID (1)
+#define VFIO_REGION_TYPE_NESTED (2)
+#define VFIO_REGION_SUBTYPE_NESTED_FAULT_PROD (1)
+#define VFIO_REGION_SUBTYPE_NESTED_FAULT_CONS (2)
+
/**
* struct vfio_region_gfx_edid - EDID region layout.
*
@@ -701,6 +705,44 @@ struct vfio_device_ioeventfd {
#define VFIO_DEVICE_IOEVENTFD _IO(VFIO_TYPE, VFIO_BASE + 16)
+
+/*
+ * Capability exposed by the Producer Fault Region
+ * @version: max fault ABI version supported by the kernel
+ */
+#define VFIO_REGION_INFO_CAP_PRODUCER_FAULT 6
+
+struct vfio_region_info_cap_fault {
+ struct vfio_info_cap_header header;
+ __u32 version;
+};
+
+/*
+ * Producer Fault Region (Read-Only from user space perspective)
+ * Contains the fault circular buffer and the producer index
+ * @version: version of the fault record uapi
+ * @entry_size: size of each fault record
+ * @offset: offset of the start of the queue
+ * @prod: producer index relative to the start of the queue
+ */
+struct vfio_region_fault_prod {
+ __u32 version;
+ __u32 nb_entries;
+ __u32 entry_size;
+ __u32 offset;
+ __u32 prod;
+};
+
+/*
+ * Consumer Fault Region (Write-Only from the user space perspective)
+ * @version: ABI version requested by the userspace
+ * @cons: consumer index relative to the start of the queue
+ */
+struct vfio_region_fault_cons {
+ __u32 version;
+ __u32 cons;
+};
+
/* -------- API for Type1 VFIO IOMMU -------- */
/**
--
2.20.1
On Sun, 26 May 2019 18:10:00 +0200
Eric Auger <[email protected]> wrote:
> This patch adds two new regions aiming to handle nested mode
> translation faults.
>
> The first region (two host kernel pages) is read-only from the
> user-space perspective. The first page contains an header
> that provides information about the circular buffer located in the
> second page. The circular buffer is put in a different page in
> the prospect to be mmappable.
>
> The max user API version supported by the kernel is returned
> through a dedicated fault region capability.
>
> The prod header contains
> - the user API version in use (potentially inferior to the one
> returned in the capability),
> - the offset of the queue within the region,
> - the producer index relative to the start of the queue
> - the max number of fault records,
> - the size of each record.
>
> The second region is write-only from the user perspective. It
> contains the version of the requested fault ABI and the consumer
> index that is updated by the userspace each time this latter has
> consumed fault records.
>
> The natural order of operation for the userspace is:
> - retrieve the highest supported fault ABI version
> - set the requested fault ABI version in the consumer region
>
> Until the ABI version is not set by the userspace, the kernel
> cannot return a comprehensive set of information inside the
> prod header (entry size and number of entries in the fault queue).
It's not clear to me why two regions are required for this. If the
first page is not mmap capable, why does it need to be read-only? If
it were not read-only couldn't the fields of the second region also fit
within this first page? If you wanted to deal with an mmap capable
writeable region, it could just be yet a 3rd page in the first region.
>
> Signed-off-by: Eric Auger <[email protected]>
>
> ---
>
> v4 -> v5
> - check cons is not null in vfio_pci_check_cons_fault
>
> v3 -> v4:
> - use 2 separate regions, respectively in read and write modes
> - add the version capability
> ---
> drivers/vfio/pci/vfio_pci.c | 105 ++++++++++++++++++++++++++++
> drivers/vfio/pci/vfio_pci_private.h | 17 +++++
> drivers/vfio/pci/vfio_pci_rdwr.c | 73 +++++++++++++++++++
> include/uapi/linux/vfio.h | 42 +++++++++++
> 4 files changed, 237 insertions(+)
>
> diff --git a/drivers/vfio/pci/vfio_pci.c b/drivers/vfio/pci/vfio_pci.c
> index cab71da46f4a..f75f61127277 100644
> --- a/drivers/vfio/pci/vfio_pci.c
> +++ b/drivers/vfio/pci/vfio_pci.c
> @@ -261,6 +261,106 @@ int vfio_pci_set_power_state(struct vfio_pci_device *vdev, pci_power_t state)
> return ret;
> }
>
> +void vfio_pci_fault_release(struct vfio_pci_device *vdev,
> + struct vfio_pci_region *region)
> +{
> +}
> +
> +static const struct vfio_pci_fault_abi fault_abi_versions[] = {
> + [0] = {
> + .entry_size = sizeof(struct iommu_fault),
> + },
> +};
> +
> +#define NR_FAULT_ABIS ARRAY_SIZE(fault_abi_versions)
This looks like it's leading to some dangerous complicated code to
support multiple user selected ABIs. How many ABIs do we plan to
support? The region capability also exposes a type, sub-type, and
version. How much of this could be exposed that way? ie. if we need
to support multiple versions, expose multiple regions.
> +
> +static int vfio_pci_fault_prod_add_capability(struct vfio_pci_device *vdev,
> + struct vfio_pci_region *region, struct vfio_info_cap *caps)
> +{
> + struct vfio_region_info_cap_fault cap = {
> + .header.id = VFIO_REGION_INFO_CAP_PRODUCER_FAULT,
> + .header.version = 1,
> + .version = NR_FAULT_ABIS,
> + };
> + return vfio_info_add_capability(caps, &cap.header, sizeof(cap));
> +}
> +
> +static const struct vfio_pci_regops vfio_pci_fault_cons_regops = {
> + .rw = vfio_pci_fault_cons_rw,
> + .release = vfio_pci_fault_release,
> +};
> +
> +static const struct vfio_pci_regops vfio_pci_fault_prod_regops = {
> + .rw = vfio_pci_fault_prod_rw,
> + .release = vfio_pci_fault_release,
> + .add_capability = vfio_pci_fault_prod_add_capability,
> +};
> +
> +static int vfio_pci_init_fault_region(struct vfio_pci_device *vdev)
> +{
> + struct vfio_region_fault_prod *header;
> + int ret;
> +
> + mutex_init(&vdev->fault_queue_lock);
> +
> + vdev->fault_pages = kzalloc(3 * PAGE_SIZE, GFP_KERNEL);
> + if (!vdev->fault_pages)
> + return -ENOMEM;
> +
> + ret = vfio_pci_register_dev_region(vdev,
> + VFIO_REGION_TYPE_NESTED,
> + VFIO_REGION_SUBTYPE_NESTED_FAULT_PROD,
> + &vfio_pci_fault_prod_regops, 2 * PAGE_SIZE,
> + VFIO_REGION_INFO_FLAG_READ, vdev->fault_pages);
If mmap isn't supported yet, why are we pushing the queue out to the
2nd page? We're just wasting space. vfio_region_fault_prod.offset
allows us to relocate it when/if it is mmap capable.
> + if (ret)
> + goto out;
> +
> + ret = vfio_pci_register_dev_region(vdev,
> + VFIO_REGION_TYPE_NESTED,
> + VFIO_REGION_SUBTYPE_NESTED_FAULT_CONS,
> + &vfio_pci_fault_cons_regops,
> + sizeof(struct vfio_region_fault_cons),
> + VFIO_REGION_INFO_FLAG_WRITE,
> + vdev->fault_pages + 2 * PAGE_SIZE);
What's the remaining (PAGE_SIZE - sizeof(struct vfio_region_fault_cons))
bytes used for?
> + if (ret)
> + goto out;
> +
> + header = (struct vfio_region_fault_prod *)vdev->fault_pages;
> + header->version = -1;
> + header->offset = PAGE_SIZE;
> + return 0;
> +out:
> + kfree(vdev->fault_pages);
> + return ret;
> +}
> +
> +int vfio_pci_check_cons_fault(struct vfio_pci_device *vdev,
> + struct vfio_region_fault_cons *cons_header)
> +{
> + struct vfio_region_fault_prod *prod_header =
> + (struct vfio_region_fault_prod *)vdev->fault_pages;
> +
> + if (cons_header->version > NR_FAULT_ABIS)
> + return -EINVAL;
> +
> + if (!vdev->fault_abi) {
> + vdev->fault_abi = cons_header->version;
> + prod_header->entry_size =
> + fault_abi_versions[vdev->fault_abi - 1].entry_size;
> + prod_header->nb_entries = PAGE_SIZE / prod_header->entry_size;
Is this sufficient for 4K hosts? Clearly a 64K host has 16x the number
of entries, so if this is a heuristic the results are vastly different.
> + return 0;
> + }
> +
> + /* Fault ABI is set */
> + if (cons_header->version != vdev->fault_abi)
> + return -EINVAL;
> +
> + if (cons_header->cons && cons_header->cons >= prod_header->nb_entries)
First test seems unnecessary.
> + return -EINVAL;
> +
> + return 0;
> +}
> +
> static int vfio_pci_enable(struct vfio_pci_device *vdev)
> {
> struct pci_dev *pdev = vdev->pdev;
> @@ -359,6 +459,10 @@ static int vfio_pci_enable(struct vfio_pci_device *vdev)
> }
> }
>
> + ret = vfio_pci_init_fault_region(vdev);
> + if (ret)
> + goto disable_exit;
> +
> vfio_pci_probe_mmaps(vdev);
>
> return 0;
> @@ -1374,6 +1478,7 @@ static void vfio_pci_remove(struct pci_dev *pdev)
>
> vfio_iommu_group_put(pdev->dev.iommu_group, &pdev->dev);
> kfree(vdev->region);
> + kfree(vdev->fault_pages);
> mutex_destroy(&vdev->ioeventfds_lock);
>
> if (!disable_idle_d3)
> diff --git a/drivers/vfio/pci/vfio_pci_private.h b/drivers/vfio/pci/vfio_pci_private.h
> index 1812cf22fc4f..8e0a55682d3f 100644
> --- a/drivers/vfio/pci/vfio_pci_private.h
> +++ b/drivers/vfio/pci/vfio_pci_private.h
> @@ -122,9 +122,12 @@ struct vfio_pci_device {
> int ioeventfds_nr;
> struct eventfd_ctx *err_trigger;
> struct eventfd_ctx *req_trigger;
> + struct mutex fault_queue_lock;
> + int fault_abi;
> struct list_head dummy_resources_list;
> struct mutex ioeventfds_lock;
> struct list_head ioeventfds_list;
> + u8 *fault_pages;
> };
>
> #define is_intx(vdev) (vdev->irq_type == VFIO_PCI_INTX_IRQ_INDEX)
> @@ -153,6 +156,18 @@ extern ssize_t vfio_pci_vga_rw(struct vfio_pci_device *vdev, char __user *buf,
> extern long vfio_pci_ioeventfd(struct vfio_pci_device *vdev, loff_t offset,
> uint64_t data, int count, int fd);
>
> +struct vfio_pci_fault_abi {
> + u32 entry_size;
> +};
> +
> +extern size_t vfio_pci_fault_cons_rw(struct vfio_pci_device *vdev,
> + char __user *buf, size_t count,
> + loff_t *ppos, bool iswrite);
> +
> +extern size_t vfio_pci_fault_prod_rw(struct vfio_pci_device *vdev,
> + char __user *buf, size_t count,
> + loff_t *ppos, bool iswrite);
> +
> extern int vfio_pci_init_perm_bits(void);
> extern void vfio_pci_uninit_perm_bits(void);
>
> @@ -166,6 +181,8 @@ extern int vfio_pci_register_dev_region(struct vfio_pci_device *vdev,
>
> extern int vfio_pci_set_power_state(struct vfio_pci_device *vdev,
> pci_power_t state);
> +extern int vfio_pci_check_cons_fault(struct vfio_pci_device *vdev,
> + struct vfio_region_fault_cons *header);
>
> #ifdef CONFIG_VFIO_PCI_IGD
> extern int vfio_pci_igd_init(struct vfio_pci_device *vdev);
> diff --git a/drivers/vfio/pci/vfio_pci_rdwr.c b/drivers/vfio/pci/vfio_pci_rdwr.c
> index a6029d0a5524..67cd9363f4e7 100644
> --- a/drivers/vfio/pci/vfio_pci_rdwr.c
> +++ b/drivers/vfio/pci/vfio_pci_rdwr.c
> @@ -277,6 +277,79 @@ ssize_t vfio_pci_vga_rw(struct vfio_pci_device *vdev, char __user *buf,
> return done;
> }
>
> +/* Read-only region */
> +size_t vfio_pci_fault_prod_rw(struct vfio_pci_device *vdev, char __user *buf,
> + size_t count, loff_t *ppos, bool iswrite)
> +{
> + unsigned int i = VFIO_PCI_OFFSET_TO_INDEX(*ppos) - VFIO_PCI_NUM_REGIONS;
> + void *base = vdev->region[i].data;
> + loff_t pos = *ppos & VFIO_PCI_OFFSET_MASK;
> + int ret = 0;
> +
> + if (iswrite)
> + return 0;
> +
> + if (!vdev->fault_abi)
> + return -EINVAL;
> +
> + if (pos >= vdev->region[i].size)
> + return -EINVAL;
> +
> + count = min(count, (size_t)(vdev->region[i].size - pos));
> +
> + mutex_lock(&vdev->fault_queue_lock);
> +
> + if (copy_to_user(buf, base + pos, count)) {
> + ret = -EFAULT;
> + goto unlock;
> + }
> + *ppos += count;
> + ret = count;
> +unlock:
> + mutex_unlock(&vdev->fault_queue_lock);
> + return ret;
> +}
> +
> +
> +/* write only */
> +size_t vfio_pci_fault_cons_rw(struct vfio_pci_device *vdev, char __user *buf,
> + size_t count, loff_t *ppos, bool iswrite)
> +{
> + unsigned int i = VFIO_PCI_OFFSET_TO_INDEX(*ppos) - VFIO_PCI_NUM_REGIONS;
> + void *base = vdev->region[i].data;
> + loff_t pos = *ppos & VFIO_PCI_OFFSET_MASK;
> + struct vfio_region_fault_cons *header;
> + struct vfio_region_fault_cons orig_header =
> + *(struct vfio_region_fault_cons *)base;
> + int ret = 0;
> +
> + if (!iswrite)
> + return 0;
> +
> + if (pos >= vdev->region[i].size)
> + return -EINVAL;
> +
> + count = min(count, (size_t)(vdev->region[i].size - pos));
> +
> + mutex_lock(&vdev->fault_queue_lock);
> +
> + if (copy_from_user(base + pos, buf, count)) {
> + ret = -EFAULT;
> + goto unlock;
> + }
> + header = (struct vfio_region_fault_cons *)base;
> + ret = vfio_pci_check_cons_fault(vdev, header);
> + if (ret) {
> + *header = orig_header;
> + goto unlock;
> + }
> + *ppos += count;
> + ret = count;
> +unlock:
> + mutex_unlock(&vdev->fault_queue_lock);
> + return ret;
> +}
> +
> static int vfio_pci_ioeventfd_handler(void *opaque, void *unused)
> {
> struct vfio_pci_ioeventfd *ioeventfd = opaque;
> diff --git a/include/uapi/linux/vfio.h b/include/uapi/linux/vfio.h
> index 2774a1ab37ae..13e041b84d48 100644
> --- a/include/uapi/linux/vfio.h
> +++ b/include/uapi/linux/vfio.h
> @@ -307,6 +307,10 @@ struct vfio_region_info_cap_type {
> #define VFIO_REGION_TYPE_GFX (1)
> #define VFIO_REGION_SUBTYPE_GFX_EDID (1)
>
> +#define VFIO_REGION_TYPE_NESTED (2)
> +#define VFIO_REGION_SUBTYPE_NESTED_FAULT_PROD (1)
> +#define VFIO_REGION_SUBTYPE_NESTED_FAULT_CONS (2)
> +
> /**
> * struct vfio_region_gfx_edid - EDID region layout.
> *
> @@ -701,6 +705,44 @@ struct vfio_device_ioeventfd {
>
> #define VFIO_DEVICE_IOEVENTFD _IO(VFIO_TYPE, VFIO_BASE + 16)
>
> +
> +/*
> + * Capability exposed by the Producer Fault Region
> + * @version: max fault ABI version supported by the kernel
> + */
> +#define VFIO_REGION_INFO_CAP_PRODUCER_FAULT 6
> +
> +struct vfio_region_info_cap_fault {
> + struct vfio_info_cap_header header;
> + __u32 version;
> +};
> +
> +/*
> + * Producer Fault Region (Read-Only from user space perspective)
> + * Contains the fault circular buffer and the producer index
> + * @version: version of the fault record uapi
> + * @entry_size: size of each fault record
> + * @offset: offset of the start of the queue
> + * @prod: producer index relative to the start of the queue
> + */
> +struct vfio_region_fault_prod {
> + __u32 version;
> + __u32 nb_entries;
> + __u32 entry_size;
> + __u32 offset;
> + __u32 prod;
> +};
> +
> +/*
> + * Consumer Fault Region (Write-Only from the user space perspective)
> + * @version: ABI version requested by the userspace
> + * @cons: consumer index relative to the start of the queue
> + */
> +struct vfio_region_fault_cons {
> + __u32 version;
> + __u32 cons;
> +};
I think there are more common semantics than this prod/cons
terminology, for example head/tail.
> +
> /* -------- API for Type1 VFIO IOMMU -------- */
>
> /**
Hi Alex,
On 6/4/19 12:31 AM, Alex Williamson wrote:
> On Sun, 26 May 2019 18:10:00 +0200
> Eric Auger <[email protected]> wrote:
>
>> This patch adds two new regions aiming to handle nested mode
>> translation faults.
>>
>> The first region (two host kernel pages) is read-only from the
>> user-space perspective. The first page contains an header
>> that provides information about the circular buffer located in the
>> second page. The circular buffer is put in a different page in
>> the prospect to be mmappable.
>>
>> The max user API version supported by the kernel is returned
>> through a dedicated fault region capability.
>>
>> The prod header contains
>> - the user API version in use (potentially inferior to the one
>> returned in the capability),
>> - the offset of the queue within the region,
>> - the producer index relative to the start of the queue
>> - the max number of fault records,
>> - the size of each record.
>>
>> The second region is write-only from the user perspective. It
>> contains the version of the requested fault ABI and the consumer
>> index that is updated by the userspace each time this latter has
>> consumed fault records.
>>
>> The natural order of operation for the userspace is:
>> - retrieve the highest supported fault ABI version
>> - set the requested fault ABI version in the consumer region
>>
>> Until the ABI version is not set by the userspace, the kernel
>> cannot return a comprehensive set of information inside the
>> prod header (entry size and number of entries in the fault queue).
>
> It's not clear to me why two regions are required for this. If the
> first page is not mmap capable, why does it need to be read-only? If
> it were not read-only couldn't the fields of the second region also fit
> within this first page? If you wanted to deal with an mmap capable
> writeable region, it could just be yet a 3rd page in the first region.
I thought it would be clearer for the userspace to have 2 separate
regions, one for the producer and one for the consumer. Otherwise I will
need to specify which fields are read-only or write-only. But this may
be more self-contained in a single region.
>
>>
>> Signed-off-by: Eric Auger <[email protected]>
>>
>> ---
>>
>> v4 -> v5
>> - check cons is not null in vfio_pci_check_cons_fault
>>
>> v3 -> v4:
>> - use 2 separate regions, respectively in read and write modes
>> - add the version capability
>> ---
>> drivers/vfio/pci/vfio_pci.c | 105 ++++++++++++++++++++++++++++
>> drivers/vfio/pci/vfio_pci_private.h | 17 +++++
>> drivers/vfio/pci/vfio_pci_rdwr.c | 73 +++++++++++++++++++
>> include/uapi/linux/vfio.h | 42 +++++++++++
>> 4 files changed, 237 insertions(+)
>>
>> diff --git a/drivers/vfio/pci/vfio_pci.c b/drivers/vfio/pci/vfio_pci.c
>> index cab71da46f4a..f75f61127277 100644
>> --- a/drivers/vfio/pci/vfio_pci.c
>> +++ b/drivers/vfio/pci/vfio_pci.c
>> @@ -261,6 +261,106 @@ int vfio_pci_set_power_state(struct vfio_pci_device *vdev, pci_power_t state)
>> return ret;
>> }
>>
>> +void vfio_pci_fault_release(struct vfio_pci_device *vdev,
>> + struct vfio_pci_region *region)
>> +{
>> +}
>> +
>> +static const struct vfio_pci_fault_abi fault_abi_versions[] = {
>> + [0] = {
>> + .entry_size = sizeof(struct iommu_fault),
>> + },
>> +};
>> +
>> +#define NR_FAULT_ABIS ARRAY_SIZE(fault_abi_versions)
>
> This looks like it's leading to some dangerous complicated code to
> support multiple user selected ABIs. How many ABIs do we plan to
> support? The region capability also exposes a type, sub-type, and
> version. How much of this could be exposed that way? ie. if we need
> to support multiple versions, expose multiple regions.
This is something that was discussed earlier and suggested by
Jean-Philippe that we may need to support several versions of the ABI
(typicallu when adding PRI support).
Exposing multiple region is an interesting idea and I will explore that
direction.
>
>> +
>> +static int vfio_pci_fault_prod_add_capability(struct vfio_pci_device *vdev,
>> + struct vfio_pci_region *region, struct vfio_info_cap *caps)
>> +{
>> + struct vfio_region_info_cap_fault cap = {
>> + .header.id = VFIO_REGION_INFO_CAP_PRODUCER_FAULT,
>> + .header.version = 1,
>> + .version = NR_FAULT_ABIS,
>> + };
>> + return vfio_info_add_capability(caps, &cap.header, sizeof(cap));
>> +}
>> +
>> +static const struct vfio_pci_regops vfio_pci_fault_cons_regops = {
>> + .rw = vfio_pci_fault_cons_rw,
>> + .release = vfio_pci_fault_release,
>> +};
>> +
>> +static const struct vfio_pci_regops vfio_pci_fault_prod_regops = {
>> + .rw = vfio_pci_fault_prod_rw,
>> + .release = vfio_pci_fault_release,
>> + .add_capability = vfio_pci_fault_prod_add_capability,
>> +};
>> +
>> +static int vfio_pci_init_fault_region(struct vfio_pci_device *vdev)
>> +{
>> + struct vfio_region_fault_prod *header;
>> + int ret;
>> +
>> + mutex_init(&vdev->fault_queue_lock);
>> +
>> + vdev->fault_pages = kzalloc(3 * PAGE_SIZE, GFP_KERNEL);
>> + if (!vdev->fault_pages)
>> + return -ENOMEM;
>> +
>> + ret = vfio_pci_register_dev_region(vdev,
>> + VFIO_REGION_TYPE_NESTED,
>> + VFIO_REGION_SUBTYPE_NESTED_FAULT_PROD,
>> + &vfio_pci_fault_prod_regops, 2 * PAGE_SIZE,
>> + VFIO_REGION_INFO_FLAG_READ, vdev->fault_pages);
>
> If mmap isn't supported yet, why are we pushing the queue out to the
> 2nd page? We're just wasting space. vfio_region_fault_prod.offset
> allows us to relocate it when/if it is mmap capable.
OK. mmap capability is introduced in 27/29 though.
>
>> + if (ret)
>> + goto out;
>> +
>> + ret = vfio_pci_register_dev_region(vdev,
>> + VFIO_REGION_TYPE_NESTED,
>> + VFIO_REGION_SUBTYPE_NESTED_FAULT_CONS,
>> + &vfio_pci_fault_cons_regops,
>> + sizeof(struct vfio_region_fault_cons),
>> + VFIO_REGION_INFO_FLAG_WRITE,
>> + vdev->fault_pages + 2 * PAGE_SIZE);
>
> What's the remaining (PAGE_SIZE - sizeof(struct vfio_region_fault_cons))
> bytes used for?
They are not used.
>
>> + if (ret)
>> + goto out;
>> +
>> + header = (struct vfio_region_fault_prod *)vdev->fault_pages;
>> + header->version = -1;
>> + header->offset = PAGE_SIZE;
>> + return 0;
>> +out:
>> + kfree(vdev->fault_pages);
>> + return ret;
>> +}
>> +
>> +int vfio_pci_check_cons_fault(struct vfio_pci_device *vdev,
>> + struct vfio_region_fault_cons *cons_header)
>> +{
>> + struct vfio_region_fault_prod *prod_header =
>> + (struct vfio_region_fault_prod *)vdev->fault_pages;
>> +
>> + if (cons_header->version > NR_FAULT_ABIS)
>> + return -EINVAL;
>> +
>> + if (!vdev->fault_abi) {
>> + vdev->fault_abi = cons_header->version;
>> + prod_header->entry_size =
>> + fault_abi_versions[vdev->fault_abi - 1].entry_size;
>> + prod_header->nb_entries = PAGE_SIZE / prod_header->entry_size;
>
> Is this sufficient for 4K hosts? Clearly a 64K host has 16x the number
> of entries, so if this is a heuristic the results are vastly different.
This series only deals with unrecoverable errors. We don't expect many
of them so I did not consider the need to have a more complicated
heuristic. Now if we consider the PRI use case we need to reconsider the
size of the fault queue. If this feature is introduced later with a new
region type, then we can handle this later?
Practically the event queue size is set by the guest SMMUv3 driver and
trapped at the SMMUV3 QEMU device level. So we could communicate this
info through IOMMU MR notifiers but that's a rather complicated chain
and I would rather avoid that complexity if not necessary.
>
>> + return 0;
>> + }
>> +
>> + /* Fault ABI is set */
>> + if (cons_header->version != vdev->fault_abi)
>> + return -EINVAL;
>> +
>> + if (cons_header->cons && cons_header->cons >= prod_header->nb_entries)
>
> First test seems unnecessary.
>
>> + return -EINVAL;
>> +
>> + return 0;
>> +}
>> +
>> static int vfio_pci_enable(struct vfio_pci_device *vdev)
>> {
>> struct pci_dev *pdev = vdev->pdev;
>> @@ -359,6 +459,10 @@ static int vfio_pci_enable(struct vfio_pci_device *vdev)
>> }
>> }
>>
>> + ret = vfio_pci_init_fault_region(vdev);
>> + if (ret)
>> + goto disable_exit;
>> +
>> vfio_pci_probe_mmaps(vdev);
>>
>> return 0;
>> @@ -1374,6 +1478,7 @@ static void vfio_pci_remove(struct pci_dev *pdev)
>>
>> vfio_iommu_group_put(pdev->dev.iommu_group, &pdev->dev);
>> kfree(vdev->region);
>> + kfree(vdev->fault_pages);
>> mutex_destroy(&vdev->ioeventfds_lock);
>>
>> if (!disable_idle_d3)
>> diff --git a/drivers/vfio/pci/vfio_pci_private.h b/drivers/vfio/pci/vfio_pci_private.h
>> index 1812cf22fc4f..8e0a55682d3f 100644
>> --- a/drivers/vfio/pci/vfio_pci_private.h
>> +++ b/drivers/vfio/pci/vfio_pci_private.h
>> @@ -122,9 +122,12 @@ struct vfio_pci_device {
>> int ioeventfds_nr;
>> struct eventfd_ctx *err_trigger;
>> struct eventfd_ctx *req_trigger;
>> + struct mutex fault_queue_lock;
>> + int fault_abi;
>> struct list_head dummy_resources_list;
>> struct mutex ioeventfds_lock;
>> struct list_head ioeventfds_list;
>> + u8 *fault_pages;
>> };
>>
>> #define is_intx(vdev) (vdev->irq_type == VFIO_PCI_INTX_IRQ_INDEX)
>> @@ -153,6 +156,18 @@ extern ssize_t vfio_pci_vga_rw(struct vfio_pci_device *vdev, char __user *buf,
>> extern long vfio_pci_ioeventfd(struct vfio_pci_device *vdev, loff_t offset,
>> uint64_t data, int count, int fd);
>>
>> +struct vfio_pci_fault_abi {
>> + u32 entry_size;
>> +};
>> +
>> +extern size_t vfio_pci_fault_cons_rw(struct vfio_pci_device *vdev,
>> + char __user *buf, size_t count,
>> + loff_t *ppos, bool iswrite);
>> +
>> +extern size_t vfio_pci_fault_prod_rw(struct vfio_pci_device *vdev,
>> + char __user *buf, size_t count,
>> + loff_t *ppos, bool iswrite);
>> +
>> extern int vfio_pci_init_perm_bits(void);
>> extern void vfio_pci_uninit_perm_bits(void);
>>
>> @@ -166,6 +181,8 @@ extern int vfio_pci_register_dev_region(struct vfio_pci_device *vdev,
>>
>> extern int vfio_pci_set_power_state(struct vfio_pci_device *vdev,
>> pci_power_t state);
>> +extern int vfio_pci_check_cons_fault(struct vfio_pci_device *vdev,
>> + struct vfio_region_fault_cons *header);
>>
>> #ifdef CONFIG_VFIO_PCI_IGD
>> extern int vfio_pci_igd_init(struct vfio_pci_device *vdev);
>> diff --git a/drivers/vfio/pci/vfio_pci_rdwr.c b/drivers/vfio/pci/vfio_pci_rdwr.c
>> index a6029d0a5524..67cd9363f4e7 100644
>> --- a/drivers/vfio/pci/vfio_pci_rdwr.c
>> +++ b/drivers/vfio/pci/vfio_pci_rdwr.c
>> @@ -277,6 +277,79 @@ ssize_t vfio_pci_vga_rw(struct vfio_pci_device *vdev, char __user *buf,
>> return done;
>> }
>>
>> +/* Read-only region */
>> +size_t vfio_pci_fault_prod_rw(struct vfio_pci_device *vdev, char __user *buf,
>> + size_t count, loff_t *ppos, bool iswrite)
>> +{
>> + unsigned int i = VFIO_PCI_OFFSET_TO_INDEX(*ppos) - VFIO_PCI_NUM_REGIONS;
>> + void *base = vdev->region[i].data;
>> + loff_t pos = *ppos & VFIO_PCI_OFFSET_MASK;
>> + int ret = 0;
>> +
>> + if (iswrite)
>> + return 0;
>> +
>> + if (!vdev->fault_abi)
>> + return -EINVAL;
>> +
>> + if (pos >= vdev->region[i].size)
>> + return -EINVAL;
>> +
>> + count = min(count, (size_t)(vdev->region[i].size - pos));
>> +
>> + mutex_lock(&vdev->fault_queue_lock);
>> +
>> + if (copy_to_user(buf, base + pos, count)) {
>> + ret = -EFAULT;
>> + goto unlock;
>> + }
>> + *ppos += count;
>> + ret = count;
>> +unlock:
>> + mutex_unlock(&vdev->fault_queue_lock);
>> + return ret;
>> +}
>> +
>> +
>> +/* write only */
>> +size_t vfio_pci_fault_cons_rw(struct vfio_pci_device *vdev, char __user *buf,
>> + size_t count, loff_t *ppos, bool iswrite)
>> +{
>> + unsigned int i = VFIO_PCI_OFFSET_TO_INDEX(*ppos) - VFIO_PCI_NUM_REGIONS;
>> + void *base = vdev->region[i].data;
>> + loff_t pos = *ppos & VFIO_PCI_OFFSET_MASK;
>> + struct vfio_region_fault_cons *header;
>> + struct vfio_region_fault_cons orig_header =
>> + *(struct vfio_region_fault_cons *)base;
>> + int ret = 0;
>> +
>> + if (!iswrite)
>> + return 0;
>> +
>> + if (pos >= vdev->region[i].size)
>> + return -EINVAL;
>> +
>> + count = min(count, (size_t)(vdev->region[i].size - pos));
>> +
>> + mutex_lock(&vdev->fault_queue_lock);
>> +
>> + if (copy_from_user(base + pos, buf, count)) {
>> + ret = -EFAULT;
>> + goto unlock;
>> + }
>> + header = (struct vfio_region_fault_cons *)base;
>> + ret = vfio_pci_check_cons_fault(vdev, header);
>> + if (ret) {
>> + *header = orig_header;
>> + goto unlock;
>> + }
>> + *ppos += count;
>> + ret = count;
>> +unlock:
>> + mutex_unlock(&vdev->fault_queue_lock);
>> + return ret;
>> +}
>> +
>> static int vfio_pci_ioeventfd_handler(void *opaque, void *unused)
>> {
>> struct vfio_pci_ioeventfd *ioeventfd = opaque;
>> diff --git a/include/uapi/linux/vfio.h b/include/uapi/linux/vfio.h
>> index 2774a1ab37ae..13e041b84d48 100644
>> --- a/include/uapi/linux/vfio.h
>> +++ b/include/uapi/linux/vfio.h
>> @@ -307,6 +307,10 @@ struct vfio_region_info_cap_type {
>> #define VFIO_REGION_TYPE_GFX (1)
>> #define VFIO_REGION_SUBTYPE_GFX_EDID (1)
>>
>> +#define VFIO_REGION_TYPE_NESTED (2)
>> +#define VFIO_REGION_SUBTYPE_NESTED_FAULT_PROD (1)
>> +#define VFIO_REGION_SUBTYPE_NESTED_FAULT_CONS (2)
>> +
>> /**
>> * struct vfio_region_gfx_edid - EDID region layout.
>> *
>> @@ -701,6 +705,44 @@ struct vfio_device_ioeventfd {
>>
>> #define VFIO_DEVICE_IOEVENTFD _IO(VFIO_TYPE, VFIO_BASE + 16)
>>
>> +
>> +/*
>> + * Capability exposed by the Producer Fault Region
>> + * @version: max fault ABI version supported by the kernel
>> + */
>> +#define VFIO_REGION_INFO_CAP_PRODUCER_FAULT 6
>> +
>> +struct vfio_region_info_cap_fault {
>> + struct vfio_info_cap_header header;
>> + __u32 version;
>> +};
>> +
>> +/*
>> + * Producer Fault Region (Read-Only from user space perspective)
>> + * Contains the fault circular buffer and the producer index
>> + * @version: version of the fault record uapi
>> + * @entry_size: size of each fault record
>> + * @offset: offset of the start of the queue
>> + * @prod: producer index relative to the start of the queue
>> + */
>> +struct vfio_region_fault_prod {
>> + __u32 version;
>> + __u32 nb_entries;
>> + __u32 entry_size;
>> + __u32 offset;
>> + __u32 prod;
>> +};
>> +
>> +/*
>> + * Consumer Fault Region (Write-Only from the user space perspective)
>> + * @version: ABI version requested by the userspace
>> + * @cons: consumer index relative to the start of the queue
>> + */
>> +struct vfio_region_fault_cons {
>> + __u32 version;
>> + __u32 cons;
>> +};
>
> I think there are more common semantics than this prod/cons
> terminology, for example head/tail.
OK. The SMMUv3 event queue uses the prod/cons terminology. Now tail/head
may be more prevalent in kernel.
Thank you for the review
Eric
>
>> +
>> /* -------- API for Type1 VFIO IOMMU -------- */
>>
>> /**
>
On 07/06/2019 09:28, Auger Eric wrote:
>>> +static const struct vfio_pci_fault_abi fault_abi_versions[] = {
>>> + [0] = {
>>> + .entry_size = sizeof(struct iommu_fault),
>>> + },
>>> +};
>>> +
>>> +#define NR_FAULT_ABIS ARRAY_SIZE(fault_abi_versions)
>>
>> This looks like it's leading to some dangerous complicated code to
>> support multiple user selected ABIs. How many ABIs do we plan to
>> support? The region capability also exposes a type, sub-type, and
>> version. How much of this could be exposed that way? ie. if we need
>> to support multiple versions, expose multiple regions.
>
> This is something that was discussed earlier and suggested by
> Jean-Philippe that we may need to support several versions of the ABI
> (typicallu when adding PRI support).
> Exposing multiple region is an interesting idea and I will explore that
> direction.
At the moment the ABI support errors and PRI. We're considering setting
the fault report structure to 64 or 128 bytes (see "[PATCH v2 2/4]
iommu: Introduce device fault data"). 64-byte allows for 2 additional
fields before we have to introduce a new ABI version, while 128 byte
should last us a while.
But that's for adding new fields to existing fault types. It's probably
a good idea to have different region types in VFIO for different fault
types, since userspace isn't necessarily prepared to deal with them. For
example right now userspace doesn't have a method to complete
recoverable faults, so we can't add them to the queue.
Thanks,
Jean
On Fri, 7 Jun 2019 10:28:06 +0200
Auger Eric <[email protected]> wrote:
> Hi Alex,
>
> On 6/4/19 12:31 AM, Alex Williamson wrote:
> > On Sun, 26 May 2019 18:10:00 +0200
> > Eric Auger <[email protected]> wrote:
> >
> >> This patch adds two new regions aiming to handle nested mode
> >> translation faults.
> >>
> >> The first region (two host kernel pages) is read-only from the
> >> user-space perspective. The first page contains an header
> >> that provides information about the circular buffer located in the
> >> second page. The circular buffer is put in a different page in
> >> the prospect to be mmappable.
> >>
> >> The max user API version supported by the kernel is returned
> >> through a dedicated fault region capability.
> >>
> >> The prod header contains
> >> - the user API version in use (potentially inferior to the one
> >> returned in the capability),
> >> - the offset of the queue within the region,
> >> - the producer index relative to the start of the queue
> >> - the max number of fault records,
> >> - the size of each record.
> >>
> >> The second region is write-only from the user perspective. It
> >> contains the version of the requested fault ABI and the consumer
> >> index that is updated by the userspace each time this latter has
> >> consumed fault records.
> >>
> >> The natural order of operation for the userspace is:
> >> - retrieve the highest supported fault ABI version
> >> - set the requested fault ABI version in the consumer region
> >>
> >> Until the ABI version is not set by the userspace, the kernel
> >> cannot return a comprehensive set of information inside the
> >> prod header (entry size and number of entries in the fault queue).
> >
> > It's not clear to me why two regions are required for this. If the
> > first page is not mmap capable, why does it need to be read-only? If
> > it were not read-only couldn't the fields of the second region also fit
> > within this first page? If you wanted to deal with an mmap capable
> > writeable region, it could just be yet a 3rd page in the first region.
> I thought it would be clearer for the userspace to have 2 separate
> regions, one for the producer and one for the consumer. Otherwise I will
> need to specify which fields are read-only or write-only. But this may
> be more self-contained in a single region.
We need to figure out read vs write anyway, but separating them to
separate regions just for that seems unnecessary. How many regions do
we expect to require for a single feature?
> >
> >>
> >> Signed-off-by: Eric Auger <[email protected]>
> >>
> >> ---
> >>
> >> v4 -> v5
> >> - check cons is not null in vfio_pci_check_cons_fault
> >>
> >> v3 -> v4:
> >> - use 2 separate regions, respectively in read and write modes
> >> - add the version capability
> >> ---
> >> drivers/vfio/pci/vfio_pci.c | 105 ++++++++++++++++++++++++++++
> >> drivers/vfio/pci/vfio_pci_private.h | 17 +++++
> >> drivers/vfio/pci/vfio_pci_rdwr.c | 73 +++++++++++++++++++
> >> include/uapi/linux/vfio.h | 42 +++++++++++
> >> 4 files changed, 237 insertions(+)
> >>
> >> diff --git a/drivers/vfio/pci/vfio_pci.c b/drivers/vfio/pci/vfio_pci.c
> >> index cab71da46f4a..f75f61127277 100644
> >> --- a/drivers/vfio/pci/vfio_pci.c
> >> +++ b/drivers/vfio/pci/vfio_pci.c
> >> @@ -261,6 +261,106 @@ int vfio_pci_set_power_state(struct vfio_pci_device *vdev, pci_power_t state)
> >> return ret;
> >> }
> >>
> >> +void vfio_pci_fault_release(struct vfio_pci_device *vdev,
> >> + struct vfio_pci_region *region)
> >> +{
> >> +}
> >> +
> >> +static const struct vfio_pci_fault_abi fault_abi_versions[] = {
> >> + [0] = {
> >> + .entry_size = sizeof(struct iommu_fault),
> >> + },
> >> +};
> >> +
> >> +#define NR_FAULT_ABIS ARRAY_SIZE(fault_abi_versions)
> >
> > This looks like it's leading to some dangerous complicated code to
> > support multiple user selected ABIs. How many ABIs do we plan to
> > support? The region capability also exposes a type, sub-type, and
> > version. How much of this could be exposed that way? ie. if we need
> > to support multiple versions, expose multiple regions.
>
> This is something that was discussed earlier and suggested by
> Jean-Philippe that we may need to support several versions of the ABI
> (typicallu when adding PRI support).
> Exposing multiple region is an interesting idea and I will explore that
> direction.
> >
> >> +
> >> +static int vfio_pci_fault_prod_add_capability(struct vfio_pci_device *vdev,
> >> + struct vfio_pci_region *region, struct vfio_info_cap *caps)
> >> +{
> >> + struct vfio_region_info_cap_fault cap = {
> >> + .header.id = VFIO_REGION_INFO_CAP_PRODUCER_FAULT,
> >> + .header.version = 1,
> >> + .version = NR_FAULT_ABIS,
> >> + };
> >> + return vfio_info_add_capability(caps, &cap.header, sizeof(cap));
> >> +}
> >> +
> >> +static const struct vfio_pci_regops vfio_pci_fault_cons_regops = {
> >> + .rw = vfio_pci_fault_cons_rw,
> >> + .release = vfio_pci_fault_release,
> >> +};
> >> +
> >> +static const struct vfio_pci_regops vfio_pci_fault_prod_regops = {
> >> + .rw = vfio_pci_fault_prod_rw,
> >> + .release = vfio_pci_fault_release,
> >> + .add_capability = vfio_pci_fault_prod_add_capability,
> >> +};
> >> +
> >> +static int vfio_pci_init_fault_region(struct vfio_pci_device *vdev)
> >> +{
> >> + struct vfio_region_fault_prod *header;
> >> + int ret;
> >> +
> >> + mutex_init(&vdev->fault_queue_lock);
> >> +
> >> + vdev->fault_pages = kzalloc(3 * PAGE_SIZE, GFP_KERNEL);
> >> + if (!vdev->fault_pages)
> >> + return -ENOMEM;
> >> +
> >> + ret = vfio_pci_register_dev_region(vdev,
> >> + VFIO_REGION_TYPE_NESTED,
> >> + VFIO_REGION_SUBTYPE_NESTED_FAULT_PROD,
> >> + &vfio_pci_fault_prod_regops, 2 * PAGE_SIZE,
> >> + VFIO_REGION_INFO_FLAG_READ, vdev->fault_pages);
> >
> > If mmap isn't supported yet, why are we pushing the queue out to the
> > 2nd page? We're just wasting space. vfio_region_fault_prod.offset
> > allows us to relocate it when/if it is mmap capable.
> OK. mmap capability is introduced in 27/29 though.
> >
> >> + if (ret)
> >> + goto out;
> >> +
> >> + ret = vfio_pci_register_dev_region(vdev,
> >> + VFIO_REGION_TYPE_NESTED,
> >> + VFIO_REGION_SUBTYPE_NESTED_FAULT_CONS,
> >> + &vfio_pci_fault_cons_regops,
> >> + sizeof(struct vfio_region_fault_cons),
> >> + VFIO_REGION_INFO_FLAG_WRITE,
> >> + vdev->fault_pages + 2 * PAGE_SIZE);
> >
> > What's the remaining (PAGE_SIZE - sizeof(struct vfio_region_fault_cons))
> > bytes used for?
> They are not used.
So we probably don't want to allocate a full page for it. Seems little
reason to separate by pages when mmap is not supported anyway.
> >> + if (ret)
> >> + goto out;
> >> +
> >> + header = (struct vfio_region_fault_prod *)vdev->fault_pages;
> >> + header->version = -1;
> >> + header->offset = PAGE_SIZE;
> >> + return 0;
> >> +out:
> >> + kfree(vdev->fault_pages);
> >> + return ret;
> >> +}
> >> +
> >> +int vfio_pci_check_cons_fault(struct vfio_pci_device *vdev,
> >> + struct vfio_region_fault_cons *cons_header)
> >> +{
> >> + struct vfio_region_fault_prod *prod_header =
> >> + (struct vfio_region_fault_prod *)vdev->fault_pages;
> >> +
> >> + if (cons_header->version > NR_FAULT_ABIS)
> >> + return -EINVAL;
> >> +
> >> + if (!vdev->fault_abi) {
> >> + vdev->fault_abi = cons_header->version;
> >> + prod_header->entry_size =
> >> + fault_abi_versions[vdev->fault_abi - 1].entry_size;
> >> + prod_header->nb_entries = PAGE_SIZE / prod_header->entry_size;
> >
> > Is this sufficient for 4K hosts? Clearly a 64K host has 16x the number
> > of entries, so if this is a heuristic the results are vastly different.
> This series only deals with unrecoverable errors. We don't expect many
> of them so I did not consider the need to have a more complicated
> heuristic. Now if we consider the PRI use case we need to reconsider the
> size of the fault queue. If this feature is introduced later with a new
> region type, then we can handle this later?
>
> Practically the event queue size is set by the guest SMMUv3 driver and
> trapped at the SMMUV3 QEMU device level. So we could communicate this
> info through IOMMU MR notifiers but that's a rather complicated chain
> and I would rather avoid that complexity if not necessary.
I'd hope that this interface accounts for both errors and page mapping
requests and that things like queue size are specified in the user
interface to account for these sorts of differences. Thanks,
Alex