Received: by 2002:a25:ab43:0:0:0:0:0 with SMTP id u61csp7294680ybi; Wed, 5 Jun 2019 14:57:51 -0700 (PDT) X-Google-Smtp-Source: APXvYqz1zQiP7HJtInI0nDEDin73d1i5yb7i5/YYxA9hwlLzx++l4BTU8l+DtCHecotRl0/1SoyL X-Received: by 2002:a17:90a:1b48:: with SMTP id q66mr260923pjq.83.1559771871871; Wed, 05 Jun 2019 14:57:51 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1559771871; cv=none; d=google.com; s=arc-20160816; b=uL38cZdV94nsmdTFTFdj8tJ3oAqCSuaJebFy1YjN3jNYwyJ1aYevPP4+TmBT9zrneJ jKkWKucg0caiXxOYHNBxmNiFhN2I04uLfgCZ8VINZn6SCzNuu3oz28iPWKONMsyHfwlA mpA3nN11SRAJYmBuELQ9rqb1Bdyk3P27QdgSXiYcgYlAJ2/B8Txd5UMuFR7QztBvMG9h GuATF6bzbgFeXVgxqmyjgaIquA00ZG0aCxen7mTI31DvarPqBl7asGe6rF0dJRwTHfrM VgQCF94oZzamKDdLZOyYxhySxE15XhPWCncH9+pL3BjTXpE5qYUyj78HwvOA3oHM6+AN 2Gbw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:content-transfer-encoding:mime-version :organization:references:in-reply-to:message-id:subject:cc:to:from :date; bh=XjKYHnGgicHeMXtH2ww63/MLSCVOcVrCsrJgIY5LZ64=; b=J7Y8G8OpNpPkcFHy1iW2gCrSd92KI+riNwrMN6IhmUo+kpxQ1I3g4EG97DAkc/P65F KIQ6szJF4cVs8AELjyMG/fZfd+r2arlAhRpaRtq8LEvEvoONlEocSJ+VvncLk4XGfW7E sq6WZwUz/HgeTbLwmAj8mHaF+RE8xP5rFxviGzdSoKDSguXHn9ZjbHPm2HOT6poY+c8r nCUitPUEovqEjTXEVuwzl45rlouh1e7Rt5CrLbF61KnrZ8VsW0Wo3nyZwLcVPp89/lcS roQeKCjOLit8XlRdjlg0GS5IYXkB116QcnbGStNMunerjkslsQnpcXZHufNWKDvwyjAa 3IvQ== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=intel.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id 3si29312276plv.102.2019.06.05.14.57.35; Wed, 05 Jun 2019 14:57:51 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=intel.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1726605AbfFEVzh (ORCPT + 99 others); Wed, 5 Jun 2019 17:55:37 -0400 Received: from mga06.intel.com ([134.134.136.31]:51548 "EHLO mga06.intel.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726537AbfFEVzg (ORCPT ); Wed, 5 Jun 2019 17:55:36 -0400 X-Amp-Result: SKIPPED(no attachment in message) X-Amp-File-Uploaded: False Received: from orsmga001.jf.intel.com ([10.7.209.18]) by orsmga104.jf.intel.com with ESMTP/TLS/DHE-RSA-AES256-GCM-SHA384; 05 Jun 2019 14:55:35 -0700 X-ExtLoop1: 1 Received: from jacob-builder.jf.intel.com (HELO jacob-builder) ([10.7.199.155]) by orsmga001.jf.intel.com with ESMTP; 05 Jun 2019 14:55:36 -0700 Date: Wed, 5 Jun 2019 14:58:41 -0700 From: Jacob Pan To: Jean-Philippe Brucker Cc: "Tian, Kevin" , "alex.williamson@redhat.com" , "robin.murphy@arm.com" , "Raj, Ashok" , "iommu@lists.linux-foundation.org" , "linux-kernel@vger.kernel.org" , "eric.auger@redhat.com" , jacob.jun.pan@linux.intel.com Subject: Re: [PATCH v2 2/4] iommu: Introduce device fault data Message-ID: <20190605145841.55cf9667@jacob-builder> In-Reply-To: <50dc3cc5-6019-ad42-6aba-d84fab4020f9@arm.com> References: <20190603145749.46347-1-jean-philippe.brucker@arm.com> <20190603145749.46347-3-jean-philippe.brucker@arm.com> <20190603150842.11070cfd@jacob-builder> <50dc3cc5-6019-ad42-6aba-d84fab4020f9@arm.com> Organization: OTC X-Mailer: Claws Mail 3.13.2 (GTK+ 2.24.30; x86_64-pc-linux-gnu) MIME-Version: 1.0 Content-Type: text/plain; charset=US-ASCII Content-Transfer-Encoding: 7bit Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Wed, 5 Jun 2019 12:24:09 +0100 Jean-Philippe Brucker wrote: > On 05/06/2019 09:51, Tian, Kevin wrote: > >> From: Jacob Pan > >> Sent: Tuesday, June 4, 2019 6:09 AM > >> > >> On Mon, 3 Jun 2019 15:57:47 +0100 > >> Jean-Philippe Brucker wrote: > >> > >>> +/** > >>> + * struct iommu_fault_page_request - Page Request data > >>> + * @flags: encodes whether the corresponding fields are valid and > >>> whether this > >>> + * is the last page in group (IOMMU_FAULT_PAGE_REQUEST_* > >>> values) > >>> + * @pasid: Process Address Space ID > >>> + * @grpid: Page Request Group Index > >>> + * @perm: requested page permissions (IOMMU_FAULT_PERM_* values) > >>> + * @addr: page address > >>> + * @private_data: device-specific private information > >>> + */ > >>> +struct iommu_fault_page_request { > >>> +#define IOMMU_FAULT_PAGE_REQUEST_PASID_VALID (1 << 0) > >>> +#define IOMMU_FAULT_PAGE_REQUEST_LAST_PAGE (1 << 1) > >>> +#define IOMMU_FAULT_PAGE_REQUEST_PRIV_DATA (1 << 2) > >>> + __u32 flags; > >>> + __u32 pasid; > >>> + __u32 grpid; > >>> + __u32 perm; > >>> + __u64 addr; > >>> + __u64 private_data[2]; > >>> +}; > >>> + > >> > >> Just a thought, for non-identity G-H PASID management. We could > >> pass on guest PASID in PRQ to save a lookup in QEMU. In this case, > >> QEMU allocate a GPASID for vIOMMU then a host PASID for pIOMMU. > >> QEMU has a G->H lookup. When PRQ comes in to the pIOMMU with > >> HPASID, IOMMU driver > >> can retrieve GPASID from the bind data then report to the guest via > >> VFIO. In this case QEMU does not need to do a H->G PASID lookup. > >> > >> Should we add a gpasid field here? or we can add a flag and field > >> later, up to you. > >> > > > > Can private_data serve this purpose? > > Isn't private_data already used for VT-d's Private Data field? > yes, as part of the PRQ. please see my explanation in the previous email. > > It's better not introducing > > gpasid awareness within host IOMMU driver. It is just a user-level > > data associated with a PASID when binding happens. Kernel doesn't > > care the actual meaning, simply record it and then return back to > > user space later upon device fault. Qemu interprets the meaning as > > gpasid in its own context. otherwise usages may use it for other > > purpose. > > Regarding a gpasid field I don't mind either way, but extending the > iommu_fault structure later won't be completely straightforward so we > could add some padding now. > > Userspace negotiate the iommu_fault struct format with VFIO, before > allocating a circular buffer of N fault structures > (). > So adding new fields requires introducing a new ABI version and a > struct iommu_fault_v2. That may be OK for disruptive changes, but > just adding a new field indicated by a flag shouldn't have to be that > complicated. > > How about setting the iommu_fault structure to 128 bytes? > > struct iommu_fault { > __u32 type; > __u32 padding; > union { > struct iommu_fault_unrecoverable event; > struct iommu_fault_page_request prm; > __u8 padding2[120]; > }; > }; > > Given that @prm is currently 40 bytes and @event 32 bytes, the padding > allows either of them to grow 10 new 64-bit fields (or 20 new 32-bit > fields, which is still representable with new flags) before we have to > upgrade the ABI version. > > A 4kB and a 64kB queue can hold respectively: > > * 85 and 1365 records when iommu_fault is 48 bytes (current format). > * 64 and 1024 records when iommu_fault is 64 bytes (but allows to grow > only 2 new 64-bit fields). > * 32 and 512 records when iommu_fault is 128 bytes. > > In comparison, > * the SMMU even queue can hold 128 and 2048 events respectively at > those sizes (and is allowed to grow up to 524k entries) > * the SMMU PRI queue can hold 256 and 4096 PR. > > But the SMMU queues have to be physically contiguous, whereas our > fault queues are in userspace memory which is less expensive. So > 128-byte records might be reasonable. What do you think? > I think though 128-byte is large enough for any future extension but 64B might be good enough and it is a cache line. PCI page request msg is only 16B :) VT-d currently uses one 4K page for PRQ, holds 128 records of PRQ descriptors. This can grow to 16K entries per spec. That is per IOMMU. The user fault queue here is per device. So we do have to be frugal about it since it will support mdev at per PASID level at some point? I have to look into Eric's patchset on how he handles queue full in the producer. If we go with 128B size in iommu_fault and 4KB size queue (32 entries as in your table), VT-d PRQ size of 128 entries can potentially cause queue full. We have to handle this VFIO queue full differently than the IOMMU queue full in that we only need to discard PRQ for one device. (Whereas IOMMU queue full has to clear out all). Anyway, I think 64B should be enough but 128B is fine too. We have to deal with queue full anyway. But queue full is expensive so we should try to avoid. > > The iommu_fault_response (patch 4/4) is a bit easier to extend because > it's userspace->kernel and userspace can just declare the size it's > using. I did add a version field in case we run out of flags or want > to change the whole thing, but I think I was being overly cautious > and it might just be a waste of space. > > Thanks, > Jean [Jacob Pan]