Received: by 2002:a25:8b91:0:0:0:0:0 with SMTP id j17csp199453ybl; Tue, 28 Jan 2020 21:57:54 -0800 (PST) X-Google-Smtp-Source: APXvYqxNCdAPDv71bJC0GtW8pNCr+4YHLKGhRaC6BxgX+fSEgd3DBWCkneeInfTnIRx1efW+ycCn X-Received: by 2002:a54:4791:: with SMTP id o17mr5178886oic.70.1580277474392; Tue, 28 Jan 2020 21:57:54 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1580277474; cv=none; d=google.com; s=arc-20160816; b=sj6Qr6HBcutXR0/qndB3/Cr8tK5LSstdPdzyBt9GUk8aRI9gA3Nj2eMfFWiz1XA7jO +q1mfdXlh9Hu3yZaHCsJCKHM7gvapS6fziyO1NiFbuIh8aBbHk5ZERHXU+zsyvDBvr2G 3M1lciddamTjfDYlZUklxm5fqjxwzmBmciaiSjzmM+wwHlMxTDDgow8FOHbwBAjzmtPQ 5Agwobc79z02TWPKUQpppCaRrLgJI2LSUvX66Q2I6nM2pGco7uA3ylUujB8v5uZT7a4c la1CA8lpx6V2TxkuBQe35GaVOstq2v4pJRAJ32aN8foCgEmU8cV7lgg2hBl7UJeqdWKU 1Vyw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:references:in-reply-to:message-id:date :subject:cc:to:from; bh=skIy4npYDjLyPB2WWtKe3c0s2LKrvhScFpfO/3EfCQw=; b=pQIz2uMOdhjYRcs8HSWnynAWchf5yuE4X5chGEEiaz4qzlDmgCd/dtPlWIXWwirR1Z yCJtz/OvrwxESg6xgK0cxxzzuvubCP7Md7qkeQz6wR/Y+qinZ8njU6qCWZvW+8BcLhga kZzK+tCfwuGicoF6WoGkWl4n2kUzwieo3vWLk331G8M4WUIkJRhF0T9cIuvyYMlhM1DK KLOrPDbeUm4QoyP3mMU2zjIyHgrd/3zUj+TlGn4YQxBrBSWZ5CXkhn3H5p6wGYrFkG/6 yX913k1pi7/OSxmIzi+xBsoqdgToVFJQNuQV+fBC2oW0IVGAk/6w7fahJoai9uAueYrU ky7A== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=intel.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id g72si647292oib.157.2020.01.28.21.57.42; Tue, 28 Jan 2020 21:57:54 -0800 (PST) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=intel.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1726330AbgA2F4t (ORCPT + 99 others); Wed, 29 Jan 2020 00:56:49 -0500 Received: from mga05.intel.com ([192.55.52.43]:23420 "EHLO mga05.intel.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726203AbgA2F4o (ORCPT ); Wed, 29 Jan 2020 00:56:44 -0500 X-Amp-Result: SKIPPED(no attachment in message) X-Amp-File-Uploaded: False Received: from orsmga008.jf.intel.com ([10.7.209.65]) by fmsmga105.fm.intel.com with ESMTP/TLS/DHE-RSA-AES256-GCM-SHA384; 28 Jan 2020 21:56:42 -0800 X-ExtLoop1: 1 X-IronPort-AV: E=Sophos;i="5.70,376,1574150400"; d="scan'208";a="222346531" Received: from jacob-builder.jf.intel.com ([10.7.199.155]) by orsmga008.jf.intel.com with ESMTP; 28 Jan 2020 21:56:41 -0800 From: Jacob Pan To: iommu@lists.linux-foundation.org, LKML , "Lu Baolu" , Joerg Roedel , David Woodhouse Cc: "Yi Liu" , "Tian, Kevin" , Raj Ashok , Alex Williamson , "Christoph Hellwig" , Jean-Philippe Brucker , Jonathan Cameron , Eric Auger , Jacob Pan , Liu@vger.kernel.org, Yi L Subject: [PATCH V9 06/10] iommu/vt-d: Add svm/sva invalidate function Date: Tue, 28 Jan 2020 22:01:49 -0800 Message-Id: <1580277713-66934-7-git-send-email-jacob.jun.pan@linux.intel.com> X-Mailer: git-send-email 2.7.4 In-Reply-To: <1580277713-66934-1-git-send-email-jacob.jun.pan@linux.intel.com> References: <1580277713-66934-1-git-send-email-jacob.jun.pan@linux.intel.com> Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org When Shared Virtual Address (SVA) is enabled for a guest OS via vIOMMU, we need to provide invalidation support at IOMMU API and driver level. This patch adds Intel VT-d specific function to implement iommu passdown invalidate API for shared virtual address. The use case is for supporting caching structure invalidation of assigned SVM capable devices. Emulated IOMMU exposes queue invalidation capability and passes down all descriptors from the guest to the physical IOMMU. The assumption is that guest to host device ID mapping should be resolved prior to calling IOMMU driver. Based on the device handle, host IOMMU driver can replace certain fields before submit to the invalidation queue. Signed-off-by: Jacob Pan Signed-off-by: Ashok Raj Signed-off-by: Liu, Yi L --- drivers/iommu/intel-iommu.c | 173 ++++++++++++++++++++++++++++++++++++++++++++ 1 file changed, 173 insertions(+) diff --git a/drivers/iommu/intel-iommu.c b/drivers/iommu/intel-iommu.c index 8a4136e805ac..b8aa6479b87f 100644 --- a/drivers/iommu/intel-iommu.c +++ b/drivers/iommu/intel-iommu.c @@ -5605,6 +5605,178 @@ static void intel_iommu_aux_detach_device(struct iommu_domain *domain, aux_domain_remove_dev(to_dmar_domain(domain), dev); } +/* + * 2D array for converting and sanitizing IOMMU generic TLB granularity to + * VT-d granularity. Invalidation is typically included in the unmap operation + * as a result of DMA or VFIO unmap. However, for assigned device where guest + * could own the first level page tables without being shadowed by QEMU. In + * this case there is no pass down unmap to the host IOMMU as a result of unmap + * in the guest. Only invalidations are trapped and passed down. + * In all cases, only first level TLB invalidation (request with PASID) can be + * passed down, therefore we do not include IOTLB granularity for request + * without PASID (second level). + * + * For an example, to find the VT-d granularity encoding for IOTLB + * type and page selective granularity within PASID: + * X: indexed by iommu cache type + * Y: indexed by enum iommu_inv_granularity + * [IOMMU_CACHE_INV_TYPE_IOTLB][IOMMU_INV_GRANU_ADDR] + * + * Granu_map array indicates validity of the table. 1: valid, 0: invalid + * + */ +const static int inv_type_granu_map[IOMMU_CACHE_INV_TYPE_NR][IOMMU_INV_GRANU_NR] = { + /* PASID based IOTLB, support PASID selective and page selective */ + {0, 1, 1}, + /* PASID based dev TLBs, only support all PASIDs or single PASID */ + {1, 1, 0}, + /* PASID cache */ + {1, 1, 0} +}; + +const static u64 inv_type_granu_table[IOMMU_CACHE_INV_TYPE_NR][IOMMU_INV_GRANU_NR] = { + /* PASID based IOTLB */ + {0, QI_GRAN_NONG_PASID, QI_GRAN_PSI_PASID}, + /* PASID based dev TLBs */ + {QI_DEV_IOTLB_GRAN_ALL, QI_DEV_IOTLB_GRAN_PASID_SEL, 0}, + /* PASID cache */ + {QI_PC_ALL_PASIDS, QI_PC_PASID_SEL, 0}, +}; + +static inline int to_vtd_granularity(int type, int granu, u64 *vtd_granu) +{ + if (type >= IOMMU_CACHE_INV_TYPE_NR || granu >= IOMMU_INV_GRANU_NR || + !inv_type_granu_map[type][granu]) + return -EINVAL; + + *vtd_granu = inv_type_granu_table[type][granu]; + + return 0; +} + +static inline u64 to_vtd_size(u64 granu_size, u64 nr_granules) +{ + u64 nr_pages = (granu_size * nr_granules) >> VTD_PAGE_SHIFT; + + /* VT-d size is encoded as 2^size of 4K pages, 0 for 4k, 9 for 2MB, etc. + * IOMMU cache invalidate API passes granu_size in bytes, and number of + * granu size in contiguous memory. + */ + return order_base_2(nr_pages); +} + +#ifdef CONFIG_INTEL_IOMMU_SVM +static int intel_iommu_sva_invalidate(struct iommu_domain *domain, + struct device *dev, struct iommu_cache_invalidate_info *inv_info) +{ + struct dmar_domain *dmar_domain = to_dmar_domain(domain); + struct device_domain_info *info; + struct intel_iommu *iommu; + unsigned long flags; + int cache_type; + u8 bus, devfn; + u16 did, sid; + int ret = 0; + u64 size; + + if (!inv_info || !dmar_domain || + inv_info->version != IOMMU_CACHE_INVALIDATE_INFO_VERSION_1) + return -EINVAL; + + if (!dev || !dev_is_pci(dev)) + return -ENODEV; + + iommu = device_to_iommu(dev, &bus, &devfn); + if (!iommu) + return -ENODEV; + + spin_lock_irqsave(&device_domain_lock, flags); + spin_lock(&iommu->lock); + info = iommu_support_dev_iotlb(dmar_domain, iommu, bus, devfn); + if (!info) { + ret = -EINVAL; + goto out_unlock; + } + did = dmar_domain->iommu_did[iommu->seq_id]; + sid = PCI_DEVID(bus, devfn); + size = to_vtd_size(inv_info->addr_info.granule_size, inv_info->addr_info.nb_granules); + + for_each_set_bit(cache_type, (unsigned long *)&inv_info->cache, IOMMU_CACHE_INV_TYPE_NR) { + u64 granu = 0; + u64 pasid = 0; + + ret = to_vtd_granularity(cache_type, inv_info->granularity, &granu); + if (ret) { + pr_err("Invalid cache type and granu combination %d/%d\n", cache_type, + inv_info->granularity); + break; + } + + /* PASID is stored in different locations based on granularity */ + if (inv_info->granularity == IOMMU_INV_GRANU_PASID) + pasid = inv_info->pasid_info.pasid; + else if (inv_info->granularity == IOMMU_INV_GRANU_ADDR) + pasid = inv_info->addr_info.pasid; + else { + pr_err("Cannot find PASID for given cache type and granularity\n"); + break; + } + + switch (BIT(cache_type)) { + case IOMMU_CACHE_INV_TYPE_IOTLB: + if (size && (inv_info->addr_info.addr & ((BIT(VTD_PAGE_SHIFT + size)) - 1))) { + pr_err("Address out of range, 0x%llx, size order %llu\n", + inv_info->addr_info.addr, size); + ret = -ERANGE; + goto out_unlock; + } + + qi_flush_piotlb(iommu, did, + pasid, + mm_to_dma_pfn(inv_info->addr_info.addr), + (granu == QI_GRAN_NONG_PASID) ? -1 : 1 << size, + inv_info->addr_info.flags & IOMMU_INV_ADDR_FLAGS_LEAF); + + /* + * Always flush device IOTLB if ATS is enabled since guest + * vIOMMU exposes CM = 1, no device IOTLB flush will be passed + * down. + */ + if (info->ats_enabled) { + qi_flush_dev_iotlb_pasid(iommu, sid, info->pfsid, + pasid, info->ats_qdep, + inv_info->addr_info.addr, size, + granu); + } + break; + case IOMMU_CACHE_INV_TYPE_DEV_IOTLB: + if (info->ats_enabled) { + qi_flush_dev_iotlb_pasid(iommu, sid, info->pfsid, + inv_info->addr_info.pasid, info->ats_qdep, + inv_info->addr_info.addr, size, + granu); + } else + pr_warn("Passdown device IOTLB flush w/o ATS!\n"); + + break; + case IOMMU_CACHE_INV_TYPE_PASID: + qi_flush_pasid_cache(iommu, did, granu, inv_info->pasid_info.pasid); + + break; + default: + dev_err(dev, "Unsupported IOMMU invalidation type %d\n", + cache_type); + ret = -EINVAL; + } + } +out_unlock: + spin_unlock(&iommu->lock); + spin_unlock_irqrestore(&device_domain_lock, flags); + + return ret; +} +#endif + static int intel_iommu_map(struct iommu_domain *domain, unsigned long iova, phys_addr_t hpa, size_t size, int iommu_prot, gfp_t gfp) @@ -6183,6 +6355,7 @@ const struct iommu_ops intel_iommu_ops = { .is_attach_deferred = intel_iommu_is_attach_deferred, .pgsize_bitmap = INTEL_IOMMU_PGSIZES, #ifdef CONFIG_INTEL_IOMMU_SVM + .cache_invalidate = intel_iommu_sva_invalidate, .sva_bind_gpasid = intel_svm_bind_gpasid, .sva_unbind_gpasid = intel_svm_unbind_gpasid, #endif -- 2.7.4