Received: by 2002:a05:6358:7058:b0:131:369:b2a3 with SMTP id 24csp8626283rwp; Wed, 19 Jul 2023 12:43:34 -0700 (PDT) X-Google-Smtp-Source: APBJJlEjo2rGug98y/V34xElCtn2aJpDg+nlVXPro01TSX8wErPQzhkOisqBlIaq7rpkq6ylsBtM X-Received: by 2002:a17:907:2da7:b0:993:eef2:5d5d with SMTP id gt39-20020a1709072da700b00993eef25d5dmr3997389ejc.27.1689795813901; Wed, 19 Jul 2023 12:43:33 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1689795813; cv=none; d=google.com; s=arc-20160816; b=qSUH/iJz4q1OQoBzaZa1bMteiDb5OYk+xrgqaW5zPp9bDuF4ts3SNdL0B2My85rW+Z xrXqUN64UmskQcGeQJhNwP8M5gTidzTwZBKsCoyWb3cAoweFk2zjWH3gjOTdjL5uuXVY CiVggXBrwhl7iWQ946dDdrMcIvrvrc0ziNU0wRy60HScYVgWG2PfmJOQNW5DgAGRwMYQ FYJS2vl2FIPmQzc8238pznRaPTHF+6WHmHga89PptX9QwaqhEb2GeTTmC4r2pMCxBHhv KKW3SZKA7XeLDkMvpsRrIt+tMmaWjQfptunG1jz0cun62qLaQOZJGBlqDU0y5H2RA7wD pwUw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:mime-version :references:in-reply-to:message-id:date:subject:cc:to:from :dkim-signature; bh=ESdqCz79AqYqq1R9n18M3NCf1nQwZKZlVuGXa3nv+14=; fh=ZDvbyS/atXdqzwPdCJNAB1SCm9CorpyeK2wk4+rW6B8=; b=QltRZCJ3t+ijk7+GUH3XM5TljZhQAIHYCpSlNjnxWQ3U4tPpHb1pdFIfei3nkJV3y5 SHjh26Dph0pSy9QapjT4CixLJq0vVqs+j8jLJsraFDiHBdtjNJs8Ibw/ZrhI1NnatfSm gW2f2ZNglj448inC+fa6UvhoOXufwrdBE88MgGxz3c4qf7Fm8HrVszVDajzFbrjddy1B c/pbKokqSKMDmxDp6z++eiZ21nANpZvOaGskpX30rbOAdD9yBVRlj3NFPFpQHeq+tuO9 VOClmg76lBrWa7bNJdu2iQjRTVPPMtSsVxVOlq9TmplFmECruC7kyfpDW2zmsfkM+WXP zo1A== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@rivosinc-com.20221208.gappssmtp.com header.s=20221208 header.b="0/zDuCQO"; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id u3-20020a170906780300b009935121ecd5si3120142ejm.646.2023.07.19.12.43.06; Wed, 19 Jul 2023 12:43:33 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; dkim=pass header.i=@rivosinc-com.20221208.gappssmtp.com header.s=20221208 header.b="0/zDuCQO"; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S229810AbjGSTf3 (ORCPT + 99 others); Wed, 19 Jul 2023 15:35:29 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:58614 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S229990AbjGSTfF (ORCPT ); Wed, 19 Jul 2023 15:35:05 -0400 Received: from mail-il1-x134.google.com (mail-il1-x134.google.com [IPv6:2607:f8b0:4864:20::134]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 6D89E1FF1 for ; Wed, 19 Jul 2023 12:34:35 -0700 (PDT) Received: by mail-il1-x134.google.com with SMTP id e9e14a558f8ab-345ff33d286so472865ab.3 for ; Wed, 19 Jul 2023 12:34:35 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=rivosinc-com.20221208.gappssmtp.com; s=20221208; t=1689795274; x=1690400074; h=content-transfer-encoding:mime-version:references:in-reply-to :message-id:date:subject:cc:to:from:from:to:cc:subject:date :message-id:reply-to; bh=ESdqCz79AqYqq1R9n18M3NCf1nQwZKZlVuGXa3nv+14=; b=0/zDuCQOfMpFDxXbQVlQLcyvvXi/F8tdUpaBbm0sOShcpabIGwxFQxtzVEr6CW5qZT qLbEmOcaWG4nRaIYFf85eKLGTmVEzllTL2dt/Cl4yPvlaPQFwO0TejYCG8CLe78e3kbN pXyzNrJP3bpLJwoU9eXFJDywKPwsYzPEdR37h2TijJfLOsLCGuux78HgzhgpSJpXRzTO Ex3WxHyQTOGbS9HPiK4OAs3jBRNV9KbyL9ndEWKjcGnYuFPj3ueMAs5r6pW2Dj0/P2/P sdIYKZUimNe0mipfF1+98uu6OSj0POrVFQnzB4Pr/mmEPgi/qd92+xDuTPyD7xNGJveX 5f5A== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20221208; t=1689795274; x=1690400074; h=content-transfer-encoding:mime-version:references:in-reply-to :message-id:date:subject:cc:to:from:x-gm-message-state:from:to:cc :subject:date:message-id:reply-to; bh=ESdqCz79AqYqq1R9n18M3NCf1nQwZKZlVuGXa3nv+14=; b=dpE7x2B/4P2WMjv6rsk5UPDlgXO1K3yljpUQpcz7FrDLcfmbESW85FCzMBcsBIgHjB 8ZZPhuTLJlNrCnMz2l02DVMgYaSUJN3K6JDvGCvAa+2aiyLbBOQCl73fREvE9gZ8Wuxe tk13mdFBBYGMzz5wvtutfR9fLMIfDI5/VIWYMldzzk0u9MlkaYs4+d+lKNMoqgYTZydS OZXS9Wr7/cg9SCCsaj12BlSfvxAzT55xDziH9eagJbVwC9MXvrpNKp5MMm17OzPJQvxL 1Y9p0ng9hj7HvxzBKlblopt2xgIjjebR9MaTThQLQ0uPdPWPkSpdNVZO8mKSPoT+42T4 52Kw== X-Gm-Message-State: ABy/qLa7RWkNeAx/j+ggKDkC9TngPShM63c7wvxJWTRcsPejLwHeR3Gu MyL0zVqA+MYqCUsf3cIJcwEtASwoZFx7sWb7emeZPA== X-Received: by 2002:a92:cd8f:0:b0:345:dc7a:7c66 with SMTP id r15-20020a92cd8f000000b00345dc7a7c66mr7383304ilb.27.1689795273798; Wed, 19 Jul 2023 12:34:33 -0700 (PDT) Received: from tjeznach.ba.rivosinc.com ([66.220.2.162]) by smtp.gmail.com with ESMTPSA id 23-20020a17090a031700b00264040322desm1591053pje.40.2023.07.19.12.34.32 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Wed, 19 Jul 2023 12:34:33 -0700 (PDT) From: Tomasz Jeznach To: Joerg Roedel , Will Deacon , Robin Murphy , Paul Walmsley Cc: Palmer Dabbelt , Albert Ou , Anup Patel , Sunil V L , Nick Kossifidis , Sebastien Boeuf , iommu@lists.linux.dev, linux-riscv@lists.infradead.org, linux-kernel@vger.kernel.org, linux@rivosinc.com, Tomasz Jeznach Subject: [PATCH 08/11] RISC-V: drivers/iommu/riscv: Add page table support Date: Wed, 19 Jul 2023 12:33:52 -0700 Message-Id: <2d5242e79a98dc75cd8fa0fefdb4e3ad37a920ba.1689792825.git.tjeznach@rivosinc.com> X-Mailer: git-send-email 2.34.1 In-Reply-To: References: MIME-Version: 1.0 Content-Type: text/plain; charset=UTF8 Content-Transfer-Encoding: 8bit X-Spam-Status: No, score=-1.9 required=5.0 tests=BAYES_00,DKIM_SIGNED, DKIM_VALID,RCVD_IN_DNSWL_BLOCKED,SPF_HELO_NONE,SPF_PASS, T_SCC_BODY_TEXT_LINE autolearn=ham autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Introduces I/O page level translation services, with 4K, 2M, 1G page size support and enables page level iommu_map/unmap domain interfaces. Co-developed-by: Sebastien Boeuf Signed-off-by: Sebastien Boeuf Signed-off-by: Tomasz Jeznach --- drivers/iommu/io-pgtable.c | 3 + drivers/iommu/riscv/Makefile | 2 +- drivers/iommu/riscv/io_pgtable.c | 266 +++++++++++++++++++++++++++++++ drivers/iommu/riscv/iommu.c | 40 +++-- drivers/iommu/riscv/iommu.h | 1 + include/linux/io-pgtable.h | 2 + 6 files changed, 297 insertions(+), 17 deletions(-) create mode 100644 drivers/iommu/riscv/io_pgtable.c diff --git a/drivers/iommu/io-pgtable.c b/drivers/iommu/io-pgtable.c index b843fcd365d2..c4807175934f 100644 --- a/drivers/iommu/io-pgtable.c +++ b/drivers/iommu/io-pgtable.c @@ -32,6 +32,9 @@ io_pgtable_init_table[IO_PGTABLE_NUM_FMTS] = { [AMD_IOMMU_V1] = &io_pgtable_amd_iommu_v1_init_fns, [AMD_IOMMU_V2] = &io_pgtable_amd_iommu_v2_init_fns, #endif +#ifdef CONFIG_RISCV_IOMMU + [RISCV_IOMMU] = &io_pgtable_riscv_init_fns, +#endif }; struct io_pgtable_ops *alloc_io_pgtable_ops(enum io_pgtable_fmt fmt, diff --git a/drivers/iommu/riscv/Makefile b/drivers/iommu/riscv/Makefile index 9523eb053cfc..13af452c3052 100644 --- a/drivers/iommu/riscv/Makefile +++ b/drivers/iommu/riscv/Makefile @@ -1 +1 @@ -obj-$(CONFIG_RISCV_IOMMU) += iommu.o iommu-pci.o iommu-platform.o iommu-sysfs.o \ No newline at end of file +obj-$(CONFIG_RISCV_IOMMU) += iommu.o iommu-pci.o iommu-platform.o iommu-sysfs.o io_pgtable.o diff --git a/drivers/iommu/riscv/io_pgtable.c b/drivers/iommu/riscv/io_pgtable.c new file mode 100644 index 000000000000..b6e603e6726e --- /dev/null +++ b/drivers/iommu/riscv/io_pgtable.c @@ -0,0 +1,266 @@ +// SPDX-License-Identifier: GPL-2.0-only +/* + * Copyright © 2022-2023 Rivos Inc. + * + * RISC-V IOMMU page table allocator. + * + * Authors: + * Tomasz Jeznach + * Sebastien Boeuf + */ + +#include +#include +#include +#include +#include +#include +#include +#include + +#include "iommu.h" + +#define io_pgtable_to_domain(x) \ + container_of((x), struct riscv_iommu_domain, pgtbl) + +#define io_pgtable_ops_to_domain(x) \ + io_pgtable_to_domain(container_of((x), struct io_pgtable, ops)) + +static inline size_t get_page_size(size_t size) +{ + if (size >= IOMMU_PAGE_SIZE_512G) + return IOMMU_PAGE_SIZE_512G; + + if (size >= IOMMU_PAGE_SIZE_1G) + return IOMMU_PAGE_SIZE_1G; + + if (size >= IOMMU_PAGE_SIZE_2M) + return IOMMU_PAGE_SIZE_2M; + + return IOMMU_PAGE_SIZE_4K; +} + +static void riscv_iommu_pt_walk_free(pmd_t * ptp, unsigned shift, bool root) +{ + pmd_t *pte, *pt_base; + int i; + + if (shift == PAGE_SHIFT) + return; + + if (root) + pt_base = ptp; + else + pt_base = + (pmd_t *) pfn_to_virt(__page_val_to_pfn(pmd_val(*ptp))); + + /* Recursively free all sub page table pages */ + for (i = 0; i < PTRS_PER_PMD; i++) { + pte = pt_base + i; + if (pmd_present(*pte) && !pmd_leaf(*pte)) + riscv_iommu_pt_walk_free(pte, shift - 9, false); + } + + /* Now free the current page table page */ + if (!root && pmd_present(*pt_base)) + free_page((unsigned long)pt_base); +} + +static void riscv_iommu_free_pgtable(struct io_pgtable *iop) +{ + struct riscv_iommu_domain *domain = io_pgtable_to_domain(iop); + riscv_iommu_pt_walk_free((pmd_t *) domain->pgd_root, PGDIR_SHIFT, true); +} + +static pte_t *riscv_iommu_pt_walk_alloc(pmd_t * ptp, unsigned long iova, + unsigned shift, bool root, + size_t pgsize, + unsigned long (*pd_alloc)(gfp_t), + gfp_t gfp) +{ + pmd_t *pte; + unsigned long pfn; + + if (root) + pte = ptp + ((iova >> shift) & (PTRS_PER_PMD - 1)); + else + pte = (pmd_t *) pfn_to_virt(__page_val_to_pfn(pmd_val(*ptp))) + + ((iova >> shift) & (PTRS_PER_PMD - 1)); + + if ((1ULL << shift) <= pgsize) { + if (pmd_present(*pte) && !pmd_leaf(*pte)) + riscv_iommu_pt_walk_free(pte, shift - 9, false); + return (pte_t *) pte; + } + + if (pmd_none(*pte)) { + pfn = pd_alloc ? virt_to_pfn(pd_alloc(gfp)) : 0; + if (!pfn) + return NULL; + set_pmd(pte, __pmd((pfn << _PAGE_PFN_SHIFT) | _PAGE_TABLE)); + } + + return riscv_iommu_pt_walk_alloc(pte, iova, shift - 9, false, + pgsize, pd_alloc, gfp); +} + +static pte_t *riscv_iommu_pt_walk_fetch(pmd_t * ptp, + unsigned long iova, unsigned shift, + bool root) +{ + pmd_t *pte; + + if (root) + pte = ptp + ((iova >> shift) & (PTRS_PER_PMD - 1)); + else + pte = (pmd_t *) pfn_to_virt(__page_val_to_pfn(pmd_val(*ptp))) + + ((iova >> shift) & (PTRS_PER_PMD - 1)); + + if (pmd_leaf(*pte)) + return (pte_t *) pte; + else if (pmd_none(*pte)) + return NULL; + else if (shift == PAGE_SHIFT) + return NULL; + + return riscv_iommu_pt_walk_fetch(pte, iova, shift - 9, false); +} + +static int riscv_iommu_map_pages(struct io_pgtable_ops *ops, + unsigned long iova, phys_addr_t phys, + size_t pgsize, size_t pgcount, int prot, + gfp_t gfp, size_t *mapped) +{ + struct riscv_iommu_domain *domain = io_pgtable_ops_to_domain(ops); + size_t size = 0; + size_t page_size = get_page_size(pgsize); + pte_t *pte; + pte_t pte_val; + pgprot_t pte_prot; + + if (domain->domain.type == IOMMU_DOMAIN_BLOCKED) + return -ENODEV; + + if (domain->domain.type == IOMMU_DOMAIN_IDENTITY) { + *mapped = pgsize * pgcount; + return 0; + } + + pte_prot = (prot & IOMMU_WRITE) ? + __pgprot(_PAGE_BASE | _PAGE_READ | _PAGE_WRITE | _PAGE_DIRTY) : + __pgprot(_PAGE_BASE | _PAGE_READ); + + while (pgcount--) { + pte = + riscv_iommu_pt_walk_alloc((pmd_t *) domain->pgd_root, iova, + PGDIR_SHIFT, true, page_size, + get_zeroed_page, gfp); + if (!pte) { + *mapped = size; + return -ENOMEM; + } + + pte_val = pfn_pte(phys_to_pfn(phys), pte_prot); + + set_pte(pte, pte_val); + + size += page_size; + iova += page_size; + phys += page_size; + } + + *mapped = size; + return 0; +} + +static size_t riscv_iommu_unmap_pages(struct io_pgtable_ops *ops, + unsigned long iova, size_t pgsize, + size_t pgcount, + struct iommu_iotlb_gather *gather) +{ + struct riscv_iommu_domain *domain = io_pgtable_ops_to_domain(ops); + size_t size = 0; + size_t page_size = get_page_size(pgsize); + pte_t *pte; + + if (domain->domain.type == IOMMU_DOMAIN_IDENTITY) + return pgsize * pgcount; + + while (pgcount--) { + pte = riscv_iommu_pt_walk_fetch((pmd_t *) domain->pgd_root, + iova, PGDIR_SHIFT, true); + if (!pte) + return size; + + set_pte(pte, __pte(0)); + + iommu_iotlb_gather_add_page(&domain->domain, gather, iova, + pgsize); + + size += page_size; + iova += page_size; + } + + return size; +} + +static phys_addr_t riscv_iommu_iova_to_phys(struct io_pgtable_ops *ops, + unsigned long iova) +{ + struct riscv_iommu_domain *domain = io_pgtable_ops_to_domain(ops); + pte_t *pte; + + if (domain->domain.type == IOMMU_DOMAIN_IDENTITY) + return (phys_addr_t) iova; + + pte = riscv_iommu_pt_walk_fetch((pmd_t *) domain->pgd_root, + iova, PGDIR_SHIFT, true); + if (!pte || !pte_present(*pte)) + return 0; + + return (pfn_to_phys(pte_pfn(*pte)) | (iova & PAGE_MASK)); +} + +static void riscv_iommu_tlb_inv_all(void *cookie) +{ +} + +static void riscv_iommu_tlb_inv_walk(unsigned long iova, size_t size, + size_t granule, void *cookie) +{ +} + +static void riscv_iommu_tlb_add_page(struct iommu_iotlb_gather *gather, + unsigned long iova, size_t granule, + void *cookie) +{ +} + +static const struct iommu_flush_ops riscv_iommu_flush_ops = { + .tlb_flush_all = riscv_iommu_tlb_inv_all, + .tlb_flush_walk = riscv_iommu_tlb_inv_walk, + .tlb_add_page = riscv_iommu_tlb_add_page, +}; + +/* NOTE: cfg should point to riscv_iommu_domain structure member pgtbl.cfg */ +static struct io_pgtable *riscv_iommu_alloc_pgtable(struct io_pgtable_cfg *cfg, + void *cookie) +{ + struct io_pgtable *iop = container_of(cfg, struct io_pgtable, cfg); + + cfg->pgsize_bitmap = SZ_4K | SZ_2M | SZ_1G; + cfg->ias = 57; // va mode, SvXX -> ias + cfg->oas = 57; // pa mode, or SvXX+4 -> oas + cfg->tlb = &riscv_iommu_flush_ops; + + iop->ops.map_pages = riscv_iommu_map_pages; + iop->ops.unmap_pages = riscv_iommu_unmap_pages; + iop->ops.iova_to_phys = riscv_iommu_iova_to_phys; + + return iop; +} + +struct io_pgtable_init_fns io_pgtable_riscv_init_fns = { + .alloc = riscv_iommu_alloc_pgtable, + .free = riscv_iommu_free_pgtable, +}; diff --git a/drivers/iommu/riscv/iommu.c b/drivers/iommu/riscv/iommu.c index 9ee7d2b222b5..2ef6952a2109 100644 --- a/drivers/iommu/riscv/iommu.c +++ b/drivers/iommu/riscv/iommu.c @@ -807,7 +807,7 @@ static struct iommu_device *riscv_iommu_probe_device(struct device *dev) /* Initial DC pointer can be NULL if IOMMU is configured in OFF or BARE mode */ ep->dc = riscv_iommu_get_dc(iommu, ep->devid); - dev_info(iommu->dev, "adding device to iommu with devid %i in domain %i\n", + dev_dbg(iommu->dev, "adding device to iommu with devid %i in domain %i\n", ep->devid, ep->domid); dev_iommu_priv_set(dev, ep); @@ -874,7 +874,10 @@ static struct iommu_domain *riscv_iommu_domain_alloc(unsigned type) { struct riscv_iommu_domain *domain; - if (type != IOMMU_DOMAIN_IDENTITY && + if (type != IOMMU_DOMAIN_DMA && + type != IOMMU_DOMAIN_DMA_FQ && + type != IOMMU_DOMAIN_UNMANAGED && + type != IOMMU_DOMAIN_IDENTITY && type != IOMMU_DOMAIN_BLOCKED) return NULL; @@ -890,7 +893,7 @@ static struct iommu_domain *riscv_iommu_domain_alloc(unsigned type) domain->pscid = ida_alloc_range(&riscv_iommu_pscids, 1, RISCV_IOMMU_MAX_PSCID, GFP_KERNEL); - printk("domain type %x alloc %u\n", type, domain->pscid); + printk("domain alloc %u\n", domain->pscid); return &domain->domain; } @@ -903,6 +906,9 @@ static void riscv_iommu_domain_free(struct iommu_domain *iommu_domain) pr_warn("IOMMU domain is not empty!\n"); } + if (domain->pgtbl.cookie) + free_io_pgtable_ops(&domain->pgtbl.ops); + if (domain->pgd_root) free_pages((unsigned long)domain->pgd_root, 0); @@ -959,6 +965,9 @@ static int riscv_iommu_domain_finalize(struct riscv_iommu_domain *domain, if (!domain->pgd_root) return -ENOMEM; + if (!alloc_io_pgtable_ops(RISCV_IOMMU, &domain->pgtbl.cfg, domain)) + return -ENOMEM; + return 0; } @@ -1006,9 +1015,8 @@ static int riscv_iommu_attach_dev(struct iommu_domain *iommu_domain, struct devi return 0; } - if (!dc) { + if (!dc) return -ENODEV; - } /* * S-Stage translation table. G-Stage remains unmodified (BARE). @@ -1104,12 +1112,11 @@ static int riscv_iommu_map_pages(struct iommu_domain *iommu_domain, { struct riscv_iommu_domain *domain = iommu_domain_to_riscv(iommu_domain); - if (domain->domain.type == IOMMU_DOMAIN_IDENTITY) { - *mapped = pgsize * pgcount; - return 0; - } + if (!domain->pgtbl.ops.map_pages) + return -ENODEV; - return -ENODEV; + return domain->pgtbl.ops.map_pages(&domain->pgtbl.ops, iova, phys, + pgsize, pgcount, prot, gfp, mapped); } static size_t riscv_iommu_unmap_pages(struct iommu_domain *iommu_domain, @@ -1118,10 +1125,11 @@ static size_t riscv_iommu_unmap_pages(struct iommu_domain *iommu_domain, { struct riscv_iommu_domain *domain = iommu_domain_to_riscv(iommu_domain); - if (domain->domain.type == IOMMU_DOMAIN_IDENTITY) - return pgsize * pgcount; + if (!domain->pgtbl.ops.unmap_pages) + return 0; - return 0; + return domain->pgtbl.ops.unmap_pages(&domain->pgtbl.ops, iova, pgsize, + pgcount, gather); } static phys_addr_t riscv_iommu_iova_to_phys(struct iommu_domain *iommu_domain, @@ -1129,10 +1137,10 @@ static phys_addr_t riscv_iommu_iova_to_phys(struct iommu_domain *iommu_domain, { struct riscv_iommu_domain *domain = iommu_domain_to_riscv(iommu_domain); - if (domain->domain.type == IOMMU_DOMAIN_IDENTITY) - return (phys_addr_t) iova; + if (!domain->pgtbl.ops.iova_to_phys) + return 0; - return 0; + return domain->pgtbl.ops.iova_to_phys(&domain->pgtbl.ops, iova); } /* diff --git a/drivers/iommu/riscv/iommu.h b/drivers/iommu/riscv/iommu.h index 9140df71e17b..fe32a4eff14e 100644 --- a/drivers/iommu/riscv/iommu.h +++ b/drivers/iommu/riscv/iommu.h @@ -88,6 +88,7 @@ struct riscv_iommu_device { struct riscv_iommu_domain { struct iommu_domain domain; + struct io_pgtable pgtbl; struct list_head endpoints; struct mutex lock; diff --git a/include/linux/io-pgtable.h b/include/linux/io-pgtable.h index 1b7a44b35616..8dd9d3a28e3a 100644 --- a/include/linux/io-pgtable.h +++ b/include/linux/io-pgtable.h @@ -19,6 +19,7 @@ enum io_pgtable_fmt { AMD_IOMMU_V2, APPLE_DART, APPLE_DART2, + RISCV_IOMMU, IO_PGTABLE_NUM_FMTS, }; @@ -258,5 +259,6 @@ extern struct io_pgtable_init_fns io_pgtable_arm_mali_lpae_init_fns; extern struct io_pgtable_init_fns io_pgtable_amd_iommu_v1_init_fns; extern struct io_pgtable_init_fns io_pgtable_amd_iommu_v2_init_fns; extern struct io_pgtable_init_fns io_pgtable_apple_dart_init_fns; +extern struct io_pgtable_init_fns io_pgtable_riscv_init_fns; #endif /* __IO_PGTABLE_H */ -- 2.34.1