Received: by 2002:a25:824b:0:0:0:0:0 with SMTP id d11csp6532824ybn; Sun, 29 Sep 2019 22:45:57 -0700 (PDT) X-Google-Smtp-Source: APXvYqy62cXvULjpOBJu2gRWNrKPZVR5zFDpj8zKKj6IYT0T+YGNfuKdApgL1sgN49jcudSUmpJc X-Received: by 2002:a17:906:b84c:: with SMTP id ga12mr18009581ejb.0.1569822356972; Sun, 29 Sep 2019 22:45:56 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1569822356; cv=none; d=google.com; s=arc-20160816; b=PhcoSLraqM6bEhZdLb1ZvQG0QXOLQ/doSTI5JEiYaMgvG1SQab/NMVdgipg50XLAZN I/IH/7MRNLe8DQFxK4he1QMb45weFEcUmQV0qDwnLx16Lr59SdfOah/8zCrZJFPCi3H3 +s7rBuQWe+RO3FJDXmHGWWy8synWsZCQeGf6Vi1PfQE+O8tVn3WZMiGatuBOlnGD4L3R aPWsEU+LJebobKa8XBHsXDwaFqYpHwSGhzZUBN4uSO/s1biFdz6XsIyXMgT0jb22tv8L qgnMH1NQdWH8h0SY2Urw6sOUoyhUCZTzT2/0F4luOkNCOx0QosCw60PTO6H0Ywei/x12 lElA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:mime-version:message-id:date:subject:cc :to:from; bh=t+rr8W0i3gT6P0tsV80vEXh8rGYJ+qi0kf5pMw/WKkQ=; b=stVxA0JXoty5OkPDd47ZBOmVHfEDr63eL03qxWzMKgKeGSQxRT0GBtvS49bzenXVeB B3JeTEDM4aQw1zXQR00y9G5B49vcGXluMRrXYQOQoZJZlv4X69eQq5SBzULxP/6IImRe TpWv7coj/9cl49o5749h3kF5rW5T7H4u/Y0mYm5GQRj5g6Kp2cW8/kmSNLIZYv95Thns oMHnh8RSCNMm6VkPg0AGALKjViI19CZn4Se/qFnFpGv912dy7bhi4EOfgofL03ExCQf7 F3QGL8FzgASzUpGOT20fRjtpAf144LuXvxTpLAGkKw26YkP+I73+EuFHCP4gHGTosIOw dS4Q== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=mediatek.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id v3si6265059edc.404.2019.09.29.22.45.32; Sun, 29 Sep 2019 22:45:56 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=mediatek.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1729524AbfI3Fmi (ORCPT + 99 others); Mon, 30 Sep 2019 01:42:38 -0400 Received: from mailgw01.mediatek.com ([210.61.82.183]:13421 "EHLO mailgw01.mediatek.com" rhost-flags-OK-FAIL-OK-FAIL) by vger.kernel.org with ESMTP id S1725767AbfI3Fmh (ORCPT ); Mon, 30 Sep 2019 01:42:37 -0400 X-UUID: a6608df61a61441dbd2460752cb632e5-20190930 X-UUID: a6608df61a61441dbd2460752cb632e5-20190930 Received: from mtkcas08.mediatek.inc [(172.21.101.126)] by mailgw01.mediatek.com (envelope-from ) (Cellopoint E-mail Firewall v4.1.10 Build 0809 with TLS) with ESMTP id 598256299; Mon, 30 Sep 2019 13:42:28 +0800 Received: from mtkcas07.mediatek.inc (172.21.101.84) by mtkmbs07n2.mediatek.inc (172.21.101.141) with Microsoft SMTP Server (TLS) id 15.0.1395.4; Mon, 30 Sep 2019 13:42:26 +0800 Received: from localhost.localdomain (10.17.3.153) by mtkcas07.mediatek.inc (172.21.101.73) with Microsoft SMTP Server id 15.0.1395.4 via Frontend Transport; Mon, 30 Sep 2019 13:42:24 +0800 From: Yong Wu To: Matthias Brugger , Joerg Roedel , Will Deacon CC: Evan Green , Robin Murphy , Tomasz Figa , , , , , , , , Nicolas Boichat , , , Subject: [PATCH] iommu/mediatek: Move the tlb_sync into tlb_flush Date: Mon, 30 Sep 2019 13:42:22 +0800 Message-ID: <1569822142-14303-1-git-send-email-yong.wu@mediatek.com> X-Mailer: git-send-email 1.9.1 MIME-Version: 1.0 Content-Type: text/plain X-MTK: N Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org The commit 4d689b619445 ("iommu/io-pgtable-arm-v7s: Convert to IOMMU API TLB sync") help move the tlb_sync of unmap from v7s into the iommu framework. It helps add a new function "mtk_iommu_iotlb_sync", But it lacked the dom->pgtlock, then it will cause the variable "tlb_flush_active" may be changed unexpectedly, we could see this warning log randomly: mtk-iommu 10205000.iommu: Partial TLB flush timed out, falling back to full flush To fix this issue, we can add dom->pgtlock in the "mtk_iommu_iotlb_sync". And when checking this issue, we find that __arm_v7s_unmap call io_pgtable_tlb_add_flush consecutively when it is supersection/largepage, this also is potential unsafe for us. There is no tlb flush queue in the MediaTek M4U HW. The HW always expect the tlb_flush/tlb_sync one by one. If v7s don't always gurarantee the sequence, Thus, In this patch I move the tlb_sync into tlb_flush(also rename the function deleting "_nosync"). and we don't care if it is leaf, rearrange the callback functions. Also, the tlb flush/sync was already finished in v7s, then iotlb_sync and iotlb_sync_all is unnecessary. Besides, there are two minor changes: a) Use writel for the register F_MMU_INV_RANGE which is for triggering the HW work. We expect all the setting(iova_start/iova_end...) have already been finished before F_MMU_INV_RANGE. b) Reduce the tlb timeout value from 100000us to 1000us. the original value is so long that affect the multimedia performance. Fixes: 4d689b619445 ("iommu/io-pgtable-arm-v7s: Convert to IOMMU API TLB sync") Signed-off-by: Chao Hao Signed-off-by: Yong Wu --- This patch looks break the logic for tlb_flush and tlb_sync. I'm not sure if it is reasonable. If someone has concern, I could change: a) Add dom->pgtlock in the mtk_iommu_iotlb_sync b) Add a io_pgtable_tlb_sync in [1]. [1] https://elixir.bootlin.com/linux/v5.3-rc1/source/drivers/iommu/io-pgtable-arm-v7s.c#L655 This patch rebase on Joerg's mediatek-smmu-merge branch which has mt8183 and Will's "Rework IOMMU API to allow for batching of invalidation". --- drivers/iommu/mtk_iommu.c | 74 ++++++++++++----------------------------------- drivers/iommu/mtk_iommu.h | 1 - 2 files changed, 19 insertions(+), 56 deletions(-) diff --git a/drivers/iommu/mtk_iommu.c b/drivers/iommu/mtk_iommu.c index 6066272..e13cc56 100644 --- a/drivers/iommu/mtk_iommu.c +++ b/drivers/iommu/mtk_iommu.c @@ -173,11 +173,12 @@ static void mtk_iommu_tlb_flush_all(void *cookie) } } -static void mtk_iommu_tlb_add_flush_nosync(unsigned long iova, size_t size, - size_t granule, bool leaf, - void *cookie) +static void mtk_iommu_tlb_add_flush(unsigned long iova, size_t size, + size_t granule, void *cookie) { struct mtk_iommu_data *data = cookie; + int ret; + u32 tmp; for_each_m4u(data) { writel_relaxed(F_INVLD_EN1 | F_INVLD_EN0, @@ -186,25 +187,15 @@ static void mtk_iommu_tlb_add_flush_nosync(unsigned long iova, size_t size, writel_relaxed(iova, data->base + REG_MMU_INVLD_START_A); writel_relaxed(iova + size - 1, data->base + REG_MMU_INVLD_END_A); - writel_relaxed(F_MMU_INV_RANGE, - data->base + REG_MMU_INVALIDATE); - data->tlb_flush_active = true; - } -} - -static void mtk_iommu_tlb_sync(void *cookie) -{ - struct mtk_iommu_data *data = cookie; - int ret; - u32 tmp; - - for_each_m4u(data) { - /* Avoid timing out if there's nothing to wait for */ - if (!data->tlb_flush_active) - return; + writel(F_MMU_INV_RANGE, data->base + REG_MMU_INVALIDATE); + /* + * There is no tlb flush queue in the HW, the HW always expect + * tlb_flush and tlb_sync one by one. Here tlb_sync always + * follows tlb_flush to avoid break the sequence. + */ ret = readl_poll_timeout_atomic(data->base + REG_MMU_CPE_DONE, - tmp, tmp != 0, 10, 100000); + tmp, tmp != 0, 10, 1000); if (ret) { dev_warn(data->dev, "Partial TLB flush timed out, falling back to full flush\n"); @@ -212,36 +203,21 @@ static void mtk_iommu_tlb_sync(void *cookie) } /* Clear the CPE status */ writel_relaxed(0, data->base + REG_MMU_CPE_DONE); - data->tlb_flush_active = false; } } -static void mtk_iommu_tlb_flush_walk(unsigned long iova, size_t size, - size_t granule, void *cookie) +static void mtk_iommu_tlb_flush_page(struct iommu_iotlb_gather *gather, + unsigned long iova, size_t granule, + void *cookie) { - mtk_iommu_tlb_add_flush_nosync(iova, size, granule, false, cookie); - mtk_iommu_tlb_sync(cookie); -} - -static void mtk_iommu_tlb_flush_leaf(unsigned long iova, size_t size, - size_t granule, void *cookie) -{ - mtk_iommu_tlb_add_flush_nosync(iova, size, granule, true, cookie); - mtk_iommu_tlb_sync(cookie); -} - -static void mtk_iommu_tlb_flush_page_nosync(struct iommu_iotlb_gather *gather, - unsigned long iova, size_t granule, - void *cookie) -{ - mtk_iommu_tlb_add_flush_nosync(iova, granule, granule, true, cookie); + mtk_iommu_tlb_add_flush(iova, granule, granule, cookie); } static const struct iommu_flush_ops mtk_iommu_flush_ops = { .tlb_flush_all = mtk_iommu_tlb_flush_all, - .tlb_flush_walk = mtk_iommu_tlb_flush_walk, - .tlb_flush_leaf = mtk_iommu_tlb_flush_leaf, - .tlb_add_page = mtk_iommu_tlb_flush_page_nosync, + .tlb_flush_walk = mtk_iommu_tlb_add_flush, + .tlb_flush_leaf = mtk_iommu_tlb_add_flush, + .tlb_add_page = mtk_iommu_tlb_flush_page, }; static irqreturn_t mtk_iommu_isr(int irq, void *dev_id) @@ -445,17 +421,6 @@ static size_t mtk_iommu_unmap(struct iommu_domain *domain, return unmapsz; } -static void mtk_iommu_flush_iotlb_all(struct iommu_domain *domain) -{ - mtk_iommu_tlb_sync(mtk_iommu_get_m4u_data()); -} - -static void mtk_iommu_iotlb_sync(struct iommu_domain *domain, - struct iommu_iotlb_gather *gather) -{ - mtk_iommu_tlb_sync(mtk_iommu_get_m4u_data()); -} - static phys_addr_t mtk_iommu_iova_to_phys(struct iommu_domain *domain, dma_addr_t iova) { @@ -574,8 +539,7 @@ static int mtk_iommu_of_xlate(struct device *dev, struct of_phandle_args *args) .detach_dev = mtk_iommu_detach_device, .map = mtk_iommu_map, .unmap = mtk_iommu_unmap, - .flush_iotlb_all = mtk_iommu_flush_iotlb_all, - .iotlb_sync = mtk_iommu_iotlb_sync, + /* No iotlb_sync here since the tlb_sync always follows the tlb_flush */ .iova_to_phys = mtk_iommu_iova_to_phys, .add_device = mtk_iommu_add_device, .remove_device = mtk_iommu_remove_device, diff --git a/drivers/iommu/mtk_iommu.h b/drivers/iommu/mtk_iommu.h index fc0f16e..24712f5 100644 --- a/drivers/iommu/mtk_iommu.h +++ b/drivers/iommu/mtk_iommu.h @@ -57,7 +57,6 @@ struct mtk_iommu_data { struct mtk_iommu_domain *m4u_dom; struct iommu_group *m4u_group; bool enable_4GB; - bool tlb_flush_active; struct iommu_device iommu; const struct mtk_iommu_plat_data *plat_data; -- 1.9.1