Received: by 2002:a25:824b:0:0:0:0:0 with SMTP id d11csp715534ybn; Wed, 2 Oct 2019 05:10:37 -0700 (PDT) X-Google-Smtp-Source: APXvYqwOM23sft2XhQ/UGAJ/yoOD12O8IjvuZQrkbEcrdfRvkbxS6OIH1WR0PX5ac3gurSadEW5I X-Received: by 2002:a17:906:fc02:: with SMTP id ov2mr2690119ejb.273.1570018236954; Wed, 02 Oct 2019 05:10:36 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1570018236; cv=none; d=google.com; s=arc-20160816; b=MJLrYqNewZQVkjojs9r5vrfgQ6MyjfbD98LwYBpffRCOPkYXVRxdfZOtK7uABhS6P9 zTkbwZO+g/2KUktwt/JnZEWfJ1EO9nzOpfkrGbB7RE2a8TUUX6P7BOqhKHbBX22zBkZc eFH2JmNkMvLWoudrsOj5EhMkJ5nX1xUzpmYQ493GeVYQDKHUG3CfCPKvRo3OwJdchNgp 3JinBkXVm0WgYoigpTKwG9EAvh1axcEqNOS0eswrjoVAQGb/S2eGZtPSLp9acIeZa16G RCGyO4tB5nNgmEGb7h9/ulfBlKd9YbMaUgKUgDQUVtDScz8IgJa9+taSpUmHlfiqkvpm B+jQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:content-transfer-encoding :content-language:in-reply-to:mime-version:user-agent:date :message-id:from:references:cc:to:subject; bh=CQ1PXufDp8CDFypm15+No0OESmcX7QmdiZ8zhOPghVY=; b=hwK1gBNc+H9Xh7kllxTQm6C8NZT4v5PANzMhpJLgFefgvfBizwCGn3hPlgJMSaIhri wN4a2JGXj5XGxTytSSW878nblj4i1DGZa1fgrFIbBwFvhO9doqmO5svl7kSXF0loPoLO Tq4MrIHENq5oRMUD0N/OU2PwdHnb2yWZhPbHy3EbvO5pqmt19dxgUEJxmcqt8bRYTI4Y T0jhHEskShOtUl4FHUfh+Y18bnAFi1XvdA6FjCbglDGgCXcr3A/jnohdmlzpZYj98pjh UnMcIyxULKHzTCeP3TFmQH71oj/WUbMJxiMLN5BGwSAXm0Mfl92t7jp5LyvUwqK3DDaA AytQ== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id d35si10657648ede.440.2019.10.02.05.10.12; Wed, 02 Oct 2019 05:10:36 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1728033AbfJBKfV (ORCPT + 99 others); Wed, 2 Oct 2019 06:35:21 -0400 Received: from foss.arm.com ([217.140.110.172]:40956 "EHLO foss.arm.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1725999AbfJBKfU (ORCPT ); Wed, 2 Oct 2019 06:35:20 -0400 Received: from usa-sjc-imap-foss1.foss.arm.com (unknown [10.121.207.14]) by usa-sjc-mx-foss1.foss.arm.com (Postfix) with ESMTP id 88AA11000; Wed, 2 Oct 2019 03:35:19 -0700 (PDT) Received: from [10.1.197.57] (e110467-lin.cambridge.arm.com [10.1.197.57]) by usa-sjc-imap-foss1.foss.arm.com (Postfix) with ESMTPSA id 2F1FC3F739; Wed, 2 Oct 2019 03:35:17 -0700 (PDT) Subject: Re: [PATCH] iommu/mediatek: Move the tlb_sync into tlb_flush To: Tomasz Figa , Yong Wu Cc: youlin.pei@mediatek.com, anan.sun@mediatek.com, Nicolas Boichat , cui.zhang@mediatek.com, srv_heupstream , Will Deacon , Linux Kernel Mailing List , Evan Green , chao.hao@mediatek.com, "list@263.net:IOMMU DRIVERS" , Joerg Roedel , iommu@lists.linux-foundation.org, "moderated list:ARM/Mediatek SoC support" , Matthias Brugger , "list@263.net:IOMMU DRIVERS" , Joerg Roedel , linux-arm-kernel@lists.infradead.org References: <1569822142-14303-1-git-send-email-yong.wu@mediatek.com> From: Robin Murphy Message-ID: <366b0bda-d874-9109-5c83-ff27301f3486@arm.com> Date: Wed, 2 Oct 2019 11:35:16 +0100 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:60.0) Gecko/20100101 Thunderbird/60.6.1 MIME-Version: 1.0 In-Reply-To: Content-Type: text/plain; charset=utf-8; format=flowed Content-Language: en-GB Content-Transfer-Encoding: 7bit Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 02/10/2019 06:18, Tomasz Figa wrote: > Hi Yong, > > On Mon, Sep 30, 2019 at 2:42 PM Yong Wu wrote: >> >> The commit 4d689b619445 ("iommu/io-pgtable-arm-v7s: Convert to IOMMU API >> TLB sync") help move the tlb_sync of unmap from v7s into the iommu >> framework. It helps add a new function "mtk_iommu_iotlb_sync", But it >> lacked the dom->pgtlock, then it will cause the variable >> "tlb_flush_active" may be changed unexpectedly, we could see this warning >> log randomly: >> > > Thanks for the patch! Please see my comments inline. > >> mtk-iommu 10205000.iommu: Partial TLB flush timed out, falling back to >> full flush >> >> To fix this issue, we can add dom->pgtlock in the "mtk_iommu_iotlb_sync". >> And when checking this issue, we find that __arm_v7s_unmap call >> io_pgtable_tlb_add_flush consecutively when it is supersection/largepage, >> this also is potential unsafe for us. There is no tlb flush queue in the >> MediaTek M4U HW. The HW always expect the tlb_flush/tlb_sync one by one. >> If v7s don't always gurarantee the sequence, Thus, In this patch I move >> the tlb_sync into tlb_flush(also rename the function deleting "_nosync"). >> and we don't care if it is leaf, rearrange the callback functions. Also, >> the tlb flush/sync was already finished in v7s, then iotlb_sync and >> iotlb_sync_all is unnecessary. > > Performance-wise, we could do much better. Instead of synchronously > syncing at the end of mtk_iommu_tlb_add_flush(), we could sync at the > beginning, if there was any previous flush still pending. We would > also have to keep the .iotlb_sync() callback, to take care of waiting > for the last flush. That would allow better pipelining with CPU in > cases like this: > > for (all pages in range) { > change page table(); > flush(); > } > > "change page table()" could execute while the IOMMU is flushing the > previous change. FWIW, given that the underlying invalidation mechanism is range-based, this driver would be an ideal candidate for making use of the new iommu_gather mechanism. As a fix for stable, though, simply ensuring that add_flush syncs any pending invalidation before issuing a new one sounds like a good idea (and probably a simpler patch too). [...] >> @@ -574,8 +539,7 @@ static int mtk_iommu_of_xlate(struct device *dev, struct of_phandle_args *args) >> .detach_dev = mtk_iommu_detach_device, >> .map = mtk_iommu_map, >> .unmap = mtk_iommu_unmap, >> - .flush_iotlb_all = mtk_iommu_flush_iotlb_all, > > Don't we still want .flush_iotlb_all()? I think it should be more > efficient in some cases than doing a big number of single flushes. > (That said, the previous implementation didn't do any flush at all. It > just waited for previously queued flushes to happen. Was that > expected?) Commit 07fdef34d2be ("iommu/arm-smmu-v3: Implement flush_iotlb_all hook") has an explanation of what the deal was there - similarly, it's probably worth this driver implementing it properly as well now (but that's really a separate patch). Robin.