Received: by 2002:a05:7412:1703:b0:e2:908c:2ebd with SMTP id dm3csp3405655rdb; Tue, 29 Aug 2023 14:40:46 -0700 (PDT) X-Google-Smtp-Source: AGHT+IE2Nm0q7MUoqGn2L6dsybKjOv5ni9JKlNYmD+G4UeED5jQ6vyFBhMCDPFDXRZNsoADCFMrA X-Received: by 2002:aa7:c0d4:0:b0:52b:d169:b374 with SMTP id j20-20020aa7c0d4000000b0052bd169b374mr448375edp.3.1693345245837; Tue, 29 Aug 2023 14:40:45 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1693345245; cv=none; d=google.com; s=arc-20160816; b=DFVdgiOfYukl70DPddhAXHQ5HIQFVvpk0w9qP7JT/uYpHYORSfIYKzUOonTYez2nnI EZfQj6iLof61xGSepbONv87X5B5asVdrezsGXe99sL+/XEaeD1PuPS14L3ypVxzZ18uU /c1R/HaeflsmAg0qFXnaXIrAeFewZIdUKVZV0ocmXY/Rl6iaz9ZFgrkZkeam+yO+uAlz c1tkOdo6ycy99H3Pi/leGCo1EcRzbDbcchqHQKOYrR51h/DySB179TAQTLa5NOib3uL6 lwbxGpA239pwc0ZxiqHZVMuM/MUXsbKK6/zjmnao5AMHZvmr3n8cigHhQTrx+ZDL65Ow N+zA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:in-reply-to:from :content-language:references:cc:to:subject:user-agent:mime-version :date:message-id; bh=t68yjao4KcphTBJLh0mvP4llyo+pYHfVWmpWq9n9vDY=; fh=U9e3T4mBVG+BBWkl/zbLRApTT+uurLBXaGQViuNB4xk=; b=Ugp0Mi+QyVFo/Nrmi5tgqIUu35q+r9ak56OwNrXK5TzkZfXJhdd4WD/tySMOWLq+Hn j2sLcKiSH0DxMIS+aknI+n9r9xzoaqJU+vGeDdpfAyiYLIHvocYs970a0j41PSZMx+0X GB2k4EDzIgn4CKx4IG05RM2GTSKcsibUtjvR4+tqqppihjWZuSNA96j+lw5sm/6MCTAP DnboABqv4/c0TPHaBVTvxh04ZvGj0dtRBPTeliXwhctSm4ixgBrsfYxcDY2bVyvgWFSy 8uuEaBxRMn69tUATVg9QBEfdNmAGizLxpTU5GsdoMFzaabtXIE8r8XJzPHJfx3q4yD+4 4XoQ== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=arm.com Return-Path: Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id a11-20020aa7cf0b000000b0052543ea081asi6839937edy.7.2023.08.29.14.40.18; Tue, 29 Aug 2023 14:40:45 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=arm.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S238633AbjH2VZX (ORCPT + 99 others); Tue, 29 Aug 2023 17:25:23 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:43424 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S239464AbjH2VZV (ORCPT ); Tue, 29 Aug 2023 17:25:21 -0400 Received: from foss.arm.com (foss.arm.com [217.140.110.172]) by lindbergh.monkeyblade.net (Postfix) with ESMTP id ED27DBD for ; Tue, 29 Aug 2023 14:25:17 -0700 (PDT) Received: from usa-sjc-imap-foss1.foss.arm.com (unknown [10.121.207.14]) by usa-sjc-mx-foss1.foss.arm.com (Postfix) with ESMTP id EC8072F4; Tue, 29 Aug 2023 14:25:56 -0700 (PDT) Received: from [10.57.2.162] (unknown [10.57.2.162]) by usa-sjc-imap-foss1.foss.arm.com (Postfix) with ESMTPSA id 2FF393F740; Tue, 29 Aug 2023 14:25:16 -0700 (PDT) Message-ID: Date: Tue, 29 Aug 2023 22:25:10 +0100 MIME-Version: 1.0 User-Agent: Mozilla/5.0 (Windows NT 10.0; rv:102.0) Gecko/20100101 Thunderbird/102.14.0 Subject: Re: [PATCH 1/3] iommu/io-pgtable-arm: Add nents_per_pgtable in struct io_pgtable_cfg To: Nicolin Chen Cc: will@kernel.org, jgg@nvidia.com, joro@8bytes.org, jean-philippe@linaro.org, apopple@nvidia.com, linux-kernel@vger.kernel.org, linux-arm-kernel@lists.infradead.org, iommu@lists.linux.dev References: <0fe68babdb3a07adf024ed471fead4e3eb7e703f.1692693557.git.nicolinc@nvidia.com> <61f9b371-7c45-26b1-ec0f-600765280c89@arm.com> Content-Language: en-GB From: Robin Murphy In-Reply-To: Content-Type: text/plain; charset=UTF-8; format=flowed Content-Transfer-Encoding: 7bit X-Spam-Status: No, score=-3.1 required=5.0 tests=BAYES_00,NICE_REPLY_A, RCVD_IN_DNSWL_BLOCKED,SPF_HELO_NONE,SPF_NONE autolearn=ham autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 2023-08-29 21:15, Nicolin Chen wrote: > On Tue, Aug 29, 2023 at 04:37:00PM +0100, Robin Murphy wrote: > >> On 2023-08-22 17:42, Nicolin Chen wrote: >>> On Tue, Aug 22, 2023 at 10:19:21AM +0100, Robin Murphy wrote: >>> >>>>> out_free_data: >>>>> @@ -1071,6 +1073,7 @@ arm_mali_lpae_alloc_pgtable(struct io_pgtable_cfg *cfg, void *cookie) >>>>> ARM_MALI_LPAE_TTBR_ADRMODE_TABLE; >>>>> if (cfg->coherent_walk) >>>>> cfg->arm_mali_lpae_cfg.transtab |= ARM_MALI_LPAE_TTBR_SHARE_OUTER; >>>>> + cfg->nents_per_pgtable = 1 << data->bits_per_level; >>>> >>>> The result of this highly complex and expensive calculation is clearly >>>> redundant with the existing bits_per_level field, so why do we need to >>>> waste space storing when the driver could simply use bits_per_level? >>> >>> bits_per_level is in the private struct arm_lpae_io_pgtable, while >>> drivers can only access struct io_pgtable_cfg. Are you suggesting >>> to move bits_per_level out of the private struct arm_lpae_io_pgtable >>> to the public struct io_pgtable_cfg? >>> >>> Or am I missing another bits_per_level? >> >> Bleh, apologies, I always confuse myself trying to remember the fiddly >> design of io-pgtable data. However, I think this then ends up proving >> the opposite point - the number of pages per table only happens to be a >> fixed constant for certain formats like LPAE, but does not necessarily >> generalise. For instance for a single v7s config it would be 1024 or 256 >> or 16 depending on what has actually been unmapped. >> >> The mechanism as proposed implicitly assumes LPAE format, so I still >> think we're better off making that assumption explicit. And at that >> point arm-smmu-v3 can then freely admit it already knows the number is >> simply 1/8th of the domain page size. > > Hmm, I am not getting that "1/8th" part, would you mind elaborating? If we know the format is LPAE, then we already know that nearly all pagetable levels are one full page, and the PTEs are 64 bits long. No magic data conduit required. > Also, what we need is actually an arbitrary number for max_tlbi_ops. > And I think it could be irrelevant to the page size, i.e. either a > 4K pgsize or a 64K pgsize could use the same max_tlbi_ops number, > because what eventually impacts the latency is the number of loops > of building/issuing commands. Although there is perhaps a counter-argument for selective invalidation, that if you're using 64K pages to improve TLB hit rates vs. 4K, then you have more to lose from overinvalidation, so maybe a single threshold isn't so appropriate for both. Yes, ultimately it all comes down to picking an arbitrary number, but the challenge is that we want to pick a *good* one, and ideally have some reasoning behind it. As Will suggested, copying what the mm layer does gives us an easy line of reasoning, even if it's just in the form of passing the buck. And that's actually quite attractive, since otherwise we'd then have to get into the question of what really is the latency of building and issuing commands, since that clearly depends on how fast the CPU is, and how fast the SMMU is, and how busy the SMMU is, and how large the command queue is, and how many other CPUs are also contending for the command queue... and very quickly it becomes hard to believe that any simple constant can be good for all possible systems. > So, combining your narrative above that nents_per_pgtable isn't so > general as we have in the tlbflush for MMU, FWIW I meant it doesn't generalise well enough to be a common io-pgtable interface; I have no issue with it forming the basis of an SMMUv3-specific heuristic when it *is* a relevant concept to all the pagetable formats SMMUv3 can possibly support. Thanks, Robin.