Received: by 2002:a05:6a10:206:0:0:0:0 with SMTP id 6csp4651325pxj; Tue, 22 Jun 2021 05:12:37 -0700 (PDT) X-Google-Smtp-Source: ABdhPJzi430VTps7MX6Iupjp7wK1P4jai1URZzEYY+S+DUKBd06Tvx47rYYtFRdVXy01ARAHxGaT X-Received: by 2002:a17:906:36d5:: with SMTP id b21mr3685656ejc.258.1624363957437; Tue, 22 Jun 2021 05:12:37 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1624363957; cv=none; d=google.com; s=arc-20160816; b=SF8tzeDJeaBn/PoTTCnYjqA983DRo+hzoXrxwsGiV4BwWUrKlYu4uzIWZRNlrEpza0 xaeXNyOVWhwhSuZd8+E4CCDDdCnJzv8k7QCAXR6ch8B7atg6REW0zZQ76WktmYcYe4Xg S4Fxvsd/l5uiqop7gYZF930hIy/75pbOAifqZ9BlqjnVk79TYrnzA9RjuU76/5jMd2rW ZpwJEzTjDCyslu8wQ5ipIOAnyt4qc5G+oDXD+UnG07XVfbI0l5wTrcYbSIqcSqETwnxd yLvMSGdILqgdkyI//kekg6OA4tvcbyFSfvLHpdjhqWvhvrI/tJGCkdIfN3KPxDQV8pLH nsWA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:content-language :in-reply-to:mime-version:user-agent:date:message-id:from:references :cc:to:subject; bh=ng8Bb6l8ramA6jW9+GcMZUR7S0TGEA0MDydMRZeqeeA=; b=FwSBgk7jyCM6xA5j5K07yw37u2soOKqKKRqNnGN7R2NkRXjRn4zmBAskn0rj1P7iea 56N4EKMWdW9fBAoitdrTfqZ8jRrzrdMGM3lc2ZsLok6ySyqwBLvFMsoRARHmHEjXjQYV tsr4aRVJc3xfdxiu6xXhNT8Z/LNKxb5FLTRfHkqpalR2XedlWpDgNwlBfpbNPBgJs+i7 favVVRuqvOn5ab0cLyzI20MW663ckB9QJtv81wjI++I7tul7ozt4oB7RlaJyT2qHZ8X9 6RvQFoDXeAbtbn7UV21ImJCAbCPqIwy+LFy7Y9TzEvA9GI1K2cs4KeJyU6nCuBi/wBub +ySg== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=arm.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id di18si19363567edb.218.2021.06.22.05.12.14; Tue, 22 Jun 2021 05:12:37 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=arm.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S231225AbhFVMNg (ORCPT + 99 others); Tue, 22 Jun 2021 08:13:36 -0400 Received: from foss.arm.com ([217.140.110.172]:48112 "EHLO foss.arm.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S229912AbhFVMNf (ORCPT ); Tue, 22 Jun 2021 08:13:35 -0400 Received: from usa-sjc-imap-foss1.foss.arm.com (unknown [10.121.207.14]) by usa-sjc-mx-foss1.foss.arm.com (Postfix) with ESMTP id D5853ED1; Tue, 22 Jun 2021 05:11:19 -0700 (PDT) Received: from [10.57.9.136] (unknown [10.57.9.136]) by usa-sjc-imap-foss1.foss.arm.com (Postfix) with ESMTPSA id 73A493F694; Tue, 22 Jun 2021 05:11:18 -0700 (PDT) Subject: Re: [PATCHv2 1/3] iommu/io-pgtable: Add a quirk to use tlb_flush_all() for partial walk flush To: Sai Prakash Ranjan Cc: Will Deacon , Joerg Roedel , Thierry Reding , linux-arm-msm@vger.kernel.org, Douglas Anderson , linux-kernel@vger.kernel.org, iommu@lists.linux-foundation.org, linux-arm-kernel@lists.infradead.org References: <904f283c-f8b1-ba84-d010-eacc87bb53c5@arm.com> From: Robin Murphy Message-ID: <2b093b93-7fab-be35-59d8-4463c199719a@arm.com> Date: Tue, 22 Jun 2021 13:11:12 +0100 User-Agent: Mozilla/5.0 (Windows NT 10.0; rv:78.0) Gecko/20100101 Thunderbird/78.10.1 MIME-Version: 1.0 In-Reply-To: Content-Type: text/plain; charset=utf-8; format=flowed Content-Language: en-GB Content-Transfer-Encoding: 7bit Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 2021-06-22 08:11, Sai Prakash Ranjan wrote: > Hi Robin, > > On 2021-06-21 21:15, Robin Murphy wrote: >> On 2021-06-18 03:51, Sai Prakash Ranjan wrote: >>> Add a quirk IO_PGTABLE_QUIRK_TLB_INV_ALL to invalidate entire context >>> with tlb_flush_all() callback in partial walk flush to improve unmap >>> performance on select few platforms where the cost of over-invalidation >>> is less than the unmap latency. >> >> I still think this doesn't belong anywhere near io-pgtable at all. >> It's a driver-internal decision how exactly it implements a non-leaf >> invalidation, and that may be more complex than a predetermined >> boolean decision. For example, I've just realised for SMMUv3 we can't >> invalidate multiple levels of table at once with a range command, >> since if we assume the whole thing is mapped at worst-case page >> granularity we may fail to invalidate any parts which are mapped as >> intermediate-level blocks. If invalidating a 1GB region (with 4KB >> granule) means having to fall back to 256K non-range commands, we may >> not want to invalidate by VA then, even though doing so for a 2MB >> region is still optimal. >> >> It's also quite feasible that drivers might want to do this for leaf >> invalidations too - if you don't like issuing 512 commands to >> invalidate 2MB, do you like issuing 511 commands to invalidate 2044KB? >> - and at that point the logic really has to be in the driver anyway. >> > > Ok I will move this to tlb_flush_walk() functions in the drivers. In the > previous > v1 thread, you suggested to make the choice in iommu_get_dma_strict() test, > I assume you meant the test in iommu_dma_init_domain() with a flag or > was it > the leaf driver(ex:arm-smmu.c) test of iommu_get_dma_strict() in > init_domain? Yes, I meant literally inside the same condition where we currently set "pgtbl_cfg.quirks |= IO_PGTABLE_QUIRK_NON_STRICT;" in arm_smmu_init_domain_context(). > I am still a bit confused on where this flag would be? Should this be a > part > of struct iommu_domain? Well, if you were to rewrite the config with an alternative set of flush_ops at that point it would be implicit. For a flag, probably either in arm_smmu_domain or arm_smmu_impl. Maybe a flag would be less useful than generalising straight to a "maximum number of by-VA invalidations it's worth sending individually" threshold value? It's clear to me what overall shape and separation of responsibility is most logical, but beyond that I don't have a particularly strong opinion on the exact implementation; I've just been chucking ideas around :) Cheers, Robin.