Received: by 2002:ab2:3141:0:b0:1ed:23cc:44d1 with SMTP id i1csp1979729lqg; Mon, 4 Mar 2024 09:11:36 -0800 (PST) X-Forwarded-Encrypted: i=3; AJvYcCUzTUqo/5dWRl7L4lKuZvR7DVwMogEGAnww3UrxXrg9D71Q/uhbszmbnEKkd43VjV14XiQHejdC7q8S+o20Ouftv2EOuOSEeOSOJSEeqg== X-Google-Smtp-Source: AGHT+IE/JGLqB+mEQCZtl6KYz9fiEe3M9zRRysbNyP475wdINw5gK1+P3qKz1ekyMDPcXwOuYuCh X-Received: by 2002:a05:6a20:1a87:b0:1a1:20b6:bf11 with SMTP id ci7-20020a056a201a8700b001a120b6bf11mr9181993pzb.50.1709572296480; Mon, 04 Mar 2024 09:11:36 -0800 (PST) ARC-Seal: i=2; a=rsa-sha256; t=1709572296; cv=pass; d=google.com; s=arc-20160816; b=LxhU1YTO3fhKMb8V0e1GkXz2KZJ8pxPYeiHJE2sgkjCGLu7oaaKRlkUoWxI3xAPNbC KwMoBOxkdIjgammDu2QNuzMelpDHIq/N8f0FOUCFG8C//S+DbYJZgILohrD/PQ3cLqSq PtzWbd25Rncsw8izrzaMoe9upl5cu9beiDWbuT0/0aVT1V0KDorWg5iDe1d3JoT55r+q jnlp1tu0G4WifWoPu69uj58pv09PEoFa0WXD5svh/w3EMOh8/ptjI0cnvyiTOoz0U/Vi GwvvCBSTx6N3xQcTzSmuCv2CdwIK7gvHojiFFn2L+5jxGZMwcgyFpxwLO2CB19r0Z0M1 UmRA== ARC-Message-Signature: i=2; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=content-transfer-encoding:in-reply-to:from:references:cc:to :content-language:subject:user-agent:mime-version:list-unsubscribe :list-subscribe:list-id:precedence:date:message-id; bh=6Ri5Htyo8onlUXTPgZ8vUXRzKt6F3uO9aQ5gS/n71Dw=; fh=f460UUGzHj4An7vgY7LBk9qN8qFFbXoypJT8XNAUk3I=; b=obiwD1LQrbggZFIOQF9KGbb177I6okadtbPn/tI+BQdOOGUjAMf5O9HCFT6hmjj1nv Zd/8/XKPfPQXhcHFLhJGJV4+9Mvo5G1KiBBK/b364cGetVlDaOZECr/FzCuMjSJObqF/ 4T5VRVZ9ayBOEQ164mrlBhLmOl1Mi/if3Ib8i8Tx8qlHmgQo+U2YV51nDNi7w9RdJr+Y rfvDFFuB2HV+5ZKV1xHmorr3s1HtGcPSKyrUwdr1SqaL/0qcZO0y3LQLHBu8EApNnxDi ZtczsTR0c44R1DCDs+hBnRgETj2wCTjZfrLdlkxCz/tHglDxWoEnGiobyHUHf4sQmzq+ wmjQ==; dara=google.com ARC-Authentication-Results: i=2; mx.google.com; arc=pass (i=1 spf=pass spfdomain=arm.com dmarc=pass fromdomain=arm.com); spf=pass (google.com: domain of linux-kernel+bounces-90989-linux.lists.archive=gmail.com@vger.kernel.org designates 139.178.88.99 as permitted sender) smtp.mailfrom="linux-kernel+bounces-90989-linux.lists.archive=gmail.com@vger.kernel.org"; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=arm.com Return-Path: Received: from sv.mirrors.kernel.org (sv.mirrors.kernel.org. [139.178.88.99]) by mx.google.com with ESMTPS id n6-20020a634d46000000b005ce08c4bff5si8611141pgl.760.2024.03.04.09.11.36 for (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Mon, 04 Mar 2024 09:11:36 -0800 (PST) Received-SPF: pass (google.com: domain of linux-kernel+bounces-90989-linux.lists.archive=gmail.com@vger.kernel.org designates 139.178.88.99 as permitted sender) client-ip=139.178.88.99; Authentication-Results: mx.google.com; arc=pass (i=1 spf=pass spfdomain=arm.com dmarc=pass fromdomain=arm.com); spf=pass (google.com: domain of linux-kernel+bounces-90989-linux.lists.archive=gmail.com@vger.kernel.org designates 139.178.88.99 as permitted sender) smtp.mailfrom="linux-kernel+bounces-90989-linux.lists.archive=gmail.com@vger.kernel.org"; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=arm.com Received: from smtp.subspace.kernel.org (wormhole.subspace.kernel.org [52.25.139.140]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by sv.mirrors.kernel.org (Postfix) with ESMTPS id 26E5F281E53 for ; Mon, 4 Mar 2024 17:11:36 +0000 (UTC) Received: from localhost.localdomain (localhost.localdomain [127.0.0.1]) by smtp.subspace.kernel.org (Postfix) with ESMTP id 632536025A; Mon, 4 Mar 2024 17:11:30 +0000 (UTC) Received: from foss.arm.com (foss.arm.com [217.140.110.172]) by smtp.subspace.kernel.org (Postfix) with ESMTP id 345755D72B for ; Mon, 4 Mar 2024 17:11:27 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=none smtp.client-ip=217.140.110.172 ARC-Seal:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1709572289; cv=none; b=dGultzV3W4VsxgpT0sKyZC23bmnbm1yDZtVo8icGOxyiHROLxuG+bKJjBCegzqg8/72n2U4s1KTGXK4Uv1cGImDZJhpARvT1Yehxl9Tqf6c0ICKl2h4bDUsWDgnM4FKJcXRSkdaQGo6mASGu7HYNwoQFfIqdZ1/foWMSGFHOfsY= ARC-Message-Signature:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1709572289; c=relaxed/simple; bh=F0AhxNrMQYa399m9SObdQwh9VeOeMmvSZ2t1OUpqLuQ=; h=Message-ID:Date:MIME-Version:Subject:To:Cc:References:From: In-Reply-To:Content-Type; b=evVOUhVb1p7tqCBy2Zyl4rAdc+5YxbKSa9tHXCTrH73cBHTrn3z/2f5zNZkvZXwZL+TDWCg9iAf2BXyptiQN20USzobS0fHChx8yilAPVNJ9qo5BHUwKy/hvFAYT1gi9OOixLYew5sGLd9/gocB/xRyRZfzTp5+8wn5xA+9oDeE= ARC-Authentication-Results:i=1; smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=arm.com; spf=pass smtp.mailfrom=arm.com; arc=none smtp.client-ip=217.140.110.172 Authentication-Results: smtp.subspace.kernel.org; dmarc=pass (p=none dis=none) header.from=arm.com Authentication-Results: smtp.subspace.kernel.org; spf=pass smtp.mailfrom=arm.com Received: from usa-sjc-imap-foss1.foss.arm.com (unknown [10.121.207.14]) by usa-sjc-mx-foss1.foss.arm.com (Postfix) with ESMTP id 32C4C1FB; Mon, 4 Mar 2024 09:12:04 -0800 (PST) Received: from [10.1.196.40] (e121345-lin.cambridge.arm.com [10.1.196.40]) by usa-sjc-imap-foss1.foss.arm.com (Postfix) with ESMTPSA id B83A73F73F; Mon, 4 Mar 2024 09:11:25 -0800 (PST) Message-ID: Date: Mon, 4 Mar 2024 17:11:23 +0000 Precedence: bulk X-Mailing-List: linux-kernel@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 User-Agent: Mozilla Thunderbird Subject: Re: [PATCH v5 6/6] swiotlb: Remove pointless stride adjustment for allocations >= PAGE_SIZE Content-Language: en-GB To: Michael Kelley , =?UTF-8?B?UGV0ciBUZXNhxZnDrWs=?= Cc: Christoph Hellwig , Will Deacon , "linux-kernel@vger.kernel.org" , Petr Tesarik , "kernel-team@android.com" , "iommu@lists.linux.dev" , Marek Szyprowski , Dexuan Cui , Nicolin Chen References: <20240228133930.15400-1-will@kernel.org> <20240228133930.15400-7-will@kernel.org> <20240229133346.GA7177@lst.de> <20240229154756.GA10137@lst.de> <20240301163927.18358ee2@meshulam.tesarici.cz> <20240301180853.5ac20b27@meshulam.tesarici.cz> <8869c8b2-29c3-41e4-8f8a-5bcf9c0d22bb@arm.com> <20240301194212.3c64c9b2@meshulam.tesarici.cz> <20240304120055.56035c21@meshulam.tesarici.cz> <20240304165506.49e3b2d3@meshulam.tesarici.cz> From: Robin Murphy In-Reply-To: Content-Type: text/plain; charset=UTF-8; format=flowed Content-Transfer-Encoding: 8bit On 04/03/2024 4:04 pm, Michael Kelley wrote: > From: Petr Tesařík Sent: Monday, March 4, 2024 7:55 AM >> >> On Mon, 4 Mar 2024 13:37:56 +0000 >> Robin Murphy wrote: >> >>> On 04/03/2024 11:00 am, Petr Tesařík wrote: >>> [...] >>>>> Here's my take on tying all the threads together. There are >>>>> four alignment combinations: >>>>> >>>>> 1. alloc_align_mask: zero; min_align_mask: zero >>>>> 2. alloc_align_mask: zero; min_align_mask: non-zero >>>>> 3. alloc_align_mask: non-zero; min_align_mask: zero/ignored >>>>> 4. alloc_align_mask: non-zero; min_align_mask: non-zero >>>> >>>> What does "min_align_mask: zero/ignored" mean? Under which >>>> circumstances should be a non-zero min_align_mask ignored? > > "Ignored" was my short-hand for the swiotlb_alloc() case where > orig_addr is zero. Even if min_align_mask is set for the device, it > doesn't have any effect when orig_addr is zero. > >>>> >>>>> xen_swiotlb_map_page() and dma_direct_map_page() are #1 or #2 >>>>> via swiotlb_map() and swiotlb_tbl_map_single() >>>>> >>>>> iommu_dma_map_page() is #3 and #4 via swiotlb_tbl_map_single() >>>>> >>>>> swiotlb_alloc() is #3, directly to swiotlb_find_slots() >>>>> >>>>> For #1, the returned physical address has no constraints if >>>>> the requested size is less than a page. For page size or >>>>> greater, the discussed historical requirement for page >>>>> alignment applies. >>>>> >>>>> For #2, min_align_mask governs the bits of the returned >>>>> physical address that must match the original address. When >>>>> needed, swiotlb must also allocate pre-padding aligned to >>>>> IO_TLB_SIZE that precedes the returned physical address. A >>>>> request size <= swiotlb_max_mapping_size() will not exceed >>>>> IO_TLB_SEGSIZE even with the padding. The historical >>>>> requirement for page alignment does not apply because the >>>>> driver has explicitly used the newer min_align_mask feature. >>>> >>>> What is the idea here? Is it the assumption that only old drivers rely >>>> on page alignment, so if they use min_align_mask, it proves that they >>>> are new and must not rely on page alignment? >>> >>> Yes, if a driver goes out of its way to set a min_align_mask which is >>> smaller than its actual alignment constraint, that is clearly the >>> driver's own bug. Strictly we only need to be sympathetic to drivers >>> which predate min_align_mask, when implicitly relying on page alignment >>> was all they had. >>> >>>>> For #3, alloc_align_mask specifies the required alignment. No >>>>> pre-padding is needed. Per earlier comments from Robin[1], >>>>> it's reasonable to assume alloc_align_mask (i.e., the granule) >>>>> is >= IO_TLB_SIZE. The original address is not relevant in >>>>> determining the alignment, and the historical page alignment >>>>> requirement does not apply since alloc_align_mask explicitly >>>>> states the alignment. >>> >>> FWIW I'm also starting to wonder about getting rid of the alloc_size >>> argument and just have SWIOTLB round the end address up to >>> alloc_align_mask itself as part of all these calculations. Seems like it >>> could potentially end up a little simpler, maybe? > > Yes, I was thinking exactly this. But my reasoning was to solve the > bug in #4 that I previously pointed out. If iommu_dma_map_page() > does *not* do > > aligned_size = iova_align(iovad, size); > > but swiotlb_tbl_map_single() rounds up the size based on > alloc_align_mask *after* adding the offset modulo > min_align_mask, then the rounded-up size won't exceed IO_TLB_SIZE, > regardless of which bits are set in orig_addr. Ah, neat, I had a gut feeling that something like that might also fall out, I just didn't feel like working through the details to see if "simpler" could lead to "objectively better" :) I guess at worst we might also need to pass an alloc_align_mask to swiotlb_max_mapping_size() as well, but even that's not necessarily a bad thing if it keeps the equivalent calculations close together within SWIOTLB and makes things more robust overall. Cheers, Robin.