Received: by 2002:a05:6a10:a852:0:0:0:0 with SMTP id d18csp406641pxy; Wed, 5 May 2021 05:17:11 -0700 (PDT) X-Google-Smtp-Source: ABdhPJz20fTw0xSEuCFA9/gPgpP2YqfW8y7ruAxKT0Ykz/IpYQxM986XuVTHN/AnoDDri1asP3Pl X-Received: by 2002:a17:902:a509:b029:ea:db56:e98 with SMTP id s9-20020a170902a509b02900eadb560e98mr31571444plq.65.1620217031727; Wed, 05 May 2021 05:17:11 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1620217031; cv=none; d=google.com; s=arc-20160816; b=B2kLUHHyH8p0uP7mjRgkaqFfZ/ElZxwtiD7Ai/Jce00vxXhrRkxiBj9qWCTxlacAaC SHIIljDgnR+qjBHJRWM3yE+STNgCMoA/iYjBOJzmPngKv09OHopTiI1ihqjC0IltGJ+U aU9+Ny+p/0LU9YX9S1KamyObROldTf4OH1L6K+/6aCPJMX3I7ynNls76muN8X29IMOj/ Z9Cec4YlIu2XkfciSIIBTJt9W/RP3sx3SBU+6hIeW6RKjrpSs9mcmwuiZWmjOZoA1eo0 FSoUyL7qTS5CmcNV8GCSq6N6dPA5yE9bdNkrcipr+KmrtzkCTFI1TcB1lMwTC6HkzLaf KxYA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:mime-version :user-agent:references:in-reply-to:message-id:date:subject:cc:to :from:dkim-signature; bh=ja7qGpR3KW5GFRWP5PnOGmZcfL7My8ZY8zvAPIX6Op8=; b=lOLJZAhspeYB2Tnlay6ooRBIbwMOf3ktpguit6x0HRUymC0/4zJ0EWqgjfYQUhw6fC sl+HiNIHrJvfhWPRHTxtl9kmY/qzJjsJouK8yyI2o9I5VBFe9epjLTjaYXQQ5WnuzH9c efVttUbOto4UcrHowPEnRHzTpJ7KhU9HJuIdvf6ZCwPOjft58j6/U/CtfrwIISXjGe/7 FI9n/YlbN8oIQZtAZ0W0GugbUTuRMeDa3EECEcHlsWljSwQntxn1N2Y7Z30a07UcQpY7 3/q14JVTVe+KI1Ry1T0gpP4TK1hxq1ANkDznVwUQECUPlyeXk04qGq5cSu5Q5lppuEl/ tRuA== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@linuxfoundation.org header.s=korg header.b="QN/+k0Mk"; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=linuxfoundation.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id d9si7518179plo.15.2021.05.05.05.16.58; Wed, 05 May 2021 05:17:11 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=pass header.i=@linuxfoundation.org header.s=korg header.b="QN/+k0Mk"; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=linuxfoundation.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S234056AbhEEMOn (ORCPT + 99 others); Wed, 5 May 2021 08:14:43 -0400 Received: from mail.kernel.org ([198.145.29.99]:53564 "EHLO mail.kernel.org" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S232613AbhEEMKg (ORCPT ); Wed, 5 May 2021 08:10:36 -0400 Received: by mail.kernel.org (Postfix) with ESMTPSA id 45A02613FE; Wed, 5 May 2021 12:09:26 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=linuxfoundation.org; s=korg; t=1620216566; bh=zkXCsQDib+WMRl0XznmjUtqgFEU6P1lwPjR+kBC5vyI=; h=From:To:Cc:Subject:Date:In-Reply-To:References:From; b=QN/+k0MkrTwIRro6T0wR1nwrPxU/BBNNc51UU5Q1wxkVKQERAU5eW1cyb5+XCej+/ 8XSaFrKpe+/OKtOU8qEeHhdKC2e0EVHKofhfKYiv/3INN16PkM+pryDJbHSEjZrnBg Hvq/gxwtZQH8bQjXdyfSoBKH5MjK8NfA8JlmBq64= From: Greg Kroah-Hartman To: linux-kernel@vger.kernel.org Cc: Greg Kroah-Hartman , stable@vger.kernel.org, Christoph Hellwig , Jianxiong Gao , Konrad Rzeszutek Wilk Subject: [PATCH 5.11 20/31] swiotlb: respect min_align_mask Date: Wed, 5 May 2021 14:06:09 +0200 Message-Id: <20210505112327.328779098@linuxfoundation.org> X-Mailer: git-send-email 2.31.1 In-Reply-To: <20210505112326.672439569@linuxfoundation.org> References: <20210505112326.672439569@linuxfoundation.org> User-Agent: quilt/0.66 MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org From: Jianxiong Gao commit: 1f221a0d0dbf0e48ef3a9c62871281d6a7819f05 swiotlb: respect min_align_mask Respect the min_align_mask in struct device_dma_parameters in swiotlb. There are two parts to it: 1) for the lower bits of the alignment inside the io tlb slot, just extent the size of the allocation and leave the start of the slot empty 2) for the high bits ensure we find a slot that matches the high bits of the alignment to avoid wasting too much memory Based on an earlier patch from Jianxiong Gao . Signed-off-by: Christoph Hellwig Acked-by: Jianxiong Gao Tested-by: Jianxiong Gao Signed-off-by: Konrad Rzeszutek Wilk Signed-off-by: Jianxiong Gao Signed-off-by: Greg Kroah-Hartman --- kernel/dma/swiotlb.c | 41 +++++++++++++++++++++++++++++++---------- 1 file changed, 31 insertions(+), 10 deletions(-) --- a/kernel/dma/swiotlb.c +++ b/kernel/dma/swiotlb.c @@ -471,6 +471,14 @@ static void swiotlb_bounce(phys_addr_t o #define slot_addr(start, idx) ((start) + ((idx) << IO_TLB_SHIFT)) /* + * Return the offset into a iotlb slot required to keep the device happy. + */ +static unsigned int swiotlb_align_offset(struct device *dev, u64 addr) +{ + return addr & dma_get_min_align_mask(dev) & (IO_TLB_SIZE - 1); +} + +/* * Carefully handle integer overflow which can occur when boundary_mask == ~0UL. */ static inline unsigned long get_max_slots(unsigned long boundary_mask) @@ -491,24 +499,29 @@ static unsigned int wrap_index(unsigned * Find a suitable number of IO TLB entries size that will fit this request and * allocate a buffer from that IO TLB pool. */ -static int find_slots(struct device *dev, size_t alloc_size) +static int find_slots(struct device *dev, phys_addr_t orig_addr, + size_t alloc_size) { unsigned long boundary_mask = dma_get_seg_boundary(dev); dma_addr_t tbl_dma_addr = phys_to_dma_unencrypted(dev, io_tlb_start) & boundary_mask; unsigned long max_slots = get_max_slots(boundary_mask); - unsigned int nslots = nr_slots(alloc_size), stride = 1; + unsigned int iotlb_align_mask = + dma_get_min_align_mask(dev) & ~(IO_TLB_SIZE - 1); + unsigned int nslots = nr_slots(alloc_size), stride; unsigned int index, wrap, count = 0, i; unsigned long flags; BUG_ON(!nslots); /* - * For mappings greater than or equal to a page, we limit the stride - * (and hence alignment) to a page size. + * For mappings with an alignment requirement don't bother looping to + * unaligned slots once we found an aligned one. For allocations of + * PAGE_SIZE or larger only look for page aligned allocations. */ + stride = (iotlb_align_mask >> IO_TLB_SHIFT) + 1; if (alloc_size >= PAGE_SIZE) - stride <<= (PAGE_SHIFT - IO_TLB_SHIFT); + stride = max(stride, stride << (PAGE_SHIFT - IO_TLB_SHIFT)); spin_lock_irqsave(&io_tlb_lock, flags); if (unlikely(nslots > io_tlb_nslabs - io_tlb_used)) @@ -516,6 +529,12 @@ static int find_slots(struct device *dev index = wrap = wrap_index(ALIGN(io_tlb_index, stride)); do { + if ((slot_addr(tbl_dma_addr, index) & iotlb_align_mask) != + (orig_addr & iotlb_align_mask)) { + index = wrap_index(index + 1); + continue; + } + /* * If we find a slot that indicates we have 'nslots' number of * contiguous buffers, we allocate the buffers from that slot @@ -559,6 +578,7 @@ phys_addr_t swiotlb_tbl_map_single(struc size_t mapping_size, size_t alloc_size, enum dma_data_direction dir, unsigned long attrs) { + unsigned int offset = swiotlb_align_offset(dev, orig_addr); unsigned int index, i; phys_addr_t tlb_addr; @@ -574,7 +594,7 @@ phys_addr_t swiotlb_tbl_map_single(struc return (phys_addr_t)DMA_MAPPING_ERROR; } - index = find_slots(dev, alloc_size); + index = find_slots(dev, orig_addr, alloc_size + offset); if (index == -1) { if (!(attrs & DMA_ATTR_NO_WARN)) dev_warn_ratelimited(dev, @@ -588,10 +608,10 @@ phys_addr_t swiotlb_tbl_map_single(struc * This is needed when we sync the memory. Then we sync the buffer if * needed. */ - for (i = 0; i < nr_slots(alloc_size); i++) + for (i = 0; i < nr_slots(alloc_size + offset); i++) io_tlb_orig_addr[index + i] = slot_addr(orig_addr, i); - tlb_addr = slot_addr(io_tlb_start, index); + tlb_addr = slot_addr(io_tlb_start, index) + offset; if (!(attrs & DMA_ATTR_SKIP_CPU_SYNC) && (dir == DMA_TO_DEVICE || dir == DMA_BIDIRECTIONAL)) swiotlb_bounce(orig_addr, tlb_addr, mapping_size, DMA_TO_DEVICE); @@ -606,8 +626,9 @@ void swiotlb_tbl_unmap_single(struct dev enum dma_data_direction dir, unsigned long attrs) { unsigned long flags; - int i, count, nslots = nr_slots(alloc_size); - int index = (tlb_addr - io_tlb_start) >> IO_TLB_SHIFT; + unsigned int offset = swiotlb_align_offset(hwdev, tlb_addr); + int i, count, nslots = nr_slots(alloc_size + offset); + int index = (tlb_addr - offset - io_tlb_start) >> IO_TLB_SHIFT; phys_addr_t orig_addr = io_tlb_orig_addr[index]; /*