Received: by 2002:a05:6a10:16a7:0:0:0:0 with SMTP id gp39csp765484pxb; Thu, 5 Nov 2020 12:22:58 -0800 (PST) X-Google-Smtp-Source: ABdhPJxxzP1mgoCeWerq1YOR1MdzgGtbDA9FDur0XZwc82uqmymD8sBxm6mPAuOuKrXw15XR+qN4 X-Received: by 2002:a17:906:134e:: with SMTP id x14mr4227454ejb.173.1604607778024; Thu, 05 Nov 2020 12:22:58 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1604607778; cv=none; d=google.com; s=arc-20160816; b=vh9Ohg6QSzEfi6GlLd9jNAGnpSIMj2X5nxL0YhcJ86JPMQPFa/Z1yUnD207mZhbest xOPslE1a1l45gTLgKM2j9pVtylYtFzJWsVDWMZo3rrkEEDmIKfiO5TLd/fy8+4l6DmhQ O/urotGzeNP7/PyHu8QolSzM3MRytEQY9MH6jL9AinrcD6ZLzA4EHGsRUTBcM+T7hl9V O9ZDOu3EGd/gMvLkCWLXEDC1nL5Dw2kzPHiu65ugQi+v4pGQIGeyALbzzw77S82RTnAf 9PrpsS4zLxmGNnU3NllbwhdnHBrfNnOYq8bOKOYIAUUA5ZRkFfeiB5dBR3LYLVmqj/CY tRbQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:user-agent:in-reply-to:content-disposition :mime-version:references:message-id:subject:cc:to:from:date :dkim-signature; bh=JfcwfHZZn7r5DpfNKWGdKjfc3q+dqsHz2Mg5eFwpXxw=; b=yMp0UnhnlmlSpGSPtwYhYZAnJdi91w0/mfvG3J/GJNMPZGNqGTUsBD3pnuVwVRjhg3 idMDKFgh3QRJ6R/X8erGrTUYRsVvLV/8l++Z3XQcKr4sKkcc2kWNdsrERhcgZCXOoX7Y UetGLHu+4esvyRntQ5lC/HliTsjaquPNDL47sdFduss5vvEfIuy9/qbHTXd9zXf0QTXt FQxiU8CjDdzcP+70+Z7C2QY8V/lAeCyGRJ7psKr7bU0KiQ8in3aMxBy/B3SQsTvLCFPC ogQ0g/wEzqlKY2JKFsU9yWOI8l5QeNgj92mGuCw9wpYnPbI2LBT9t4PzaF2MQ0Ytddt7 jsyQ== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@oracle.com header.s=corp-2020-01-29 header.b=CYr59WuZ; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=oracle.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id z17si1864464eja.372.2020.11.05.12.22.35; Thu, 05 Nov 2020 12:22:58 -0800 (PST) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=pass header.i=@oracle.com header.s=corp-2020-01-29 header.b=CYr59WuZ; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=oracle.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1731965AbgKEUU4 (ORCPT + 99 others); Thu, 5 Nov 2020 15:20:56 -0500 Received: from aserp2130.oracle.com ([141.146.126.79]:36234 "EHLO aserp2130.oracle.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726801AbgKEUUz (ORCPT ); Thu, 5 Nov 2020 15:20:55 -0500 Received: from pps.filterd (aserp2130.oracle.com [127.0.0.1]) by aserp2130.oracle.com (8.16.0.42/8.16.0.42) with SMTP id 0A5KK3XS125713; Thu, 5 Nov 2020 20:20:25 GMT DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=oracle.com; h=date : from : to : cc : subject : message-id : references : mime-version : content-type : in-reply-to; s=corp-2020-01-29; bh=JfcwfHZZn7r5DpfNKWGdKjfc3q+dqsHz2Mg5eFwpXxw=; b=CYr59WuZ0GgoLMPkpW1gsmXPeL6v8/t3I6Tjt5YpuUbzG4HYmxddg9mgIuMKMzBOCvjg H/6qtsP5O9rOPE3CjyO+hHAJttScfIN9l3zR0Es0hLdz+Xcw6l4o3xWOpFza+AwCta8g QJoRX+VKmRl6QMdlej38cxQ7F2m53i3QZvOy6sMaqJWbaoY+Z8Q/FStVCSK4MKfz9uK3 tV09aAcgRxdIOH+q6PQW36ZzXwiswNzd9yqD4YkC56hjtnXkOwtujljSenvRU7mSIxmt rdMvb8DPeLSKYzimow6KgNGsJ3zEp58rs9LfHwuVtX4ChiaEZXA/11s4JyYEUjuSUBAx MA== Received: from userp3030.oracle.com (userp3030.oracle.com [156.151.31.80]) by aserp2130.oracle.com with ESMTP id 34hhb2dxfv-1 (version=TLSv1.2 cipher=ECDHE-RSA-AES256-GCM-SHA384 bits=256 verify=FAIL); Thu, 05 Nov 2020 20:20:25 +0000 Received: from pps.filterd (userp3030.oracle.com [127.0.0.1]) by userp3030.oracle.com (8.16.0.42/8.16.0.42) with SMTP id 0A5JxpX1140140; Thu, 5 Nov 2020 20:18:25 GMT Received: from userv0122.oracle.com (userv0122.oracle.com [156.151.31.75]) by userp3030.oracle.com with ESMTP id 34hvs17tts-1 (version=TLSv1.2 cipher=ECDHE-RSA-AES256-GCM-SHA384 bits=256 verify=OK); Thu, 05 Nov 2020 20:18:24 +0000 Received: from abhmp0018.oracle.com (abhmp0018.oracle.com [141.146.116.24]) by userv0122.oracle.com (8.14.4/8.14.4) with ESMTP id 0A5KINwh026279; Thu, 5 Nov 2020 20:18:23 GMT Received: from char.us.oracle.com (/10.152.32.25) by default (Oracle Beehive Gateway v4.0) with ESMTP ; Thu, 05 Nov 2020 12:18:22 -0800 Received: by char.us.oracle.com (Postfix, from userid 1000) id CCF006A00F4; Thu, 5 Nov 2020 15:20:07 -0500 (EST) Date: Thu, 5 Nov 2020 15:20:07 -0500 From: Konrad Rzeszutek Wilk To: Ashish Kalra Cc: hch@lst.de, tglx@linutronix.de, mingo@redhat.com, bp@alien8.de, hpa@zytor.com, x86@kernel.org, luto@kernel.org, peterz@infradead.org, dave.hansen@linux-intel.com, iommu@lists.linux-foundation.org, linux-kernel@vger.kernel.org, brijesh.singh@amd.com, Thomas.Lendacky@amd.com, ssg.sos.patches@amd.com Subject: Re: [PATCH v3] swiotlb: Adjust SWIOTBL bounce buffer size for SEV guests. Message-ID: <20201105202007.GA6370@char.us.oracle.com> References: <20201104220804.21026-1-Ashish.Kalra@amd.com> <20201104221452.GA26079@char.us.oracle.com> <20201104223913.GA25311@ashkalra_ubuntu_server> <20201105174317.GA4294@char.us.oracle.com> <20201105184115.GA25261@ashkalra_ubuntu_server> <20201105190649.GB5366@char.us.oracle.com> <20201105193828.GA25303@ashkalra_ubuntu_server> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20201105193828.GA25303@ashkalra_ubuntu_server> User-Agent: Mutt/1.9.1 (2017-09-22) X-Proofpoint-Virus-Version: vendor=nai engine=6000 definitions=9796 signatures=668682 X-Proofpoint-Spam-Details: rule=notspam policy=default score=0 bulkscore=0 adultscore=0 mlxscore=0 malwarescore=0 mlxlogscore=999 suspectscore=0 spamscore=0 phishscore=0 classifier=spam adjust=0 reason=mlx scancount=1 engine=8.12.0-2009150000 definitions=main-2011050129 X-Proofpoint-Virus-Version: vendor=nai engine=6000 definitions=9796 signatures=668682 X-Proofpoint-Spam-Details: rule=notspam policy=default score=0 bulkscore=0 phishscore=0 suspectscore=0 clxscore=1015 mlxlogscore=999 impostorscore=0 malwarescore=0 lowpriorityscore=0 adultscore=0 spamscore=0 priorityscore=1501 mlxscore=0 classifier=spam adjust=0 reason=mlx scancount=1 engine=8.12.0-2009150000 definitions=main-2011050130 Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Thu, Nov 05, 2020 at 07:38:28PM +0000, Ashish Kalra wrote: > On Thu, Nov 05, 2020 at 02:06:49PM -0500, Konrad Rzeszutek Wilk wrote: > > . > > > > Right, so I am wondering if we can do this better. > > > > > > > > That is you are never going to get any 32-bit devices with SEV right? That > > > > is there is nothing that bounds you to always use the memory below 4GB? > > > > > > > > > > We do support 32-bit PCIe passthrough devices with SEV. > > > > Ewww.. Which devices would this be? > > That will be difficult to predict as customers could be doing > passthrough of all kinds of devices. But SEV is not on some 1990 hardware. It has PCIe, there is no PCI slots in there. Is it really possible to have a PCIe device that can't do more than 32-bit DMA? > > > > > > > Therefore, we can't just depend on >4G memory for SWIOTLB bounce buffering > > > when there is I/O pressure, because we do need to support device > > > passthrough of 32-bit devices. > > > > Presumarily there is just a handful of them? > > > Again, it will be incorrect to assume this. > > > > > > > Considering this, we believe that this patch needs to adjust/extend > > > boot-allocation of SWIOTLB and we want to keep it simple to do this > > > within a range detemined by amount of allocated guest memory. > > > > I would prefer to not have to revert this in a year as customers > > complain about "I paid $$$ and I am wasting half a gig on something > > I am not using" and giving customers knobs to tweak this instead of > > doing the right thing from the start. > > Currently, we face a lot of situations where we have to tell our > internal teams/external customers to explicitly increase SWIOTLB buffer > via the swiotlb parameter on the kernel command line, especially to > get better I/O performance numbers with SEV. Presumarily these are 64-bit? And what devices do you speak off that are actually affected by this performance? Increasing the SWIOTLB just means we have more memory, which in mind means you can have _more_ devices in the guest that won't handle the fact that DMA mapping returns an error. Not neccessarily that one device suddenly can go faster. > > So by having this SWIOTLB size adjustment done implicitly (even using a > static logic) is a great win-win situation. In other words, having even > a simple and static default increase of SWIOTLB buffer size for SEV is > really useful for us. > > We can always think of adding all kinds of heuristics to this, but that > just adds too much complexity without any predictable performance gain. > > And to add, the patch extends the SWIOTLB size as an architecture > specific callback, currently it is a simple and static logic for SEV/x86 > specific, but there is always an option to tweak/extend it with > additional logic in the future. Right, and that is what I would like to talk about as I think you are going to disappear (aka, busy with other stuff) after this patch goes in. I need to understand this more than "performance" and "internal teams" requirements to come up with a better way going forward as surely other platforms will hit the same issue anyhow. Lets break this down: How does the performance improve for one single device if you increase the SWIOTLB? Is there a specific device/driver that you can talk about that improve with this patch? > > Thanks, > Ashish' > > > > > That is the right thing being something less static. > > > > Can you work with me on what that could be please? > > > > > > > > Thanks, > > > Ashish > > > > > > > What I wonder is if we can combine the boot-allocation of the SWIOTLB > > > > with the post-boot-allocation of SWIOLTB to stitch together > > > > continous physical ranges. > > > > > > > > That way you have the flexibility at the start of using 64MB but if there > > > > is pressure, we grow to a bigger size? > > > > > > > > > > > > > > Thanks, > > > > > Ashish > > > > > > > > > > > > memory. > > > > > > > > > > > > > > Using late_initcall() interface to invoke > > > > > > > swiotlb_adjust() does not work as the size > > > > > > > adjustment needs to be done before mem_encrypt_init() > > > > > > > and reserve_crashkernel() which use the allocated > > > > > > > SWIOTLB buffer size, hence calling it explicitly > > > > > > > from setup_arch(). > > > > > > > > > > > > > > The SWIOTLB default size adjustment is added as an > > > > > > > architecture specific interface/callback to allow > > > > > > > architectures such as those supporting memory > > > > > > > encryption to adjust/expand SWIOTLB size for their > > > > > > > use. > > > > > > > > > > > > > > Signed-off-by: Ashish Kalra > > > > > > > --- > > > > > > > arch/x86/kernel/setup.c | 2 ++ > > > > > > > arch/x86/mm/mem_encrypt.c | 42 +++++++++++++++++++++++++++++++++++++++ > > > > > > > include/linux/swiotlb.h | 1 + > > > > > > > kernel/dma/swiotlb.c | 27 +++++++++++++++++++++++++ > > > > > > > 4 files changed, 72 insertions(+) > > > > > > > > > > > > > > diff --git a/arch/x86/kernel/setup.c b/arch/x86/kernel/setup.c > > > > > > > index 3511736fbc74..b073d58dd4a3 100644 > > > > > > > --- a/arch/x86/kernel/setup.c > > > > > > > +++ b/arch/x86/kernel/setup.c > > > > > > > @@ -1166,6 +1166,8 @@ void __init setup_arch(char **cmdline_p) > > > > > > > if (boot_cpu_has(X86_FEATURE_GBPAGES)) > > > > > > > hugetlb_cma_reserve(PUD_SHIFT - PAGE_SHIFT); > > > > > > > > > > > > > > + swiotlb_adjust(); > > > > > > > + > > > > > > > /* > > > > > > > * Reserve memory for crash kernel after SRAT is parsed so that it > > > > > > > * won't consume hotpluggable memory. > > > > > > > diff --git a/arch/x86/mm/mem_encrypt.c b/arch/x86/mm/mem_encrypt.c > > > > > > > index 3f248f0d0e07..e0deb157cddd 100644 > > > > > > > --- a/arch/x86/mm/mem_encrypt.c > > > > > > > +++ b/arch/x86/mm/mem_encrypt.c > > > > > > > @@ -489,7 +489,49 @@ static void print_mem_encrypt_feature_info(void) > > > > > > > pr_cont("\n"); > > > > > > > } > > > > > > > > > > > > > > +#define TOTAL_MEM_1G 0x40000000UL > > > > > > > +#define TOTAL_MEM_4G 0x100000000UL > > > > > > > + > > > > > > > +#define SIZE_128M (128UL<<20) > > > > > > > +#define SIZE_256M (256UL<<20) > > > > > > > +#define SIZE_512M (512UL<<20) > > > > > > > + > > > > > > > /* Architecture __weak replacement functions */ > > > > > > > +unsigned long __init arch_swiotlb_adjust(unsigned long iotlb_default_size) > > > > > > > +{ > > > > > > > + unsigned long size = 0; > > > > > > > + > > > > > > > + /* > > > > > > > + * For SEV, all DMA has to occur via shared/unencrypted pages. > > > > > > > + * SEV uses SWOTLB to make this happen without changing device > > > > > > > + * drivers. However, depending on the workload being run, the > > > > > > > + * default 64MB of SWIOTLB may not be enough & SWIOTLB may > > > > > > > + * run out of buffers for DMA, resulting in I/O errors and/or > > > > > > > + * performance degradation especially with high I/O workloads. > > > > > > > + * Increase the default size of SWIOTLB for SEV guests using > > > > > > > + * a minimum value of 128MB and a maximum value of 512MB, > > > > > > > + * depending on amount of provisioned guest memory. > > > > > > > + */ > > > > > > > + if (sev_active()) { > > > > > > > + phys_addr_t total_mem = memblock_phys_mem_size(); > > > > > > > + > > > > > > > + if (total_mem <= TOTAL_MEM_1G) > > > > > > > + size = clamp(iotlb_default_size * 2, SIZE_128M, > > > > > > > + SIZE_128M); > > > > > > > + else if (total_mem <= TOTAL_MEM_4G) > > > > > > > + size = clamp(iotlb_default_size * 4, SIZE_256M, > > > > > > > + SIZE_256M); > > > > > > > + else > > > > > > > + size = clamp(iotlb_default_size * 8, SIZE_512M, > > > > > > > + SIZE_512M); > > > > > > > + > > > > > > > + pr_info("SEV adjusted max SWIOTLB size = %luMB", > > > > > > > + size >> 20); > > > > > > > + } > > > > > > > + > > > > > > > + return size; > > > > > > > +} > > > > > > > + > > > > > > > void __init mem_encrypt_init(void) > > > > > > > { > > > > > > > if (!sme_me_mask) > > > > > > > diff --git a/include/linux/swiotlb.h b/include/linux/swiotlb.h > > > > > > > index 046bb94bd4d6..01ae6d891327 100644 > > > > > > > --- a/include/linux/swiotlb.h > > > > > > > +++ b/include/linux/swiotlb.h > > > > > > > @@ -33,6 +33,7 @@ extern void swiotlb_init(int verbose); > > > > > > > int swiotlb_init_with_tbl(char *tlb, unsigned long nslabs, int verbose); > > > > > > > extern unsigned long swiotlb_nr_tbl(void); > > > > > > > unsigned long swiotlb_size_or_default(void); > > > > > > > +extern void __init swiotlb_adjust(void); > > > > > > > extern int swiotlb_late_init_with_tbl(char *tlb, unsigned long nslabs); > > > > > > > extern void __init swiotlb_update_mem_attributes(void); > > > > > > > > > > > > > > diff --git a/kernel/dma/swiotlb.c b/kernel/dma/swiotlb.c > > > > > > > index c19379fabd20..66a9e627bb51 100644 > > > > > > > --- a/kernel/dma/swiotlb.c > > > > > > > +++ b/kernel/dma/swiotlb.c > > > > > > > @@ -163,6 +163,33 @@ unsigned long swiotlb_size_or_default(void) > > > > > > > return size ? size : (IO_TLB_DEFAULT_SIZE); > > > > > > > } > > > > > > > > > > > > > > +unsigned long __init __weak arch_swiotlb_adjust(unsigned long size) > > > > > > > +{ > > > > > > > + return 0; > > > > > > > +} > > > > > > > + > > > > > > > +void __init swiotlb_adjust(void) > > > > > > > +{ > > > > > > > + unsigned long size; > > > > > > > + > > > > > > > + /* > > > > > > > + * If swiotlb parameter has not been specified, give a chance to > > > > > > > + * architectures such as those supporting memory encryption to > > > > > > > + * adjust/expand SWIOTLB size for their use. > > > > > > > + */ > > > > > > > + if (!io_tlb_nslabs) { > > > > > > > + size = arch_swiotlb_adjust(IO_TLB_DEFAULT_SIZE); > > > > > > > + if (size) { > > > > > > > + size = ALIGN(size, 1 << IO_TLB_SHIFT); > > > > > > > + io_tlb_nslabs = size >> IO_TLB_SHIFT; > > > > > > > + io_tlb_nslabs = ALIGN(io_tlb_nslabs, IO_TLB_SEGSIZE); > > > > > > > + > > > > > > > + pr_info("architecture adjusted SWIOTLB slabs = %lu\n", > > > > > > > + io_tlb_nslabs); > > > > > > > + } > > > > > > > + } > > > > > > > +} > > > > > > > + > > > > > > > void swiotlb_print_info(void) > > > > > > > { > > > > > > > unsigned long bytes = io_tlb_nslabs << IO_TLB_SHIFT; > > > > > > > -- > > > > > > > 2.17.1 > > > > > > >