Received: by 2002:a05:6a10:a0d1:0:0:0:0 with SMTP id j17csp1671623pxa; Thu, 20 Aug 2020 18:06:00 -0700 (PDT) X-Google-Smtp-Source: ABdhPJwsqykKQvsCwZ9bgsnIALp2F7V9JoVakh951HSQ/zmKdw7U/FV/YXHBx5GLH4Qk6UL+ucZ6 X-Received: by 2002:a17:906:c310:: with SMTP id s16mr599935ejz.466.1597971960120; Thu, 20 Aug 2020 18:06:00 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1597971960; cv=none; d=google.com; s=arc-20160816; b=jxGbNhtS/butI4eYv6PbnGi0kneaZdSVXCEV0oy7+nXgJvVHs3OfK54U02907CGlk/ H2Jp+ultTQ1rbEK20+Z1WxsDGdiXnP4KgKvEULUgS6rY6ZkWOCU7QnT/FMUI6+zrJ7aL DMaoI/TMPn8L1sGQ5GC1NlQn4z9OoNG2UkRQIEDi8ZvqMo5GKzBV+kwHBtelVwb48JJZ l43ylekeaCAwTVU169fYCaJxg1xWf4Mi77fX8KkWlBErN2+iv+zXSIzAl8CAMpTRagC8 5SSr8S2/eyCxhT1q9UcHYb24I0qstTBV+0M0PJR1/VB78aLWXovgOeXfsYbqYYPJlJm2 MS8w== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:in-reply-to:content-disposition :mime-version:references:message-id:subject:cc:to:from:date; bh=10EVGpbyv4bRBsp0oH1hQaDm+6st4Pj/EQTrxHYSo1s=; b=JWIt3H8PxJUq7y4kOHY/mWdATWAj0MS3ffAc611jipci6nNNR70/CGUHx6c7I829Nh qsQeyxnGL8bGD2npDZhDaaSWVEORo9U1hL8VHyrF/M3EoXcIvAZla7TeJEqRByEHdhsB VHMU0z5TWbuWjPACDYx1nN20gIuRvdOP/8m7bNYF5e2kCtoQuHP3f9PBmVYLHMwxC2jI yOXFmVxJ5cywCalxOe3BMif9htTjsmuih1oPEAiIMGgf5/+ZGtb6lubH/8ofqcUENpO9 1SeG5FRk5nrwlHQf8SVbiOZums0N71JNleoQPb9sWHHgcTFuH6+2EmL1S1z6Bi1OZmGB tygw== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id re10si156810ejb.140.2020.08.20.18.05.36; Thu, 20 Aug 2020 18:06:00 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1726897AbgHUBDt (ORCPT + 99 others); Thu, 20 Aug 2020 21:03:49 -0400 Received: from mail104.syd.optusnet.com.au ([211.29.132.246]:58194 "EHLO mail104.syd.optusnet.com.au" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726309AbgHUBDp (ORCPT ); Thu, 20 Aug 2020 21:03:45 -0400 Received: from dread.disaster.area (pa49-181-146-199.pa.nsw.optusnet.com.au [49.181.146.199]) by mail104.syd.optusnet.com.au (Postfix) with ESMTPS id AF65F823059; Fri, 21 Aug 2020 11:03:31 +1000 (AEST) Received: from dave by dread.disaster.area with local (Exim 4.92.3) (envelope-from ) id 1k8vSs-0007mS-Td; Fri, 21 Aug 2020 11:03:30 +1000 Date: Fri, 21 Aug 2020 11:03:30 +1000 From: Dave Chinner To: Gao Xiang Cc: Andrew Morton , linux-mm@kvack.org, linux-kernel@vger.kernel.org, Carlos Maiolino , Eric Sandeen , "Huang, Ying" , Yang Shi , Rafael Aquini , stable Subject: Re: [PATCH v2] mm, THP, swap: fix allocating cluster for swapfile by mistake Message-ID: <20200821010330.GC7728@dread.disaster.area> References: <20200820045323.7809-1-hsiangkao@redhat.com> <20200820233446.GB7728@dread.disaster.area> <20200821002145.GA28298@xiangao.remote.csb> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20200821002145.GA28298@xiangao.remote.csb> X-Optus-CM-Score: 0 X-Optus-CM-Analysis: v=2.3 cv=QIgWuTDL c=1 sm=1 tr=0 cx=a_idp_d a=GorAHYkI+xOargNMzM6qxQ==:117 a=GorAHYkI+xOargNMzM6qxQ==:17 a=kj9zAlcOel0A:10 a=y4yBn9ojGxQA:10 a=7-415B0cAAAA:8 a=6Q28Y-3DynG6Ewtw8AoA:9 a=CjuIK1q_8ugA:10 a=biEYGPWJfzWAr4FL6Ov7:22 Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Fri, Aug 21, 2020 at 08:21:45AM +0800, Gao Xiang wrote: > Hi Dave, > > On Fri, Aug 21, 2020 at 09:34:46AM +1000, Dave Chinner wrote: > > On Thu, Aug 20, 2020 at 12:53:23PM +0800, Gao Xiang wrote: > > > SWP_FS is used to make swap_{read,write}page() go through > > > the filesystem, and it's only used for swap files over > > > NFS. So, !SWP_FS means non NFS for now, it could be either > > > file backed or device backed. Something similar goes with > > > legacy SWP_FILE. > > > > > > So in order to achieve the goal of the original patch, > > > SWP_BLKDEV should be used instead. > > > > > > FS corruption can be observed with SSD device + XFS + > > > fragmented swapfile due to CONFIG_THP_SWAP=y. > > > > > > I reproduced the issue with the following details: > > > > > > Environment: > > > QEMU + upstream kernel + buildroot + NVMe (2 GB) > > > > > > Kernel config: > > > CONFIG_BLK_DEV_NVME=y > > > CONFIG_THP_SWAP=y > > > > Ok, so at it's core this is a swap file extent versus THP swap > > cluster alignment issue? > > I think yes. > > > > > > diff --git a/mm/swapfile.c b/mm/swapfile.c > > > index 6c26916e95fd..2937daf3ca02 100644 > > > --- a/mm/swapfile.c > > > +++ b/mm/swapfile.c > > > @@ -1074,7 +1074,7 @@ int get_swap_pages(int n_goal, swp_entry_t swp_entries[], int entry_size) > > > goto nextsi; > > > } > > > if (size == SWAPFILE_CLUSTER) { > > > - if (!(si->flags & SWP_FS)) > > > + if (si->flags & SWP_BLKDEV) > > > n_ret = swap_alloc_cluster(si, swp_entries); > > > } else > > > n_ret = scan_swap_map_slots(si, SWAP_HAS_CACHE, > > > > IOWs, if you don't make this change, does the corruption problem go > > away if you align swap extents in iomap_swapfile_add_extent() to > > (SWAPFILE_CLUSTER * PAGE_SIZE) instead of just PAGE_SIZE? > > > > I.e. if the swapfile extents are aligned correctly to huge page swap > > cluster size and alignment, does the swap clustering optimisations > > for swapping THP pages work correctly? And, if so, is there any > > performance benefit we get from enabling proper THP swap clustering > > on swapfiles? > > > > Yeah, I once think about some similiar thing as well. My thought for now is > > - First, SWAP THP doesn't claim to support such swapfile for now. > And the original author tried to explicitly avoid the whole thing in > > f0eea189e8e9 ("mm, THP, swap: Don't allocate huge cluster for file backed swap device") > > So such thing would be considered as some new feature and need > more testing at least. But for now I think we just need a quick > fix to fix the commit f0eea189e8e9 to avoid regression and for > backport use. Sure, a quick fix is fine for the current issue. I'm asking questions about the design/architecture of how THP_SWAP is supposed to work and whether swapfiles are violating some other undocumented assumption about swapping THP files... > - It is hard for users to control swapfile in > SWAPFILE_CLUSTER * PAGE_SIZE extents, especially users' > disk are fragmented or have some on-disk metadata limitation or > something. It's very hard for users to utilize this and arrange > their swapfile physical addr alignment and fragments for now. This isn't something users control. The swapfile extent mapping code rounds the swap extents inwards so that the parts of the on-disk extents that are not aligned or cannot hold a full page are omitted from the ranges of the file that can be swapped to. i.e. a file that extents aligned to 4kB is fine for a 4KB page size machine, but needs additional alignment to allow swapping to work on a 64kB page size machine. Hence the swap code rounds the file extents inwards to PAGE_SIZE to align them correctly. We really should be doing this for THP page size rather than PAGE_SIZE if THP_SWAP is enabled, regardless of whether swap clustering is enabled or not... Cheers, Dave. -- Dave Chinner david@fromorbit.com