Received: by 2002:a05:6a10:1d13:0:0:0:0 with SMTP id pp19csp1499999pxb; Tue, 17 Aug 2021 13:23:51 -0700 (PDT) X-Google-Smtp-Source: ABdhPJzGAU5BnZTJ4WU6lo8XI2+FY4AIBWhmOMR4/YKexnH58m1IWRmn8aLXM8SgwB/+GfUwBHJ5 X-Received: by 2002:a17:906:9747:: with SMTP id o7mr5850563ejy.486.1629231831608; Tue, 17 Aug 2021 13:23:51 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1629231831; cv=none; d=google.com; s=arc-20160816; b=RJGVPX8nknsdVqeaNihmNQ4P/J/ZiBbOKkhApvZbb2A0HXhzCeTfk7Brk55GM0fiLS jng6hj5wmh1pKnkI6GAU4je5bIoDDVP/dtdNWAglIeIsCTkWYGjSVK9/04w6hySqnq6/ NZTkZ4DUzTdzj/EQOMCwdaKrCLNs0lI1Pcw4ANcMI2oOi9V71UyMOxbtqyE7KjRPBvRO ZUWPC+ZPRX4B1QJAZOjdACTDR30Qzy1gqBbQoG7nzJfrH2ikhtYruznFHHr8Bnd7iG6n ZEDZBD947UDPU//IXETZJrQTgtWYveNF9NuApEsX5Y2nwjYZA4bye8fGpEnKmJulVNau Yk4w== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:cc:to:subject:message-id:date:from:in-reply-to :references:mime-version:dkim-signature; bh=zCBx9edfMztVack3J/dOyrn64HMiBBn0SZqe8UufBb4=; b=wfmOh7N2rKfhklw2W+HWXpK55Zuw3h6qUAqHko+LJiYihWkrGc9Yak0rGIZODBM++P G8eTZ3cT8ISCwGTD27DiCapVjt5RmXhmd/KfxlhksMIHDWiktZL2apkPvEWlf2CzML23 1pSUeZB+4FEGKtmtw09qTr1l/3ZrkhHLAsTAF/cfYhavgTXy3AJ+BYlYfy3/G11GUH/E blwUJiWi3qfzstxzF/hmcQw1Gfn5Fvm77nFjMSwHc4QmBef0Y6nh6iscZhHZ5rzaQK3N y9EIMuoM981FTfdgmhewQIOKutlDgJODiSnBqx8EjO/H08vjr5fCNDBdBLIQJ+YtvCYB Jcjw== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@gmail.com header.s=20161025 header.b=Vuhg74MR; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=gmail.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id bs18si3197013ejb.674.2021.08.17.13.23.26; Tue, 17 Aug 2021 13:23:51 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=pass header.i=@gmail.com header.s=20161025 header.b=Vuhg74MR; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=gmail.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S236372AbhHQUVd (ORCPT + 99 others); Tue, 17 Aug 2021 16:21:33 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:52700 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S234799AbhHQUVX (ORCPT ); Tue, 17 Aug 2021 16:21:23 -0400 Received: from mail-ej1-x629.google.com (mail-ej1-x629.google.com [IPv6:2a00:1450:4864:20::629]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 9910DC0698CF for ; Tue, 17 Aug 2021 13:16:25 -0700 (PDT) Received: by mail-ej1-x629.google.com with SMTP id w5so81199ejq.2 for ; Tue, 17 Aug 2021 13:16:25 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20161025; h=mime-version:references:in-reply-to:from:date:message-id:subject:to :cc; bh=zCBx9edfMztVack3J/dOyrn64HMiBBn0SZqe8UufBb4=; b=Vuhg74MRDEM2sRYPpMt7341EGU8CNxTrApQJ9QE+boi8CXhx+UechGUOOrQrg5cG6Q 9B/a8yZITa6CKN4WSbWMkP5fbmFDDRMzUNrYNHojGOde+mjgUdaxZ7uhpVhDCf+ySeGX QVr2mItMF8INxG6j7/SnOKj5LiT31CkZzb8bCPenN5LE+Ntgkwse8H9FN2fUuTsUbJb1 BKm9t1vHbzm6vm3lcFNM28MX04Z5l/Yf5g/DyyLhJTnCe9xxyhxvT2BG92SoKkgZaNyb BX+fxq9I29pOjhKbyQSpvdBmJKEQNX/KcXERhmOMX5epSSfYO+L+UyJU2/Udw/YNZuXJ Miew== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:mime-version:references:in-reply-to:from:date :message-id:subject:to:cc; bh=zCBx9edfMztVack3J/dOyrn64HMiBBn0SZqe8UufBb4=; b=PzIj/PPJMwVn/p1O97lgM5q5b1aJe7ILZz9yBXElZSfRkIbZJYdDR05l4mgT8PzpTP VFy7iGV2ulEujP9okIbquriq+z8WeCh5/7sD0b4MzETfzNTGIS8YqHXXw3sxY72wfab0 YoUaupWuHzqn8ssG6XeVlHaZOZ2PGGkGxmxEmu/KV97e+jnEVv4SfeaYIu+V9kpIIguD rayEX3emopS2lGEDPYZ+lK6Ju7KE479gcWdi8Ffv0i/nC/0Zl4i5Mrmkpn+GZIezLHmg QK7KxewGkV+D8zDTxBE0VYiCkXklP5tiHvTndoiI5VSKWqTWLhmknGuvy7SJNBCJ83Qi kWFw== X-Gm-Message-State: AOAM532YrxPl4hcXXqMuhXD+4WiBq1g2PqN9CMR4xOID+vPAoL06Ptzy aIsMjwG/nt4zX/seTVY/9w03+V+uPbPkM5ny88g= X-Received: by 2002:a17:907:76a1:: with SMTP id jw1mr5806404ejc.238.1629231384198; Tue, 17 Aug 2021 13:16:24 -0700 (PDT) MIME-Version: 1.0 References: <23a77889-2ddc-b030-75cd-44ca27fd4d1@google.com> In-Reply-To: <23a77889-2ddc-b030-75cd-44ca27fd4d1@google.com> From: Yang Shi Date: Tue, 17 Aug 2021 13:16:12 -0700 Message-ID: Subject: Re: [PATCH 7/9] huge tmpfs: shmem_is_huge(vma, inode, index) To: Hugh Dickins Cc: Andrew Morton , Shakeel Butt , "Kirill A. Shutemov" , Miaohe Lin , Mike Kravetz , Michal Hocko , Rik van Riel , Matthew Wilcox , Linux Kernel Mailing List , Linux MM Content-Type: text/plain; charset="UTF-8" Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Tue, Aug 17, 2021 at 1:19 AM Hugh Dickins wrote: > > Extend shmem_huge_enabled(vma) to shmem_is_huge(vma, inode, index), so > that a consistent set of checks can be applied, even when the inode is > accessed through read/write syscalls (with NULL vma) instead of mmaps > (the index argument is seldom of interest, but required by mount option > "huge=within_size"). Clean up and rearrange the checks a little. > > This then replaces the checks which shmem_fault() and shmem_getpage_gfp() > were making, and eliminates the SGP_HUGE and SGP_NOHUGE modes. > > Replace a couple of 0s by explicit SHMEM_HUGE_NEVERs; and replace the > obscure !shmem_mapping() symlink check by explicit S_ISLNK() - nothing > else needs that symlink check, so leave it there in shmem_getpage_gfp(). > > Signed-off-by: Hugh Dickins Reviewed-by: Yang Shi > --- > include/linux/shmem_fs.h | 9 +++-- > mm/shmem.c | 84 ++++++++++++---------------------------- > 2 files changed, 31 insertions(+), 62 deletions(-) > > diff --git a/include/linux/shmem_fs.h b/include/linux/shmem_fs.h > index 7d97b15a2f7a..60c6e4eac275 100644 > --- a/include/linux/shmem_fs.h > +++ b/include/linux/shmem_fs.h > @@ -86,7 +86,12 @@ extern void shmem_truncate_range(struct inode *inode, loff_t start, loff_t end); > extern int shmem_unuse(unsigned int type, bool frontswap, > unsigned long *fs_pages_to_unuse); > > -extern bool shmem_huge_enabled(struct vm_area_struct *vma); > +extern bool shmem_is_huge(struct vm_area_struct *vma, > + struct inode *inode, pgoff_t index); > +static inline bool shmem_huge_enabled(struct vm_area_struct *vma) > +{ > + return shmem_is_huge(vma, file_inode(vma->vm_file), vma->vm_pgoff); > +} > extern unsigned long shmem_swap_usage(struct vm_area_struct *vma); > extern unsigned long shmem_partial_swap_usage(struct address_space *mapping, > pgoff_t start, pgoff_t end); > @@ -96,8 +101,6 @@ enum sgp_type { > SGP_READ, /* don't exceed i_size, don't allocate page */ > SGP_NOALLOC, /* similar, but fail on hole or use fallocated page */ > SGP_CACHE, /* don't exceed i_size, may allocate page */ > - SGP_NOHUGE, /* like SGP_CACHE, but no huge pages */ > - SGP_HUGE, /* like SGP_CACHE, huge pages preferred */ > SGP_WRITE, /* may exceed i_size, may allocate !Uptodate page */ > SGP_FALLOC, /* like SGP_WRITE, but make existing page Uptodate */ > }; > diff --git a/mm/shmem.c b/mm/shmem.c > index 226ac3a911e9..56ee56b1cab6 100644 > --- a/mm/shmem.c > +++ b/mm/shmem.c > @@ -474,39 +474,35 @@ static bool shmem_confirm_swap(struct address_space *mapping, > #ifdef CONFIG_TRANSPARENT_HUGEPAGE > /* ifdef here to avoid bloating shmem.o when not necessary */ > > -static int shmem_huge __read_mostly; > +static int shmem_huge __read_mostly = SHMEM_HUGE_NEVER; > > -bool shmem_huge_enabled(struct vm_area_struct *vma) > +bool shmem_is_huge(struct vm_area_struct *vma, > + struct inode *inode, pgoff_t index) > { > - struct inode *inode = file_inode(vma->vm_file); > - struct shmem_sb_info *sbinfo = SHMEM_SB(inode->i_sb); > loff_t i_size; > - pgoff_t off; > > - if ((vma->vm_flags & VM_NOHUGEPAGE) || > - test_bit(MMF_DISABLE_THP, &vma->vm_mm->flags)) > - return false; > - if (shmem_huge == SHMEM_HUGE_FORCE) > - return true; > if (shmem_huge == SHMEM_HUGE_DENY) > return false; > - switch (sbinfo->huge) { > - case SHMEM_HUGE_NEVER: > + if (vma && ((vma->vm_flags & VM_NOHUGEPAGE) || > + test_bit(MMF_DISABLE_THP, &vma->vm_mm->flags))) > return false; > + if (shmem_huge == SHMEM_HUGE_FORCE) > + return true; > + > + switch (SHMEM_SB(inode->i_sb)->huge) { > case SHMEM_HUGE_ALWAYS: > return true; > case SHMEM_HUGE_WITHIN_SIZE: > - off = round_up(vma->vm_pgoff, HPAGE_PMD_NR); > + index = round_up(index, HPAGE_PMD_NR); > i_size = round_up(i_size_read(inode), PAGE_SIZE); > - if (i_size >= HPAGE_PMD_SIZE && > - i_size >> PAGE_SHIFT >= off) > + if (i_size >= HPAGE_PMD_SIZE && (i_size >> PAGE_SHIFT) >= index) > return true; > fallthrough; > case SHMEM_HUGE_ADVISE: > - /* TODO: implement fadvise() hints */ > - return (vma->vm_flags & VM_HUGEPAGE); > + if (vma && (vma->vm_flags & VM_HUGEPAGE)) > + return true; > + fallthrough; > default: > - VM_BUG_ON(1); > return false; > } > } > @@ -680,6 +676,12 @@ static long shmem_unused_huge_count(struct super_block *sb, > > #define shmem_huge SHMEM_HUGE_DENY > > +bool shmem_is_huge(struct vm_area_struct *vma, > + struct inode *inode, pgoff_t index) > +{ > + return false; > +} > + > static unsigned long shmem_unused_huge_shrink(struct shmem_sb_info *sbinfo, > struct shrink_control *sc, unsigned long nr_to_split) > { > @@ -1829,7 +1831,6 @@ static int shmem_getpage_gfp(struct inode *inode, pgoff_t index, > struct shmem_sb_info *sbinfo; > struct mm_struct *charge_mm; > struct page *page; > - enum sgp_type sgp_huge = sgp; > pgoff_t hindex = index; > gfp_t huge_gfp; > int error; > @@ -1838,8 +1839,6 @@ static int shmem_getpage_gfp(struct inode *inode, pgoff_t index, > > if (index > (MAX_LFS_FILESIZE >> PAGE_SHIFT)) > return -EFBIG; > - if (sgp == SGP_NOHUGE || sgp == SGP_HUGE) > - sgp = SGP_CACHE; > repeat: > if (sgp <= SGP_CACHE && > ((loff_t)index << PAGE_SHIFT) >= i_size_read(inode)) { > @@ -1903,36 +1902,12 @@ static int shmem_getpage_gfp(struct inode *inode, pgoff_t index, > return 0; > } > > - /* shmem_symlink() */ > - if (!shmem_mapping(mapping)) > - goto alloc_nohuge; > - if (shmem_huge == SHMEM_HUGE_DENY || sgp_huge == SGP_NOHUGE) > + /* Never use a huge page for shmem_symlink() */ > + if (S_ISLNK(inode->i_mode)) > goto alloc_nohuge; > - if (shmem_huge == SHMEM_HUGE_FORCE) > - goto alloc_huge; > - switch (sbinfo->huge) { > - case SHMEM_HUGE_NEVER: > + if (!shmem_is_huge(vma, inode, index)) > goto alloc_nohuge; > - case SHMEM_HUGE_WITHIN_SIZE: { > - loff_t i_size; > - pgoff_t off; > - > - off = round_up(index, HPAGE_PMD_NR); > - i_size = round_up(i_size_read(inode), PAGE_SIZE); > - if (i_size >= HPAGE_PMD_SIZE && > - i_size >> PAGE_SHIFT >= off) > - goto alloc_huge; > > - fallthrough; > - } > - case SHMEM_HUGE_ADVISE: > - if (sgp_huge == SGP_HUGE) > - goto alloc_huge; > - /* TODO: implement fadvise() hints */ > - goto alloc_nohuge; > - } > - > -alloc_huge: > huge_gfp = vma_thp_gfp_mask(vma); > huge_gfp = limit_gfp_mask(huge_gfp, gfp); > page = shmem_alloc_and_acct_page(huge_gfp, inode, index, true); > @@ -2088,7 +2063,6 @@ static vm_fault_t shmem_fault(struct vm_fault *vmf) > struct vm_area_struct *vma = vmf->vma; > struct inode *inode = file_inode(vma->vm_file); > gfp_t gfp = mapping_gfp_mask(inode->i_mapping); > - enum sgp_type sgp; > int err; > vm_fault_t ret = VM_FAULT_LOCKED; > > @@ -2151,15 +2125,7 @@ static vm_fault_t shmem_fault(struct vm_fault *vmf) > spin_unlock(&inode->i_lock); > } > > - sgp = SGP_CACHE; > - > - if ((vma->vm_flags & VM_NOHUGEPAGE) || > - test_bit(MMF_DISABLE_THP, &vma->vm_mm->flags)) > - sgp = SGP_NOHUGE; > - else if (vma->vm_flags & VM_HUGEPAGE) > - sgp = SGP_HUGE; > - > - err = shmem_getpage_gfp(inode, vmf->pgoff, &vmf->page, sgp, > + err = shmem_getpage_gfp(inode, vmf->pgoff, &vmf->page, SGP_CACHE, > gfp, vma, vmf, &ret); > if (err) > return vmf_error(err); > @@ -3966,7 +3932,7 @@ int __init shmem_init(void) > if (has_transparent_hugepage() && shmem_huge > SHMEM_HUGE_DENY) > SHMEM_SB(shm_mnt->mnt_sb)->huge = shmem_huge; > else > - shmem_huge = 0; /* just in case it was patched */ > + shmem_huge = SHMEM_HUGE_NEVER; /* just in case it was patched */ > #endif > return 0; > > -- > 2.26.2 >