Received: by 2002:a05:6a10:1287:0:0:0:0 with SMTP id d7csp654532pxv; Thu, 22 Jul 2021 09:02:47 -0700 (PDT) X-Google-Smtp-Source: ABdhPJxPMKF2lJpKQjWnbLhnD0mF9IzJZXO8NOJswzhBLKbYwFQYr2T8/MFKRIPgAjOKGhZ0ZQXW X-Received: by 2002:a5d:8453:: with SMTP id w19mr329406ior.105.1626969767056; Thu, 22 Jul 2021 09:02:47 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1626969767; cv=none; d=google.com; s=arc-20160816; b=dAaNRIZiymSMSHbeK/ES4vYLBoyrptw8YkIR1RvpzHl8A6o1a/35f2gyuHDTgFqxgI 7QcbGlz1Nig2yo3dt9DX+8T2LOw2H59MGMlgR/hyZbUtiLPe3o7cwi6GzSUEOFWGt0ZO k/oe31k+/yS/xzIOp3KV6L9X3DnaAnIYNv8pKIMo+dXX9BYH+uLU6WPiBf1ZoJ9mjtHH kQgbJdta2TcpV0j6t6ggKe1M1x02/SDXHAeUCFFAJzVK6PeKVzkz7ZNcskYcExFQlY6L 6scNWHV0e/wjfLcgbaq73InsrMrM9QfZiv9LucmbIm6wMbxckXtLbPUFbROJfSjfZfwC W8AA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:cc:to:subject:message-id:date:from:in-reply-to :references:mime-version:dkim-signature; bh=2xUb5fAQ0tDlb4VRhzOtn9r/D+EbOQjl6xjOKizdBOw=; b=g2xZIvJD3P+Mu04Pa7sQnJdl+Km3oREbvzAihvCy9DC7y8wZmYwJDYSORmH7lx5+cT 44SkxW7gNcYycFecVXE/SZ729U/wo5W78g+Pu+ELfDjtbAgkNTwqI2fz071umQ7tFrTO QhjzjURRSGiNI4L3iwoZd19FAcGQjgYELgkYGPFgcbk2GrM2mZRxobdvU4feviYISgTw PsuckeZWLHCk/OcNjUHG8B6gdPCzi6Hbk+ni6KxN/0kobO0zvedO4E3xuq11JWQoK8bo vxQx0DUQptD85mOG587R18//n4EahiVInFkmkaZQRAuaXZpA9FNfCkp7JDoG7J6c+3t4 VA6A== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@google.com header.s=20161025 header.b=rUZJn2b3; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=REJECT sp=REJECT dis=NONE) header.from=google.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id v1si18890063ioj.78.2021.07.22.09.02.33; Thu, 22 Jul 2021 09:02:47 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=pass header.i=@google.com header.s=20161025 header.b=rUZJn2b3; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=REJECT sp=REJECT dis=NONE) header.from=google.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S229751AbhGVPUn (ORCPT + 99 others); Thu, 22 Jul 2021 11:20:43 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:38656 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S229660AbhGVPUm (ORCPT ); Thu, 22 Jul 2021 11:20:42 -0400 Received: from mail-ot1-x329.google.com (mail-ot1-x329.google.com [IPv6:2607:f8b0:4864:20::329]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 312FDC061575 for ; Thu, 22 Jul 2021 09:01:16 -0700 (PDT) Received: by mail-ot1-x329.google.com with SMTP id o72-20020a9d224e0000b02904bb9756274cso5731714ota.6 for ; Thu, 22 Jul 2021 09:01:16 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=20161025; h=mime-version:references:in-reply-to:from:date:message-id:subject:to :cc; bh=2xUb5fAQ0tDlb4VRhzOtn9r/D+EbOQjl6xjOKizdBOw=; b=rUZJn2b3p9GysD8IF02q/IQnRrGRErASUdGLvTdpex8dp40RIqR0ewQB6BcDB4Iif6 m8aYZLrgzuroF+Gftjziv1SuhG19GIpT1me5Mm6Yl7xJmHruEyS+Yyz75tBAm2LnCWiy iH7hg57gLQh7tfM4V24uTQ1kupC0BNs2D5X+fXJeqAkIi8QiKDngqpXmpcitB1untHZ/ CQkXzdAJ/FVgrkS94oRNrBAF+0bx+5+E9ZUe00HUSKQCM5WEFlboW8or9RALtl+EcjJ5 uzos3mUwgEkaOXJ1VwJTxtbfOAiQgd0iujYJaj9B1hoHdXyghNS4tCvSgSesTGzOziYP XF9A== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:mime-version:references:in-reply-to:from:date :message-id:subject:to:cc; bh=2xUb5fAQ0tDlb4VRhzOtn9r/D+EbOQjl6xjOKizdBOw=; b=YpfNUyvocYTMYQtpYr2Pw7C5tyE4FQlMbV+n2AQLHMFV67jgPCtHgiRVQegqlFj5k5 gdw1OPelY12vkfzlxUXDji3waYcc/+Mq3q9PXAPNRyGcbpVVfDoxIbQFMKEPfHf0iZZd Z7QN0zZBpdB1jdaqt7KGGnch4VGiAD5ylg+wLE/cFl9tnF8fWfHAcUdOabTKA0O4Re9G WTFXxFE5UkONJdqx2aCTF+uUuoaYtSQOYhcIHf9XFFfy/E8l1EzCnLKpDZDsJZ9qd4xF N06ADB1kIBEoa0GBTRBgOGSZQfTnqVqbNIYKHnAzPAF+yCEyR+a4DnqAp/OQBTjzY29D zLEA== X-Gm-Message-State: AOAM532miLDJGyvRfQwXGb4Ih3oqODug2pS7LjmiKGcFyGNjkIh222hk m1detuk5U2aLdViidMe/lj858zgeHlqtAAjJYN+/2QiQKSSeejPO X-Received: by 2002:a05:6830:1495:: with SMTP id s21mr302037otq.86.1626969673875; Thu, 22 Jul 2021 09:01:13 -0700 (PDT) MIME-Version: 1.0 References: <20210721131320.522061-1-dima@arista.com> In-Reply-To: From: Brian Geffon Date: Thu, 22 Jul 2021 12:00:36 -0400 Message-ID: Subject: Re: [PATCH v2] mm/mremap: Don't account pages in vma_to_resize() To: Dmitry Safonov Cc: LKML , Dmitry Safonov <0x7f454c46@gmail.com>, Alexander Viro , Andrew Morton , Andy Lutomirski , Catalin Marinas , Chen Wandun , Dan Carpenter , Dan Williams , Dave Jiang , Hugh Dickins , Ingo Molnar , Jason Gunthorpe , John Hubbard , Kefeng Wang , "Kirill A. Shutemov" , Mike Kravetz , Minchan Kim , Ralph Campbell , Russell King , Thomas Bogendoerfer , Thomas Gleixner , Vishal Verma , Vlastimil Babka , Wei Yongjun , Will Deacon Content-Type: text/plain; charset="UTF-8" Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Whoops I missed that you did remove it from vma_to_resize() so then it looks good to me. Acked-By: Brian Geffon On Thu, Jul 22, 2021 at 11:58 AM Brian Geffon wrote: > > On Wed, Jul 21, 2021 at 9:13 AM Dmitry Safonov wrote: > > > > All this vm_unacct_memory(charged) dance seems to complicate the life > > without a good reason. Furthermore, it seems not always done right on > > error-pathes in mremap_to(). > > And worse than that: this `charged' difference is sometimes > > double-accounted for growing MREMAP_DONTUNMAP mremap()s in move_vma(): > > : if (security_vm_enough_memory_mm(mm, new_len >> PAGE_SHIFT)) > > > > Let's not do this. > > Account memory in mremap() fast-path for growing VMAs or in move_vma() > > for actually moving things. The same simpler way as it's done by > > vm_stat_account(), but with a difference to call > > security_vm_enough_memory_mm() before copying/adjusting VMA. > > > > Originally noticed by Chen Wandun: > > https://lkml.kernel.org/r/20210717101942.120607-1-chenwandun@huawei.com > > > > Cc: Alexander Viro > > Cc: Andrew Morton > > Cc: Andy Lutomirski > > Cc: Brian Geffon > > Cc: Catalin Marinas > > Cc: Chen Wandun > > Cc: Dan Carpenter > > Cc: Dan Williams > > Cc: Dave Jiang > > Cc: Hugh Dickins > > Cc: Ingo Molnar > > Cc: Jason Gunthorpe > > Cc: John Hubbard > > Cc: Kefeng Wang > > Cc: "Kirill A. Shutemov" > > Cc: Mike Kravetz > > Cc: Minchan Kim > > Cc: Ralph Campbell > > Cc: Russell King > > Cc: Thomas Bogendoerfer > > Cc: Thomas Gleixner > > Cc: Vishal Verma > > Cc: Vlastimil Babka > > Cc: Wei Yongjun > > Cc: Will Deacon > > Fixes: e346b3813067 ("mm/mremap: add MREMAP_DONTUNMAP to mremap()") > > Signed-off-by: Dmitry Safonov > > --- > > mm/mremap.c | 50 ++++++++++++++++++++++---------------------------- > > 1 file changed, 22 insertions(+), 28 deletions(-) > > > > diff --git a/mm/mremap.c b/mm/mremap.c > > index 5989d3990020..b90c8e732e29 100644 > > --- a/mm/mremap.c > > +++ b/mm/mremap.c > > @@ -565,6 +565,7 @@ static unsigned long move_vma(struct vm_area_struct *vma, > > bool *locked, unsigned long flags, > > struct vm_userfaultfd_ctx *uf, struct list_head *uf_unmap) > > { > > + long to_account = new_len - old_len; > > struct mm_struct *mm = vma->vm_mm; > > struct vm_area_struct *new_vma; > > unsigned long vm_flags = vma->vm_flags; > > @@ -583,6 +584,9 @@ static unsigned long move_vma(struct vm_area_struct *vma, > > if (mm->map_count >= sysctl_max_map_count - 3) > > return -ENOMEM; > > > > + if (unlikely(flags & MREMAP_DONTUNMAP)) > > + to_account = new_len; > > + > > if (vma->vm_ops && vma->vm_ops->may_split) { > > if (vma->vm_start != old_addr) > > err = vma->vm_ops->may_split(vma, old_addr); > > @@ -604,8 +608,8 @@ static unsigned long move_vma(struct vm_area_struct *vma, > > if (err) > > return err; > > > > - if (unlikely(flags & MREMAP_DONTUNMAP && vm_flags & VM_ACCOUNT)) { > > - if (security_vm_enough_memory_mm(mm, new_len >> PAGE_SHIFT)) > > + if (vm_flags & VM_ACCOUNT) { > > + if (security_vm_enough_memory_mm(mm, to_account >> PAGE_SHIFT)) > > return -ENOMEM; > > } > > > > @@ -613,8 +617,8 @@ static unsigned long move_vma(struct vm_area_struct *vma, > > new_vma = copy_vma(&vma, new_addr, new_len, new_pgoff, > > &need_rmap_locks); > > if (!new_vma) { > > - if (unlikely(flags & MREMAP_DONTUNMAP && vm_flags & VM_ACCOUNT)) > > - vm_unacct_memory(new_len >> PAGE_SHIFT); > > + if (vm_flags & VM_ACCOUNT) > > + vm_unacct_memory(to_account >> PAGE_SHIFT); > > return -ENOMEM; > > } > > > > @@ -708,8 +712,7 @@ static unsigned long move_vma(struct vm_area_struct *vma, > > } > > > > static struct vm_area_struct *vma_to_resize(unsigned long addr, > > - unsigned long old_len, unsigned long new_len, unsigned long flags, > > - unsigned long *p) > > + unsigned long old_len, unsigned long new_len, unsigned long flags) > > { > > struct mm_struct *mm = current->mm; > > struct vm_area_struct *vma; > > @@ -768,13 +771,6 @@ static struct vm_area_struct *vma_to_resize(unsigned long addr, > > (new_len - old_len) >> PAGE_SHIFT)) > > return ERR_PTR(-ENOMEM); > > > > - if (vma->vm_flags & VM_ACCOUNT) { > > - unsigned long charged = (new_len - old_len) >> PAGE_SHIFT; > > - if (security_vm_enough_memory_mm(mm, charged)) > > - return ERR_PTR(-ENOMEM); > > - *p = charged; > > - } > > - > > return vma; > > } > > > > @@ -787,7 +783,6 @@ static unsigned long mremap_to(unsigned long addr, unsigned long old_len, > > struct mm_struct *mm = current->mm; > > struct vm_area_struct *vma; > > unsigned long ret = -EINVAL; > > - unsigned long charged = 0; > > unsigned long map_flags = 0; > > > > if (offset_in_page(new_addr)) > > @@ -830,7 +825,7 @@ static unsigned long mremap_to(unsigned long addr, unsigned long old_len, > > old_len = new_len; > > } > > > > - vma = vma_to_resize(addr, old_len, new_len, flags, &charged); > > + vma = vma_to_resize(addr, old_len, new_len, flags); > > if (IS_ERR(vma)) { > > ret = PTR_ERR(vma); > > goto out; > > @@ -853,7 +848,7 @@ static unsigned long mremap_to(unsigned long addr, unsigned long old_len, > > ((addr - vma->vm_start) >> PAGE_SHIFT), > > map_flags); > > if (IS_ERR_VALUE(ret)) > > - goto out1; > > + goto out; > > > > /* We got a new mapping */ > > if (!(flags & MREMAP_FIXED)) > > @@ -862,12 +857,6 @@ static unsigned long mremap_to(unsigned long addr, unsigned long old_len, > > ret = move_vma(vma, addr, old_len, new_len, new_addr, locked, flags, uf, > > uf_unmap); > > > > - if (!(offset_in_page(ret))) > > - goto out; > > - > > -out1: > > - vm_unacct_memory(charged); > > - > > out: > > return ret; > > } > > @@ -899,7 +888,6 @@ SYSCALL_DEFINE5(mremap, unsigned long, addr, unsigned long, old_len, > > struct mm_struct *mm = current->mm; > > struct vm_area_struct *vma; > > unsigned long ret = -EINVAL; > > - unsigned long charged = 0; > > bool locked = false; > > bool downgraded = false; > > struct vm_userfaultfd_ctx uf = NULL_VM_UFFD_CTX; > > @@ -981,7 +969,7 @@ SYSCALL_DEFINE5(mremap, unsigned long, addr, unsigned long, old_len, > > /* > > * Ok, we need to grow.. > > */ > > - vma = vma_to_resize(addr, old_len, new_len, flags, &charged); > > + vma = vma_to_resize(addr, old_len, new_len, flags); > > if (IS_ERR(vma)) { > > ret = PTR_ERR(vma); > > goto out; > > @@ -992,10 +980,18 @@ SYSCALL_DEFINE5(mremap, unsigned long, addr, unsigned long, old_len, > > if (old_len == vma->vm_end - addr) { > > /* can we just expand the current mapping? */ > > if (vma_expandable(vma, new_len - old_len)) { > > - int pages = (new_len - old_len) >> PAGE_SHIFT; > > + long pages = (new_len - old_len) >> PAGE_SHIFT; > > + > > + if (vma->vm_flags & VM_ACCOUNT) { > > + if (security_vm_enough_memory_mm(mm, pages)) { > > + ret = -ENOMEM; > > + goto out; > > + } > > + } > > Wasn't this whole check already performed in vma_to_resize()? > > > > > if (vma_adjust(vma, vma->vm_start, addr + new_len, > > vma->vm_pgoff, NULL)) { > > + vm_unacct_memory(pages); > > ret = -ENOMEM; > > goto out; > > } > > @@ -1034,10 +1030,8 @@ SYSCALL_DEFINE5(mremap, unsigned long, addr, unsigned long, old_len, > > &locked, flags, &uf, &uf_unmap); > > } > > out: > > - if (offset_in_page(ret)) { > > - vm_unacct_memory(charged); > > + if (offset_in_page(ret)) > > locked = false; > > - } > > if (downgraded) > > mmap_read_unlock(current->mm); > > else > > -- > > 2.32.0 > >