Received: by 2002:a05:6a10:1287:0:0:0:0 with SMTP id d7csp652606pxv; Thu, 22 Jul 2021 09:00:44 -0700 (PDT) X-Google-Smtp-Source: ABdhPJzK3sLuhHBUOkSW0IP/5SGLAvp8M4Ej61SBXhLnugIlMPLWN0lv9Jina0F8YhRsVLxvREcE X-Received: by 2002:a92:d9c6:: with SMTP id n6mr376063ilq.142.1626969644624; Thu, 22 Jul 2021 09:00:44 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1626969644; cv=none; d=google.com; s=arc-20160816; b=u9V8D6esKrayuhCVX0zm/s9RItjc6j2jlb0V2g3VLZMyIX3ZO6r4wVnXzMvcb/Wbj9 QUrdr9KFrYmhVmk8yhscVkFi4Hlt2Eq+OU1pxgiNP1aTmQKU+jIo7AlTxMqsuHFbViKw 4FSl5IYnss5pBX4E7qqMLddZ3buF6rwO7lGP6J5tObQYskVn5xdAwyPcAaUdd3txSTxk gNO9ct3e/9fhjkSist8xFX0RqYJ+BOqZDHz38kCZFpSdI/HBKyc8oirZ4PbGnOLM5jKy p3W2dMwFiaKheRo/71YL3LP9GoPUszwSGJMyOm2wp6U7G41Pz4oPQIxbaGfnHwucgFQZ e8lQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:cc:to:subject:message-id:date:from:in-reply-to :references:mime-version:dkim-signature; bh=QFoiqBuewKwLijpRlkePQSSHPtlNuZZX3ekB9lmCZZI=; b=rpbzC2aQjNA5y/xzA5mwJM6Gdyly/angoTUZh2BCHVN47nV1ibYy2NXE+1+jCOWVHB sQv22X4anR1s8tGKKZEX9aTyzMVwCbb1mj+DCTiEMNwQ3TYLGVZCgi3AA+3cCyuCY0MC /MFPAqkSC6wnXJmxwe3L5BxQUf/ssYoXZAdisWB0KsmKawRyoOiGBF+8IPkB0yLtW3t0 kse4ZVwThKw9JwaBZqZO/YpWZda/0zleXY/cRc1E1vES2L19viNssI/DCNfs82k/GoiK SUA7FXxiDrnUtQqYEehAs3d1qJArh/UOn48453q0OgpA7DYRUhAJl65Cq1W/s5uo7CoI tW+g== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@google.com header.s=20161025 header.b=BpR31Ssu; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=REJECT sp=REJECT dis=NONE) header.from=google.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id i13si34695573iov.29.2021.07.22.09.00.31; Thu, 22 Jul 2021 09:00:44 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=pass header.i=@google.com header.s=20161025 header.b=BpR31Ssu; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=REJECT sp=REJECT dis=NONE) header.from=google.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S232421AbhGVPSp (ORCPT + 99 others); Thu, 22 Jul 2021 11:18:45 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:38218 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S230343AbhGVPSp (ORCPT ); Thu, 22 Jul 2021 11:18:45 -0400 Received: from mail-ot1-x32d.google.com (mail-ot1-x32d.google.com [IPv6:2607:f8b0:4864:20::32d]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id D8AA8C061575 for ; Thu, 22 Jul 2021 08:59:19 -0700 (PDT) Received: by mail-ot1-x32d.google.com with SMTP id o17-20020a9d76510000b02903eabfc221a9so5780717otl.0 for ; Thu, 22 Jul 2021 08:59:19 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=20161025; h=mime-version:references:in-reply-to:from:date:message-id:subject:to :cc; bh=QFoiqBuewKwLijpRlkePQSSHPtlNuZZX3ekB9lmCZZI=; b=BpR31SsuvCrbfJ6UE8IP8oUtk+Qz7dxKXqfvXMUkSKRkIjT1dE8pKOdTHpWw0fkdVy j7ulxIZZ5yjJDUhQ/GaezCTwgPEAEb/o5Oy3oTqM7RXw05DhLvqJwxfT4S3t5jHUiEUY W0W4QiUpIMwNzOf0tQlPujKwtwK+WjVgRu9HLQYTjtZw3MrGX7IP/FMecjzifKZ/Ws26 gybGSIn6daT1xw7tQDOECNLs8k1xhbLqeF3cTCaIL2fNymV6j+hYFtbIq/2u3E9bftXb jrd3H0t6nkJ2+QxX7mE9r0SIQLxL70VInTV8BPUoqb4Tyrfb7rb2NbtsYon0bz9yUce2 YGZA== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:mime-version:references:in-reply-to:from:date :message-id:subject:to:cc; bh=QFoiqBuewKwLijpRlkePQSSHPtlNuZZX3ekB9lmCZZI=; b=BbhAbIfgk8RJv+YrsX3WAOJeRMeus6fW58SLMrSWYCtcqEPfdFf5FUxzJbbw/l+v79 a5jIUd7fE0Cs5iCNL+Y9ztSem51xsfXFN451jB9piKSkR+oMZwEmfpBiCu7nKENLdpvj 4fgIPlzHT7KMWFDm1MTbRQ+J0ukdk4n47xcWV6WKYn9AqFQ5AdpmgtxcX4L3z6jYZTFx kUOw9TCP2mhwHlTVbi+A/EsObac2KuzpADMOSbcxO3PMZ/7P1w48Yt1L3Hhtp97jrGjJ WH9L5LLYxc5ZiZFse+9gzyWhc9NwHO7ryTqNkON5sA06Ij7wsmJweec/jKXOPBceboVz ynLQ== X-Gm-Message-State: AOAM531x8ImoIqMq7BcZeGPPJv+IwegwbqoK3pxiSZ0Fn5bRGIpghjB+ PGgnnhUpOO2bzdbn5nJWCuo480XzoM2kNoJG+GBk5A== X-Received: by 2002:a05:6830:1495:: with SMTP id s21mr296363otq.86.1626969558998; Thu, 22 Jul 2021 08:59:18 -0700 (PDT) MIME-Version: 1.0 References: <20210721131320.522061-1-dima@arista.com> In-Reply-To: <20210721131320.522061-1-dima@arista.com> From: Brian Geffon Date: Thu, 22 Jul 2021 11:58:42 -0400 Message-ID: Subject: Re: [PATCH v2] mm/mremap: Don't account pages in vma_to_resize() To: Dmitry Safonov Cc: LKML , Dmitry Safonov <0x7f454c46@gmail.com>, Alexander Viro , Andrew Morton , Andy Lutomirski , Catalin Marinas , Chen Wandun , Dan Carpenter , Dan Williams , Dave Jiang , Hugh Dickins , Ingo Molnar , Jason Gunthorpe , John Hubbard , Kefeng Wang , "Kirill A. Shutemov" , Mike Kravetz , Minchan Kim , Ralph Campbell , Russell King , Thomas Bogendoerfer , Thomas Gleixner , Vishal Verma , Vlastimil Babka , Wei Yongjun , Will Deacon Content-Type: text/plain; charset="UTF-8" Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Wed, Jul 21, 2021 at 9:13 AM Dmitry Safonov wrote: > > All this vm_unacct_memory(charged) dance seems to complicate the life > without a good reason. Furthermore, it seems not always done right on > error-pathes in mremap_to(). > And worse than that: this `charged' difference is sometimes > double-accounted for growing MREMAP_DONTUNMAP mremap()s in move_vma(): > : if (security_vm_enough_memory_mm(mm, new_len >> PAGE_SHIFT)) > > Let's not do this. > Account memory in mremap() fast-path for growing VMAs or in move_vma() > for actually moving things. The same simpler way as it's done by > vm_stat_account(), but with a difference to call > security_vm_enough_memory_mm() before copying/adjusting VMA. > > Originally noticed by Chen Wandun: > https://lkml.kernel.org/r/20210717101942.120607-1-chenwandun@huawei.com > > Cc: Alexander Viro > Cc: Andrew Morton > Cc: Andy Lutomirski > Cc: Brian Geffon > Cc: Catalin Marinas > Cc: Chen Wandun > Cc: Dan Carpenter > Cc: Dan Williams > Cc: Dave Jiang > Cc: Hugh Dickins > Cc: Ingo Molnar > Cc: Jason Gunthorpe > Cc: John Hubbard > Cc: Kefeng Wang > Cc: "Kirill A. Shutemov" > Cc: Mike Kravetz > Cc: Minchan Kim > Cc: Ralph Campbell > Cc: Russell King > Cc: Thomas Bogendoerfer > Cc: Thomas Gleixner > Cc: Vishal Verma > Cc: Vlastimil Babka > Cc: Wei Yongjun > Cc: Will Deacon > Fixes: e346b3813067 ("mm/mremap: add MREMAP_DONTUNMAP to mremap()") > Signed-off-by: Dmitry Safonov > --- > mm/mremap.c | 50 ++++++++++++++++++++++---------------------------- > 1 file changed, 22 insertions(+), 28 deletions(-) > > diff --git a/mm/mremap.c b/mm/mremap.c > index 5989d3990020..b90c8e732e29 100644 > --- a/mm/mremap.c > +++ b/mm/mremap.c > @@ -565,6 +565,7 @@ static unsigned long move_vma(struct vm_area_struct *vma, > bool *locked, unsigned long flags, > struct vm_userfaultfd_ctx *uf, struct list_head *uf_unmap) > { > + long to_account = new_len - old_len; > struct mm_struct *mm = vma->vm_mm; > struct vm_area_struct *new_vma; > unsigned long vm_flags = vma->vm_flags; > @@ -583,6 +584,9 @@ static unsigned long move_vma(struct vm_area_struct *vma, > if (mm->map_count >= sysctl_max_map_count - 3) > return -ENOMEM; > > + if (unlikely(flags & MREMAP_DONTUNMAP)) > + to_account = new_len; > + > if (vma->vm_ops && vma->vm_ops->may_split) { > if (vma->vm_start != old_addr) > err = vma->vm_ops->may_split(vma, old_addr); > @@ -604,8 +608,8 @@ static unsigned long move_vma(struct vm_area_struct *vma, > if (err) > return err; > > - if (unlikely(flags & MREMAP_DONTUNMAP && vm_flags & VM_ACCOUNT)) { > - if (security_vm_enough_memory_mm(mm, new_len >> PAGE_SHIFT)) > + if (vm_flags & VM_ACCOUNT) { > + if (security_vm_enough_memory_mm(mm, to_account >> PAGE_SHIFT)) > return -ENOMEM; > } > > @@ -613,8 +617,8 @@ static unsigned long move_vma(struct vm_area_struct *vma, > new_vma = copy_vma(&vma, new_addr, new_len, new_pgoff, > &need_rmap_locks); > if (!new_vma) { > - if (unlikely(flags & MREMAP_DONTUNMAP && vm_flags & VM_ACCOUNT)) > - vm_unacct_memory(new_len >> PAGE_SHIFT); > + if (vm_flags & VM_ACCOUNT) > + vm_unacct_memory(to_account >> PAGE_SHIFT); > return -ENOMEM; > } > > @@ -708,8 +712,7 @@ static unsigned long move_vma(struct vm_area_struct *vma, > } > > static struct vm_area_struct *vma_to_resize(unsigned long addr, > - unsigned long old_len, unsigned long new_len, unsigned long flags, > - unsigned long *p) > + unsigned long old_len, unsigned long new_len, unsigned long flags) > { > struct mm_struct *mm = current->mm; > struct vm_area_struct *vma; > @@ -768,13 +771,6 @@ static struct vm_area_struct *vma_to_resize(unsigned long addr, > (new_len - old_len) >> PAGE_SHIFT)) > return ERR_PTR(-ENOMEM); > > - if (vma->vm_flags & VM_ACCOUNT) { > - unsigned long charged = (new_len - old_len) >> PAGE_SHIFT; > - if (security_vm_enough_memory_mm(mm, charged)) > - return ERR_PTR(-ENOMEM); > - *p = charged; > - } > - > return vma; > } > > @@ -787,7 +783,6 @@ static unsigned long mremap_to(unsigned long addr, unsigned long old_len, > struct mm_struct *mm = current->mm; > struct vm_area_struct *vma; > unsigned long ret = -EINVAL; > - unsigned long charged = 0; > unsigned long map_flags = 0; > > if (offset_in_page(new_addr)) > @@ -830,7 +825,7 @@ static unsigned long mremap_to(unsigned long addr, unsigned long old_len, > old_len = new_len; > } > > - vma = vma_to_resize(addr, old_len, new_len, flags, &charged); > + vma = vma_to_resize(addr, old_len, new_len, flags); > if (IS_ERR(vma)) { > ret = PTR_ERR(vma); > goto out; > @@ -853,7 +848,7 @@ static unsigned long mremap_to(unsigned long addr, unsigned long old_len, > ((addr - vma->vm_start) >> PAGE_SHIFT), > map_flags); > if (IS_ERR_VALUE(ret)) > - goto out1; > + goto out; > > /* We got a new mapping */ > if (!(flags & MREMAP_FIXED)) > @@ -862,12 +857,6 @@ static unsigned long mremap_to(unsigned long addr, unsigned long old_len, > ret = move_vma(vma, addr, old_len, new_len, new_addr, locked, flags, uf, > uf_unmap); > > - if (!(offset_in_page(ret))) > - goto out; > - > -out1: > - vm_unacct_memory(charged); > - > out: > return ret; > } > @@ -899,7 +888,6 @@ SYSCALL_DEFINE5(mremap, unsigned long, addr, unsigned long, old_len, > struct mm_struct *mm = current->mm; > struct vm_area_struct *vma; > unsigned long ret = -EINVAL; > - unsigned long charged = 0; > bool locked = false; > bool downgraded = false; > struct vm_userfaultfd_ctx uf = NULL_VM_UFFD_CTX; > @@ -981,7 +969,7 @@ SYSCALL_DEFINE5(mremap, unsigned long, addr, unsigned long, old_len, > /* > * Ok, we need to grow.. > */ > - vma = vma_to_resize(addr, old_len, new_len, flags, &charged); > + vma = vma_to_resize(addr, old_len, new_len, flags); > if (IS_ERR(vma)) { > ret = PTR_ERR(vma); > goto out; > @@ -992,10 +980,18 @@ SYSCALL_DEFINE5(mremap, unsigned long, addr, unsigned long, old_len, > if (old_len == vma->vm_end - addr) { > /* can we just expand the current mapping? */ > if (vma_expandable(vma, new_len - old_len)) { > - int pages = (new_len - old_len) >> PAGE_SHIFT; > + long pages = (new_len - old_len) >> PAGE_SHIFT; > + > + if (vma->vm_flags & VM_ACCOUNT) { > + if (security_vm_enough_memory_mm(mm, pages)) { > + ret = -ENOMEM; > + goto out; > + } > + } Wasn't this whole check already performed in vma_to_resize()? > > if (vma_adjust(vma, vma->vm_start, addr + new_len, > vma->vm_pgoff, NULL)) { > + vm_unacct_memory(pages); > ret = -ENOMEM; > goto out; > } > @@ -1034,10 +1030,8 @@ SYSCALL_DEFINE5(mremap, unsigned long, addr, unsigned long, old_len, > &locked, flags, &uf, &uf_unmap); > } > out: > - if (offset_in_page(ret)) { > - vm_unacct_memory(charged); > + if (offset_in_page(ret)) > locked = false; > - } > if (downgraded) > mmap_read_unlock(current->mm); > else > -- > 2.32.0 >