Received: by 2002:a25:1506:0:0:0:0:0 with SMTP id 6csp2463502ybv; Fri, 21 Feb 2020 16:25:19 -0800 (PST) X-Google-Smtp-Source: APXvYqxrv868sJenV66b46qjBCKP6Piczmzxd77E7Dr9aLCy1gyVWHl6JRI+R0VGy5F/BBHCJXnY X-Received: by 2002:a05:6830:154a:: with SMTP id l10mr30915056otp.44.1582331119585; Fri, 21 Feb 2020 16:25:19 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1582331119; cv=none; d=google.com; s=arc-20160816; b=qmp4Idv05yy1KNFDESy1HPhxgS6AWlyi0JLzy4GLHhHi1sCjYhIlOVEX1rIT0DgM9B Qt3lSxKPDwHX7gMVu6UcJfpoJ47fqPytd6GsjFIw2YRv3k0Q8bPAnOzdxCuQWOuFaEXF lv7IjIMb3ncCQLUqWVCFb3oWpO8iYRpapjQhaEuq9hMgs9XSKDoqeutn0QUma2tgU+Pi 1vnjKmQwjpZBNJdVnhDcwK0rSmPxnG66kMDL/1jrDCbZxrS/UIj4auVAcOncnEEvKCvu Fy0ogVUEMb6B2gMC/L9yv1/9CGWJKDQ6hwkLrdEFBV4qgKkjYbqQ99MFiH+pcpFfgDrK IY8w== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:cc:to:subject:message-id:date:from :in-reply-to:references:mime-version:dkim-signature; bh=8A1YE2lgpZQL8AL3OpYvUcklk0ocIiXfWyGf6voJc/k=; b=Ix2Z1hgSI83Xiu+h0ath/x8ZbIbHMVYsagRxhEbNqzMr7+RF+b2ZOyaP9Qp+63kosv WzMUhfGx3HlDogmKlquqiIKCVVRWC99N0USAE54kRcnApqg95H8ZO/n/+PqcL/MZ2gDN qBnbTNsitEVnDg9hvpY0KJxNNkDqSdHT9vyYiipQxeg7n0UjLTOg/NI4FLR0GvFSAKqM wAay3zwhQSkmOggiRtmCAgt0ylk0blAvsFE4F/Eri7d4RAad8H2XMG9EycOkOpLkuZ/o jyXr1N5rB6vxsMwUygMe7r/fLwUjKLx2m4S2nDplCpnpWRjJ026HSgx4PLrBEVpAKYpm TetQ== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@gmail.com header.s=20161025 header.b=A468YAmh; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=gmail.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id a6si1461585oia.33.2020.02.21.16.25.06; Fri, 21 Feb 2020 16:25:19 -0800 (PST) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; dkim=pass header.i=@gmail.com header.s=20161025 header.b=A468YAmh; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=gmail.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1726777AbgBVAYx (ORCPT + 99 others); Fri, 21 Feb 2020 19:24:53 -0500 Received: from mail-il1-f194.google.com ([209.85.166.194]:38616 "EHLO mail-il1-f194.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726045AbgBVAYx (ORCPT ); Fri, 21 Feb 2020 19:24:53 -0500 Received: by mail-il1-f194.google.com with SMTP id f5so3123520ilq.5 for ; Fri, 21 Feb 2020 16:24:52 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20161025; h=mime-version:references:in-reply-to:from:date:message-id:subject:to :cc; bh=8A1YE2lgpZQL8AL3OpYvUcklk0ocIiXfWyGf6voJc/k=; b=A468YAmhVxOg4ZCPKOO+HngxVCenqPG6tj44flBbVH8f0eC10zN6F7so5Iev26O7H3 Dcsyd/DKYvl/F4qOhbUwZDjGaZHo+I6NqfzROPeoW+GPOiXj/5HtYN6ANTurUUlsuJaN 8f0DGvEnng1nQwEWMBgCZAYHTG4b+WAwOJ5fgYtWTlP8+p9yMFax/5n8UAOOXpuytHyB 2hE4m8cteF5vy3AH2wICcLz8Mi8V2SgRiC4nKCHUD2Ne+Bs1bClF0dMrRD6rQIo48kBv VxPIWQ5NlFCu2saASVViOuZxbRZfzoDCCWgikDXNTqOVuvbkg1tkP6SF84FmR+cL7RXg ohFg== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:mime-version:references:in-reply-to:from:date :message-id:subject:to:cc; bh=8A1YE2lgpZQL8AL3OpYvUcklk0ocIiXfWyGf6voJc/k=; b=rDaa187VTMhWDkyBKRGgDrgA+wXMBRjJiDkX1hX6ZrqiFAGCXMhJNB17piY0Q9W0SG rZXVzStOpgjxOU+IUvIEfe6PVPDgvwdihYOt/voroTSYz595x0gEMi+DYMK0J9NpMza1 NgQnEVK41x1fC/7bLsLgttJiEb9ed7UlzhKHu39h+8g06JYHiZZ0n9IcsX9OEdrfEKPw owQ/LbpHDgMr7U9tbzKEd9drq4AucJLzSK38J2EjwOGewTBcNvWGaHXZ2Z3VsBXv9PQY xSTz+8WDoAnlESG/PBi4uJ/RxFW67P5EAUmolXERxFxW3ZdxvSguzH+AOzhEXu9ekE9h /cPQ== X-Gm-Message-State: APjAAAVI7m5By8OyJP9hL0y2vI58dOiCh4ei4Ra397Ypu0oev+H/5ypH QrV+uzZXIxJ/j3iq15n6QeNOWH4d0/qg7MVCsgs= X-Received: by 2002:a92:7402:: with SMTP id p2mr38368278ilc.64.1582331092334; Fri, 21 Feb 2020 16:24:52 -0800 (PST) MIME-Version: 1.0 References: <1575420174-19171-1-git-send-email-yang.shi@linux.alibaba.com> <9b8ff9ca-75b0-c256-cf37-885ccd786de7@linux.alibaba.com> In-Reply-To: <9b8ff9ca-75b0-c256-cf37-885ccd786de7@linux.alibaba.com> From: Alexander Duyck Date: Fri, 21 Feb 2020 16:24:41 -0800 Message-ID: Subject: Re: [v2 PATCH] mm: shmem: allow split THP when truncating THP partially To: Yang Shi Cc: "Michael S. Tsirkin" , David Hildenbrand , Hugh Dickins , "Kirill A. Shutemov" , Andrea Arcangeli , Andrew Morton , linux-mm , LKML Content-Type: text/plain; charset="UTF-8" Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Fri, Feb 21, 2020 at 10:24 AM Yang Shi wrote: > > > > On 2/20/20 10:16 AM, Alexander Duyck wrote: > > On Tue, Dec 3, 2019 at 4:43 PM Yang Shi wrote: > >> Currently when truncating shmem file, if the range is partial of THP > >> (start or end is in the middle of THP), the pages actually will just get > >> cleared rather than being freed unless the range cover the whole THP. > >> Even though all the subpages are truncated (randomly or sequentially), > >> the THP may still be kept in page cache. This might be fine for some > >> usecases which prefer preserving THP. > >> > >> But, when doing balloon inflation in QEMU, QEMU actually does hole punch > >> or MADV_DONTNEED in base page size granulairty if hugetlbfs is not used. > >> So, when using shmem THP as memory backend QEMU inflation actually doesn't > >> work as expected since it doesn't free memory. But, the inflation > >> usecase really needs get the memory freed. Anonymous THP will not get > >> freed right away too but it will be freed eventually when all subpages are > >> unmapped, but shmem THP would still stay in page cache. > >> > >> Split THP right away when doing partial hole punch, and if split fails > >> just clear the page so that read to the hole punched area would return > >> zero. > >> > >> Cc: Hugh Dickins > >> Cc: Kirill A. Shutemov > >> Cc: Andrea Arcangeli > >> Signed-off-by: Yang Shi > > One question I would have is if this is really the desired behavior we > > are looking for? > > > > By proactively splitting the THP you are likely going to see a > > performance regression with the virtio-balloon driver enabled in QEMU. > > I would suspect the response to that would be to update the QEMU code > > to identify the page size of the shared memory ramblock. At that > > point I suspect it would start behaving the same as how it currently > > handles anonymous memory, and the work done here would essentially > > have been wasted other than triggering the desire to resolve this in > > QEMU to avoid a performance regression. > > > > The code for inflating a the balloon in virtio-balloon in QEMU can be > > found here: > > https://github.com/qemu/qemu/blob/master/hw/virtio/virtio-balloon.c#L66 > > > > If there is a way for us to just populate the value obtained via > > qemu_ram_pagesize with the THP page size instead of leaving it at 4K, > > which is the size I am assuming it is at since you indicated that it > > is just freeing the base page size, then we could address the same > > issue and likely get the desired outcome of freeing the entire THP > > page when it is no longer used. > > If qemu could punch hole (this is how qemu free file-backed memory) in > THP unit, either w/ or w/o the patch the THP won't get split since the > whole THP will get truncated. But, if qemu has to free memory in sub-THP > size due to whatever reason (for example, 1MB for every 2MB section), > then we have to split THP otherwise no memory will be freed actually > with the current code. It is not about performance, it is about really > giving memory back to host. I get that, but at the same time I am not sure if everyone will be happy with the trade-off. That is my concern. You may want to change the patch description above if that is the case. Based on the description above it makes it sound as if the issue is that QEMU is using hole punch or MADV_DONTNEED with the wrong granularity. Based on your comment here it sounds like you want to have the ability to break up the larger THP page as soon as you want to push out a single 4K page from it. I am not sure the description for the behavior of anonymous THP with respect to QEMU makes sense either. Based on the description you made it sound like it was somehow using the same process used for huge pages. That isn't the case right? My understanding is that in the case of an anonymous THP it is getting broken into 4K subpages and then those are freed individually. That should leave you with the same performance regression that I had brought up earlier. So if anything it sounds like what you are really wanting is feature parity with anonymous THP which will split the anonymous THP page when a single 4K page within the THP is hit with MADV_DONTNEED. - Alex