Received: by 2002:a25:e74b:0:0:0:0:0 with SMTP id e72csp1689823ybh; Thu, 16 Jul 2020 20:53:58 -0700 (PDT) X-Google-Smtp-Source: ABdhPJwvpy9UxrgztOi8rER5jUWL2ori4H40BRkFB3DIkadfbjfU+KTLVxk9KAzuewuhpsdSe/5J X-Received: by 2002:a17:906:1499:: with SMTP id x25mr6664465ejc.406.1594958038078; Thu, 16 Jul 2020 20:53:58 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1594958038; cv=none; d=google.com; s=arc-20160816; b=yn8Ki0Ns/XNc1BxnkOTxU1LaDuDgvawYopfo9UH9nBpXSeTdA8L1RquluHumG3B5FW YaT2iogx9R3f+03wE9XdzvhWo9rmn7xD4Ifg6xICrYgfGEBJiphq9F2BOr7cqjDrTH04 Jt7S0UpnqsY8TY/xciX8dpDD8AobQutEy4VRHcNbLzW1Eg1C8gZ6gBL0X+bqI4+bgtNU CSTduNiOYWu4fXXeGWjQiCYCdg5DkVQP8Fhbox+V2wufr3CQZuF70sFWgPLksyadsAc7 xK0PieBUeXqy4gAQXT26lRN3ycruXMrzFv04gsMK+41Ad9rl4I8kCSKN4Cv//FhSxPaX nLlA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:to:references:message-id :content-transfer-encoding:cc:date:in-reply-to:from:subject :mime-version; bh=NoxY5qgvHHP+6dWoWuNg90nhDn5uQrTG+AZWQOaMa88=; b=J6l66bu98eSZN3Ozhs80HfjbVvDg9S2Wo4ze4+GIvgxR1sFmBU5rjOtXy6tl9uZHMZ 0np8SN+mkxZeCY1gE0SlTmfgYqbYd2nVlZfnfUus/ajiXnBm77sEECp/7L0O6mQffhis nGt4+HU4iZ+3omh7/k72uSHMNaoqIm3UUPZkRxjcTPWlp7l88NU9d3QrCd88eWahod1h t9X23naWUZPdR6MkDflUr0eYO0NGFyXudXr0Q3rKm5xOuifFkkb3QHoyhaEcjXcQ7ViU EuLIv9kbUsYykazuJVMwaGXfN4AZ1fgekitTGEiPUYZ13jWKS0LoDJWZednieDUa4nY1 NE8Q== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=alibaba.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id y24si1811437ejj.675.2020.07.16.20.53.35; Thu, 16 Jul 2020 20:53:58 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=alibaba.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1726859AbgGQDw6 convert rfc822-to-8bit (ORCPT + 99 others); Thu, 16 Jul 2020 23:52:58 -0400 Received: from out30-42.freemail.mail.aliyun.com ([115.124.30.42]:44675 "EHLO out30-42.freemail.mail.aliyun.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726530AbgGQDw5 (ORCPT ); Thu, 16 Jul 2020 23:52:57 -0400 X-Alimail-AntiSpam: AC=PASS;BC=-1|-1;BR=01201311R141e4;CH=green;DM=||false|;DS=||;FP=0|-1|-1|-1|0|-1|-1|-1;HT=e01e01355;MF=teawaterz@linux.alibaba.com;NM=1;PH=DS;RN=11;SR=0;TI=SMTPD_---0U2yEEgS_1594957963; Received: from 127.0.0.1(mailfrom:teawaterz@linux.alibaba.com fp:SMTPD_---0U2yEEgS_1594957963) by smtp.aliyun-inc.com(127.0.0.1); Fri, 17 Jul 2020 11:52:53 +0800 Content-Type: text/plain; charset=utf-8 Mime-Version: 1.0 (Mac OS X Mail 13.4 \(3608.80.23.2.2\)) Subject: Re: [virtio-dev] [RFC for Linux v4 0/2] virtio_balloon: Add VIRTIO_BALLOON_F_CONT_PAGES to report continuous pages From: teawater In-Reply-To: <20200716064340-mutt-send-email-mst@kernel.org> Date: Fri, 17 Jul 2020 11:52:42 +0800 Cc: Hui Zhu , David Hildenbrand , Jason Wang , Andrew Morton , virtualization@lists.linux-foundation.org, linux-kernel@vger.kernel.org, linux-mm@kvack.org, qemu-devel@nongnu.org, virtio-dev@lists.oasis-open.org, Andrea Arcangeli Content-Transfer-Encoding: 8BIT Message-Id: References: <1594867315-8626-1-git-send-email-teawater@gmail.com> <20200716021929-mutt-send-email-mst@kernel.org> <744230FA-78D8-4568-8188-683087065E84@linux.alibaba.com> <20200716064340-mutt-send-email-mst@kernel.org> To: "Michael S. Tsirkin" X-Mailer: Apple Mail (2.3608.80.23.2.2) Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org > 2020年7月16日 18:45,Michael S. Tsirkin 写道: > > On Thu, Jul 16, 2020 at 03:01:18PM +0800, teawater wrote: >> >> >>> 2020年7月16日 14:38,Michael S. Tsirkin 写道: >>> >>> On Thu, Jul 16, 2020 at 10:41:50AM +0800, Hui Zhu wrote: >>>> The first, second and third version are in [1], [2] and [3]. >>>> Code of current version for Linux and qemu is available in [4] and [5]. >>>> Update of this version: >>>> 1. Report continuous pages will increase the speed. So added deflate >>>> continuous pages. >>>> 2. According to the comments from David in [6], added 2 new vqs inflate_cont_vq >>>> and deflate_cont_vq to report continuous pages with format 32 bits pfn and 32 >>>> bits size. >>>> Following is the introduction of the function. >>>> These patches add VIRTIO_BALLOON_F_CONT_PAGES to virtio_balloon. With this >>>> flag, balloon tries to use continuous pages to inflate and deflate. >>>> Opening this flag can bring two benefits: >>>> 1. Report continuous pages will increase memory report size of each time >>>> call tell_host. Then it will increase the speed of balloon inflate and >>>> deflate. >>>> 2. Host THPs will be splitted when qemu release the page of balloon inflate. >>>> Inflate balloon with continuous pages will let QEMU release the pages >>>> of same THPs. That will help decrease the splitted THPs number in >>>> the host. >>>> Following is an example in a VM with 1G memory 1CPU. This test setups an >>>> environment that has a lot of fragmentation pages. Then inflate balloon will >>>> split the THPs. >> >> >>>> // This is the THP number before VM execution in the host. >>>> // None use THP. >>>> cat /proc/meminfo | grep AnonHugePages: >>>> AnonHugePages: 0 kB >> These lines are from host. >> >>>> // After VM start, use usemem >>>> // (https://git.kernel.org/pub/scm/linux/kernel/git/wfg/vm-scalability.git) >>>> // punch-holes function generates 400m fragmentation pages in the guest >>>> // kernel. >>>> usemem --punch-holes -s -1 800m & >> These lines are from guest. They setups the environment that has a lot of fragmentation pages. >> >>>> // This is the THP number after this command in the host. >>>> // Some THP is used by VM because usemem will access 800M memory >>>> // in the guest. >>>> cat /proc/meminfo | grep AnonHugePages: >>>> AnonHugePages: 911360 kB >> These lines are from host. >> >>>> // Connect to the QEMU monitor, setup balloon, and set it size to 600M. >>>> (qemu) device_add virtio-balloon-pci,id=balloon1 >>>> (qemu) info balloon >>>> balloon: actual=1024 >>>> (qemu) balloon 600 >>>> (qemu) info balloon >>>> balloon: actual=600 >> These lines are from host. >> >>>> // This is the THP number after inflate the balloon in the host. >>>> cat /proc/meminfo | grep AnonHugePages: >>>> AnonHugePages: 88064 kB >> These lines are from host. >> >>>> // Set the size back to 1024M in the QEMU monitor. >>>> (qemu) balloon 1024 >>>> (qemu) info balloon >>>> balloon: actual=1024 >> These lines are from host. >> >>>> // Use usemem to increase the memory usage of QEMU. >>>> killall usemem >>>> usemem 800m >> These lines are from guest. >> >>>> // This is the THP number after this operation. >>>> cat /proc/meminfo | grep AnonHugePages: >>>> AnonHugePages: 65536 kB >> These lines are from host. >> >> >> >>>> >>>> Following example change to use continuous pages balloon. The number of >>>> splitted THPs is decreased. >>>> // This is the THP number before VM execution in the host. >>>> // None use THP. >>>> cat /proc/meminfo | grep AnonHugePages: >>>> AnonHugePages: 0 kB >> These lines are from host. >> >>>> // After VM start, use usemem punch-holes function generates 400M >>>> // fragmentation pages in the guest kernel. >>>> usemem --punch-holes -s -1 800m & >> These lines are from guest. They setups the environment that has a lot of fragmentation pages. >> >>>> // This is the THP number after this command in the host. >>>> // Some THP is used by VM because usemem will access 800M memory >>>> // in the guest. >>>> cat /proc/meminfo | grep AnonHugePages: >>>> AnonHugePages: 911360 kB >> These lines are from host. >> >>>> // Connect to the QEMU monitor, setup balloon, and set it size to 600M. >>>> (qemu) device_add virtio-balloon-pci,id=balloon1,cont-pages=on >>>> (qemu) info balloon >>>> balloon: actual=1024 >>>> (qemu) balloon 600 >>>> (qemu) info balloon >>>> balloon: actual=600 >> These lines are from host. >> >>>> // This is the THP number after inflate the balloon in the host. >>>> cat /proc/meminfo | grep AnonHugePages: >>>> AnonHugePages: 616448 kB >>>> // Set the size back to 1024M in the QEMU monitor. >>>> (qemu) balloon 1024 >>>> (qemu) info balloon >>>> balloon: actual=1024 >> These lines are from host. >> >>>> // Use usemem to increase the memory usage of QEMU. >>>> killall usemem >>>> usemem 800m >> These lines are from guest. >> >>>> // This is the THP number after this operation. >>>> cat /proc/meminfo | grep AnonHugePages: >>>> AnonHugePages: 907264 kB >> These lines are from host. >> >>> >>> I'm a bit confused about which of the above run within guest, >>> and which run within host. Could you explain pls? >>> >>> >> >> I added some introduction to show where these lines is get from. >> >> Best, >> Hui > > > OK so we see host has more free THPs. But guest has presumably less now - so > the total page table depth is the same. Did we gain anything? > cat /proc/meminfo | grep AnonHugePages: This command will output how many THPs is used by current system. There is no program using THPs except qemu. So this command will show how many THPs is used by qemu. The last outout of “cat /proc/meminfo | grep AnonHugePages:” show how many THPs is used by qemu when this 2 qemu’s anon page number is same. Without “cont-pages=on”, qemu keep 65536kb THPs. Wiht “cont-pages=on”, qemu keep 907264kb THPs. Keep more THPs will make memory access speed high. This is a test record use this 1G 1 cpu qemu after the fragmentation balloon test: Without “cont-pages=on”, qemu keep 81920kB THPs. / # usemem 800m 943718400 bytes / 489412 usecs = 1883076 KB/s 18725 usecs to free memory / # usemem 800m 943718400 bytes / 487070 usecs = 1892130 KB/s 18913 usecs to free memory / # usemem 800m 943718400 bytes / 484234 usecs = 1903212 KB/s 18538 usecs to free memory / # usemem 800m 943718400 bytes / 486568 usecs = 1894082 KB/s 18982 usecs to free memory With “cont-pages=on”, qemu keep 907264kb THPs. / # usemem 800m 943718400 bytes / 479098 usecs = 1923614 KB/s 18980 usecs to free memory / # usemem 800m 943718400 bytes / 477433 usecs = 1930323 KB/s 18562 usecs to free memory / # usemem 800m 943718400 bytes / 479790 usecs = 1920840 KB/s 18663 usecs to free memory / # usemem 800m 943718400 bytes / 480253 usecs = 1918988 KB/s 19011 usecs to free memory Best, Hui >> >>> >>>> [1] https://lkml.org/lkml/2020/3/12/144 >>>> [2] https://lore.kernel.org/linux-mm/1584893097-12317-1-git-send-email-teawater@gmail.com/ >>>> [3] https://lkml.org/lkml/2020/5/12/324 >>>> [4] https://github.com/teawater/linux/tree/balloon_conts >>>> [5] https://github.com/teawater/qemu/tree/balloon_conts >>>> [6] https://lkml.org/lkml/2020/5/13/1211 >>>> >>>> Hui Zhu (2): >>>> virtio_balloon: Add VIRTIO_BALLOON_F_CONT_PAGES and inflate_cont_vq >>>> virtio_balloon: Add deflate_cont_vq to deflate continuous pages >>>> >>>> drivers/virtio/virtio_balloon.c | 180 +++++++++++++++++++++++++++++++----- >>>> include/linux/balloon_compaction.h | 12 ++ >>>> include/uapi/linux/virtio_balloon.h | 1 >>>> mm/balloon_compaction.c | 117 +++++++++++++++++++++-- >>>> 4 files changed, 280 insertions(+), 30 deletions(-) >>> >>> >>> --------------------------------------------------------------------- >>> To unsubscribe, e-mail: virtio-dev-unsubscribe@lists.oasis-open.org >>> For additional commands, e-mail: virtio-dev-help@lists.oasis-open.org