Received: by 2002:a25:86ce:0:0:0:0:0 with SMTP id y14csp737974ybm; Tue, 21 May 2019 02:42:23 -0700 (PDT) X-Google-Smtp-Source: APXvYqybQGmU8T8c1a2FYBkZxnsTNGR9m2r6RXSLVJT9Y79i3VE8U91t3t4jjAvB50R/QEPSV42+ X-Received: by 2002:a63:560d:: with SMTP id k13mr80581338pgb.124.1558431743701; Tue, 21 May 2019 02:42:23 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1558431743; cv=none; d=google.com; s=arc-20160816; b=Xz5GCsM5ElmDwJ0e5D4wVvyHPvlWnO8iJO2pA6fxH3bZKaIHrEFFPHcFCfHBioUsv6 HT4wLGXLrA1Wq4YVgbKEYsn5y472gVia43lE7rQzdXjcSkjpQlWpJY4uHtOa3PY/ger3 SRMEpSoJxUa56YlHn+/IH95Ns0qpo+redrbQdYQuRraMT72cuwKgcmsWkadjn8RHunHT rzt7hWoCwxufNVhp/HvpcZpBPd7mH4lYTH9HRO2h8dPxfqwirxa8eGdlNSviHT7mFuF3 /P7whdmbGHskb8fLEZaflfNOiQzVeXz4PIwGG4GzFN3CB5dPSI1gJ/4tT+Lw9ms3PjN4 cQXA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:references:in-reply-to:message-id:date :subject:cc:to:from; bh=+E9tZ5jftajU2k/zpNBz2bssx03Et5KHZum3c3z5GqI=; b=fjCglzaV4IHzLWbwH9fuooWnUul5DYyPD3wPzxANnbuM2D/VhV7W+i5FfTYJDIS59A lKtH1HoRR0v59ndisYylTlXk7T1bEWEAsoVMoDm1voPuebY4uZNJtG8CIpIRyiOxU3Vz G43IK6vpfkXNaysmZZ//onkRwNKEGVQGXljALhFoQQ1yINBy8ec5LYH+DZ/+crFGEqGE c/QGJlrGeXavTQXiaLK7WhC557uihNFGJz0g3d6qWUW6yuSTt0sb4RCtpWBZko0UQOrR LzdascA53Pt2eNLiOT/5yryajTQcuRWaWqE+AWOr6gGWigLKkPvtDStUqhqm/bctzbF9 XeDQ== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=alibaba.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id f1si21559063pln.23.2019.05.21.02.42.07; Tue, 21 May 2019 02:42:23 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=alibaba.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1727269AbfEUJlB (ORCPT + 99 others); Tue, 21 May 2019 05:41:01 -0400 Received: from out30-133.freemail.mail.aliyun.com ([115.124.30.133]:36783 "EHLO out30-133.freemail.mail.aliyun.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726006AbfEUJk7 (ORCPT ); Tue, 21 May 2019 05:40:59 -0400 X-Alimail-AntiSpam: AC=PASS;BC=-1|-1;BR=01201311R241e4;CH=green;DM=||false|;FP=0|-1|-1|-1|0|-1|-1|-1;HT=e01e04400;MF=yang.shi@linux.alibaba.com;NM=1;PH=DS;RN=12;SR=0;TI=SMTPD_---0TSIe59t_1558431642; Received: from e19h19392.et15sqa.tbsite.net(mailfrom:yang.shi@linux.alibaba.com fp:SMTPD_---0TSIe59t_1558431642) by smtp.aliyun-inc.com(127.0.0.1); Tue, 21 May 2019 17:40:55 +0800 From: Yang Shi To: ying.huang@intel.com, hannes@cmpxchg.org, mhocko@suse.com, mgorman@techsingularity.net, kirill.shutemov@linux.intel.com, josef@toxicpanda.com, hughd@google.com, shakeelb@google.com, akpm@linux-foundation.org Cc: yang.shi@linux.alibaba.com, linux-mm@kvack.org, linux-kernel@vger.kernel.org Subject: [v3 PATCH 2/2] mm: vmscan: correct some vmscan counters for THP swapout Date: Tue, 21 May 2019 17:40:42 +0800 Message-Id: <1558431642-52120-2-git-send-email-yang.shi@linux.alibaba.com> X-Mailer: git-send-email 1.8.3.1 In-Reply-To: <1558431642-52120-1-git-send-email-yang.shi@linux.alibaba.com> References: <1558431642-52120-1-git-send-email-yang.shi@linux.alibaba.com> Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Since commit bd4c82c22c36 ("mm, THP, swap: delay splitting THP after swapped out"), THP can be swapped out in a whole. But, nr_reclaimed and some other vm counters still get inc'ed by one even though a whole THP (512 pages) gets swapped out. This doesn't make too much sense to memory reclaim. For example, direct reclaim may just need reclaim SWAP_CLUSTER_MAX pages, reclaiming one THP could fulfill it. But, if nr_reclaimed is not increased correctly, direct reclaim may just waste time to reclaim more pages, SWAP_CLUSTER_MAX * 512 pages in worst case. And, it may cause pgsteal_{kswapd|direct} is greater than pgscan_{kswapd|direct}, like the below: pgsteal_kswapd 122933 pgsteal_direct 26600225 pgscan_kswapd 174153 pgscan_direct 14678312 nr_reclaimed and nr_scanned must be fixed in parallel otherwise it would break some page reclaim logic, e.g. vmpressure: this looks at the scanned/reclaimed ratio so it won't change semantics as long as scanned & reclaimed are fixed in parallel. compaction/reclaim: compaction wants a certain number of physical pages freed up before going back to compacting. kswapd priority raising: kswapd raises priority if we scan fewer pages than the reclaim target (which itself is obviously expressed in order-0 pages). As a result, kswapd can falsely raise its aggressiveness even when it's making great progress. Other than nr_scanned and nr_reclaimed, some other counters, e.g. pgactivate, nr_skipped, nr_ref_keep and nr_unmap_fail need to be fixed too since they are user visible via cgroup, /proc/vmstat or trace points, otherwise they would be underreported. When isolating pages from LRUs, nr_taken has been accounted in base page, but nr_scanned and nr_skipped are still accounted in THP. It doesn't make too much sense too since this may cause trace point underreport the numbers as well. So accounting those counters in base page instead of accounting THP as one page. This change may result in lower steal/scan ratio in some cases since THP may get split during page reclaim, then a part of tail pages get reclaimed instead of the whole 512 pages, but nr_scanned is accounted by 512, particularly for direct reclaim. But, this should be not a significant issue. Cc: "Huang, Ying" Cc: Johannes Weiner Cc: Michal Hocko Cc: Mel Gorman Cc: "Kirill A . Shutemov" Cc: Hugh Dickins Cc: Shakeel Butt Signed-off-by: Yang Shi --- v3: Removed Shakeel's Reviewed-by since the patch has been changed significantly Switched back to use compound_order per Matthew Fixed more counters per Johannes v2: Added Shakeel's Reviewed-by Use hpage_nr_pages instead of compound_order per Huang Ying and William Kucharski mm/vmscan.c | 40 ++++++++++++++++++++++++++++------------ 1 file changed, 28 insertions(+), 12 deletions(-) diff --git a/mm/vmscan.c b/mm/vmscan.c index b65bc50..1044834 100644 --- a/mm/vmscan.c +++ b/mm/vmscan.c @@ -1250,7 +1250,7 @@ static unsigned long shrink_page_list(struct list_head *page_list, case PAGEREF_ACTIVATE: goto activate_locked; case PAGEREF_KEEP: - stat->nr_ref_keep++; + stat->nr_ref_keep += (1 << compound_order(page)); goto keep_locked; case PAGEREF_RECLAIM: case PAGEREF_RECLAIM_CLEAN: @@ -1294,6 +1294,17 @@ static unsigned long shrink_page_list(struct list_head *page_list, goto activate_locked; } + /* + * Account all tail pages when THP is added + * into swap cache successfully. + * The head page has been accounted at the + * first place. + */ + if (PageTransHuge(page)) + sc->nr_scanned += + ((1 << compound_order(page)) - + 1); + may_enter_fs = 1; /* Adding to swap updated mapping */ @@ -1315,7 +1326,8 @@ static unsigned long shrink_page_list(struct list_head *page_list, if (unlikely(PageTransHuge(page))) flags |= TTU_SPLIT_HUGE_PMD; if (!try_to_unmap(page, flags)) { - stat->nr_unmap_fail++; + stat->nr_unmap_fail += + (1 << compound_order(page)); goto activate_locked; } } @@ -1442,7 +1454,11 @@ static unsigned long shrink_page_list(struct list_head *page_list, unlock_page(page); free_it: - nr_reclaimed++; + /* + * THP may get swapped out in a whole, need account + * all base pages. + */ + nr_reclaimed += (1 << compound_order(page)); /* * Is there need to periodically free_page_list? It would @@ -1464,7 +1480,6 @@ static unsigned long shrink_page_list(struct list_head *page_list, if (!PageMlocked(page)) { int type = page_is_file_cache(page); SetPageActive(page); - pgactivate++; stat->nr_activate[type] += hpage_nr_pages(page); count_memcg_page_event(page, PGACTIVATE); } @@ -1475,6 +1490,8 @@ static unsigned long shrink_page_list(struct list_head *page_list, VM_BUG_ON_PAGE(PageLRU(page) || PageUnevictable(page), page); } + pgactivate = stat->nr_activate[0] + stat->nr_activate[1]; + mem_cgroup_uncharge_list(&free_pages); try_to_unmap_flush(); free_unref_page_list(&free_pages); @@ -1642,14 +1659,12 @@ static unsigned long isolate_lru_pages(unsigned long nr_to_scan, unsigned long nr_zone_taken[MAX_NR_ZONES] = { 0 }; unsigned long nr_skipped[MAX_NR_ZONES] = { 0, }; unsigned long skipped = 0; - unsigned long scan, total_scan, nr_pages; + unsigned long scan, nr_pages; LIST_HEAD(pages_skipped); isolate_mode_t mode = (sc->may_unmap ? 0 : ISOLATE_UNMAPPED); scan = 0; - for (total_scan = 0; - scan < nr_to_scan && nr_taken < nr_to_scan && !list_empty(src); - total_scan++) { + while (scan < nr_to_scan && nr_taken < nr_to_scan && !list_empty(src)) { struct page *page; page = lru_to_page(src); @@ -1659,7 +1674,8 @@ static unsigned long isolate_lru_pages(unsigned long nr_to_scan, if (page_zonenum(page) > sc->reclaim_idx) { list_move(&page->lru, &pages_skipped); - nr_skipped[page_zonenum(page)]++; + nr_skipped[page_zonenum(page)] += + (1 << compound_order(page)); continue; } @@ -1669,7 +1685,7 @@ static unsigned long isolate_lru_pages(unsigned long nr_to_scan, * ineligible pages. This causes the VM to not reclaim any * pages, triggering a premature OOM. */ - scan++; + scan += (1 << compound_order(page)); switch (__isolate_lru_page(page, mode)) { case 0: nr_pages = hpage_nr_pages(page); @@ -1707,9 +1723,9 @@ static unsigned long isolate_lru_pages(unsigned long nr_to_scan, skipped += nr_skipped[zid]; } } - *nr_scanned = total_scan; + *nr_scanned = scan; trace_mm_vmscan_lru_isolate(sc->reclaim_idx, sc->order, nr_to_scan, - total_scan, skipped, nr_taken, mode, lru); + scan, skipped, nr_taken, mode, lru); update_lru_sizes(lruvec, lru, nr_zone_taken); return nr_taken; } -- 1.8.3.1