Received: by 2002:a05:7412:a9a2:b0:e2:908c:2ebd with SMTP id o34csp2463243rdh; Sun, 29 Oct 2023 18:14:07 -0700 (PDT) X-Google-Smtp-Source: AGHT+IFG3m8IPL284HF53Q6pNV8gZCUxJTMGlpmJeq/285nXx5tLzX5q37IhkkP1T82L1wZuhzxR X-Received: by 2002:a05:6a00:2d82:b0:6b2:6835:2a7f with SMTP id fb2-20020a056a002d8200b006b268352a7fmr10095757pfb.22.1698628447561; Sun, 29 Oct 2023 18:14:07 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1698628447; cv=none; d=google.com; s=arc-20160816; b=YCZ+DRuCAefjuM/TH0eb7E7JGy7QCwZHj2a2VtMu5+6oGahFMTlBcWwCgquTPCB/au baUxpNt20pQ2Lbfw9UnLKm/zgFhGTumEqa2nbmbzr8I7C9T7bqOt7bG5N34gTOfcE7Ur l6RvxnnMjyftBDzHY0fw0Cw2YtGoTUN6EbvBSxTZgQjIkK/CaDT8Z2YxjTWudUxTAp9x UCUbrL/vhA0muamsRA7qHbcWeUBm+BmVmI6WQrBi8xNp1vSvIHcdgKWZHutnc04Kf6mk LH3Poo5Mgna+OAsN5qsaFZmW/6r8gp+11PNv3RmsM/6xBcYXhhfJiDyA0tI7u0cp2gVv TvNA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:mime-version :message-id:date:subject:cc:to:from; bh=JXanDA/ABYcoIZYeJL4efLBSA/his1lJ3Qr9kTwd/O4=; fh=sO6tabB9Wkbm/jrGt/hKe1LmwgPxoQCGTD1JlX/xdDc=; b=ODKM/2ZlFhdirF5Y13lriONmQ+iFpsW9qSqa+7tDj8JoDg+8Qtt4wj8/1vXpOKy4in Vf4IbP78JBmlLepqPA+pNYIKfLV5EAie2bl+DZwvytb69+Q8IRWU8HEjvcOIzJ22wNyw 4SAAueoJ/h2Ng4AxtcZcxCiD9CzY3KKAdQk4R8kpUYAgEb9Gnm1yRQRCpIhAd1ZpSfr4 eHT+K4RnkCV1A700c9Zt266x58rp+BRccssCaYhy21/Vf+RDIVSd8ZCed1NkjESofdsh nuQ3RMXN8psDT4+uRBWvGT3ITdY4x36K8pp2QKrA+j6ZlNVH5BmskpwDns8pO1MGxl6a mQgQ== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::3:8 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=alibaba.com Return-Path: Received: from fry.vger.email (fry.vger.email. [2620:137:e000::3:8]) by mx.google.com with ESMTPS id a3-20020a63d203000000b00565ecee8793si4194921pgg.875.2023.10.29.18.14.07 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Sun, 29 Oct 2023 18:14:07 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::3:8 as permitted sender) client-ip=2620:137:e000::3:8; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::3:8 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=alibaba.com Received: from out1.vger.email (depot.vger.email [IPv6:2620:137:e000::3:0]) by fry.vger.email (Postfix) with ESMTP id 3B60780952C5; Sun, 29 Oct 2023 18:14:05 -0700 (PDT) X-Virus-Status: Clean X-Virus-Scanned: clamav-milter 0.103.10 at fry.vger.email Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S230421AbjJ3BMH (ORCPT + 99 others); Sun, 29 Oct 2023 21:12:07 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:47790 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S229470AbjJ3BMF (ORCPT ); Sun, 29 Oct 2023 21:12:05 -0400 Received: from out30-124.freemail.mail.aliyun.com (out30-124.freemail.mail.aliyun.com [115.124.30.124]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id B587DBD for ; Sun, 29 Oct 2023 18:12:01 -0700 (PDT) X-Alimail-AntiSpam: AC=PASS;BC=-1|-1;BR=01201311R381e4;CH=green;DM=||false|;DS=||;FP=0|-1|-1|-1|0|-1|-1|-1;HT=ay29a033018045192;MF=baolin.wang@linux.alibaba.com;NM=1;PH=DS;RN=6;SR=0;TI=SMTPD_---0Vv4ZEFb_1698628318; Received: from localhost(mailfrom:baolin.wang@linux.alibaba.com fp:SMTPD_---0Vv4ZEFb_1698628318) by smtp.aliyun-inc.com; Mon, 30 Oct 2023 09:11:59 +0800 From: Baolin Wang To: akpm@linux-foundation.org Cc: shy828301@gmail.com, ying.huang@intel.com, baolin.wang@linux.alibaba.com, linux-mm@kvack.org, linux-kernel@vger.kernel.org Subject: [PATCH] mm: huge_memory: batch tlb flush when splitting a pte-mapped THP Date: Mon, 30 Oct 2023 09:11:47 +0800 Message-Id: <431d9fb6823036369dcb1d3b2f63732f01df21a7.1698488264.git.baolin.wang@linux.alibaba.com> X-Mailer: git-send-email 2.39.3 MIME-Version: 1.0 Content-Transfer-Encoding: 8bit X-Spam-Status: No, score=-0.7 required=5.0 tests=HEADER_FROM_DIFFERENT_DOMAINS, MAILING_LIST_MULTI,SPF_HELO_NONE,SPF_PASS,UNPARSEABLE_RELAY autolearn=unavailable autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on fry.vger.email Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org X-Greylist: Sender passed SPF test, not delayed by milter-greylist-4.6.4 (fry.vger.email [0.0.0.0]); Sun, 29 Oct 2023 18:14:05 -0700 (PDT) I can observe an obvious tlb flush hotpot when splitting a pte-mapped THP on my ARM64 server, and the distribution of this hotspot is as follows: - 16.85% split_huge_page_to_list + 7.80% down_write - 7.49% try_to_migrate - 7.48% rmap_walk_anon 7.23% ptep_clear_flush + 1.52% __split_huge_page The reason is that the split_huge_page_to_list() will build migration entries for each subpage of a pte-mapped Anon THP by try_to_migrate(), or unmap for file THP, and it will clear and tlb flush for each subpage's pte. Moreover, the split_huge_page_to_list() will set TTU_SPLIT_HUGE_PMD flag to ensure the THP is already a pte-mapped THP before splitting it to some normal pages. Actually, there is no need to flush tlb for each subpage immediately, instead we can batch tlb flush for the pte-mapped THP to improve the performance. After this patch, we can see the batch tlb flush can improve the latency obviously when running thpscale. k6.5-base patched Amean fault-both-1 1071.17 ( 0.00%) 901.83 * 15.81%* Amean fault-both-3 2386.08 ( 0.00%) 1865.32 * 21.82%* Amean fault-both-5 2851.10 ( 0.00%) 2273.84 * 20.25%* Amean fault-both-7 3679.91 ( 0.00%) 2881.66 * 21.69%* Amean fault-both-12 5916.66 ( 0.00%) 4369.55 * 26.15%* Amean fault-both-18 7981.36 ( 0.00%) 6303.57 * 21.02%* Amean fault-both-24 10950.79 ( 0.00%) 8752.56 * 20.07%* Amean fault-both-30 14077.35 ( 0.00%) 10170.01 * 27.76%* Amean fault-both-32 13061.57 ( 0.00%) 11630.08 * 10.96%* Signed-off-by: Baolin Wang --- mm/huge_memory.c | 4 +++- 1 file changed, 3 insertions(+), 1 deletion(-) diff --git a/mm/huge_memory.c b/mm/huge_memory.c index f31f02472396..0e4c14bf6872 100644 --- a/mm/huge_memory.c +++ b/mm/huge_memory.c @@ -2379,7 +2379,7 @@ void vma_adjust_trans_huge(struct vm_area_struct *vma, static void unmap_folio(struct folio *folio) { enum ttu_flags ttu_flags = TTU_RMAP_LOCKED | TTU_SPLIT_HUGE_PMD | - TTU_SYNC; + TTU_SYNC | TTU_BATCH_FLUSH; VM_BUG_ON_FOLIO(!folio_test_large(folio), folio); @@ -2392,6 +2392,8 @@ static void unmap_folio(struct folio *folio) try_to_migrate(folio, ttu_flags); else try_to_unmap(folio, ttu_flags | TTU_IGNORE_MLOCK); + + try_to_unmap_flush(); } static void remap_page(struct folio *folio, unsigned long nr) -- 2.39.3