Received: by 2002:a05:6a10:206:0:0:0:0 with SMTP id 6csp3939368pxj; Tue, 8 Jun 2021 02:26:35 -0700 (PDT) X-Google-Smtp-Source: ABdhPJz3+K06Qe4VfGmdBHurCeMVJhhaIl4foV3OWi8vnhjC1A9nlnCtQGUJyg8rp0fEllsC14jF X-Received: by 2002:a05:6402:40c3:: with SMTP id z3mr2738741edb.187.1623144394812; Tue, 08 Jun 2021 02:26:34 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1623144394; cv=none; d=google.com; s=arc-20160816; b=ez5cGc9Yl0fZUZgZVd5/ul+LrANwoFxK+lGZ0tLj3z4WF4H7L2zC/2Bv6PGRUPWMRF 8ZLmpTSiA7UwjHKLYhHqkLKTb+Y2K7sY4jnrVYLVhvlDbaBrXmhtcXkUI+GrKAaEGjTA 553AJEGHsuQviWGvONEjCqmUiflX6w+NY49dOlg9lpBGOOmjChdt6kdLci6ahT8rJybo 4x3mAMIMzLSVqZQ9O5Ylv9It6SP9JHzS9O0Ml9jadPf9KUpBFm0UzaCZgqhBUJGjOm5n VkOwYnlfURuOW9MaGXMq04+BzAKfVp2HQ/1I0oce7ATN4+sbkuVB5o/cVGUjEX6OgTvJ mrhw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:mime-version :message-id:date:subject:cc:to:from; bh=k0KrYTKY40K5k1VuaDo9DvRFPJcnk3Eqr1CbXt43NHE=; b=zIfbk2nNKpxb7dls8FZ8RUFPDssYqrDpgUe3lB3GV47r4TffrUM80HAgN+aVNkgGFe bO1/OnqGvZ2aWevyNo0D4WRjjqtow8gd/Xp/jwi4Z+ZcMfbL+XgbrbVuyoIF03NjW3CT qD0ZRNj0uhVVQZjSKbjz9qLfIxFHkZoOAk01oavpk4pBARMCdGG0Rm+aaDCjyGg4TAds UadbwendK6DWk8sU2G8NvanSco5FSl31YuSOJ+hsOI/tY+G1HeKODt93MRs3glSrjQQG HnPuneDtP3P+keqYNSU7AAj20FX0Sv8h+nQbFdu1EswLTedBld0a53p6wy8LOIq8tjWO y2Vw== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=alibaba.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id q7si8083574edd.99.2021.06.08.02.26.08; Tue, 08 Jun 2021 02:26:34 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=alibaba.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S229792AbhFHJYj (ORCPT + 99 others); Tue, 8 Jun 2021 05:24:39 -0400 Received: from out30-56.freemail.mail.aliyun.com ([115.124.30.56]:44065 "EHLO out30-56.freemail.mail.aliyun.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S229507AbhFHJYj (ORCPT ); Tue, 8 Jun 2021 05:24:39 -0400 X-Alimail-AntiSpam: AC=PASS;BC=-1|-1;BR=01201311R201e4;CH=green;DM=||false|;DS=||;FP=0|-1|-1|-1|0|-1|-1|-1;HT=e01e04400;MF=xuyu@linux.alibaba.com;NM=1;PH=DS;RN=5;SR=0;TI=SMTPD_---0UbkfDnd_1623144164; Received: from localhost(mailfrom:xuyu@linux.alibaba.com fp:SMTPD_---0UbkfDnd_1623144164) by smtp.aliyun-inc.com(127.0.0.1); Tue, 08 Jun 2021 17:22:45 +0800 From: Xu Yu To: linux-mm@kvack.org, linux-kernel@vger.kernel.org, hughd@google.com Cc: akpm@linux-foundation.org, gavin.dg@linux.alibaba.com Subject: [PATCH v2] mm, thp: use head page in __migration_entry_wait Date: Tue, 8 Jun 2021 17:22:39 +0800 Message-Id: X-Mailer: git-send-email 2.20.1.2432.ga663e714 MIME-Version: 1.0 Content-Transfer-Encoding: 8bit Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org We notice that hung task happens in a conner but practical scenario when CONFIG_PREEMPT_NONE is enabled, as follows. Process 0 Process 1 Process 2..Inf split_huge_page_to_list unmap_page split_huge_pmd_address __migration_entry_wait(head) __migration_entry_wait(tail) remap_page (roll back) remove_migration_ptes rmap_walk_anon cond_resched Where __migration_entry_wait(tail) is occurred in kernel space, e.g., copy_to_user in fstat, which will immediately fault again without rescheduling, and thus occupy the cpu fully. When there are too many processes performing __migration_entry_wait on tail page, remap_page will never be done after cond_resched. This makes __migration_entry_wait operate on the compound head page, thus waits for remap_page to complete, whether the THP is split successfully or roll back. Note that put_and_wait_on_page_locked helps to drop the page reference acquired with get_page_unless_zero, as soon as the page is on the wait queue, before actually waiting. So splitting the THP is only prevented for a brief interval. Fixes: ba98828088ad ("thp: add option to setup migration entries during PMD split") Suggested-by: Hugh Dickins Signed-off-by: Gang Deng Signed-off-by: Xu Yu --- mm/migrate.c | 1 + 1 file changed, 1 insertion(+) diff --git a/mm/migrate.c b/mm/migrate.c index b234c3f3acb7..41ff2c9896c4 100644 --- a/mm/migrate.c +++ b/mm/migrate.c @@ -295,6 +295,7 @@ void __migration_entry_wait(struct mm_struct *mm, pte_t *ptep, goto out; page = migration_entry_to_page(entry); + page = compound_head(page); /* * Once page cache replacement of page migration started, page_count -- 2.20.1.2432.ga663e714