Received: by 2002:a05:6a10:206:0:0:0:0 with SMTP id 6csp3049684pxj; Mon, 7 Jun 2021 00:27:16 -0700 (PDT) X-Google-Smtp-Source: ABdhPJylK/7SSwwpu4VRaGiC1SFl6mUJueJzPSHQzI4/YOQpeU00eJKEM93w1jyq+fZQ35Iw1vaR X-Received: by 2002:aa7:dd12:: with SMTP id i18mr18285260edv.278.1623050836016; Mon, 07 Jun 2021 00:27:16 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1623050836; cv=none; d=google.com; s=arc-20160816; b=W5l9jLuTHYLdKyYIICJ29BZzQvcX/ZrWePlmqoa9vkYXPibZgV2vkjanbP9NAc56sS 5vLe8S8TGp/0s5mvKHOMWgRtMKGzXfJ6L/itZro3DEazdrsl/POFNdlaUk2DCJB6E4GK NtKkqIXyz1l5T16Jmk71I1L5NutMrQlAfzNqunJQTVfiEldVzQju9rgtRxKxnygZ4NwJ 5p/3Nox7sA7YTgOU7ieTEgd4zwjJlFij1i6+/f8aE+YRZ8zgEI8H+1mOYP0RDioZeQlA nWM3OK5AxQKh3OeaH/Kww0m6rf1Y4ag687kPP2gv42QzpDYt5ampTAbnbuuyU9VbS1ZX he5g== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:content-language :in-reply-to:mime-version:user-agent:date:message-id:from:references :cc:to:subject; bh=PuhEloIWffdytVaeD6KGTA231oXpIPk8sCd3DcCV0yg=; b=aVyNH/qRK0/yLNUnD48kot/OlJRPgsxCT7Kb+aFxxpYzX9xDEXqn6C/BCtNqv+4AmF CgppKJ8b0Y0n8jccvbCYtzVMIXVaqz4CNKfIZhDdxJM714Wfb5/Uo0dOQ/h+/4HCc8KA HbArdiPdaTVWVUOcPyoM061SRz906SiVbv4gpHxWz4IDtMowAYQ+5XDz+39viRcOEnwn Wp7udNpey5RHOH08OVeb7QYFxEMnKMZBreBlt/InuZ7SAGomVOTxSh8dGRrWjQ4/2XA/ vMHEfOTxCUupqGZ0g4Na3FVpAukKU9Eb39JAmDUcUWNMq53u294bUA4yM7LCuW6VMPh7 FWnQ== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=alibaba.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id hp35si8969762ejc.369.2021.06.07.00.26.53; Mon, 07 Jun 2021 00:27:16 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=alibaba.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S229436AbhFGH0h (ORCPT + 99 others); Mon, 7 Jun 2021 03:26:37 -0400 Received: from out30-56.freemail.mail.aliyun.com ([115.124.30.56]:58041 "EHLO out30-56.freemail.mail.aliyun.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S230197AbhFGH0g (ORCPT ); Mon, 7 Jun 2021 03:26:36 -0400 X-Alimail-AntiSpam: AC=PASS;BC=-1|-1;BR=01201311R211e4;CH=green;DM=||false|;DS=||;FP=0|-1|-1|-1|0|-1|-1|-1;HT=e01e04395;MF=xuyu@linux.alibaba.com;NM=1;PH=DS;RN=11;SR=0;TI=SMTPD_---0UbXVknN_1623050678; Received: from xuyu-mbp15.local(mailfrom:xuyu@linux.alibaba.com fp:SMTPD_---0UbXVknN_1623050678) by smtp.aliyun-inc.com(127.0.0.1); Mon, 07 Jun 2021 15:24:43 +0800 Subject: Re: [PATCH] mm, thp: relax migration wait when failed to get tail page To: Hugh Dickins Cc: linux-mm@kvack.org, linux-kernel@vger.kernel.org, akpm@linux-foundation.org, gavin.dg@linux.alibaba.com, Greg Thelen , Wei Xu , Matthew Wilcox , Nicholas Piggin , Vlastimil Babka References: From: Yu Xu Message-ID: <6c4e0df7-1f06-585f-d113-f38db6c819b5@linux.alibaba.com> Date: Mon, 7 Jun 2021 15:24:41 +0800 User-Agent: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.15; rv:78.0) Gecko/20100101 Thunderbird/78.10.2 MIME-Version: 1.0 In-Reply-To: Content-Type: text/plain; charset=utf-8; format=flowed Content-Language: en-US Content-Transfer-Encoding: 7bit Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 6/2/21 11:57 PM, Hugh Dickins wrote: > On Wed, 2 Jun 2021, Yu Xu wrote: >> On 6/2/21 12:55 AM, Hugh Dickins wrote: >>> On Wed, 2 Jun 2021, Xu Yu wrote: >>> >>>> We notice that hung task happens in a conner but practical scenario when >>>> CONFIG_PREEMPT_NONE is enabled, as follows. >>>> >>>> Process 0 Process 1 Process >>>> 2..Inf >>>> split_huge_page_to_list >>>> unmap_page >>>> split_huge_pmd_address >>>> __migration_entry_wait(head) >>>> __migration_entry_wait(tail) >>>> remap_page (roll back) >>>> remove_migration_ptes >>>> rmap_walk_anon >>>> cond_resched >>>> >>>> Where __migration_entry_wait(tail) is occurred in kernel space, e.g., >>>> copy_to_user, which will immediately fault again without rescheduling, >>>> and thus occupy the cpu fully. >>>> >>>> When there are too many processes performing __migration_entry_wait on >>>> tail page, remap_page will never be done after cond_resched. >>>> >>>> This relaxes __migration_entry_wait on tail page, thus gives remap_page >>>> a chance to complete. >>>> >>>> Signed-off-by: Gang Deng >>>> Signed-off-by: Xu Yu >>> >>> Well caught: you're absolutely right that there's a bug there. >>> But isn't cond_resched() just papering over the real bug, and >>> what it should do is a "page = compound_head(page);" before the >>> get_page_unless_zero()? How does that work out in your testing? >> >> compound_head works. The patched kernel is alive for hours under >> our reproducer, which usually makes the vanilla kernel hung after >> tens of minutes at most. > > Oh, that's good news, thanks. > > (It's still likely that a well-placed cond_resched() somewhere in > mm/gup.c would also be a good idea, but none of us have yet got > around to identifying where.) We neither. If really have to do it outside of __migration_entry_wait, return value of __migration_entry_wait is needed, and many related functions have to updated, which may be undesirable. > >> >> If we use compound_head, the behavior of __migration_entry_wait(tail) >> changes from "retry fault" to "prevent THP from being split". Is that >> right? Then which is preferred? If it were me, I would prefer "retry >> fault". > > As Matthew remarked, you are asking very good questions, and split > migration entries are difficult to think about. But I believe you'll > find it works out okay. > > The point of *put_and_* wait_on_page_locked() is that it does drop > the page reference you acquired with get_page_unless_zero, as soon > as the page is on the wait queue, before actually waiting. > > So splitting the THP is only prevented for a brief interval. Now, > it's true that if there are very many tasks faulting on portions > of the huge page, in that interval between inserting the migration > entries and freezing the huge page's refcount to 0, they can reduce > the chance of splitting considerably. But that's not an excuse for > for doing get_page_unless_zero() on the wrong thing, as it was doing. We finally come to your solution, i.e., compound_head. In that case, who should resend the compound_head patch to this issue? shall we do with your s.o.b? > > Hugh > -- Thanks, Yu