Received: by 2002:a05:6358:3188:b0:123:57c1:9b43 with SMTP id q8csp2343416rwd; Fri, 9 Jun 2023 09:46:31 -0700 (PDT) X-Google-Smtp-Source: ACHHUZ5wCf/jD5SY7PDRjwr2t1R3MGbfysIDLqGZOI+BT7VTnCw755fhqL0Vq13oy8YhMrXS9vqJ X-Received: by 2002:a05:6a20:3d12:b0:103:b0f9:7110 with SMTP id y18-20020a056a203d1200b00103b0f97110mr1716364pzi.11.1686329191305; Fri, 09 Jun 2023 09:46:31 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1686329191; cv=none; d=google.com; s=arc-20160816; b=Q09Hju9C2+aRvAKkQB5518SzdntW2Ub1YZh2NyRDzGuv8NUB5NUDuOo2uBe0YDdR+E XiBRa3aMVSsQZtLiKgD+DxIu7dl9qeXUbi7eiESBqP7fPNzLE843giXUI5V/UX6bItB9 ve6Z4rx3+u5Kn35IdG/flYKml8unlYoQDqPPMHgKq7CZbKy52uHuxxxzVQzKYaMCGSit VD1bLG34HjVAo8xu3zMR/WrqLY6wB5Guk3JuH25svlVHXUhuMTUiyGZV8cJNdrf8GFPJ YvxtR+nSTj7LpCxdX1xEKW6trhlQmNYI19Ievd6sV4c8QIkcNq3hIME+Gsr2P6roPdmj A4gg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:mime-version:content-transfer-encoding :message-id:date:subject:cc:to:from:dkim-signature; bh=L/xGwhX2AoTt7EqDkVYEXtV8BmGP3M8VcDdHfDlDTOo=; b=oTNceFQgfxrr1S+CNQWXA5PJoW1+KAnem3zuanq1BCPzP+n9mj6+caI3rubWkHAKlI hAQwL3bueCjoka2NjPZ2oNnF/t8oJ/gIcKufSPmNsCqTDFWuLuG8G/k9IyMWpthlsBOh 5QBpUN2vy9wKY6R7HylOgFMp7OjTIo5mMj7/+mvlGctqbWbI33sB4ETTQvpFDGHrkcJm OZ4qHbSmB5RNCp83kpYdN//K8fC5ij/CyFVyxUOerFoqXAJXNcOF+XQ8JVLBdt2rey/L cM2Th6CmoN7mGocv1wNZMD7xSeJAmfH0G6xDLJGqeF36vEg0AoHH6JbsLGyCeJiM2Zdo HQTg== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@ibm.com header.s=pp1 header.b=e2i546JP; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=REJECT sp=NONE dis=NONE) header.from=ibm.com Return-Path: Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id e2-20020a63aa02000000b00502d825633asi2829948pgf.639.2023.06.09.09.46.19; Fri, 09 Jun 2023 09:46:31 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; dkim=pass header.i=@ibm.com header.s=pp1 header.b=e2i546JP; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=REJECT sp=NONE dis=NONE) header.from=ibm.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S229726AbjFIQ34 (ORCPT + 99 others); Fri, 9 Jun 2023 12:29:56 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:42364 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S229609AbjFIQ3n (ORCPT ); Fri, 9 Jun 2023 12:29:43 -0400 Received: from mx0b-001b2d01.pphosted.com (mx0b-001b2d01.pphosted.com [148.163.158.5]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id BAFBB1BEB for ; Fri, 9 Jun 2023 09:29:42 -0700 (PDT) Received: from pps.filterd (m0353722.ppops.net [127.0.0.1]) by mx0a-001b2d01.pphosted.com (8.17.1.19/8.17.1.19) with ESMTP id 359GNitX031946; Fri, 9 Jun 2023 16:29:21 GMT DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=ibm.com; h=from : to : cc : subject : date : message-id : content-transfer-encoding : mime-version; s=pp1; bh=L/xGwhX2AoTt7EqDkVYEXtV8BmGP3M8VcDdHfDlDTOo=; b=e2i546JPXmvZj+fZ6iNoyn2N8m0y5grJ95Jf0Xs/MOS9CoVsAp7DGDjGHQRlNz1f0RPj BCtj95bzAfcJWxJDLLaK3Ki5vLcr5OX4MYajfdL6hNyHsleamYLooN7eIMQ583a+jzkX NDLu7qjanToUL5t5hEhEKpBrQoAubRydd3YwdeR6IS92RTgwnbrTSoIOVace8UJvd99A GjwAkPTr1FZFzrJ8z/aS8OZ5gHnKgBcC9hp2kQ7HCMJeCReP5euzrQflfo7jf7xp43QG vLLDxm6NeaqljK5rZ+WSwVm6qf94E0yWLGLV3iIjkOELpE68rdYSaDHMreFo/qBk2i5C jg== Received: from ppma03ams.nl.ibm.com (62.31.33a9.ip4.static.sl-reverse.com [169.51.49.98]) by mx0a-001b2d01.pphosted.com (PPS) with ESMTPS id 3r47jc032h-1 (version=TLSv1.2 cipher=ECDHE-RSA-AES256-GCM-SHA384 bits=256 verify=NOT); Fri, 09 Jun 2023 16:29:20 +0000 Received: from pps.filterd (ppma03ams.nl.ibm.com [127.0.0.1]) by ppma03ams.nl.ibm.com (8.17.1.19/8.17.1.19) with ESMTP id 3599SOsk021810; Fri, 9 Jun 2023 16:29:18 GMT Received: from smtprelay02.fra02v.mail.ibm.com ([9.218.2.226]) by ppma03ams.nl.ibm.com (PPS) with ESMTPS id 3r2a76a22f-1 (version=TLSv1.2 cipher=ECDHE-RSA-AES256-GCM-SHA384 bits=256 verify=NOT); Fri, 09 Jun 2023 16:29:17 +0000 Received: from smtpav07.fra02v.mail.ibm.com (smtpav07.fra02v.mail.ibm.com [10.20.54.106]) by smtprelay02.fra02v.mail.ibm.com (8.14.9/8.14.9/NCO v10.0) with ESMTP id 359GTEpF21234262 (version=TLSv1/SSLv3 cipher=DHE-RSA-AES256-GCM-SHA384 bits=256 verify=OK); Fri, 9 Jun 2023 16:29:14 GMT Received: from smtpav07.fra02v.mail.ibm.com (unknown [127.0.0.1]) by IMSVA (Postfix) with ESMTP id 828C220043; Fri, 9 Jun 2023 16:29:14 +0000 (GMT) Received: from smtpav07.fra02v.mail.ibm.com (unknown [127.0.0.1]) by IMSVA (Postfix) with ESMTP id 4843020040; Fri, 9 Jun 2023 16:29:12 +0000 (GMT) Received: from tarunpc.ibmuc.com (unknown [9.43.86.205]) by smtpav07.fra02v.mail.ibm.com (Postfix) with ESMTP; Fri, 9 Jun 2023 16:29:12 +0000 (GMT) From: Tarun Sahu To: linux-mm@kvack.org Cc: akpm@linux-foundation.org, muchun.song@linux.dev, mike.kravetz@oracle.com, aneesh.kumar@linux.ibm.com, willy@infradead.org, sidhartha.kumar@oracle.com, gerald.schaefer@linux.ibm.com, linux-kernel@vger.kernel.org, jaypatel@linux.ibm.com, tsahu@linux.ibm.com Subject: [PATCH v3] mm/folio: Avoid special handling for order value 0 in folio_set_order Date: Fri, 9 Jun 2023 21:59:07 +0530 Message-Id: <20230609162907.111756-1-tsahu@linux.ibm.com> X-Mailer: git-send-email 2.31.1 X-TM-AS-GCONF: 00 X-Proofpoint-ORIG-GUID: ILg9SmIf6gmUc4pFU8EEijKVkf_BPtjl X-Proofpoint-GUID: ILg9SmIf6gmUc4pFU8EEijKVkf_BPtjl Content-Transfer-Encoding: 8bit X-Proofpoint-UnRewURL: 0 URL was un-rewritten MIME-Version: 1.0 X-Proofpoint-Virus-Version: vendor=baseguard engine=ICAP:2.0.254,Aquarius:18.0.957,Hydra:6.0.573,FMLib:17.11.176.26 definitions=2023-06-09_11,2023-06-09_01,2023-05-22_02 X-Proofpoint-Spam-Details: rule=outbound_notspam policy=outbound score=0 lowpriorityscore=0 malwarescore=0 suspectscore=0 phishscore=0 mlxlogscore=999 priorityscore=1501 impostorscore=0 spamscore=0 bulkscore=0 adultscore=0 clxscore=1015 mlxscore=0 classifier=spam adjust=0 reason=mlx scancount=1 engine=8.12.0-2305260000 definitions=main-2306090135 X-Spam-Status: No, score=-2.0 required=5.0 tests=BAYES_00,DKIM_SIGNED, DKIM_VALID,DKIM_VALID_EF,RCVD_IN_MSPIKE_H5,RCVD_IN_MSPIKE_WL, SPF_HELO_NONE,SPF_PASS,T_SCC_BODY_TEXT_LINE autolearn=ham autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org folio_set_order(folio, 0) is used in kernel at two places __destroy_compound_gigantic_folio and __prep_compound_gigantic_folio. Currently, It is called to clear out the folio->_folio_nr_pages and folio->_folio_order. For __destroy_compound_gigantic_folio: In past, folio_set_order(folio, 0) was needed because page->mapping used to overlap with _folio_nr_pages and _folio_order. So if these fields were left uncleared during freeing gigantic hugepages, they were causing "BUG: bad page state" due to non-zero page->mapping. Now, After Commit a01f43901cfb ("hugetlb: be sure to free demoted CMA pages to CMA") page->mapping has explicitly been cleared out for tail pages. Also, _folio_order and _folio_nr_pages no longer overlaps with page->mapping. So, folio_set_order(folio, 0) can be removed from freeing gigantic folio path (__destroy_compound_gigantic_folio). Another place, folio_set_order(folio, 0) is called inside __prep_compound_gigantic_folio during error path. Here, folio_set_order(folio, 0) can also be removed if we move folio_set_order(folio, order) after for loop. The patch also moves _folio_set_head call in __prep_compound_gigantic_folio() such that we avoid clearing them in the error path. Also, as Mike pointed out: "It would actually be better to move the calls _folio_set_head and folio_set_order in __prep_compound_gigantic_folio() as suggested here. Why? In the current code, the ref count on the 'head page' is still 1 (or more) while those calls are made. So, someone could take a speculative ref on the page BEFORE the tail pages are set up." This way, folio_set_order(folio, 0) is no more needed. And it will also helps removing the confusion of folio order being set to 0 (as _folio_order field is part of first tail page). Testing: I have run LTP tests, which all passes. and also I have written the test in LTP which tests the bug caused by compound_nr and page->mapping overlapping. https://github.com/linux-test-project/ltp/blob/master/testcases/kernel/mem/hugetlb/hugemmap/hugemmap32.c Running on older kernel ( < 5.10-rc7) with the above bug this fails while on newer kernel and, also with this patch it passes. Signed-off-by: Tarun Sahu --- v2->v3 - removed the copy of page/folio definition from commit msg v1->v2 - Reword the commit message mm/hugetlb.c | 9 +++------ mm/internal.h | 8 ++------ 2 files changed, 5 insertions(+), 12 deletions(-) diff --git a/mm/hugetlb.c b/mm/hugetlb.c index f154019e6b84..607553445855 100644 --- a/mm/hugetlb.c +++ b/mm/hugetlb.c @@ -1489,7 +1489,6 @@ static void __destroy_compound_gigantic_folio(struct folio *folio, set_page_refcounted(p); } - folio_set_order(folio, 0); __folio_clear_head(folio); } @@ -1951,9 +1950,6 @@ static bool __prep_compound_gigantic_folio(struct folio *folio, struct page *p; __folio_clear_reserved(folio); - __folio_set_head(folio); - /* we rely on prep_new_hugetlb_folio to set the destructor */ - folio_set_order(folio, order); for (i = 0; i < nr_pages; i++) { p = folio_page(folio, i); @@ -1999,6 +1995,9 @@ static bool __prep_compound_gigantic_folio(struct folio *folio, if (i != 0) set_compound_head(p, &folio->page); } + __folio_set_head(folio); + /* we rely on prep_new_hugetlb_folio to set the destructor */ + folio_set_order(folio, order); atomic_set(&folio->_entire_mapcount, -1); atomic_set(&folio->_nr_pages_mapped, 0); atomic_set(&folio->_pincount, 0); @@ -2017,8 +2016,6 @@ static bool __prep_compound_gigantic_folio(struct folio *folio, p = folio_page(folio, j); __ClearPageReserved(p); } - folio_set_order(folio, 0); - __folio_clear_head(folio); return false; } diff --git a/mm/internal.h b/mm/internal.h index 68410c6d97ac..c59fe08c5b39 100644 --- a/mm/internal.h +++ b/mm/internal.h @@ -425,16 +425,12 @@ int split_free_page(struct page *free_page, */ static inline void folio_set_order(struct folio *folio, unsigned int order) { - if (WARN_ON_ONCE(!folio_test_large(folio))) + if (WARN_ON_ONCE(!order || !folio_test_large(folio))) return; folio->_folio_order = order; #ifdef CONFIG_64BIT - /* - * When hugetlb dissolves a folio, we need to clear the tail - * page, rather than setting nr_pages to 1. - */ - folio->_folio_nr_pages = order ? 1U << order : 0; + folio->_folio_nr_pages = 1U << order; #endif } -- 2.31.1