Received: by 2002:a05:6a10:206:0:0:0:0 with SMTP id 6csp2230214pxj; Sun, 9 May 2021 20:06:16 -0700 (PDT) X-Google-Smtp-Source: ABdhPJw9faZsflqxtjucp4tmOkZRTdmof3wNFA2FN28ZgEEWmGQX0xoN4BXWFdVt520q7U5mQ5+8 X-Received: by 2002:a02:3506:: with SMTP id k6mr1575528jaa.39.1620615976421; Sun, 09 May 2021 20:06:16 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1620615976; cv=none; d=google.com; s=arc-20160816; b=UV3T3tlXA/gKqo/wt3rMx/Lpat2aPZyMLhvWWNJCA6zQGrYz3atvdyrfJvCw8vw2Br I1zYgX5RpEqA4DHpQoHSYtbOXD/gr20MKY1ae0Mi6kD51NSpBrYMpec1snUmX8+xoii3 FgjUPUHhjmUWqixO7W/oJfTj16BBWUru6VtJKV4LwhFfL8LahgqnGDFnQEwscr9ETtwE 41dy3J9sjPjtbVrVcJlp4G8cfY+juEUP/6xRqHrfPMCtkXN/rl0ZuQW8EzIZxPHBziMl nBvPuB4qxPRuUnMuvyo6ERPc71eXAVePA60wShsUHI48e53jUT56OHZvNzVClgxPwvAU VXTw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:mime-version :references:in-reply-to:message-id:date:subject:cc:to:from :dkim-signature; bh=VCnSzaZHCKqH61wVuHk0o1TT9bY3u6rBTZhvsMh9WN0=; b=hIA1J9jEmvH0IQoy6T6EAn8dgH21nDXcJX11Wx1JpfHvHvbLGemklyPUUWa8OQmwPo Is82raHks2GSrWYNBjOq7v151a8KJYYxV9g2UPH8fNBMeMYvUS755VSEKgbe06r9m4ry lybz+2gU14NvwbIDvUF1ZDdDpzyMpj9UosKvy8XlUfCJLeEmbH8MZ0StLcfi0J8/UXFw VBT/vQHYWdi7+SJfX9/vZU8qVU6FtmE2abIQjn8c2G/A8r4u7NSbeP0EyHWfVUuAd3b+ uwU6+3iT5E7/X9lt+iYOhG6Z5TPlkJAlUaaCPtl39n65Qp3aP57DM3A3emNCj/35xCif gm+w== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@bytedance-com.20150623.gappssmtp.com header.s=20150623 header.b=Pb7qNWAh; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=bytedance.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id z8si14783219ilh.46.2021.05.09.20.06.04; Sun, 09 May 2021 20:06:16 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=pass header.i=@bytedance-com.20150623.gappssmtp.com header.s=20150623 header.b=Pb7qNWAh; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=bytedance.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S230285AbhEJDE7 (ORCPT + 99 others); Sun, 9 May 2021 23:04:59 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:52286 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S230227AbhEJDE6 (ORCPT ); Sun, 9 May 2021 23:04:58 -0400 Received: from mail-pj1-x102d.google.com (mail-pj1-x102d.google.com [IPv6:2607:f8b0:4864:20::102d]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id A29C0C06175F for ; Sun, 9 May 2021 20:03:54 -0700 (PDT) Received: by mail-pj1-x102d.google.com with SMTP id gc22-20020a17090b3116b02901558435aec1so9441765pjb.4 for ; Sun, 09 May 2021 20:03:54 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=bytedance-com.20150623.gappssmtp.com; s=20150623; h=from:to:cc:subject:date:message-id:in-reply-to:references :mime-version:content-transfer-encoding; bh=VCnSzaZHCKqH61wVuHk0o1TT9bY3u6rBTZhvsMh9WN0=; b=Pb7qNWAha1OVfKqauvyElSqUG/Zcs/Q/nfC1DWzJov+/LLrLA1Tc0pNo3gubEKiSV1 +jkuSAD6yhtvsfzwgnXK1vqAEuZykHKCdymsi7yNXRdLQR6lH3Iy1NfyD+7/LH01r2n+ J0rc4mGI7L/pYmA3460MU+Q6dGVe665RlQ9tzI+DkW6MB7JNnHvn6Uk1TBS81eQp6UpM k6ppyBSnsjxOlVr6WvKvT2XaK3RQ3VDqrFGp0jZYZkAbiS43eud3divw3FhPyiUHDZPZ HpykkOTgeMU2+Gr2OHYtnbAu3vloFSnKdrsHTbONo9mGZmb707FdNqSKcgZAh7xxd+ta /dlA== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:from:to:cc:subject:date:message-id:in-reply-to :references:mime-version:content-transfer-encoding; bh=VCnSzaZHCKqH61wVuHk0o1TT9bY3u6rBTZhvsMh9WN0=; b=JW7r2iTd3YXl4yD0Q66MCNxHp7ZNZwZoXonC8CRAphFL4eMGRVFMrM+WUeZdmILwM2 kfOEzpH3ssKhVztcYVm1XcjlfaN6C/NaMTVJt3gnAypKHZ3SmQa/sWR7hurnC6iDI/GI a9YanVLxS87RKZhB5SoENFQzFEHEkljyXVTmIRdejZrtJ80LCqeGDPlrSbxfuCfrqfO+ 53KRnzd7y3xsC/UI0ymZZod3wsgGISev3COi2k1maLragaOeME0O5GDb1N80zQLhwqcS TwOK2mTtM5buueOYZw0NQcTOnj6gGlS6OOLy9nl5q+VE1kfvbA+zKezhMHDL+Osp0Aac gZ+w== X-Gm-Message-State: AOAM533shaVOXkuTrZVYaNMEBNJcyHWAMr4Yj6yK4Wcbb8AngC+Hfpk3 l8CVQ3bXYT1jKkvIlWaUHvhKQg== X-Received: by 2002:a17:90a:684b:: with SMTP id e11mr25764342pjm.87.1620615833983; Sun, 09 May 2021 20:03:53 -0700 (PDT) Received: from localhost.localdomain ([139.177.225.236]) by smtp.gmail.com with ESMTPSA id a128sm9777003pfd.115.2021.05.09.20.03.42 (version=TLS1_2 cipher=ECDHE-ECDSA-AES128-GCM-SHA256 bits=128/128); Sun, 09 May 2021 20:03:53 -0700 (PDT) From: Muchun Song To: corbet@lwn.net, mike.kravetz@oracle.com, tglx@linutronix.de, mingo@redhat.com, bp@alien8.de, x86@kernel.org, hpa@zytor.com, dave.hansen@linux.intel.com, luto@kernel.org, peterz@infradead.org, viro@zeniv.linux.org.uk, akpm@linux-foundation.org, paulmck@kernel.org, pawan.kumar.gupta@linux.intel.com, rdunlap@infradead.org, oneukum@suse.com, anshuman.khandual@arm.com, jroedel@suse.de, almasrymina@google.com, rientjes@google.com, willy@infradead.org, osalvador@suse.de, mhocko@suse.com, song.bao.hua@hisilicon.com, david@redhat.com, naoya.horiguchi@nec.com, joao.m.martins@oracle.com Cc: duanxiongchun@bytedance.com, fam.zheng@bytedance.com, zhengqi.arch@bytedance.com, linux-doc@vger.kernel.org, linux-kernel@vger.kernel.org, linux-mm@kvack.org, linux-fsdevel@vger.kernel.org, Muchun Song , Miaohe Lin , Chen Huang , Bodeddula Balasubramaniam Subject: [PATCH v23 3/9] mm: hugetlb: gather discrete indexes of tail page Date: Mon, 10 May 2021 11:00:21 +0800 Message-Id: <20210510030027.56044-4-songmuchun@bytedance.com> X-Mailer: git-send-email 2.21.0 (Apple Git-122) In-Reply-To: <20210510030027.56044-1-songmuchun@bytedance.com> References: <20210510030027.56044-1-songmuchun@bytedance.com> MIME-Version: 1.0 Content-Transfer-Encoding: 8bit Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org For HugeTLB page, there are more metadata to save in the struct page. But the head struct page cannot meet our needs, so we have to abuse other tail struct page to store the metadata. In order to avoid conflicts caused by subsequent use of more tail struct pages, we can gather these discrete indexes of tail struct page. In this case, it will be easier to add a new tail page index later. Signed-off-by: Muchun Song Reviewed-by: Oscar Salvador Reviewed-by: Miaohe Lin Tested-by: Chen Huang Tested-by: Bodeddula Balasubramaniam Acked-by: Michal Hocko Reviewed-by: Mike Kravetz --- include/linux/hugetlb.h | 21 +++++++++++++++++++-- include/linux/hugetlb_cgroup.h | 19 +++++++++++-------- 2 files changed, 30 insertions(+), 10 deletions(-) diff --git a/include/linux/hugetlb.h b/include/linux/hugetlb.h index b92f25ccef58..d523a345dc86 100644 --- a/include/linux/hugetlb.h +++ b/include/linux/hugetlb.h @@ -29,6 +29,23 @@ typedef struct { unsigned long pd; } hugepd_t; #include #include +/* + * For HugeTLB page, there are more metadata to save in the struct page. But + * the head struct page cannot meet our needs, so we have to abuse other tail + * struct page to store the metadata. In order to avoid conflicts caused by + * subsequent use of more tail struct pages, we gather these discrete indexes + * of tail struct page here. + */ +enum { + SUBPAGE_INDEX_SUBPOOL = 1, /* reuse page->private */ +#ifdef CONFIG_CGROUP_HUGETLB + SUBPAGE_INDEX_CGROUP, /* reuse page->private */ + SUBPAGE_INDEX_CGROUP_RSVD, /* reuse page->private */ + __MAX_CGROUP_SUBPAGE_INDEX = SUBPAGE_INDEX_CGROUP_RSVD, +#endif + __NR_USED_SUBPAGE, +}; + struct hugepage_subpool { spinlock_t lock; long count; @@ -627,13 +644,13 @@ extern unsigned int default_hstate_idx; */ static inline struct hugepage_subpool *hugetlb_page_subpool(struct page *hpage) { - return (struct hugepage_subpool *)(hpage+1)->private; + return (void *)page_private(hpage + SUBPAGE_INDEX_SUBPOOL); } static inline void hugetlb_set_page_subpool(struct page *hpage, struct hugepage_subpool *subpool) { - set_page_private(hpage+1, (unsigned long)subpool); + set_page_private(hpage + SUBPAGE_INDEX_SUBPOOL, (unsigned long)subpool); } static inline struct hstate *hstate_file(struct file *f) diff --git a/include/linux/hugetlb_cgroup.h b/include/linux/hugetlb_cgroup.h index 0bff345c4bc6..0b8d1fdda3a1 100644 --- a/include/linux/hugetlb_cgroup.h +++ b/include/linux/hugetlb_cgroup.h @@ -21,15 +21,16 @@ struct hugetlb_cgroup; struct resv_map; struct file_region; +#ifdef CONFIG_CGROUP_HUGETLB /* * Minimum page order trackable by hugetlb cgroup. * At least 4 pages are necessary for all the tracking information. - * The second tail page (hpage[2]) is the fault usage cgroup. - * The third tail page (hpage[3]) is the reservation usage cgroup. + * The second tail page (hpage[SUBPAGE_INDEX_CGROUP]) is the fault + * usage cgroup. The third tail page (hpage[SUBPAGE_INDEX_CGROUP_RSVD]) + * is the reservation usage cgroup. */ -#define HUGETLB_CGROUP_MIN_ORDER 2 +#define HUGETLB_CGROUP_MIN_ORDER order_base_2(__MAX_CGROUP_SUBPAGE_INDEX + 1) -#ifdef CONFIG_CGROUP_HUGETLB enum hugetlb_memory_event { HUGETLB_MAX, HUGETLB_NR_MEMORY_EVENTS, @@ -66,9 +67,9 @@ __hugetlb_cgroup_from_page(struct page *page, bool rsvd) if (compound_order(page) < HUGETLB_CGROUP_MIN_ORDER) return NULL; if (rsvd) - return (struct hugetlb_cgroup *)page[3].private; + return (void *)page_private(page + SUBPAGE_INDEX_CGROUP_RSVD); else - return (struct hugetlb_cgroup *)page[2].private; + return (void *)page_private(page + SUBPAGE_INDEX_CGROUP); } static inline struct hugetlb_cgroup *hugetlb_cgroup_from_page(struct page *page) @@ -90,9 +91,11 @@ static inline int __set_hugetlb_cgroup(struct page *page, if (compound_order(page) < HUGETLB_CGROUP_MIN_ORDER) return -1; if (rsvd) - page[3].private = (unsigned long)h_cg; + set_page_private(page + SUBPAGE_INDEX_CGROUP_RSVD, + (unsigned long)h_cg); else - page[2].private = (unsigned long)h_cg; + set_page_private(page + SUBPAGE_INDEX_CGROUP, + (unsigned long)h_cg); return 0; } -- 2.11.0