Received: by 2002:ad5:474a:0:0:0:0:0 with SMTP id i10csp8132674imu; Tue, 4 Dec 2018 03:34:23 -0800 (PST) X-Google-Smtp-Source: AFSGD/UjTv3QFiDC3ruVjcrmBX0f4Y5+CGBSshQaTzFVFsaSqPwJ3KfCZtdsHgAUyYwtf/S3GB6d X-Received: by 2002:a63:eb0e:: with SMTP id t14mr16699007pgh.445.1543923263076; Tue, 04 Dec 2018 03:34:23 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1543923263; cv=none; d=google.com; s=arc-20160816; b=VY8wytXlgSv/Ff8uVwUI09QAwwwyjnCYSP2ZAVYKnrgOXj12o5b8sy5nBk7hbXfYPU gOq8PComvxAI9wZkyYwJbf3DIp5wSf/J99Gg5CdS3I2etiu7OH0QDTFgCKK5tKviFkUg /VLYPlSRYrFBwKVjot9pZX/GFqMYZ/LgpGPpriFmDUWTLZcnGvsZY6MrV8FTiZXSCE7N CJV+dYcvv6B+O+z9BqA1ZpOGWoZuWAGLqmeIoEywMpcyFsAkMUbeB6pVpLx/mVbD7zca QDPqiZBgqzOvTujGnfeFdBactodt62MCwS5zM+whJdO76DDQxuGyGJfaoK5MgfkDEwXw HjoQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:content-transfer-encoding:mime-version :user-agent:references:in-reply-to:message-id:date:subject:cc:to :from:dkim-signature; bh=wIjvlGLQHthL7Q5eiMe4ZHE76hRljnkeM5H0BdK3RRU=; b=kFLZt6yJdJwApvp27JfkLFFRGBUjDd0L/k+gpd4Q70QDyqCWswhpBjWskd0jdqTLUP 5BoEyY80GWT4IgAr/mZJWUkL9mce4kKIjVLoz8SMDLKGbRkC/hQn1bhW1D/L/bghbgCL bSGkGPHd/MlPcCXc6DojG3FETk1wVBC76UcgRnZNbuJf5UTKufeLZPs4SgNVl6irz2Ij tSFy1XXvOUI+cAgPf4vjeUzzzXLe/Ld+NQ1E90jss4cFFkgFU5BvVGAyIhSVxIh3O7EW f9UtFllZhqFw+x6sh2EsW5vtM/uwwz8FcPJZLSTO2Nhy/ks42gHjiFSXhjVOBR7bSeQ8 lUCg== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@kernel.org header.s=default header.b=ddMBMzR1; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id d4si17045259pls.348.2018.12.04.03.34.08; Tue, 04 Dec 2018 03:34:23 -0800 (PST) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; dkim=pass header.i=@kernel.org header.s=default header.b=ddMBMzR1; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1726643AbeLDLcU (ORCPT + 99 others); Tue, 4 Dec 2018 06:32:20 -0500 Received: from mail.kernel.org ([198.145.29.99]:39298 "EHLO mail.kernel.org" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1725902AbeLDK4L (ORCPT ); Tue, 4 Dec 2018 05:56:11 -0500 Received: from localhost (5356596B.cm-6-7b.dynamic.ziggo.nl [83.86.89.107]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by mail.kernel.org (Postfix) with ESMTPSA id DF874214DA; Tue, 4 Dec 2018 10:56:09 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=kernel.org; s=default; t=1543920970; bh=E5e+OhVZ9zT/Mm7mEMiy9yW4hT06gOx3cNXG7tPy4oU=; h=From:To:Cc:Subject:Date:In-Reply-To:References:From; b=ddMBMzR1QJNlogGjiRoc+SuRbG2QDEwQej48N85GnB8CIRLkvTwG42NVhn3CE4QO+ Dx16pGmM98m4bydoPRARVsDhThdAq8EmSsT3fZlKJRLwR1AJ1ao1qsJRTAzpJqn9aY pIN84icz84KBe8N93aAgrfBJSks+qmE2c8RjNN5E= From: Greg Kroah-Hartman To: linux-kernel@vger.kernel.org Cc: Greg Kroah-Hartman , stable@vger.kernel.org, Hugh Dickins , "Kirill A. Shutemov" , Jerome Glisse , Konstantin Khlebnikov , Matthew Wilcox , Andrew Morton , Linus Torvalds , Sasha Levin Subject: [PATCH 4.19 003/139] mm/huge_memory: fix lockdep complaint on 32-bit i_size_read() Date: Tue, 4 Dec 2018 11:48:04 +0100 Message-Id: <20181204103650.092375118@linuxfoundation.org> X-Mailer: git-send-email 2.19.2 In-Reply-To: <20181204103649.950154335@linuxfoundation.org> References: <20181204103649.950154335@linuxfoundation.org> User-Agent: quilt/0.65 X-stable: review X-Patchwork-Hint: ignore MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org 4.19-stable review patch. If anyone has any objections, please let me know. ------------------ commit 006d3ff27e884f80bd7d306b041afc415f63598f upstream. Huge tmpfs testing, on 32-bit kernel with lockdep enabled, showed that __split_huge_page() was using i_size_read() while holding the irq-safe lru_lock and page tree lock, but the 32-bit i_size_read() uses an irq-unsafe seqlock which should not be nested inside them. Instead, read the i_size earlier in split_huge_page_to_list(), and pass the end offset down to __split_huge_page(): all while holding head page lock, which is enough to prevent truncation of that extent before the page tree lock has been taken. Link: http://lkml.kernel.org/r/alpine.LSU.2.11.1811261520070.2275@eggly.anvils Fixes: baa355fd33142 ("thp: file pages support for split_huge_page()") Signed-off-by: Hugh Dickins Acked-by: Kirill A. Shutemov Cc: Jerome Glisse Cc: Konstantin Khlebnikov Cc: Matthew Wilcox Cc: [4.8+] Signed-off-by: Andrew Morton Signed-off-by: Linus Torvalds Signed-off-by: Sasha Levin --- mm/huge_memory.c | 19 +++++++++++++------ 1 file changed, 13 insertions(+), 6 deletions(-) diff --git a/mm/huge_memory.c b/mm/huge_memory.c index c12b441a99f9..15310f14c25e 100644 --- a/mm/huge_memory.c +++ b/mm/huge_memory.c @@ -2410,12 +2410,11 @@ static void __split_huge_page_tail(struct page *head, int tail, } static void __split_huge_page(struct page *page, struct list_head *list, - unsigned long flags) + pgoff_t end, unsigned long flags) { struct page *head = compound_head(page); struct zone *zone = page_zone(head); struct lruvec *lruvec; - pgoff_t end = -1; int i; lruvec = mem_cgroup_page_lruvec(head, zone->zone_pgdat); @@ -2423,9 +2422,6 @@ static void __split_huge_page(struct page *page, struct list_head *list, /* complete memcg works before add pages to LRU */ mem_cgroup_split_huge_fixup(head); - if (!PageAnon(page)) - end = DIV_ROUND_UP(i_size_read(head->mapping->host), PAGE_SIZE); - for (i = HPAGE_PMD_NR - 1; i >= 1; i--) { __split_huge_page_tail(head, i, lruvec, list); /* Some pages can be beyond i_size: drop them from page cache */ @@ -2597,6 +2593,7 @@ int split_huge_page_to_list(struct page *page, struct list_head *list) int count, mapcount, extra_pins, ret; bool mlocked; unsigned long flags; + pgoff_t end; VM_BUG_ON_PAGE(is_huge_zero_page(page), page); VM_BUG_ON_PAGE(!PageLocked(page), page); @@ -2619,6 +2616,7 @@ int split_huge_page_to_list(struct page *page, struct list_head *list) ret = -EBUSY; goto out; } + end = -1; mapping = NULL; anon_vma_lock_write(anon_vma); } else { @@ -2632,6 +2630,15 @@ int split_huge_page_to_list(struct page *page, struct list_head *list) anon_vma = NULL; i_mmap_lock_read(mapping); + + /* + *__split_huge_page() may need to trim off pages beyond EOF: + * but on 32-bit, i_size_read() takes an irq-unsafe seqlock, + * which cannot be nested inside the page tree lock. So note + * end now: i_size itself may be changed at any moment, but + * head page lock is good enough to serialize the trimming. + */ + end = DIV_ROUND_UP(i_size_read(mapping->host), PAGE_SIZE); } /* @@ -2681,7 +2688,7 @@ int split_huge_page_to_list(struct page *page, struct list_head *list) if (mapping) __dec_node_page_state(page, NR_SHMEM_THPS); spin_unlock(&pgdata->split_queue_lock); - __split_huge_page(page, list, flags); + __split_huge_page(page, list, end, flags); if (PageSwapCache(head)) { swp_entry_t entry = { .val = page_private(head) }; -- 2.17.1