Received: by 10.213.65.68 with SMTP id h4csp4361658imn; Tue, 10 Apr 2018 13:41:57 -0700 (PDT) X-Google-Smtp-Source: AIpwx48Go2lfivtyvN61mTTE/YldQYm1fmFBb0XXDWh0uMTgG/GF4+bQUfbT/waPhUxUpmHJNoKb X-Received: by 2002:a17:902:5497:: with SMTP id e23-v6mr1976703pli.308.1523392917784; Tue, 10 Apr 2018 13:41:57 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1523392917; cv=none; d=google.com; s=arc-20160816; b=FCoUpDw/An0SyUlsiGS1QHHd2tvNFMkh6hubS25DlRG+HTfzo4WTQpUVV4zAtsWIpb iIDFoBKIBmeDIE2sBxd7o0jgs74pD9f5DlBD/H0rkrbKsBOyJoMlArUj6633JrIpJG64 IEyLEHdJBgXK25gF5VTX4/GT7SWxdo9gJUvj2pxypBXM3mys8b4wyVYVq/6NvBtqPtpK crwvd1hXTrEcQq3lIF9LvScwGEtVodajC6QQRMHBlcOTMPJgBPgbZdeJMhAtkABxmOaZ mJ2X4DNGLlHXWNZ1Y2jkB7i8BdcIZOSPT1Y3naPUHUMiLtsej11/1NfaK6n66x4s7ycG FQqw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:content-transfer-encoding:mime-version :references:in-reply-to:message-id:subject:cc:to:from:date :arc-authentication-results; bh=hFiiRZ5mTGR6kQ+lcmmv0akgGXNZxnd23W84Q9MysbY=; b=QO45Py7AqEJxrpMMCKyNwDJfmgHcKXNCNGsEo1HDi615acxAGGxge92o6zEY89bQbL 0X5sZYcFcIh3ZGZohhhNUtuw7cX2G/lRSS7VwBJAuTDaO2S28O48H/smtiuaN0h+biSO cSfxYI+YNxE+nLbsESDdKIaBfwZ7dF+DcZ54QYE1Xakak6lju3S5KGUY3XUGE4OTntFx Ra6YCqNtD+kBbvRpt9xPcyCwVdcgxDyAoARNgSTI/MQjS1erBkzyxzsKR79/hJ7Rc1fM y5fcKxS5D1v+RF4sDoF1fQgF9ML/cQtfQZUOTZabjuJTGGx+2w4Tnc7och8oVuR0kni6 zX3A== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id f3-v6si3414786plf.446.2018.04.10.13.41.20; Tue, 10 Apr 2018 13:41:57 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1753507AbeDJUiC (ORCPT + 99 others); Tue, 10 Apr 2018 16:38:02 -0400 Received: from mail.linuxfoundation.org ([140.211.169.12]:36188 "EHLO mail.linuxfoundation.org" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1751629AbeDJUiB (ORCPT ); Tue, 10 Apr 2018 16:38:01 -0400 Received: from akpm3.svl.corp.google.com (unknown [104.133.9.71]) by mail.linuxfoundation.org (Postfix) with ESMTPSA id 7095FC12; Tue, 10 Apr 2018 20:38:00 +0000 (UTC) Date: Tue, 10 Apr 2018 13:37:59 -0700 From: Andrew Morton To: Greg Thelen Cc: Wang Long , Michal Hocko , Johannes Weiner , Tejun Heo , npiggin@gmail.com, linux-kernel@vger.kernel.org, linux-mm@kvack.org Subject: Re: [PATCH v3] writeback: safer lock nesting Message-Id: <20180410133759.8ffd3170e5aaa7eb7eddcba6@linux-foundation.org> In-Reply-To: <20180410005908.167976-1-gthelen@google.com> References: <201804080259.VS5U0mKT%fengguang.wu@intel.com> <20180410005908.167976-1-gthelen@google.com> X-Mailer: Sylpheed 3.5.1 (GTK+ 2.24.31; x86_64-pc-linux-gnu) Mime-Version: 1.0 Content-Type: text/plain; charset=US-ASCII Content-Transfer-Encoding: 7bit Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Mon, 9 Apr 2018 17:59:08 -0700 Greg Thelen wrote: > lock_page_memcg()/unlock_page_memcg() use spin_lock_irqsave/restore() if > the page's memcg is undergoing move accounting, which occurs when a > process leaves its memcg for a new one that has > memory.move_charge_at_immigrate set. > > unlocked_inode_to_wb_begin,end() use spin_lock_irq/spin_unlock_irq() if the > given inode is switching writeback domains. Switches occur when enough > writes are issued from a new domain. > > This existing pattern is thus suspicious: > lock_page_memcg(page); > unlocked_inode_to_wb_begin(inode, &locked); > ... > unlocked_inode_to_wb_end(inode, locked); > unlock_page_memcg(page); > > If both inode switch and process memcg migration are both in-flight then > unlocked_inode_to_wb_end() will unconditionally enable interrupts while > still holding the lock_page_memcg() irq spinlock. This suggests the > possibility of deadlock if an interrupt occurs before > unlock_page_memcg(). > > truncate > __cancel_dirty_page > lock_page_memcg > unlocked_inode_to_wb_begin > unlocked_inode_to_wb_end > > > end_page_writeback > test_clear_page_writeback > lock_page_memcg > > unlock_page_memcg > > Due to configuration limitations this deadlock is not currently possible > because we don't mix cgroup writeback (a cgroupv2 feature) and > memory.move_charge_at_immigrate (a cgroupv1 feature). > > If the kernel is hacked to always claim inode switching and memcg > moving_account, then this script triggers lockup in less than a minute: > cd /mnt/cgroup/memory > mkdir a b > echo 1 > a/memory.move_charge_at_immigrate > echo 1 > b/memory.move_charge_at_immigrate > ( > echo $BASHPID > a/cgroup.procs > while true; do > dd if=/dev/zero of=/mnt/big bs=1M count=256 > done > ) & > while true; do > sync > done & > sleep 1h & > SLEEP=$! > while true; do > echo $SLEEP > a/cgroup.procs > echo $SLEEP > b/cgroup.procs > done > > Given the deadlock is not currently possible, it's debatable if there's > any reason to modify the kernel. I suggest we should to prevent future > surprises. > > ... > > Changelog since v2: > - explicitly initialize wb_lock_cookie to silence compiler warnings. But only in some places. What's up with that? > > ... > > --- a/include/linux/backing-dev.h > +++ b/include/linux/backing-dev.h > @@ -346,7 +346,7 @@ static inline struct bdi_writeback *inode_to_wb(const struct inode *inode) > /** > * unlocked_inode_to_wb_begin - begin unlocked inode wb access transaction > * @inode: target inode > - * @lockedp: temp bool output param, to be passed to the end function > + * @cookie: output param, to be passed to the end function > * > * The caller wants to access the wb associated with @inode but isn't > * holding inode->i_lock, mapping->tree_lock or wb->list_lock. This > @@ -354,12 +354,11 @@ static inline struct bdi_writeback *inode_to_wb(const struct inode *inode) > * association doesn't change until the transaction is finished with > * unlocked_inode_to_wb_end(). > * > - * The caller must call unlocked_inode_to_wb_end() with *@lockdep > - * afterwards and can't sleep during transaction. IRQ may or may not be > - * disabled on return. > + * The caller must call unlocked_inode_to_wb_end() with *@cookie afterwards and > + * can't sleep during transaction. IRQ may or may not be disabled on return. > */ Grammar is a bit awkward here, > > ... > > --- a/mm/page-writeback.c > +++ b/mm/page-writeback.c > @@ -2501,13 +2501,13 @@ void account_page_redirty(struct page *page) > if (mapping && mapping_cap_account_dirty(mapping)) { > struct inode *inode = mapping->host; > struct bdi_writeback *wb; > - bool locked; > + struct wb_lock_cookie cookie = {0}; Trivia: it's better to use "= {}" here. That has the same effect and it doesn't assume that the first field is a scalar. And indeed, the first field is a bool so it should be {false}! So... --- a/include/linux/backing-dev.h~writeback-safer-lock-nesting-fix +++ a/include/linux/backing-dev.h @@ -355,7 +355,8 @@ static inline struct bdi_writeback *inod * unlocked_inode_to_wb_end(). * * The caller must call unlocked_inode_to_wb_end() with *@cookie afterwards and - * can't sleep during transaction. IRQ may or may not be disabled on return. + * can't sleep during the transaction. IRQs may or may not be disabled on + * return. */ static inline struct bdi_writeback * unlocked_inode_to_wb_begin(struct inode *inode, struct wb_lock_cookie *cookie) --- a/mm/page-writeback.c~writeback-safer-lock-nesting-fix +++ a/mm/page-writeback.c @@ -2501,7 +2501,7 @@ void account_page_redirty(struct page *p if (mapping && mapping_cap_account_dirty(mapping)) { struct inode *inode = mapping->host; struct bdi_writeback *wb; - struct wb_lock_cookie cookie = {0}; + struct wb_lock_cookie cookie = {}; wb = unlocked_inode_to_wb_begin(inode, &cookie); current->nr_dirtied--; @@ -2613,7 +2613,7 @@ void __cancel_dirty_page(struct page *pa if (mapping_cap_account_dirty(mapping)) { struct inode *inode = mapping->host; struct bdi_writeback *wb; - struct wb_lock_cookie cookie = {0}; + struct wb_lock_cookie cookie = {}; lock_page_memcg(page); wb = unlocked_inode_to_wb_begin(inode, &cookie); @@ -2653,7 +2653,7 @@ int clear_page_dirty_for_io(struct page if (mapping && mapping_cap_account_dirty(mapping)) { struct inode *inode = mapping->host; struct bdi_writeback *wb; - struct wb_lock_cookie cookie = {0}; + struct wb_lock_cookie cookie = {}; /* * Yes, Virginia, this is indeed insane. But I wonder about the remaining uninitialized wb_lock_cookies?