Received: by 10.213.65.68 with SMTP id h4csp3600356imn; Tue, 10 Apr 2018 01:22:57 -0700 (PDT) X-Google-Smtp-Source: AIpwx4/nt0d7t8JsQ+je9pCqytN+ObVeB6c80a0x16b/bLuMHXcI0Ix8KZBx/tm6JvkW5ii3nVTI X-Received: by 2002:a17:902:51ce:: with SMTP id y72-v6mr41756069plh.157.1523348577032; Tue, 10 Apr 2018 01:22:57 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1523348576; cv=none; d=google.com; s=arc-20160816; b=SuyJF2qdUzQcCqYzX7lBMQfQ6Ddl265v7QMzdZZgHg+383yVVv6lSGNCYYa05vrJgw kc5TU7+9AJjA76JBa4BRqAwfdBpGqO659lnvKBRJKsv1VNsel5gOCtVfd1vL6Gs0YK2o 0K+UmuThbhFiKIsPvi8hvJKS/LYNED9ADqbDF5eWH9iQ39trSrPk92Up9YB+zuGqJpSo PkOe5j7zV/hJtWnPBz1LsFL8IB9wEhaPYbSd9rmHfKQNO4PVa8CfXFrXFshgJzyiErRo Ud06TY3Oju7yA0ZMciAJmabKhT7j7U3dtIOIkGxjMjexm3Nay5WyZevsDf7jFV/Ih+ZZ 8Rqg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:content-transfer-encoding :content-language:in-reply-to:mime-version:user-agent:date :message-id:from:references:cc:to:subject:dkim-signature:dkim-filter :arc-authentication-results; bh=6+8PXN28POHuULm+z6BLJSSV3L+ImCFpfgdi/u3CQCY=; b=AofxBmBgLqSlPhQIGN8PaAp7/I7yFNsi3kIMMEjjPoDm/5EbCoG/iuAMzZ5y5yz1gR u2rAFHRb9FxRSZBDI1A5RSac1CPSUPwfJ14MKM/+ZWmFEoqUlTdblSwWUObKer5i/Ji9 rke+rS768aBVn118Zsx6kYTrm9dhos8MYSW2ebjVkwM8GyrxqIfgn/4BeFEz8+3r+IGs hZXmBmCRcXZQ0ZJT7ny80UjLgoTH8askRiBE7EuXbc8LjyJW7MacA+2CtGRLgcTmGozm ns+hI4vZ5eWwTlrGBHHV4fF3a+2uNRaUxETdnZjhvWqtHhKEJ/+4Bkw8T0uWSxd3r3IL 4SOw== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@meituan.com header.s=20130113 header.b=fJw+yhdS; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=QUARANTINE sp=QUARANTINE dis=NONE) header.from=meituan.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id e92-v6si2198876pld.16.2018.04.10.01.22.20; Tue, 10 Apr 2018 01:22:56 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; dkim=pass header.i=@meituan.com header.s=20130113 header.b=fJw+yhdS; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=QUARANTINE sp=QUARANTINE dis=NONE) header.from=meituan.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1752406AbeDJIOt (ORCPT + 99 others); Tue, 10 Apr 2018 04:14:49 -0400 Received: from mx-fe5-210.meituan.com ([103.37.138.210]:45699 "EHLO mx02.meituan.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1751825AbeDJIOs (ORCPT ); Tue, 10 Apr 2018 04:14:48 -0400 Received: from localhost (localhost [127.0.0.1]) by dx-it-mx02.dx.sankuai.com (Postfix) with ESMTP id DF13029758E9; Tue, 10 Apr 2018 16:14:44 +0800 (CST) Received: from mx02.meituan.com ([127.0.0.1]) by localhost (dx-it-mx02.dx.sankuai.com [127.0.0.1]) (amavisd-new, port 10032) with ESMTP id dFJZhAMMwGeQ; Tue, 10 Apr 2018 16:14:44 +0800 (CST) Received: from localhost (localhost [127.0.0.1]) by dx-it-mx02.dx.sankuai.com (Postfix) with ESMTP id 32C62297593A; Tue, 10 Apr 2018 16:14:38 +0800 (CST) DKIM-Filter: OpenDKIM Filter v2.9.2 dx-it-mx02.dx.sankuai.com 32C62297593A DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=meituan.com; s=20130113; t=1523348084; bh=6+8PXN28POHuULm+z6BLJSSV3L+ImCFpfgdi/u3CQCY=; h=Subject:To:From:Message-ID:Date:MIME-Version:Content-Type: Content-Transfer-Encoding; b=fJw+yhdSQm50iPY0A9AY4A9bq/TZMk7YFxI9TPd81KaRL7LvEKzmj0tWiZndrMMmf 64gZb242tGPMFAosT7xBhqcyK5bLHBpGC4+f5A2Del3HaeZtx7WvhRviRcj5BucEKy nPv2EYqWNNPM5cyMwi8u9y86Iv1VdsyY83rVmvzU= X-Virus-Scanned: amavisd-new at dx-it-mx02.dx.sankuai.com Received: from mx02.meituan.com ([127.0.0.1]) by localhost (dx-it-mx02.dx.sankuai.com [127.0.0.1]) (amavisd-new, port 10026) with ESMTP id xaQaYU1nDqQA; Tue, 10 Apr 2018 16:14:38 +0800 (CST) Received: from wanglongs-MacBook-Pro.local (unknown [103.37.140.18]) by dx-it-mx02.dx.sankuai.com (Postfix) with ESMTPSA id 32D4F297596C; Tue, 10 Apr 2018 16:14:25 +0800 (CST) Subject: Re: [PATCH v3] writeback: safer lock nesting To: Greg Thelen , Michal Hocko , Andrew Morton , Johannes Weiner , Tejun Heo Cc: npiggin@gmail.com, linux-kernel@vger.kernel.org, linux-mm@kvack.org References: <201804080259.VS5U0mKT%fengguang.wu@intel.com> <20180410005908.167976-1-gthelen@google.com> From: Wang Long Message-ID: <55efb2c6-04c5-d2bb-738e-8308aa0eaf8f@meituan.com> Date: Tue, 10 Apr 2018 16:14:25 +0800 User-Agent: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.12; rv:52.0) Gecko/20100101 Thunderbird/52.7.0 MIME-Version: 1.0 In-Reply-To: <20180410005908.167976-1-gthelen@google.com> Content-Type: text/plain; charset=utf-8; format=flowed Content-Language: en-US Content-Transfer-Encoding: quoted-printable Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org > lock_page_memcg()/unlock_page_memcg() use spin_lock_irqsave/restore() i= f > the page's memcg is undergoing move accounting, which occurs when a > process leaves its memcg for a new one that has > memory.move_charge_at_immigrate set. > > unlocked_inode_to_wb_begin,end() use spin_lock_irq/spin_unlock_irq() if= the > given inode is switching writeback domains. Switches occur when enough > writes are issued from a new domain. > > This existing pattern is thus suspicious: > lock_page_memcg(page); > unlocked_inode_to_wb_begin(inode, &locked); > ... > unlocked_inode_to_wb_end(inode, locked); > unlock_page_memcg(page); > > If both inode switch and process memcg migration are both in-flight the= n > unlocked_inode_to_wb_end() will unconditionally enable interrupts while > still holding the lock_page_memcg() irq spinlock. This suggests the > possibility of deadlock if an interrupt occurs before > unlock_page_memcg(). > > truncate > __cancel_dirty_page > lock_page_memcg > unlocked_inode_to_wb_begin > unlocked_inode_to_wb_end > > > end_page_writeback > test_clear_page_writeback > lock_page_memcg > > unlock_page_memcg > > Due to configuration limitations this deadlock is not currently possibl= e > because we don't mix cgroup writeback (a cgroupv2 feature) and > memory.move_charge_at_immigrate (a cgroupv1 feature). > > If the kernel is hacked to always claim inode switching and memcg > moving_account, then this script triggers lockup in less than a minute: > cd /mnt/cgroup/memory > mkdir a b > echo 1 > a/memory.move_charge_at_immigrate > echo 1 > b/memory.move_charge_at_immigrate > ( > echo $BASHPID > a/cgroup.procs > while true; do > dd if=3D/dev/zero of=3D/mnt/big bs=3D1M count=3D256 > done > ) & > while true; do > sync > done & > sleep 1h & > SLEEP=3D$! > while true; do > echo $SLEEP > a/cgroup.procs > echo $SLEEP > b/cgroup.procs > done > > Given the deadlock is not currently possible, it's debatable if there's > any reason to modify the kernel. I suggest we should to prevent future > surprises. This deadlock occurs three times in our environment=EF=BC=8C this deadlock occurs three times in our environment. It is better to cc s= table kernel and backport it. Acked-by: Wang Long thanks > Reported-by: Wang Long > Signed-off-by: Greg Thelen > Change-Id: Ibb773e8045852978f6207074491d262f1b3fb613 > --- > Changelog since v2: > - explicitly initialize wb_lock_cookie to silence compiler warnings. > > Changelog since v1: > - add wb_lock_cookie to record lock context. > > fs/fs-writeback.c | 7 ++++--- > include/linux/backing-dev-defs.h | 5 +++++ > include/linux/backing-dev.h | 30 ++++++++++++++++-------------- > mm/page-writeback.c | 18 +++++++++--------- > 4 files changed, 34 insertions(+), 26 deletions(-) > > diff --git a/fs/fs-writeback.c b/fs/fs-writeback.c > index 1280f915079b..f4b2f6625913 100644 > --- a/fs/fs-writeback.c > +++ b/fs/fs-writeback.c > @@ -745,11 +745,12 @@ int inode_congested(struct inode *inode, int cong= _bits) > */ > if (inode && inode_to_wb_is_valid(inode)) { > struct bdi_writeback *wb; > - bool locked, congested; > + struct wb_lock_cookie lock_cookie; > + bool congested; > =20 > - wb =3D unlocked_inode_to_wb_begin(inode, &locked); > + wb =3D unlocked_inode_to_wb_begin(inode, &lock_cookie); > congested =3D wb_congested(wb, cong_bits); > - unlocked_inode_to_wb_end(inode, locked); > + unlocked_inode_to_wb_end(inode, &lock_cookie); > return congested; > } > =20 > diff --git a/include/linux/backing-dev-defs.h b/include/linux/backing-d= ev-defs.h > index bfe86b54f6c1..0bd432a4d7bd 100644 > --- a/include/linux/backing-dev-defs.h > +++ b/include/linux/backing-dev-defs.h > @@ -223,6 +223,11 @@ static inline void set_bdi_congested(struct backin= g_dev_info *bdi, int sync) > set_wb_congested(bdi->wb.congested, sync); > } > =20 > +struct wb_lock_cookie { > + bool locked; > + unsigned long flags; > +}; > + > #ifdef CONFIG_CGROUP_WRITEBACK > =20 > /** > diff --git a/include/linux/backing-dev.h b/include/linux/backing-dev.h > index 3e4ce54d84ab..1d744c61d996 100644 > --- a/include/linux/backing-dev.h > +++ b/include/linux/backing-dev.h > @@ -346,7 +346,7 @@ static inline struct bdi_writeback *inode_to_wb(con= st struct inode *inode) > /** > * unlocked_inode_to_wb_begin - begin unlocked inode wb access transa= ction > * @inode: target inode > - * @lockedp: temp bool output param, to be passed to the end function > + * @cookie: output param, to be passed to the end function > * > * The caller wants to access the wb associated with @inode but isn't > * holding inode->i_lock, mapping->tree_lock or wb->list_lock. This > @@ -354,12 +354,11 @@ static inline struct bdi_writeback *inode_to_wb(c= onst struct inode *inode) > * association doesn't change until the transaction is finished with > * unlocked_inode_to_wb_end(). > * > - * The caller must call unlocked_inode_to_wb_end() with *@lockdep > - * afterwards and can't sleep during transaction. IRQ may or may not = be > - * disabled on return. > + * The caller must call unlocked_inode_to_wb_end() with *@cookie after= wards and > + * can't sleep during transaction. IRQ may or may not be disabled on = return. > */ > static inline struct bdi_writeback * > -unlocked_inode_to_wb_begin(struct inode *inode, bool *lockedp) > +unlocked_inode_to_wb_begin(struct inode *inode, struct wb_lock_cookie = *cookie) > { > rcu_read_lock(); > =20 > @@ -367,10 +366,10 @@ unlocked_inode_to_wb_begin(struct inode *inode, b= ool *lockedp) > * Paired with store_release in inode_switch_wb_work_fn() and > * ensures that we see the new wb if we see cleared I_WB_SWITCH. > */ > - *lockedp =3D smp_load_acquire(&inode->i_state) & I_WB_SWITCH; > + cookie->locked =3D smp_load_acquire(&inode->i_state) & I_WB_SWITCH; > =20 > - if (unlikely(*lockedp)) > - spin_lock_irq(&inode->i_mapping->tree_lock); > + if (unlikely(cookie->locked)) > + spin_lock_irqsave(&inode->i_mapping->tree_lock, cookie->flags); > =20 > /* > * Protected by either !I_WB_SWITCH + rcu_read_lock() or tree_lock. > @@ -382,12 +381,14 @@ unlocked_inode_to_wb_begin(struct inode *inode, b= ool *lockedp) > /** > * unlocked_inode_to_wb_end - end inode wb access transaction > * @inode: target inode > - * @locked: *@lockedp from unlocked_inode_to_wb_begin() > + * @cookie: @cookie from unlocked_inode_to_wb_begin() > */ > -static inline void unlocked_inode_to_wb_end(struct inode *inode, bool = locked) > +static inline void unlocked_inode_to_wb_end(struct inode *inode, > + struct wb_lock_cookie *cookie) > { > - if (unlikely(locked)) > - spin_unlock_irq(&inode->i_mapping->tree_lock); > + if (unlikely(cookie->locked)) > + spin_unlock_irqrestore(&inode->i_mapping->tree_lock, > + cookie->flags); > =20 > rcu_read_unlock(); > } > @@ -434,12 +435,13 @@ static inline struct bdi_writeback *inode_to_wb(s= truct inode *inode) > } > =20 > static inline struct bdi_writeback * > -unlocked_inode_to_wb_begin(struct inode *inode, bool *lockedp) > +unlocked_inode_to_wb_begin(struct inode *inode, struct wb_lock_cookie = *cookie) > { > return inode_to_wb(inode); > } > =20 > -static inline void unlocked_inode_to_wb_end(struct inode *inode, bool = locked) > +static inline void unlocked_inode_to_wb_end(struct inode *inode, > + struct wb_lock_cookie *cookie) > { > } > =20 > diff --git a/mm/page-writeback.c b/mm/page-writeback.c > index 586f31261c83..bc38a2a7a597 100644 > --- a/mm/page-writeback.c > +++ b/mm/page-writeback.c > @@ -2501,13 +2501,13 @@ void account_page_redirty(struct page *page) > if (mapping && mapping_cap_account_dirty(mapping)) { > struct inode *inode =3D mapping->host; > struct bdi_writeback *wb; > - bool locked; > + struct wb_lock_cookie cookie =3D {0}; > =20 > - wb =3D unlocked_inode_to_wb_begin(inode, &locked); > + wb =3D unlocked_inode_to_wb_begin(inode, &cookie); > current->nr_dirtied--; > dec_node_page_state(page, NR_DIRTIED); > dec_wb_stat(wb, WB_DIRTIED); > - unlocked_inode_to_wb_end(inode, locked); > + unlocked_inode_to_wb_end(inode, &cookie); > } > } > EXPORT_SYMBOL(account_page_redirty); > @@ -2613,15 +2613,15 @@ void __cancel_dirty_page(struct page *page) > if (mapping_cap_account_dirty(mapping)) { > struct inode *inode =3D mapping->host; > struct bdi_writeback *wb; > - bool locked; > + struct wb_lock_cookie cookie =3D {0}; > =20 > lock_page_memcg(page); > - wb =3D unlocked_inode_to_wb_begin(inode, &locked); > + wb =3D unlocked_inode_to_wb_begin(inode, &cookie); > =20 > if (TestClearPageDirty(page)) > account_page_cleaned(page, mapping, wb); > =20 > - unlocked_inode_to_wb_end(inode, locked); > + unlocked_inode_to_wb_end(inode, &cookie); > unlock_page_memcg(page); > } else { > ClearPageDirty(page); > @@ -2653,7 +2653,7 @@ int clear_page_dirty_for_io(struct page *page) > if (mapping && mapping_cap_account_dirty(mapping)) { > struct inode *inode =3D mapping->host; > struct bdi_writeback *wb; > - bool locked; > + struct wb_lock_cookie cookie =3D {0}; > =20 > /* > * Yes, Virginia, this is indeed insane. > @@ -2690,14 +2690,14 @@ int clear_page_dirty_for_io(struct page *page) > * always locked coming in here, so we get the desired > * exclusion. > */ > - wb =3D unlocked_inode_to_wb_begin(inode, &locked); > + wb =3D unlocked_inode_to_wb_begin(inode, &cookie); > if (TestClearPageDirty(page)) { > dec_lruvec_page_state(page, NR_FILE_DIRTY); > dec_zone_page_state(page, NR_ZONE_WRITE_PENDING); > dec_wb_stat(wb, WB_RECLAIMABLE); > ret =3D 1; > } > - unlocked_inode_to_wb_end(inode, locked); > + unlocked_inode_to_wb_end(inode, &cookie); > return ret; > } > return TestClearPageDirty(page);