Received: by 2002:a25:c593:0:0:0:0:0 with SMTP id v141csp5553063ybe; Tue, 10 Sep 2019 05:39:43 -0700 (PDT) X-Google-Smtp-Source: APXvYqwvFpqDXKBqsZxcYlfIltl5l6UEcgbHPicH1Lev7Ecm5qm/5ToSHgU4t6du9ziwR26EpEBp X-Received: by 2002:a50:fc94:: with SMTP id f20mr29427296edq.175.1568119183418; Tue, 10 Sep 2019 05:39:43 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1568119183; cv=none; d=google.com; s=arc-20160816; b=loTIKp1ZgVQByLmuvYN6Cw6DKAEAcWfWv93aL1Ugf7P5YgfU4rkheprP3gUeCuFfyI a38NtxXjA131U9CKOfh6QPc1sVT8Pf3CzcXq05Bf4WAPtXOc7LSfb/5GoKsf7///vJEy qK+B0/+7zD2a5qlr/9iMza0up5DUiOVQ3Og8755uVGDOhKIoOhx8w13QlHdBNAvZSccj o869oNr21sSqFnaCwZLxJSI80mQf2vcLEsabEn/5BtC88XXyVpkt0EY3DWBJjcDWN6Uc o8RcwHazVMzWiONK3bHjUhahpiMFazUTqPbgiphJehNgFoZYWbhQkhae1C3doxTw46Br XF/A== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:content-transfer-encoding:mime-version :message-id:subject:cc:to:from:date:dkim-signature; bh=x0YgKmdVtumN8tvDl4QalfpDpMB5gNsbs7jNvv+lUoo=; b=beREzWokKI+v5+LPpYl+XCbYHPajqDX1CQ8sGqLfF6VZDKlKCc6d/cAXX06FpIR95e mdAUUtBV9pfiGiJXJyT5JK44bu/RXKnH71wsC7JG6e4UN28jwS6JYuWvHSx3qUK5T6dr XcAZQSu7TSN/KSLr9JZcE+Hah97pB14coarTcLwrkGOh7aj2coD31VZ5o0Aw/kw4eM5N ClNL6Wo3VHG10PpO9/2KshX1gAn4UAqfDlhLV6+n6WGkC8jLtPLAj00a9g0XzCa5qLUn DxzNmGG/942nWWyPAxV3M7VoXqAEf3DeD4sZKg2/a1UnXWxtlQBEKbumrjr9pFhA844Y fRSQ== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@gmail.com header.s=20161025 header.b=gaM1Wphg; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=gmail.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id i14si9362171ejb.271.2019.09.10.05.39.19; Tue, 10 Sep 2019 05:39:43 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; dkim=pass header.i=@gmail.com header.s=20161025 header.b=gaM1Wphg; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=gmail.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S2392567AbfIJKbr (ORCPT + 99 others); Tue, 10 Sep 2019 06:31:47 -0400 Received: from mail-lj1-f195.google.com ([209.85.208.195]:35433 "EHLO mail-lj1-f195.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726482AbfIJKbq (ORCPT ); Tue, 10 Sep 2019 06:31:46 -0400 Received: by mail-lj1-f195.google.com with SMTP id q22so11296591ljj.2 for ; Tue, 10 Sep 2019 03:31:44 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20161025; h=date:from:to:cc:subject:message-id:mime-version :content-transfer-encoding; bh=x0YgKmdVtumN8tvDl4QalfpDpMB5gNsbs7jNvv+lUoo=; b=gaM1WphgA3BYhyQ3v2TEVs/84kZ3SRr/EFx6UBDMAK2XgnVFpWeEjEJ862aNXmdqX+ dQUMj1mCe2LtTH/gsVlXkse22+SQO5olU/+1krJhRdIr3m0aYomsgIaAUYD8d0qe+47z W94Xx0kuTUV1dUo/HIOoaUMtKPIZBOeEfxQFkA+JFrnigXIzS88yC7viA2cNjUaaxbLu CSBl6IfY5g9k/ttH8CRbpqWgLMLTux6c/sp+7d3m3wIEQ8GeIc42d5BkhdzSv9p0v0qb /NNZJWLGK+CV9BOczlqpIH8edaw8bWg/Pt/GIpChdu3n6IUbt4l1prNwJHlUaW0vm7ha XQUA== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:date:from:to:cc:subject:message-id:mime-version :content-transfer-encoding; bh=x0YgKmdVtumN8tvDl4QalfpDpMB5gNsbs7jNvv+lUoo=; b=XHO85MQ6w69pLoNKsiAGihpG8AuUfzxX2iFtUzbu5UZn61E7mxfGwZszuagNy6NIhD otiEDbVzPupkbBRaGBuJKdKJTQNLEr6QFqlPzR3h9jzGIaFrjoI3A4MpYg6VeLDDOwl/ tywyURBpD0lkOXfmyD/6YPaOI5SxSwhVsGF00qJ5sNE+V3HimfWZ7SUBO5EAP7TvhEfE MceKI5VJTmfvIpgdE6I2/lzbZzgYayPGggaTE7W6T61SH5RlS6n+M9ql8TB+KqAWG6Sj tlFtgvNy7jIOadKM62U+P3wdtZDZKPNY82QHRVXGZ38viY27pi/u1Tt78VkpXU7/FA92 lP3Q== X-Gm-Message-State: APjAAAVSeFEI0KPBiudODzLFPn7HbTzZdhmJ769uSTcJNttkhiJl5Lxk LFkpWhA7e42iEtedkMJ8W/o= X-Received: by 2002:a2e:86c5:: with SMTP id n5mr12600118ljj.153.1568111503915; Tue, 10 Sep 2019 03:31:43 -0700 (PDT) Received: from seldlx21914.corpusers.net ([37.139.156.40]) by smtp.gmail.com with ESMTPSA id q11sm244474ljc.27.2019.09.10.03.31.43 (version=TLS1_2 cipher=ECDHE-RSA-AES128-GCM-SHA256 bits=128/128); Tue, 10 Sep 2019 03:31:43 -0700 (PDT) Date: Tue, 10 Sep 2019 12:31:42 +0200 From: Vitaly Wool To: Linux-MM , linux-kernel@vger.kernel.org Cc: Andrew Morton , =?ISO-8859-1?Q?Agust=EDn?= Dall'Alba , Henry Burns , Vlastimil Babka Subject: [PATCH] Revert "mm/z3fold.c: fix race between migration and destruction" Message-Id: <20190910123142.7a9c8d2de4d0acbc0977c602@gmail.com> X-Mailer: Sylpheed 3.7.0 (GTK+ 2.24.30; x86_64-unknown-linux-gnu) Mime-Version: 1.0 Content-Type: text/plain; charset=ISO-8859-1 Content-Transfer-Encoding: quoted-printable Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org With the original commit applied, z3fold_zpool_destroy() may get blocked on wait_event() for indefinite time. Revert this commit for the time being to get rid of this problem since the issue the original commit addresses is less severe. This reverts commit d776aaa9895eb6eb770908e899cb7f5bd5025b3c. Reported-by: Agust=EDn Dall'Alba Signed-off-by: Vitaly Wool --- mm/z3fold.c | 90 ----------------------------------------------------- 1 file changed, 90 deletions(-) diff --git a/mm/z3fold.c b/mm/z3fold.c index 75b7962439ff..ed19d98c9dcd 100644 --- a/mm/z3fold.c +++ b/mm/z3fold.c @@ -41,7 +41,6 @@ #include #include #include -#include #include #include =20 @@ -146,8 +145,6 @@ struct z3fold_header { * @release_wq: workqueue for safe page release * @work: work_struct for safe page release * @inode: inode for z3fold pseudo filesystem - * @destroying: bool to stop migration once we start destruction - * @isolated: int to count the number of pages currently in isolation * * This structure is allocated at pool creation time and maintains metadata * pertaining to a particular z3fold pool. @@ -166,11 +163,8 @@ struct z3fold_pool { const struct zpool_ops *zpool_ops; struct workqueue_struct *compact_wq; struct workqueue_struct *release_wq; - struct wait_queue_head isolate_wait; struct work_struct work; struct inode *inode; - bool destroying; - int isolated; }; =20 /* @@ -775,7 +769,6 @@ static struct z3fold_pool *z3fold_create_pool(const cha= r *name, gfp_t gfp, goto out_c; spin_lock_init(&pool->lock); spin_lock_init(&pool->stale_lock); - init_waitqueue_head(&pool->isolate_wait); pool->unbuddied =3D __alloc_percpu(sizeof(struct list_head)*NCHUNKS, 2); if (!pool->unbuddied) goto out_pool; @@ -815,15 +808,6 @@ static struct z3fold_pool *z3fold_create_pool(const ch= ar *name, gfp_t gfp, return NULL; } =20 -static bool pool_isolated_are_drained(struct z3fold_pool *pool) -{ - bool ret; - - spin_lock(&pool->lock); - ret =3D pool->isolated =3D=3D 0; - spin_unlock(&pool->lock); - return ret; -} /** * z3fold_destroy_pool() - destroys an existing z3fold pool * @pool: the z3fold pool to be destroyed @@ -833,22 +817,6 @@ static bool pool_isolated_are_drained(struct z3fold_po= ol *pool) static void z3fold_destroy_pool(struct z3fold_pool *pool) { kmem_cache_destroy(pool->c_handle); - /* - * We set pool-> destroying under lock to ensure that - * z3fold_page_isolate() sees any changes to destroying. This way we - * avoid the need for any memory barriers. - */ - - spin_lock(&pool->lock); - pool->destroying =3D true; - spin_unlock(&pool->lock); - - /* - * We need to ensure that no pages are being migrated while we destroy - * these workqueues, as migration can queue work on either of the - * workqueues. - */ - wait_event(pool->isolate_wait, !pool_isolated_are_drained(pool)); =20 /* * We need to destroy pool->compact_wq before pool->release_wq, @@ -1339,28 +1307,6 @@ static u64 z3fold_get_pool_size(struct z3fold_pool *= pool) return atomic64_read(&pool->pages_nr); } =20 -/* - * z3fold_dec_isolated() expects to be called while pool->lock is held. - */ -static void z3fold_dec_isolated(struct z3fold_pool *pool) -{ - assert_spin_locked(&pool->lock); - VM_BUG_ON(pool->isolated <=3D 0); - pool->isolated--; - - /* - * If we have no more isolated pages, we have to see if - * z3fold_destroy_pool() is waiting for a signal. - */ - if (pool->isolated =3D=3D 0 && waitqueue_active(&pool->isolate_wait)) - wake_up_all(&pool->isolate_wait); -} - -static void z3fold_inc_isolated(struct z3fold_pool *pool) -{ - pool->isolated++; -} - static bool z3fold_page_isolate(struct page *page, isolate_mode_t mode) { struct z3fold_header *zhdr; @@ -1387,34 +1333,6 @@ static bool z3fold_page_isolate(struct page *page, i= solate_mode_t mode) spin_lock(&pool->lock); if (!list_empty(&page->lru)) list_del(&page->lru); - /* - * We need to check for destruction while holding pool->lock, as - * otherwise destruction could see 0 isolated pages, and - * proceed. - */ - if (unlikely(pool->destroying)) { - spin_unlock(&pool->lock); - /* - * If this page isn't stale, somebody else holds a - * reference to it. Let't drop our refcount so that they - * can call the release logic. - */ - if (unlikely(kref_put(&zhdr->refcount, - release_z3fold_page_locked))) { - /* - * If we get here we have kref problems, so we - * should freak out. - */ - WARN(1, "Z3fold is experiencing kref problems\n"); - z3fold_page_unlock(zhdr); - return false; - } - z3fold_page_unlock(zhdr); - return false; - } - - - z3fold_inc_isolated(pool); spin_unlock(&pool->lock); z3fold_page_unlock(zhdr); return true; @@ -1483,10 +1401,6 @@ static int z3fold_page_migrate(struct address_space = *mapping, struct page *newpa =20 queue_work_on(new_zhdr->cpu, pool->compact_wq, &new_zhdr->work); =20 - spin_lock(&pool->lock); - z3fold_dec_isolated(pool); - spin_unlock(&pool->lock); - page_mapcount_reset(page); put_page(page); return 0; @@ -1506,14 +1420,10 @@ static void z3fold_page_putback(struct page *page) INIT_LIST_HEAD(&page->lru); if (kref_put(&zhdr->refcount, release_z3fold_page_locked)) { atomic64_dec(&pool->pages_nr); - spin_lock(&pool->lock); - z3fold_dec_isolated(pool); - spin_unlock(&pool->lock); return; } spin_lock(&pool->lock); list_add(&page->lru, &pool->lru); - z3fold_dec_isolated(pool); spin_unlock(&pool->lock); z3fold_page_unlock(zhdr); } --=20 2.17.1