Return-Path: Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1752942AbcD0A5Y (ORCPT ); Tue, 26 Apr 2016 20:57:24 -0400 Received: from mail-pf0-f176.google.com ([209.85.192.176]:32940 "EHLO mail-pf0-f176.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1751980AbcD0A5V (ORCPT ); Tue, 26 Apr 2016 20:57:21 -0400 Date: Wed, 27 Apr 2016 09:58:53 +0900 From: Sergey Senozhatsky To: Dan Streetman Cc: Yu Zhao , Andrew Morton , Seth Jennings , Sergey Senozhatsky , Minchan Kim , Nitin Gupta , Linux-MM , Sergey Senozhatsky , linux-kernel , Dan Streetman Subject: Re: [PATCH] mm/zswap: use workqueue to destroy pool Message-ID: <20160427005853.GD4782@swordfish> References: <1461619210-10057-1-git-send-email-ddstreet@ieee.org> <1461704891-15272-1-git-send-email-ddstreet@ieee.org> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <1461704891-15272-1-git-send-email-ddstreet@ieee.org> User-Agent: Mutt/1.6.0 (2016-04-01) Sender: linux-kernel-owner@vger.kernel.org List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Content-Length: 1640 Lines: 56 Hello, On (04/26/16 17:08), Dan Streetman wrote: [..] > -static void __zswap_pool_release(struct rcu_head *head) > +static void __zswap_pool_release(struct work_struct *work) > { > - struct zswap_pool *pool = container_of(head, typeof(*pool), rcu_head); > + struct zswap_pool *pool = container_of(work, typeof(*pool), work); > + > + synchronize_rcu(); > > /* nobody should have been able to get a kref... */ > WARN_ON(kref_get_unless_zero(&pool->kref)); > @@ -674,7 +676,9 @@ static void __zswap_pool_empty(struct kref *kref) > WARN_ON(pool == zswap_pool_current()); > > list_del_rcu(&pool->list); > - call_rcu(&pool->rcu_head, __zswap_pool_release); > + > + INIT_WORK(&pool->work, __zswap_pool_release); > + schedule_work(&pool->work); so in general the patch look good to me. it's either I didn't have enough coffee yet (which is true) or _IN THEORY_ it creates a tiny race condition; which is hard (and unlikely) to hit, but still. and the problem being is CONFIG_ZSMALLOC_STAT. zsmalloc stats are exported via debugfs which is getting init during pool set up in zs_pool_stat_create() -> debugfs_create_dir() zsmalloc. so, once again, in theory, since zswap has the same , debugfs dir will have the same for different pool, so a series of zpool changes via user space knob zsmalloc > zpool zbud > zpool zsmalloc > zpool can result in release zsmalloc0 switch to zbud switch to zsmalloc __zswap_pool_release() schedule_work() ... zs_create_pool() zs_pool_stat_create() << zsmalloc0 still exists >> work is finally scheduled zs_destroy_pool() zs_pool_stat_destroy() -ss