Received: by 2002:a25:683:0:0:0:0:0 with SMTP id 125csp88688ybg; Thu, 11 Jun 2020 18:05:28 -0700 (PDT) X-Google-Smtp-Source: ABdhPJzRVEGt+trj3fmtD1SwvycSd4R/IHgGGIeZzcFrvWxt/0KzJeW0uTH8Seth7vsdZ8gyOd8h X-Received: by 2002:a17:906:3041:: with SMTP id d1mr11176079ejd.7.1591923928621; Thu, 11 Jun 2020 18:05:28 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1591923928; cv=none; d=google.com; s=arc-20160816; b=I8wwwlXniUCmVP099nMA7XBXGgcKnGOd1hriVhdgKnr1QNQf6SiYYHims+M20dGFCe DV/8S0KyKKf6Ca/6FnkQDpDZEzPPUmbdOpFbhqZzf9FcYVHJMW92aM7oGc0UTSOIZO8W /lLO1bq72b4cNrLc8FQpDKErow3KQu2Ifk/noQ8v5JxvfcK3ushiTN8yK8ov8jsBmIdb Ma3CS/z62A/v7HAuvvsW/bj2L9Z+nuBPOn6065qLlJ38aZdvkGBrlqQDsAuGbEQIyFoq H7kKrqibEDciwyKbDrbgeqBZkWdf3KUQOY1MM2nsOePwBONJH6+bVwZxVoWz6we0UUH5 WuZA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:content-transfer-encoding :content-language:mime-version:user-agent:date:message-id:subject :from:cc:to; bh=ctz5OpELLFnnIJOizQvrv7DRS+Y5xl3J1T+Qcyxvta8=; b=mBgeefgXB1fF+IikY25z4baHAkUHWSCtZ7UeNObMV3t76hmYySikeBPWYvv1MzCiG3 VX6JfB+4iJxPUS9KgyjFz3JY5pGEJOPftY+LrwzmH0vkK7dzCvn9nyaEKEAOPubudABh EdEOrD16JMdFeJuAarTAtTMEjVIX6qgBxI4ye4+rxs4z0a5Ra0QM+pk4hPP89rQPEmeY +3AyMhLWm1FyCmlhPKXGea2LfQ5DCGYXcaSPE5dMjb54Eqb52TqJ7ewqdX1bmk6Y3Klf sNYm2Sm/Lvd1dcHLCLu599f+ENbMG7rbU5FmHOPb5kncGDlWNuNuUNJjlNNuz25DJ2t2 ZCUg== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id n25si3065417ejs.551.2020.06.11.18.05.06; Thu, 11 Jun 2020 18:05:28 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1726386AbgFLBCw (ORCPT + 99 others); Thu, 11 Jun 2020 21:02:52 -0400 Received: from szxga07-in.huawei.com ([45.249.212.35]:44454 "EHLO huawei.com" rhost-flags-OK-OK-OK-FAIL) by vger.kernel.org with ESMTP id S1726305AbgFLBCv (ORCPT ); Thu, 11 Jun 2020 21:02:51 -0400 Received: from DGGEMS405-HUB.china.huawei.com (unknown [172.30.72.58]) by Forcepoint Email with ESMTP id CFC2F10795EDF25801FC; Fri, 12 Jun 2020 09:02:49 +0800 (CST) Received: from [127.0.0.1] (10.166.215.235) by DGGEMS405-HUB.china.huawei.com (10.3.19.205) with Microsoft SMTP Server id 14.3.487.0; Fri, 12 Jun 2020 09:02:41 +0800 To: Coly Li , Kent Overstreet , CC: "linux-kernel@vger.kernel.org" , "wubo (T)" , Mingfangsen , Yanxiaodan , linfeilong , renxudong From: Zhiqiang Liu Subject: [PATCH V3] bcache: fix potential deadlock problem in btree_gc_coalesce Message-ID: Date: Fri, 12 Jun 2020 09:02:40 +0800 User-Agent: Mozilla/5.0 (Windows NT 10.0; WOW64; rv:68.0) Gecko/20100101 Thunderbird/68.2.2 MIME-Version: 1.0 Content-Type: text/plain; charset="utf-8" Content-Language: en-US Content-Transfer-Encoding: 8bit X-Originating-IP: [10.166.215.235] X-CFilter-Loop: Reflected Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org coccicheck reports: drivers/md//bcache/btree.c:1538:1-7: preceding lock on line 1417 In btree_gc_coalesce func, if the coalescing process fails, we will goto to out_nocoalesce tag directly without releasing new_nodes[i]->write_lock. Then, it will cause a deadlock when trying to acquire new_nodes[i]->write_lock for freeing new_nodes[i] before return. btree_gc_coalesce func details as follows: if alloc new_nodes[i] fails: goto out_nocoalesce; mutex_lock(&new_nodes[i]->write_lock) // obtain new_nodes[i]->write_lock for (i = nodes - 1; i > 0; --i) // main coalescing process …… if coalescing process fails: goto out_nocoalesce; // Here, directly goto out_nocoalesce // tag will cause a deadlock …… mutex_unlock(&new_nodes[i]->write_lock) // release new_nodes[i]->write_lock return; // coalesing succ, return out_nocoalesce: btree_node_free(new_nodes[i]) // free new_nodes[i] mutex_lock(&new_nodes[i]->write_lock); // obtain new_nodes[i]->write_lock clear_bit(BTREE_NODE_dirty, &ew_nodes[i]->flags); // set flag for reuse mutex_unlock(&new_nodes[i]->write_lock); // release new_nodes[i]->write_lock To fix the problem, we add a new tag 'out_unlock_nocoalesce' for releasing new_nodes[i]->write_lock before out_nocoalesce tag. If coalescing process fails, we will go to out_unlock_nocoalesce tag for releasing new_nodes[i]->write_lock before free new_nodes[i] in out_nocoalesce tag. Fixes: 2a285686c109816 ("bcache: btree locking rework") Signed-off-by: Zhiqiang Liu --- Changelog: V3: improve commit log again (suggested by Markus and Coly) V2: improve commit log V1: initial version drivers/md/bcache/btree.c | 8 ++++++-- 1 file changed, 6 insertions(+), 2 deletions(-) diff --git a/drivers/md/bcache/btree.c b/drivers/md/bcache/btree.c index 72856e5f23a3..fd1f288fd801 100644 --- a/drivers/md/bcache/btree.c +++ b/drivers/md/bcache/btree.c @@ -1389,7 +1389,7 @@ static int btree_gc_coalesce(struct btree *b, struct btree_op *op, if (__set_blocks(n1, n1->keys + n2->keys, block_bytes(b->c)) > btree_blocks(new_nodes[i])) - goto out_nocoalesce; + goto out_unlock_nocoalesce; keys = n2->keys; /* Take the key of the node we're getting rid of */ @@ -1418,7 +1418,7 @@ static int btree_gc_coalesce(struct btree *b, struct btree_op *op, if (__bch_keylist_realloc(&keylist, bkey_u64s(&new_nodes[i]->key))) - goto out_nocoalesce; + goto out_unlock_nocoalesce; bch_btree_node_write(new_nodes[i], &cl); bch_keylist_add(&keylist, &new_nodes[i]->key); @@ -1464,6 +1464,10 @@ static int btree_gc_coalesce(struct btree *b, struct btree_op *op, /* Invalidated our iterator */ return -EINTR; +out_unlock_nocoalesce: + for (i = 0; i < nodes; i++) + mutex_unlock(&new_nodes[i]->write_lock); + out_nocoalesce: closure_sync(&cl); -- 2.19.1