Received: by 10.223.185.116 with SMTP id b49csp860155wrg; Wed, 21 Feb 2018 08:05:08 -0800 (PST) X-Google-Smtp-Source: AH8x2241+cgGOKkCL154OQgQlmgM6842F2jcpigGdLtTmTetiWWW4zL+g+hqNvZh5Dogug4PoKJd X-Received: by 10.98.61.73 with SMTP id k70mr800866pfa.10.1519229108212; Wed, 21 Feb 2018 08:05:08 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1519229108; cv=none; d=google.com; s=arc-20160816; b=IoZpyRIk2dq/QPc4AFveLHyA6ioZsKrbopH14XXamXv9FE16wqh6r4Dx5HPqabQHi3 7VXYR6Vt9TapCrRUNI21PUmxtddyNoQb2xL48yaFPekQZtMXNKm1NAo39cRxPBLRRmO1 7T7uqe0MU1zoHoGkBMM6TMzuANNAWgsoqaswAoZR50DRsZ01CQgdMbG12vo/Ek0P/+Xc iFZWNCV4aRPHE2cOWq1BHG1cRMOWDiSNrafvxhslKVoGVWk7/6dwhRfzdBPtetV9SVLH e9LaBkhslmoHFXqcWrecU7BVo8+/SLC8My45K6toXE0alC3UKgJ/e/mRiQx5zmyTsM4W 3bcw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:content-transfer-encoding:mime-version :references:in-reply-to:message-id:date:subject:cc:to:from :dkim-signature:arc-authentication-results; bh=QNwysN3CzqPH9kxSwDkQO66vVLQZzqXhdjeIJpMvnaI=; b=pcs9tI5N0eNLG0O3ZQDizF8jgqsr3YRr9EaW9LjzuAF6oVGlsLGcbDv+usLHir45wM pXJd863oeh9N2FWscQGYRJ23R8b0S2AKvh1tXdne0c1G7Hp+qo+UDCLsBdsRdn/+CQfn dWLz3nx2P2zJnZpL2t9Nx1SVWAajxJF09AIpP5/NoM32gPhf8UT00CgaTvIyEjm8xst6 hQBI262KlvdKU/7UTsby6emo0Jzl4SjaS3nh75MxPihlD4kgf2giWWXXE2Ou2LFbKkQV aKUYVQxdI/OWrHP4T4OJU6p4ws7rKkLqgOd8i3uNbCq35L3gBm+AN08w7Wv30lrgQhpz rBcw== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@lightnvm-io.20150623.gappssmtp.com header.s=20150623 header.b=Kw1p7Qs8; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id e22si6810577pfj.33.2018.02.21.08.04.53; Wed, 21 Feb 2018 08:05:08 -0800 (PST) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; dkim=pass header.i=@lightnvm-io.20150623.gappssmtp.com header.s=20150623 header.b=Kw1p7Qs8; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S932314AbeBUJ2G (ORCPT + 99 others); Wed, 21 Feb 2018 04:28:06 -0500 Received: from mail-wr0-f195.google.com ([209.85.128.195]:42908 "EHLO mail-wr0-f195.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S932104AbeBUJ1m (ORCPT ); Wed, 21 Feb 2018 04:27:42 -0500 Received: by mail-wr0-f195.google.com with SMTP id k9so2435137wre.9 for ; Wed, 21 Feb 2018 01:27:42 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=lightnvm-io.20150623.gappssmtp.com; s=20150623; h=from:to:cc:subject:date:message-id:in-reply-to:references :mime-version:content-transfer-encoding; bh=QNwysN3CzqPH9kxSwDkQO66vVLQZzqXhdjeIJpMvnaI=; b=Kw1p7Qs8pRyb2qDX/8N8BQj+Em7COYaCbH0uSZsIvfKQQcBJFDv8UrVWPE3uM1nnmd nHQGAgTAWkNoxYcmL8k4m1xjbjZ3y+ZwpCa2eIjH/Bf0vJj/xbssomGd/+NbLlQh7ctB pIHgfmfCYbLxYGH3dwzNr2GFkZLJKkkCfmykn1knmgtmWNvskkmQEtld2TguAUDLRAqs THAhuDkC+WBdldj0bB4ZS7aYxGwz56XGiuzQYpglzITrb/JIDqLb44MUKA9+lKFujqf5 V977QeIgcXUEqZ6VzeKz/4K99r2McMfkqGb+MNwLBpUVIzveTwKZXuR6aF/ssJ/YQ+zn uQww== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:from:to:cc:subject:date:message-id:in-reply-to :references:mime-version:content-transfer-encoding; bh=QNwysN3CzqPH9kxSwDkQO66vVLQZzqXhdjeIJpMvnaI=; b=LIX/4nSU1blzCqeS03zX0oQYQ7yC1bYd71Fx3JLaiBopuFWQEmv8IDjlyFSsaiGTJw ro81pVP1SblmvRm4111WELSENpxt6DIo6RShpm3lO7r7/DMEsGTYtIPouWgF0nnNW+tW u39QftJ8jfw8YIKDWkCNqcLeb6qDPgAiyhbMubTr+o+dXVYgqV4vEG2nVTBKXtll1RMD iLui/UALVvZCf5XO7xMik89gCEVI8MQntTtmieKc9L7RGT57znKrKFW69fd/HkkgyC1f vyOm7Y3uY5GHLiSc8xQehrMzpTNnWFzcA+Hymc3kLjei7+JQNubWHdFTbXeo64Tu5i78 Kxnw== X-Gm-Message-State: APf1xPDMkBqRlrdZrAq5jFhbE2jVhgpIQkFM5XmwYIivSTzZhgnJkOCu L2ysyi235DGns7qZG44KwBHLQg== X-Received: by 10.80.224.139 with SMTP id f11mr4074430edl.50.1519205261129; Wed, 21 Feb 2018 01:27:41 -0800 (PST) Received: from uHalley.cnexlabs.com (6164211-cl69.boa.fiberby.dk. [193.106.164.211]) by smtp.gmail.com with ESMTPSA id a17sm3389955eda.42.2018.02.21.01.27.39 (version=TLS1_2 cipher=ECDHE-RSA-AES128-SHA bits=128/128); Wed, 21 Feb 2018 01:27:40 -0800 (PST) From: "=?UTF-8?q?Javier=20Gonz=C3=A1lez?=" X-Google-Original-From: =?UTF-8?q?Javier=20Gonz=C3=A1lez?= To: mb@lightnvm.io Cc: linux-block@vger.kernel.org, linux-kernel@vger.kernel.org, linux-nvme@lists.infradead.org, =?UTF-8?q?Javier=20Gonz=C3=A1lez?= Subject: [PATCH 19/20] lightnvm: pblk: refactor init/exit sequences Date: Wed, 21 Feb 2018 10:26:57 +0100 Message-Id: <1519205218-26994-20-git-send-email-javier@cnexlabs.com> X-Mailer: git-send-email 2.7.4 In-Reply-To: <1519205218-26994-1-git-send-email-javier@cnexlabs.com> References: <1519205218-26994-1-git-send-email-javier@cnexlabs.com> MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Refactor init and exit sequences to improve readability. In the way, fix bad free ordering on the init error path. Signed-off-by: Javier González --- drivers/lightnvm/pblk-init.c | 503 ++++++++++++++++++++++--------------------- 1 file changed, 254 insertions(+), 249 deletions(-) diff --git a/drivers/lightnvm/pblk-init.c b/drivers/lightnvm/pblk-init.c index 1618e923ef5f..c5c304a37554 100644 --- a/drivers/lightnvm/pblk-init.c +++ b/drivers/lightnvm/pblk-init.c @@ -103,7 +103,40 @@ static void pblk_l2p_free(struct pblk *pblk) vfree(pblk->trans_map); } -static int pblk_l2p_init(struct pblk *pblk) +static int pblk_l2p_recover(struct pblk *pblk, bool factory_init) +{ + struct pblk_line *line = NULL; + + if (factory_init) { + pblk_setup_uuid(pblk); + } else { + line = pblk_recov_l2p(pblk); + if (IS_ERR(line)) { + pr_err("pblk: could not recover l2p table\n"); + return -EFAULT; + } + } + +#ifdef CONFIG_NVM_DEBUG + pr_info("pblk init: L2P CRC: %x\n", pblk_l2p_crc(pblk)); +#endif + + /* Free full lines directly as GC has not been started yet */ + pblk_gc_free_full_lines(pblk); + + if (!line) { + /* Configure next line for user data */ + line = pblk_line_get_first_data(pblk); + if (!line) { + pr_err("pblk: line list corrupted\n"); + return -EFAULT; + } + } + + return 0; +} + +static int pblk_l2p_init(struct pblk *pblk, bool factory_init) { sector_t i; struct ppa_addr ppa; @@ -119,7 +152,7 @@ static int pblk_l2p_init(struct pblk *pblk) for (i = 0; i < pblk->rl.nr_secs; i++) pblk_trans_map_set(pblk, i, ppa); - return 0; + return pblk_l2p_recover(pblk, factory_init); } static void pblk_rwb_free(struct pblk *pblk) @@ -268,86 +301,113 @@ static int pblk_core_init(struct pblk *pblk) { struct nvm_tgt_dev *dev = pblk->dev; struct nvm_geo *geo = &dev->geo; + int max_write_ppas; + + atomic64_set(&pblk->user_wa, 0); + atomic64_set(&pblk->pad_wa, 0); + atomic64_set(&pblk->gc_wa, 0); + pblk->user_rst_wa = 0; + pblk->pad_rst_wa = 0; + pblk->gc_rst_wa = 0; + + atomic_long_set(&pblk->nr_flush, 0); + pblk->nr_flush_rst = 0; pblk->pgs_in_buffer = geo->c.mw_cunits * geo->c.ws_opt * geo->all_luns; + pblk->min_write_pgs = geo->c.ws_opt * (geo->c.csecs / PAGE_SIZE); + max_write_ppas = pblk->min_write_pgs * geo->all_luns; + pblk->max_write_pgs = (max_write_ppas < NVM_MAX_VLBA) ? + max_write_ppas : NVM_MAX_VLBA; + pblk_set_sec_per_write(pblk, pblk->min_write_pgs); + + if (pblk->max_write_pgs > PBLK_MAX_REQ_ADDRS) { + pr_err("pblk: cannot support device max_phys_sect\n"); + return -EINVAL; + } + + pblk->pad_dist = kzalloc((pblk->min_write_pgs - 1) * sizeof(atomic64_t), + GFP_KERNEL); + if (!pblk->pad_dist) + return -ENOMEM; + if (pblk_init_global_caches(pblk)) - return -ENOMEM; + goto fail_free_pad_dist; /* Internal bios can be at most the sectors signaled by the device. */ pblk->page_bio_pool = mempool_create_page_pool(NVM_MAX_VLBA, 0); if (!pblk->page_bio_pool) - goto free_global_caches; + goto fail_free_global_caches; pblk->gen_ws_pool = mempool_create_slab_pool(PBLK_GEN_WS_POOL_SIZE, pblk_ws_cache); if (!pblk->gen_ws_pool) - goto free_page_bio_pool; + goto fail_free_page_bio_pool; pblk->rec_pool = mempool_create_slab_pool(geo->all_luns, pblk_rec_cache); if (!pblk->rec_pool) - goto free_gen_ws_pool; + goto fail_free_gen_ws_pool; pblk->r_rq_pool = mempool_create_slab_pool(geo->all_luns, pblk_g_rq_cache); if (!pblk->r_rq_pool) - goto free_rec_pool; + goto fail_free_rec_pool; pblk->e_rq_pool = mempool_create_slab_pool(geo->all_luns, pblk_g_rq_cache); if (!pblk->e_rq_pool) - goto free_r_rq_pool; + goto fail_free_r_rq_pool; pblk->w_rq_pool = mempool_create_slab_pool(geo->all_luns, pblk_w_rq_cache); if (!pblk->w_rq_pool) - goto free_e_rq_pool; + goto fail_free_e_rq_pool; pblk->close_wq = alloc_workqueue("pblk-close-wq", WQ_MEM_RECLAIM | WQ_UNBOUND, PBLK_NR_CLOSE_JOBS); if (!pblk->close_wq) - goto free_w_rq_pool; + goto fail_free_w_rq_pool; pblk->bb_wq = alloc_workqueue("pblk-bb-wq", WQ_MEM_RECLAIM | WQ_UNBOUND, 0); if (!pblk->bb_wq) - goto free_close_wq; + goto fail_free_close_wq; pblk->r_end_wq = alloc_workqueue("pblk-read-end-wq", WQ_MEM_RECLAIM | WQ_UNBOUND, 0); if (!pblk->r_end_wq) - goto free_bb_wq; + goto fail_free_bb_wq; if (pblk_set_addrf(pblk)) - goto free_r_end_wq; - - if (pblk_rwb_init(pblk)) - goto free_r_end_wq; + goto fail_free_r_end_wq; INIT_LIST_HEAD(&pblk->compl_list); + return 0; -free_r_end_wq: +fail_free_r_end_wq: destroy_workqueue(pblk->r_end_wq); -free_bb_wq: +fail_free_bb_wq: destroy_workqueue(pblk->bb_wq); -free_close_wq: +fail_free_close_wq: destroy_workqueue(pblk->close_wq); -free_w_rq_pool: +fail_free_w_rq_pool: mempool_destroy(pblk->w_rq_pool); -free_e_rq_pool: +fail_free_e_rq_pool: mempool_destroy(pblk->e_rq_pool); -free_r_rq_pool: +fail_free_r_rq_pool: mempool_destroy(pblk->r_rq_pool); -free_rec_pool: +fail_free_rec_pool: mempool_destroy(pblk->rec_pool); -free_gen_ws_pool: +fail_free_gen_ws_pool: mempool_destroy(pblk->gen_ws_pool); -free_page_bio_pool: +fail_free_page_bio_pool: mempool_destroy(pblk->page_bio_pool); -free_global_caches: +fail_free_global_caches: pblk_free_global_caches(pblk); +fail_free_pad_dist: + kfree(pblk->pad_dist); return -ENOMEM; } @@ -369,9 +429,8 @@ static void pblk_core_free(struct pblk *pblk) mempool_destroy(pblk->e_rq_pool); mempool_destroy(pblk->w_rq_pool); - pblk_rwb_free(pblk); - pblk_free_global_caches(pblk); + kfree(pblk->pad_dist); } static void pblk_luns_free(struct pblk *pblk) @@ -393,8 +452,6 @@ static void pblk_line_mg_free(struct pblk *pblk) pblk_mfree(l_mg->eline_meta[i]->buf, l_mg->emeta_alloc_type); kfree(l_mg->eline_meta[i]); } - - kfree(pblk->lines); } static void pblk_line_meta_free(struct pblk_line *line) @@ -418,6 +475,10 @@ static void pblk_lines_free(struct pblk *pblk) pblk_line_meta_free(line); } spin_unlock(&l_mg->free_lock); + + pblk_line_mg_free(pblk); + + kfree(pblk->lines); } static int pblk_bb_get_tbl(struct nvm_tgt_dev *dev, struct pblk_lun *rlun, @@ -515,38 +576,6 @@ static int pblk_luns_init(struct pblk *pblk, struct ppa_addr *luns) return 0; } -static int pblk_lines_configure(struct pblk *pblk, int flags) -{ - struct pblk_line *line = NULL; - int ret = 0; - - if (!(flags & NVM_TARGET_FACTORY)) { - line = pblk_recov_l2p(pblk); - if (IS_ERR(line)) { - pr_err("pblk: could not recover l2p table\n"); - ret = -EFAULT; - } - } - -#ifdef CONFIG_NVM_DEBUG - pr_info("pblk init: L2P CRC: %x\n", pblk_l2p_crc(pblk)); -#endif - - /* Free full lines directly as GC has not been started yet */ - pblk_gc_free_full_lines(pblk); - - if (!line) { - /* Configure next line for user data */ - line = pblk_line_get_first_data(pblk); - if (!line) { - pr_err("pblk: line list corrupted\n"); - ret = -EFAULT; - } - } - - return ret; -} - /* See comment over struct line_emeta definition */ static unsigned int calc_emeta_len(struct pblk *pblk) { @@ -612,81 +641,6 @@ static void pblk_set_provision(struct pblk *pblk, long nr_free_blks) atomic_set(&pblk->rl.free_user_blocks, nr_free_blks); } -static int pblk_lines_alloc_metadata(struct pblk *pblk) -{ - struct pblk_line_mgmt *l_mg = &pblk->l_mg; - struct pblk_line_meta *lm = &pblk->lm; - int i; - - /* smeta is always small enough to fit on a kmalloc memory allocation, - * emeta depends on the number of LUNs allocated to the pblk instance - */ - for (i = 0; i < PBLK_DATA_LINES; i++) { - l_mg->sline_meta[i] = kmalloc(lm->smeta_len, GFP_KERNEL); - if (!l_mg->sline_meta[i]) - goto fail_free_smeta; - } - - /* emeta allocates three different buffers for managing metadata with - * in-memory and in-media layouts - */ - for (i = 0; i < PBLK_DATA_LINES; i++) { - struct pblk_emeta *emeta; - - emeta = kmalloc(sizeof(struct pblk_emeta), GFP_KERNEL); - if (!emeta) - goto fail_free_emeta; - - if (lm->emeta_len[0] > KMALLOC_MAX_CACHE_SIZE) { - l_mg->emeta_alloc_type = PBLK_VMALLOC_META; - - emeta->buf = vmalloc(lm->emeta_len[0]); - if (!emeta->buf) { - kfree(emeta); - goto fail_free_emeta; - } - - emeta->nr_entries = lm->emeta_sec[0]; - l_mg->eline_meta[i] = emeta; - } else { - l_mg->emeta_alloc_type = PBLK_KMALLOC_META; - - emeta->buf = kmalloc(lm->emeta_len[0], GFP_KERNEL); - if (!emeta->buf) { - kfree(emeta); - goto fail_free_emeta; - } - - emeta->nr_entries = lm->emeta_sec[0]; - l_mg->eline_meta[i] = emeta; - } - } - - l_mg->vsc_list = kcalloc(l_mg->nr_lines, sizeof(__le32), GFP_KERNEL); - if (!l_mg->vsc_list) - goto fail_free_emeta; - - for (i = 0; i < l_mg->nr_lines; i++) - l_mg->vsc_list[i] = cpu_to_le32(EMPTY_ENTRY); - - return 0; - -fail_free_emeta: - while (--i >= 0) { - if (l_mg->emeta_alloc_type == PBLK_VMALLOC_META) - vfree(l_mg->eline_meta[i]->buf); - else - kfree(l_mg->eline_meta[i]->buf); - kfree(l_mg->eline_meta[i]); - } - -fail_free_smeta: - for (i = 0; i < PBLK_DATA_LINES; i++) - kfree(l_mg->sline_meta[i]); - - return -ENOMEM; -} - static int pblk_setup_line_meta_12(struct pblk *pblk, struct pblk_line *line, void *chunk_log) { @@ -830,29 +784,13 @@ static int pblk_alloc_line_meta(struct pblk *pblk, struct pblk_line *line) return 0; } -static int pblk_lines_init(struct pblk *pblk) +static int pblk_line_mg_init(struct pblk *pblk) { struct nvm_tgt_dev *dev = pblk->dev; struct nvm_geo *geo = &dev->geo; struct pblk_line_mgmt *l_mg = &pblk->l_mg; struct pblk_line_meta *lm = &pblk->lm; - struct pblk_line *line; - void *chunk_log; - unsigned int smeta_len, emeta_len; - long nr_free_chks = 0; - int bb_distance, max_write_ppas; - int i, ret; - - pblk->min_write_pgs = geo->c.ws_opt * (geo->c.csecs / PAGE_SIZE); - max_write_ppas = pblk->min_write_pgs * geo->all_luns; - pblk->max_write_pgs = min_t(int, max_write_ppas, NVM_MAX_VLBA); - pblk_set_sec_per_write(pblk, pblk->min_write_pgs); - - if (pblk->max_write_pgs > PBLK_MAX_REQ_ADDRS) { - pr_err("pblk: vector list too big(%u > %u)\n", - pblk->max_write_pgs, PBLK_MAX_REQ_ADDRS); - return -EINVAL; - } + int i, bb_distance; l_mg->nr_lines = geo->c.num_chk; l_mg->log_line = l_mg->data_line = NULL; @@ -861,6 +799,119 @@ static int pblk_lines_init(struct pblk *pblk) atomic_set(&l_mg->sysfs_line_state, -1); bitmap_zero(&l_mg->meta_bitmap, PBLK_DATA_LINES); + INIT_LIST_HEAD(&l_mg->free_list); + INIT_LIST_HEAD(&l_mg->corrupt_list); + INIT_LIST_HEAD(&l_mg->bad_list); + INIT_LIST_HEAD(&l_mg->gc_full_list); + INIT_LIST_HEAD(&l_mg->gc_high_list); + INIT_LIST_HEAD(&l_mg->gc_mid_list); + INIT_LIST_HEAD(&l_mg->gc_low_list); + INIT_LIST_HEAD(&l_mg->gc_empty_list); + + INIT_LIST_HEAD(&l_mg->emeta_list); + + l_mg->gc_lists[0] = &l_mg->gc_high_list; + l_mg->gc_lists[1] = &l_mg->gc_mid_list; + l_mg->gc_lists[2] = &l_mg->gc_low_list; + + spin_lock_init(&l_mg->free_lock); + spin_lock_init(&l_mg->close_lock); + spin_lock_init(&l_mg->gc_lock); + + l_mg->vsc_list = kcalloc(l_mg->nr_lines, sizeof(__le32), GFP_KERNEL); + if (!l_mg->vsc_list) + goto fail; + + l_mg->bb_template = kzalloc(lm->sec_bitmap_len, GFP_KERNEL); + if (!l_mg->bb_template) + goto fail_free_vsc_list; + + l_mg->bb_aux = kzalloc(lm->sec_bitmap_len, GFP_KERNEL); + if (!l_mg->bb_aux) + goto fail_free_bb_template; + + /* smeta is always small enough to fit on a kmalloc memory allocation, + * emeta depends on the number of LUNs allocated to the pblk instance + */ + for (i = 0; i < PBLK_DATA_LINES; i++) { + l_mg->sline_meta[i] = kmalloc(lm->smeta_len, GFP_KERNEL); + if (!l_mg->sline_meta[i]) + goto fail_free_smeta; + } + + /* emeta allocates three different buffers for managing metadata with + * in-memory and in-media layouts + */ + for (i = 0; i < PBLK_DATA_LINES; i++) { + struct pblk_emeta *emeta; + + emeta = kmalloc(sizeof(struct pblk_emeta), GFP_KERNEL); + if (!emeta) + goto fail_free_emeta; + + if (lm->emeta_len[0] > KMALLOC_MAX_CACHE_SIZE) { + l_mg->emeta_alloc_type = PBLK_VMALLOC_META; + + emeta->buf = vmalloc(lm->emeta_len[0]); + if (!emeta->buf) { + kfree(emeta); + goto fail_free_emeta; + } + + emeta->nr_entries = lm->emeta_sec[0]; + l_mg->eline_meta[i] = emeta; + } else { + l_mg->emeta_alloc_type = PBLK_KMALLOC_META; + + emeta->buf = kmalloc(lm->emeta_len[0], GFP_KERNEL); + if (!emeta->buf) { + kfree(emeta); + goto fail_free_emeta; + } + + emeta->nr_entries = lm->emeta_sec[0]; + l_mg->eline_meta[i] = emeta; + } + } + + for (i = 0; i < l_mg->nr_lines; i++) + l_mg->vsc_list[i] = cpu_to_le32(EMPTY_ENTRY); + + bb_distance = (geo->all_luns) * geo->c.ws_opt; + for (i = 0; i < lm->sec_per_line; i += bb_distance) + bitmap_set(l_mg->bb_template, i, geo->c.ws_opt); + + return 0; + +fail_free_emeta: + while (--i >= 0) { + if (l_mg->emeta_alloc_type == PBLK_VMALLOC_META) + vfree(l_mg->eline_meta[i]->buf); + else + kfree(l_mg->eline_meta[i]->buf); + kfree(l_mg->eline_meta[i]); + } +fail_free_smeta: + kfree(l_mg->bb_aux); + + for (i = 0; i < PBLK_DATA_LINES; i++) + kfree(l_mg->sline_meta[i]); +fail_free_bb_template: + kfree(l_mg->bb_template); +fail_free_vsc_list: + kfree(l_mg->vsc_list); +fail: + return -ENOMEM; +} + +static int pblk_line_meta_init(struct pblk *pblk) +{ + struct nvm_tgt_dev *dev = pblk->dev; + struct nvm_geo *geo = &dev->geo; + struct pblk_line_meta *lm = &pblk->lm; + unsigned int smeta_len, emeta_len; + int i; + lm->sec_per_line = geo->c.clba * geo->all_luns; lm->blk_per_line = geo->all_luns; lm->blk_bitmap_len = BITS_TO_LONGS(geo->all_luns) * sizeof(long); @@ -911,58 +962,38 @@ static int pblk_lines_init(struct pblk *pblk) return -EINVAL; } - ret = pblk_lines_alloc_metadata(pblk); + return 0; +} + +static int pblk_lines_init(struct pblk *pblk) +{ + struct pblk_line_mgmt *l_mg = &pblk->l_mg; + struct pblk_line *line; + void *chunk_log; + long nr_free_chks = 0; + int i, ret; + + ret = pblk_line_meta_init(pblk); if (ret) return ret; - l_mg->bb_template = kzalloc(lm->sec_bitmap_len, GFP_KERNEL); - if (!l_mg->bb_template) { - ret = -ENOMEM; - goto fail_free_meta; - } - - l_mg->bb_aux = kzalloc(lm->sec_bitmap_len, GFP_KERNEL); - if (!l_mg->bb_aux) { - ret = -ENOMEM; - goto fail_free_bb_template; - } - - bb_distance = (geo->all_luns) * geo->c.ws_opt; - for (i = 0; i < lm->sec_per_line; i += bb_distance) - bitmap_set(l_mg->bb_template, i, geo->c.ws_opt); - - INIT_LIST_HEAD(&l_mg->free_list); - INIT_LIST_HEAD(&l_mg->corrupt_list); - INIT_LIST_HEAD(&l_mg->bad_list); - INIT_LIST_HEAD(&l_mg->gc_full_list); - INIT_LIST_HEAD(&l_mg->gc_high_list); - INIT_LIST_HEAD(&l_mg->gc_mid_list); - INIT_LIST_HEAD(&l_mg->gc_low_list); - INIT_LIST_HEAD(&l_mg->gc_empty_list); - - INIT_LIST_HEAD(&l_mg->emeta_list); - - l_mg->gc_lists[0] = &l_mg->gc_high_list; - l_mg->gc_lists[1] = &l_mg->gc_mid_list; - l_mg->gc_lists[2] = &l_mg->gc_low_list; - - spin_lock_init(&l_mg->free_lock); - spin_lock_init(&l_mg->close_lock); - spin_lock_init(&l_mg->gc_lock); - - pblk->lines = kcalloc(l_mg->nr_lines, sizeof(struct pblk_line), - GFP_KERNEL); - if (!pblk->lines) { - ret = -ENOMEM; - goto fail_free_bb_aux; - } + ret = pblk_line_mg_init(pblk); + if (ret) + return ret; chunk_log = pblk_chunk_get_log(pblk); if (IS_ERR(chunk_log)) { pr_err("pblk: could not get chunk log (%lu)\n", PTR_ERR(chunk_log)); ret = PTR_ERR(chunk_log); - goto fail_free_lines; + goto fail_free_meta; + } + + pblk->lines = kcalloc(l_mg->nr_lines, sizeof(struct pblk_line), + GFP_KERNEL); + if (!pblk->lines) { + ret = -ENOMEM; + goto fail_free_chunk_log; } for (i = 0; i < l_mg->nr_lines; i++) { @@ -970,7 +1001,7 @@ static int pblk_lines_init(struct pblk *pblk) ret = pblk_alloc_line_meta(pblk, line); if (ret) - goto fail_free_chunk_log; + goto fail_free_lines; nr_free_chks += pblk_setup_line_meta(pblk, line, chunk_log, i); } @@ -980,16 +1011,12 @@ static int pblk_lines_init(struct pblk *pblk) kfree(chunk_log); return 0; -fail_free_chunk_log: - kfree(chunk_log); +fail_free_lines: while (--i >= 0) pblk_line_meta_free(&pblk->lines[i]); -fail_free_lines: kfree(pblk->lines); -fail_free_bb_aux: - kfree(l_mg->bb_aux); -fail_free_bb_template: - kfree(l_mg->bb_template); +fail_free_chunk_log: + kfree(chunk_log); fail_free_meta: pblk_line_mg_free(pblk); @@ -1032,12 +1059,11 @@ static void pblk_writer_stop(struct pblk *pblk) static void pblk_free(struct pblk *pblk) { - pblk_luns_free(pblk); pblk_lines_free(pblk); - kfree(pblk->pad_dist); - pblk_line_mg_free(pblk); - pblk_core_free(pblk); pblk_l2p_free(pblk); + pblk_rwb_free(pblk); + pblk_core_free(pblk); + pblk_luns_free(pblk); kfree(pblk); } @@ -1108,19 +1134,6 @@ static void *pblk_init(struct nvm_tgt_dev *dev, struct gendisk *tdisk, spin_lock_init(&pblk->trans_lock); spin_lock_init(&pblk->lock); - if (flags & NVM_TARGET_FACTORY) - pblk_setup_uuid(pblk); - - atomic64_set(&pblk->user_wa, 0); - atomic64_set(&pblk->pad_wa, 0); - atomic64_set(&pblk->gc_wa, 0); - pblk->user_rst_wa = 0; - pblk->pad_rst_wa = 0; - pblk->gc_rst_wa = 0; - - atomic_long_set(&pblk->nr_flush, 0); - pblk->nr_flush_rst = 0; - #ifdef CONFIG_NVM_DEBUG atomic_long_set(&pblk->inflight_writes, 0); atomic_long_set(&pblk->padded_writes, 0); @@ -1144,48 +1157,42 @@ static void *pblk_init(struct nvm_tgt_dev *dev, struct gendisk *tdisk, atomic_long_set(&pblk->write_failed, 0); atomic_long_set(&pblk->erase_failed, 0); + ret = pblk_luns_init(pblk, dev->luns); if (ret) { pr_err("pblk: could not initialize luns\n"); goto fail; } - ret = pblk_lines_init(pblk); - if (ret) { - pr_err("pblk: could not initialize lines\n"); - goto fail_free_luns; - } - - pblk->pad_dist = kzalloc((pblk->min_write_pgs - 1) * sizeof(atomic64_t), - GFP_KERNEL); - if (!pblk->pad_dist) { - ret = -ENOMEM; - goto fail_free_line_meta; - } - ret = pblk_core_init(pblk); if (ret) { pr_err("pblk: could not initialize core\n"); - goto fail_free_pad_dist; + goto fail_free_luns; } - ret = pblk_l2p_init(pblk); + ret = pblk_lines_init(pblk); if (ret) { - pr_err("pblk: could not initialize maps\n"); + pr_err("pblk: could not initialize lines\n"); goto fail_free_core; } - ret = pblk_lines_configure(pblk, flags); + ret = pblk_rwb_init(pblk); if (ret) { - pr_err("pblk: could not configure lines\n"); - goto fail_free_l2p; + pr_err("pblk: could not initialize write buffer\n"); + goto fail_free_lines; + } + + ret = pblk_l2p_init(pblk, flags & NVM_TARGET_FACTORY); + if (ret) { + pr_err("pblk: could not initialize maps\n"); + goto fail_free_rwb; } ret = pblk_writer_init(pblk); if (ret) { if (ret != -EINTR) pr_err("pblk: could not initialize write thread\n"); - goto fail_free_lines; + goto fail_free_l2p; } ret = pblk_gc_init(pblk); @@ -1220,16 +1227,14 @@ static void *pblk_init(struct nvm_tgt_dev *dev, struct gendisk *tdisk, fail_stop_writer: pblk_writer_stop(pblk); -fail_free_lines: - pblk_lines_free(pblk); fail_free_l2p: pblk_l2p_free(pblk); +fail_free_rwb: + pblk_rwb_free(pblk); +fail_free_lines: + pblk_lines_free(pblk); fail_free_core: pblk_core_free(pblk); -fail_free_pad_dist: - kfree(pblk->pad_dist); -fail_free_line_meta: - pblk_line_mg_free(pblk); fail_free_luns: pblk_luns_free(pblk); fail: -- 2.7.4