Received: by 2002:a05:7412:cfc7:b0:fc:a2b0:25d7 with SMTP id by7csp2096303rdb; Tue, 20 Feb 2024 17:55:26 -0800 (PST) X-Forwarded-Encrypted: i=3; AJvYcCWkXD2aiqgyn8A94+zsQTjJxTIpYSzb/beeWkac9erBqdwHeOkwiSBzSu985qG7jGQ1l+9E6ocLK+QwNBbxQVC153ULViVPPppvf35zqg== X-Google-Smtp-Source: AGHT+IHf6S90pLk4NJw0clkbl1rgmVnQPpFDq2JYvuQRDWzzjnMC9wWWL48Jy6pPnquQJojtH2/R X-Received: by 2002:a17:902:c1c4:b0:1db:e6a1:9e87 with SMTP id c4-20020a170902c1c400b001dbe6a19e87mr7267920plc.46.1708480526321; Tue, 20 Feb 2024 17:55:26 -0800 (PST) ARC-Seal: i=2; a=rsa-sha256; t=1708480526; cv=pass; d=google.com; s=arc-20160816; b=X/HLmXCNwPDAqKO5xfkDlAtpWkT9Rj7EgzX6q8wXa4+IVuxGOIS+0EPw9XOu80EYST HohudrIvH22WJ9Sug0Z8aGe1Jns5+6PvzlqAChQy+5rfj4yWBYZkxFSrTsYweLuH/kbp qK1OPkx9h+SNVFQd3BD8pb2DpM0k7mMubhTbFUf53Csuabcf49lZ5qxMX7raMgON5gMF M/xhGgs1mxE3yKRv0ABuMt53g2x4ymni2yFa+RA3mXKnlD9iYFoeQi8iLYsCkSAJ3BXm HMSkog6ZoeC1ZAr1dFRTMqE1Mvu+nPO13DgGI3iD87R3yNC0c5SAOxCX4FwXhkLAQI66 /JFQ== ARC-Message-Signature: i=2; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=in-reply-to:content-disposition:mime-version:list-unsubscribe :list-subscribe:list-id:precedence:references:message-id:subject:cc :to:from:date:dkim-signature; bh=iwfZBvCaNO9U2vBFLEJzWvQPS0djkKc63sis3fvmoHU=; fh=KuS+1UlV5gqHbJ37XPD8MCZ1v8KBm0nhjTGQI/haxiI=; b=QIV/LWmF2huPviW08jxNz3rTw/B8DTbc3iqCnruGceJog4EfWL3yVA99Ab67XeNjjT ELb9qWTIXbtYB75xiaURvaBCeum6VbwYlYyEpusZT+pJIQYxbcD+QPoDu6aQlOnQrKy2 axQZSfupPkK28EhQxtRz89/c/Jy3FHgDtkfyK4y8K3KlQOFzxXn+wJ3Mii3Zvit1DGyc h4HmftAo19W1xvfn2UWlow6I1r8iVqrLIO23KeeH87WE8q/ye8sDDRxsJtXXSIKG1vCB i5J7EYYe9ZR2kFGsIPAI7YAVqJ8OXLBpJPGn2ogeCkgBqPFcsrtgxJQXhvvf9vMH54xL yIDg==; dara=google.com ARC-Authentication-Results: i=2; mx.google.com; dkim=pass header.i=@kernel.org header.s=k20201202 header.b="IBCyuQ/x"; arc=pass (i=1 dkim=pass dkdomain=kernel.org); spf=pass (google.com: domain of linux-kernel+bounces-73919-linux.lists.archive=gmail.com@vger.kernel.org designates 2604:1380:40f1:3f00::1 as permitted sender) smtp.mailfrom="linux-kernel+bounces-73919-linux.lists.archive=gmail.com@vger.kernel.org"; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=kernel.org Return-Path: Received: from sy.mirrors.kernel.org (sy.mirrors.kernel.org. [2604:1380:40f1:3f00::1]) by mx.google.com with ESMTPS id z1-20020a170902708100b001db730d3f16si6975870plk.621.2024.02.20.17.55.25 for (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Tue, 20 Feb 2024 17:55:26 -0800 (PST) Received-SPF: pass (google.com: domain of linux-kernel+bounces-73919-linux.lists.archive=gmail.com@vger.kernel.org designates 2604:1380:40f1:3f00::1 as permitted sender) client-ip=2604:1380:40f1:3f00::1; Authentication-Results: mx.google.com; dkim=pass header.i=@kernel.org header.s=k20201202 header.b="IBCyuQ/x"; arc=pass (i=1 dkim=pass dkdomain=kernel.org); spf=pass (google.com: domain of linux-kernel+bounces-73919-linux.lists.archive=gmail.com@vger.kernel.org designates 2604:1380:40f1:3f00::1 as permitted sender) smtp.mailfrom="linux-kernel+bounces-73919-linux.lists.archive=gmail.com@vger.kernel.org"; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=kernel.org Received: from smtp.subspace.kernel.org (wormhole.subspace.kernel.org [52.25.139.140]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by sy.mirrors.kernel.org (Postfix) with ESMTPS id 6042EB238DB for ; Wed, 21 Feb 2024 01:55:12 +0000 (UTC) Received: from localhost.localdomain (localhost.localdomain [127.0.0.1]) by smtp.subspace.kernel.org (Postfix) with ESMTP id 1E5FB6FCA; Wed, 21 Feb 2024 01:55:00 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; dkim=pass (2048-bit key) header.d=kernel.org header.i=@kernel.org header.b="IBCyuQ/x" Received: from smtp.kernel.org (aws-us-west-2-korg-mail-1.web.codeaurora.org [10.30.226.201]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by smtp.subspace.kernel.org (Postfix) with ESMTPS id 7D5BE6AB9 for ; Wed, 21 Feb 2024 01:54:57 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=none smtp.client-ip=10.30.226.201 ARC-Seal:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1708480497; cv=none; b=HYUx4koT4qp0cboGjL9iA5CS0W4RTDigjhGZnVDy9JX5huw91Ns4Eq48cl07cEqv+aHSm3PeV3PfiUHq1Tvhm38e3Lww3PVNOXRwQYM2lWahfgeKoiDJo9TBdtkY3sydH/+0QbPiIzwlPTxzShC2vTsGY7iXkqfY8JaQ14rf35I= ARC-Message-Signature:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1708480497; c=relaxed/simple; bh=rE/WB4EKwYKjR3zbJN9nhSMok+Z6Yspc3k4u3ui3WVk=; h=Date:From:To:Cc:Subject:Message-ID:References:MIME-Version: Content-Type:Content-Disposition:In-Reply-To; b=B6+I/SJSd+7MNX/58qn+e79BF0kvD7feS1F6czTjLVvY0T2Opcj3v0Rr5e93hOO7NTAFfsT+ETZhq2c5VOP0uDOtBoIfmGbi1FLjaPFLfMg/KcCwh8Ls8S9mMw1gNiEYb3HTuxvtAe7me3+2gP2SwFVYKfQ0UiG2t7+IdUGkh2o= ARC-Authentication-Results:i=1; smtp.subspace.kernel.org; dkim=pass (2048-bit key) header.d=kernel.org header.i=@kernel.org header.b=IBCyuQ/x; arc=none smtp.client-ip=10.30.226.201 Received: by smtp.kernel.org (Postfix) with ESMTPSA id DF09CC43390; Wed, 21 Feb 2024 01:54:56 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=kernel.org; s=k20201202; t=1708480497; bh=rE/WB4EKwYKjR3zbJN9nhSMok+Z6Yspc3k4u3ui3WVk=; h=Date:From:To:Cc:Subject:References:In-Reply-To:From; b=IBCyuQ/x61sZ/e7Wwo0mp/YZ3zvOzqtIXXBJ+ygU8ZHB0NEiyFH2M4jIHb/l/CSeo cPzpb8upRsppkfSM7M9J/80sAVJbfHSpzhRuy+r8XUo6YGeRYH5W7j29YtkRRMPHHi xSCl0w4CIEUlblzMqcTAAB5+xYnKsYDX4F+yTJinxSZEGUDDU7i+Bi1NxfRGbeOzM9 4fupe76oT3hor9m+iOXRvawHLp9YSiQFVaX2rjOw0FVmlrsW4jIdAkK0s383KYn5Ok ftt613ufdHbSGSkdrSxk+wzC26OSVT8spRSAAQhcYyNgLWwxtejuCTCO/nTM4Nv2NE VvEHWv6bpIVtg== Date: Tue, 20 Feb 2024 17:54:55 -0800 From: Jaegeuk Kim To: Chao Yu Cc: linux-kernel@vger.kernel.org, linux-f2fs-devel@lists.sourceforge.net Subject: Re: [f2fs-dev] [PATCH 2/3 v3] f2fs: use BLKS_PER_SEG, BLKS_PER_SEC, and SEGS_PER_SEC Message-ID: References: <20240207005105.3744811-1-jaegeuk@kernel.org> <20240207005105.3744811-2-jaegeuk@kernel.org> <18ce2359-f2ee-4b33-8ba7-d6444d56be56@kernel.org> Precedence: bulk X-Mailing-List: linux-kernel@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <18ce2359-f2ee-4b33-8ba7-d6444d56be56@kernel.org> On 02/21, Chao Yu wrote: > On 2024/2/21 4:39, Jaegeuk Kim wrote: > > No functional change. > > > > Reviewed-by: Daeho Jeong > > Signed-off-by: Jaegeuk Kim > > --- > > Change log from v2: > > - cover more cases > > Change log from v1: > > - use div_u64 > > fs/f2fs/checkpoint.c | 10 ++-- > > fs/f2fs/debug.c | 6 +-- > > fs/f2fs/f2fs.h | 21 +++++---- > > fs/f2fs/file.c | 16 +++---- > > fs/f2fs/gc.c | 40 ++++++++-------- > > fs/f2fs/node.c | 4 +- > > fs/f2fs/node.h | 4 +- > > fs/f2fs/recovery.c | 2 +- > > fs/f2fs/segment.c | 110 +++++++++++++++++++++---------------------- > > fs/f2fs/segment.h | 44 +++++++++-------- > > fs/f2fs/super.c | 8 ++-- > > fs/f2fs/sysfs.c | 6 +-- > > 12 files changed, 135 insertions(+), 136 deletions(-) > > > > diff --git a/fs/f2fs/checkpoint.c b/fs/f2fs/checkpoint.c > > index b85820e70f5e..a09a9609e228 100644 > > --- a/fs/f2fs/checkpoint.c > > +++ b/fs/f2fs/checkpoint.c > > @@ -900,7 +900,7 @@ static struct page *validate_checkpoint(struct f2fs_sb_info *sbi, > > cp_blocks = le32_to_cpu(cp_block->cp_pack_total_block_count); > > - if (cp_blocks > sbi->blocks_per_seg || cp_blocks <= F2FS_CP_PACKS) { > > + if (cp_blocks > BLKS_PER_SEG(sbi) || cp_blocks <= F2FS_CP_PACKS) { > > f2fs_warn(sbi, "invalid cp_pack_total_block_count:%u", > > le32_to_cpu(cp_block->cp_pack_total_block_count)); > > goto invalid_cp; > > @@ -1335,7 +1335,7 @@ static void update_ckpt_flags(struct f2fs_sb_info *sbi, struct cp_control *cpc) > > if (cpc->reason & CP_UMOUNT) { > > if (le32_to_cpu(ckpt->cp_pack_total_block_count) + > > - NM_I(sbi)->nat_bits_blocks > sbi->blocks_per_seg) { > > + NM_I(sbi)->nat_bits_blocks > BLKS_PER_SEG(sbi)) { > > clear_ckpt_flags(sbi, CP_NAT_BITS_FLAG); > > f2fs_notice(sbi, "Disable nat_bits due to no space"); > > } else if (!is_set_ckpt_flags(sbi, CP_NAT_BITS_FLAG) && > > @@ -1538,7 +1538,7 @@ static int do_checkpoint(struct f2fs_sb_info *sbi, struct cp_control *cpc) > > cp_ver |= ((__u64)crc32 << 32); > > *(__le64 *)nm_i->nat_bits = cpu_to_le64(cp_ver); > > - blk = start_blk + sbi->blocks_per_seg - nm_i->nat_bits_blocks; > > + blk = start_blk + BLKS_PER_SEG(sbi) - nm_i->nat_bits_blocks; > > for (i = 0; i < nm_i->nat_bits_blocks; i++) > > f2fs_update_meta_page(sbi, nm_i->nat_bits + > > (i << F2FS_BLKSIZE_BITS), blk + i); > > @@ -1741,9 +1741,9 @@ void f2fs_init_ino_entry_info(struct f2fs_sb_info *sbi) > > im->ino_num = 0; > > } > > - sbi->max_orphans = (sbi->blocks_per_seg - F2FS_CP_PACKS - > > + sbi->max_orphans = (BLKS_PER_SEG(sbi) - F2FS_CP_PACKS - > > NR_CURSEG_PERSIST_TYPE - __cp_payload(sbi)) * > > - F2FS_ORPHANS_PER_BLOCK; > > + F2FS_ORPHANS_PER_BLOCK; > > } > > int __init f2fs_create_checkpoint_caches(void) > > diff --git a/fs/f2fs/debug.c b/fs/f2fs/debug.c > > index fdbf994f1271..0d02224b99b7 100644 > > --- a/fs/f2fs/debug.c > > +++ b/fs/f2fs/debug.c > > @@ -41,7 +41,7 @@ void f2fs_update_sit_info(struct f2fs_sb_info *sbi) > > total_vblocks = 0; > > blks_per_sec = CAP_BLKS_PER_SEC(sbi); > > hblks_per_sec = blks_per_sec / 2; > > - for (segno = 0; segno < MAIN_SEGS(sbi); segno += sbi->segs_per_sec) { > > + for (segno = 0; segno < MAIN_SEGS(sbi); segno += SEGS_PER_SEC(sbi)) { > > vblocks = get_valid_blocks(sbi, segno, true); > > dist = abs(vblocks - hblks_per_sec); > > bimodal += dist * dist; > > @@ -135,7 +135,7 @@ static void update_general_status(struct f2fs_sb_info *sbi) > > si->cur_ckpt_time = sbi->cprc_info.cur_time; > > si->peak_ckpt_time = sbi->cprc_info.peak_time; > > spin_unlock(&sbi->cprc_info.stat_lock); > > - si->total_count = (int)sbi->user_block_count / sbi->blocks_per_seg; > > + si->total_count = (int)sbi->user_block_count / BLKS_PER_SEG(sbi); > > si->rsvd_segs = reserved_segments(sbi); > > si->overp_segs = overprovision_segments(sbi); > > si->valid_count = valid_user_blocks(sbi); > > @@ -208,7 +208,7 @@ static void update_general_status(struct f2fs_sb_info *sbi) > > if (!blks) > > continue; > > - if (blks == sbi->blocks_per_seg) > > + if (blks == BLKS_PER_SEG(sbi)) > > si->full_seg[type]++; > > else > > si->dirty_seg[type]++; > > diff --git a/fs/f2fs/f2fs.h b/fs/f2fs/f2fs.h > > index 4c52136cbc10..50e7890cc6a5 100644 > > --- a/fs/f2fs/f2fs.h > > +++ b/fs/f2fs/f2fs.h > > @@ -1813,6 +1813,14 @@ struct f2fs_sb_info { > > #endif > > }; > > +/* Definitions to access f2fs_sb_info */ > > +#define BLKS_PER_SEG(sbi) \ > > + ((sbi)->blocks_per_seg) > > +#define BLKS_PER_SEC(sbi) \ > > + ((sbi)->segs_per_sec << (sbi)->log_blocks_per_seg) > > +#define SEGS_PER_SEC(sbi) \ > > + ((sbi)->segs_per_sec) > > I can see lots of 'xxx << sbi->log_blocks_per_seg', how about introducing > a new macro to wrap it, maybe: > > #define SEGMENTS_TO_BLKS(sbi, segments) \ > (segments << (sbi)->log_blocks_per_seg) Can you write up in a separate patch, since this is already quite huge? > > Thanks, > > > + > > __printf(3, 4) > > void f2fs_printk(struct f2fs_sb_info *sbi, bool limit_rate, const char *fmt, ...); > > @@ -2511,11 +2519,8 @@ static inline int get_dirty_pages(struct inode *inode) > > static inline int get_blocktype_secs(struct f2fs_sb_info *sbi, int block_type) > > { > > - unsigned int pages_per_sec = sbi->segs_per_sec * sbi->blocks_per_seg; > > - unsigned int segs = (get_pages(sbi, block_type) + pages_per_sec - 1) >> > > - sbi->log_blocks_per_seg; > > - > > - return segs / sbi->segs_per_sec; > > + return div_u64(get_pages(sbi, block_type) + BLKS_PER_SEC(sbi) - 1, > > + BLKS_PER_SEC(sbi)); > > } > > static inline block_t valid_user_blocks(struct f2fs_sb_info *sbi) > > @@ -2579,7 +2584,7 @@ static inline block_t __start_cp_addr(struct f2fs_sb_info *sbi) > > block_t start_addr = le32_to_cpu(F2FS_RAW_SUPER(sbi)->cp_blkaddr); > > if (sbi->cur_cp_pack == 2) > > - start_addr += sbi->blocks_per_seg; > > + start_addr += BLKS_PER_SEG(sbi); > > return start_addr; > > } > > @@ -2588,7 +2593,7 @@ static inline block_t __start_cp_next_addr(struct f2fs_sb_info *sbi) > > block_t start_addr = le32_to_cpu(F2FS_RAW_SUPER(sbi)->cp_blkaddr); > > if (sbi->cur_cp_pack == 1) > > - start_addr += sbi->blocks_per_seg; > > + start_addr += BLKS_PER_SEG(sbi); > > return start_addr; > > } > > @@ -3458,7 +3463,7 @@ static inline __le32 *get_dnode_addr(struct inode *inode, > > sizeof((f2fs_inode)->field)) \ > > <= (F2FS_OLD_ATTRIBUTE_SIZE + (extra_isize))) \ > > -#define __is_large_section(sbi) ((sbi)->segs_per_sec > 1) > > +#define __is_large_section(sbi) (SEGS_PER_SEC(sbi) > 1) > > #define __is_meta_io(fio) (PAGE_TYPE_OF_BIO((fio)->type) == META) > > diff --git a/fs/f2fs/file.c b/fs/f2fs/file.c > > index c6cd9474ba2d..767d16c74bb6 100644 > > --- a/fs/f2fs/file.c > > +++ b/fs/f2fs/file.c > > @@ -2580,7 +2580,6 @@ static int f2fs_defragment_range(struct f2fs_sb_info *sbi, > > .m_may_create = false }; > > struct extent_info ei = {}; > > pgoff_t pg_start, pg_end, next_pgofs; > > - unsigned int blk_per_seg = sbi->blocks_per_seg; > > unsigned int total = 0, sec_num; > > block_t blk_end = 0; > > bool fragmented = false; > > @@ -2689,7 +2688,8 @@ static int f2fs_defragment_range(struct f2fs_sb_info *sbi, > > set_inode_flag(inode, FI_SKIP_WRITES); > > idx = map.m_lblk; > > - while (idx < map.m_lblk + map.m_len && cnt < blk_per_seg) { > > + while (idx < map.m_lblk + map.m_len && > > + cnt < BLKS_PER_SEG(sbi)) { > > struct page *page; > > page = f2fs_get_lock_data_page(inode, idx, true); > > @@ -2709,7 +2709,7 @@ static int f2fs_defragment_range(struct f2fs_sb_info *sbi, > > map.m_lblk = idx; > > check: > > - if (map.m_lblk < pg_end && cnt < blk_per_seg) > > + if (map.m_lblk < pg_end && cnt < BLKS_PER_SEG(sbi)) > > goto do_map; > > clear_inode_flag(inode, FI_SKIP_WRITES); > > @@ -2978,8 +2978,8 @@ static int f2fs_ioc_flush_device(struct file *filp, unsigned long arg) > > if (!f2fs_is_multi_device(sbi) || sbi->s_ndevs - 1 <= range.dev_num || > > __is_large_section(sbi)) { > > - f2fs_warn(sbi, "Can't flush %u in %d for segs_per_sec %u != 1", > > - range.dev_num, sbi->s_ndevs, sbi->segs_per_sec); > > + f2fs_warn(sbi, "Can't flush %u in %d for SEGS_PER_SEC %u != 1", > > + range.dev_num, sbi->s_ndevs, SEGS_PER_SEC(sbi)); > > return -EINVAL; > > } > > @@ -4081,7 +4081,6 @@ static int f2fs_ioc_decompress_file(struct file *filp) > > struct f2fs_sb_info *sbi = F2FS_I_SB(inode); > > struct f2fs_inode_info *fi = F2FS_I(inode); > > pgoff_t page_idx = 0, last_idx; > > - unsigned int blk_per_seg = sbi->blocks_per_seg; > > int cluster_size = fi->i_cluster_size; > > int count, ret; > > @@ -4125,7 +4124,7 @@ static int f2fs_ioc_decompress_file(struct file *filp) > > if (ret < 0) > > break; > > - if (get_dirty_pages(inode) >= blk_per_seg) { > > + if (get_dirty_pages(inode) >= BLKS_PER_SEG(sbi)) { > > ret = filemap_fdatawrite(inode->i_mapping); > > if (ret < 0) > > break; > > @@ -4160,7 +4159,6 @@ static int f2fs_ioc_compress_file(struct file *filp) > > struct inode *inode = file_inode(filp); > > struct f2fs_sb_info *sbi = F2FS_I_SB(inode); > > pgoff_t page_idx = 0, last_idx; > > - unsigned int blk_per_seg = sbi->blocks_per_seg; > > int cluster_size = F2FS_I(inode)->i_cluster_size; > > int count, ret; > > @@ -4203,7 +4201,7 @@ static int f2fs_ioc_compress_file(struct file *filp) > > if (ret < 0) > > break; > > - if (get_dirty_pages(inode) >= blk_per_seg) { > > + if (get_dirty_pages(inode) >= BLKS_PER_SEG(sbi)) { > > ret = filemap_fdatawrite(inode->i_mapping); > > if (ret < 0) > > break; > > diff --git a/fs/f2fs/gc.c b/fs/f2fs/gc.c > > index 6899f434ad68..d61a60c1c844 100644 > > --- a/fs/f2fs/gc.c > > +++ b/fs/f2fs/gc.c > > @@ -259,7 +259,7 @@ static void select_policy(struct f2fs_sb_info *sbi, int gc_type, > > p->ofs_unit = 1; > > } else { > > p->gc_mode = select_gc_type(sbi, gc_type); > > - p->ofs_unit = sbi->segs_per_sec; > > + p->ofs_unit = SEGS_PER_SEC(sbi); > > if (__is_large_section(sbi)) { > > p->dirty_bitmap = dirty_i->dirty_secmap; > > p->max_search = count_bits(p->dirty_bitmap, > > @@ -282,7 +282,8 @@ static void select_policy(struct f2fs_sb_info *sbi, int gc_type, > > /* let's select beginning hot/small space first in no_heap mode*/ > > if (f2fs_need_rand_seg(sbi)) > > - p->offset = get_random_u32_below(MAIN_SECS(sbi) * sbi->segs_per_sec); > > + p->offset = get_random_u32_below(MAIN_SECS(sbi) * > > + SEGS_PER_SEC(sbi)); > > else if (test_opt(sbi, NOHEAP) && > > (type == CURSEG_HOT_DATA || IS_NODESEG(type))) > > p->offset = 0; > > @@ -295,13 +296,13 @@ static unsigned int get_max_cost(struct f2fs_sb_info *sbi, > > { > > /* SSR allocates in a segment unit */ > > if (p->alloc_mode == SSR) > > - return sbi->blocks_per_seg; > > + return BLKS_PER_SEG(sbi); > > else if (p->alloc_mode == AT_SSR) > > return UINT_MAX; > > /* LFS */ > > if (p->gc_mode == GC_GREEDY) > > - return 2 * sbi->blocks_per_seg * p->ofs_unit; > > + return 2 * BLKS_PER_SEG(sbi) * p->ofs_unit; > > else if (p->gc_mode == GC_CB) > > return UINT_MAX; > > else if (p->gc_mode == GC_AT) > > @@ -496,9 +497,9 @@ static void add_victim_entry(struct f2fs_sb_info *sbi, > > return; > > } > > - for (i = 0; i < sbi->segs_per_sec; i++) > > + for (i = 0; i < SEGS_PER_SEC(sbi); i++) > > mtime += get_seg_entry(sbi, start + i)->mtime; > > - mtime = div_u64(mtime, sbi->segs_per_sec); > > + mtime = div_u64(mtime, SEGS_PER_SEC(sbi)); > > /* Handle if the system time has changed by the user */ > > if (mtime < sit_i->min_mtime) > > @@ -599,7 +600,6 @@ static void atssr_lookup_victim(struct f2fs_sb_info *sbi, > > unsigned long long age; > > unsigned long long max_mtime = sit_i->dirty_max_mtime; > > unsigned long long min_mtime = sit_i->dirty_min_mtime; > > - unsigned int seg_blocks = sbi->blocks_per_seg; > > unsigned int vblocks; > > unsigned int dirty_threshold = max(am->max_candidate_count, > > am->candidate_ratio * > > @@ -629,7 +629,7 @@ static void atssr_lookup_victim(struct f2fs_sb_info *sbi, > > f2fs_bug_on(sbi, !vblocks); > > /* rare case */ > > - if (vblocks == seg_blocks) > > + if (vblocks == BLKS_PER_SEG(sbi)) > > goto skip_node; > > iter++; > > @@ -755,7 +755,7 @@ int f2fs_get_victim(struct f2fs_sb_info *sbi, unsigned int *result, > > int ret = 0; > > mutex_lock(&dirty_i->seglist_lock); > > - last_segment = MAIN_SECS(sbi) * sbi->segs_per_sec; > > + last_segment = MAIN_SECS(sbi) * SEGS_PER_SEC(sbi); > > p.alloc_mode = alloc_mode; > > p.age = age; > > @@ -896,7 +896,7 @@ int f2fs_get_victim(struct f2fs_sb_info *sbi, unsigned int *result, > > else > > sm->last_victim[p.gc_mode] = segno + p.ofs_unit; > > sm->last_victim[p.gc_mode] %= > > - (MAIN_SECS(sbi) * sbi->segs_per_sec); > > + (MAIN_SECS(sbi) * SEGS_PER_SEC(sbi)); > > break; > > } > > } > > @@ -1670,7 +1670,7 @@ static int do_garbage_collect(struct f2fs_sb_info *sbi, > > struct f2fs_summary_block *sum; > > struct blk_plug plug; > > unsigned int segno = start_segno; > > - unsigned int end_segno = start_segno + sbi->segs_per_sec; > > + unsigned int end_segno = start_segno + SEGS_PER_SEC(sbi); > > int seg_freed = 0, migrated = 0; > > unsigned char type = IS_DATASEG(get_seg_entry(sbi, segno)->type) ? > > SUM_TYPE_DATA : SUM_TYPE_NODE; > > @@ -1678,7 +1678,7 @@ static int do_garbage_collect(struct f2fs_sb_info *sbi, > > int submitted = 0; > > if (__is_large_section(sbi)) > > - end_segno = rounddown(end_segno, sbi->segs_per_sec); > > + end_segno = rounddown(end_segno, SEGS_PER_SEC(sbi)); > > /* > > * zone-capacity can be less than zone-size in zoned devices, > > @@ -1686,7 +1686,7 @@ static int do_garbage_collect(struct f2fs_sb_info *sbi, > > * calculate the end segno in the zone which can be garbage collected > > */ > > if (f2fs_sb_has_blkzoned(sbi)) > > - end_segno -= sbi->segs_per_sec - > > + end_segno -= SEGS_PER_SEC(sbi) - > > f2fs_usable_segs_in_sec(sbi, segno); > > sanity_check_seg_type(sbi, get_seg_entry(sbi, segno)->type); > > @@ -1986,7 +1986,7 @@ static int free_segment_range(struct f2fs_sb_info *sbi, > > /* Force block allocation for GC */ > > MAIN_SECS(sbi) -= secs; > > - start = MAIN_SECS(sbi) * sbi->segs_per_sec; > > + start = MAIN_SECS(sbi) * SEGS_PER_SEC(sbi); > > end = MAIN_SEGS(sbi) - 1; > > mutex_lock(&DIRTY_I(sbi)->seglist_lock); > > @@ -2004,7 +2004,7 @@ static int free_segment_range(struct f2fs_sb_info *sbi, > > f2fs_allocate_segment_for_resize(sbi, type, start, end); > > /* do GC to move out valid blocks in the range */ > > - for (segno = start; segno <= end; segno += sbi->segs_per_sec) { > > + for (segno = start; segno <= end; segno += SEGS_PER_SEC(sbi)) { > > struct gc_inode_list gc_list = { > > .ilist = LIST_HEAD_INIT(gc_list.ilist), > > .iroot = RADIX_TREE_INIT(gc_list.iroot, GFP_NOFS), > > @@ -2048,7 +2048,7 @@ static void update_sb_metadata(struct f2fs_sb_info *sbi, int secs) > > int segment_count; > > int segment_count_main; > > long long block_count; > > - int segs = secs * sbi->segs_per_sec; > > + int segs = secs * SEGS_PER_SEC(sbi); > > f2fs_down_write(&sbi->sb_lock); > > @@ -2061,7 +2061,7 @@ static void update_sb_metadata(struct f2fs_sb_info *sbi, int secs) > > raw_sb->segment_count = cpu_to_le32(segment_count + segs); > > raw_sb->segment_count_main = cpu_to_le32(segment_count_main + segs); > > raw_sb->block_count = cpu_to_le64(block_count + > > - (long long)segs * sbi->blocks_per_seg); > > + (long long)(segs << sbi->log_blocks_per_seg)); > > if (f2fs_is_multi_device(sbi)) { > > int last_dev = sbi->s_ndevs - 1; > > int dev_segs = > > @@ -2076,8 +2076,8 @@ static void update_sb_metadata(struct f2fs_sb_info *sbi, int secs) > > static void update_fs_metadata(struct f2fs_sb_info *sbi, int secs) > > { > > - int segs = secs * sbi->segs_per_sec; > > - long long blks = (long long)segs * sbi->blocks_per_seg; > > + int segs = secs * SEGS_PER_SEC(sbi); > > + long long blks = (long long)(segs << sbi->log_blocks_per_seg); > > long long user_block_count = > > le64_to_cpu(F2FS_CKPT(sbi)->user_block_count); > > @@ -2119,7 +2119,7 @@ int f2fs_resize_fs(struct file *filp, __u64 block_count) > > int last_dev = sbi->s_ndevs - 1; > > __u64 last_segs = FDEV(last_dev).total_segments; > > - if (block_count + last_segs * sbi->blocks_per_seg <= > > + if (block_count + (last_segs << sbi->log_blocks_per_seg) <= > > old_block_count) > > return -EINVAL; > > } > > diff --git a/fs/f2fs/node.c b/fs/f2fs/node.c > > index 1d898a16f05a..51241996b9ec 100644 > > --- a/fs/f2fs/node.c > > +++ b/fs/f2fs/node.c > > @@ -2841,7 +2841,7 @@ int f2fs_restore_node_summary(struct f2fs_sb_info *sbi, > > int i, idx, last_offset, nrpages; > > /* scan the node segment */ > > - last_offset = sbi->blocks_per_seg; > > + last_offset = BLKS_PER_SEG(sbi); > > addr = START_BLOCK(sbi, segno); > > sum_entry = &sum->entries[0]; > > @@ -3158,7 +3158,7 @@ static int __get_nat_bitmaps(struct f2fs_sb_info *sbi) > > if (!is_set_ckpt_flags(sbi, CP_NAT_BITS_FLAG)) > > return 0; > > - nat_bits_addr = __start_cp_addr(sbi) + sbi->blocks_per_seg - > > + nat_bits_addr = __start_cp_addr(sbi) + BLKS_PER_SEG(sbi) - > > nm_i->nat_bits_blocks; > > for (i = 0; i < nm_i->nat_bits_blocks; i++) { > > struct page *page; > > diff --git a/fs/f2fs/node.h b/fs/f2fs/node.h > > index 5bd16a95eef8..6aea13024ac1 100644 > > --- a/fs/f2fs/node.h > > +++ b/fs/f2fs/node.h > > @@ -208,10 +208,10 @@ static inline pgoff_t current_nat_addr(struct f2fs_sb_info *sbi, nid_t start) > > block_addr = (pgoff_t)(nm_i->nat_blkaddr + > > (block_off << 1) - > > - (block_off & (sbi->blocks_per_seg - 1))); > > + (block_off & (BLKS_PER_SEG(sbi) - 1))); > > if (f2fs_test_bit(block_off, nm_i->nat_bitmap)) > > - block_addr += sbi->blocks_per_seg; > > + block_addr += BLKS_PER_SEG(sbi); > > return block_addr; > > } > > diff --git a/fs/f2fs/recovery.c b/fs/f2fs/recovery.c > > index aad1d1a9b3d6..b3baec666afe 100644 > > --- a/fs/f2fs/recovery.c > > +++ b/fs/f2fs/recovery.c > > @@ -354,7 +354,7 @@ static unsigned int adjust_por_ra_blocks(struct f2fs_sb_info *sbi, > > if (blkaddr + 1 == next_blkaddr) > > ra_blocks = min_t(unsigned int, RECOVERY_MAX_RA_BLOCKS, > > ra_blocks * 2); > > - else if (next_blkaddr % sbi->blocks_per_seg) > > + else if (next_blkaddr % BLKS_PER_SEG(sbi)) > > ra_blocks = max_t(unsigned int, RECOVERY_MIN_RA_BLOCKS, > > ra_blocks / 2); > > return ra_blocks; > > diff --git a/fs/f2fs/segment.c b/fs/f2fs/segment.c > > index e5759813276a..1518f1287c28 100644 > > --- a/fs/f2fs/segment.c > > +++ b/fs/f2fs/segment.c > > @@ -448,8 +448,8 @@ static inline bool excess_dirty_threshold(struct f2fs_sb_info *sbi) > > unsigned int nodes = get_pages(sbi, F2FS_DIRTY_NODES); > > unsigned int meta = get_pages(sbi, F2FS_DIRTY_META); > > unsigned int imeta = get_pages(sbi, F2FS_DIRTY_IMETA); > > - unsigned int threshold = sbi->blocks_per_seg * factor * > > - DEFAULT_DIRTY_THRESHOLD; > > + unsigned int threshold = (factor * DEFAULT_DIRTY_THRESHOLD) << > > + sbi->log_blocks_per_seg; > > unsigned int global_threshold = threshold * 3 / 2; > > if (dents >= threshold || qdata >= threshold || > > @@ -1134,8 +1134,7 @@ static void __check_sit_bitmap(struct f2fs_sb_info *sbi, > > struct seg_entry *sentry; > > unsigned int segno; > > block_t blk = start; > > - unsigned long offset, size, max_blocks = sbi->blocks_per_seg; > > - unsigned long *map; > > + unsigned long offset, size, *map; > > while (blk < end) { > > segno = GET_SEGNO(sbi, blk); > > @@ -1145,7 +1144,7 @@ static void __check_sit_bitmap(struct f2fs_sb_info *sbi, > > if (end < START_BLOCK(sbi, segno + 1)) > > size = GET_BLKOFF_FROM_SEG0(sbi, end); > > else > > - size = max_blocks; > > + size = BLKS_PER_SEG(sbi); > > map = (unsigned long *)(sentry->cur_valid_map); > > offset = __find_rev_next_bit(map, size, offset); > > f2fs_bug_on(sbi, offset != size); > > @@ -2044,7 +2043,6 @@ static bool add_discard_addrs(struct f2fs_sb_info *sbi, struct cp_control *cpc, > > bool check_only) > > { > > int entries = SIT_VBLOCK_MAP_SIZE / sizeof(unsigned long); > > - int max_blocks = sbi->blocks_per_seg; > > struct seg_entry *se = get_seg_entry(sbi, cpc->trim_start); > > unsigned long *cur_map = (unsigned long *)se->cur_valid_map; > > unsigned long *ckpt_map = (unsigned long *)se->ckpt_valid_map; > > @@ -2056,8 +2054,9 @@ static bool add_discard_addrs(struct f2fs_sb_info *sbi, struct cp_control *cpc, > > struct list_head *head = &SM_I(sbi)->dcc_info->entry_list; > > int i; > > - if (se->valid_blocks == max_blocks || !f2fs_hw_support_discard(sbi) || > > - !f2fs_block_unit_discard(sbi)) > > + if (se->valid_blocks == BLKS_PER_SEG(sbi) || > > + !f2fs_hw_support_discard(sbi) || > > + !f2fs_block_unit_discard(sbi)) > > return false; > > if (!force) { > > @@ -2074,13 +2073,14 @@ static bool add_discard_addrs(struct f2fs_sb_info *sbi, struct cp_control *cpc, > > while (force || SM_I(sbi)->dcc_info->nr_discards <= > > SM_I(sbi)->dcc_info->max_discards) { > > - start = __find_rev_next_bit(dmap, max_blocks, end + 1); > > - if (start >= max_blocks) > > + start = __find_rev_next_bit(dmap, BLKS_PER_SEG(sbi), end + 1); > > + if (start >= BLKS_PER_SEG(sbi)) > > break; > > - end = __find_rev_next_zero_bit(dmap, max_blocks, start + 1); > > - if (force && start && end != max_blocks > > - && (end - start) < cpc->trim_minlen) > > + end = __find_rev_next_zero_bit(dmap, > > + BLKS_PER_SEG(sbi), start + 1); > > + if (force && start && end != BLKS_PER_SEG(sbi) && > > + (end - start) < cpc->trim_minlen) > > continue; > > if (check_only) > > @@ -2162,8 +2162,8 @@ void f2fs_clear_prefree_segments(struct f2fs_sb_info *sbi, > > start + 1); > > if (section_alignment) { > > - start = rounddown(start, sbi->segs_per_sec); > > - end = roundup(end, sbi->segs_per_sec); > > + start = rounddown(start, SEGS_PER_SEC(sbi)); > > + end = roundup(end, SEGS_PER_SEC(sbi)); > > } > > for (i = start; i < end; i++) { > > @@ -2191,9 +2191,9 @@ void f2fs_clear_prefree_segments(struct f2fs_sb_info *sbi, > > if (!IS_CURSEC(sbi, secno) && > > !get_valid_blocks(sbi, start, true)) > > f2fs_issue_discard(sbi, START_BLOCK(sbi, start_segno), > > - sbi->segs_per_sec << sbi->log_blocks_per_seg); > > + BLKS_PER_SEC(sbi)); > > - start = start_segno + sbi->segs_per_sec; > > + start = start_segno + SEGS_PER_SEC(sbi); > > if (start < end) > > goto next; > > else > > @@ -2212,7 +2212,7 @@ void f2fs_clear_prefree_segments(struct f2fs_sb_info *sbi, > > find_next: > > if (is_valid) { > > next_pos = find_next_zero_bit_le(entry->discard_map, > > - sbi->blocks_per_seg, cur_pos); > > + BLKS_PER_SEG(sbi), cur_pos); > > len = next_pos - cur_pos; > > if (f2fs_sb_has_blkzoned(sbi) || > > @@ -2224,13 +2224,13 @@ void f2fs_clear_prefree_segments(struct f2fs_sb_info *sbi, > > total_len += len; > > } else { > > next_pos = find_next_bit_le(entry->discard_map, > > - sbi->blocks_per_seg, cur_pos); > > + BLKS_PER_SEG(sbi), cur_pos); > > } > > skip: > > cur_pos = next_pos; > > is_valid = !is_valid; > > - if (cur_pos < sbi->blocks_per_seg) > > + if (cur_pos < BLKS_PER_SEG(sbi)) > > goto find_next; > > release_discard_addr(entry); > > @@ -2279,7 +2279,7 @@ static int create_discard_cmd_control(struct f2fs_sb_info *sbi) > > dcc->max_ordered_discard = DEFAULT_MAX_ORDERED_DISCARD_GRANULARITY; > > dcc->discard_io_aware = DPOLICY_IO_AWARE_ENABLE; > > if (F2FS_OPTION(sbi).discard_unit == DISCARD_UNIT_SEGMENT) > > - dcc->discard_granularity = sbi->blocks_per_seg; > > + dcc->discard_granularity = BLKS_PER_SEG(sbi); > > else if (F2FS_OPTION(sbi).discard_unit == DISCARD_UNIT_SECTION) > > dcc->discard_granularity = BLKS_PER_SEC(sbi); > > @@ -2542,7 +2542,7 @@ static unsigned short f2fs_curseg_valid_blocks(struct f2fs_sb_info *sbi, int typ > > struct curseg_info *curseg = CURSEG_I(sbi, type); > > if (sbi->ckpt->alloc_type[type] == SSR) > > - return sbi->blocks_per_seg; > > + return BLKS_PER_SEG(sbi); > > return curseg->next_blkoff; > > } > > @@ -2630,7 +2630,7 @@ static int is_next_segment_free(struct f2fs_sb_info *sbi, > > unsigned int segno = curseg->segno + 1; > > struct free_segmap_info *free_i = FREE_I(sbi); > > - if (segno < MAIN_SEGS(sbi) && segno % sbi->segs_per_sec) > > + if (segno < MAIN_SEGS(sbi) && segno % SEGS_PER_SEC(sbi)) > > return !test_bit(segno, free_i->free_segmap); > > return 0; > > } > > @@ -2654,7 +2654,7 @@ static void get_new_segment(struct f2fs_sb_info *sbi, > > spin_lock(&free_i->segmap_lock); > > - if (!new_sec && ((*newseg + 1) % sbi->segs_per_sec)) { > > + if (!new_sec && ((*newseg + 1) % SEGS_PER_SEC(sbi))) { > > segno = find_next_zero_bit(free_i->free_segmap, > > GET_SEG_FROM_SEC(sbi, hint + 1), *newseg + 1); > > if (segno < GET_SEG_FROM_SEC(sbi, hint + 1)) > > @@ -2757,9 +2757,8 @@ static unsigned int __get_next_segno(struct f2fs_sb_info *sbi, int type) > > sanity_check_seg_type(sbi, seg_type); > > if (f2fs_need_rand_seg(sbi)) > > - return get_random_u32_below(MAIN_SECS(sbi) * sbi->segs_per_sec); > > + return get_random_u32_below(MAIN_SECS(sbi) * SEGS_PER_SEC(sbi)); > > - /* if segs_per_sec is large than 1, we need to keep original policy. */ > > if (__is_large_section(sbi)) > > return curseg->segno; > > @@ -2827,7 +2826,7 @@ static int __next_free_blkoff(struct f2fs_sb_info *sbi, > > for (i = 0; i < entries; i++) > > target_map[i] = ckpt_map[i] | cur_map[i]; > > - return __find_rev_next_zero_bit(target_map, sbi->blocks_per_seg, start); > > + return __find_rev_next_zero_bit(target_map, BLKS_PER_SEG(sbi), start); > > } > > static int f2fs_find_next_ssr_block(struct f2fs_sb_info *sbi, > > @@ -2838,7 +2837,7 @@ static int f2fs_find_next_ssr_block(struct f2fs_sb_info *sbi, > > bool f2fs_segment_has_free_slot(struct f2fs_sb_info *sbi, int segno) > > { > > - return __next_free_blkoff(sbi, segno, 0) < sbi->blocks_per_seg; > > + return __next_free_blkoff(sbi, segno, 0) < BLKS_PER_SEG(sbi); > > } > > /* > > @@ -3238,8 +3237,8 @@ int f2fs_trim_fs(struct f2fs_sb_info *sbi, struct fstrim_range *range) > > end_segno = (end >= MAX_BLKADDR(sbi)) ? MAIN_SEGS(sbi) - 1 : > > GET_SEGNO(sbi, end); > > if (need_align) { > > - start_segno = rounddown(start_segno, sbi->segs_per_sec); > > - end_segno = roundup(end_segno + 1, sbi->segs_per_sec) - 1; > > + start_segno = rounddown(start_segno, SEGS_PER_SEC(sbi)); > > + end_segno = roundup(end_segno + 1, SEGS_PER_SEC(sbi)) - 1; > > } > > cpc.reason = CP_DISCARD; > > @@ -3437,7 +3436,7 @@ void f2fs_allocate_data_block(struct f2fs_sb_info *sbi, struct page *page, > > } > > *new_blkaddr = NEXT_FREE_BLKADDR(sbi, curseg); > > - f2fs_bug_on(sbi, curseg->next_blkoff >= sbi->blocks_per_seg); > > + f2fs_bug_on(sbi, curseg->next_blkoff >= BLKS_PER_SEG(sbi)); > > f2fs_wait_discard_bio(sbi, *new_blkaddr); > > @@ -3881,7 +3880,7 @@ static int read_compacted_summaries(struct f2fs_sb_info *sbi) > > seg_i->next_blkoff = blk_off; > > if (seg_i->alloc_type == SSR) > > - blk_off = sbi->blocks_per_seg; > > + blk_off = BLKS_PER_SEG(sbi); > > for (j = 0; j < blk_off; j++) { > > struct f2fs_summary *s; > > @@ -3949,7 +3948,7 @@ static int read_normal_summaries(struct f2fs_sb_info *sbi, int type) > > struct f2fs_summary *ns = &sum->entries[0]; > > int i; > > - for (i = 0; i < sbi->blocks_per_seg; i++, ns++) { > > + for (i = 0; i < BLKS_PER_SEG(sbi); i++, ns++) { > > ns->version = 0; > > ns->ofs_in_node = 0; > > } > > @@ -4582,21 +4581,20 @@ static int build_sit_entries(struct f2fs_sb_info *sbi) > > sit_valid_blocks[SE_PAGETYPE(se)] += se->valid_blocks; > > - if (f2fs_block_unit_discard(sbi)) { > > - /* build discard map only one time */ > > - if (is_set_ckpt_flags(sbi, CP_TRIMMED_FLAG)) { > > - memset(se->discard_map, 0xff, > > + if (!f2fs_block_unit_discard(sbi)) > > + goto init_discard_map_done; > > + > > + /* build discard map only one time */ > > + if (is_set_ckpt_flags(sbi, CP_TRIMMED_FLAG)) { > > + memset(se->discard_map, 0xff, > > SIT_VBLOCK_MAP_SIZE); > > - } else { > > - memcpy(se->discard_map, > > - se->cur_valid_map, > > + goto init_discard_map_done; > > + } > > + memcpy(se->discard_map, se->cur_valid_map, > > SIT_VBLOCK_MAP_SIZE); > > - sbi->discard_blks += > > - sbi->blocks_per_seg - > > + sbi->discard_blks += BLKS_PER_SEG(sbi) - > > se->valid_blocks; > > - } > > - } > > - > > +init_discard_map_done: > > if (__is_large_section(sbi)) > > get_sec_entry(sbi, start)->valid_blocks += > > se->valid_blocks; > > @@ -4736,7 +4734,7 @@ static void init_dirty_segmap(struct f2fs_sb_info *sbi) > > return; > > mutex_lock(&dirty_i->seglist_lock); > > - for (segno = 0; segno < MAIN_SEGS(sbi); segno += sbi->segs_per_sec) { > > + for (segno = 0; segno < MAIN_SEGS(sbi); segno += SEGS_PER_SEC(sbi)) { > > valid_blocks = get_valid_blocks(sbi, segno, true); > > secno = GET_SEC_FROM_SEG(sbi, segno); > > @@ -4835,7 +4833,7 @@ static int sanity_check_curseg(struct f2fs_sb_info *sbi) > > if (curseg->alloc_type == SSR) > > continue; > > - for (blkofs += 1; blkofs < sbi->blocks_per_seg; blkofs++) { > > + for (blkofs += 1; blkofs < BLKS_PER_SEG(sbi); blkofs++) { > > if (!f2fs_test_bit(blkofs, se->cur_valid_map)) > > continue; > > out: > > @@ -5114,7 +5112,7 @@ static inline unsigned int f2fs_usable_zone_blks_in_seg( > > unsigned int secno; > > if (!sbi->unusable_blocks_per_sec) > > - return sbi->blocks_per_seg; > > + return BLKS_PER_SEG(sbi); > > secno = GET_SEC_FROM_SEG(sbi, segno); > > seg_start = START_BLOCK(sbi, segno); > > @@ -5129,10 +5127,10 @@ static inline unsigned int f2fs_usable_zone_blks_in_seg( > > */ > > if (seg_start >= sec_cap_blkaddr) > > return 0; > > - if (seg_start + sbi->blocks_per_seg > sec_cap_blkaddr) > > + if (seg_start + BLKS_PER_SEG(sbi) > sec_cap_blkaddr) > > return sec_cap_blkaddr - seg_start; > > - return sbi->blocks_per_seg; > > + return BLKS_PER_SEG(sbi); > > } > > #else > > int f2fs_fix_curseg_write_pointer(struct f2fs_sb_info *sbi) > > @@ -5158,7 +5156,7 @@ unsigned int f2fs_usable_blks_in_seg(struct f2fs_sb_info *sbi, > > if (f2fs_sb_has_blkzoned(sbi)) > > return f2fs_usable_zone_blks_in_seg(sbi, segno); > > - return sbi->blocks_per_seg; > > + return BLKS_PER_SEG(sbi); > > } > > unsigned int f2fs_usable_segs_in_sec(struct f2fs_sb_info *sbi, > > @@ -5167,7 +5165,7 @@ unsigned int f2fs_usable_segs_in_sec(struct f2fs_sb_info *sbi, > > if (f2fs_sb_has_blkzoned(sbi)) > > return CAP_SEGS_PER_SEC(sbi); > > - return sbi->segs_per_sec; > > + return SEGS_PER_SEC(sbi); > > } > > /* > > @@ -5182,14 +5180,14 @@ static void init_min_max_mtime(struct f2fs_sb_info *sbi) > > sit_i->min_mtime = ULLONG_MAX; > > - for (segno = 0; segno < MAIN_SEGS(sbi); segno += sbi->segs_per_sec) { > > + for (segno = 0; segno < MAIN_SEGS(sbi); segno += SEGS_PER_SEC(sbi)) { > > unsigned int i; > > unsigned long long mtime = 0; > > - for (i = 0; i < sbi->segs_per_sec; i++) > > + for (i = 0; i < SEGS_PER_SEC(sbi); i++) > > mtime += get_seg_entry(sbi, segno + i)->mtime; > > - mtime = div_u64(mtime, sbi->segs_per_sec); > > + mtime = div_u64(mtime, SEGS_PER_SEC(sbi)); > > if (sit_i->min_mtime > mtime) > > sit_i->min_mtime = mtime; > > @@ -5228,7 +5226,7 @@ int f2fs_build_segment_manager(struct f2fs_sb_info *sbi) > > sm_info->ipu_policy = BIT(F2FS_IPU_FSYNC); > > sm_info->min_ipu_util = DEF_MIN_IPU_UTIL; > > sm_info->min_fsync_blocks = DEF_MIN_FSYNC_BLOCKS; > > - sm_info->min_seq_blocks = sbi->blocks_per_seg; > > + sm_info->min_seq_blocks = BLKS_PER_SEG(sbi); > > sm_info->min_hot_blocks = DEF_MIN_HOT_BLOCKS; > > sm_info->min_ssr_sections = reserved_sections(sbi); > > diff --git a/fs/f2fs/segment.h b/fs/f2fs/segment.h > > index 8129be788bd5..febcfbadcdfa 100644 > > --- a/fs/f2fs/segment.h > > +++ b/fs/f2fs/segment.h > > @@ -48,21 +48,21 @@ static inline void sanity_check_seg_type(struct f2fs_sb_info *sbi, > > #define IS_CURSEC(sbi, secno) \ > > (((secno) == CURSEG_I(sbi, CURSEG_HOT_DATA)->segno / \ > > - (sbi)->segs_per_sec) || \ > > + SEGS_PER_SEC(sbi)) || \ > > ((secno) == CURSEG_I(sbi, CURSEG_WARM_DATA)->segno / \ > > - (sbi)->segs_per_sec) || \ > > + SEGS_PER_SEC(sbi)) || \ > > ((secno) == CURSEG_I(sbi, CURSEG_COLD_DATA)->segno / \ > > - (sbi)->segs_per_sec) || \ > > + SEGS_PER_SEC(sbi)) || \ > > ((secno) == CURSEG_I(sbi, CURSEG_HOT_NODE)->segno / \ > > - (sbi)->segs_per_sec) || \ > > + SEGS_PER_SEC(sbi)) || \ > > ((secno) == CURSEG_I(sbi, CURSEG_WARM_NODE)->segno / \ > > - (sbi)->segs_per_sec) || \ > > + SEGS_PER_SEC(sbi)) || \ > > ((secno) == CURSEG_I(sbi, CURSEG_COLD_NODE)->segno / \ > > - (sbi)->segs_per_sec) || \ > > + SEGS_PER_SEC(sbi)) || \ > > ((secno) == CURSEG_I(sbi, CURSEG_COLD_DATA_PINNED)->segno / \ > > - (sbi)->segs_per_sec) || \ > > + SEGS_PER_SEC(sbi)) || \ > > ((secno) == CURSEG_I(sbi, CURSEG_ALL_DATA_ATGC)->segno / \ > > - (sbi)->segs_per_sec)) > > + SEGS_PER_SEC(sbi))) > > #define MAIN_BLKADDR(sbi) \ > > (SM_I(sbi) ? SM_I(sbi)->main_blkaddr : \ > > @@ -93,24 +93,22 @@ static inline void sanity_check_seg_type(struct f2fs_sb_info *sbi, > > #define GET_SEGNO_FROM_SEG0(sbi, blk_addr) \ > > (GET_SEGOFF_FROM_SEG0(sbi, blk_addr) >> (sbi)->log_blocks_per_seg) > > #define GET_BLKOFF_FROM_SEG0(sbi, blk_addr) \ > > - (GET_SEGOFF_FROM_SEG0(sbi, blk_addr) & ((sbi)->blocks_per_seg - 1)) > > + (GET_SEGOFF_FROM_SEG0(sbi, blk_addr) & (BLKS_PER_SEG(sbi) - 1)) > > #define GET_SEGNO(sbi, blk_addr) \ > > ((!__is_valid_data_blkaddr(blk_addr)) ? \ > > NULL_SEGNO : GET_L2R_SEGNO(FREE_I(sbi), \ > > GET_SEGNO_FROM_SEG0(sbi, blk_addr))) > > -#define BLKS_PER_SEC(sbi) \ > > - ((sbi)->segs_per_sec * (sbi)->blocks_per_seg) > > #define CAP_BLKS_PER_SEC(sbi) \ > > - ((sbi)->segs_per_sec * (sbi)->blocks_per_seg - \ > > + (SEGS_PER_SEC(sbi) * BLKS_PER_SEG(sbi) - \ > > (sbi)->unusable_blocks_per_sec) > > #define CAP_SEGS_PER_SEC(sbi) \ > > - ((sbi)->segs_per_sec - ((sbi)->unusable_blocks_per_sec >>\ > > + (SEGS_PER_SEC(sbi) - ((sbi)->unusable_blocks_per_sec >> \ > > (sbi)->log_blocks_per_seg)) > > #define GET_SEC_FROM_SEG(sbi, segno) \ > > - (((segno) == -1) ? -1 : (segno) / (sbi)->segs_per_sec) > > + (((segno) == -1) ? -1 : (segno) / SEGS_PER_SEC(sbi)) > > #define GET_SEG_FROM_SEC(sbi, secno) \ > > - ((secno) * (sbi)->segs_per_sec) > > + ((secno) * SEGS_PER_SEC(sbi)) > > #define GET_ZONE_FROM_SEC(sbi, secno) \ > > (((secno) == -1) ? -1 : (secno) / (sbi)->secs_per_zone) > > #define GET_ZONE_FROM_SEG(sbi, segno) \ > > @@ -364,7 +362,7 @@ static inline unsigned int get_ckpt_valid_blocks(struct f2fs_sb_info *sbi, > > unsigned int blocks = 0; > > int i; > > - for (i = 0; i < sbi->segs_per_sec; i++, start_segno++) { > > + for (i = 0; i < SEGS_PER_SEC(sbi); i++, start_segno++) { > > struct seg_entry *se = get_seg_entry(sbi, start_segno); > > blocks += se->ckpt_valid_blocks; > > @@ -449,7 +447,7 @@ static inline void __set_free(struct f2fs_sb_info *sbi, unsigned int segno) > > free_i->free_segments++; > > next = find_next_bit(free_i->free_segmap, > > - start_segno + sbi->segs_per_sec, start_segno); > > + start_segno + SEGS_PER_SEC(sbi), start_segno); > > if (next >= start_segno + usable_segs) { > > clear_bit(secno, free_i->free_secmap); > > free_i->free_sections++; > > @@ -485,7 +483,7 @@ static inline void __set_test_and_free(struct f2fs_sb_info *sbi, > > if (!inmem && IS_CURSEC(sbi, secno)) > > goto skip_free; > > next = find_next_bit(free_i->free_segmap, > > - start_segno + sbi->segs_per_sec, start_segno); > > + start_segno + SEGS_PER_SEC(sbi), start_segno); > > if (next >= start_segno + usable_segs) { > > if (test_and_clear_bit(secno, free_i->free_secmap)) > > free_i->free_sections++; > > @@ -793,10 +791,10 @@ static inline int check_block_count(struct f2fs_sb_info *sbi, > > return -EFSCORRUPTED; > > } > > - if (usable_blks_per_seg < sbi->blocks_per_seg) > > + if (usable_blks_per_seg < BLKS_PER_SEG(sbi)) > > f2fs_bug_on(sbi, find_next_bit_le(&raw_sit->valid_map, > > - sbi->blocks_per_seg, > > - usable_blks_per_seg) != sbi->blocks_per_seg); > > + BLKS_PER_SEG(sbi), > > + usable_blks_per_seg) != BLKS_PER_SEG(sbi)); > > /* check segment usage, and check boundary of a given segment number */ > > if (unlikely(GET_SIT_VBLOCKS(raw_sit) > usable_blks_per_seg > > @@ -915,9 +913,9 @@ static inline int nr_pages_to_skip(struct f2fs_sb_info *sbi, int type) > > return 0; > > if (type == DATA) > > - return sbi->blocks_per_seg; > > + return BLKS_PER_SEG(sbi); > > else if (type == NODE) > > - return 8 * sbi->blocks_per_seg; > > + return 8 * BLKS_PER_SEG(sbi); > > else if (type == META) > > return 8 * BIO_MAX_VECS; > > else > > diff --git a/fs/f2fs/super.c b/fs/f2fs/super.c > > index f1516fd5088a..c0688c124aa7 100644 > > --- a/fs/f2fs/super.c > > +++ b/fs/f2fs/super.c > > @@ -3643,7 +3643,7 @@ int f2fs_sanity_check_ckpt(struct f2fs_sb_info *sbi) > > } > > main_segs = le32_to_cpu(raw_super->segment_count_main); > > - blocks_per_seg = sbi->blocks_per_seg; > > + blocks_per_seg = BLKS_PER_SEG(sbi); > > for (i = 0; i < NR_CURSEG_NODE_TYPE; i++) { > > if (le32_to_cpu(ckpt->cur_node_segno[i]) >= main_segs || > > @@ -3756,8 +3756,8 @@ static void init_sb_info(struct f2fs_sb_info *sbi) > > sbi->secs_per_zone = le32_to_cpu(raw_super->secs_per_zone); > > sbi->total_sections = le32_to_cpu(raw_super->section_count); > > sbi->total_node_count = > > - (le32_to_cpu(raw_super->segment_count_nat) / 2) > > - * sbi->blocks_per_seg * NAT_ENTRY_PER_BLOCK; > > + ((le32_to_cpu(raw_super->segment_count_nat) / 2) * > > + NAT_ENTRY_PER_BLOCK) << sbi->log_blocks_per_seg; > > F2FS_ROOT_INO(sbi) = le32_to_cpu(raw_super->root_ino); > > F2FS_NODE_INO(sbi) = le32_to_cpu(raw_super->node_ino); > > F2FS_META_INO(sbi) = le32_to_cpu(raw_super->meta_ino); > > @@ -3766,7 +3766,7 @@ static void init_sb_info(struct f2fs_sb_info *sbi) > > sbi->next_victim_seg[BG_GC] = NULL_SEGNO; > > sbi->next_victim_seg[FG_GC] = NULL_SEGNO; > > sbi->max_victim_search = DEF_MAX_VICTIM_SEARCH; > > - sbi->migration_granularity = sbi->segs_per_sec; > > + sbi->migration_granularity = SEGS_PER_SEC(sbi); > > sbi->seq_file_ra_mul = MIN_RA_MUL; > > sbi->max_fragment_chunk = DEF_FRAGMENT_SIZE; > > sbi->max_fragment_hole = DEF_FRAGMENT_SIZE; > > diff --git a/fs/f2fs/sysfs.c b/fs/f2fs/sysfs.c > > index a7ec55c7bb20..906d2af2d849 100644 > > --- a/fs/f2fs/sysfs.c > > +++ b/fs/f2fs/sysfs.c > > @@ -493,8 +493,8 @@ static ssize_t __sbi_store(struct f2fs_attr *a, > > spin_lock(&sbi->stat_lock); > > if (t > (unsigned long)(sbi->user_block_count - > > F2FS_OPTION(sbi).root_reserved_blocks - > > - sbi->blocks_per_seg * > > - SM_I(sbi)->additional_reserved_segments)) { > > + (SM_I(sbi)->additional_reserved_segments << > > + sbi->log_blocks_per_seg))) { > > spin_unlock(&sbi->stat_lock); > > return -EINVAL; > > } > > @@ -551,7 +551,7 @@ static ssize_t __sbi_store(struct f2fs_attr *a, > > } > > if (!strcmp(a->attr.name, "migration_granularity")) { > > - if (t == 0 || t > sbi->segs_per_sec) > > + if (t == 0 || t > SEGS_PER_SEC(sbi)) > > return -EINVAL; > > }