Received: by 2002:a05:7412:cfc7:b0:fc:a2b0:25d7 with SMTP id by7csp2087727rdb; Tue, 20 Feb 2024 17:30:59 -0800 (PST) X-Forwarded-Encrypted: i=3; AJvYcCWoU7s2eWL5puTeNjIb6cdI/Kg29CGfFbTb6JR6TRnoTeicb2nyqL95xGWtzRtI3tUkGbAqHEFn+5TdvbU24nuiJsF+IQhF928fV99BsQ== X-Google-Smtp-Source: AGHT+IFPXIWPdaKBJl23v7GOAJMcuVqEHoZf9MegZ/OmQLnsBsV18Dmz7kU37Ibb4z8+okG373m6 X-Received: by 2002:a25:414f:0:b0:dcc:ae3:d8a0 with SMTP id o76-20020a25414f000000b00dcc0ae3d8a0mr8747228yba.48.1708479059250; Tue, 20 Feb 2024 17:30:59 -0800 (PST) ARC-Seal: i=2; a=rsa-sha256; t=1708479059; cv=pass; d=google.com; s=arc-20160816; b=zQl5rdOTGZIsd5nuZiW6C0MSMG/1dcLDNJ2fXvp6Xnqsu0yf2heOnfmZMLS0ovZJyz 0mlGjQKOhJCw3A7ZCrkrB5HBq62MyGzwGsfTLoOrEdN+j6vLCPEWEx/glupNa4S4u8Uz CO2oLJToiMDax2xph3CDKcnCKMYAJxxNbiAhzUE+a9KIPt9jZguu7nudz0lNuibMmyMh 9aQfCerc6QZkhCtdrnN/4K66HTksCtp2C69Kzk7F5absJdhMLyDu80THKCas2lbTfeAH aphsol6rYVTEajDDdtnFvNmXWCPDsHybDW3YST+6QmckRNc/OZiIS3ga0Rww6c+4oPE6 GhGA== ARC-Message-Signature: i=2; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=content-transfer-encoding:in-reply-to:from:references:to :content-language:subject:user-agent:mime-version:list-unsubscribe :list-subscribe:list-id:precedence:date:message-id:dkim-signature; bh=nBAEKXit2nGyBRJyEq5mYcWSbU9EKQ4uMwEz0kU0o0s=; fh=PHq4OuhOUUP7V1I+gDdMCZNzaT4vH9shSoLBu78e8bA=; b=nsxeEkxKxrcw77bU+wnqLFBnSOrzUETiyfxs3I0HaGpndIYXfLQQr3DGdeqMwhFHhV vba67RgQ73+04pKqdyff/iWNb9mhaMNoNbIF8lfcRbkYn2dMsO/6S5pLyQl5bpXWPjbj gJ21/vfiUtpIhMydzUC7xM+1dAxBeStwN1e4dtu+K2YENRMtrMaCkPYst5r5oOHDcHzd b7/hlEW0czJ9JKj6Q66XsGUV3TdMucLR0ZXxfnQ8YXY2r9fQND0amSfRzSCBFmKtYhEg c5sx4JiAC+kYum2V3k3MTCXTwmK79GiC/7Lo8nRkPM8NEpFcZxDlgVt61/1lqqk7MMyU ID6g==; dara=google.com ARC-Authentication-Results: i=2; mx.google.com; dkim=pass header.i=@kernel.org header.s=k20201202 header.b=bFhDitHX; arc=pass (i=1 dkim=pass dkdomain=kernel.org); spf=pass (google.com: domain of linux-kernel+bounces-73907-linux.lists.archive=gmail.com@vger.kernel.org designates 147.75.199.223 as permitted sender) smtp.mailfrom="linux-kernel+bounces-73907-linux.lists.archive=gmail.com@vger.kernel.org"; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=kernel.org Return-Path: Received: from ny.mirrors.kernel.org (ny.mirrors.kernel.org. [147.75.199.223]) by mx.google.com with ESMTPS id b24-20020a05620a127800b00785a70ff265si9453629qkl.710.2024.02.20.17.30.59 for (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Tue, 20 Feb 2024 17:30:59 -0800 (PST) Received-SPF: pass (google.com: domain of linux-kernel+bounces-73907-linux.lists.archive=gmail.com@vger.kernel.org designates 147.75.199.223 as permitted sender) client-ip=147.75.199.223; Authentication-Results: mx.google.com; dkim=pass header.i=@kernel.org header.s=k20201202 header.b=bFhDitHX; arc=pass (i=1 dkim=pass dkdomain=kernel.org); spf=pass (google.com: domain of linux-kernel+bounces-73907-linux.lists.archive=gmail.com@vger.kernel.org designates 147.75.199.223 as permitted sender) smtp.mailfrom="linux-kernel+bounces-73907-linux.lists.archive=gmail.com@vger.kernel.org"; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=kernel.org Received: from smtp.subspace.kernel.org (wormhole.subspace.kernel.org [52.25.139.140]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by ny.mirrors.kernel.org (Postfix) with ESMTPS id B8C631C22A00 for ; Wed, 21 Feb 2024 01:30:58 +0000 (UTC) Received: from localhost.localdomain (localhost.localdomain [127.0.0.1]) by smtp.subspace.kernel.org (Postfix) with ESMTP id 12EE4468D; Wed, 21 Feb 2024 01:30:53 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; dkim=pass (2048-bit key) header.d=kernel.org header.i=@kernel.org header.b="bFhDitHX" Received: from smtp.kernel.org (aws-us-west-2-korg-mail-1.web.codeaurora.org [10.30.226.201]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by smtp.subspace.kernel.org (Postfix) with ESMTPS id 8896217D5 for ; Wed, 21 Feb 2024 01:30:50 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=none smtp.client-ip=10.30.226.201 ARC-Seal:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1708479050; cv=none; b=K3Jzhi74qcK3OQDMd00yXNJyVwv4fOTEMcIjzgL9L3bTvrO/BrUuQrU/VxD5SloQDu6+2ix62Gh4mnUJ9PeR7+ORgDeGN0RmeZYQ9vNwzMjqh9YtQ6MLOuu6+Pepgh+wY2A34PQGABYaFmJ0b4FX9U7mCFfOHG+/BtshNMLrX78= ARC-Message-Signature:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1708479050; c=relaxed/simple; bh=798HVcUc3wIX9Yue/WxK2scs9h+wrka/s4hhfq0/KbY=; h=Message-ID:Date:MIME-Version:Subject:To:References:From: In-Reply-To:Content-Type; b=hkGw06RBLIn9wJtwdt8uEPCkpHudlb1qAQfthieE/ht09ZFU0Wm3tdOSg8o2yk1TsX455ba1nHp1c1QIKHFRROxFExpCoYptG4swKT+iJszS4dHvGOyroBayxz9VqDEIoletu31p6xndNTx4YZpKDLGYUX185xifY+0TWFO7988= ARC-Authentication-Results:i=1; smtp.subspace.kernel.org; dkim=pass (2048-bit key) header.d=kernel.org header.i=@kernel.org header.b=bFhDitHX; arc=none smtp.client-ip=10.30.226.201 Received: by smtp.kernel.org (Postfix) with ESMTPSA id 870F0C433C7; Wed, 21 Feb 2024 01:30:48 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=kernel.org; s=k20201202; t=1708479049; bh=798HVcUc3wIX9Yue/WxK2scs9h+wrka/s4hhfq0/KbY=; h=Date:Subject:To:References:From:In-Reply-To:From; b=bFhDitHXF//IUb5cFiLduPqvt2ucIt4vHbla2Eeq3N+YEe/d5WQZ5Ak7RMQaBlPXS faTNfZ5iPEBeQIvroso05i/uPMFPc2LMDsamjm+pfwpLha2kAE380lv0zqoQk2Zu04 M+wTXAigT6FTNlPj/cZP4J2sSv/6cOG7I+6ihq08q/HRZlhBOQqG21WMcNyAVVwbKZ wYPqs5FY3EIn29gEKolH4sYDw4Yl1D8gu/qROH0VzO8qIasUn6bLPyiNzUhYeHLXUS nqhaejMOE59R5jhaGIUEv5mp7DHoYlk3jPZV/EYr1DI94NlcLYRLzUOoCjkEGqsxqH samK0FKZWgG9A== Message-ID: <18ce2359-f2ee-4b33-8ba7-d6444d56be56@kernel.org> Date: Wed, 21 Feb 2024 09:30:45 +0800 Precedence: bulk X-Mailing-List: linux-kernel@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 User-Agent: Mozilla Thunderbird Subject: Re: [f2fs-dev] [PATCH 2/3 v3] f2fs: use BLKS_PER_SEG, BLKS_PER_SEC, and SEGS_PER_SEC Content-Language: en-US To: Jaegeuk Kim , linux-kernel@vger.kernel.org, linux-f2fs-devel@lists.sourceforge.net References: <20240207005105.3744811-1-jaegeuk@kernel.org> <20240207005105.3744811-2-jaegeuk@kernel.org> From: Chao Yu In-Reply-To: Content-Type: text/plain; charset=UTF-8; format=flowed Content-Transfer-Encoding: 7bit On 2024/2/21 4:39, Jaegeuk Kim wrote: > No functional change. > > Reviewed-by: Daeho Jeong > Signed-off-by: Jaegeuk Kim > --- > Change log from v2: > - cover more cases > > Change log from v1: > - use div_u64 > > fs/f2fs/checkpoint.c | 10 ++-- > fs/f2fs/debug.c | 6 +-- > fs/f2fs/f2fs.h | 21 +++++---- > fs/f2fs/file.c | 16 +++---- > fs/f2fs/gc.c | 40 ++++++++-------- > fs/f2fs/node.c | 4 +- > fs/f2fs/node.h | 4 +- > fs/f2fs/recovery.c | 2 +- > fs/f2fs/segment.c | 110 +++++++++++++++++++++---------------------- > fs/f2fs/segment.h | 44 +++++++++-------- > fs/f2fs/super.c | 8 ++-- > fs/f2fs/sysfs.c | 6 +-- > 12 files changed, 135 insertions(+), 136 deletions(-) > > diff --git a/fs/f2fs/checkpoint.c b/fs/f2fs/checkpoint.c > index b85820e70f5e..a09a9609e228 100644 > --- a/fs/f2fs/checkpoint.c > +++ b/fs/f2fs/checkpoint.c > @@ -900,7 +900,7 @@ static struct page *validate_checkpoint(struct f2fs_sb_info *sbi, > > cp_blocks = le32_to_cpu(cp_block->cp_pack_total_block_count); > > - if (cp_blocks > sbi->blocks_per_seg || cp_blocks <= F2FS_CP_PACKS) { > + if (cp_blocks > BLKS_PER_SEG(sbi) || cp_blocks <= F2FS_CP_PACKS) { > f2fs_warn(sbi, "invalid cp_pack_total_block_count:%u", > le32_to_cpu(cp_block->cp_pack_total_block_count)); > goto invalid_cp; > @@ -1335,7 +1335,7 @@ static void update_ckpt_flags(struct f2fs_sb_info *sbi, struct cp_control *cpc) > > if (cpc->reason & CP_UMOUNT) { > if (le32_to_cpu(ckpt->cp_pack_total_block_count) + > - NM_I(sbi)->nat_bits_blocks > sbi->blocks_per_seg) { > + NM_I(sbi)->nat_bits_blocks > BLKS_PER_SEG(sbi)) { > clear_ckpt_flags(sbi, CP_NAT_BITS_FLAG); > f2fs_notice(sbi, "Disable nat_bits due to no space"); > } else if (!is_set_ckpt_flags(sbi, CP_NAT_BITS_FLAG) && > @@ -1538,7 +1538,7 @@ static int do_checkpoint(struct f2fs_sb_info *sbi, struct cp_control *cpc) > cp_ver |= ((__u64)crc32 << 32); > *(__le64 *)nm_i->nat_bits = cpu_to_le64(cp_ver); > > - blk = start_blk + sbi->blocks_per_seg - nm_i->nat_bits_blocks; > + blk = start_blk + BLKS_PER_SEG(sbi) - nm_i->nat_bits_blocks; > for (i = 0; i < nm_i->nat_bits_blocks; i++) > f2fs_update_meta_page(sbi, nm_i->nat_bits + > (i << F2FS_BLKSIZE_BITS), blk + i); > @@ -1741,9 +1741,9 @@ void f2fs_init_ino_entry_info(struct f2fs_sb_info *sbi) > im->ino_num = 0; > } > > - sbi->max_orphans = (sbi->blocks_per_seg - F2FS_CP_PACKS - > + sbi->max_orphans = (BLKS_PER_SEG(sbi) - F2FS_CP_PACKS - > NR_CURSEG_PERSIST_TYPE - __cp_payload(sbi)) * > - F2FS_ORPHANS_PER_BLOCK; > + F2FS_ORPHANS_PER_BLOCK; > } > > int __init f2fs_create_checkpoint_caches(void) > diff --git a/fs/f2fs/debug.c b/fs/f2fs/debug.c > index fdbf994f1271..0d02224b99b7 100644 > --- a/fs/f2fs/debug.c > +++ b/fs/f2fs/debug.c > @@ -41,7 +41,7 @@ void f2fs_update_sit_info(struct f2fs_sb_info *sbi) > total_vblocks = 0; > blks_per_sec = CAP_BLKS_PER_SEC(sbi); > hblks_per_sec = blks_per_sec / 2; > - for (segno = 0; segno < MAIN_SEGS(sbi); segno += sbi->segs_per_sec) { > + for (segno = 0; segno < MAIN_SEGS(sbi); segno += SEGS_PER_SEC(sbi)) { > vblocks = get_valid_blocks(sbi, segno, true); > dist = abs(vblocks - hblks_per_sec); > bimodal += dist * dist; > @@ -135,7 +135,7 @@ static void update_general_status(struct f2fs_sb_info *sbi) > si->cur_ckpt_time = sbi->cprc_info.cur_time; > si->peak_ckpt_time = sbi->cprc_info.peak_time; > spin_unlock(&sbi->cprc_info.stat_lock); > - si->total_count = (int)sbi->user_block_count / sbi->blocks_per_seg; > + si->total_count = (int)sbi->user_block_count / BLKS_PER_SEG(sbi); > si->rsvd_segs = reserved_segments(sbi); > si->overp_segs = overprovision_segments(sbi); > si->valid_count = valid_user_blocks(sbi); > @@ -208,7 +208,7 @@ static void update_general_status(struct f2fs_sb_info *sbi) > if (!blks) > continue; > > - if (blks == sbi->blocks_per_seg) > + if (blks == BLKS_PER_SEG(sbi)) > si->full_seg[type]++; > else > si->dirty_seg[type]++; > diff --git a/fs/f2fs/f2fs.h b/fs/f2fs/f2fs.h > index 4c52136cbc10..50e7890cc6a5 100644 > --- a/fs/f2fs/f2fs.h > +++ b/fs/f2fs/f2fs.h > @@ -1813,6 +1813,14 @@ struct f2fs_sb_info { > #endif > }; > > +/* Definitions to access f2fs_sb_info */ > +#define BLKS_PER_SEG(sbi) \ > + ((sbi)->blocks_per_seg) > +#define BLKS_PER_SEC(sbi) \ > + ((sbi)->segs_per_sec << (sbi)->log_blocks_per_seg) > +#define SEGS_PER_SEC(sbi) \ > + ((sbi)->segs_per_sec) I can see lots of 'xxx << sbi->log_blocks_per_seg', how about introducing a new macro to wrap it, maybe: #define SEGMENTS_TO_BLKS(sbi, segments) \ (segments << (sbi)->log_blocks_per_seg) Thanks, > + > __printf(3, 4) > void f2fs_printk(struct f2fs_sb_info *sbi, bool limit_rate, const char *fmt, ...); > > @@ -2511,11 +2519,8 @@ static inline int get_dirty_pages(struct inode *inode) > > static inline int get_blocktype_secs(struct f2fs_sb_info *sbi, int block_type) > { > - unsigned int pages_per_sec = sbi->segs_per_sec * sbi->blocks_per_seg; > - unsigned int segs = (get_pages(sbi, block_type) + pages_per_sec - 1) >> > - sbi->log_blocks_per_seg; > - > - return segs / sbi->segs_per_sec; > + return div_u64(get_pages(sbi, block_type) + BLKS_PER_SEC(sbi) - 1, > + BLKS_PER_SEC(sbi)); > } > > static inline block_t valid_user_blocks(struct f2fs_sb_info *sbi) > @@ -2579,7 +2584,7 @@ static inline block_t __start_cp_addr(struct f2fs_sb_info *sbi) > block_t start_addr = le32_to_cpu(F2FS_RAW_SUPER(sbi)->cp_blkaddr); > > if (sbi->cur_cp_pack == 2) > - start_addr += sbi->blocks_per_seg; > + start_addr += BLKS_PER_SEG(sbi); > return start_addr; > } > > @@ -2588,7 +2593,7 @@ static inline block_t __start_cp_next_addr(struct f2fs_sb_info *sbi) > block_t start_addr = le32_to_cpu(F2FS_RAW_SUPER(sbi)->cp_blkaddr); > > if (sbi->cur_cp_pack == 1) > - start_addr += sbi->blocks_per_seg; > + start_addr += BLKS_PER_SEG(sbi); > return start_addr; > } > > @@ -3458,7 +3463,7 @@ static inline __le32 *get_dnode_addr(struct inode *inode, > sizeof((f2fs_inode)->field)) \ > <= (F2FS_OLD_ATTRIBUTE_SIZE + (extra_isize))) \ > > -#define __is_large_section(sbi) ((sbi)->segs_per_sec > 1) > +#define __is_large_section(sbi) (SEGS_PER_SEC(sbi) > 1) > > #define __is_meta_io(fio) (PAGE_TYPE_OF_BIO((fio)->type) == META) > > diff --git a/fs/f2fs/file.c b/fs/f2fs/file.c > index c6cd9474ba2d..767d16c74bb6 100644 > --- a/fs/f2fs/file.c > +++ b/fs/f2fs/file.c > @@ -2580,7 +2580,6 @@ static int f2fs_defragment_range(struct f2fs_sb_info *sbi, > .m_may_create = false }; > struct extent_info ei = {}; > pgoff_t pg_start, pg_end, next_pgofs; > - unsigned int blk_per_seg = sbi->blocks_per_seg; > unsigned int total = 0, sec_num; > block_t blk_end = 0; > bool fragmented = false; > @@ -2689,7 +2688,8 @@ static int f2fs_defragment_range(struct f2fs_sb_info *sbi, > set_inode_flag(inode, FI_SKIP_WRITES); > > idx = map.m_lblk; > - while (idx < map.m_lblk + map.m_len && cnt < blk_per_seg) { > + while (idx < map.m_lblk + map.m_len && > + cnt < BLKS_PER_SEG(sbi)) { > struct page *page; > > page = f2fs_get_lock_data_page(inode, idx, true); > @@ -2709,7 +2709,7 @@ static int f2fs_defragment_range(struct f2fs_sb_info *sbi, > > map.m_lblk = idx; > check: > - if (map.m_lblk < pg_end && cnt < blk_per_seg) > + if (map.m_lblk < pg_end && cnt < BLKS_PER_SEG(sbi)) > goto do_map; > > clear_inode_flag(inode, FI_SKIP_WRITES); > @@ -2978,8 +2978,8 @@ static int f2fs_ioc_flush_device(struct file *filp, unsigned long arg) > > if (!f2fs_is_multi_device(sbi) || sbi->s_ndevs - 1 <= range.dev_num || > __is_large_section(sbi)) { > - f2fs_warn(sbi, "Can't flush %u in %d for segs_per_sec %u != 1", > - range.dev_num, sbi->s_ndevs, sbi->segs_per_sec); > + f2fs_warn(sbi, "Can't flush %u in %d for SEGS_PER_SEC %u != 1", > + range.dev_num, sbi->s_ndevs, SEGS_PER_SEC(sbi)); > return -EINVAL; > } > > @@ -4081,7 +4081,6 @@ static int f2fs_ioc_decompress_file(struct file *filp) > struct f2fs_sb_info *sbi = F2FS_I_SB(inode); > struct f2fs_inode_info *fi = F2FS_I(inode); > pgoff_t page_idx = 0, last_idx; > - unsigned int blk_per_seg = sbi->blocks_per_seg; > int cluster_size = fi->i_cluster_size; > int count, ret; > > @@ -4125,7 +4124,7 @@ static int f2fs_ioc_decompress_file(struct file *filp) > if (ret < 0) > break; > > - if (get_dirty_pages(inode) >= blk_per_seg) { > + if (get_dirty_pages(inode) >= BLKS_PER_SEG(sbi)) { > ret = filemap_fdatawrite(inode->i_mapping); > if (ret < 0) > break; > @@ -4160,7 +4159,6 @@ static int f2fs_ioc_compress_file(struct file *filp) > struct inode *inode = file_inode(filp); > struct f2fs_sb_info *sbi = F2FS_I_SB(inode); > pgoff_t page_idx = 0, last_idx; > - unsigned int blk_per_seg = sbi->blocks_per_seg; > int cluster_size = F2FS_I(inode)->i_cluster_size; > int count, ret; > > @@ -4203,7 +4201,7 @@ static int f2fs_ioc_compress_file(struct file *filp) > if (ret < 0) > break; > > - if (get_dirty_pages(inode) >= blk_per_seg) { > + if (get_dirty_pages(inode) >= BLKS_PER_SEG(sbi)) { > ret = filemap_fdatawrite(inode->i_mapping); > if (ret < 0) > break; > diff --git a/fs/f2fs/gc.c b/fs/f2fs/gc.c > index 6899f434ad68..d61a60c1c844 100644 > --- a/fs/f2fs/gc.c > +++ b/fs/f2fs/gc.c > @@ -259,7 +259,7 @@ static void select_policy(struct f2fs_sb_info *sbi, int gc_type, > p->ofs_unit = 1; > } else { > p->gc_mode = select_gc_type(sbi, gc_type); > - p->ofs_unit = sbi->segs_per_sec; > + p->ofs_unit = SEGS_PER_SEC(sbi); > if (__is_large_section(sbi)) { > p->dirty_bitmap = dirty_i->dirty_secmap; > p->max_search = count_bits(p->dirty_bitmap, > @@ -282,7 +282,8 @@ static void select_policy(struct f2fs_sb_info *sbi, int gc_type, > > /* let's select beginning hot/small space first in no_heap mode*/ > if (f2fs_need_rand_seg(sbi)) > - p->offset = get_random_u32_below(MAIN_SECS(sbi) * sbi->segs_per_sec); > + p->offset = get_random_u32_below(MAIN_SECS(sbi) * > + SEGS_PER_SEC(sbi)); > else if (test_opt(sbi, NOHEAP) && > (type == CURSEG_HOT_DATA || IS_NODESEG(type))) > p->offset = 0; > @@ -295,13 +296,13 @@ static unsigned int get_max_cost(struct f2fs_sb_info *sbi, > { > /* SSR allocates in a segment unit */ > if (p->alloc_mode == SSR) > - return sbi->blocks_per_seg; > + return BLKS_PER_SEG(sbi); > else if (p->alloc_mode == AT_SSR) > return UINT_MAX; > > /* LFS */ > if (p->gc_mode == GC_GREEDY) > - return 2 * sbi->blocks_per_seg * p->ofs_unit; > + return 2 * BLKS_PER_SEG(sbi) * p->ofs_unit; > else if (p->gc_mode == GC_CB) > return UINT_MAX; > else if (p->gc_mode == GC_AT) > @@ -496,9 +497,9 @@ static void add_victim_entry(struct f2fs_sb_info *sbi, > return; > } > > - for (i = 0; i < sbi->segs_per_sec; i++) > + for (i = 0; i < SEGS_PER_SEC(sbi); i++) > mtime += get_seg_entry(sbi, start + i)->mtime; > - mtime = div_u64(mtime, sbi->segs_per_sec); > + mtime = div_u64(mtime, SEGS_PER_SEC(sbi)); > > /* Handle if the system time has changed by the user */ > if (mtime < sit_i->min_mtime) > @@ -599,7 +600,6 @@ static void atssr_lookup_victim(struct f2fs_sb_info *sbi, > unsigned long long age; > unsigned long long max_mtime = sit_i->dirty_max_mtime; > unsigned long long min_mtime = sit_i->dirty_min_mtime; > - unsigned int seg_blocks = sbi->blocks_per_seg; > unsigned int vblocks; > unsigned int dirty_threshold = max(am->max_candidate_count, > am->candidate_ratio * > @@ -629,7 +629,7 @@ static void atssr_lookup_victim(struct f2fs_sb_info *sbi, > f2fs_bug_on(sbi, !vblocks); > > /* rare case */ > - if (vblocks == seg_blocks) > + if (vblocks == BLKS_PER_SEG(sbi)) > goto skip_node; > > iter++; > @@ -755,7 +755,7 @@ int f2fs_get_victim(struct f2fs_sb_info *sbi, unsigned int *result, > int ret = 0; > > mutex_lock(&dirty_i->seglist_lock); > - last_segment = MAIN_SECS(sbi) * sbi->segs_per_sec; > + last_segment = MAIN_SECS(sbi) * SEGS_PER_SEC(sbi); > > p.alloc_mode = alloc_mode; > p.age = age; > @@ -896,7 +896,7 @@ int f2fs_get_victim(struct f2fs_sb_info *sbi, unsigned int *result, > else > sm->last_victim[p.gc_mode] = segno + p.ofs_unit; > sm->last_victim[p.gc_mode] %= > - (MAIN_SECS(sbi) * sbi->segs_per_sec); > + (MAIN_SECS(sbi) * SEGS_PER_SEC(sbi)); > break; > } > } > @@ -1670,7 +1670,7 @@ static int do_garbage_collect(struct f2fs_sb_info *sbi, > struct f2fs_summary_block *sum; > struct blk_plug plug; > unsigned int segno = start_segno; > - unsigned int end_segno = start_segno + sbi->segs_per_sec; > + unsigned int end_segno = start_segno + SEGS_PER_SEC(sbi); > int seg_freed = 0, migrated = 0; > unsigned char type = IS_DATASEG(get_seg_entry(sbi, segno)->type) ? > SUM_TYPE_DATA : SUM_TYPE_NODE; > @@ -1678,7 +1678,7 @@ static int do_garbage_collect(struct f2fs_sb_info *sbi, > int submitted = 0; > > if (__is_large_section(sbi)) > - end_segno = rounddown(end_segno, sbi->segs_per_sec); > + end_segno = rounddown(end_segno, SEGS_PER_SEC(sbi)); > > /* > * zone-capacity can be less than zone-size in zoned devices, > @@ -1686,7 +1686,7 @@ static int do_garbage_collect(struct f2fs_sb_info *sbi, > * calculate the end segno in the zone which can be garbage collected > */ > if (f2fs_sb_has_blkzoned(sbi)) > - end_segno -= sbi->segs_per_sec - > + end_segno -= SEGS_PER_SEC(sbi) - > f2fs_usable_segs_in_sec(sbi, segno); > > sanity_check_seg_type(sbi, get_seg_entry(sbi, segno)->type); > @@ -1986,7 +1986,7 @@ static int free_segment_range(struct f2fs_sb_info *sbi, > > /* Force block allocation for GC */ > MAIN_SECS(sbi) -= secs; > - start = MAIN_SECS(sbi) * sbi->segs_per_sec; > + start = MAIN_SECS(sbi) * SEGS_PER_SEC(sbi); > end = MAIN_SEGS(sbi) - 1; > > mutex_lock(&DIRTY_I(sbi)->seglist_lock); > @@ -2004,7 +2004,7 @@ static int free_segment_range(struct f2fs_sb_info *sbi, > f2fs_allocate_segment_for_resize(sbi, type, start, end); > > /* do GC to move out valid blocks in the range */ > - for (segno = start; segno <= end; segno += sbi->segs_per_sec) { > + for (segno = start; segno <= end; segno += SEGS_PER_SEC(sbi)) { > struct gc_inode_list gc_list = { > .ilist = LIST_HEAD_INIT(gc_list.ilist), > .iroot = RADIX_TREE_INIT(gc_list.iroot, GFP_NOFS), > @@ -2048,7 +2048,7 @@ static void update_sb_metadata(struct f2fs_sb_info *sbi, int secs) > int segment_count; > int segment_count_main; > long long block_count; > - int segs = secs * sbi->segs_per_sec; > + int segs = secs * SEGS_PER_SEC(sbi); > > f2fs_down_write(&sbi->sb_lock); > > @@ -2061,7 +2061,7 @@ static void update_sb_metadata(struct f2fs_sb_info *sbi, int secs) > raw_sb->segment_count = cpu_to_le32(segment_count + segs); > raw_sb->segment_count_main = cpu_to_le32(segment_count_main + segs); > raw_sb->block_count = cpu_to_le64(block_count + > - (long long)segs * sbi->blocks_per_seg); > + (long long)(segs << sbi->log_blocks_per_seg)); > if (f2fs_is_multi_device(sbi)) { > int last_dev = sbi->s_ndevs - 1; > int dev_segs = > @@ -2076,8 +2076,8 @@ static void update_sb_metadata(struct f2fs_sb_info *sbi, int secs) > > static void update_fs_metadata(struct f2fs_sb_info *sbi, int secs) > { > - int segs = secs * sbi->segs_per_sec; > - long long blks = (long long)segs * sbi->blocks_per_seg; > + int segs = secs * SEGS_PER_SEC(sbi); > + long long blks = (long long)(segs << sbi->log_blocks_per_seg); > long long user_block_count = > le64_to_cpu(F2FS_CKPT(sbi)->user_block_count); > > @@ -2119,7 +2119,7 @@ int f2fs_resize_fs(struct file *filp, __u64 block_count) > int last_dev = sbi->s_ndevs - 1; > __u64 last_segs = FDEV(last_dev).total_segments; > > - if (block_count + last_segs * sbi->blocks_per_seg <= > + if (block_count + (last_segs << sbi->log_blocks_per_seg) <= > old_block_count) > return -EINVAL; > } > diff --git a/fs/f2fs/node.c b/fs/f2fs/node.c > index 1d898a16f05a..51241996b9ec 100644 > --- a/fs/f2fs/node.c > +++ b/fs/f2fs/node.c > @@ -2841,7 +2841,7 @@ int f2fs_restore_node_summary(struct f2fs_sb_info *sbi, > int i, idx, last_offset, nrpages; > > /* scan the node segment */ > - last_offset = sbi->blocks_per_seg; > + last_offset = BLKS_PER_SEG(sbi); > addr = START_BLOCK(sbi, segno); > sum_entry = &sum->entries[0]; > > @@ -3158,7 +3158,7 @@ static int __get_nat_bitmaps(struct f2fs_sb_info *sbi) > if (!is_set_ckpt_flags(sbi, CP_NAT_BITS_FLAG)) > return 0; > > - nat_bits_addr = __start_cp_addr(sbi) + sbi->blocks_per_seg - > + nat_bits_addr = __start_cp_addr(sbi) + BLKS_PER_SEG(sbi) - > nm_i->nat_bits_blocks; > for (i = 0; i < nm_i->nat_bits_blocks; i++) { > struct page *page; > diff --git a/fs/f2fs/node.h b/fs/f2fs/node.h > index 5bd16a95eef8..6aea13024ac1 100644 > --- a/fs/f2fs/node.h > +++ b/fs/f2fs/node.h > @@ -208,10 +208,10 @@ static inline pgoff_t current_nat_addr(struct f2fs_sb_info *sbi, nid_t start) > > block_addr = (pgoff_t)(nm_i->nat_blkaddr + > (block_off << 1) - > - (block_off & (sbi->blocks_per_seg - 1))); > + (block_off & (BLKS_PER_SEG(sbi) - 1))); > > if (f2fs_test_bit(block_off, nm_i->nat_bitmap)) > - block_addr += sbi->blocks_per_seg; > + block_addr += BLKS_PER_SEG(sbi); > > return block_addr; > } > diff --git a/fs/f2fs/recovery.c b/fs/f2fs/recovery.c > index aad1d1a9b3d6..b3baec666afe 100644 > --- a/fs/f2fs/recovery.c > +++ b/fs/f2fs/recovery.c > @@ -354,7 +354,7 @@ static unsigned int adjust_por_ra_blocks(struct f2fs_sb_info *sbi, > if (blkaddr + 1 == next_blkaddr) > ra_blocks = min_t(unsigned int, RECOVERY_MAX_RA_BLOCKS, > ra_blocks * 2); > - else if (next_blkaddr % sbi->blocks_per_seg) > + else if (next_blkaddr % BLKS_PER_SEG(sbi)) > ra_blocks = max_t(unsigned int, RECOVERY_MIN_RA_BLOCKS, > ra_blocks / 2); > return ra_blocks; > diff --git a/fs/f2fs/segment.c b/fs/f2fs/segment.c > index e5759813276a..1518f1287c28 100644 > --- a/fs/f2fs/segment.c > +++ b/fs/f2fs/segment.c > @@ -448,8 +448,8 @@ static inline bool excess_dirty_threshold(struct f2fs_sb_info *sbi) > unsigned int nodes = get_pages(sbi, F2FS_DIRTY_NODES); > unsigned int meta = get_pages(sbi, F2FS_DIRTY_META); > unsigned int imeta = get_pages(sbi, F2FS_DIRTY_IMETA); > - unsigned int threshold = sbi->blocks_per_seg * factor * > - DEFAULT_DIRTY_THRESHOLD; > + unsigned int threshold = (factor * DEFAULT_DIRTY_THRESHOLD) << > + sbi->log_blocks_per_seg; > unsigned int global_threshold = threshold * 3 / 2; > > if (dents >= threshold || qdata >= threshold || > @@ -1134,8 +1134,7 @@ static void __check_sit_bitmap(struct f2fs_sb_info *sbi, > struct seg_entry *sentry; > unsigned int segno; > block_t blk = start; > - unsigned long offset, size, max_blocks = sbi->blocks_per_seg; > - unsigned long *map; > + unsigned long offset, size, *map; > > while (blk < end) { > segno = GET_SEGNO(sbi, blk); > @@ -1145,7 +1144,7 @@ static void __check_sit_bitmap(struct f2fs_sb_info *sbi, > if (end < START_BLOCK(sbi, segno + 1)) > size = GET_BLKOFF_FROM_SEG0(sbi, end); > else > - size = max_blocks; > + size = BLKS_PER_SEG(sbi); > map = (unsigned long *)(sentry->cur_valid_map); > offset = __find_rev_next_bit(map, size, offset); > f2fs_bug_on(sbi, offset != size); > @@ -2044,7 +2043,6 @@ static bool add_discard_addrs(struct f2fs_sb_info *sbi, struct cp_control *cpc, > bool check_only) > { > int entries = SIT_VBLOCK_MAP_SIZE / sizeof(unsigned long); > - int max_blocks = sbi->blocks_per_seg; > struct seg_entry *se = get_seg_entry(sbi, cpc->trim_start); > unsigned long *cur_map = (unsigned long *)se->cur_valid_map; > unsigned long *ckpt_map = (unsigned long *)se->ckpt_valid_map; > @@ -2056,8 +2054,9 @@ static bool add_discard_addrs(struct f2fs_sb_info *sbi, struct cp_control *cpc, > struct list_head *head = &SM_I(sbi)->dcc_info->entry_list; > int i; > > - if (se->valid_blocks == max_blocks || !f2fs_hw_support_discard(sbi) || > - !f2fs_block_unit_discard(sbi)) > + if (se->valid_blocks == BLKS_PER_SEG(sbi) || > + !f2fs_hw_support_discard(sbi) || > + !f2fs_block_unit_discard(sbi)) > return false; > > if (!force) { > @@ -2074,13 +2073,14 @@ static bool add_discard_addrs(struct f2fs_sb_info *sbi, struct cp_control *cpc, > > while (force || SM_I(sbi)->dcc_info->nr_discards <= > SM_I(sbi)->dcc_info->max_discards) { > - start = __find_rev_next_bit(dmap, max_blocks, end + 1); > - if (start >= max_blocks) > + start = __find_rev_next_bit(dmap, BLKS_PER_SEG(sbi), end + 1); > + if (start >= BLKS_PER_SEG(sbi)) > break; > > - end = __find_rev_next_zero_bit(dmap, max_blocks, start + 1); > - if (force && start && end != max_blocks > - && (end - start) < cpc->trim_minlen) > + end = __find_rev_next_zero_bit(dmap, > + BLKS_PER_SEG(sbi), start + 1); > + if (force && start && end != BLKS_PER_SEG(sbi) && > + (end - start) < cpc->trim_minlen) > continue; > > if (check_only) > @@ -2162,8 +2162,8 @@ void f2fs_clear_prefree_segments(struct f2fs_sb_info *sbi, > start + 1); > > if (section_alignment) { > - start = rounddown(start, sbi->segs_per_sec); > - end = roundup(end, sbi->segs_per_sec); > + start = rounddown(start, SEGS_PER_SEC(sbi)); > + end = roundup(end, SEGS_PER_SEC(sbi)); > } > > for (i = start; i < end; i++) { > @@ -2191,9 +2191,9 @@ void f2fs_clear_prefree_segments(struct f2fs_sb_info *sbi, > if (!IS_CURSEC(sbi, secno) && > !get_valid_blocks(sbi, start, true)) > f2fs_issue_discard(sbi, START_BLOCK(sbi, start_segno), > - sbi->segs_per_sec << sbi->log_blocks_per_seg); > + BLKS_PER_SEC(sbi)); > > - start = start_segno + sbi->segs_per_sec; > + start = start_segno + SEGS_PER_SEC(sbi); > if (start < end) > goto next; > else > @@ -2212,7 +2212,7 @@ void f2fs_clear_prefree_segments(struct f2fs_sb_info *sbi, > find_next: > if (is_valid) { > next_pos = find_next_zero_bit_le(entry->discard_map, > - sbi->blocks_per_seg, cur_pos); > + BLKS_PER_SEG(sbi), cur_pos); > len = next_pos - cur_pos; > > if (f2fs_sb_has_blkzoned(sbi) || > @@ -2224,13 +2224,13 @@ void f2fs_clear_prefree_segments(struct f2fs_sb_info *sbi, > total_len += len; > } else { > next_pos = find_next_bit_le(entry->discard_map, > - sbi->blocks_per_seg, cur_pos); > + BLKS_PER_SEG(sbi), cur_pos); > } > skip: > cur_pos = next_pos; > is_valid = !is_valid; > > - if (cur_pos < sbi->blocks_per_seg) > + if (cur_pos < BLKS_PER_SEG(sbi)) > goto find_next; > > release_discard_addr(entry); > @@ -2279,7 +2279,7 @@ static int create_discard_cmd_control(struct f2fs_sb_info *sbi) > dcc->max_ordered_discard = DEFAULT_MAX_ORDERED_DISCARD_GRANULARITY; > dcc->discard_io_aware = DPOLICY_IO_AWARE_ENABLE; > if (F2FS_OPTION(sbi).discard_unit == DISCARD_UNIT_SEGMENT) > - dcc->discard_granularity = sbi->blocks_per_seg; > + dcc->discard_granularity = BLKS_PER_SEG(sbi); > else if (F2FS_OPTION(sbi).discard_unit == DISCARD_UNIT_SECTION) > dcc->discard_granularity = BLKS_PER_SEC(sbi); > > @@ -2542,7 +2542,7 @@ static unsigned short f2fs_curseg_valid_blocks(struct f2fs_sb_info *sbi, int typ > struct curseg_info *curseg = CURSEG_I(sbi, type); > > if (sbi->ckpt->alloc_type[type] == SSR) > - return sbi->blocks_per_seg; > + return BLKS_PER_SEG(sbi); > return curseg->next_blkoff; > } > > @@ -2630,7 +2630,7 @@ static int is_next_segment_free(struct f2fs_sb_info *sbi, > unsigned int segno = curseg->segno + 1; > struct free_segmap_info *free_i = FREE_I(sbi); > > - if (segno < MAIN_SEGS(sbi) && segno % sbi->segs_per_sec) > + if (segno < MAIN_SEGS(sbi) && segno % SEGS_PER_SEC(sbi)) > return !test_bit(segno, free_i->free_segmap); > return 0; > } > @@ -2654,7 +2654,7 @@ static void get_new_segment(struct f2fs_sb_info *sbi, > > spin_lock(&free_i->segmap_lock); > > - if (!new_sec && ((*newseg + 1) % sbi->segs_per_sec)) { > + if (!new_sec && ((*newseg + 1) % SEGS_PER_SEC(sbi))) { > segno = find_next_zero_bit(free_i->free_segmap, > GET_SEG_FROM_SEC(sbi, hint + 1), *newseg + 1); > if (segno < GET_SEG_FROM_SEC(sbi, hint + 1)) > @@ -2757,9 +2757,8 @@ static unsigned int __get_next_segno(struct f2fs_sb_info *sbi, int type) > > sanity_check_seg_type(sbi, seg_type); > if (f2fs_need_rand_seg(sbi)) > - return get_random_u32_below(MAIN_SECS(sbi) * sbi->segs_per_sec); > + return get_random_u32_below(MAIN_SECS(sbi) * SEGS_PER_SEC(sbi)); > > - /* if segs_per_sec is large than 1, we need to keep original policy. */ > if (__is_large_section(sbi)) > return curseg->segno; > > @@ -2827,7 +2826,7 @@ static int __next_free_blkoff(struct f2fs_sb_info *sbi, > for (i = 0; i < entries; i++) > target_map[i] = ckpt_map[i] | cur_map[i]; > > - return __find_rev_next_zero_bit(target_map, sbi->blocks_per_seg, start); > + return __find_rev_next_zero_bit(target_map, BLKS_PER_SEG(sbi), start); > } > > static int f2fs_find_next_ssr_block(struct f2fs_sb_info *sbi, > @@ -2838,7 +2837,7 @@ static int f2fs_find_next_ssr_block(struct f2fs_sb_info *sbi, > > bool f2fs_segment_has_free_slot(struct f2fs_sb_info *sbi, int segno) > { > - return __next_free_blkoff(sbi, segno, 0) < sbi->blocks_per_seg; > + return __next_free_blkoff(sbi, segno, 0) < BLKS_PER_SEG(sbi); > } > > /* > @@ -3238,8 +3237,8 @@ int f2fs_trim_fs(struct f2fs_sb_info *sbi, struct fstrim_range *range) > end_segno = (end >= MAX_BLKADDR(sbi)) ? MAIN_SEGS(sbi) - 1 : > GET_SEGNO(sbi, end); > if (need_align) { > - start_segno = rounddown(start_segno, sbi->segs_per_sec); > - end_segno = roundup(end_segno + 1, sbi->segs_per_sec) - 1; > + start_segno = rounddown(start_segno, SEGS_PER_SEC(sbi)); > + end_segno = roundup(end_segno + 1, SEGS_PER_SEC(sbi)) - 1; > } > > cpc.reason = CP_DISCARD; > @@ -3437,7 +3436,7 @@ void f2fs_allocate_data_block(struct f2fs_sb_info *sbi, struct page *page, > } > *new_blkaddr = NEXT_FREE_BLKADDR(sbi, curseg); > > - f2fs_bug_on(sbi, curseg->next_blkoff >= sbi->blocks_per_seg); > + f2fs_bug_on(sbi, curseg->next_blkoff >= BLKS_PER_SEG(sbi)); > > f2fs_wait_discard_bio(sbi, *new_blkaddr); > > @@ -3881,7 +3880,7 @@ static int read_compacted_summaries(struct f2fs_sb_info *sbi) > seg_i->next_blkoff = blk_off; > > if (seg_i->alloc_type == SSR) > - blk_off = sbi->blocks_per_seg; > + blk_off = BLKS_PER_SEG(sbi); > > for (j = 0; j < blk_off; j++) { > struct f2fs_summary *s; > @@ -3949,7 +3948,7 @@ static int read_normal_summaries(struct f2fs_sb_info *sbi, int type) > struct f2fs_summary *ns = &sum->entries[0]; > int i; > > - for (i = 0; i < sbi->blocks_per_seg; i++, ns++) { > + for (i = 0; i < BLKS_PER_SEG(sbi); i++, ns++) { > ns->version = 0; > ns->ofs_in_node = 0; > } > @@ -4582,21 +4581,20 @@ static int build_sit_entries(struct f2fs_sb_info *sbi) > > sit_valid_blocks[SE_PAGETYPE(se)] += se->valid_blocks; > > - if (f2fs_block_unit_discard(sbi)) { > - /* build discard map only one time */ > - if (is_set_ckpt_flags(sbi, CP_TRIMMED_FLAG)) { > - memset(se->discard_map, 0xff, > + if (!f2fs_block_unit_discard(sbi)) > + goto init_discard_map_done; > + > + /* build discard map only one time */ > + if (is_set_ckpt_flags(sbi, CP_TRIMMED_FLAG)) { > + memset(se->discard_map, 0xff, > SIT_VBLOCK_MAP_SIZE); > - } else { > - memcpy(se->discard_map, > - se->cur_valid_map, > + goto init_discard_map_done; > + } > + memcpy(se->discard_map, se->cur_valid_map, > SIT_VBLOCK_MAP_SIZE); > - sbi->discard_blks += > - sbi->blocks_per_seg - > + sbi->discard_blks += BLKS_PER_SEG(sbi) - > se->valid_blocks; > - } > - } > - > +init_discard_map_done: > if (__is_large_section(sbi)) > get_sec_entry(sbi, start)->valid_blocks += > se->valid_blocks; > @@ -4736,7 +4734,7 @@ static void init_dirty_segmap(struct f2fs_sb_info *sbi) > return; > > mutex_lock(&dirty_i->seglist_lock); > - for (segno = 0; segno < MAIN_SEGS(sbi); segno += sbi->segs_per_sec) { > + for (segno = 0; segno < MAIN_SEGS(sbi); segno += SEGS_PER_SEC(sbi)) { > valid_blocks = get_valid_blocks(sbi, segno, true); > secno = GET_SEC_FROM_SEG(sbi, segno); > > @@ -4835,7 +4833,7 @@ static int sanity_check_curseg(struct f2fs_sb_info *sbi) > if (curseg->alloc_type == SSR) > continue; > > - for (blkofs += 1; blkofs < sbi->blocks_per_seg; blkofs++) { > + for (blkofs += 1; blkofs < BLKS_PER_SEG(sbi); blkofs++) { > if (!f2fs_test_bit(blkofs, se->cur_valid_map)) > continue; > out: > @@ -5114,7 +5112,7 @@ static inline unsigned int f2fs_usable_zone_blks_in_seg( > unsigned int secno; > > if (!sbi->unusable_blocks_per_sec) > - return sbi->blocks_per_seg; > + return BLKS_PER_SEG(sbi); > > secno = GET_SEC_FROM_SEG(sbi, segno); > seg_start = START_BLOCK(sbi, segno); > @@ -5129,10 +5127,10 @@ static inline unsigned int f2fs_usable_zone_blks_in_seg( > */ > if (seg_start >= sec_cap_blkaddr) > return 0; > - if (seg_start + sbi->blocks_per_seg > sec_cap_blkaddr) > + if (seg_start + BLKS_PER_SEG(sbi) > sec_cap_blkaddr) > return sec_cap_blkaddr - seg_start; > > - return sbi->blocks_per_seg; > + return BLKS_PER_SEG(sbi); > } > #else > int f2fs_fix_curseg_write_pointer(struct f2fs_sb_info *sbi) > @@ -5158,7 +5156,7 @@ unsigned int f2fs_usable_blks_in_seg(struct f2fs_sb_info *sbi, > if (f2fs_sb_has_blkzoned(sbi)) > return f2fs_usable_zone_blks_in_seg(sbi, segno); > > - return sbi->blocks_per_seg; > + return BLKS_PER_SEG(sbi); > } > > unsigned int f2fs_usable_segs_in_sec(struct f2fs_sb_info *sbi, > @@ -5167,7 +5165,7 @@ unsigned int f2fs_usable_segs_in_sec(struct f2fs_sb_info *sbi, > if (f2fs_sb_has_blkzoned(sbi)) > return CAP_SEGS_PER_SEC(sbi); > > - return sbi->segs_per_sec; > + return SEGS_PER_SEC(sbi); > } > > /* > @@ -5182,14 +5180,14 @@ static void init_min_max_mtime(struct f2fs_sb_info *sbi) > > sit_i->min_mtime = ULLONG_MAX; > > - for (segno = 0; segno < MAIN_SEGS(sbi); segno += sbi->segs_per_sec) { > + for (segno = 0; segno < MAIN_SEGS(sbi); segno += SEGS_PER_SEC(sbi)) { > unsigned int i; > unsigned long long mtime = 0; > > - for (i = 0; i < sbi->segs_per_sec; i++) > + for (i = 0; i < SEGS_PER_SEC(sbi); i++) > mtime += get_seg_entry(sbi, segno + i)->mtime; > > - mtime = div_u64(mtime, sbi->segs_per_sec); > + mtime = div_u64(mtime, SEGS_PER_SEC(sbi)); > > if (sit_i->min_mtime > mtime) > sit_i->min_mtime = mtime; > @@ -5228,7 +5226,7 @@ int f2fs_build_segment_manager(struct f2fs_sb_info *sbi) > sm_info->ipu_policy = BIT(F2FS_IPU_FSYNC); > sm_info->min_ipu_util = DEF_MIN_IPU_UTIL; > sm_info->min_fsync_blocks = DEF_MIN_FSYNC_BLOCKS; > - sm_info->min_seq_blocks = sbi->blocks_per_seg; > + sm_info->min_seq_blocks = BLKS_PER_SEG(sbi); > sm_info->min_hot_blocks = DEF_MIN_HOT_BLOCKS; > sm_info->min_ssr_sections = reserved_sections(sbi); > > diff --git a/fs/f2fs/segment.h b/fs/f2fs/segment.h > index 8129be788bd5..febcfbadcdfa 100644 > --- a/fs/f2fs/segment.h > +++ b/fs/f2fs/segment.h > @@ -48,21 +48,21 @@ static inline void sanity_check_seg_type(struct f2fs_sb_info *sbi, > > #define IS_CURSEC(sbi, secno) \ > (((secno) == CURSEG_I(sbi, CURSEG_HOT_DATA)->segno / \ > - (sbi)->segs_per_sec) || \ > + SEGS_PER_SEC(sbi)) || \ > ((secno) == CURSEG_I(sbi, CURSEG_WARM_DATA)->segno / \ > - (sbi)->segs_per_sec) || \ > + SEGS_PER_SEC(sbi)) || \ > ((secno) == CURSEG_I(sbi, CURSEG_COLD_DATA)->segno / \ > - (sbi)->segs_per_sec) || \ > + SEGS_PER_SEC(sbi)) || \ > ((secno) == CURSEG_I(sbi, CURSEG_HOT_NODE)->segno / \ > - (sbi)->segs_per_sec) || \ > + SEGS_PER_SEC(sbi)) || \ > ((secno) == CURSEG_I(sbi, CURSEG_WARM_NODE)->segno / \ > - (sbi)->segs_per_sec) || \ > + SEGS_PER_SEC(sbi)) || \ > ((secno) == CURSEG_I(sbi, CURSEG_COLD_NODE)->segno / \ > - (sbi)->segs_per_sec) || \ > + SEGS_PER_SEC(sbi)) || \ > ((secno) == CURSEG_I(sbi, CURSEG_COLD_DATA_PINNED)->segno / \ > - (sbi)->segs_per_sec) || \ > + SEGS_PER_SEC(sbi)) || \ > ((secno) == CURSEG_I(sbi, CURSEG_ALL_DATA_ATGC)->segno / \ > - (sbi)->segs_per_sec)) > + SEGS_PER_SEC(sbi))) > > #define MAIN_BLKADDR(sbi) \ > (SM_I(sbi) ? SM_I(sbi)->main_blkaddr : \ > @@ -93,24 +93,22 @@ static inline void sanity_check_seg_type(struct f2fs_sb_info *sbi, > #define GET_SEGNO_FROM_SEG0(sbi, blk_addr) \ > (GET_SEGOFF_FROM_SEG0(sbi, blk_addr) >> (sbi)->log_blocks_per_seg) > #define GET_BLKOFF_FROM_SEG0(sbi, blk_addr) \ > - (GET_SEGOFF_FROM_SEG0(sbi, blk_addr) & ((sbi)->blocks_per_seg - 1)) > + (GET_SEGOFF_FROM_SEG0(sbi, blk_addr) & (BLKS_PER_SEG(sbi) - 1)) > > #define GET_SEGNO(sbi, blk_addr) \ > ((!__is_valid_data_blkaddr(blk_addr)) ? \ > NULL_SEGNO : GET_L2R_SEGNO(FREE_I(sbi), \ > GET_SEGNO_FROM_SEG0(sbi, blk_addr))) > -#define BLKS_PER_SEC(sbi) \ > - ((sbi)->segs_per_sec * (sbi)->blocks_per_seg) > #define CAP_BLKS_PER_SEC(sbi) \ > - ((sbi)->segs_per_sec * (sbi)->blocks_per_seg - \ > + (SEGS_PER_SEC(sbi) * BLKS_PER_SEG(sbi) - \ > (sbi)->unusable_blocks_per_sec) > #define CAP_SEGS_PER_SEC(sbi) \ > - ((sbi)->segs_per_sec - ((sbi)->unusable_blocks_per_sec >>\ > + (SEGS_PER_SEC(sbi) - ((sbi)->unusable_blocks_per_sec >> \ > (sbi)->log_blocks_per_seg)) > #define GET_SEC_FROM_SEG(sbi, segno) \ > - (((segno) == -1) ? -1 : (segno) / (sbi)->segs_per_sec) > + (((segno) == -1) ? -1 : (segno) / SEGS_PER_SEC(sbi)) > #define GET_SEG_FROM_SEC(sbi, secno) \ > - ((secno) * (sbi)->segs_per_sec) > + ((secno) * SEGS_PER_SEC(sbi)) > #define GET_ZONE_FROM_SEC(sbi, secno) \ > (((secno) == -1) ? -1 : (secno) / (sbi)->secs_per_zone) > #define GET_ZONE_FROM_SEG(sbi, segno) \ > @@ -364,7 +362,7 @@ static inline unsigned int get_ckpt_valid_blocks(struct f2fs_sb_info *sbi, > unsigned int blocks = 0; > int i; > > - for (i = 0; i < sbi->segs_per_sec; i++, start_segno++) { > + for (i = 0; i < SEGS_PER_SEC(sbi); i++, start_segno++) { > struct seg_entry *se = get_seg_entry(sbi, start_segno); > > blocks += se->ckpt_valid_blocks; > @@ -449,7 +447,7 @@ static inline void __set_free(struct f2fs_sb_info *sbi, unsigned int segno) > free_i->free_segments++; > > next = find_next_bit(free_i->free_segmap, > - start_segno + sbi->segs_per_sec, start_segno); > + start_segno + SEGS_PER_SEC(sbi), start_segno); > if (next >= start_segno + usable_segs) { > clear_bit(secno, free_i->free_secmap); > free_i->free_sections++; > @@ -485,7 +483,7 @@ static inline void __set_test_and_free(struct f2fs_sb_info *sbi, > if (!inmem && IS_CURSEC(sbi, secno)) > goto skip_free; > next = find_next_bit(free_i->free_segmap, > - start_segno + sbi->segs_per_sec, start_segno); > + start_segno + SEGS_PER_SEC(sbi), start_segno); > if (next >= start_segno + usable_segs) { > if (test_and_clear_bit(secno, free_i->free_secmap)) > free_i->free_sections++; > @@ -793,10 +791,10 @@ static inline int check_block_count(struct f2fs_sb_info *sbi, > return -EFSCORRUPTED; > } > > - if (usable_blks_per_seg < sbi->blocks_per_seg) > + if (usable_blks_per_seg < BLKS_PER_SEG(sbi)) > f2fs_bug_on(sbi, find_next_bit_le(&raw_sit->valid_map, > - sbi->blocks_per_seg, > - usable_blks_per_seg) != sbi->blocks_per_seg); > + BLKS_PER_SEG(sbi), > + usable_blks_per_seg) != BLKS_PER_SEG(sbi)); > > /* check segment usage, and check boundary of a given segment number */ > if (unlikely(GET_SIT_VBLOCKS(raw_sit) > usable_blks_per_seg > @@ -915,9 +913,9 @@ static inline int nr_pages_to_skip(struct f2fs_sb_info *sbi, int type) > return 0; > > if (type == DATA) > - return sbi->blocks_per_seg; > + return BLKS_PER_SEG(sbi); > else if (type == NODE) > - return 8 * sbi->blocks_per_seg; > + return 8 * BLKS_PER_SEG(sbi); > else if (type == META) > return 8 * BIO_MAX_VECS; > else > diff --git a/fs/f2fs/super.c b/fs/f2fs/super.c > index f1516fd5088a..c0688c124aa7 100644 > --- a/fs/f2fs/super.c > +++ b/fs/f2fs/super.c > @@ -3643,7 +3643,7 @@ int f2fs_sanity_check_ckpt(struct f2fs_sb_info *sbi) > } > > main_segs = le32_to_cpu(raw_super->segment_count_main); > - blocks_per_seg = sbi->blocks_per_seg; > + blocks_per_seg = BLKS_PER_SEG(sbi); > > for (i = 0; i < NR_CURSEG_NODE_TYPE; i++) { > if (le32_to_cpu(ckpt->cur_node_segno[i]) >= main_segs || > @@ -3756,8 +3756,8 @@ static void init_sb_info(struct f2fs_sb_info *sbi) > sbi->secs_per_zone = le32_to_cpu(raw_super->secs_per_zone); > sbi->total_sections = le32_to_cpu(raw_super->section_count); > sbi->total_node_count = > - (le32_to_cpu(raw_super->segment_count_nat) / 2) > - * sbi->blocks_per_seg * NAT_ENTRY_PER_BLOCK; > + ((le32_to_cpu(raw_super->segment_count_nat) / 2) * > + NAT_ENTRY_PER_BLOCK) << sbi->log_blocks_per_seg; > F2FS_ROOT_INO(sbi) = le32_to_cpu(raw_super->root_ino); > F2FS_NODE_INO(sbi) = le32_to_cpu(raw_super->node_ino); > F2FS_META_INO(sbi) = le32_to_cpu(raw_super->meta_ino); > @@ -3766,7 +3766,7 @@ static void init_sb_info(struct f2fs_sb_info *sbi) > sbi->next_victim_seg[BG_GC] = NULL_SEGNO; > sbi->next_victim_seg[FG_GC] = NULL_SEGNO; > sbi->max_victim_search = DEF_MAX_VICTIM_SEARCH; > - sbi->migration_granularity = sbi->segs_per_sec; > + sbi->migration_granularity = SEGS_PER_SEC(sbi); > sbi->seq_file_ra_mul = MIN_RA_MUL; > sbi->max_fragment_chunk = DEF_FRAGMENT_SIZE; > sbi->max_fragment_hole = DEF_FRAGMENT_SIZE; > diff --git a/fs/f2fs/sysfs.c b/fs/f2fs/sysfs.c > index a7ec55c7bb20..906d2af2d849 100644 > --- a/fs/f2fs/sysfs.c > +++ b/fs/f2fs/sysfs.c > @@ -493,8 +493,8 @@ static ssize_t __sbi_store(struct f2fs_attr *a, > spin_lock(&sbi->stat_lock); > if (t > (unsigned long)(sbi->user_block_count - > F2FS_OPTION(sbi).root_reserved_blocks - > - sbi->blocks_per_seg * > - SM_I(sbi)->additional_reserved_segments)) { > + (SM_I(sbi)->additional_reserved_segments << > + sbi->log_blocks_per_seg))) { > spin_unlock(&sbi->stat_lock); > return -EINVAL; > } > @@ -551,7 +551,7 @@ static ssize_t __sbi_store(struct f2fs_attr *a, > } > > if (!strcmp(a->attr.name, "migration_granularity")) { > - if (t == 0 || t > sbi->segs_per_sec) > + if (t == 0 || t > SEGS_PER_SEC(sbi)) > return -EINVAL; > } >