Received: by 2002:a05:6a10:206:0:0:0:0 with SMTP id 6csp5068625pxj; Wed, 26 May 2021 01:57:30 -0700 (PDT) X-Google-Smtp-Source: ABdhPJzP+mu3p/LrcZ6G5jKKqrWNWWMAwPnrAfzrE/aylItlUMyDyr4lfb022EzIOzU/QO+CNKN4 X-Received: by 2002:a17:907:760b:: with SMTP id jx11mr32958799ejc.344.1622019450512; Wed, 26 May 2021 01:57:30 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1622019450; cv=none; d=google.com; s=arc-20160816; b=pu2RURW3Ttp9sQd0wk+h9pbC/L7MDRBmBsZoU3DnFmNdZ6UVG36Drd/Hbo4vDCzShW U7S025MrkfgNroK0+wRhIPHaYxR+3GcMHKb6DGRR6zQwNrqqgWj7uXM2ch2vWXvkSSXi sBr7vwcSuxWKm5zyYKZ0rY4iq9mjTQWM10UlXFb/N+RmrCY/YOukT7WIk2x8QVsF6BKE hRuxo7knoU5eXkLaborZI12FDbsXCvSZLX7qhXc1vXiRsiC0BtmEaXJqufMmNjy8OSQ4 TjCnLGGSpZPUQwO7K2i7S0RvKk8N/F6JiUd/a0uKQYIpQYBVRteX+5pNkfrAqYVm38M/ oGCA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:content-language :in-reply-to:mime-version:user-agent:date:message-id:from:references :cc:to:subject; bh=1g/ePnFr1a1lQc4K9YZkKN3h6ree6MlxEvjguyf02Fc=; b=ftiWtDF6208Gord9B8p5zaW7oE/cRuhxvoM16a5ve20QItt8vqhPZuJyui70rVaBZ/ IhyJuarokvSonPBmvh8EnRi7aVydUygzbEUCytAPipM3HdfvuHW1th12IlZFDMgF6UX1 nOUfSXGfJtq5j28WkJBJPhaKFbznd2cvVrTac7+xIqNtzqTPea6ZOFDNhnwBb5qWS38b B/qyykHgSdKmit8FOwWb2Aq5KUrKMGFsFE8BWFXZy94vFIY/D0KGBu1T4lxRrF3diT5D GGCAtOtpTq5mCdVF02ldi9No+QggMBCYc2xzsa403KUzMtRvCqLUbVE0/u2ag+dyGxhd UWzg== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=huawei.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id m17si15548634edc.326.2021.05.26.01.57.02; Wed, 26 May 2021 01:57:30 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=huawei.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S233312AbhEZIyT (ORCPT + 99 others); Wed, 26 May 2021 04:54:19 -0400 Received: from szxga05-in.huawei.com ([45.249.212.191]:3973 "EHLO szxga05-in.huawei.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S232995AbhEZIyS (ORCPT ); Wed, 26 May 2021 04:54:18 -0400 Received: from dggems706-chm.china.huawei.com (unknown [172.30.72.59]) by szxga05-in.huawei.com (SkyGuard) with ESMTP id 4Fql4321NwzQs2x; Wed, 26 May 2021 16:49:07 +0800 (CST) Received: from dggemx753-chm.china.huawei.com (10.0.44.37) by dggems706-chm.china.huawei.com (10.3.19.183) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256) id 15.1.2176.2; Wed, 26 May 2021 16:52:44 +0800 Received: from [10.136.110.154] (10.136.110.154) by dggemx753-chm.china.huawei.com (10.0.44.37) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256_P256) id 15.1.2176.2; Wed, 26 May 2021 16:52:44 +0800 Subject: Re: [f2fs-dev] [PATCH v6] f2fs: compress: add compress_inode to cache compressed blocks To: Jaegeuk Kim , Chao Yu CC: , References: <20210520115150.4718-1-chao@kernel.org> <2c4db877-b5e6-1139-98b5-be2d9e7872be@kernel.org> From: Chao Yu Message-ID: Date: Wed, 26 May 2021 16:52:43 +0800 User-Agent: Mozilla/5.0 (Windows NT 10.0; WOW64; rv:52.0) Gecko/20100101 Thunderbird/52.9.1 MIME-Version: 1.0 In-Reply-To: Content-Type: text/plain; charset="windows-1252"; format=flowed Content-Language: en-US Content-Transfer-Encoding: 7bit X-Originating-IP: [10.136.110.154] X-ClientProxiedBy: dggemx703-chm.china.huawei.com (10.1.199.50) To dggemx753-chm.china.huawei.com (10.0.44.37) X-CFilter-Loop: Reflected Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 2021/5/25 22:01, Jaegeuk Kim wrote: > On 05/25, Chao Yu wrote: >> On 2021/5/25 21:02, Jaegeuk Kim wrote: >>> On 05/25, Jaegeuk Kim wrote: >>>> On 05/25, Chao Yu wrote: >>>>> Also, and queue this? >>>> >>>> Easy to get this? >>> >>> need GFP_NOFS? >> >> Not sure, I use __GFP_IO intentionally here to avoid __GFP_RECLAIM from >> GFP_NOFS, because in low memory case, I don't want to instead page cache >> of normal file with page cache of sbi->compress_inode. >> >> What is memory size in your vm? > > 4GB. If I set GFP_NOFS, I don't see the error anymore, at least. I applied below patch and don't see the warning message anymore. --- fs/f2fs/compress.c | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/fs/f2fs/compress.c b/fs/f2fs/compress.c index 701dd0f6f4ec..ed5b7fabc604 100644 --- a/fs/f2fs/compress.c +++ b/fs/f2fs/compress.c @@ -1703,7 +1703,7 @@ void f2fs_cache_compressed_page(struct f2fs_sb_info *sbi, struct page *page, avail_ram = si.totalram - si.totalhigh; /* free memory is lower than watermark, deny caching compress page */ - if (free_ram <= sbi->compress_watermark / 100 * avail_ram) + if (free_ram <= avail_ram / 100 * sbi->compress_watermark) return; /* cached page count exceed threshold, deny caching compress page */ -- 2.29.2 Thanks, > >> >> Thanks, >> >>> >>>> >>>> [ 1204.287099] kworker/u17:0: page allocation failure: order:0, mode:0x40(__GFP_IO), nodemask=(null),cpuset=/,mems_allowed=0 >>>> [ 1204.296932] CPU: 1 PID: 158 Comm: kworker/u17:0 Tainted: G OE 5.13.0-rc1-custom #1 >>>> [ 1204.300746] Hardware name: QEMU Standard PC (i440FX + PIIX, 1996), BIOS 1.14.0-2 04/01/2014 >>>> [ 1204.303458] Workqueue: f2fs_post_read_wq f2fs_post_read_work [f2fs] >>>> [ 1204.305772] Call Trace: >>>> [ 1204.307103] dump_stack+0x7d/0x9c >>>> [ 1204.308613] warn_alloc.cold+0x7b/0xdf >>>> [ 1204.310167] __alloc_pages_slowpath.constprop.0+0xd57/0xd80 >>>> [ 1204.312214] __alloc_pages+0x30e/0x330 >>>> [ 1204.313780] alloc_pages+0x87/0x110 >>>> [ 1204.315265] f2fs_cache_compressed_page+0x136/0x2d0 [f2fs] >>>> [ 1204.317142] ? dequeue_entity+0xdb/0x450 >>>> [ 1204.318708] f2fs_end_read_compressed_page+0x5c/0x70 [f2fs] >>>> [ 1204.320659] f2fs_post_read_work+0x11f/0x180 [f2fs] >>>> [ 1204.322442] process_one_work+0x220/0x3c0 >>>> [ 1204.324091] worker_thread+0x53/0x420 >>>> [ 1204.325577] kthread+0x12f/0x150 >>>> >>>>> >>>>> On 2021/5/20 19:51, Chao Yu wrote: >>>>>> From: Chao Yu >>>>>> >>>>>> Support to use address space of inner inode to cache compressed block, >>>>>> in order to improve cache hit ratio of random read. >>>>>> >>>>>> Signed-off-by: Chao Yu >>>>>> --- >>>>>> v6: >>>>>> - fix to cover COMPRESS_MAPPING() with CONFIG_F2FS_FS_COMPRESSION >>>>>> Documentation/filesystems/f2fs.rst | 3 + >>>>>> fs/f2fs/compress.c | 180 ++++++++++++++++++++++++++++- >>>>>> fs/f2fs/data.c | 41 ++++++- >>>>>> fs/f2fs/debug.c | 13 +++ >>>>>> fs/f2fs/f2fs.h | 71 +++++++++++- >>>>>> fs/f2fs/gc.c | 1 + >>>>>> fs/f2fs/inode.c | 21 +++- >>>>>> fs/f2fs/segment.c | 6 +- >>>>>> fs/f2fs/super.c | 35 +++++- >>>>>> include/linux/f2fs_fs.h | 1 + >>>>>> 10 files changed, 358 insertions(+), 14 deletions(-) >>>>>> >>>>>> diff --git a/Documentation/filesystems/f2fs.rst b/Documentation/filesystems/f2fs.rst >>>>>> index 992bf91eeec8..809c4d0a696f 100644 >>>>>> --- a/Documentation/filesystems/f2fs.rst >>>>>> +++ b/Documentation/filesystems/f2fs.rst >>>>>> @@ -289,6 +289,9 @@ compress_mode=%s Control file compression mode. This supports "fs" and "user" >>>>>> choosing the target file and the timing. The user can do manual >>>>>> compression/decompression on the compression enabled files using >>>>>> ioctls. >>>>>> +compress_cache Support to use address space of a filesystem managed inode to >>>>>> + cache compressed block, in order to improve cache hit ratio of >>>>>> + random read. >>>>>> inlinecrypt When possible, encrypt/decrypt the contents of encrypted >>>>>> files using the blk-crypto framework rather than >>>>>> filesystem-layer encryption. This allows the use of >>>>>> diff --git a/fs/f2fs/compress.c b/fs/f2fs/compress.c >>>>>> index d4f7371fb0d8..25e785e0d9fc 100644 >>>>>> --- a/fs/f2fs/compress.c >>>>>> +++ b/fs/f2fs/compress.c >>>>>> @@ -12,9 +12,11 @@ >>>>>> #include >>>>>> #include >>>>>> #include >>>>>> +#include >>>>>> #include "f2fs.h" >>>>>> #include "node.h" >>>>>> +#include "segment.h" >>>>>> #include >>>>>> static struct kmem_cache *cic_entry_slab; >>>>>> @@ -736,7 +738,7 @@ static int f2fs_compress_pages(struct compress_ctx *cc) >>>>>> return ret; >>>>>> } >>>>>> -static void f2fs_decompress_cluster(struct decompress_io_ctx *dic) >>>>>> +void f2fs_decompress_cluster(struct decompress_io_ctx *dic) >>>>>> { >>>>>> struct f2fs_sb_info *sbi = F2FS_I_SB(dic->inode); >>>>>> struct f2fs_inode_info *fi = F2FS_I(dic->inode); >>>>>> @@ -835,7 +837,8 @@ static void f2fs_decompress_cluster(struct decompress_io_ctx *dic) >>>>>> * page being waited on in the cluster, and if so, it decompresses the cluster >>>>>> * (or in the case of a failure, cleans up without actually decompressing). >>>>>> */ >>>>>> -void f2fs_end_read_compressed_page(struct page *page, bool failed) >>>>>> +void f2fs_end_read_compressed_page(struct page *page, bool failed, >>>>>> + block_t blkaddr) >>>>>> { >>>>>> struct decompress_io_ctx *dic = >>>>>> (struct decompress_io_ctx *)page_private(page); >>>>>> @@ -845,6 +848,9 @@ void f2fs_end_read_compressed_page(struct page *page, bool failed) >>>>>> if (failed) >>>>>> WRITE_ONCE(dic->failed, true); >>>>>> + else if (blkaddr) >>>>>> + f2fs_cache_compressed_page(sbi, page, >>>>>> + dic->inode->i_ino, blkaddr); >>>>>> if (atomic_dec_and_test(&dic->remaining_pages)) >>>>>> f2fs_decompress_cluster(dic); >>>>>> @@ -1660,6 +1666,176 @@ void f2fs_put_page_dic(struct page *page) >>>>>> f2fs_put_dic(dic); >>>>>> } >>>>>> +const struct address_space_operations f2fs_compress_aops = { >>>>>> + .releasepage = f2fs_release_page, >>>>>> + .invalidatepage = f2fs_invalidate_page, >>>>>> +}; >>>>>> + >>>>>> +struct address_space *COMPRESS_MAPPING(struct f2fs_sb_info *sbi) >>>>>> +{ >>>>>> + return sbi->compress_inode->i_mapping; >>>>>> +} >>>>>> + >>>>>> +void f2fs_invalidate_compress_page(struct f2fs_sb_info *sbi, block_t blkaddr) >>>>>> +{ >>>>>> + if (!sbi->compress_inode) >>>>>> + return; >>>>>> + invalidate_mapping_pages(COMPRESS_MAPPING(sbi), blkaddr, blkaddr); >>>>>> +} >>>>>> + >>>>>> +void f2fs_cache_compressed_page(struct f2fs_sb_info *sbi, struct page *page, >>>>>> + nid_t ino, block_t blkaddr) >>>>>> +{ >>>>>> + struct page *cpage; >>>>>> + int ret; >>>>>> + struct sysinfo si; >>>>>> + unsigned long free_ram, avail_ram; >>>>>> + >>>>>> + if (!test_opt(sbi, COMPRESS_CACHE)) >>>>>> + return; >>>>>> + >>>>>> + if (!f2fs_is_valid_blkaddr(sbi, blkaddr, DATA_GENERIC_ENHANCE_READ)) >>>>>> + return; >>>>>> + >>>>>> + si_meminfo(&si); >>>>>> + free_ram = si.freeram; >>>>>> + avail_ram = si.totalram - si.totalhigh; >>>>>> + >>>>>> + /* free memory is lower than watermark, deny caching compress page */ >>>>>> + if (free_ram <= sbi->compress_watermark / 100 * avail_ram) >>>>>> + return; >>>>>> + >>>>>> + /* cached page count exceed threshold, deny caching compress page */ >>>>>> + if (COMPRESS_MAPPING(sbi)->nrpages >= >>>>>> + free_ram / 100 * sbi->compress_percent) >>>>>> + return; >>>>>> + >>>>>> + cpage = find_get_page(COMPRESS_MAPPING(sbi), blkaddr); >>>>>> + if (cpage) { >>>>>> + f2fs_put_page(cpage, 0); >>>>>> + return; >>>>>> + } >>>>>> + >>>>>> + cpage = alloc_page(__GFP_IO); >>>>>> + if (!cpage) >>>>>> + return; >>>>>> + >>>>>> + ret = add_to_page_cache_lru(cpage, COMPRESS_MAPPING(sbi), >>>>>> + blkaddr, GFP_NOFS); >>>>>> + if (ret) { >>>>>> + f2fs_put_page(cpage, 0); >>>>>> + return; >>>>>> + } >>>>>> + >>>>>> + set_page_private_data(cpage, ino); >>>>>> + >>>>>> + if (!f2fs_is_valid_blkaddr(sbi, blkaddr, DATA_GENERIC_ENHANCE_READ)) >>>>>> + goto out; >>>>>> + >>>>>> + memcpy(page_address(cpage), page_address(page), PAGE_SIZE); >>>>>> + SetPageUptodate(cpage); >>>>>> +out: >>>>>> + f2fs_put_page(cpage, 1); >>>>>> +} >>>>>> + >>>>>> +bool f2fs_load_compressed_page(struct f2fs_sb_info *sbi, struct page *page, >>>>>> + block_t blkaddr) >>>>>> +{ >>>>>> + struct page *cpage; >>>>>> + bool hitted = false; >>>>>> + >>>>>> + if (!test_opt(sbi, COMPRESS_CACHE)) >>>>>> + return false; >>>>>> + >>>>>> + cpage = f2fs_pagecache_get_page(COMPRESS_MAPPING(sbi), >>>>>> + blkaddr, FGP_LOCK | FGP_NOWAIT, GFP_NOFS); >>>>>> + if (cpage) { >>>>>> + if (PageUptodate(cpage)) { >>>>>> + atomic_inc(&sbi->compress_page_hit); >>>>>> + memcpy(page_address(page), >>>>>> + page_address(cpage), PAGE_SIZE); >>>>>> + hitted = true; >>>>>> + } >>>>>> + f2fs_put_page(cpage, 1); >>>>>> + } >>>>>> + >>>>>> + return hitted; >>>>>> +} >>>>>> + >>>>>> +void f2fs_invalidate_compress_pages(struct f2fs_sb_info *sbi, nid_t ino) >>>>>> +{ >>>>>> + struct address_space *mapping = sbi->compress_inode->i_mapping; >>>>>> + struct pagevec pvec; >>>>>> + pgoff_t index = 0; >>>>>> + pgoff_t end = MAX_BLKADDR(sbi); >>>>>> + >>>>>> + if (!mapping->nrpages) >>>>>> + return; >>>>>> + >>>>>> + pagevec_init(&pvec); >>>>>> + >>>>>> + do { >>>>>> + unsigned int nr_pages; >>>>>> + int i; >>>>>> + >>>>>> + nr_pages = pagevec_lookup_range(&pvec, mapping, >>>>>> + &index, end - 1); >>>>>> + if (!nr_pages) >>>>>> + break; >>>>>> + >>>>>> + for (i = 0; i < nr_pages; i++) { >>>>>> + struct page *page = pvec.pages[i]; >>>>>> + >>>>>> + if (page->index > end) >>>>>> + break; >>>>>> + >>>>>> + lock_page(page); >>>>>> + if (page->mapping != mapping) { >>>>>> + unlock_page(page); >>>>>> + continue; >>>>>> + } >>>>>> + >>>>>> + if (ino != get_page_private_data(page)) { >>>>>> + unlock_page(page); >>>>>> + continue; >>>>>> + } >>>>>> + >>>>>> + generic_error_remove_page(mapping, page); >>>>>> + unlock_page(page); >>>>>> + } >>>>>> + pagevec_release(&pvec); >>>>>> + cond_resched(); >>>>>> + } while (index < end); >>>>>> +} >>>>>> + >>>>>> +int f2fs_init_compress_inode(struct f2fs_sb_info *sbi) >>>>>> +{ >>>>>> + struct inode *inode; >>>>>> + >>>>>> + if (!test_opt(sbi, COMPRESS_CACHE)) >>>>>> + return 0; >>>>>> + >>>>>> + inode = f2fs_iget(sbi->sb, F2FS_COMPRESS_INO(sbi)); >>>>>> + if (IS_ERR(inode)) >>>>>> + return PTR_ERR(inode); >>>>>> + sbi->compress_inode = inode; >>>>>> + >>>>>> + sbi->compress_percent = COMPRESS_PERCENT; >>>>>> + sbi->compress_watermark = COMPRESS_WATERMARK; >>>>>> + >>>>>> + atomic_set(&sbi->compress_page_hit, 0); >>>>>> + >>>>>> + return 0; >>>>>> +} >>>>>> + >>>>>> +void f2fs_destroy_compress_inode(struct f2fs_sb_info *sbi) >>>>>> +{ >>>>>> + if (!sbi->compress_inode) >>>>>> + return; >>>>>> + iput(sbi->compress_inode); >>>>>> + sbi->compress_inode = NULL; >>>>>> +} >>>>>> + >>>>>> int f2fs_init_page_array_cache(struct f2fs_sb_info *sbi) >>>>>> { >>>>>> dev_t dev = sbi->sb->s_bdev->bd_dev; >>>>>> diff --git a/fs/f2fs/data.c b/fs/f2fs/data.c >>>>>> index d4795eda12fa..3058c7e28b11 100644 >>>>>> --- a/fs/f2fs/data.c >>>>>> +++ b/fs/f2fs/data.c >>>>>> @@ -132,7 +132,7 @@ static void f2fs_finish_read_bio(struct bio *bio) >>>>>> if (f2fs_is_compressed_page(page)) { >>>>>> if (bio->bi_status) >>>>>> - f2fs_end_read_compressed_page(page, true); >>>>>> + f2fs_end_read_compressed_page(page, true, 0); >>>>>> f2fs_put_page_dic(page); >>>>>> continue; >>>>>> } >>>>>> @@ -228,15 +228,19 @@ static void f2fs_handle_step_decompress(struct bio_post_read_ctx *ctx) >>>>>> struct bio_vec *bv; >>>>>> struct bvec_iter_all iter_all; >>>>>> bool all_compressed = true; >>>>>> + block_t blkaddr = SECTOR_TO_BLOCK(ctx->bio->bi_iter.bi_sector); >>>>>> bio_for_each_segment_all(bv, ctx->bio, iter_all) { >>>>>> struct page *page = bv->bv_page; >>>>>> /* PG_error was set if decryption failed. */ >>>>>> if (f2fs_is_compressed_page(page)) >>>>>> - f2fs_end_read_compressed_page(page, PageError(page)); >>>>>> + f2fs_end_read_compressed_page(page, PageError(page), >>>>>> + blkaddr); >>>>>> else >>>>>> all_compressed = false; >>>>>> + >>>>>> + blkaddr++; >>>>>> } >>>>>> /* >>>>>> @@ -1352,9 +1356,11 @@ static int __allocate_data_block(struct dnode_of_data *dn, int seg_type) >>>>>> old_blkaddr = dn->data_blkaddr; >>>>>> f2fs_allocate_data_block(sbi, NULL, old_blkaddr, &dn->data_blkaddr, >>>>>> &sum, seg_type, NULL); >>>>>> - if (GET_SEGNO(sbi, old_blkaddr) != NULL_SEGNO) >>>>>> + if (GET_SEGNO(sbi, old_blkaddr) != NULL_SEGNO) { >>>>>> invalidate_mapping_pages(META_MAPPING(sbi), >>>>>> old_blkaddr, old_blkaddr); >>>>>> + f2fs_invalidate_compress_page(sbi, old_blkaddr); >>>>>> + } >>>>>> f2fs_update_data_blkaddr(dn, dn->data_blkaddr); >>>>>> /* >>>>>> @@ -2174,7 +2180,7 @@ int f2fs_read_multi_pages(struct compress_ctx *cc, struct bio **bio_ret, >>>>>> goto out_put_dnode; >>>>>> } >>>>>> - for (i = 0; i < dic->nr_cpages; i++) { >>>>>> + for (i = 0; i < cc->nr_cpages; i++) { >>>>>> struct page *page = dic->cpages[i]; >>>>>> block_t blkaddr; >>>>>> struct bio_post_read_ctx *ctx; >>>>>> @@ -2182,6 +2188,14 @@ int f2fs_read_multi_pages(struct compress_ctx *cc, struct bio **bio_ret, >>>>>> blkaddr = data_blkaddr(dn.inode, dn.node_page, >>>>>> dn.ofs_in_node + i + 1); >>>>>> + f2fs_wait_on_block_writeback(inode, blkaddr); >>>>>> + >>>>>> + if (f2fs_load_compressed_page(sbi, page, blkaddr)) { >>>>>> + if (atomic_dec_and_test(&dic->remaining_pages)) >>>>>> + f2fs_decompress_cluster(dic); >>>>>> + continue; >>>>>> + } >>>>>> + >>>>>> if (bio && (!page_is_mergeable(sbi, bio, >>>>>> *last_block_in_bio, blkaddr) || >>>>>> !f2fs_crypt_mergeable_bio(bio, inode, page->index, NULL))) { >>>>>> @@ -2203,8 +2217,6 @@ int f2fs_read_multi_pages(struct compress_ctx *cc, struct bio **bio_ret, >>>>>> } >>>>>> } >>>>>> - f2fs_wait_on_block_writeback(inode, blkaddr); >>>>>> - >>>>>> if (bio_add_page(bio, page, blocksize, 0) < blocksize) >>>>>> goto submit_and_realloc; >>>>>> @@ -3618,6 +3630,13 @@ void f2fs_invalidate_page(struct page *page, unsigned int offset, >>>>>> clear_page_private_gcing(page); >>>>>> + if (test_opt(sbi, COMPRESS_CACHE)) { >>>>>> + if (f2fs_compressed_file(inode)) >>>>>> + f2fs_invalidate_compress_pages(sbi, inode->i_ino); >>>>>> + if (inode->i_ino == F2FS_COMPRESS_INO(sbi)) >>>>>> + clear_page_private_data(page); >>>>>> + } >>>>>> + >>>>>> if (page_private_atomic(page)) >>>>>> return f2fs_drop_inmem_page(inode, page); >>>>>> @@ -3635,6 +3654,16 @@ int f2fs_release_page(struct page *page, gfp_t wait) >>>>>> if (page_private_atomic(page)) >>>>>> return 0; >>>>>> + if (test_opt(F2FS_P_SB(page), COMPRESS_CACHE)) { >>>>>> + struct f2fs_sb_info *sbi = F2FS_P_SB(page); >>>>>> + struct inode *inode = page->mapping->host; >>>>>> + >>>>>> + if (f2fs_compressed_file(inode)) >>>>>> + f2fs_invalidate_compress_pages(sbi, inode->i_ino); >>>>>> + if (inode->i_ino == F2FS_COMPRESS_INO(sbi)) >>>>>> + clear_page_private_data(page); >>>>>> + } >>>>>> + >>>>>> clear_page_private_gcing(page); >>>>>> detach_page_private(page); >>>>>> diff --git a/fs/f2fs/debug.c b/fs/f2fs/debug.c >>>>>> index c03949a7ccff..833325038ef3 100644 >>>>>> --- a/fs/f2fs/debug.c >>>>>> +++ b/fs/f2fs/debug.c >>>>>> @@ -152,6 +152,12 @@ static void update_general_status(struct f2fs_sb_info *sbi) >>>>>> si->node_pages = NODE_MAPPING(sbi)->nrpages; >>>>>> if (sbi->meta_inode) >>>>>> si->meta_pages = META_MAPPING(sbi)->nrpages; >>>>>> +#ifdef CONFIG_F2FS_FS_COMPRESSION >>>>>> + if (sbi->compress_inode) { >>>>>> + si->compress_pages = COMPRESS_MAPPING(sbi)->nrpages; >>>>>> + si->compress_page_hit = atomic_read(&sbi->compress_page_hit); >>>>>> + } >>>>>> +#endif >>>>>> si->nats = NM_I(sbi)->nat_cnt[TOTAL_NAT]; >>>>>> si->dirty_nats = NM_I(sbi)->nat_cnt[DIRTY_NAT]; >>>>>> si->sits = MAIN_SEGS(sbi); >>>>>> @@ -309,6 +315,12 @@ static void update_mem_info(struct f2fs_sb_info *sbi) >>>>>> si->page_mem += (unsigned long long)npages << PAGE_SHIFT; >>>>>> } >>>>>> +#ifdef CONFIG_F2FS_FS_COMPRESSION >>>>>> + if (sbi->compress_inode) { >>>>>> + unsigned npages = COMPRESS_MAPPING(sbi)->nrpages; >>>>>> + si->page_mem += (unsigned long long)npages << PAGE_SHIFT; >>>>>> + } >>>>>> +#endif >>>>>> } >>>>>> static int stat_show(struct seq_file *s, void *v) >>>>>> @@ -476,6 +488,7 @@ static int stat_show(struct seq_file *s, void *v) >>>>>> "volatile IO: %4d (Max. %4d)\n", >>>>>> si->inmem_pages, si->aw_cnt, si->max_aw_cnt, >>>>>> si->vw_cnt, si->max_vw_cnt); >>>>>> + seq_printf(s, " - compress: %4d, hit:%8d\n", si->compress_pages, si->compress_page_hit); >>>>>> seq_printf(s, " - nodes: %4d in %4d\n", >>>>>> si->ndirty_node, si->node_pages); >>>>>> seq_printf(s, " - dents: %4d in dirs:%4d (%4d)\n", >>>>>> diff --git a/fs/f2fs/f2fs.h b/fs/f2fs/f2fs.h >>>>>> index c0bead0df66a..70c0bd563732 100644 >>>>>> --- a/fs/f2fs/f2fs.h >>>>>> +++ b/fs/f2fs/f2fs.h >>>>>> @@ -98,6 +98,7 @@ extern const char *f2fs_fault_name[FAULT_MAX]; >>>>>> #define F2FS_MOUNT_ATGC 0x08000000 >>>>>> #define F2FS_MOUNT_MERGE_CHECKPOINT 0x10000000 >>>>>> #define F2FS_MOUNT_GC_MERGE 0x20000000 >>>>>> +#define F2FS_MOUNT_COMPRESS_CACHE 0x40000000 >>>>>> #define F2FS_OPTION(sbi) ((sbi)->mount_opt) >>>>>> #define clear_opt(sbi, option) (F2FS_OPTION(sbi).opt &= ~F2FS_MOUNT_##option) >>>>>> @@ -1371,6 +1372,37 @@ PAGE_PRIVATE_CLEAR_FUNC(gcing, ONGOING_MIGRATION); >>>>>> PAGE_PRIVATE_CLEAR_FUNC(atomic, ATOMIC_WRITE); >>>>>> PAGE_PRIVATE_CLEAR_FUNC(dummy, DUMMY_WRITE); >>>>>> +static inline unsigned long get_page_private_data(struct page *page) >>>>>> +{ >>>>>> + unsigned long data = page_private(page); >>>>>> + >>>>>> + if (!test_bit(PAGE_PRIVATE_NOT_POINTER, &data)) >>>>>> + return 0; >>>>>> + return data >> PAGE_PRIVATE_MAX; >>>>>> +} >>>>>> + >>>>>> +static inline void set_page_private_data(struct page *page, unsigned long data) >>>>>> +{ >>>>>> + if (!PagePrivate(page)) { >>>>>> + get_page(page); >>>>>> + SetPagePrivate(page); >>>>>> + } >>>>>> + set_bit(PAGE_PRIVATE_NOT_POINTER, &page_private(page)); >>>>>> + page_private(page) |= data << PAGE_PRIVATE_MAX; >>>>>> +} >>>>>> + >>>>>> +static inline void clear_page_private_data(struct page *page) >>>>>> +{ >>>>>> + page_private(page) &= (1 << PAGE_PRIVATE_MAX) - 1; >>>>>> + if (page_private(page) == 1 << PAGE_PRIVATE_NOT_POINTER) { >>>>>> + set_page_private(page, 0); >>>>>> + if (PagePrivate(page)) { >>>>>> + ClearPagePrivate(page); >>>>>> + put_page(page); >>>>>> + } >>>>>> + } >>>>>> +} >>>>>> + >>>>>> /* For compression */ >>>>>> enum compress_algorithm_type { >>>>>> COMPRESS_LZO, >>>>>> @@ -1385,6 +1417,9 @@ enum compress_flag { >>>>>> COMPRESS_MAX_FLAG, >>>>>> }; >>>>>> +#define COMPRESS_WATERMARK 20 >>>>>> +#define COMPRESS_PERCENT 20 >>>>>> + >>>>>> #define COMPRESS_DATA_RESERVED_SIZE 4 >>>>>> struct compress_data { >>>>>> __le32 clen; /* compressed data size */ >>>>>> @@ -1694,6 +1729,12 @@ struct f2fs_sb_info { >>>>>> u64 compr_written_block; >>>>>> u64 compr_saved_block; >>>>>> u32 compr_new_inode; >>>>>> + >>>>>> + /* For compressed block cache */ >>>>>> + struct inode *compress_inode; /* cache compressed blocks */ >>>>>> + unsigned int compress_percent; /* cache page percentage */ >>>>>> + unsigned int compress_watermark; /* cache page watermark */ >>>>>> + atomic_t compress_page_hit; /* cache hit count */ >>>>>> #endif >>>>>> }; >>>>>> @@ -3660,7 +3701,8 @@ struct f2fs_stat_info { >>>>>> unsigned int bimodal, avg_vblocks; >>>>>> int util_free, util_valid, util_invalid; >>>>>> int rsvd_segs, overp_segs; >>>>>> - int dirty_count, node_pages, meta_pages; >>>>>> + int dirty_count, node_pages, meta_pages, compress_pages; >>>>>> + int compress_page_hit; >>>>>> int prefree_count, call_count, cp_count, bg_cp_count; >>>>>> int tot_segs, node_segs, data_segs, free_segs, free_secs; >>>>>> int bg_node_segs, bg_data_segs; >>>>>> @@ -3996,7 +4038,9 @@ void f2fs_compress_write_end_io(struct bio *bio, struct page *page); >>>>>> bool f2fs_is_compress_backend_ready(struct inode *inode); >>>>>> int f2fs_init_compress_mempool(void); >>>>>> void f2fs_destroy_compress_mempool(void); >>>>>> -void f2fs_end_read_compressed_page(struct page *page, bool failed); >>>>>> +void f2fs_decompress_cluster(struct decompress_io_ctx *dic); >>>>>> +void f2fs_end_read_compressed_page(struct page *page, bool failed, >>>>>> + block_t blkaddr); >>>>>> bool f2fs_cluster_is_empty(struct compress_ctx *cc); >>>>>> bool f2fs_cluster_can_merge_page(struct compress_ctx *cc, pgoff_t index); >>>>>> void f2fs_compress_ctx_add_page(struct compress_ctx *cc, struct page *page); >>>>>> @@ -4014,10 +4058,19 @@ void f2fs_put_page_dic(struct page *page); >>>>>> int f2fs_init_compress_ctx(struct compress_ctx *cc); >>>>>> void f2fs_destroy_compress_ctx(struct compress_ctx *cc, bool reuse); >>>>>> void f2fs_init_compress_info(struct f2fs_sb_info *sbi); >>>>>> +int f2fs_init_compress_inode(struct f2fs_sb_info *sbi); >>>>>> +void f2fs_destroy_compress_inode(struct f2fs_sb_info *sbi); >>>>>> int f2fs_init_page_array_cache(struct f2fs_sb_info *sbi); >>>>>> void f2fs_destroy_page_array_cache(struct f2fs_sb_info *sbi); >>>>>> int __init f2fs_init_compress_cache(void); >>>>>> void f2fs_destroy_compress_cache(void); >>>>>> +struct address_space *COMPRESS_MAPPING(struct f2fs_sb_info *sbi); >>>>>> +void f2fs_invalidate_compress_page(struct f2fs_sb_info *sbi, block_t blkaddr); >>>>>> +void f2fs_cache_compressed_page(struct f2fs_sb_info *sbi, struct page *page, >>>>>> + nid_t ino, block_t blkaddr); >>>>>> +bool f2fs_load_compressed_page(struct f2fs_sb_info *sbi, struct page *page, >>>>>> + block_t blkaddr); >>>>>> +void f2fs_invalidate_compress_pages(struct f2fs_sb_info *sbi, nid_t ino); >>>>>> #define inc_compr_inode_stat(inode) \ >>>>>> do { \ >>>>>> struct f2fs_sb_info *sbi = F2FS_I_SB(inode); \ >>>>>> @@ -4046,7 +4099,9 @@ static inline struct page *f2fs_compress_control_page(struct page *page) >>>>>> } >>>>>> static inline int f2fs_init_compress_mempool(void) { return 0; } >>>>>> static inline void f2fs_destroy_compress_mempool(void) { } >>>>>> -static inline void f2fs_end_read_compressed_page(struct page *page, bool failed) >>>>>> +static inline void f2fs_decompress_cluster(struct decompress_io_ctx *dic) { } >>>>>> +static inline void f2fs_end_read_compressed_page(struct page *page, >>>>>> + bool failed, block_t blkaddr) >>>>>> { >>>>>> WARN_ON_ONCE(1); >>>>>> } >>>>>> @@ -4054,10 +4109,20 @@ static inline void f2fs_put_page_dic(struct page *page) >>>>>> { >>>>>> WARN_ON_ONCE(1); >>>>>> } >>>>>> +static inline int f2fs_init_compress_inode(struct f2fs_sb_info *sbi) { return 0; } >>>>>> +static inline void f2fs_destroy_compress_inode(struct f2fs_sb_info *sbi) { } >>>>>> static inline int f2fs_init_page_array_cache(struct f2fs_sb_info *sbi) { return 0; } >>>>>> static inline void f2fs_destroy_page_array_cache(struct f2fs_sb_info *sbi) { } >>>>>> static inline int __init f2fs_init_compress_cache(void) { return 0; } >>>>>> static inline void f2fs_destroy_compress_cache(void) { } >>>>>> +static inline void f2fs_invalidate_compress_page(struct f2fs_sb_info *sbi, >>>>>> + block_t blkaddr) { } >>>>>> +static inline void f2fs_cache_compressed_page(struct f2fs_sb_info *sbi, >>>>>> + struct page *page, nid_t ino, block_t blkaddr) { } >>>>>> +static inline bool f2fs_load_compressed_page(struct f2fs_sb_info *sbi, >>>>>> + struct page *page, block_t blkaddr) { return false; } >>>>>> +static inline void f2fs_invalidate_compress_pages(struct f2fs_sb_info *sbi, >>>>>> + nid_t ino) { } >>>>>> #define inc_compr_inode_stat(inode) do { } while (0) >>>>>> #endif >>>>>> diff --git a/fs/f2fs/gc.c b/fs/f2fs/gc.c >>>>>> index bcb3b488dbca..f3d2bed746b0 100644 >>>>>> --- a/fs/f2fs/gc.c >>>>>> +++ b/fs/f2fs/gc.c >>>>>> @@ -1261,6 +1261,7 @@ static int move_data_block(struct inode *inode, block_t bidx, >>>>>> f2fs_put_page(mpage, 1); >>>>>> invalidate_mapping_pages(META_MAPPING(fio.sbi), >>>>>> fio.old_blkaddr, fio.old_blkaddr); >>>>>> + f2fs_invalidate_compress_page(fio.sbi, fio.old_blkaddr); >>>>>> set_page_dirty(fio.encrypted_page); >>>>>> if (clear_page_dirty_for_io(fio.encrypted_page)) >>>>>> diff --git a/fs/f2fs/inode.c b/fs/f2fs/inode.c >>>>>> index cbda7ca3b3be..9141147b5bb0 100644 >>>>>> --- a/fs/f2fs/inode.c >>>>>> +++ b/fs/f2fs/inode.c >>>>>> @@ -18,6 +18,10 @@ >>>>>> #include >>>>>> +#ifdef CONFIG_F2FS_FS_COMPRESSION >>>>>> +extern const struct address_space_operations f2fs_compress_aops; >>>>>> +#endif >>>>>> + >>>>>> void f2fs_mark_inode_dirty_sync(struct inode *inode, bool sync) >>>>>> { >>>>>> if (is_inode_flag_set(inode, FI_NEW_INODE)) >>>>>> @@ -494,6 +498,11 @@ struct inode *f2fs_iget(struct super_block *sb, unsigned long ino) >>>>>> if (ino == F2FS_NODE_INO(sbi) || ino == F2FS_META_INO(sbi)) >>>>>> goto make_now; >>>>>> +#ifdef CONFIG_F2FS_FS_COMPRESSION >>>>>> + if (ino == F2FS_COMPRESS_INO(sbi)) >>>>>> + goto make_now; >>>>>> +#endif >>>>>> + >>>>>> ret = do_read_inode(inode); >>>>>> if (ret) >>>>>> goto bad_inode; >>>>>> @@ -504,6 +513,12 @@ struct inode *f2fs_iget(struct super_block *sb, unsigned long ino) >>>>>> } else if (ino == F2FS_META_INO(sbi)) { >>>>>> inode->i_mapping->a_ops = &f2fs_meta_aops; >>>>>> mapping_set_gfp_mask(inode->i_mapping, GFP_NOFS); >>>>>> + } else if (ino == F2FS_COMPRESS_INO(sbi)) { >>>>>> +#ifdef CONFIG_F2FS_FS_COMPRESSION >>>>>> + inode->i_mapping->a_ops = &f2fs_compress_aops; >>>>>> +#endif >>>>>> + mapping_set_gfp_mask(inode->i_mapping, >>>>>> + GFP_NOFS | __GFP_HIGHMEM | __GFP_MOVABLE); >>>>>> } else if (S_ISREG(inode->i_mode)) { >>>>>> inode->i_op = &f2fs_file_inode_operations; >>>>>> inode->i_fop = &f2fs_file_operations; >>>>>> @@ -723,8 +738,12 @@ void f2fs_evict_inode(struct inode *inode) >>>>>> trace_f2fs_evict_inode(inode); >>>>>> truncate_inode_pages_final(&inode->i_data); >>>>>> + if (test_opt(sbi, COMPRESS_CACHE) && f2fs_compressed_file(inode)) >>>>>> + f2fs_invalidate_compress_pages(sbi, inode->i_ino); >>>>>> + >>>>>> if (inode->i_ino == F2FS_NODE_INO(sbi) || >>>>>> - inode->i_ino == F2FS_META_INO(sbi)) >>>>>> + inode->i_ino == F2FS_META_INO(sbi) || >>>>>> + inode->i_ino == F2FS_COMPRESS_INO(sbi)) >>>>>> goto out_clear; >>>>>> f2fs_bug_on(sbi, get_dirty_pages(inode)); >>>>>> diff --git a/fs/f2fs/segment.c b/fs/f2fs/segment.c >>>>>> index 8668df7870d0..406a6b244782 100644 >>>>>> --- a/fs/f2fs/segment.c >>>>>> +++ b/fs/f2fs/segment.c >>>>>> @@ -2322,6 +2322,7 @@ void f2fs_invalidate_blocks(struct f2fs_sb_info *sbi, block_t addr) >>>>>> return; >>>>>> invalidate_mapping_pages(META_MAPPING(sbi), addr, addr); >>>>>> + f2fs_invalidate_compress_page(sbi, addr); >>>>>> /* add it into sit main buffer */ >>>>>> down_write(&sit_i->sentry_lock); >>>>>> @@ -3469,9 +3470,11 @@ static void do_write_page(struct f2fs_summary *sum, struct f2fs_io_info *fio) >>>>>> reallocate: >>>>>> f2fs_allocate_data_block(fio->sbi, fio->page, fio->old_blkaddr, >>>>>> &fio->new_blkaddr, sum, type, fio); >>>>>> - if (GET_SEGNO(fio->sbi, fio->old_blkaddr) != NULL_SEGNO) >>>>>> + if (GET_SEGNO(fio->sbi, fio->old_blkaddr) != NULL_SEGNO) { >>>>>> invalidate_mapping_pages(META_MAPPING(fio->sbi), >>>>>> fio->old_blkaddr, fio->old_blkaddr); >>>>>> + f2fs_invalidate_compress_page(fio->sbi, fio->old_blkaddr); >>>>>> + } >>>>>> /* writeout dirty page into bdev */ >>>>>> f2fs_submit_page_write(fio); >>>>>> @@ -3661,6 +3664,7 @@ void f2fs_do_replace_block(struct f2fs_sb_info *sbi, struct f2fs_summary *sum, >>>>>> if (GET_SEGNO(sbi, old_blkaddr) != NULL_SEGNO) { >>>>>> invalidate_mapping_pages(META_MAPPING(sbi), >>>>>> old_blkaddr, old_blkaddr); >>>>>> + f2fs_invalidate_compress_page(sbi, old_blkaddr); >>>>>> if (!from_gc) >>>>>> update_segment_mtime(sbi, old_blkaddr, 0); >>>>>> update_sit_entry(sbi, old_blkaddr, -1); >>>>>> diff --git a/fs/f2fs/super.c b/fs/f2fs/super.c >>>>>> index 096492caaa6b..5056b8cfe919 100644 >>>>>> --- a/fs/f2fs/super.c >>>>>> +++ b/fs/f2fs/super.c >>>>>> @@ -150,6 +150,7 @@ enum { >>>>>> Opt_compress_extension, >>>>>> Opt_compress_chksum, >>>>>> Opt_compress_mode, >>>>>> + Opt_compress_cache, >>>>>> Opt_atgc, >>>>>> Opt_gc_merge, >>>>>> Opt_nogc_merge, >>>>>> @@ -224,6 +225,7 @@ static match_table_t f2fs_tokens = { >>>>>> {Opt_compress_extension, "compress_extension=%s"}, >>>>>> {Opt_compress_chksum, "compress_chksum"}, >>>>>> {Opt_compress_mode, "compress_mode=%s"}, >>>>>> + {Opt_compress_cache, "compress_cache"}, >>>>>> {Opt_atgc, "atgc"}, >>>>>> {Opt_gc_merge, "gc_merge"}, >>>>>> {Opt_nogc_merge, "nogc_merge"}, >>>>>> @@ -1066,12 +1068,16 @@ static int parse_options(struct super_block *sb, char *options, bool is_remount) >>>>>> } >>>>>> kfree(name); >>>>>> break; >>>>>> + case Opt_compress_cache: >>>>>> + set_opt(sbi, COMPRESS_CACHE); >>>>>> + break; >>>>>> #else >>>>>> case Opt_compress_algorithm: >>>>>> case Opt_compress_log_size: >>>>>> case Opt_compress_extension: >>>>>> case Opt_compress_chksum: >>>>>> case Opt_compress_mode: >>>>>> + case Opt_compress_cache: >>>>>> f2fs_info(sbi, "compression options not supported"); >>>>>> break; >>>>>> #endif >>>>>> @@ -1403,6 +1409,8 @@ static void f2fs_put_super(struct super_block *sb) >>>>>> f2fs_bug_on(sbi, sbi->fsync_node_num); >>>>>> + f2fs_destroy_compress_inode(sbi); >>>>>> + >>>>>> iput(sbi->node_inode); >>>>>> sbi->node_inode = NULL; >>>>>> @@ -1672,6 +1680,9 @@ static inline void f2fs_show_compress_options(struct seq_file *seq, >>>>>> seq_printf(seq, ",compress_mode=%s", "fs"); >>>>>> else if (F2FS_OPTION(sbi).compress_mode == COMPR_MODE_USER) >>>>>> seq_printf(seq, ",compress_mode=%s", "user"); >>>>>> + >>>>>> + if (test_opt(sbi, COMPRESS_CACHE)) >>>>>> + seq_puts(seq, ",compress_cache"); >>>>>> } >>>>>> #endif >>>>>> @@ -1949,6 +1960,7 @@ static int f2fs_remount(struct super_block *sb, int *flags, char *data) >>>>>> bool disable_checkpoint = test_opt(sbi, DISABLE_CHECKPOINT); >>>>>> bool no_io_align = !F2FS_IO_ALIGNED(sbi); >>>>>> bool no_atgc = !test_opt(sbi, ATGC); >>>>>> + bool no_compress_cache = !test_opt(sbi, COMPRESS_CACHE); >>>>>> bool checkpoint_changed; >>>>>> #ifdef CONFIG_QUOTA >>>>>> int i, j; >>>>>> @@ -2041,6 +2053,12 @@ static int f2fs_remount(struct super_block *sb, int *flags, char *data) >>>>>> goto restore_opts; >>>>>> } >>>>>> + if (no_compress_cache == !!test_opt(sbi, COMPRESS_CACHE)) { >>>>>> + err = -EINVAL; >>>>>> + f2fs_warn(sbi, "switch compress_cache option is not allowed"); >>>>>> + goto restore_opts; >>>>>> + } >>>>>> + >>>>>> if ((*flags & SB_RDONLY) && test_opt(sbi, DISABLE_CHECKPOINT)) { >>>>>> err = -EINVAL; >>>>>> f2fs_warn(sbi, "disabling checkpoint not compatible with read-only"); >>>>>> @@ -3940,10 +3958,14 @@ static int f2fs_fill_super(struct super_block *sb, void *data, int silent) >>>>>> goto free_node_inode; >>>>>> } >>>>>> - err = f2fs_register_sysfs(sbi); >>>>>> + err = f2fs_init_compress_inode(sbi); >>>>>> if (err) >>>>>> goto free_root_inode; >>>>>> + err = f2fs_register_sysfs(sbi); >>>>>> + if (err) >>>>>> + goto free_compress_inode; >>>>>> + >>>>>> #ifdef CONFIG_QUOTA >>>>>> /* Enable quota usage during mount */ >>>>>> if (f2fs_sb_has_quota_ino(sbi) && !f2fs_readonly(sb)) { >>>>>> @@ -4084,6 +4106,8 @@ static int f2fs_fill_super(struct super_block *sb, void *data, int silent) >>>>>> /* evict some inodes being cached by GC */ >>>>>> evict_inodes(sb); >>>>>> f2fs_unregister_sysfs(sbi); >>>>>> +free_compress_inode: >>>>>> + f2fs_destroy_compress_inode(sbi); >>>>>> free_root_inode: >>>>>> dput(sb->s_root); >>>>>> sb->s_root = NULL; >>>>>> @@ -4162,6 +4186,15 @@ static void kill_f2fs_super(struct super_block *sb) >>>>>> f2fs_stop_gc_thread(sbi); >>>>>> f2fs_stop_discard_thread(sbi); >>>>>> +#ifdef CONFIG_F2FS_FS_COMPRESSION >>>>>> + /* >>>>>> + * latter evict_inode() can bypass checking and invalidating >>>>>> + * compress inode cache. >>>>>> + */ >>>>>> + if (test_opt(sbi, COMPRESS_CACHE)) >>>>>> + truncate_inode_pages_final(COMPRESS_MAPPING(sbi)); >>>>>> +#endif >>>>>> + >>>>>> if (is_sbi_flag_set(sbi, SBI_IS_DIRTY) || >>>>>> !is_set_ckpt_flags(sbi, CP_UMOUNT_FLAG)) { >>>>>> struct cp_control cpc = { >>>>>> diff --git a/include/linux/f2fs_fs.h b/include/linux/f2fs_fs.h >>>>>> index 5487a80617a3..0021ea8f7c3b 100644 >>>>>> --- a/include/linux/f2fs_fs.h >>>>>> +++ b/include/linux/f2fs_fs.h >>>>>> @@ -34,6 +34,7 @@ >>>>>> #define F2FS_ROOT_INO(sbi) ((sbi)->root_ino_num) >>>>>> #define F2FS_NODE_INO(sbi) ((sbi)->node_ino_num) >>>>>> #define F2FS_META_INO(sbi) ((sbi)->meta_ino_num) >>>>>> +#define F2FS_COMPRESS_INO(sbi) (NM_I(sbi)->max_nid) >>>>>> #define F2FS_MAX_QUOTAS 3 >>>>>> >>>> >>>> >>>> _______________________________________________ >>>> Linux-f2fs-devel mailing list >>>> Linux-f2fs-devel@lists.sourceforge.net >>>> https://lists.sourceforge.net/lists/listinfo/linux-f2fs-devel >>> >>> >>> _______________________________________________ >>> Linux-f2fs-devel mailing list >>> Linux-f2fs-devel@lists.sourceforge.net >>> https://lists.sourceforge.net/lists/listinfo/linux-f2fs-devel >>> > . >