Received: by 2002:a6b:fb09:0:0:0:0:0 with SMTP id h9csp271737iog; Mon, 13 Jun 2022 02:17:41 -0700 (PDT) X-Google-Smtp-Source: ABdhPJxMdzzqgG9fLOcOvKiQnPGjAWVueJ+UjDg7pykavSCAbknO8HESevfN8FMSgmu64wXrlLQy X-Received: by 2002:a17:90b:3c87:b0:1ea:95be:6b7d with SMTP id pv7-20020a17090b3c8700b001ea95be6b7dmr11541661pjb.203.1655111860743; Mon, 13 Jun 2022 02:17:40 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1655111860; cv=none; d=google.com; s=arc-20160816; b=0ck1FJA2sOUFl2exoz3VMevfwOCmWC0o6eHfWLnAW7+z+xzID61oB4/5TSp2uDaAiL keckDMT5eibm1N9mHbsY0LzHKWN1hzBmKLL95iJ44hbyH7aYRLfsrbusDrG9X0VCk1D2 4MfxKGu8Ayqgrt0VNXIQ/HYsc8omDa8UPIhyxUFMH39iAcB6srO0hJ17Cy6NoAb6exJo epXdgdz0sxafRKUv0eF2cfiWXI50Dkqc7+LR0hbio0IY3gLXUUsfqfnQyqLcCoqnXdD1 x76A8w8+8aB3d/+ZEJvjkwvvb13wOeORjAhbUn9s3i6LnpZUqEQKiuOSlpMUahqOnExY ABFw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:cc:to:subject:message-id:date:from:in-reply-to :references:mime-version:dkim-signature; bh=OdEpZ4ZctY+bmfnlyy9Zc7Me3QTB1QCGRf37pj7dMLs=; b=xf/SCrwv6uiyUHGJ81Y9TOxwkEZtUCFvwtKG32OHpYor51kjvxUzhYHNj+4TqTUxTx evdKNzfKZWD7LV+Ea+LTMfdEe6fym6tYJv0NVvjWYOhekVSldKCXBJlt5I3FNXI94g8l Y8kwZbP73PVh+8UlXP5JfWMEwexZFZY3FHLgDWnHylbKCO9OUSV8FAep1Ip6u58GLZXf J6hNdrD9NYueCYX9GMEay4mf9YzwQQNXLSBg0lt6HMM/sraGPGHwyo1lR5AdK0HU3Xy5 Vt6hcoEwqO7cyMpYphD64ncm78LKwfNqURg0Yvui7dyLw+IwvGFPZ4dsJAYftH94LYXP jPHQ== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@chromium.org header.s=google header.b=kyZPFd4X; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=chromium.org Return-Path: Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id t11-20020a63eb0b000000b003f8a48ed403si9981249pgh.580.2022.06.13.02.17.26; Mon, 13 Jun 2022 02:17:40 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; dkim=pass header.i=@chromium.org header.s=google header.b=kyZPFd4X; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=chromium.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S237220AbiFMIgH (ORCPT + 99 others); Mon, 13 Jun 2022 04:36:07 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:44936 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S240116AbiFMIgF (ORCPT ); Mon, 13 Jun 2022 04:36:05 -0400 Received: from mail-ed1-x532.google.com (mail-ed1-x532.google.com [IPv6:2a00:1450:4864:20::532]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 4517D18E0E for ; Mon, 13 Jun 2022 01:36:03 -0700 (PDT) Received: by mail-ed1-x532.google.com with SMTP id o10so6119797edi.1 for ; Mon, 13 Jun 2022 01:36:03 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=chromium.org; s=google; h=mime-version:references:in-reply-to:from:date:message-id:subject:to :cc; bh=OdEpZ4ZctY+bmfnlyy9Zc7Me3QTB1QCGRf37pj7dMLs=; b=kyZPFd4XHm4C1krejJG5ESgszoHa1O5CSIWsqoiR4A4eu2kzkAKPd2K7h4CLEzh1q1 XTERnLTo6nxneJzcxZUDnMa70pBijrUSmBwIH4aO6kX7vdzLRgPdAM23pXo8LH5ZJQ1h bsNDV4aFOLrjEo1PcWpXJ7IcIvswtiriO3fPU= X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20210112; h=x-gm-message-state:mime-version:references:in-reply-to:from:date :message-id:subject:to:cc; bh=OdEpZ4ZctY+bmfnlyy9Zc7Me3QTB1QCGRf37pj7dMLs=; b=Z/OwJUSSoYi0r2RDUURWI1kBw8fdsbhUtI2B0VQP2hWiF5FrBNc00Vs5UjA9Ba/DWx j7jQNYSqxqZevqu4TItG3jV7PVljdA+D0H/A0IFUKpddfUkd6WditjWDIDm5qTUCOr8B v4w1Gdhu4uhGV3ujiyq/o9V4XfK33iPPeLbqqow5wlUf5/69kaFCPTl1m9SkIA2RK3X+ qe7FutHLCRno+5TjuTkTgZs+lGtx4oLKJd1r9AuoHAHm17ZLs560GLlSt/evoI4Xw/wX JWmX6F4LYR5nfUGh/1E1D4FP3IXeuuNO04B0MclWSQYWwjY54e7WmjU/pWLGep9nagkB NXLw== X-Gm-Message-State: AOAM532vCxza13/I8WsHNp9ROZnSJ1FCUbeYkO8SnFmKGgwUDUFzysgB aknWs73UAasoRZFk4zb2eYGOGk0IObS3gD4BGHEIOg== X-Received: by 2002:a05:6402:23a3:b0:42e:251a:c963 with SMTP id j35-20020a05640223a300b0042e251ac963mr59167039eda.173.1655109361642; Mon, 13 Jun 2022 01:36:01 -0700 (PDT) MIME-Version: 1.0 References: <20220606150305.1883410-1-hsinyi@chromium.org> <20220606150305.1883410-4-hsinyi@chromium.org> In-Reply-To: From: Hsin-Yi Wang Date: Mon, 13 Jun 2022 16:35:35 +0800 Message-ID: Subject: Re: [PATCH v5 3/3] squashfs: implement readahead To: Xiongwei Song Cc: Phillip Lougher , Matthew Wilcox , Xiongwei Song , Marek Szyprowski , Andrew Morton , Zheng Liang , Zhang Yi , Hou Tao , Miao Xie , "linux-mm @ kvack . org" , "squashfs-devel @ lists . sourceforge . net" , Linux Kernel Mailing List Content-Type: text/plain; charset="UTF-8" X-Spam-Status: No, score=-3.3 required=5.0 tests=BAYES_00,DKIMWL_WL_HIGH, DKIM_SIGNED,DKIM_VALID,DKIM_VALID_AU,DKIM_VALID_EF,RCVD_IN_DNSWL_NONE, SPF_HELO_NONE,SPF_PASS,T_SCC_BODY_TEXT_LINE autolearn=ham autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Mon, Jun 13, 2022 at 9:36 AM Xiongwei Song wrote: > > Hi, > > On Fri, Jun 10, 2022 at 3:42 PM Phillip Lougher wrote: > > > > On 09/06/2022 15:46, Xiongwei Song wrote: > > > This version is bad for my test. I ran the test below > > > "for cnt in $(seq 0 9); do echo 3 > /proc/sys/vm/drop_caches; echo > > > "Loop ${cnt}:"; time -v find /software/test[0-9][0-9] | xargs -P 24 -i > > > cat {} > /dev/null 2>/dev/null; echo ""; done" > > > in 90 partitions. > > > > > > With 9eec1d897139 reverted: > > > 1:06.18 (1m + 6.18s) > > > 1:05.65 > > > 1:06.34 > > > 1:06.88 > > > 1:06.52 > > > 1:06.78 > > > 1:06.61 > > > 1:06.99 > > > 1:06.60 > > > 1:06.79 > > > > > > With this version: > > > 2:36.85 (2m + 36.85s) > > > 2:28.89 > > > 1:43.46 > > > 1:41.50 > > > 1:42.75 > > > 1:43.46 > > > 1:43.67 > > > 1:44.41 > > > 1:44.91 > > > 1:45.44 > > > > > > Any thoughts? > > > > Thank-you for your latest test results, and they tend to > > imply that the latest version of the patch hasn't improved > > performance in your use-case. > > > > One thing which is becoming clear here is that the devil is in > > the detail, and your results being summaries are not capturing > > enough detail to understand what is happening. They show > > something is wrong, but, don't give any guidance as to what > > is happening. > > > > I think it will be difficult to capture more details from > > your test case. But, detail can be captured from summaries, by > > varying the input and extrapolating from the results. > > > > By that I mean have you tried changing anything, and observed any > > changed results? > > > > For instance have you tried any of the following > > > > 1. Changing the parallelism of your test from 24 read threads. > > Does 1, 2, 4 etc parallel read threads change the observed > > behaviour? In other words is the slow-down observed across > > all degrees of parallelism, or is there a critical point. > > > > 2. Does the Squashfs parallelism options in the kernel configuration > > change the behaviour? Knowing if the number of "decompressors" > > available changes the difference in performance could be important. > > > > 3. Are your Squashfs filesystems built using fragments, or without > > fragments? Rebuilding the filesystems without fragments, and > > observing any different performance, would help to pinpoint > > where the issue lies. > > > > 4. What is the block size used in your Squashfs filesystems. Have > > you tried changing the block size, and seen what effect > > it has on the difference in performance between the patches? > > > > 5. You don't mention where your Squashfs filesystems are stored. > > Is this slow media or fast media? Have you tried moving > > the Squashfs filesystems onto different media and observed > > any difference in performance between the patches? > > > > Thanks for your response and inputs. I really appreciated your help. > I can try these things but can't provide the detailed results for > now because I'm busy with a few things, hence It's hard to focus > on this one thing for me. > > > The fact of the matter is there are many over-lapping factors > > which affect the performance of Squashfs filesystems (like any > > reasonably complex code), which may be elsewhere. It can only > > take a small change somewhere to have a dramatic affect on > > performance. > > > > This is particularly the case with embedded systems, which > > may be short on CPU performance, short on RAM, and have low > > performance media, and be effectively operating on the "edge". > > It can only take a small change, an update for instance, to > > change from performing well to badly. > > Totally agree. > > > > > I speak from experience, having spent over ten years in embedded > > Linux as a senior engineer and then as a consultant. I have > > my own horror tales as a consultant, dealing with systems pushed > > beyond the edge (with hacks), and the customer insisting they > > didn't do anything to cause the system to finally break. > > > > Maybe it is off topic here. But, I remember one instance where > > a customer had a system out in the field, which "inexplicably" > > started to lock up every 6 months or so. This system had regular > > updates "over the air", and I discovered the "lock up" only > > started happening after the latest update. It turns out the new version > > of the application had grown a new feature which needed more > > RAM than normal. This feature wasn't used very often, but, > > if it coincided with an infrequent "house-keeping" background task, > > the system ran out of memory and locked up (they had disabled the OOM > > killer). This was so rare it might only coincide after six months. No > > bug, but a slow growth in working set RAM over a number of versions. > > > > In other words we may be looking at a knock-on side effect of > > readahead, which is either caused by issues elsewhere or is > > causing issues elsewhere. > > > > Dealing with it in isolation, as bug in the readahead code is going > > to get us nowhere, looking for something that isn't there. > > > > I'm not saying that this is the case here. But, the more detail > > you can provide, and the more test variants you can provide will > > help to determine what is the problem. > > Thanks for your sharing. I will provide detail later. > Hi, Thanks for testing on v5. I've sent v6 which is based on the series: https://patchwork.kernel.org/project/linux-mm/cover/20220611032133.5743-1-phillip@squashfs.org.uk/. v6: https://patchwork.kernel.org/project/linux-mm/cover/20220613082802.1301238-1-hsinyi@chromium.org/ To apply the patch on linux-next, one might have to revert ca1505bf4805 ("squashfs: implement readahead") and 9d58b94aa73a ("squashfs: always build "file direct" version of page actor") first. > Regards, > Xiongwei > > > > > Thanks > > > > Phillip > > > > > > > > > > Regards, > > > Xiongwei > > > > > > On Mon, Jun 6, 2022 at 11:03 PM Hsin-Yi Wang wrote: > > >> > > >> Implement readahead callback for squashfs. It will read datablocks > > >> which cover pages in readahead request. For a few cases it will > > >> not mark page as uptodate, including: > > >> - file end is 0. > > >> - zero filled blocks. > > >> - current batch of pages isn't in the same datablock. > > >> - decompressor error. > > >> Otherwise pages will be marked as uptodate. The unhandled pages will be > > >> updated by readpage later. > > >> > > >> Suggested-by: Matthew Wilcox > > >> Signed-off-by: Hsin-Yi Wang > > >> Reported-by: Matthew Wilcox > > >> Reported-by: Phillip Lougher > > >> Reported-by: Xiongwei Song > > >> Reported-by: Marek Szyprowski > > >> Reported-by: Andrew Morton > > >> --- > > >> v4->v5: > > >> - Handle short file cases reported by Marek and Matthew. > > >> - Fix checkpatch error reported by Andrew. > > >> > > >> v4: https://lore.kernel.org/lkml/20220601103922.1338320-4-hsinyi@chromium.org/ > > >> v3: https://lore.kernel.org/lkml/20220523065909.883444-4-hsinyi@chromium.org/ > > >> v2: https://lore.kernel.org/lkml/20220517082650.2005840-4-hsinyi@chromium.org/ > > >> v1: https://lore.kernel.org/lkml/20220516105100.1412740-3-hsinyi@chromium.org/ > > >> --- > > >> fs/squashfs/file.c | 124 ++++++++++++++++++++++++++++++++++++++++++++- > > >> 1 file changed, 123 insertions(+), 1 deletion(-) > > >> > > >> diff --git a/fs/squashfs/file.c b/fs/squashfs/file.c > > >> index a8e495d8eb86..fbd096cd15f4 100644 > > >> --- a/fs/squashfs/file.c > > >> +++ b/fs/squashfs/file.c > > >> @@ -39,6 +39,7 @@ > > >> #include "squashfs_fs_sb.h" > > >> #include "squashfs_fs_i.h" > > >> #include "squashfs.h" > > >> +#include "page_actor.h" > > >> > > >> /* > > >> * Locate cache slot in range [offset, index] for specified inode. If > > >> @@ -495,7 +496,128 @@ static int squashfs_read_folio(struct file *file, struct folio *folio) > > >> return 0; > > >> } > > >> > > >> +static void squashfs_readahead(struct readahead_control *ractl) > > >> +{ > > >> + struct inode *inode = ractl->mapping->host; > > >> + struct squashfs_sb_info *msblk = inode->i_sb->s_fs_info; > > >> + size_t mask = (1UL << msblk->block_log) - 1; > > >> + unsigned short shift = msblk->block_log - PAGE_SHIFT; > > >> + loff_t start = readahead_pos(ractl) & ~mask; > > >> + size_t len = readahead_length(ractl) + readahead_pos(ractl) - start; > > >> + struct squashfs_page_actor *actor; > > >> + unsigned int nr_pages = 0; > > >> + struct page **pages; > > >> + int i, file_end = i_size_read(inode) >> msblk->block_log; > > >> + unsigned int max_pages = 1UL << shift; > > >> + > > >> + readahead_expand(ractl, start, (len | mask) + 1); > > >> + > > >> + if (file_end == 0) > > >> + return; > > >> + > > >> + pages = kmalloc_array(max_pages, sizeof(void *), GFP_KERNEL); > > >> + if (!pages) > > >> + return; > > >> + > > >> + actor = squashfs_page_actor_init_special(pages, max_pages, 0); > > >> + if (!actor) > > >> + goto out; > > >> + > > >> + for (;;) { > > >> + pgoff_t index; > > >> + int res, bsize; > > >> + u64 block = 0; > > >> + unsigned int expected; > > >> + > > >> + nr_pages = __readahead_batch(ractl, pages, max_pages); > > >> + if (!nr_pages) > > >> + break; > > >> + > > >> + if (readahead_pos(ractl) >= i_size_read(inode)) > > >> + goto skip_pages; > > >> + > > >> + index = pages[0]->index >> shift; > > >> + if ((pages[nr_pages - 1]->index >> shift) != index) > > >> + goto skip_pages; > > >> + > > >> + expected = index == file_end ? > > >> + (i_size_read(inode) & (msblk->block_size - 1)) : > > >> + msblk->block_size; > > >> + > > >> + bsize = read_blocklist(inode, index, &block); > > >> + if (bsize == 0) > > >> + goto skip_pages; > > >> + > > >> + if (nr_pages < max_pages) { > > >> + struct squashfs_cache_entry *buffer; > > >> + unsigned int block_mask = max_pages - 1; > > >> + int offset = pages[0]->index - (pages[0]->index & ~block_mask); > > >> + > > >> + buffer = squashfs_get_datablock(inode->i_sb, block, > > >> + bsize); > > >> + if (buffer->error) { > > >> + squashfs_cache_put(buffer); > > >> + goto skip_pages; > > >> + } > > >> + > > >> + expected -= offset * PAGE_SIZE; > > >> + for (i = 0; i < nr_pages && expected > 0; i++, > > >> + expected -= PAGE_SIZE, offset++) { > > >> + int avail = min_t(int, expected, PAGE_SIZE); > > >> + > > >> + squashfs_fill_page(pages[i], buffer, > > >> + offset * PAGE_SIZE, avail); > > >> + unlock_page(pages[i]); > > >> + } > > >> + > > >> + squashfs_cache_put(buffer); > > >> + continue; > > >> + } > > >> + > > >> + res = squashfs_read_data(inode->i_sb, block, bsize, NULL, > > >> + actor); > > >> + > > >> + if (res == expected) { > > >> + int bytes; > > >> + > > >> + /* Last page may have trailing bytes not filled */ > > >> + bytes = res % PAGE_SIZE; > > >> + if (bytes) { > > >> + void *pageaddr; > > >> + > > >> + pageaddr = kmap_atomic(pages[nr_pages - 1]); > > >> + memset(pageaddr + bytes, 0, PAGE_SIZE - bytes); > > >> + kunmap_atomic(pageaddr); > > >> + } > > >> + > > >> + for (i = 0; i < nr_pages; i++) { > > >> + flush_dcache_page(pages[i]); > > >> + SetPageUptodate(pages[i]); > > >> + } > > >> + } > > >> + > > >> + for (i = 0; i < nr_pages; i++) { > > >> + unlock_page(pages[i]); > > >> + put_page(pages[i]); > > >> + } > > >> + } > > >> + > > >> + kfree(actor); > > >> + kfree(pages); > > >> + return; > > >> + > > >> +skip_pages: > > >> + for (i = 0; i < nr_pages; i++) { > > >> + unlock_page(pages[i]); > > >> + put_page(pages[i]); > > >> + } > > >> + > > >> + kfree(actor); > > >> +out: > > >> + kfree(pages); > > >> +} > > >> > > >> const struct address_space_operations squashfs_aops = { > > >> - .read_folio = squashfs_read_folio > > >> + .read_folio = squashfs_read_folio, > > >> + .readahead = squashfs_readahead > > >> }; > > >> -- > > >> 2.36.1.255.ge46751e96f-goog > > >> > > >> > >