Received: by 2002:a25:86ce:0:0:0:0:0 with SMTP id y14csp1195727ybm; Wed, 22 May 2019 19:31:25 -0700 (PDT) X-Google-Smtp-Source: APXvYqzm/vxYvOQlvBn02oy7pQcyDrsPriNuPrEnY1VUlXqTSQvdLibCYA4ck2EOTyzToAFZLvtg X-Received: by 2002:a63:f10e:: with SMTP id f14mr92916655pgi.226.1558578685224; Wed, 22 May 2019 19:31:25 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1558578685; cv=none; d=google.com; s=arc-20160816; b=ms8UkRkU7YcexEGTf30DX10lTgi7T+dVvJknkGxzqH38cPqblRFwoToq/5Hr3DHz5l 7t2B3SOwag6eSF42qTMH7BvXTZ/DKbUfJ3wPYZMOu7LArmMn7HBxpZTipN3XGkKwHhyL 5TyVH2UciaFz8R/juXjV5uESUMOWKklCgWF7W5DZHUf6IBx3JtZ+BOkb40KZgEWcEsVX rzWgRFS8Jo0mMptgZC+P9uNEZGlNggaKy2mQqG2p03fwWIe470u6bZhHJuLPTgib0+p9 Db/yQE32XFDi7FAhp0xf8FfdhRSOJoZS+7s70E1ESVlFl2CP72tchnNOcHJEha6VgzQ3 +Fpg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:content-transfer-encoding :content-language:in-reply-to:mime-version:user-agent:date :message-id:from:references:cc:to:subject; bh=yC31A6OBJJZK7i5zQRuRndtvnyYZO7dnYUN41043veM=; b=nMArdKfYNjt7YJGheF1CzaY9y5NR+YBmBBvsVZNTIWGzRKC/eueFNWiJqXFaNGR+5j 7nzdK5WmaoSuQOLEd/IosvFHE28srjqUcRYQeBZ2Sycew+LVZ9yXjGOdEsSbZQcA1Ztp ryc4SY3rF6ZAB+8v+rg43inZsKo2MdDCKykOWxer2HMPItdQJYCGD0uVfyknm56oAQmJ EWdRuCdGcZkFcqyR9VcPLvTSGzjNk5McZpz6w8xGJHA0EPAE7UX9Tt5sjlRj+gqU21ce xjK9o9sO9kgManJV7EoUO7pamWb/pdUFRLE4WEVsgsZ1YDyZvBRtgwUU5YE/3d+YUR5h CReg== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-ext4-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-ext4-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=csail.mit.edu Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id m14si22490035pls.393.2019.05.22.19.31.10; Wed, 22 May 2019 19:31:25 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-ext4-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-ext4-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-ext4-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=csail.mit.edu Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1729057AbfEWCa1 (ORCPT + 99 others); Wed, 22 May 2019 22:30:27 -0400 Received: from outgoing-stata.csail.mit.edu ([128.30.2.210]:36078 "EHLO outgoing-stata.csail.mit.edu" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1728022AbfEWCa1 (ORCPT ); Wed, 22 May 2019 22:30:27 -0400 Received: from [4.30.142.84] (helo=srivatsab-a01.vmware.com) by outgoing-stata.csail.mit.edu with esmtpsa (TLS1.2:RSA_AES_128_CBC_SHA1:128) (Exim 4.82) (envelope-from ) id 1hTdUn-000CNl-Vx; Wed, 22 May 2019 22:30:18 -0400 Subject: Re: CFQ idling kills I/O performance on ext4 with blkio cgroup controller To: Paolo Valente Cc: linux-fsdevel@vger.kernel.org, linux-block , linux-ext4@vger.kernel.org, cgroups@vger.kernel.org, kernel list , Jens Axboe , Jan Kara , jmoyer@redhat.com, Theodore Ts'o , amakhalov@vmware.com, anishs@vmware.com, srivatsab@vmware.com References: <8d72fcf7-bbb4-2965-1a06-e9fc177a8938@csail.mit.edu> <1812E450-14EF-4D5A-8F31-668499E13652@linaro.org> <46c6a4be-f567-3621-2e16-0e341762b828@csail.mit.edu> <07D11833-8285-49C2-943D-E4C1D23E8859@linaro.org> <5B6570A2-541A-4CF8-98E0-979EA6E3717D@linaro.org> <2CB39B34-21EE-4A95-A073-8633CF2D187C@linaro.org> <0e3fdf31-70d9-26eb-7b42-2795d4b03722@csail.mit.edu> <686D6469-9DE7-4738-B92A-002144C3E63E@linaro.org> <01d55216-5718-767a-e1e6-aadc67b632f4@csail.mit.edu> From: "Srivatsa S. Bhat" Message-ID: Date: Wed, 22 May 2019 19:30:14 -0700 User-Agent: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.13; rv:60.0) Gecko/20100101 Thunderbird/60.6.1 MIME-Version: 1.0 In-Reply-To: Content-Type: text/plain; charset=windows-1252 Content-Language: en-US Content-Transfer-Encoding: 7bit Sender: linux-ext4-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-ext4@vger.kernel.org On 5/22/19 3:54 AM, Paolo Valente wrote: > > >> Il giorno 22 mag 2019, alle ore 12:01, Srivatsa S. Bhat ha scritto: >> >> On 5/22/19 2:09 AM, Paolo Valente wrote: >>> >>> First, thank you very much for testing my patches, and, above all, for >>> sharing those huge traces! >>> >>> According to the your traces, the residual 20% lower throughput that you >>> record is due to the fact that the BFQ injection mechanism takes a few >>> hundredths of seconds to stabilize, at the beginning of the workload. >>> During that setup time, the throughput is equal to the dreadful ~60-90 KB/s >>> that you see without this new patch. After that time, there >>> seems to be no loss according to the trace. >>> >>> The problem is that a loss lasting only a few hundredths of seconds is >>> however not negligible for a write workload that lasts only 3-4 >>> seconds. Could you please try writing a larger file? >>> >> >> I tried running dd for longer (about 100 seconds), but still saw around >> 1.4 MB/s throughput with BFQ, and between 1.5 MB/s - 1.6 MB/s with >> mq-deadline and noop. > > Ok, then now the cause is the periodic reset of the mechanism. > > It would be super easy to fill this gap, by just gearing the mechanism > toward a very aggressive injection. The problem is maintaining > control. As you can imagine from the performance gap between CFQ (or > BFQ with malfunctioning injection) and BFQ with this fix, it is very > hard to succeed in maximizing the throughput while at the same time > preserving control on per-group I/O. > Ah, I see. Just to make sure that this fix doesn't overly optimize for total throughput (because of the testcase we've been using) and end up causing regressions in per-group I/O control, I ran a test with multiple simultaneous dd instances, each writing to a different portion of the filesystem (well separated, to induce seeks), and each dd task bound to its own blkio cgroup. I saw similar results with and without this patch, and the throughput was equally distributed among all the dd tasks. > On the bright side, you might be interested in one of the benefits > that BFQ gives in return for this ~10% loss of throughput, in a > scenario that may be important for you (according to affiliation you > report): from ~500% to ~1000% higher throughput when you have to serve > the I/O of multiple VMs, and to guarantee at least no starvation to > any VM [1]. The same holds with multiple clients or containers, and > in general with any set of entities that may compete for storage. > > [1] https://www.linaro.org/blog/io-bandwidth-management-for-production-quality-services/ > Great article! :) Thank you for sharing it! >> But I'm not too worried about that difference. >> >>> In addition, I wanted to ask you whether you measured BFQ throughput >>> with traces disabled. This may make a difference. >>> >> >> The above result (1.4 MB/s) was obtained with traces disabled. >> >>> After trying writing a larger file, you can try with low_latency on. >>> On my side, it causes results to become a little unstable across >>> repetitions (which is expected). >>> >> With low_latency on, I get between 60 KB/s - 100 KB/s. >> > > Gosh, full regression. Fortunately, it is simply meaningless to use > low_latency in a scenario where the goal is to guarantee per-group > bandwidths. Low-latency heuristics, to reach their (low-latency) > goals, modify the I/O schedule compared to the best schedule for > honoring group weights and boosting throughput. So, as recommended in > BFQ documentation, just switch low_latency off if you want to control > I/O with groups. It may still make sense to leave low_latency on > in some specific case, which I don't want to bother you about. > My main concern here is about Linux's I/O performance out-of-the-box, i.e., with all default settings, which are: - cgroups and blkio enabled (systemd default) - blkio non-root cgroups in use (this is the implicit systemd behavior if docker is installed; i.e., it runs tasks under user.slice) - I/O scheduler with blkio group sched support: bfq - bfq default configuration: low_latency = 1 If this yields a throughput that is 10x-30x slower than what is achievable, I think we should either fix the code (if possible) or change the defaults such that they don't lead to this performance collapse (perhaps default low_latency to 0 if bfq group scheduling is in use?) > However, I feel bad with such a low throughput :) Would you be so > kind to provide me with a trace? > Certainly! Short runs of dd resulted in a lot of variation in the throughput (between 60 KB/s - 1 MB/s), so I increased dd's runtime to get repeatable numbers (~70 KB/s). As a result, the trace file (trace-bfq-boost-injection-low-latency-71KBps) is quite large, and is available here: https://www.dropbox.com/s/svqfbv0idcg17pn/bfq-traces.tar.gz?dl=0 Also, I'm very happy to run additional tests or experiments to help track down this issue. So, please don't hesitate to let me know if you'd like me to try anything else or get you additional traces etc. :) Thank you! Regards, Srivatsa VMware Photon OS