Received: by 2002:a05:6902:102b:0:0:0:0 with SMTP id x11csp1018408ybt; Fri, 19 Jun 2020 21:47:13 -0700 (PDT) X-Google-Smtp-Source: ABdhPJyXaAXNnbcys4oXsX/7ha9vgRlk3aHEdFQxIQnTuBjjnvLMQ+PPlpBZNPDatt9D5NHWybef X-Received: by 2002:a17:906:a889:: with SMTP id ha9mr6906146ejb.429.1592628433159; Fri, 19 Jun 2020 21:47:13 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1592628433; cv=none; d=google.com; s=arc-20160816; b=u6SUU8EA0szMynj8pjQQVfB57gUNqZmRhYiQm4gtIK/LRLnjF6S+Tf+gVk490PLJGc uZ3OU0uRmawlkpEbAVEEHqLujxEv47IDtrgE7Az5Rbb35nUIWpmXKPmXSynZwdJN8bEW urzpst1bqswx96HRAvsiN2gE9v/zKVgDba2eb7kfXJZC81ZjBQfDRC1wTwagwS58k4eg 1qIxO+QuvMPEt+RYIomiQCBdinKDO5zqQyXf2PrSaW2L9rYPeVaVHhuPWVxUS+lulxde r0Qb2FQCkfXGHFuvgTIwd5QFFojWb3ErchupgbtAckJsoeFHdqe8wbKcmJ/KbwRFv/10 eslQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:content-transfer-encoding :content-language:in-reply-to:mime-version:user-agent:date :message-id:from:references:cc:to:subject:dkim-signature; bh=gbwiB4/hC7J/wbW7mR3PpzNGlPblQy2+As+ZOMxuqyQ=; b=rsIpyy8WQBXu4My/t/FrKc53rW7iL7dgQwwg0BZqwUFmqSj92ybS+5Fo44Th2goPQv vky8W+xnzToW9RV+2xx8w/mRM2n01SMPpqIx9GBIXwuxMnYWC7jzactBlpR+NDcbiv41 fodllEAX1HXeor4S1OYR1t9Kcq9j45Zsyo1xohSsBwVfhmRvkVlzI5KQJbvFxOFnKKop 639fPZBGNxrSNPMvzatZLOqXxqKE1lvbqXl92sROjzntjVutoYbXIBJ3e1+1oqMB13xu qUwQYGsEvtgy2v8vdwj8F7SlgUq0TVremvNQzrmXUp536qhD4euJwLLrGoLagpQhdeyJ IYJQ== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@kernel-dk.20150623.gappssmtp.com header.s=20150623 header.b=UAdp4bJm; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id n10si5221190eja.130.2020.06.19.21.46.50; Fri, 19 Jun 2020 21:47:13 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=pass header.i=@kernel-dk.20150623.gappssmtp.com header.s=20150623 header.b=UAdp4bJm; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S2391309AbgFSU4N (ORCPT + 99 others); Fri, 19 Jun 2020 16:56:13 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:54342 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S2389020AbgFSU4M (ORCPT ); Fri, 19 Jun 2020 16:56:12 -0400 Received: from mail-pf1-x444.google.com (mail-pf1-x444.google.com [IPv6:2607:f8b0:4864:20::444]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 236FAC06174E for ; Fri, 19 Jun 2020 13:56:12 -0700 (PDT) Received: by mail-pf1-x444.google.com with SMTP id x207so4944220pfc.5 for ; Fri, 19 Jun 2020 13:56:12 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=kernel-dk.20150623.gappssmtp.com; s=20150623; h=subject:to:cc:references:from:message-id:date:user-agent :mime-version:in-reply-to:content-language:content-transfer-encoding; bh=gbwiB4/hC7J/wbW7mR3PpzNGlPblQy2+As+ZOMxuqyQ=; b=UAdp4bJm3lhtD+q8SGUaWB4JL3VYWJToLBOt6hVQpCPIwbgfXWMtEqY6M3y0vraEaB EXVD4kLFSKzbR/6Ed9mNLkcCyDPYLC4M2y3s9k0SYHeivx+9ZR6USxN9d7guuFBjpIcp zvQrculOAuSSDi+3odE35LX0RcPmNaOsMYLHfGGRYgnwIy2k0TuQv7/iGJbr4aJ/z++1 LFwjS6ZftTkt6TMR+77lSH6k2Fcgd/TtfXSdIdaPZgQxy0nBC4sy93A5bNvpbXOnxS5n BzxDGhcfO6nuGwuMgk8uZtTGyVX0KORMwTIEDJUfPX82DEjAbxi7Bm1TA/FDhHEjGIlx S2xA== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:subject:to:cc:references:from:message-id:date :user-agent:mime-version:in-reply-to:content-language :content-transfer-encoding; bh=gbwiB4/hC7J/wbW7mR3PpzNGlPblQy2+As+ZOMxuqyQ=; b=Eq57RJgE1ucqaEji/mwN+KqWzXt4X3CbVxUea2l0LpIjmbTbyyPDuq47NnnAU7KYqw 1PJZ8ryNQfIeG/zvAo//nQ+m+l8fE0+GQB/dBCATqChjQ0E7PI+FazkDpOw1mu8J+Bd4 PsYX1IgE4gUYgdjU402ghGPUadN3z3Q31FDsHS7b4xCmZvvMtRltKAKyo8FOqp2sU/1W wZeLEXN2OVn1BI0XGh+Ns75/m9DathWiQ3da4Ikx/L6BGRCnahNdxrDj55Xmm9zKAPn0 +DZlvlrbPGf1qDI2csrE6HfUkrVRmFSc5s0ICQKIQCAWeNQKcZ5Co6lz67jdDOOFVK51 CR1A== X-Gm-Message-State: AOAM533uElCK1KAp+xup1PxT2fHnC/83fGbBMnDCHTkY/8Hfp3PzQIGz uh+xqKgIzIfR5Wu9KZqceq3/kA== X-Received: by 2002:a05:6a00:15c7:: with SMTP id o7mr9403224pfu.51.1592600171270; Fri, 19 Jun 2020 13:56:11 -0700 (PDT) Received: from ?IPv6:2620:10d:c085:21c1::1762? ([2620:10d:c090:400::5:a8e1]) by smtp.gmail.com with ESMTPSA id m20sm7298847pfk.52.2020.06.19.13.56.09 (version=TLS1_3 cipher=TLS_AES_128_GCM_SHA256 bits=128/128); Fri, 19 Jun 2020 13:56:10 -0700 (PDT) Subject: Re: [PATCH v3] docs: block: Create blk-mq documentation To: =?UTF-8?Q?Andr=c3=a9_Almeida?= , corbet@lwn.net, linux-block@vger.kernel.org, linux-doc@vger.kernel.org Cc: linux-kernel@vger.kernel.org, kernel@collabora.com, krisman@collabora.com, rdunlap@infradead.org, dongli.zhang@oracle.com References: <20200619200114.6896-1-andrealmeid@collabora.com> From: Jens Axboe Message-ID: Date: Fri, 19 Jun 2020 14:56:08 -0600 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:68.0) Gecko/20100101 Thunderbird/68.8.0 MIME-Version: 1.0 In-Reply-To: <20200619200114.6896-1-andrealmeid@collabora.com> Content-Type: text/plain; charset=utf-8 Content-Language: en-US Content-Transfer-Encoding: 8bit Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 6/19/20 2:01 PM, André Almeida wrote: > Create a documentation providing a background and explanation around the > operation of the Multi-Queue Block IO Queueing Mechanism (blk-mq). > > The reference for writing this documentation was the source code and > "Linux Block IO: Introducing Multi-queue SSD Access on Multi-core > Systems", by Axboe et al. > > Signed-off-by: André Almeida > --- > Changes from v2: > - More fixed typos > - Once again, reworked the definition of `blk_mq_hw_ctx` in "Hardware > dispatch queues" section > > Changes from v1: > - Fixed typos > - Reworked blk_mq_hw_ctx > > Hello, > > This commit was tested using "make htmldocs" and the HTML output has > been verified. > > Thanks, > André > --- > Documentation/block/blk-mq.rst | 155 +++++++++++++++++++++++++++++++++ > Documentation/block/index.rst | 1 + > 2 files changed, 156 insertions(+) > create mode 100644 Documentation/block/blk-mq.rst > > diff --git a/Documentation/block/blk-mq.rst b/Documentation/block/blk-mq.rst > new file mode 100644 > index 000000000000..d1b8f04a822d > --- /dev/null > +++ b/Documentation/block/blk-mq.rst > @@ -0,0 +1,155 @@ > +.. SPDX-License-Identifier: GPL-2.0 > + > +================================================ > +Multi-Queue Block IO Queueing Mechanism (blk-mq) > +================================================ > + > +The Multi-Queue Block IO Queueing Mechanism is an API to enable fast storage > +devices to achieve a huge number of input/output operations per second (IOPS) > +through queueing and submitting IO requests to block devices simultaneously, > +benefiting from the parallelism offered by modern storage devices. > + > +Introduction > +============ > + > +Background > +---------- > + > +Magnetic hard disks have been the de facto standard from the beginning of the > +development of the kernel. The Block IO subsystem aimed to achieve the best > +performance possible for those devices with a high penalty when doing random > +access, and the bottleneck was the mechanical moving parts, a lot slower than > +any layer on the storage stack. One example of such optimization technique > +involves ordering read/write requests according to the current position of the > +hard disk head. > + > +However, with the development of Solid State Drives and Non-Volatile Memories > +without mechanical parts nor random access penalty and capable of performing > +high parallel access, the bottleneck of the stack had moved from the storage > +device to the operating system. In order to take advantage of the parallelism > +in those devices' design, the multi-queue mechanism was introduced. > + > +The former design had a single queue to store block IO requests with a single > +lock. That did not scale well in SMP systems due to dirty data in cache and the > +bottleneck of having a single lock for multiple processors. This setup also > +suffered with congestion when different processes (or the same process, moving > +to different CPUs) wanted to perform block IO. Instead of this, the blk-mq API > +spawns multiple queues with individual entry points local to the CPU, removing > +the need for a lock. A deeper explanation on how this works is covered in the > +following section (`Operation`_). > + > +Operation > +--------- > + > +When the userspace performs IO to a block device (reading or writing a file, > +for instance), blk-mq takes action: it will store and manage IO requests to > +the block device, acting as middleware between the userspace (and a file > +system, if present) and the block device driver. > + > +blk-mq has two group of queues: software staging queues and hardware dispatch > +queues. When the request arrives at the block layer, it will try the shortest > +path possible: send it directly to the hardware queue. However, there are two > +cases that it might not do that: if there's an IO scheduler attached at the > +layer or if we want to try to merge requests. In both cases, requests will be > +sent to the software queue. > + > +Then, after the requests are processed by software queues, they will be placed > +at the hardware queue, a second stage queue were the hardware has direct access > +to process those requests. However, if the hardware does not have enough > +resources to accept more requests, blk-mq will places requests on a temporary > +queue, to be sent in the future, when the hardware is able. > + > +Software staging queues > +~~~~~~~~~~~~~~~~~~~~~~~ > + > +The block IO subsystem adds requests (represented by struct > +:c:type:`blk_mq_ctx`) in the software staging queues in case that they weren't This reads a bit funny, did you want to put the blk_mq_ctx thing after the "software staging queues"? Right now it looks like the requests are of that type, which of course isn't true. > +sent directly to the driver. A request is a collection of BIOs. They arrived at I'd say "one or more BIOs", as there can be just one. > +IO Schedulers > +^^^^^^^^^^^^^ > + > +There are several schedulers implemented by the block layer, each one following > +a heuristic to improve the IO performance. They are "pluggable" (as in plug > +and play), in the sense of they can be selected at run time using sysfs. You > +can read more about Linux's IO schedulers `here > +`_. The scheduling > +happens only between requests in the same queue, so it is not possible to merge > +requests from different queues, otherwise there would be cache trashing and a > +need to have a lock for each queue. After the scheduling, the requests are > +eligible to be sent to the hardware. One of the possible schedulers to be > +selected is the NOOP scheduler, the most straightforward one, that implements a > +simple FIFO, without performing any reordering. This is useful in the following NOOP is a relic from the single queue days, the basic "doesn't do much" scheduler is NONE these days. And it doesn't provide FIFO ordering, requests will basically just end up in whatever software queue the process is running on. When someone runs the hardware queue, the software queues mapped to that hardware queue will be drained in sequence according to their mapping (generally from 0..N, if 0..N are mapped to that hardware queue). -- Jens Axboe