Received: by 2002:a25:683:0:0:0:0:0 with SMTP id 125csp2324323ybg; Fri, 5 Jun 2020 11:00:02 -0700 (PDT) X-Google-Smtp-Source: ABdhPJx78hn6TZMJ26bmjrvmdHwefrBlLl4iWGCT095skyrZ0h7kI/Nt8jjm8N53y4CQoH6AY8eL X-Received: by 2002:a17:906:8684:: with SMTP id g4mr9327354ejx.431.1591380002215; Fri, 05 Jun 2020 11:00:02 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1591380002; cv=none; d=google.com; s=arc-20160816; b=S4YIboQ1Rgw8gcUwYT97mDq7pbykZ2yRMxI9exiiwfwVpfjwLqwn8pyIOGYZqeyORR n8nxucYMjJKnrPzbBYSrJaOJQNAtRk0kq5CcbftCRgEkORslmE0AtAaFNn63N5likl7r h0Echb/eBjUfnyqZdZvLrkCNTm9BYbmwZ7SWvI875oMVCXb8Tz7b14FoxPuOKv7vS1pD USmsqNkG9ROuV9rioGlrPePocY1otFgmsZXfXULMIueQyf/oUQ75nImUwXIxj0iBG+tb SryFf4dD+1VAI0A779XeZ9PRhkASv7SnE1Pr5sQY6PMiFc8o3bT+Atxla6h8ZgqHOW/k /Rww== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:content-transfer-encoding:mime-version :message-id:date:subject:cc:to:from; bh=aQzK76duIKiA/sHzzog7x9jK4qxu5LJUb+CLGBJJbm8=; b=o7pmMeJQiXNjX1fjzstVYOT86qtiH0FOPXSe79xI3VdmDMGxbheOlxxs3hpke+fUBN j4gz90t1x8xWP3B39vlBx2IFMFqNWx9REeQUPLGO3Nmv63loHeQkhzhnQGoXuSgxf0sw PuKBCDwE+3woF0Ad07OAYdCRq965RZLClvDdxBp1u7ATHK73BE9Hf/qKofx0WPdaLwn6 DS4tanzApg0Nz/fob6ImWZLeqif+WnBvfe4B0/2+dMJVkbqlBSz0YfzIik9xi/MqgEZu wvzTS2jQU1GkNPwiIrVKApbW0yNkXPFznxzQsPPcHkaHISdHhaqfJEmdfuExU/EL3pfL aAtQ== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=collabora.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id d14si4242244edp.573.2020.06.05.10.59.39; Fri, 05 Jun 2020 11:00:02 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=collabora.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1727993AbgFERzs (ORCPT + 99 others); Fri, 5 Jun 2020 13:55:48 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:54972 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726245AbgFERzs (ORCPT ); Fri, 5 Jun 2020 13:55:48 -0400 Received: from bhuna.collabora.co.uk (bhuna.collabora.co.uk [IPv6:2a00:1098:0:82:1000:25:2eeb:e3e3]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id EAF7BC08C5C2; Fri, 5 Jun 2020 10:55:47 -0700 (PDT) Received: from [127.0.0.1] (localhost [127.0.0.1]) (Authenticated sender: tonyk) with ESMTPSA id 301462A45BD From: =?UTF-8?q?Andr=C3=A9=20Almeida?= To: axboe@kernel.dk, corbet@lwn.net, linux-block@vger.kernel.org, linux-doc@vger.kernel.org Cc: linux-kernel@vger.kernel.org, kernel@collabora.com, krisman@collabora.com, rdunlap@infradead.org, =?UTF-8?q?Andr=C3=A9=20Almeida?= Subject: [PATCH v2] docs: block: Create blk-mq documentation Date: Fri, 5 Jun 2020 14:55:36 -0300 Message-Id: <20200605175536.19681-1-andrealmeid@collabora.com> X-Mailer: git-send-email 2.27.0 MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Create a documentation providing a background and explanation around the operation of the Multi-Queue Block IO Queueing Mechanism (blk-mq). The reference for writing this documentation was the source code and "Linux Block IO: Introducing Multi-queue SSD Access on Multi-core Systems", by Axboe et al. Signed-off-by: André Almeida --- Changes from v1: - Fixed typos - Reworked blk_mq_hw_ctx Hello, This commit was tested using "make htmldocs" and the HTML output has been verified. Thanks, André --- Documentation/block/blk-mq.rst | 154 +++++++++++++++++++++++++++++++++ Documentation/block/index.rst | 1 + 2 files changed, 155 insertions(+) create mode 100644 Documentation/block/blk-mq.rst diff --git a/Documentation/block/blk-mq.rst b/Documentation/block/blk-mq.rst new file mode 100644 index 000000000000..1f702adbc577 --- /dev/null +++ b/Documentation/block/blk-mq.rst @@ -0,0 +1,154 @@ +.. SPDX-License-Identifier: GPL-2.0 + +================================================ +Multi-Queue Block IO Queueing Mechanism (blk-mq) +================================================ + +The Multi-Queue Block IO Queueing Mechanism is an API to enable fast storage +devices to achieve a huge number of input/output operations per second (IOPS) +through queueing and submitting IO requests to block devices simultaneously, +benefiting from the parallelism offered by modern storage devices. + +Introduction +============ + +Background +---------- + +Magnetic hard disks have been the de facto standard from the beginning of the +development of the kernel. The Block IO subsystem aimed to achieve the best +performance possible for those devices with a high penalty when doing random +access, and the bottleneck was the mechanical moving parts, a lot more slower +than any layer on the storage stack. One example of such optimization technique +involves ordering read/write requests accordingly to the current position of +the hard disk head. + +However, with the development of Solid State Drives and Non-Volatile Memories +without mechanical parts nor random access penalty and capable of performing +high parallel access, the bottleneck of the stack had moved from the storage +device to the operating system. In order to take advantage of the parallelism +in those devices design, the multi-queue mechanism was introduced. + +The former design had a single queue to store block IO requests with a single +lock. That did not scale well in SMP systems due to dirty data in cache and the +bottleneck of having a single lock for multiple processors. This setup also +suffered with congestion when different processes (or the same process, moving +to different CPUs) wanted to perform block IO. Instead of this, the blk-mq API +spawns multiple queues with individual entry points local to the CPU, removing +the need for a lock. A deeper explanation on how this works is covered in the +following section (`Operation`_). + +Operation +--------- + +When the userspace performs IO to a block device (reading or writing a file, +for instance), blk-mq takes action: it will store and manage IO requests to +the block device, acting as middleware between the userspace (and a file +system, if present) and the block device driver. + +blk-mq has two group of queues: software staging queues and hardware dispatch +queues. When the request arrives at the block layer, it will try the shortest +path possible: send it directly to the hardware queue. However, there are two +cases that it might not do that: if there's an IO scheduler attached at the +layer or if we want to try to merge requests. In both cases, requests will be +sent to the software queue. + +Then, after the requests are processed by software queues, they will be placed +at the hardware queue, a second stage queue were the hardware has direct access +to process those requests. However, if the hardware does not have enough +resources to accept more requests, blk-mq will places requests on a temporary +queue, to be sent in the future, when the hardware is able. + +Software staging queues +~~~~~~~~~~~~~~~~~~~~~~~ + +The block IO subsystem adds requests (represented by struct +:c:type:`blk_mq_ctx`) in the software staging queues in case that they weren't +sent directly to the driver. A request is a collection of BIOs. They arrived at +the block layer through the data structure struct :c:type:`bio`. The block +layer will then build a new structure from it, the struct :c:type:`request` +that will be used to communicate with the device driver. Each queue has its +own lock and the number of queues is defined by a per-CPU or per-node basis. + +The staging queue can be used to merge requests for adjacent sectors. For +instance, requests for sector 3-6, 6-7, 7-9 can become one request for 3-9. +Even if random access to SSDs and NVMs have the same time of response compared +to sequential access, grouped requests for sequential access decreases the +number of individual requests. This technique of merging requests is called +plugging. + +Along with that, the requests can be reordered to ensure fairness of system +resources (e.g. to ensure that no application suffers from starvation) and/or to +improve IO performance, by an IO scheduler. + +IO Schedulers +^^^^^^^^^^^^^ + +There are several schedulers implemented by the block layer, each one following +a heuristic to improve the IO performance. They are "pluggable" (as in plug +and play), in the sense of they can be selected at run time using sysfs. You +can read more about Linux's IO schedulers `here +`_. The scheduling +happens only between requests in the same queue, so it is not possible to merge +requests from different queues, otherwise there would be cache trashing and a +need to have a lock for each queue. After the scheduling, the requests are +eligible to be sent to the hardware. One of the possible schedulers to be +selected is the NOOP scheduler, the most straightforward one, that implements a +simple FIFO, without performing any reordering. This is useful in the following +scenarios: when scheduling will be performed in a next step somewhere in the +stack, like block device controllers; the actual sector position of blocks are +transparent for the host, meaning it hasn't enough information to take a proper +decision; or the overhead of reordering is higher than the handicap of +non-sequential accesses. + +Hardware dispatch queues +~~~~~~~~~~~~~~~~~~~~~~~~ + +The hardware queues (represented by struct :c:type:`blk_mq_hw_ctx`) have a 1:1 +correspondence to the device driver's submission queues, and are the last step +of the block layer submission code before the low level device driver taking +ownership of the request. To run this queue, the block layer removes requests +from the associated software queues and tries to dispatch to the hardware. + +If it's not possible to send the requests directly to hardware, they will be +added to a linked list (:c:type:`hctx->dispatch`) of requests. Then, +next time the block layer runs a queue, it will send the requests laying at the +:c:type:`dispatch` list first, to ensure a fairness dispatch with those +requests that were ready to be sent first. The number of hardware queues +depends on the number of hardware contexts supported by the hardware and its +device driver, but it will not be more than the number of cores of the system. +There is no reordering at this stage, and each software queue has a set of +hardware queues to send requests for. + +.. note:: + + Neither the block layer nor the device protocols guarantee + the order of completion of requests. This must be handled by + higher layers, like the filesystem. + +Tag-based completion +~~~~~~~~~~~~~~~~~~~~ + +In order to indicate which request has been completed, every request is +identified by an integer, ranging from 0 to the dispatch queue size. This tag +is generated by the block layer and later reused by the device driver, removing +the need to create a redundant identifier. When a request is completed in the +drive, the tag is sent back to the block layer to notify it of the finalization. +This removes the need to do a linear search to find out which IO has been +completed. + +Further reading +--------------- + +- `Linux Block IO: Introducing Multi-queue SSD Access on Multi-core Systems `_ + +- `NOOP scheduler `_ + +- `Null block device driver `_ + +Source code documentation +========================= + +.. kernel-doc:: include/linux/blk-mq.h + +.. kernel-doc:: block/blk-mq.c diff --git a/Documentation/block/index.rst b/Documentation/block/index.rst index 3fa7a52fafa4..3a3f38322185 100644 --- a/Documentation/block/index.rst +++ b/Documentation/block/index.rst @@ -10,6 +10,7 @@ Block bfq-iosched biodoc biovecs + blk-mq capability cmdline-partition data-integrity -- 2.27.0