Received: by 2002:a05:6902:102b:0:0:0:0 with SMTP id x11csp3676179ybt; Tue, 23 Jun 2020 08:10:16 -0700 (PDT) X-Google-Smtp-Source: ABdhPJyKngwmkFKLKjX+OpNwcQog4QpM1RidOQjtA6a0bUXdhaMhkOYYho8IhgFbkkhYQi1s/lwG X-Received: by 2002:a17:906:4c2:: with SMTP id g2mr15708215eja.299.1592925016151; Tue, 23 Jun 2020 08:10:16 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1592925016; cv=none; d=google.com; s=arc-20160816; b=kMpFyCk6RttOc7PbypK6XofRuK3pYGoBjn26ZqYnONaq8376Sjzd965kOWUa/7K9o1 NJ/dix+N5DhtnnY45GwubB9EPJcJetoaVDujho7NfzDtWqFZfYRNdtajImjwAi+8SpWb 4HGhKITsI/8P6OC1X27edYZXc8LmVjLptpUzQOxiaj1SBYWF0LecwIvnJA1UxRcprqWV gIOeE+qOjgKzqois2TXwMWJ9H72fYx9e5OyhdE/XKCLdfidf+hIsftUT45sz05AePzFz VRwovyQiYSvRsY4P5Y1I7wryGUwnjQ/z4Eu1PF93/hYHBAunmq7mx9ZYuJU3cFa9fEmA 92pg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:cc:to:subject:message-id:date:from :in-reply-to:references:mime-version:dkim-signature; bh=fOJ4r5eRLu7pYwZ7nb72+ethCqIdEc1b6ZQTMBOmtl0=; b=FMSdCL+TooL7NrtnIvftt4WQPl4HQe18G0t2fbsbZOH0pBmPnvQqnqQLV7SW3C3Vof 0LgBEAWQdvF0lyZIyNTBl0VOGRKLMJjsYT5f2+Y70N3o7Nni05MVI7Stz1xI9y1rhIeb m1ySTTIZanE1Qt2umgM/1bfWBtldOhULEY9kUKqnAzY0U1ybpZNRjf0u8QJvIE3NdPM6 gzlWv9qgV9jnZxjpeKYh3nxCPA2xQnEsUT+qPIdHv1jtLKdpRzmfI2qIkFKhkqJslrNK /5KKGUu1AegrJngD2TayLBYeGgYydaLxc/plxXeyNAVjCrhRjANIXOC3+A6vnIJfeNEQ ksYQ== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@cloudflare.com header.s=google header.b=eyGFPsb3; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=REJECT sp=REJECT dis=NONE) header.from=cloudflare.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id k4si10636929ejg.386.2020.06.23.08.09.51; Tue, 23 Jun 2020 08:10:16 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=pass header.i=@cloudflare.com header.s=google header.b=eyGFPsb3; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=REJECT sp=REJECT dis=NONE) header.from=cloudflare.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1732947AbgFWPHs (ORCPT + 99 others); Tue, 23 Jun 2020 11:07:48 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:60466 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1732868AbgFWPHs (ORCPT ); Tue, 23 Jun 2020 11:07:48 -0400 Received: from mail-lj1-x241.google.com (mail-lj1-x241.google.com [IPv6:2a00:1450:4864:20::241]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id CC488C061573 for ; Tue, 23 Jun 2020 08:07:47 -0700 (PDT) Received: by mail-lj1-x241.google.com with SMTP id e4so23838183ljn.4 for ; Tue, 23 Jun 2020 08:07:47 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=cloudflare.com; s=google; h=mime-version:references:in-reply-to:from:date:message-id:subject:to :cc; bh=fOJ4r5eRLu7pYwZ7nb72+ethCqIdEc1b6ZQTMBOmtl0=; b=eyGFPsb3Ecv6WIff2bnyEcA8O2/8mOk3tPZLQUq2/H/f8Sd85M7TXgnPiylYEf8dyu gUB03dQE5U2CvM4Ksj/DXZU45UuQfwrgX4Ev7MiRI2FRei2SetffW0alu4zcr+XB9Uoi d3IZsPcLIR5PSKN5vvLNqDrjl234SQc5VASr0= X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:mime-version:references:in-reply-to:from:date :message-id:subject:to:cc; bh=fOJ4r5eRLu7pYwZ7nb72+ethCqIdEc1b6ZQTMBOmtl0=; b=VlLiioxbQB6vJkQnQn6+2pGTS+rMIC4ElFhcF5elP7n92RyYs/sb2D/XcFgh9aB0CE GOblNrJV/LZP+1IzoB5kbmm2x/NdIgz3Rgc4hYIqbLm7uI/avoG+73/u4FJ4eBkNZWvu 0p+9p7/spOYDDj9VfURj8VHLw/l67X9Jz/D0Fwk+DAUqpi75BaUIYKML+LvT6UMHXO+K qax2ROwNh4dr6HPmSGLto3nnswhaNnb+raIzbqse0F4DPqzrsvdSGkcLKdakh5D12N3L OBWLqtYGfgxosQwVMQDnKOeD4/PFSelfLMW2TTuofLU4hlgv/Q3RfcwWyqtcABV6bLtA 16IQ== X-Gm-Message-State: AOAM532xmsmEjRa7d6Czr1fPPuUNGN9rRKKOQ/bF9skETzCVHdlx0+c/ t/I7tXroAuT00n7hde5tSV31uQ/G9wUyCfecBIOBcQ== X-Received: by 2002:a05:651c:2d0:: with SMTP id f16mr10958567ljo.387.1592924866020; Tue, 23 Jun 2020 08:07:46 -0700 (PDT) MIME-Version: 1.0 References: <20200619164132.1648-1-ignat@cloudflare.com> <20200619165548.GA24779@redhat.com> <20200623150118.GA19657@redhat.com> In-Reply-To: <20200623150118.GA19657@redhat.com> From: Ignat Korchagin Date: Tue, 23 Jun 2020 16:07:33 +0100 Message-ID: Subject: Re: [RFC PATCH 0/1] dm-crypt excessive overhead To: Mike Snitzer , Damien Le Moal Cc: Mikulas Patocka , "dm-crypt@saout.de" , "dm-devel@redhat.com" , "linux-kernel@vger.kernel.org" , "agk@redhat.com" , "kernel-team@cloudflare.com" Content-Type: text/plain; charset="UTF-8" Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Do you think it may be better to break it in two flags: one for read path and one for write? So, depending on the needs and workflow these could be enabled independently? Regards, Ignat On Tue, Jun 23, 2020 at 4:01 PM Mike Snitzer wrote: > > On Sun, Jun 21 2020 at 8:45pm -0400, > Damien Le Moal wrote: > > > On 2020/06/20 1:56, Mike Snitzer wrote: > > > On Fri, Jun 19 2020 at 12:41pm -0400, > > > Ignat Korchagin wrote: > > > > > >> This is a follow up from the long-forgotten [1], but with some more convincing > > >> evidence. Consider the following script: > > >> > > >> #!/bin/bash -e > > >> > > >> # create 4G ramdisk > > >> sudo modprobe brd rd_nr=1 rd_size=4194304 > > >> > > >> # create a dm-crypt device with NULL cipher on top of /dev/ram0 > > >> echo '0 8388608 crypt capi:ecb(cipher_null) - 0 /dev/ram0 0' | sudo dmsetup create eram0 > > >> > > >> # create a dm-crypt device with NULL cipher and custom force_inline flag > > >> echo '0 8388608 crypt capi:ecb(cipher_null) - 0 /dev/ram0 0 1 force_inline' | sudo dmsetup create inline-eram0 > > >> > > >> # read all data from /dev/ram0 > > >> sudo dd if=/dev/ram0 bs=4k iflag=direct | sha256sum > > >> > > >> # read the same data from /dev/mapper/eram0 > > >> sudo dd if=/dev/mapper/eram0 bs=4k iflag=direct | sha256sum > > >> > > >> # read the same data from /dev/mapper/inline-eram0 > > >> sudo dd if=/dev/mapper/inline-eram0 bs=4k iflag=direct | sha256sum > > >> > > >> This script creates a ramdisk (to eliminate hardware bias in the benchmark) and > > >> two dm-crypt instances on top. Both dm-crypt instances use the NULL cipher > > >> to eliminate potentially expensive crypto bias (the NULL cipher just uses memcpy > > >> for "encyption"). The first instance is the current dm-crypt implementation from > > >> 5.8-rc1, the second is the dm-crypt instance with a custom new flag enabled from > > >> the patch attached to this thread. On my VM (Debian in VirtualBox with 4 cores > > >> on 2.8 GHz Quad-Core Intel Core i7) I get the following output (formatted for > > >> better readability): > > >> > > >> # plain ram0 > > >> 1048576+0 records in > > >> 1048576+0 records out > > >> 4294967296 bytes (4.3 GB, 4.0 GiB) copied, 21.2305 s, 202 MB/s > > >> 8479e43911dc45e89f934fe48d01297e16f51d17aa561d4d1c216b1ae0fcddca - > > >> > > >> # eram0 (current dm-crypt) > > >> 1048576+0 records in > > >> 1048576+0 records out > > >> 4294967296 bytes (4.3 GB, 4.0 GiB) copied, 53.2212 s, 80.7 MB/s > > >> 8479e43911dc45e89f934fe48d01297e16f51d17aa561d4d1c216b1ae0fcddca - > > >> > > >> # inline-eram0 (patched dm-crypt) > > >> 1048576+0 records in > > >> 1048576+0 records out > > >> 4294967296 bytes (4.3 GB, 4.0 GiB) copied, 21.3472 s, 201 MB/s > > >> 8479e43911dc45e89f934fe48d01297e16f51d17aa561d4d1c216b1ae0fcddca - > > >> > > >> As we can see, current dm-crypt implementation creates a significant IO > > >> performance overhead (at least on small IO block sizes) for both latency and > > >> throughput. We suspect offloading IO request processing into workqueues and > > >> async threads is more harmful these days with the modern fast storage. I also > > >> did some digging into the dm-crypt git history and much of this async processing > > >> is not needed anymore, because the reasons it was added are mostly gone from the > > >> kernel. More details can be found in [2] (see "Git archeology" section). > > >> > > >> We have been running the attached patch on different hardware generations in > > >> more than 200 datacentres on both SATA SSDs and NVME SSDs and so far were very > > >> happy with the performance benefits. > > >> > > >> [1]: https://www.spinics.net/lists/dm-crypt/msg07516.html > > >> [2]: https://blog.cloudflare.com/speeding-up-linux-disk-encryption/ > > >> > > >> Ignat Korchagin (1): > > >> Add DM_CRYPT_FORCE_INLINE flag to dm-crypt target > > >> > > >> drivers/md/dm-crypt.c | 55 +++++++++++++++++++++++++++++++++---------- > > >> 1 file changed, 43 insertions(+), 12 deletions(-) > > >> > > >> -- > > >> 2.20.1 > > >> > > > > > > Hi, > > > > > > I saw [2] and have been expecting something from cloudflare ever since. > > > Nice to see this submission. > > > > > > There is useful context in your 0th patch header. I'll likely merge > > > parts of this patch header with the more terse 1/1 header (reality is > > > there only needed to be a single patch submission). > > > > > > Will review and stage accordingly if all looks fine to me. Mikulas, > > > please have a look too. > > > > Very timely: I was about to send a couple of patches to add zoned block device > > support to dm-crypt :) > > > > I used [1] work as a base to have all _write_ requests be processed inline in > > the submitter context so that the submission order is preserved, avoiding the > > potential reordering of sequential writes that the normal workqueue based > > processing can generate. This inline processing is done only for writes. Reads > > are unaffected. > > > > To do this, I added a "inline_io" flag to struct convert_context which is > > initialized in crypt_io_init() based on the BIO op. > > kcryptd_crypt_write_io_submit() then uses this flag to call directly > > generic_make_request() if inline_io is true. > > > > This simplifies things compared to [1] since reads can still be processed as is, > > so there are no issued with irq context and no need for a tasklet. > > > > Should I send these patches as RFC to see what can be merged ? Or I can wait for > > these patches and rebase on top. > > It'd be ideal for this inline capability to address both Ignat's and > your needs. Given Ignat's changes _should_ enable yours (and Ignat > clarified that having reads inline is actually important) then I think it > best if you review Ignat's patch closely, rebase on it and test that it > meets your needs. > > I'll wait for you to do this work so that I can get your feedback on > whether Ignat's changes look good for you too. We have some time before > the 5.9 merge window opens, lets just keep the communication going and > make sure what we send upstream addresses everyone's needs and concerns. > > Thanks, > Mike >