Received: by 2002:a05:6359:c8b:b0:c7:702f:21d4 with SMTP id go11csp693426rwb; Sat, 17 Sep 2022 12:59:00 -0700 (PDT) X-Google-Smtp-Source: AMsMyM5F1iKxRLT9lpl749KgXInBtujqm13rO8eeMpiyFREk0Isc0DUgW3tW+gMz8Iw46sTm8IY+ X-Received: by 2002:a05:6a00:a1d:b0:536:90d3:1930 with SMTP id p29-20020a056a000a1d00b0053690d31930mr11100316pfh.25.1663444739796; Sat, 17 Sep 2022 12:58:59 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1663444739; cv=none; d=google.com; s=arc-20160816; b=cV9yIFhUpDhTYPKyl3dpKY9n3YgS/qF/+CpLbQqt7szMLS8Pr6rLrYFV4VjBdFPSGb 1+gAHZGOmAp3TQe9CWey7yxfLDQbY9TdiTZCRXYBrOVxN7P0uJbOyvGhKZTNLEEaPeY5 GTv8HFulovh1a5T/Yk69bMd7QwcLcIkLSTMMbooTuw1sMHQqYf0t03rU/wNlf0KgHjnl j7EuxNi10Q5XIieBUci7xJHpss9F469MdEr4A9wMDLrSOJI2NR5UMWP53q0YoouFeJtH QP8N4r4myp7/HsN8jXtWhAsJ+G/wAjdGuq9Zzp8J0j+l3GC/UbOuC6TzYtJeddHTJKQ9 SJDQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:cc:to:subject :message-id:date:from:in-reply-to:references:mime-version :dkim-signature; bh=YyeNIHUzNbtSHe6o09PMb3YQ7MMyCYWkOVqFjUHdZoU=; b=HciJBvhmMERyVcshl+2ib0gzQYHQyE+ST1yOwwigcpA8acrP0atVAZYqun6tPgzeCP uzRNiHYWyriD0Uz+1o8h8QuO5mJ2anpZuGmbSm/8yVuAxb5r9sOcDhzNLQ6vuOsCTREb 4zvVK5dzqiajPWqL/vff2sZ0iWaP+bdcSXBOwtx57T8axSGZG3wxkYD06nnJ8b51HlsJ PRkq9UbXoPb519cdr9O58swVE2+94nntxJPwXwjdOQ2zTgQ9Wg3jxQ8Gf33ZxXv5WlCx yF6vdfu640M+WSY6NejVeRaqrIDEY5jWqcEoogZ75yBEw1+8ql8wDL5gP1+jszIK+3zZ WDlw== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@chromium.org header.s=google header.b="iffWO/0y"; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=chromium.org Return-Path: Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id o6-20020a637306000000b0041dcd13c180si25668092pgc.366.2022.09.17.12.58.44; Sat, 17 Sep 2022 12:58:59 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; dkim=pass header.i=@chromium.org header.s=google header.b="iffWO/0y"; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=chromium.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S229455AbiIQTqu (ORCPT + 99 others); Sat, 17 Sep 2022 15:46:50 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:43436 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S229436AbiIQTqs (ORCPT ); Sat, 17 Sep 2022 15:46:48 -0400 Received: from mail-ed1-x532.google.com (mail-ed1-x532.google.com [IPv6:2a00:1450:4864:20::532]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id AE5502BB24 for ; Sat, 17 Sep 2022 12:46:46 -0700 (PDT) Received: by mail-ed1-x532.google.com with SMTP id 29so35977001edv.2 for ; Sat, 17 Sep 2022 12:46:46 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=chromium.org; s=google; h=content-transfer-encoding:cc:to:subject:message-id:date:from :in-reply-to:references:mime-version:from:to:cc:subject:date; bh=YyeNIHUzNbtSHe6o09PMb3YQ7MMyCYWkOVqFjUHdZoU=; b=iffWO/0yAzSK5e6E1WnNnRN8CWp03XEgokCUrm8VcMYG56l+/wiV3OVIX/MzEsVhZz Y3EZ6Hhdpb0rcQXB970URz0oZuHLHSHcPAwjzs9lGhNjN6vrSmxINM5D1mh++walEPBg hKBSM50RAdbX+Dlan4RT0ACVQ7geepnsAmoGE= X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20210112; h=content-transfer-encoding:cc:to:subject:message-id:date:from :in-reply-to:references:mime-version:x-gm-message-state:from:to:cc :subject:date; bh=YyeNIHUzNbtSHe6o09PMb3YQ7MMyCYWkOVqFjUHdZoU=; b=LFFHBYjNC/dQPIxSZ09FuE2sj1JgNDJ0dBK9vHi1nKxNYnEr/qloKPd73FyP4AwsHw Td+Z/v+cQ7xMlpHOZbX4r0yBz7c/FuuV+O1Z+0lsxMwJnpjnDkdcsnddgZTLPadkjMcY G83c7HwMymnSd5mONvYX9dm4bL93w2yBz/c4n+2WRJ2KEBw/tCfTakOlXilYpZqO1kLv I7QuEUxCO7ggqg45Fj47N9QMltLtvV/LHSiqfUWqssJOnYmQM+quGc5fJrfeEoyobZyr YBD2/ikFIT3j2ivpuG0R1SoXyvJY3ajFLpV/83q4jxGL0IJMq7De4FVSznkebllTZ232 qouQ== X-Gm-Message-State: ACrzQf3o8Y3TouYkV+tNUse7rxcH9w6PmWZbkVkCB0f62LKKyCVKuNnU ldkphabc5d/9p3hED7hrzMBWObceLhn1O/+vTBN3RQ== X-Received: by 2002:a05:6402:90a:b0:443:8b10:bcad with SMTP id g10-20020a056402090a00b004438b10bcadmr8770066edz.416.1663444005137; Sat, 17 Sep 2022 12:46:45 -0700 (PDT) MIME-Version: 1.0 References: <20220915164826.1396245-1-sarthakkukreti@google.com> In-Reply-To: From: Sarthak Kukreti Date: Sat, 17 Sep 2022 12:46:33 -0700 Message-ID: Subject: Re: [dm-devel] [PATCH RFC 0/8] Introduce provisioning primitives for thinly provisioned storage To: "Darrick J. Wong" Cc: dm-devel@redhat.com, linux-block@vger.kernel.org, linux-ext4@vger.kernel.org, linux-kernel@vger.kernel.org, virtualization@lists.linux-foundation.org, Jens Axboe , Gwendal Grignou , "Theodore Ts'o" , "Michael S . Tsirkin" , Jason Wang , Bart Van Assche , Mike Snitzer , Evan Green , Andreas Dilger , Daniil Lunev , Stefan Hajnoczi , Paolo Bonzini , Alasdair Kergon Content-Type: text/plain; charset="UTF-8" Content-Transfer-Encoding: quoted-printable X-Spam-Status: No, score=-2.1 required=5.0 tests=BAYES_00,DKIMWL_WL_HIGH, DKIM_SIGNED,DKIM_VALID,DKIM_VALID_AU,DKIM_VALID_EF,RCVD_IN_DNSWL_NONE, SPF_HELO_NONE,SPF_PASS autolearn=ham autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Fri, Sep 16, 2022 at 8:03 PM Darrick J. Wong wrote: > > On Thu, Sep 15, 2022 at 09:48:18AM -0700, Sarthak Kukreti wrote: > > From: Sarthak Kukreti > > > > Hi, > > > > This patch series is an RFC of a mechanism to pass through provision > > requests on stacked thinly provisioned storage devices/filesystems. > > [Reflowed text] > > > The linux kernel provides several mechanisms to set up thinly > > provisioned block storage abstractions (eg. dm-thin, loop devices over > > sparse files), either directly as block devices or backing storage for > > filesystems. Currently, short of writing data to either the device or > > filesystem, there is no way for users to pre-allocate space for use in > > such storage setups. Consider the following use-cases: > > > > 1) Suspend-to-disk and resume from a dm-thin device: In order to > > ensure that the underlying thinpool metadata is not modified during > > the suspend mechanism, the dm-thin device needs to be fully > > provisioned. > > 2) If a filesystem uses a loop device over a sparse file, fallocate() > > on the filesystem will allocate blocks for files but the underlying > > sparse file will remain intact. > > 3) Another example is virtual machine using a sparse file/dm-thin as a > > storage device; by default, allocations within the VM boundaries will > > not affect the host. > > 4) Several storage standards support mechanisms for thin provisioning > > on real hardware devices. For example: > > a. The NVMe spec 1.0b section 2.1.1 loosely talks about thin > > provisioning: "When the THINP bit in the NSFEAT field of the > > Identify Namespace data structure is set to =E2=80=981=E2=80=99, the = controller ... > > shall track the number of allocated blocks in the Namespace > > Utilization field" > > b. The SCSi Block Commands reference - 4 section references "Thin > > provisioned logical units", > > c. UFS 3.0 spec section 13.3.3 references "Thin provisioning". > > > > In all of the above situations, currently the only way for > > pre-allocating space is to issue writes (or use > > WRITE_ZEROES/WRITE_SAME). However, that does not scale well with > > larger pre-allocation sizes. > > > > This patchset introduces primitives to support block-level > > provisioning (note: the term 'provisioning' is used to prevent > > overloading the term 'allocations/pre-allocations') requests across > > filesystems and block devices. This allows fallocate() and file > > creation requests to reserve space across stacked layers of block > > devices and filesystems. Currently, the patchset covers a prototype on > > the device-mapper targets, loop device and ext4, but the same > > mechanism can be extended to other filesystems/block devices as well > > as extended for use with devices in 4 a-c. > > If you call REQ_OP_PROVISION on an unmapped LBA range of a block device > and then try to read the provisioned blocks, what do you get? Zeroes? > Random stale disk contents? > > I think I saw elsewhere in the thread that any mapped LBAs within the > provisioning range are left alone (i.e. not zeroed) so I'll proceed on > that basis. > For block devices, I'd say it's definitely possible to get stale data, depe= nding on the implementation of the allocation layer; for example, with dm-thinpoo= l, the default setting via using LVM2 tools is to zero out blocks on allocatio= n. But that's configurable and can be turned off to improve performance. Similarly, for actual devices that end up supporting thin provisioning, unl= ess the specification absolutely mandates that an LBA contains zeroes post allocation, some implementations will definitely miss out on that (probably similar to the semantics of discard_zeroes_data today). I'm operating under the assumption that it's possible to get stale data from LBAs allocated usi= ng provision requests at the block layer and trying to see if we can create a safe default operating model from that. > > Patch 1 introduces REQ_OP_PROVISION as a new request type. The > > provision request acts like the inverse of a discard request; instead > > of notifying lower layers that the block range will no longer be used, > > provision acts as a request to lower layers to provision disk space > > for the given block range. Real hardware storage devices will > > currently disable the provisioing capability but for the standards > > listed in 4a.-c., REQ_OP_PROVISION can be overloaded for use as the > > provisioing primitive for future devices. > > > > Patch 2 implements REQ_OP_PROVISION handling for some of the > > device-mapper targets. This additionally adds support for > > pre-allocating space for thinly provisioned logical volumes via > > fallocate() > > > > Patch 3 implements the handling for virtio-blk. > > > > Patch 4 introduces an fallocate() mode (FALLOC_FL_PROVISION) that > > sends a provision request to the underlying block device (and beyond). > > This acts as the primary mechanism for file-level provisioing. > > Personally, I think it's well within the definition of fallocate mode=3D= =3D0 > (aka preallocate) for XFS to call REQ_OP_PROVISION on the blocks that it > preallocates? XFS always sets the unwritten flag on the file mapping, > so it doesn't matter if the device provisions space without zeroing the > contents. > > That said, if devices are really allowed to expose stale disk blocks > then for blkdev fallocate I think you could get away with reusin > FALLOC_FL_NO_HIDE_STALE instead of introducing a new fallocate flag. > For filesystems, I think it's reasonable to support the mode if and only if the filesystem can guarantee that unwritten extents return zero. For instan= ce, in the current ext4 implementation, the provisioned extents are still marke= d as unwritten, which means a read from the file would still show all zeroes (wh= ich I think differs from the original FALLOC_FL_NO_HIDE implementation). That might be one more reason to keep the mode separate from the regular modes though; to drive home the point that it is only acceptable under the above conditions and that there's more to it than just adding blkdev_issue_provision(..) at the end of fs_fallocate(). Best Sarthak > > Patch 5 wires up the loop device handling of REQ_OP_PROVISION. > > > > Patches 6-8 cover a prototype implementation for ext4, which includes > > wiring up the fallocate() implementation, introducing a filesystem > > level option (called 'provision') to control the default allocation > > behaviour and finally a file level override to retain current > > handling, even on filesystems mounted with 'provision' > > Hmm, I'll have a look. > > > Testing: > > -------- > > - A backport of this patch series was tested on ChromiumOS using a > > 5.10 kernel. > > - File on ext4 on a thin logical volume: > > fallocate(FALLOC_FL_PROVISION) : 4.6s, dd if=3D/dev/zero of=3D...: 6 mi= ns. > > > > TODOs: > > ------ > > 1) The stacked block devices (dm-*, loop etc.) currently > > unconditionally pass through provision requests. Add support for > > provision, similar to how discard handling is set up (with options to > > disable, passdown or passthrough requests). > > 2) Blktests and Xfstests for validating provisioning. > > Yes.... > > --D > > > -- > > dm-devel mailing list > > dm-devel@redhat.com > > https://listman.redhat.com/mailman/listinfo/dm-devel