Received: by 2002:a05:6a10:af89:0:0:0:0 with SMTP id iu9csp4623522pxb; Tue, 25 Jan 2022 14:48:37 -0800 (PST) X-Google-Smtp-Source: ABdhPJwK91+p5mZ1RF02W91islekAZvqsN8QlELDVl2arRsXpneXhVMmo6hYHM3AuKevs0essy3+ X-Received: by 2002:a17:903:1108:b0:149:9c03:23ce with SMTP id n8-20020a170903110800b001499c0323cemr20328899plh.140.1643150916860; Tue, 25 Jan 2022 14:48:36 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1643150916; cv=none; d=google.com; s=arc-20160816; b=x6Z1zStJcaDfQfgIWIuxT8tFKPbTYwth5CF7Ux9ivpd7UCZHHPcdJR/jEi1vjjX1Si V+TZH5XmWXZKj40uuBIsmJXmrFQavO7JNlAVbGC0FAuuXp25sZDTWOETYpYnhLVRCDd4 l78OWsCbsm5xovZdrfTBMCJijEPLhNagkSRC75jNHLAXEg+gCIdpX1KzPr5xQ2ZlGuHW bBncykpeZUAxTkFA+J4t3uEzZu2otJIV92crmM6woUWO/oIo/SLHbVx/XABd25Re7SLB M0bKajot8C4FVXEdLWQa98iJjoIezNOpcIpVXoSuRri7BKiW2QSoaPFbP5oqVlEJl9fu U7yg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:cc:to:subject:message-id:date:from:in-reply-to :references:mime-version:dkim-signature; bh=FmCu5q/9cyd6jLwDAKHOnptOvmHmNnDiPHpjTLMqXXc=; b=owwnzz6rZazIkYS8XLh9XGjR7/vrYQexDRP7J3itLJ2/fNQsLEbBsMZUNqfHr0Ek2u LqotE9iK9+naFts9irZWQmc+WNkkoCkfn1z+py0rmRzFK61a28zIlTDxpn3jx0GtgE4t RU5DZ4u81B6QdT3jxEaPxOzlDazm+HkqfdDnX8JBq6082/0bPIPnJaCUajlE7QWF9KNg OTKyz7nxGGHhdV7dozXExAvo5/74AR/fD4eKfDhX58OEBDI6kq5WlSH1O/pcfzeTp9UE dxqqp25yExQTJohX2nyaZRmiqmmvcc+kPyiry85vfEwe2R+ZczmzezNfOuu3IrKeRPrj ua+g== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@dneg.com header.s=google header.b=FYMmG0bW; spf=pass (google.com: domain of linux-nfs-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-nfs-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=dneg.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id y2si92623pfi.316.2022.01.25.14.48.24; Tue, 25 Jan 2022 14:48:36 -0800 (PST) Received-SPF: pass (google.com: domain of linux-nfs-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=pass header.i=@dneg.com header.s=google header.b=FYMmG0bW; spf=pass (google.com: domain of linux-nfs-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-nfs-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=dneg.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1345552AbiAYP1D (ORCPT + 99 others); Tue, 25 Jan 2022 10:27:03 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:50778 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1349626AbiAYPYq (ORCPT ); Tue, 25 Jan 2022 10:24:46 -0500 Received: from mail-ej1-x635.google.com (mail-ej1-x635.google.com [IPv6:2a00:1450:4864:20::635]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id A4EA5C06173D for ; Tue, 25 Jan 2022 07:24:42 -0800 (PST) Received: by mail-ej1-x635.google.com with SMTP id ah7so30930485ejc.4 for ; Tue, 25 Jan 2022 07:24:42 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=dneg.com; s=google; h=mime-version:references:in-reply-to:from:date:message-id:subject:to :cc; bh=FmCu5q/9cyd6jLwDAKHOnptOvmHmNnDiPHpjTLMqXXc=; b=FYMmG0bW7AcVJozcIb4UL5jaq8gx/U3N8bfYx6LTr8RR0tl5lC7CI/OgS+3IA4EHwS V/5uM0CkfAtvMvBW0hbzxYMfl0HnSZQ3/o7F65e/NJzgkO6ffIWPdtCuKAtBDp4vR9RP GaE93WvFVGl3kyFEf0jbkwOKBp6YyQOMA7ta5GO3Q/UNQt7gLYhQY+sITf0fNXiD7VTS JFJCHKkp9mvtq7lfMvLRBiLgAR+qnRo5tQEFNtl1RC/wgVtzdu50rdBVstykoaXzpLfo 88VTt+lfnBAGE5/a4bBhINihqjbz+X1jUrA78EmCPlYttwEe/6yUCxs85gbYf2BOHtq9 qm/w== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20210112; h=x-gm-message-state:mime-version:references:in-reply-to:from:date :message-id:subject:to:cc; bh=FmCu5q/9cyd6jLwDAKHOnptOvmHmNnDiPHpjTLMqXXc=; b=kZ8pM0SGamRmXToqfv/2Q5cMDLP01OMMuATilmzGCXKAKELBs5KrVPPUpCrVpWm3cA 3dMk+KKkZdg6T516KgYvELPpvG/BycOIs8z+X2X2/EKF6JMjymeSrQKDTDyfSJ67paMv xbNv4yY+RIusvyUY5rAckt/JT6hUtJhRm54mQSH7ftIK8l1j0wtp2G/ki0jJvwTHjIzt HlI2+/JkFUSZPa1phkwhJsKuQP02S4A6XUaY8dinYcwnH038b4cUWXpsutX1ktsp6bv0 yTvfCx/ZEkJkC8oEOyOKTEDrEhnFSH7XuMYigEFCQDy808YBI3klEHCjE/ddm56Mwbp7 gOFA== X-Gm-Message-State: AOAM533K2hzbft3sQzGhwGp/Xcn5olbfXFV5Wq1mYerb4rKziYQmRGYU EpjvPF7mnUA/U+7vc2o8KJFZAHjo8Uwgv3OafBSTm8C3vEJ1Kg== X-Received: by 2002:a17:906:2ec8:: with SMTP id s8mr16699667eji.746.1643124281229; Tue, 25 Jan 2022 07:24:41 -0800 (PST) MIME-Version: 1.0 References: <20220124193759.GA4975@fieldses.org> <20220124205045.GB4975@fieldses.org> <20220125135959.GA15537@fieldses.org> In-Reply-To: <20220125135959.GA15537@fieldses.org> From: Daire Byrne Date: Tue, 25 Jan 2022 15:24:05 +0000 Message-ID: Subject: Re: parallel file create rates (+high latency) To: "J. Bruce Fields" Cc: linux-nfs Content-Type: text/plain; charset="UTF-8" Precedence: bulk List-ID: X-Mailing-List: linux-nfs@vger.kernel.org On Tue, 25 Jan 2022 at 14:00, J. Bruce Fields wrote: > > On Tue, Jan 25, 2022 at 12:52:46PM +0000, Daire Byrne wrote: > > Yea, it does seem like the server is the ultimate arbitrar and the > > fact that multiple clients can achieve much higher rates of > > parallelism does suggest that the VFS locking per client is somewhat > > redundant and limiting (in this super niche case). > > It doesn't seem *so* weird to have a server with fast storage a long > round-trip time away, in which case the client-side operation could take > several orders of magnitude longer than the server. Yea, I'm fine with the speed of light constraints for a single process far away. But the best way to achieve aggregate performance in such environments is to have multiple parallel streams in flight at once (preferably bulk transfers). Because I am writing through a single re-export server, I just so happen to be killing any parallelism for single directory files creates even though it works reasonably well for opens, reads, writes and stat (and anything cacheable) which all retain a certain amount of useful parallelism over a high latency network (nconnect helps too). Each of our batch jobs all read 95% of the same files each time and they all tend to run within the same short (hour) time periods so highly cacheable. > Though even if the client locking wasn't a factor, you might still have > to do some work to take advantage of that. (E.g. if your workload is > just a single "untar"--it still waits for one create before doing the > next one). Yep. Again I'm okay with each client of a re-export server doing 3 creates per second, my problem is that all N instances together do 3 creates per second aggregate (in a single directory). But I guess for this kind of workload, win some lose some. I just need to figure out if I can engineer it to be less of a loser... Of course, Hammerspace have a different approach to this kind of problem with their global namespace and replicated MDT servers. And that is probably a much more sensible way of going about this kind of thing. Daire