Received: by 2002:a05:6358:9144:b0:117:f937:c515 with SMTP id r4csp1656881rwr; Fri, 5 May 2023 18:44:15 -0700 (PDT) X-Google-Smtp-Source: ACHHUZ6PEsE9EXLV4v1K9U1BVejqG8/8kItlYu7gNmV1DD+4BZ0yaRXxE/exHU+5XC3JWrs/5JMX X-Received: by 2002:a05:6a20:244d:b0:f3:5cf7:581b with SMTP id t13-20020a056a20244d00b000f35cf7581bmr4402992pzc.28.1683337454775; Fri, 05 May 2023 18:44:14 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1683337454; cv=none; d=google.com; s=arc-20160816; b=jNI0/Ml9ghuxyVhuEJyLzZ4NOhK7neEy5dZ63JepjASfFGLGD2udtRiZRRZdo8WixB m1lSEUbD3m+solMEJ3wwppMfwczDiZgq9yOEoPMmVt3S7gv2lSvikLEXfLQWTKYJWFC9 wathRwbFuwWbTlBwO/qWTa5l0vCASVTiAFxjUOSdMFyZDZLnBayxXLiRWGb5cZVT1neK EkH6Y0R66ArqKg7MjdTU08LrX2cHV9oJZtiHIpznLpnx1Ob+4gbqZG7XfrJonMyCo2It OOdQhdZhFLeToWdIONvWqfUH7VUsVAHHyKTW/3U8aP4yvYTkoxQP1inO+8eOvyVSuyGj AjPg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:mime-version :message-id:references:in-reply-to:cc:subject:to:from:date; bh=V5bU4nQmy0K+JSrXqTFdWBQapv8pulWTNDsHXeSgfAM=; b=pNL7v8oKOJwg28TtNTv97qJJlJfJq6ulKbmLyV9In5uBILqJbKPX+DcVCCbnH5I5wk mhzqOhL+xHdUeoLsCIs18Tj1gnP4MVupckWlioJmKNiBYJsM0ye2RhIQC0XNVX6pGhaV 6QCCz96vc/gvtAGxcenstHIUgwwbshYXLoLliWc6JxR7/sbTHSIzk1MeA4Ypl3+52Sg+ 2NFyLlxlbilj41mtPc272wJ1c7P6Mk6bgqFeYXVMDfLzUTE9GPQQk+HETnlRJbngH3ly Ne40RPoGvJO3JU88CwgK6JcB0ksH8esH/UimmcfnNwASmIFsxn0hsoljCm0LwO6uJ092 DoCQ== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id v25-20020a63b959000000b0051b83af33desi3250034pgo.246.2023.05.05.18.44.00; Fri, 05 May 2023 18:44:14 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S231986AbjEFBkX (ORCPT + 99 others); Fri, 5 May 2023 21:40:23 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:59984 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S230196AbjEFBkW (ORCPT ); Fri, 5 May 2023 21:40:22 -0400 Received: from out28-55.mail.aliyun.com (out28-55.mail.aliyun.com [115.124.28.55]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id E8AD944AD; Fri, 5 May 2023 18:40:19 -0700 (PDT) X-Alimail-AntiSpam: AC=CONTINUE;BC=0.04436267|-1;CH=green;DM=|CONTINUE|false|;DS=CONTINUE|ham_regular_dialog|0.00445034-0.000320935-0.995229;FP=0|0|0|0|0|-1|-1|-1;HT=ay29a033018047187;MF=wangyugui@e16-tech.com;NM=1;PH=DS;RN=8;RT=8;SR=0;TI=SMTPD_---.SZWvAUP_1683337213; Received: from 192.168.2.112(mailfrom:wangyugui@e16-tech.com fp:SMTPD_---.SZWvAUP_1683337213) by smtp.aliyun-inc.com; Sat, 06 May 2023 09:40:14 +0800 Date: Sat, 06 May 2023 09:40:14 +0800 From: Wang Yugui To: Tejun Heo Subject: Re: [PATCH v2 16/22] btrfs: Use alloc_ordered_workqueue() to create ordered workqueues Cc: jiangshanlai@gmail.com, linux-kernel@vger.kernel.org, kernel-team@meta.com, Chris Mason , Josef Bacik , David Sterba , linux-btrfs@vger.kernel.org In-Reply-To: References: <20230430124006.49D2.409509F4@e16-tech.com> Message-Id: <20230506094013.29A6.409509F4@e16-tech.com> MIME-Version: 1.0 Content-Type: text/plain; charset="US-ASCII" Content-Transfer-Encoding: 7bit X-Mailer: Becky! ver. 2.81.04 [en] X-Spam-Status: No, score=-1.9 required=5.0 tests=BAYES_00,SPF_HELO_NONE, SPF_PASS,T_SCC_BODY_TEXT_LINE,UNPARSEABLE_RELAY autolearn=ham autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Hi, > BACKGROUND > ========== > > When multiple work items are queued to a workqueue, their execution order > doesn't match the queueing order. They may get executed in any order and > simultaneously. When fully serialized execution - one by one in the queueing > order - is needed, an ordered workqueue should be used which can be created > with alloc_ordered_workqueue(). > > However, alloc_ordered_workqueue() was a later addition. Before it, an > ordered workqueue could be obtained by creating an UNBOUND workqueue with > @max_active==1. This originally was an implementation side-effect which was > broken by 4c16bd327c74 ("workqueue: restore WQ_UNBOUND/max_active==1 to be > ordered"). Because there were users that depended on the ordered execution, > 5c0338c68706 ("workqueue: restore WQ_UNBOUND/max_active==1 to be ordered") > made workqueue allocation path to implicitly promote UNBOUND workqueues w/ > @max_active==1 to ordered workqueues. > > While this has worked okay, overloading the UNBOUND allocation interface > this way creates other issues. It's difficult to tell whether a given > workqueue actually needs to be ordered and users that legitimately want a > min concurrency level wq unexpectedly gets an ordered one instead. With > planned UNBOUND workqueue updates to improve execution locality and more > prevalence of chiplet designs which can benefit from such improvements, this > isn't a state we wanna be in forever. > > This patch series audits all callsites that create an UNBOUND workqueue w/ > @max_active==1 and converts them to alloc_ordered_workqueue() as necessary. > > WHAT TO LOOK FOR > ================ > > The conversions are from > > alloc_workqueue(WQ_UNBOUND | flags, 1, args..) > > to > > alloc_ordered_workqueue(flags, args...) > > which don't cause any functional changes. If you know that fully ordered > execution is not ncessary, please let me know. I'll drop the conversion and > instead add a comment noting the fact to reduce confusion while conversion > is in progress. > > If you aren't fully sure, it's completely fine to let the conversion > through. The behavior will stay exactly the same and we can always > reconsider later. > > As there are follow-up workqueue core changes, I'd really appreciate if the > patch can be routed through the workqueue tree w/ your acks. Thanks. > > v2: btrfs_alloc_workqueue() updated too as suggested by Wang. > > Signed-off-by: Tejun Heo > Cc: Wang Yugui > Cc: Chris Mason > Cc: Josef Bacik > Cc: David Sterba > Cc: linux-btrfs@vger.kernel.org > --- > Hello, > > Wang, yeah, that's a helper that can't tell whether the caller wants an > ordered wq or not, so it needs to be updated too. How does this look? > > Thanks. > > fs/btrfs/async-thread.c | 7 +++++-- > fs/btrfs/disk-io.c | 2 +- > fs/btrfs/scrub.c | 6 ++++-- > 3 files changed, 10 insertions(+), 5 deletions(-) > > --- a/fs/btrfs/async-thread.c > +++ b/fs/btrfs/async-thread.c > @@ -99,8 +99,11 @@ struct btrfs_workqueue *btrfs_alloc_work > ret->thresh = thresh; > } > > - ret->normal_wq = alloc_workqueue("btrfs-%s", flags, ret->current_active, > - name); > + if (ret->current_active == 1) > + ret->normal_wq = alloc_ordered_workqueue("btrfs-%s", flags, name); > + else > + ret->normal_wq = alloc_workqueue("btrfs-%s", flags, > + ret->current_active, name); > if (!ret->normal_wq) { > kfree(ret); > return NULL; by test, I noticed some warning caused by void workqueue_set_max_active(struct workqueue_struct *wq, int max_active) if (WARN_ON(wq->flags & __WQ_ORDERED_EXPLICIT)) return; so I tested again with the flowing fix diff --git a/fs/btrfs/async-thread.c b/fs/btrfs/async-thread.c index 43c8995..e4b68e9 100644 --- a/fs/btrfs/async-thread.c +++ b/fs/btrfs/async-thread.c @@ -99,8 +99,11 @@ struct btrfs_workqueue *btrfs_alloc_workqueue(struct btrfs_fs_info *fs_info, ret->thresh = thresh; } - ret->normal_wq = alloc_workqueue("btrfs-%s", flags, ret->current_active, - name); + if(limit_active == 1) + ret->normal_wq = alloc_ordered_workqueue("btrfs-%s", flags, name); + else + ret->normal_wq = alloc_workqueue("btrfs-%s", flags, + ret->current_active, name); if (!ret->normal_wq) { kfree(ret); return NULL; @@ -139,7 +139,7 @@ static inline void thresh_exec_hook(struct btrfs_workqueue *wq) long pending; int need_change = 0; - if (wq->thresh == NO_THRESHOLD) + if (wq->thresh == NO_THRESHOLD || wq->limit_active == 1) return; atomic_dec(&wq->pending); we need 'limit_active' at 2nd postition, so I used 'limit_active' and 1st postition too. Best Regards Wang Yugui (wangyugui@e16-tech.com) 2023/05/06 > --- a/fs/btrfs/disk-io.c > +++ b/fs/btrfs/disk-io.c > @@ -2218,7 +2218,7 @@ static int btrfs_init_workqueues(struct > fs_info->qgroup_rescan_workers = > btrfs_alloc_workqueue(fs_info, "qgroup-rescan", flags, 1, 0); > fs_info->discard_ctl.discard_workers = > - alloc_workqueue("btrfs_discard", WQ_UNBOUND | WQ_FREEZABLE, 1); > + alloc_ordered_workqueue("btrfs_discard", WQ_FREEZABLE); > > if (!(fs_info->workers && fs_info->hipri_workers && > fs_info->delalloc_workers && fs_info->flush_workers && > --- a/fs/btrfs/scrub.c > +++ b/fs/btrfs/scrub.c > @@ -4245,8 +4245,10 @@ static noinline_for_stack int scrub_work > if (refcount_inc_not_zero(&fs_info->scrub_workers_refcnt)) > return 0; > > - scrub_workers = alloc_workqueue("btrfs-scrub", flags, > - is_dev_replace ? 1 : max_active); > + if (is_dev_replace) > + scrub_workers = alloc_ordered_workqueue("btrfs-scrub", flags); > + else > + scrub_workers = alloc_workqueue("btrfs-scrub", flags, max_active); > if (!scrub_workers) > goto fail_scrub_workers; >