Received: by 2002:ad5:4acb:0:0:0:0:0 with SMTP id n11csp1904imw; Wed, 13 Jul 2022 17:34:56 -0700 (PDT) X-Google-Smtp-Source: AGRyM1sJQxd3419rt9VNy+4NySk+X47405RG3LAsnVly0dGgeZFVS/ZjKQYzCiKkoz6R/gphYFr5 X-Received: by 2002:a63:b443:0:b0:40c:fbf9:8366 with SMTP id n3-20020a63b443000000b0040cfbf98366mr5116968pgu.308.1657758896054; Wed, 13 Jul 2022 17:34:56 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1657758896; cv=none; d=google.com; s=arc-20160816; b=xBdJ9OY+zwD2UC6amd4aN4vtMN4fUCPjejboaxwtV1z+Kjq4VUTQjJLXzJBE86h1hy nk2oRgH5MtkGE2F/WxKK1hqACvXlzp94Nua0jUdPNqeXL3kWm21o0f1/me1uqUIMuPIL LeZImaze7qOUIggz9oghSNQnpYoJkgNoWo1CV59q+ESqUpGSyrs9bpphHSY+uXv1v/wp hLnkcut5QBvSPPNlfeh5DDNB09FnmsTjHGrRqjlmzjCfISeUrW3h+ddDQPkfp7o/cqHz DPLeqh9kZlg8+PreN8QrrcaebRZ4uZw+G91FQ41FZOGw7+xhnOi3IKTZGhH04vKgd3JN Twlg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:cc:to:subject:message-id:date:from:in-reply-to :references:mime-version:dkim-signature; bh=cRXGKXu4ProwL4B2fSS8bEMaPpUQ1vx76Iq2sHTQhlg=; b=KMKzGEzb2CcwKZFnneq/ZkY2RRtQYSgkpNEq2UBKnT/o2UxTK3+YQH0QYOEvwN0hSn rbinAanD2JsGq66RHxF8LK7KHgfP6NJP2VdNP9Lev7n6IA+Q5kllVp0UBznnYeE+wQm8 qw/iAG2AwMchdrOBKSYlUfjtV0LJ1uIW5slnEOdm50kKgzMKxoHeemH0Ut3HaoJWc0DQ QVv8lxvXvFvx6EmJ94AyvvKMwd7j5lRXWGnXHXiGpbRZVM1C9beHM4sxfxd/OAax2QkJ J7ozQesfueIw+2uiKJdMbkYeuX0dNAO4FN03ZnpvLOck4KiLfsweDQPwbSe69tqF0Yxv 4qOw== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@gmail.com header.s=20210112 header.b=oEk5EWsu; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=gmail.com Return-Path: Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id r30-20020a635d1e000000b00407e25d4528si19459543pgb.5.2022.07.13.17.34.24; Wed, 13 Jul 2022 17:34:56 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; dkim=pass header.i=@gmail.com header.s=20210112 header.b=oEk5EWsu; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=gmail.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S229842AbiGMXzu (ORCPT + 99 others); Wed, 13 Jul 2022 19:55:50 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:33158 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S229703AbiGMXzt (ORCPT ); Wed, 13 Jul 2022 19:55:49 -0400 Received: from mail-wm1-x330.google.com (mail-wm1-x330.google.com [IPv6:2a00:1450:4864:20::330]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 39E5B25C68 for ; Wed, 13 Jul 2022 16:55:48 -0700 (PDT) Received: by mail-wm1-x330.google.com with SMTP id c131-20020a1c3589000000b003a2cc290135so2140671wma.2 for ; Wed, 13 Jul 2022 16:55:48 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20210112; h=mime-version:references:in-reply-to:from:date:message-id:subject:to :cc; bh=cRXGKXu4ProwL4B2fSS8bEMaPpUQ1vx76Iq2sHTQhlg=; b=oEk5EWsuFg2b0cV8tBTgj0Z3Z9EJ66H7WF3QmtBNL1ueucc3ebqEOwW0eDtnNL101O BMTEpv8axsesFWJQ2a30FNcmxB3NXpw49yIGa6T4A7zfPpr+2M5nYfnlIRsGL+kWGCmX 7nacKgRaxYmfEyhV0UTK5sZNYIHHF1NfN09XCvTg4KtBaiKFonzIn2aMazKYUIHo7bPg TQB99EXGxOfH02ND0H8zfpgo6EwpPO8yWHuGWLOhq9BqV3rIxZG2hlB3xNFW63abg+7/ IjHhkrqyVBUcthVzagBlltbDq4/GnB4QgwYkTpoYuIT65ms/GMOhc0waNXHkeyXp5C84 AACA== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20210112; h=x-gm-message-state:mime-version:references:in-reply-to:from:date :message-id:subject:to:cc; bh=cRXGKXu4ProwL4B2fSS8bEMaPpUQ1vx76Iq2sHTQhlg=; b=O+0MaAb8pb4i39R4B3wE0pFe94MqCWb6nY5cJfPVzS5d44BYctiZqy1xj7brtyQuHK cxsBJCA53VO6Aermf6cugwMaIg+P3UjsB1qt4KafWCSvYh/gQxhzPrbeeeBwb8DXUcRI lfUwVwkUfVQVWNbK3ZGwn6EN06XRJpthCQ86NT3EZb7tRN2A8duSDIipYFtHgWF5rOaV Zj2V7GNrZyFRm25TqRJRPtDt8nD/X4Ov6mX9KIygHH9x8ptDRXTdYxmsEktqsDEAEunA DB7T5jF95CioLdIsmWtLN3cbcPH41dMoNrIR/1ctDCcKTevdBWoNe71rOfcksOia0Yau y6Sw== X-Gm-Message-State: AJIora+rjILqaZ7HqPXQra/J4W+W8FStoiVUAFm7hkUyHF7LsUN73m9y 8X7Emx61QaUc4C5LOOY7K+mbfZEJBePBJXM4PI/PaOow1vHbnw== X-Received: by 2002:a05:600c:1c88:b0:3a3:1ae:5784 with SMTP id k8-20020a05600c1c8800b003a301ae5784mr116535wms.161.1657756546588; Wed, 13 Jul 2022 16:55:46 -0700 (PDT) MIME-Version: 1.0 References: <20220713125123.8701-1-schspa@gmail.com> In-Reply-To: <20220713125123.8701-1-schspa@gmail.com> From: Lai Jiangshan Date: Thu, 14 Jul 2022 07:55:35 +0800 Message-ID: Subject: Re: [PATCH v2] workqueue: Use active mask for new worker when pool is DISASSOCIATED To: Schspa Shi Cc: Tejun Heo , LKML , zhaohui.shi@horizon.ai Content-Type: text/plain; charset="UTF-8" X-Spam-Status: No, score=-2.1 required=5.0 tests=BAYES_00,DKIM_SIGNED, DKIM_VALID,DKIM_VALID_AU,DKIM_VALID_EF,FREEMAIL_FROM, RCVD_IN_DNSWL_NONE,SPF_HELO_NONE,SPF_PASS,T_SCC_BODY_TEXT_LINE autolearn=ham autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Wed, Jul 13, 2022 at 8:51 PM Schspa Shi wrote: > > When CPU-[un]hotplugs, all workers will be bound to active CPU via > unbind_workers(). > > But the unbound worker still has a chance to create a new worker, which > has bound the newly created task to pool->attrs->cpumask. But the CPU has > been unplugged. > > Please refer to the following scenarios. > > CPU0 CPU1 > ------------------------------------------------------------------ > sched_cpu_deactivate(cpu_active_mask clear) > workqueue_offline_cpu(work pool POOL_DISASSOCIATED) > -- all worker will migrate to another cpu -- > worker_thread > -- will create new worker if > pool->worklist is not empty > create_worker() > -- new kworker will bound to CPU0 > (pool->attrs->cpumask will be mask of CPU0). > kworker/0:x will running on rq > > sched_cpu_dying > if (rq->nr_running != 1 || rq_has_pinned_tasks(rq)) > WARN(true, "Dying CPU not properly vacated!"); > ---------OOPS------------- > > And the crash log as fellowing. > > The crash log is as follows: > [ 1622.829074] ------------[ cut here ]------------ > [ 1622.829081] Dying CPU not properly vacated! > [ 1622.829091] WARNING: CPU: 3 PID: 31 at kernel/sched/core.c:7756 sched_cpu_dying+0x74/0x204 > [ 1622.829374] CPU: 3 PID: 31 Comm: migration/3 Tainted: P O 5.10.59-rt52 #2 > [ 1622.829386] Stopper: multi_cpu_stop+0x0/0x160 <- 0x0 > [ 1622.829400] pstate: 60c00009 (nZCv daif +PAN +UAO -TCO BTYPE=--) > [ 1622.829408] pc : sched_cpu_dying+0x74/0x204 > [ 1622.829415] lr : sched_cpu_dying+0x74/0x204 > [ 1622.829421] sp : ffff800012933c70 > [ 1622.829424] pmr_save: 00000060 > [ 1622.829426] x29: ffff800012933c70 x28: 0000000000000000 > [ 1622.829435] x27: 0000000000000000 x26: 0000000000000001 > [ 1622.829444] x25: 0000000000000000 x24: ffff800018353c2c > [ 1622.829452] x23: 0000000000000003 x22: 0000000000000003 > [ 1622.829460] x21: 0000000000000059 x20: 0000000000000000 > [ 1622.829468] x19: ffff00027ee68a80 x18: 0000000000000000 > [ 1622.829477] x17: 0000000000000000 x16: 0000000000000000 > [ 1622.829485] x15: ffffffffffffffff x14: ffff80001169ae30 > [ 1622.829493] x13: ffffffffffc38c07 x12: ffffffffffffffff > [ 1622.829501] x11: ffffffffffe00000 x10: ffff80001169ae58 > [ 1622.829510] x9 : 000000000000001e x8 : ffff80001169ae30 > [ 1622.829518] x7 : ffff800012933ab0 x6 : 00000000ffff0e20 > [ 1622.829526] x5 : ffff00027ee62988 x4 : 00000000ffff0e20 > [ 1622.829535] x3 : ffff800011e30180 x2 : 0000000100000002 > [ 1622.829543] x1 : 0000000000000000 x0 : 0000000000000000 > [ 1622.829552] Call trace: > [ 1622.829555] sched_cpu_dying+0x74/0x204 > [ 1622.829562] cpuhp_invoke_callback+0xc0/0x1b0 > [ 1622.829571] take_cpu_down+0xbc/0xd4 > [ 1622.829577] multi_cpu_stop+0x138/0x160 > [ 1622.829584] cpu_stopper_thread+0x9c/0x118 > [ 1622.829591] smpboot_thread_fn+0x1e8/0x1ec > [ 1622.829598] kthread+0x114/0x124 > [ 1622.829606] ret_from_fork+0x10/0x30 > [ 1622.829615] ---[ end trace 0000000000000002 ]--- > [ 1623.830273] CPU3 enqueued tasks (2 total): > [ 1623.830291] pid: 31, name: migration/3 > [ 1623.830440] pid: 25654, name: kworker/3:0 > [ 1623.830444] task:kworker/3:0 state:R running task stack: 0 pid:25654 ppid: 2 flags:0x00000028 > [ 1623.830458] Call trace: > [ 1623.830460] __switch_to+0x164/0x17c > [ 1623.830472] __schedule+0x4cc/0x5c0 > [ 1623.830483] schedule+0x7c/0xcc > [ 1623.830491] schedule_preempt_disabled+0x14/0x24 > [ 1623.830500] kthread+0xd8/0x124 > [ 1623.830509] ret_from_fork+0x10/0x30 > > In the crash log, the error enqueued kworker(pid: 25654, name: kworker/3:0) > is a newly created thread, and have bind to a unpluged CPU 3. > > Can crash dump can also verify this. > crash> task -R nr_cpus_allowed,cpus_mask 25654 > PID: 25654 TASK: ffff000181ff0000 CPU: 3 COMMAND: "kworker/3:0" > nr_cpus_allowed = 1, > cpus_mask = { > bits = {8, 0} > }, > > crash> struct worker_pool.cpu,nr_workers,attrs 0xffff00027ee68380 > cpu = 3, > nr_workers = 3, > attrs = 0xffff000180004480, > crash> struct workqueue_attrs 0xffff000180004480 > struct workqueue_attrs { > nice = 0, > cpumask = {{ > bits = {8, 0} > }}, > no_numa = false > } > > The stack trace of the bad running task was dumped via the following patch: > Link: https://lore.kernel.org/all/20220519161125.41144-1-schspa@gmail.com/ > And I think this debug patch needs to be added to the mainline, > it can help us to debug this kind of problem > > To fix it, we can use cpu_active_mask when work pool is DISASSOCIATED. > > Signed-off-by: Schspa Shi > > -- > > Changelog: > v1 -> v2: > - Move worker task bind to worker_attach_to_pool, remove extra > wq_pool_attach_mutex added. > - Add a timing diagram to make this question clearer. > --- > kernel/workqueue.c | 9 ++++----- > 1 file changed, 4 insertions(+), 5 deletions(-) > > diff --git a/kernel/workqueue.c b/kernel/workqueue.c > index 1ea50f6be843..3f6e6d884793 100644 > --- a/kernel/workqueue.c > +++ b/kernel/workqueue.c > @@ -1867,13 +1867,13 @@ static void worker_attach_to_pool(struct worker *worker, > * stable across this function. See the comments above the flag > * definition for details. > */ > - if (pool->flags & POOL_DISASSOCIATED) > + if (pool->flags & POOL_DISASSOCIATED) { > worker->flags |= WORKER_UNBOUND; > - else > + set_cpus_allowed_ptr(worker->task, cpu_active_mask); > + } else { > kthread_set_per_cpu(worker->task, pool->cpu); > - > - if (worker->rescue_wq) > set_cpus_allowed_ptr(worker->task, pool->attrs->cpumask); > + } For unbound pools (get_unbound_pool(), which also has POOL_DISASSOCIATED), pool->attrs->cpumask should be used if pool->attrs->cpumask has active cpu. if (cpumask_intersects(pool->attrs->cpumask, cpu_active_mask)) set_cpus_allowed_ptr(worker->task, pool->attrs->cpumask); else set_cpus_allowed_ptr(worker->task, wq_unbound_cpumask); And put it above the "if (pool->flags & POOL_DISASSOCIATED)" code block. > > list_add_tail(&worker->node, &pool->workers); > worker->pool = pool; > @@ -1952,7 +1952,6 @@ static struct worker *create_worker(struct worker_pool *pool) > goto fail; > > set_user_nice(worker->task, pool->attrs->nice); > - kthread_bind_mask(worker->task, pool->attrs->cpumask); worker->flags |= PF_NO_SETAFFINITY; PF_NO_SETAFFINITY needs to be set for workers. > > /* successful, attach the worker to the pool */ > worker_attach_to_pool(worker, pool); > -- > 2.29.0 >