Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by smtp.lore.kernel.org (Postfix) with ESMTP id 2D6B4C61DA4 for ; Fri, 24 Feb 2023 02:37:15 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S229794AbjBXChN (ORCPT ); Thu, 23 Feb 2023 21:37:13 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:51798 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S229644AbjBXChL (ORCPT ); Thu, 23 Feb 2023 21:37:11 -0500 Received: from mail-lf1-x12a.google.com (mail-lf1-x12a.google.com [IPv6:2a00:1450:4864:20::12a]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 68C5B5EEDF for ; Thu, 23 Feb 2023 18:37:07 -0800 (PST) Received: by mail-lf1-x12a.google.com with SMTP id i9so16038231lfc.6 for ; Thu, 23 Feb 2023 18:37:07 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=joelfernandes.org; s=google; h=cc:to:subject:message-id:date:from:in-reply-to:references :mime-version:from:to:cc:subject:date:message-id:reply-to; bh=9eCkpzltIlGRTF6sYL0yBS3gJolIQPZMfLq+lqf6a/c=; b=XOVv3jtQtWJzCrX/HzfJ1iu/sanBxru7gFmbG938gEue8PMGJGvrmcrOKGfIAjzwlp ShY5wb6VYf28OfbqwQwdXownt4kVAYKJGaHZGAQgTUdh3UETxXimgnRh2iD5AI1Jj/RA mCluCpxFmkGJtXQiGCT7oGzUThGaxX75paCuk= X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20210112; h=cc:to:subject:message-id:date:from:in-reply-to:references :mime-version:x-gm-message-state:from:to:cc:subject:date:message-id :reply-to; bh=9eCkpzltIlGRTF6sYL0yBS3gJolIQPZMfLq+lqf6a/c=; b=4u1X6hTYSeT9Y581DPow/HXfRsp5VVvSXGrtA6hpC5RWtLq1YTauLEoPyjLK2XFXgy ynHSkr8WZobVr7g6a5cjXFXBfcPvAcIzi8OmKJcUrtcFUothDpSo5C4Go0PtwvmhxmWm cUKHGrWbHuYwD1l5xckhfwGsbaSgt0vfPXuQloqVmewch0n4dxo/GUg/QTia/2vn8qgI 6dUlaxIHyp12Gls9bwbLPpux1M6OgRe+W8T6AO2IJj5AwHgEKvOtTCZ10aBrRKz2oCr3 T5xsJ58X+zOycLWMZ2YN7a5LLWmqMHasjxmzduOrEfDqAiCK6cDb/QuvxY77a73Y41oc k/9g== X-Gm-Message-State: AO0yUKU6kYUu9pgn3Ink8756mkL24kL/Jl9ex/j3d3nkO46wtusPURyP Gk9sH/yklaW8wstsN/jPRzpoaWP1yfkxGsViRbWVKg== X-Google-Smtp-Source: AK7set/949p9CR8H2KeCDKiHocUC07GWv3yNQbyW3H0ZtmQ2vDGp6nIHLWFO+k17RLsMOCG9AAs+C2gRQl8RJtAsbd4= X-Received: by 2002:a05:6512:208:b0:4d5:ca32:9bdf with SMTP id a8-20020a056512020800b004d5ca329bdfmr4479433lfo.11.1677206225585; Thu, 23 Feb 2023 18:37:05 -0800 (PST) MIME-Version: 1.0 References: <20230223063022.2592212-1-qiang1.zhang@intel.com> In-Reply-To: From: Joel Fernandes Date: Thu, 23 Feb 2023 21:36:54 -0500 Message-ID: Subject: Re: [PATCH] rcu-tasks: Directly invoke rcuwait_wake_up() in call_rcu_tasks_generic() To: "Zhang, Qiang1" Cc: "Zhuo, Qiuxu" , "paulmck@kernel.org" , "frederic@kernel.org" , "quic_neeraju@quicinc.com" , "rcu@vger.kernel.org" , "linux-kernel@vger.kernel.org" Content-Type: text/plain; charset="UTF-8" Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Thu, Feb 23, 2023 at 9:35 PM Joel Fernandes wrote: > > On Thu, Feb 23, 2023 at 9:25 PM Joel Fernandes wrote: > > > > On Fri, Feb 24, 2023 at 12:36:05AM +0000, Zhang, Qiang1 wrote: > > > On Thu, Feb 23, 2023 at 08:43:05AM +0000, Zhang, Qiang1 wrote: > > > > > From: Zqiang > > > > > Sent: Thursday, February 23, 2023 2:30 PM > > > > > To: paulmck@kernel.org; frederic@kernel.org; quic_neeraju@quicinc.com; > > > > > joel@joelfernandes.org > > > > > Cc: rcu@vger.kernel.org; linux-kernel@vger.kernel.org > > > > > Subject: [PATCH] rcu-tasks: Directly invoke rcuwait_wake_up() in > > > > > call_rcu_tasks_generic() > > > > > > > > > > According to commit '3063b33a347c ("Avoid raw-spinlocked wakeups from > > > > > call_rcu_tasks_generic()")', the grace-period kthread is delayed to wakeup > > > > > using irq_work_queue() is because if the caller of > > > > > call_rcu_tasks_generic() holds a raw spinlock, when the kernel is built with > > > > > CONFIG_PROVE_RAW_LOCK_NESTING=y, due to a spinlock will be hold in > > > > > wake_up(), so the lockdep splats will happen. but now using > > > > > rcuwait_wake_up() to wakeup grace-period kthread instead of wake_up(), in > > > > > rcuwait_wake_up() no spinlock will be acquired, so this commit remove using > > > > > > > > > >There are still spinlock-acquisition and spinlock-release invocations within the call path from rcuwait_wake_up(). > > > > > > > > > >rcuwait_wake_up() -> wake_up_process() -> try_to_wake_up(), then: > > > > > > > > > > raw_spin_lock_irqsave() > > > > > ... > > > > > raw_spin_unlock_irqrestore > > > > > > > > Yes, but this is raw_spinlock acquisition and release(note: spinlock will convert to > > > > sleepable lock in Preempt-RT kernel, but raw spinlock is not change). > > > > > > > > acquire raw_spinlock -> acquire spinlock will trigger lockdep warning. > > > > > > > >Is this really safe in the long run though? I seem to remember there are > > > >weird locking dependencies if RCU is used from within the scheduler [1]. > > > > > > > > > > > > > I have been running rcutorture with rcutorture.type = tasks-tracing, > > > so far no problems have been found. > > > > > > > > > >I prefer to keep it as irq_work_queue() unless you are seeing some benefit. > > > >Generally, there has to be a 'win' or other justification for adding more > > > >risk. > > > > > > > >thanks, > > > > > > > >- Joel > > > >[1] http://www.joelfernandes.org/rcu/scheduler/locking/2019/09/02/rcu-schedlocks.html > > > > > > > > > The problem in this link, in an earlier RCU version, rcu_read_unlock_special() > > > Invoke wakeup and enter scheduler can lead to deadlock, but my modification is for > > > call_rcu_tasks_generic(), even if there is a lock dependency problem, we should pay > > > more attention to rcu_read_unlock_trace_special() > > > > Consider ABBA deadlocks as well, not just self-deadlocks (which IIRC is what > > the straight-RCU rcu_read_unlock() issues were about). > > > > What prevents the following scenario? > > > > In the scheduler you have code like this: > > rq = task_rq_lock(p, &rf); > > trace_sched_wait_task(p); > > > > Someone can hook up a BPF program to that tracepoint that then calls > > rcu_read_unlock_trace() -> rcu_read_unlock_trace_special(). All of > > this while holding the rq and pi scheduler locks. > > > > That's A (rq lock) -> B (rtpcp lock). > > > > In another path, your change adds the following dependency due to doing > > wakeup under the rtpcp lock. > > > > That's call_rcu_tasks_generic() -> B (rtpcp lock) -> A (rq lock). > > I would like to correct this last statement. That cannot happen but > the concern I guess is, can the following happen due to the change? > > call_rcu_tasks_generic() -> B (some BPF lock) -> A (rq lock) > Aaargh, one more correction: B (some BPF lock) -> call_rcu_tasks_generic() -> -> A (rq lock) ;-) -Joel