Received: by 2002:ac0:bc90:0:0:0:0:0 with SMTP id a16csp802586img; Fri, 22 Mar 2019 08:53:23 -0700 (PDT) X-Google-Smtp-Source: APXvYqw0gjZySh6t2/T3mDbmAhFr4yAvPn3sDCpuIIubWXs8jUhF7gUD7eZQRls5FjMkjeZ+xLB3 X-Received: by 2002:a17:902:d70f:: with SMTP id w15mr10304154ply.234.1553270003382; Fri, 22 Mar 2019 08:53:23 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1553270003; cv=none; d=google.com; s=arc-20160816; b=LLRlNhIK7IWkE8y6V4V+tJW2H76SbYTLBNA9LVbmzFJ++3yD6VWZMCVuMsQFb2BvYv NwSc9AXPDiFi+42wEr6vpoVvA0hEFfNGYOdoz4Kumwo8Mc9gYnAcrnJ/pLPPHLPbbBeJ BuDiSX1fO85x4ER7K8OBmnPa8IMwpvHyu0xvwQTMfmPZxTKuzh34BTwjRWZC3rFJRDAx IOsCeFHis9CsYN4O2OpHimNocCcACQOQrb0pqtg0czi6yokkm8LkrYE2a7GZ3ppCM8+/ wy7HEcfx49PzGyQlYOuWgvaFIiEkma2Ghga9g8VU+VBn5ELktCVsJoJ57+qeZAgTuD4G VFqw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:user-agent:in-reply-to :content-disposition:mime-version:references:message-id:subject:cc :to:from:date:dkim-signature; bh=AM2zwyWctDQ2UX6SHcdI5QG21PJCaZEAyYeGKJYkSzU=; b=STkI4mLEyMn/ZCb5Qtd/sBbRGstjFgbUkHvhBMyhH72stkYKHueWGoqayEeFh79fHg j+S9gAVc7RQIXfWsLC1utVjcflZhfN4JTYeHn3fPTF76si0NxPEs/gfsnOotPdBqwWNi dDhufjrkwpTKB3aVdNoxYQVCbHfUH/jecQV5DlzLIJsCZsmUSaYu0yFGQOFS4AWmRLUD LHnE0lWgyOw/ZMi58HAIZzZUDNz4G8ZTKSqW6TjqpEKLLu9FD0rdyjzVf8fuol++y0+j l2XkmMDIj9WFbYbDMM6XssgScyqF7zuk7BCHFZBmpaqunTWJArkKfITRGzBJJRb1ru+P 3ZTQ== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@joelfernandes.org header.s=google header.b=DhfsOzmg; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id m3si6752761pgp.263.2019.03.22.08.53.05; Fri, 22 Mar 2019 08:53:23 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; dkim=pass header.i=@joelfernandes.org header.s=google header.b=DhfsOzmg; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1727908AbfCVPux (ORCPT + 99 others); Fri, 22 Mar 2019 11:50:53 -0400 Received: from mail-pg1-f195.google.com ([209.85.215.195]:33463 "EHLO mail-pg1-f195.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1727363AbfCVPuw (ORCPT ); Fri, 22 Mar 2019 11:50:52 -0400 Received: by mail-pg1-f195.google.com with SMTP id b12so1820949pgk.0 for ; Fri, 22 Mar 2019 08:50:52 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=joelfernandes.org; s=google; h=date:from:to:cc:subject:message-id:references:mime-version :content-disposition:in-reply-to:user-agent; bh=AM2zwyWctDQ2UX6SHcdI5QG21PJCaZEAyYeGKJYkSzU=; b=DhfsOzmg8c0pXR0EphcZ6NyAftJb5+sZo/AsfxUdwz00vJ1BxEmOsOeHk3ZYQZYiF5 I12uB+SBhIums7U7IKigFO3I9sFdwZHu4LLM7swWD4JDAMgoRoOKAr774f3V7koUKSKy 2Lq7+vKxpKYixXWvcSzMJLUglpo5RchxFn4g4= X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:date:from:to:cc:subject:message-id:references :mime-version:content-disposition:in-reply-to:user-agent; bh=AM2zwyWctDQ2UX6SHcdI5QG21PJCaZEAyYeGKJYkSzU=; b=NnFuk2Ppjy8q8RQhbY/P3ObWD+4MlqxEDIkhfXBDdB/tOqaer3OAltoeevL7vpSW2M Reswa77+KFVEA5mHcva3Aotl0722nM70lVYv14SN0oc0gsyP7oZ9Y4Urd+hGm/ws0JoJ rz+oQlHWDlMawLJr2YZ84Lk1CpfC+h8VCK6rKjdOUhJdwBHfoeuxE6xXGrzO87MWXUyW pHMlc4EdLu8DqeDT4riWx2O452AVp9t0aC/P59j5voAZM9ZELIrZXV3NV6CeBpICxLqb cWTKNsYGv8UyGG5X2LoOxK4BKKXF25WpzDp/YLd+Q87sC04z7SfD8kBrcI+D6W3VqaaA nmsQ== X-Gm-Message-State: APjAAAVEaD5ytj6AwyQhgGBnfvWvWgnfe/+Sy1m1IUPhruojlN6udOCI mYtHrDJaTmEqYRY2a2/uuUwJYQ== X-Received: by 2002:a63:fc49:: with SMTP id r9mr9427942pgk.447.1553269851637; Fri, 22 Mar 2019 08:50:51 -0700 (PDT) Received: from localhost ([2620:15c:6:12:9c46:e0da:efbf:69cc]) by smtp.gmail.com with ESMTPSA id x13sm9251540pgr.87.2019.03.22.08.50.49 (version=TLS1_2 cipher=ECDHE-RSA-CHACHA20-POLY1305 bits=256/256); Fri, 22 Mar 2019 08:50:50 -0700 (PDT) Date: Fri, 22 Mar 2019 11:50:49 -0400 From: Joel Fernandes To: "Paul E. McKenney" Cc: Sebastian Andrzej Siewior , linux-kernel@vger.kernel.org, Josh Triplett , Steven Rostedt , Mathieu Desnoyers , Lai Jiangshan , tglx@linutronix.de, Mike Galbraith Subject: Re: [PATCH v3] rcu: Allow to eliminate softirq processing from rcutree Message-ID: <20190322155049.GA86662@google.com> References: <20190320163532.mr32oi53iaueuizw@linutronix.de> <20190320173001.GM4102@linux.ibm.com> <20190320175952.yh6yfy64vaiurszw@linutronix.de> <20190320181210.GO4102@linux.ibm.com> <20190320181435.x3qyutwqllmq5zbk@linutronix.de> <20190320211333.eq7pwxnte7la67ph@linutronix.de> <20190320234601.GQ4102@linux.ibm.com> <20190321233244.GA11476@linux.ibm.com> <20190322134207.GA56461@google.com> <20190322145823.GM4102@linux.ibm.com> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20190322145823.GM4102@linux.ibm.com> User-Agent: Mutt/1.10.1 (2018-07-13) Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Fri, Mar 22, 2019 at 07:58:23AM -0700, Paul E. McKenney wrote: [snip] > > > > #ifdef CONFIG_RCU_NOCB_CPU > > > > static cpumask_var_t rcu_nocb_mask; /* CPUs to have callbacks offloaded. */ > > > > @@ -94,6 +72,8 @@ static void __init rcu_bootup_announce_oddness(void) > > > > pr_info("\tRCU debug GP init slowdown %d jiffies.\n", gp_init_delay); > > > > if (gp_cleanup_delay) > > > > pr_info("\tRCU debug GP init slowdown %d jiffies.\n", gp_cleanup_delay); > > > > + if (!use_softirq) > > > > + pr_info("\tRCU_SOFTIRQ processing moved to rcuc kthreads.\n"); > > > > if (IS_ENABLED(CONFIG_RCU_EQS_DEBUG)) > > > > pr_info("\tRCU debug extended QS entry/exit.\n"); > > > > rcupdate_announce_bootup_oddness(); > > > > @@ -629,7 +609,10 @@ static void rcu_read_unlock_special(struct task_struct *t) > > > > /* Need to defer quiescent state until everything is enabled. */ > > > > if (irqs_were_disabled) { > > > > /* Enabling irqs does not reschedule, so... */ > > > > - raise_softirq_irqoff(RCU_SOFTIRQ); > > > > + if (!use_softirq) > > > > + raise_softirq_irqoff(RCU_SOFTIRQ); > > > > + else > > > > + invoke_rcu_core(); > > > > > > This can result in deadlock. This happens when the scheduler invokes > > > rcu_read_unlock() with one of the rq or pi locks held, which means that > > > interrupts are disabled. And it also means that the wakeup done in > > > invoke_rcu_core() could go after the same rq or pi lock. > > > > > > What we really need here is some way to make soemthing happen on this > > > CPU just after interrupts are re-enabled. Here are the options I see: > > > > > > 1. Do set_tsk_need_resched() and set_preempt_need_resched(), > > > just like in the "else" clause below. This sort of works, but > > > relies on some later interrupt or similar to get things started. > > > This is just fine for normal grace periods, but not so much for > > > expedited grace periods. > > > > > > 2. IPI some other CPU and have it IPI us back. Not such a good plan > > > when running an SMP kernel on a single CPU. > > > > > > 3. Have a "stub" RCU_SOFTIRQ that contains only the following: > > > > > > /* Report any deferred quiescent states if preemption enabled. */ > > > if (!(preempt_count() & PREEMPT_MASK)) { > > > rcu_preempt_deferred_qs(current); > > > } else if (rcu_preempt_need_deferred_qs(current)) { > > > set_tsk_need_resched(current); > > > set_preempt_need_resched(); > > > } > > > > > > 4. Except that raise_softirq_irqoff() could potentially have this > > > same problem if rcu_read_unlock() is invoked at process level > > > from the scheduler with either rq or pi locks held. :-/ > > > > > > Which raises the question "why aren't I seeing hangs and > > > lockdep splats?" > > > > Interesting, could it be you're not seeing a hang in the regular case, > > because enqueuing ksoftirqd on the same CPU as where the rcu_read_unlock is > > happening is a rare event? First, ksoftirqd has to even be awakened in the > > first place. On the other hand, with the new code the thread is always awaked > > and is more likely to run into the issue you found? > > No, in many cases, including the self-deadlock that showed up last night, > raise_softirq_irqoff() will simply set a bit in a per-CPU variable. > One case where this happens is when called from an interrupt handler. I think we are saying the same thing, in some cases ksoftirqd will be awakened and some case it will not. I will go through all scenarios to convince myself it is safe, if I find some issue I will let you know. > > The lockdep splats should be a more common occurence though IMO. If you could > > let me know which RCU config is hanging, I can try to debug this at my end as > > well. > > TREE01, TREE02, TREE03, and TREE09. I would guess that TREE08 would also > do the same thing, given that it also sets PREEMPT=y and tests Tree RCU. > > Please see the patch I posted and tested overnight. I suspect that there > is a better fix, but this does at least seem to suppress the error. Ok, will do. > > > Also, having lots of non-migratable timers might be considered unfriendly, > > > though they shouldn't be -that- heavily utilized. Yet, anyway... > > > I could try adding logic to local_irq_enable() and local_irq_restore(), > > > but that probably wouldn't go over all that well. Besides, sometimes > > > interrupt enabling happens in assembly language. > > > > > > It is quite likely that delays to expedited grace periods wouldn't > > > happen all that often. First, the grace period has to start while > > > the CPU itself (not some blocked task) is in an RCU read-side critical > > > section, second, that critical section cannot be preempted, and third > > > the rcu_read_unlock() must run with interrupts disabled. > > > > > > Ah, but that sequence of events is not supposed to happen with the > > > scheduler lock! > > > > > > From Documentation/RCU/Design/Requirements/Requirements.html: > > > > > > It is forbidden to hold any of scheduler's runqueue or > > > priority-inheritance spinlocks across an rcu_read_unlock() > > > unless interrupts have been disabled across the entire RCU > > > read-side critical section, that is, up to and including the > > > matching rcu_read_lock(). > > > > > > Here are the reasons we even get to rcu_read_unlock_special(): > > > > > > 1. The just-ended RCU read-side critical section was preempted. > > > This clearly cannot happen if interrupts are disabled across > > > the entire critical section. > > > > > > 2. The scheduling-clock interrupt noticed that this critical > > > section has been taking a long time. But scheduling-clock > > > interrupts also cannot happen while interrupts are disabled. > > > > > > 3. An expedited grace periods started during this critical > > > section. But if that happened, the corresponding IPI would > > > have waited until this CPU enabled interrupts, so this > > > cannot happen either. > > > > > > So the call to invoke_rcu_core() should be OK after all. > > > > > > Which is a bit of a disappointment, given that I am still seeing hangs! > > > > Oh ok, discount whatever I just said then ;-) Indeed I remember this > > requirement too now. Your neat documentation skills are indeed life saving :D > > No, this did turn out to be the problem area. Or at least one of the > problem areas. Again, see my earlier email. Ok. Too many emails so I got confused :-D. I also forgot which version of the patch are we testing since I don't think an updated one was posted. But I will refer to your last night diff dig out the base patch from your git tree, no problem. > > > I might replace this invoke_rcu_core() with set_tsk_need_resched() and > > > set_preempt_need_resched() to see if that gets rid of the hangs, but > > > first... > > > > Could we use the NMI watchdog to dump the stack at the time of the hang? May > > be a deadlock will present on the stack (I think its config is called > > HARDLOCKUP_DETECTOR or something). > > Another approach would be to instrument the locking code that notices > the recursive acquisition. Or to run lockdep... Because none of the > failing scenarios enable lockdep! ;-) I was wondering why lockdep is not always turned on in your testing. Is it due to performance concerns? thanks, - Joel