Received: by 2002:a25:8b12:0:0:0:0:0 with SMTP id i18csp581ybl; Thu, 15 Aug 2019 11:31:22 -0700 (PDT) X-Google-Smtp-Source: APXvYqw9KTycRq3apb7Z6YqmaG5hU+TJxVLbSNqC8BH3u350lydW2smMHO9beANhnygxsxKN9lWq X-Received: by 2002:a62:5883:: with SMTP id m125mr6749044pfb.248.1565893882229; Thu, 15 Aug 2019 11:31:22 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1565893882; cv=none; d=google.com; s=arc-20160816; b=zLa8bL5SPmUglGOWq1R/NWmD3xlF4J3p6kgvMb9ls2GCm0xGwIfPZfPp16AqIMZW0w G03Lbj0Je4KuDpyufPFOZ6jLIYrAQsbE9StwuWwDlC8MKNS1NuFJE+8ywODZeZnPE2fk zR1kyZWaJRXabseQLYesjV2mXWE9lEEuhuKNZ7vQla8Rx+7OSJdGrWaecrisbe8dHU9m CzsRfOCaPkZQ0+yjfBdBJ8Jdjd8t93UgKY6lV2g47RYJaVbnZAakGHypFKwm/E6LryE1 FQzDpCXZ0JxIN7ScV48Jsom6h2jbxTzv2cgs76Q8svFBsShW8MFb351+QvD6lB5y+Xzd zGzg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:user-agent:in-reply-to :content-disposition:mime-version:references:message-id:subject:cc :to:from:date:dkim-signature; bh=wLsgqXU1lI5ukS7nuzUPXn2TwlmqZhK7yyq3cYkhveo=; b=khLWLqn0T5jzoICj97AJKIuHMjBC8sHwqUUfq4FrLtZeVRNGzTw4nXRIx4ToNfwLm4 Pz8LguIIuSyY1PTyQhRLBaa0VZTNfhkV4VwAF6qFdQQYfCHSCy0JloOao90FrzZZtBFD zI8JclyLFuWRmha1CKm0uSDoKYP9DVskZnhWlAgn+/Vdnffwa6Ev9SK6k9XS8HQKR24S TN7uqqXde8I0LsMxuEnETkb2VkfQbYOk/Zu0n8LOZEBSYQOu1HmDgcp2PXnbP10XFEV1 ORZyjWyK6hiS+sRNeUdOCr/GSKujy5qrTwrWssSj2JTfGoGReJtJuYa3lnoz1XU8GAaX U0aA== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@joelfernandes.org header.s=google header.b="p5/r3BDq"; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id ay1si2124398plb.83.2019.08.15.11.31.05; Thu, 15 Aug 2019 11:31:22 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; dkim=pass header.i=@joelfernandes.org header.s=google header.b="p5/r3BDq"; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1732645AbfHOSPU (ORCPT + 99 others); Thu, 15 Aug 2019 14:15:20 -0400 Received: from mail-pg1-f195.google.com ([209.85.215.195]:34177 "EHLO mail-pg1-f195.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1729204AbfHOSPU (ORCPT ); Thu, 15 Aug 2019 14:15:20 -0400 Received: by mail-pg1-f195.google.com with SMTP id n9so1649599pgc.1 for ; Thu, 15 Aug 2019 11:15:19 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=joelfernandes.org; s=google; h=date:from:to:cc:subject:message-id:references:mime-version :content-disposition:in-reply-to:user-agent; bh=wLsgqXU1lI5ukS7nuzUPXn2TwlmqZhK7yyq3cYkhveo=; b=p5/r3BDqyePLRWZQ4kkU58Z4NXUAlgxniz8FZG8FodttPDb6DqKAwtfqBI8v/T7JKf IKwqVHBxHrJsav3WVNp0XCsWhr3VNrX2dwvYrtE8BkHFbjfzVsVSOJvH8NDMtIz8jyi+ iGdcy7h2WlDyP2NxctB4VG8nteB3Nl7KObP+A= X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:date:from:to:cc:subject:message-id:references :mime-version:content-disposition:in-reply-to:user-agent; bh=wLsgqXU1lI5ukS7nuzUPXn2TwlmqZhK7yyq3cYkhveo=; b=KcIPb+ZN/CJol/uNrN4YUF8Llf/VQBt1M8fVnL1jCFbLXIpegZYFhlhYswWx8O/Oc0 h98S6eUNsOADFstrKeTJB/xnpqXFdVhRbWcd0BMRKl/CtcNGi3k7CM5PoNqFRX3n4gJU DJd5CaDLK4HU3Cvc68JqFGIUKgCB/MUZBJX0wCodVJGAaosWUHsm5QdZ+brNtX/cl8lZ yV3TSxmkbaO13ZMmAI6NIb4UEfVdbMb/HmJ5SHSs18a1nuktD4V/KCQdhrufHYXll+Ox XY2jR4HdYWlo8ves0lLwUlxb24X7v1RASRO5ZqHHVz2LAKYwk84g0mBWEfwnPDEogS18 pvKw== X-Gm-Message-State: APjAAAWz8351U5lsg3vj1BLmerGs8cun9m/8nj0VK/5p9X4gWKNwL65+ +zfnR5O8L5k+vPOlmjLf+cJKnQ== X-Received: by 2002:aa7:93aa:: with SMTP id x10mr6846163pff.83.1565892918857; Thu, 15 Aug 2019 11:15:18 -0700 (PDT) Received: from localhost ([172.19.216.18]) by smtp.gmail.com with ESMTPSA id a128sm3419007pfb.185.2019.08.15.11.15.17 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Thu, 15 Aug 2019 11:15:17 -0700 (PDT) Date: Thu, 15 Aug 2019 14:15:00 -0400 From: Joel Fernandes To: "Paul E. McKenney" Cc: Frederic Weisbecker , rcu@vger.kernel.org, linux-kernel@vger.kernel.org, mingo@kernel.org, jiangshanlai@gmail.com, dipankar@in.ibm.com, akpm@linux-foundation.org, mathieu.desnoyers@efficios.com, josh@joshtriplett.org, tglx@linutronix.de, peterz@infradead.org, rostedt@goodmis.org, dhowells@redhat.com, edumazet@google.com, fweisbec@gmail.com, oleg@redhat.com Subject: Re: [PATCH RFC tip/core/rcu 14/14] rcu/nohz: Make multi_cpu_stop() enable tick on all online CPUs Message-ID: <20190815181500.GC12078@google.com> References: <20190802151435.GA1081@linux.ibm.com> <20190802151501.13069-14-paulmck@linux.ibm.com> <20190812210232.GA3648@lenoir> <20190812232316.GT28441@linux.ibm.com> <20190813123016.GA11455@lenoir> <20190813144809.GB28441@linux.ibm.com> <20190814175546.GB68498@google.com> <20190814220516.GY28441@linux.ibm.com> <20190815150735.GA12078@google.com> <20190815172351.GI28441@linux.ibm.com> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20190815172351.GI28441@linux.ibm.com> User-Agent: Mutt/1.10.1 (2018-07-13) Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Thu, Aug 15, 2019 at 10:23:51AM -0700, Paul E. McKenney wrote: > On Thu, Aug 15, 2019 at 11:07:35AM -0400, Joel Fernandes wrote: > > On Wed, Aug 14, 2019 at 03:05:16PM -0700, Paul E. McKenney wrote: > > [snip] > > > > > If so, perhaps that monitoring could periodically invoke an RCU function > > > > > that I provide for deciding when to turn the tick on. We would also need > > > > > to work out how to turn the tick off in a timely fashion once the CPU got > > > > > out of kernel mode, perhaps in rcu_user_enter() or rcu_nmi_exit_common(). > > > > > > > > > > If this would be called only every second or so, the separate grace-period > > > > > checking is still needed for its shorter timespan, though. > > > > > > > > > > Thoughts? > > > > > > > > Do you want me to test the below patch to see if it fixes the issue with my > > > > other test case (where I had a nohz full CPU holding up a grace period). > > > > > > Please! > > > > I tried the patch below, but it did not seem to make a difference to the > > issue I was seeing. My test tree is here in case you can spot anything I did > > not do right: https://github.com/joelagnel/linux-kernel/commits/rcu/nohz-test > > The main patch is here: > > https://github.com/joelagnel/linux-kernel/commit/4dc282b559d918a0be826936f997db0bdad7abb3 > > That is more aggressive that rcutorture's rcu_torture_fwd_prog_nr(), so > I am guessing that I need to up rcu_torture_fwd_prog_nr()'s game. I am > currently testing that. > > > On the trace output, I grep something like: egrep "(rcu_perf|cpu 3|3d)". I > > see a few ticks after 300ms, but then there are no more ticks and just a > > periodic resched_cpu() from rcu_implicit_dynticks_qs(): > > > > [ 19.534107] rcu_perf-165 12.... 2276436us : rcu_perf_writer: Start of rcuperf test > > [ 19.557968] rcu_pree-10 0d..1 2287973us : rcu_implicit_dynticks_qs: Sending urgent resched to cpu 3 > > [ 20.136222] rcu_perf-165 3d.h. 2591894us : rcu_sched_clock_irq: sched-tick > > [ 20.137185] rcu_perf-165 3d.h2 2591906us : rcu_sched_clock_irq: sched-tick > > [ 20.138149] rcu_perf-165 3d.h. 2591911us : rcu_sched_clock_irq: sched-tick > > [ 20.139106] rcu_perf-165 3d.h. 2591915us : rcu_sched_clock_irq: sched-tick [snip] > > [ 20.147797] rcu_perf-165 3d.h. 2591953us : rcu_sched_clock_irq: sched-tick > > [ 20.148759] rcu_perf-165 3d.h. 2591957us : rcu_sched_clock_irq: sched-tick > > [ 20.151655] rcu_pree-10 0d..1 2591979us : rcu_implicit_dynticks_qs: Sending urgent resched to cpu 3 > > [ 20.732938] rcu_pree-10 0d..1 2895960us : rcu_implicit_dynticks_qs: Sending urgent resched to cpu 3 [snip] > > [ 26.566100] rcu_pree-10 0d..1 5935982us : rcu_implicit_dynticks_qs: Sending urgent resched to cpu 3 > > [ 27.144497] rcu_pree-10 0d..1 6239973us : rcu_implicit_dynticks_qs: Sending urgent resched to cpu 3 > > [ 27.192661] rcu_perf-165 3d.h. 6276923us : rcu_sched_clock_irq: sched-tick > > [ 27.705789] rcu_pree-10 0d..1 6541901us : rcu_implicit_dynticks_qs: Sending urgent resched to cpu 3 > > [ 28.292155] rcu_pree-10 0d..1 6845974us : rcu_implicit_dynticks_qs: Sending urgent resched to cpu 3 > > [ 28.874049] rcu_pree-10 0d..1 7149972us : rcu_implicit_dynticks_qs: Sending urgent resched to cpu 3 > > [ 29.112646] rcu_perf-165 3.... 7275951us : rcu_perf_writer: End of rcuperf test > > That would be due to my own stupidity. I forgot to clear ->rcu_forced_tick > in rcu_disable_tick_upon_qs() inside the "if" statement. This of course > prevents rcu_nmi_exit_common() from ever re-enabling it. > > Excellent catch! Thank you for testing this!!! Ah I missed it too. Happy to help! I tried setting it as below but getting same results: +/* + * If the scheduler-clock interrupt was enabled on a nohz_full CPU + * in order to get to a quiescent state, disable it. + */ +void rcu_disable_tick_upon_qs(struct rcu_data *rdp) +{ + if (tick_nohz_full_cpu(rdp->cpu) && rdp->rcu_forced_tick) + tick_dep_clear_cpu(rdp->cpu, TICK_DEP_MASK_RCU); + rdp->rcu_forced_tick = false; +} + > > [snip] > > > > > if (rnp->qsmask & mask) { /* RCU waiting on incoming CPU? */ > > > > > + rcu_disable_tick_upon_qs(rdp); > > > > > /* Report QS -after- changing ->qsmaskinitnext! */ > > > > > rcu_report_qs_rnp(mask, rnp, rnp->gp_seq, flags); > > > > > > > > Just curious about the existing code. If a CPU is just starting up (after > > > > bringing it online), how can RCU be waiting on it? I thought RCU would not be > > > > watching offline CPUs. > > > > > > Well, neither grace periods nor CPU-hotplug operations are atomic, > > > and each can take significant time to complete. > > > > > > So suppose we have a large system with multiple leaf rcu_node structures > > > (not that 17 CPUs is all that many these days, but please bear with me). > > > Suppose just after a new grace period initializes a given leaf rcu_node > > > structure, one of its CPUs goes offline (yes, that CPU would have to > > > have waited on a grace period, but that might have been the previous > > > grace period). But before the FQS scan notices that RCU is waiting on > > > an offline CPU, the CPU comes back online. > > > > > > That situation is exactly what the above code is intended to handle. > > > > That makes sense! > > > > > Without that code, RCU can give false-positive splats at various points > > > in its processing. ("Wait! How can a task be blocked waiting on a > > > grace period that hasn't even started yet???") > > > > I did not fully understand the question in brackets though, a task can be on > > a different CPU though which has nothing to do with the CPU that's going > > offline/online so it could totally be waiting on a grace period right? > > > > Also waiting on a grace period that hasn't even started is totally possible: > > > > GP1 GP2 > > |<--------->|<-------->| > > ^ ^ > > | |____ task gets unblocked > > task blocks > > on synchronize_rcu > > but is waiting on > > GP2 which hasn't started > > > > Or did I misunderstand the question? > > There is a ->gp_tasks field in the leaf rcu_node structures that > references a list of tasks blocking the current grace period. When there > is no grace period in progress (as is the case from the end of GP1 to > the beginning of GP2, the RCU code expects ->gp_tasks to be NULL. > Without the curiosity code you pointed out above, ->gp_tasks could > in fact end up being non-NULL when no grace period was in progress. > > And did end up being non-NULL from time to time, initially every few > hundred hours of a particular rcutorture scenario. Oh ok! I will think more about it. I am not yet able to connect the gp_tasks being non-NULL to the CPU going offline/online scenario though. Maybe I should delete this code, run an experiment and trace for this condition (gp_tasks != NULL)? I love it how you found these issues by heavy testing and fixed them. thanks, - Joel