Received: by 2002:a05:6a10:a0d1:0:0:0:0 with SMTP id j17csp3164162pxa; Tue, 25 Aug 2020 13:16:25 -0700 (PDT) X-Google-Smtp-Source: ABdhPJwHGnu0OYQVwkp1C2AFbE/XFPkNL3WnbY6MVkABsVK19MeHBJWEB2rXe37QX4Uf8QUjxpl7 X-Received: by 2002:a17:906:3e1a:: with SMTP id k26mr11202782eji.188.1598386585532; Tue, 25 Aug 2020 13:16:25 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1598386585; cv=none; d=google.com; s=arc-20160816; b=rYxw8mWQK29UlvAsHlKBbeH3VSDMn0oB7KSmglkvJu9Q+0Nug6nhGn0a8mV0jOUFVT sD7Su1WSQq+1d72DWIAX5oDHDKKWIXiJOJULbdNeX5AOA5jEKQvVlwElrL19YAdw86Dd sOjva4wcMIfrem4N58bR24ESY/ShsuKzuCZxaKYr9nJ3XIqTm4D1rJ8fGlcXPmWOMrsf ULhfRuTa6vbqznbIPf786mbOfrQAc+5CtQv4PE7crEC66YY5dLk2G6/cRVO1F4OJupFb sVHVK6ah8Nm30lFti/sSfc+qyfck+k3jp3tSeTboHlJyasACu3aHPHJwrB5I7BgXs6kT ehaw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:user-agent:in-reply-to :content-disposition:mime-version:references:reply-to:message-id :subject:cc:to:from:date:dkim-signature; bh=112P8kzOcPSxB2QdZJK1QKz5RilHzfLYg/o/FK1/Lxg=; b=zBNIagXDcYfD1vEZa1xtOCxel3zDFPglu9GQxFNKSXgPyXgcalyZYd/b/Ph4TWtWrf 4v1ctrSGWBPsXokSsl07SmokQbF9CEseI1cRZbhNTUQ7yBnU1syyLX4+6UdqOWP9UAeC IYcsjnL/Twu8MgR+dfK7Wwn29uFBsd3x3tsTYtQEaPPiHUbNQKDa4hiZWSyif1DOttdI IIf7mbMTKI9i7kKPQQ9fZaAIy6DhIGoRY4ZMxNbiki6xnQ+EqyUPNXJjKNGxMGot/LzP 75dSpl79snapr5Ep/eLwma0o2ilnew7CDnhX+5ZIJP6NrxgXEFNm+Ejf6/eLKp6vlKsJ v0gQ== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@kernel.org header.s=default header.b=EEOc6Dfw; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id z15si88706ejr.114.2020.08.25.13.16.02; Tue, 25 Aug 2020 13:16:25 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; dkim=pass header.i=@kernel.org header.s=default header.b=EEOc6Dfw; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1726541AbgHYUN4 (ORCPT + 99 others); Tue, 25 Aug 2020 16:13:56 -0400 Received: from mail.kernel.org ([198.145.29.99]:53830 "EHLO mail.kernel.org" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726149AbgHYUN4 (ORCPT ); Tue, 25 Aug 2020 16:13:56 -0400 Received: from paulmck-ThinkPad-P72.home (unknown [50.45.173.55]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by mail.kernel.org (Postfix) with ESMTPSA id 813D220738; Tue, 25 Aug 2020 20:13:55 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=kernel.org; s=default; t=1598386435; bh=AKpUwXbrvLHVdIrs1d1s9f5rMKCIf8DxBfGkM26WAbM=; h=Date:From:To:Cc:Subject:Reply-To:References:In-Reply-To:From; b=EEOc6Dfw6VquQL7OMVibtTzxHXT9ts/F+dcrk71UOtFuhnp99SpAL5rAHxOD9o+7R 25/Yj8RGZyVH6GkO69UO9zaXlc0GlE54y1kklT32A4bJYt3n6U0LeBtVHIwrcEeTzQ iV1GJDmmnE295TAaM7CMfj8yvsWA7b0pVouV2fkY= Received: by paulmck-ThinkPad-P72.home (Postfix, from userid 1000) id 3B7EC35226AE; Tue, 25 Aug 2020 13:13:55 -0700 (PDT) Date: Tue, 25 Aug 2020 13:13:55 -0700 From: "Paul E. McKenney" To: "Joel Fernandes (Google)" Cc: linux-kernel@vger.kernel.org, boqun.feng@gmail.com, dave@stgolabs.net, Ingo Molnar , Josh Triplett , Lai Jiangshan , Madhuparna Bhowmik , Mathieu Desnoyers , neeraj.iitr10@gmail.com, rcu@vger.kernel.org, Steven Rostedt , "Uladzislau Rezki (Sony)" , vineethrp@gmail.com Subject: Re: [PATCH v4 -rcu 2/4] rcu/tree: Make rcu_do_batch count how many callbacks were executed Message-ID: <20200825201355.GX2855@paulmck-ThinkPad-P72> Reply-To: paulmck@kernel.org References: <20200825024842.3408659-1-joel@joelfernandes.org> <20200825024842.3408659-3-joel@joelfernandes.org> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20200825024842.3408659-3-joel@joelfernandes.org> User-Agent: Mutt/1.9.4 (2018-02-28) Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Mon, Aug 24, 2020 at 10:48:40PM -0400, Joel Fernandes (Google) wrote: > Currently, rcu_do_batch() depends on the unsegmented callback list's len field > to know how many CBs are executed. This fields counts down from 0 as CBs are > dequeued. It is possible that all CBs could not be run because of reaching > limits in which case the remaining unexecuted callbacks are requeued in the > CPU's segcblist. Again, please mention why the ->cblist.len fields are preserved, namely due to interactions with rcu_barrier(). > The number of callbacks that were not requeued are then the negative count (how > many CBs were run) stored in the rcl->len which has been counting down on every > dequeue. This negative count is then added to the per-cpu segmented callback > list's to correct its count. > > Such a design works against future efforts to track the length of each segment > of the segmented callback list. The reason is because > rcu_segcblist_extract_done_cbs() will have to store the length of the callback > list in rcl->len to make rcu_segcblist_merge() work. > > Also, the design of counting down from 0 is confusing and error-prone IMHO. > > This commit therefore explicitly counts have many callbacks were executed in > rcu_do_batch() itself, and uses that to update the per-CPU segcb list's len > field, without relying on the negativity of rcl->len. This last, as noted in response to 1/4, should allow rcu_segcblist_merge() to avoid carrying the callback count separately. > Signed-off-by: Joel Fernandes (Google) > --- > kernel/rcu/rcu_segcblist.c | 2 +- > kernel/rcu/rcu_segcblist.h | 1 + > kernel/rcu/tree.c | 9 ++++----- > 3 files changed, 6 insertions(+), 6 deletions(-) > > diff --git a/kernel/rcu/rcu_segcblist.c b/kernel/rcu/rcu_segcblist.c > index b70d4154433c..076337ae2e50 100644 > --- a/kernel/rcu/rcu_segcblist.c > +++ b/kernel/rcu/rcu_segcblist.c > @@ -95,7 +95,7 @@ static void rcu_segcblist_set_len(struct rcu_segcblist *rsclp, long v) > * This increase is fully ordered with respect to the callers accesses > * both before and after. > */ > -static void rcu_segcblist_add_len(struct rcu_segcblist *rsclp, long v) > +void rcu_segcblist_add_len(struct rcu_segcblist *rsclp, long v) > { > #ifdef CONFIG_RCU_NOCB_CPU > smp_mb__before_atomic(); /* Up to the caller! */ > diff --git a/kernel/rcu/rcu_segcblist.h b/kernel/rcu/rcu_segcblist.h > index 5c293afc07b8..b90725f81d77 100644 > --- a/kernel/rcu/rcu_segcblist.h > +++ b/kernel/rcu/rcu_segcblist.h > @@ -76,6 +76,7 @@ static inline bool rcu_segcblist_restempty(struct rcu_segcblist *rsclp, int seg) > } > > void rcu_segcblist_inc_len(struct rcu_segcblist *rsclp); > +void rcu_segcblist_add_len(struct rcu_segcblist *rsclp, long v); > void rcu_segcblist_init(struct rcu_segcblist *rsclp); > void rcu_segcblist_disable(struct rcu_segcblist *rsclp); > void rcu_segcblist_offload(struct rcu_segcblist *rsclp); > diff --git a/kernel/rcu/tree.c b/kernel/rcu/tree.c > index 548404489c04..51348144a4ea 100644 > --- a/kernel/rcu/tree.c > +++ b/kernel/rcu/tree.c > @@ -2419,7 +2419,7 @@ static void rcu_do_batch(struct rcu_data *rdp) > rcu_segcblist_is_offloaded(&rdp->cblist); > struct rcu_head *rhp; > struct rcu_cblist rcl = RCU_CBLIST_INITIALIZER(rcl); > - long bl, count; > + long bl, count = 0; > long pending, tlimit = 0; > > /* If no callbacks are ready, just return. */ > @@ -2464,6 +2464,7 @@ static void rcu_do_batch(struct rcu_data *rdp) > for (; rhp; rhp = rcu_cblist_dequeue(&rcl)) { > rcu_callback_t f; > > + count++; > debug_rcu_head_unqueue(rhp); > > rcu_lock_acquire(&rcu_callback_map); > @@ -2477,9 +2478,8 @@ static void rcu_do_batch(struct rcu_data *rdp) > > /* > * Stop only if limit reached and CPU has something to do. > - * Note: The rcl structure counts down from zero. > */ > - if (-rcl.len >= bl && !offloaded && > + if (count >= bl && !offloaded && > (need_resched() || > (!is_idle_task(current) && !rcu_is_callbacks_kthread()))) > break; Please also replace "-rcl.len" in this if statement with "count": /* only call local_clock() every 32 callbacks */ if (likely((-rcl.len & 31) || local_clock() < tlimit)) continue; /* Exceeded the time limit, so leave. */ break; Yeah, it does work either way, but having "count" consistently throughout is easier on the people reading the code. And I haven't heard many complaints about the code being too easy to read... Thanx, Paul > @@ -2502,7 +2502,6 @@ static void rcu_do_batch(struct rcu_data *rdp) > > local_irq_save(flags); > rcu_nocb_lock(rdp); > - count = -rcl.len; > rdp->n_cbs_invoked += count; > trace_rcu_batch_end(rcu_state.name, count, !!rcl.head, need_resched(), > is_idle_task(current), rcu_is_callbacks_kthread()); > @@ -2510,7 +2509,7 @@ static void rcu_do_batch(struct rcu_data *rdp) > /* Update counts and requeue any remaining callbacks. */ > rcu_segcblist_insert_done_cbs(&rdp->cblist, &rcl); > smp_mb(); /* List handling before counting for rcu_barrier(). */ > - rcu_segcblist_insert_count(&rdp->cblist, &rcl); > + rcu_segcblist_add_len(&rdp->cblist, -count); > > /* Reinstate batch limit if we have worked down the excess. */ > count = rcu_segcblist_n_cbs(&rdp->cblist); > -- > 2.28.0.297.g1956fa8f8d-goog >