Received: by 2002:a05:6358:11c7:b0:104:8066:f915 with SMTP id i7csp4045873rwl; Tue, 28 Mar 2023 01:36:22 -0700 (PDT) X-Google-Smtp-Source: AKy350ZyXyca4SLMarEbeK1kJoMVaHMLaRQf6UJ92n/kJepnqCxYkeOcXeK/bjkrUa52tOAyS15r X-Received: by 2002:a17:903:1c2:b0:1a2:22e4:bbb2 with SMTP id e2-20020a17090301c200b001a222e4bbb2mr17520126plh.58.1679992582110; Tue, 28 Mar 2023 01:36:22 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1679992582; cv=none; d=google.com; s=arc-20160816; b=YiSnT/xxYojS9kEb0la6i+n74KXnLbl2kcmA2PITKHZqlCQFJBo5VJ5z7c4lLYqlxA LPfYRwWIFTYbm7qv+OoH8QUe4e2hCR2SwR3yWomuQqo4RAnCHmE0lpd0m52UORNACGiY JY/3zTqIGA8Se6ALZy0ef1i426eLRPoWl66bsi+okVpakaZCQUtZxksDsfk/TOSKGPHe T1vAAhb108KcSgiENXkGiZpx6s6WY3VNeFnrYUdw/j56+zJhh7d6OBWBCmRW/g9lGcSY Fb/A7fIzHv9OjJ11Xvb4KBoI0Mw0y9ad/26CFbCTXnKBLSw2zDvE+FTL4eRKxCVAlpn+ feHg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:robot-unsubscribe :robot-id:message-id:mime-version:references:in-reply-to:cc:subject :to:reply-to:sender:from:dkim-signature:dkim-signature:date; bh=1tUILXd7hnx0zQEHBSwqkcnBfj7Vwkx3YWbcPQ6UQd8=; b=bf2TAupbRyBvsgOerC5uaryLEqMg5OXUAz5RGKi6kYKgvOMMkWwugbPK36ugUF/AvZ 4Ihka34x5RT5TSGmok2j4rshGd087ROIYAWI7snJ/e+c2aHkSgv9X/GT1j63Wv2j/XgV VJLcHoIyQiIA+Aj9F8YUsqALxQvqyaTTx4yRZ2nlsGSb0w/jHThmwdlBEaFkC3+nA7hG rrOiIsKRIHUNSNPfhFqQoIYd24L0N5JHDMj6Rmk7Q5/xt1fV/hqFLT89r7/bjKzaNpxQ AQCpK0ZVI5N71srEwjkPB0EaKH8gpf1Efg8R3XWSNozlZS9C5lCMWSHGmv6QTaaphB2H DElA== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@linutronix.de header.s=2020 header.b=BK8AqibD; dkim=neutral (no key) header.i=@linutronix.de; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=linutronix.de Return-Path: Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id e5-20020a17090301c500b0019e4154578esi31063756plh.76.2023.03.28.01.36.10; Tue, 28 Mar 2023 01:36:22 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; dkim=pass header.i=@linutronix.de header.s=2020 header.b=BK8AqibD; dkim=neutral (no key) header.i=@linutronix.de; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=QUARANTINE dis=NONE) header.from=linutronix.de Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S232372AbjC1Iew (ORCPT + 99 others); Tue, 28 Mar 2023 04:34:52 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:51736 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S230155AbjC1Iet (ORCPT ); Tue, 28 Mar 2023 04:34:49 -0400 Received: from galois.linutronix.de (Galois.linutronix.de [193.142.43.55]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 896143C33; Tue, 28 Mar 2023 01:34:47 -0700 (PDT) Date: Tue, 28 Mar 2023 08:34:44 -0000 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linutronix.de; s=2020; t=1679992485; h=from:from:sender:sender:reply-to:reply-to:subject:subject:date:date: message-id:message-id:to:to:cc:cc:mime-version:mime-version: content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=1tUILXd7hnx0zQEHBSwqkcnBfj7Vwkx3YWbcPQ6UQd8=; b=BK8AqibDR3CgHdcAQO9RiHJowcyabFUcLVqAV+FEKtGqg8hC25OrSnWhwPxLkkKpikiRAT jusO145RpzhO4OftC9+WI3wSF3oJ6yA52bfQCnQy02okkaszZFm4MYYICMw88weq3usdle 5YxEGveikii5FuVLCTY4WMlwPAv4hkOg/fsXWdDdIK6qhr0oirgTcDwN5ZKNaMNQL9+8Vy eIPuWStpvF8SwDFeawIlYNWr0TN5uwAwbAd2GWYVBQq2e4SDgLewcjARG5EHgDgxW6MzGE FudiRtcXaJIvASRIpGb11FhAnNVSE2d/WWfRMI6+FU6Sa/HmssOH+Bf74ar24Q== DKIM-Signature: v=1; a=ed25519-sha256; c=relaxed/relaxed; d=linutronix.de; s=2020e; t=1679992485; h=from:from:sender:sender:reply-to:reply-to:subject:subject:date:date: message-id:message-id:to:to:cc:cc:mime-version:mime-version: content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=1tUILXd7hnx0zQEHBSwqkcnBfj7Vwkx3YWbcPQ6UQd8=; b=d1mbeOjiMFFtpl/SCGEPp43CpTbjLiSEVUSne8uyqHxkqPBcs1RO7r9dhUN7DD8sTLBvxf 6N+y8r1RbSuGDcCg== From: "tip-bot2 for Valentin Schneider" Sender: tip-bot2@linutronix.de Reply-to: linux-kernel@vger.kernel.org To: linux-tip-commits@vger.kernel.org Subject: [tip: smp/core] sched, smp: Trace smp callback causing an IPI Cc: Valentin Schneider , "Peter Zijlstra (Intel)" , x86@kernel.org, linux-kernel@vger.kernel.org In-Reply-To: <20230307143558.294354-8-vschneid@redhat.com> References: <20230307143558.294354-8-vschneid@redhat.com> MIME-Version: 1.0 Message-ID: <167999248462.5837.17270622315487282939.tip-bot2@tip-bot2> Robot-ID: Robot-Unsubscribe: Contact to get blacklisted from these emails Content-Type: text/plain; charset="utf-8" Content-Transfer-Encoding: 7bit X-Spam-Status: No, score=-2.5 required=5.0 tests=DKIM_SIGNED,DKIM_VALID, DKIM_VALID_AU,DKIM_VALID_EF,RCVD_IN_DNSWL_MED,SPF_HELO_NONE,SPF_PASS autolearn=unavailable autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org The following commit has been merged into the smp/core branch of tip: Commit-ID: 68f4ff04dbada18dad79659c266a8e5e29e458cd Gitweb: https://git.kernel.org/tip/68f4ff04dbada18dad79659c266a8e5e29e458cd Author: Valentin Schneider AuthorDate: Tue, 07 Mar 2023 14:35:58 Committer: Peter Zijlstra CommitterDate: Fri, 24 Mar 2023 11:01:29 +01:00 sched, smp: Trace smp callback causing an IPI Context ======= The newly-introduced ipi_send_cpumask tracepoint has a "callback" parameter which so far has only been fed with NULL. While CSD_TYPE_SYNC/ASYNC and CSD_TYPE_IRQ_WORK share a similar backing struct layout (meaning their callback func can be accessed without caring about the actual CSD type), CSD_TYPE_TTWU doesn't even have a function attached to its struct. This means we need to check the type of a CSD before eventually dereferencing its associated callback. This isn't as trivial as it sounds: the CSD type is stored in __call_single_node.u_flags, which get cleared right before the callback is executed via csd_unlock(). This implies checking the CSD type before it is enqueued on the call_single_queue, as the target CPU's queue can be flushed before we get to sending an IPI. Furthermore, send_call_function_single_ipi() only has a CPU parameter, and would need to have an additional argument to trickle down the invoked function. This is somewhat silly, as the extra argument will always be pushed down to the function even when nothing is being traced, which is unnecessary overhead. Changes ======= send_call_function_single_ipi() is only used by smp.c, and is defined in sched/core.c as it contains scheduler-specific ops (set_nr_if_polling() of a CPU's idle task). Split it into two parts: the scheduler bits remain in sched/core.c, and the actual IPI emission is moved into smp.c. This lets us define an __always_inline helper function that can take the related callback as parameter without creating useless register pressure in the non-traced path which only gains a (disabled) static branch. Do the same thing for the multi IPI case. Signed-off-by: Valentin Schneider Signed-off-by: Peter Zijlstra (Intel) Link: https://lore.kernel.org/r/20230307143558.294354-8-vschneid@redhat.com --- kernel/sched/core.c | 18 +++++++++------- kernel/sched/smp.h | 2 +- kernel/smp.c | 49 ++++++++++++++++++++++++++++++++++++-------- 3 files changed, 53 insertions(+), 16 deletions(-) diff --git a/kernel/sched/core.c b/kernel/sched/core.c index c26a2cd..b0a48cf 100644 --- a/kernel/sched/core.c +++ b/kernel/sched/core.c @@ -3829,16 +3829,20 @@ void sched_ttwu_pending(void *arg) rq_unlock_irqrestore(rq, &rf); } -void send_call_function_single_ipi(int cpu) +/* + * Prepare the scene for sending an IPI for a remote smp_call + * + * Returns true if the caller can proceed with sending the IPI. + * Returns false otherwise. + */ +bool call_function_single_prep_ipi(int cpu) { - struct rq *rq = cpu_rq(cpu); - - if (!set_nr_if_polling(rq->idle)) { - trace_ipi_send_cpumask(cpumask_of(cpu), _RET_IP_, NULL); - arch_send_call_function_single_ipi(cpu); - } else { + if (set_nr_if_polling(cpu_rq(cpu)->idle)) { trace_sched_wake_idle_without_ipi(cpu); + return false; } + + return true; } /* diff --git a/kernel/sched/smp.h b/kernel/sched/smp.h index 2eb23dd..21ac444 100644 --- a/kernel/sched/smp.h +++ b/kernel/sched/smp.h @@ -6,7 +6,7 @@ extern void sched_ttwu_pending(void *arg); -extern void send_call_function_single_ipi(int cpu); +extern bool call_function_single_prep_ipi(int cpu); #ifdef CONFIG_SMP extern void flush_smp_call_function_queue(void); diff --git a/kernel/smp.c b/kernel/smp.c index 6bbfabb..37e9613 100644 --- a/kernel/smp.c +++ b/kernel/smp.c @@ -104,9 +104,18 @@ void __init call_function_init(void) } static __always_inline void -send_call_function_ipi_mask(struct cpumask *mask) +send_call_function_single_ipi(int cpu, smp_call_func_t func) { - trace_ipi_send_cpumask(mask, _RET_IP_, NULL); + if (call_function_single_prep_ipi(cpu)) { + trace_ipi_send_cpumask(cpumask_of(cpu), _RET_IP_, func); + arch_send_call_function_single_ipi(cpu); + } +} + +static __always_inline void +send_call_function_ipi_mask(struct cpumask *mask, smp_call_func_t func) +{ + trace_ipi_send_cpumask(mask, _RET_IP_, func); arch_send_call_function_ipi_mask(mask); } @@ -307,9 +316,8 @@ static __always_inline void csd_unlock(struct __call_single_data *csd) smp_store_release(&csd->node.u_flags, 0); } -static DEFINE_PER_CPU_SHARED_ALIGNED(call_single_data_t, csd_data); - -void __smp_call_single_queue(int cpu, struct llist_node *node) +static __always_inline void +raw_smp_call_single_queue(int cpu, struct llist_node *node, smp_call_func_t func) { /* * The list addition should be visible to the target CPU when it pops @@ -324,7 +332,32 @@ void __smp_call_single_queue(int cpu, struct llist_node *node) * equipped to do the right thing... */ if (llist_add(node, &per_cpu(call_single_queue, cpu))) - send_call_function_single_ipi(cpu); + send_call_function_single_ipi(cpu, func); +} + +static DEFINE_PER_CPU_SHARED_ALIGNED(call_single_data_t, csd_data); + +void __smp_call_single_queue(int cpu, struct llist_node *node) +{ + /* + * We have to check the type of the CSD before queueing it, because + * once queued it can have its flags cleared by + * flush_smp_call_function_queue() + * even if we haven't sent the smp_call IPI yet (e.g. the stopper + * executes migration_cpu_stop() on the remote CPU). + */ + if (trace_ipi_send_cpumask_enabled()) { + call_single_data_t *csd; + smp_call_func_t func; + + csd = container_of(node, call_single_data_t, node.llist); + func = CSD_TYPE(csd) == CSD_TYPE_TTWU ? + sched_ttwu_pending : csd->func; + + raw_smp_call_single_queue(cpu, node, func); + } else { + raw_smp_call_single_queue(cpu, node, NULL); + } } /* @@ -768,9 +801,9 @@ static void smp_call_function_many_cond(const struct cpumask *mask, * provided mask. */ if (nr_cpus == 1) - send_call_function_single_ipi(last_cpu); + send_call_function_single_ipi(last_cpu, func); else if (likely(nr_cpus > 1)) - send_call_function_ipi_mask(cfd->cpumask_ipi); + send_call_function_ipi_mask(cfd->cpumask_ipi, func); } if (run_local && (!cond_func || cond_func(this_cpu, info))) {