Received: by 2002:a05:6358:3188:b0:123:57c1:9b43 with SMTP id q8csp701944rwd; Mon, 12 Jun 2023 21:59:04 -0700 (PDT) X-Google-Smtp-Source: ACHHUZ5ccUDZEdsQmK/clBT4DSS2xMp5Oen7wVK2X+mkTNq7GIF65xC4wZwVxKecHgbc/iXiD1EX X-Received: by 2002:a17:90a:350:b0:25b:f0f7:404f with SMTP id 16-20020a17090a035000b0025bf0f7404fmr3125547pjf.28.1686632343962; Mon, 12 Jun 2023 21:59:03 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1686632343; cv=none; d=google.com; s=arc-20160816; b=U4COOv7AXOcc26oOm3XEMG66PKqAopjFlGTfvWLIAxF7dnEfOcGScEOytU5ccfNBCv GhqIVelUFwDjc+fBktnhm5SiSHZpGMRVPFU7ivAV+FOs4sD+eUm8lf+p6Kdt36TClW4R LmiS0JBTNOvpu++aYl9SQ5149dZ8b+VjgeDxDUEetGtl+ovHKL7JOk5tSRxJ6L1Q5mVI 1fYK6sV1cxnfvi1Ys6h8hrqyYTVofdX88tqWlS3kj0V5USojQ33YHtzPVlgaN8C8IOmg 60dMUkndrOYvH3poNBpqltbM8V+EWlJMLX+2pvlN4dHDQHKRTWTVuq1sGrtJSdamNReU 6ZhQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:references:in-reply-to:message-id:date:subject :cc:to:from:dkim-signature; bh=OYFeMQYBnvjBlDs1NLcV1GtJe9SX9ALefezoVgDUrzs=; b=HHRGPelf48w/0vTz7xtXroKPoZfLq/KXNinEa069l8C5P99ruaGK1bwTQQjZ2ggHKV oEGihXZ9DZEu3FeK7Zu1kEeO/arhSvMcgM92N+5ldc7noLL3KH5XZFhXw4jlI9ihbcgZ GTpLcHHNFWjEwOkc04AEnILEPPWNcT3PL6DkaBVmJ6vIg6QCcRoT3RxluPnYAk1zVDiK itcTdhLSNs6dXS9EnlNV0wSM2kb9btfdqOOZzCt4aMphUGyVqTGaziSxmNn197fm8Oig oNjsqZ9xhoS2JVRZ2ZDdb8yzeTpuoh3p6LLS6yehnJF5uLuDDp4O1rAdkfNeRaY+20gj C1mg== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@intel.com header.s=Intel header.b=EWVioQCC; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=intel.com Return-Path: Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id hg19-20020a17090b301300b00247b1a05438si4391083pjb.16.2023.06.12.21.58.52; Mon, 12 Jun 2023 21:59:03 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; dkim=pass header.i=@intel.com header.s=Intel header.b=EWVioQCC; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=pass (p=NONE sp=NONE dis=NONE) header.from=intel.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S239810AbjFMEZV (ORCPT + 99 others); Tue, 13 Jun 2023 00:25:21 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:45646 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S239825AbjFMEYt (ORCPT ); Tue, 13 Jun 2023 00:24:49 -0400 Received: from mga14.intel.com (mga14.intel.com [192.55.52.115]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 5E1172693; Mon, 12 Jun 2023 21:22:21 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=intel.com; i=@intel.com; q=dns/txt; s=Intel; t=1686630141; x=1718166141; h=from:to:cc:subject:date:message-id:in-reply-to: references; bh=40EcOZxxZvEM9rJzNTuBarR/nR+bap62+RifKT4WIh0=; b=EWVioQCCzCHgSWQrhh+A1sXBGbS2Nh213ZPWTe5WDsdN5wgFBXpdYe+O ewrEX2eqDfWO4xh+PdQqd8PSkthtQDW2ecnGhf/j/uj1ix+Ny1r8IYTjI zFote07seVZL+XdQOJIQgxdNbwSrNox8iQICdZdHgV1OKHe3y6/x1rUN6 SYk6X+iu7J07Yi/Yw+A6jcskjViQG/iFZVdtn0AkRzkCE9kfeA3PNPmsl XM7I0SJ2/kvq5Fz8BePXNWTrKsuv5fna3VyAUbGzr/qKaKkq3FyGS+lO7 uYOOZjut80D9E6mLJlZA6sQ6ekiziIYQdvDz+trzrwgrJw+ZLQsknpqzj Q==; X-IronPort-AV: E=McAfee;i="6600,9927,10739"; a="358222265" X-IronPort-AV: E=Sophos;i="6.00,238,1681196400"; d="scan'208";a="358222265" Received: from orsmga003.jf.intel.com ([10.7.209.27]) by fmsmga103.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 12 Jun 2023 21:22:01 -0700 X-ExtLoop1: 1 X-IronPort-AV: E=McAfee;i="6600,9927,10739"; a="661854995" X-IronPort-AV: E=Sophos;i="6.00,238,1681196400"; d="scan'208";a="661854995" Received: from ranerica-svr.sc.intel.com ([172.25.110.23]) by orsmga003.jf.intel.com with ESMTP; 12 Jun 2023 21:21:59 -0700 From: Ricardo Neri To: "Peter Zijlstra (Intel)" , Juri Lelli , Vincent Guittot Cc: Ricardo Neri , "Ravi V. Shankar" , Ben Segall , Daniel Bristot de Oliveira , Dietmar Eggemann , Len Brown , Mel Gorman , "Rafael J. Wysocki" , Srinivas Pandruvada , Steven Rostedt , Tim Chen , Valentin Schneider , Lukasz Luba , Ionela Voinescu , Zhao Liu , "Yuan, Perry" , x86@kernel.org, "Joel Fernandes (Google)" , linux-kernel@vger.kernel.org, linux-pm@vger.kernel.org, Ricardo Neri , "Tim C . Chen" , Zhao Liu Subject: [PATCH v4 20/24] sched/fair: Introduce sched_smt_siblings_idle() Date: Mon, 12 Jun 2023 21:24:18 -0700 Message-Id: <20230613042422.5344-21-ricardo.neri-calderon@linux.intel.com> X-Mailer: git-send-email 2.17.1 In-Reply-To: <20230613042422.5344-1-ricardo.neri-calderon@linux.intel.com> References: <20230613042422.5344-1-ricardo.neri-calderon@linux.intel.com> X-Spam-Status: No, score=-4.3 required=5.0 tests=BAYES_00,DKIMWL_WL_HIGH, DKIM_SIGNED,DKIM_VALID,DKIM_VALID_EF,RCVD_IN_DNSWL_MED,SPF_HELO_NONE, SPF_NONE,T_SCC_BODY_TEXT_LINE,URIBL_BLOCKED autolearn=ham autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org X86 needs to know the idle state of the SMT siblings of a CPU to improve the accuracy of IPCC classification. X86 implements support for IPC classes in the thermal HFI driver. Rename is_core_idle() as sched_smt_siblings_idle() and make it available outside the scheduler code. Cc: Ben Segall Cc: Daniel Bristot de Oliveira Cc: Dietmar Eggemann Cc: Len Brown Cc: Mel Gorman Cc: Perry Yuan Cc: Rafael J. Wysocki Cc: Srinivas Pandruvada Cc: Steven Rostedt Cc: Tim C. Chen Cc: Valentin Schneider Cc: Zhao Liu Cc: x86@kernel.org Cc: linux-kernel@vger.kernel.org Signed-off-by: Ricardo Neri --- is_core_idle() is no longer an inline function after this patch. To rule out performance degradation, I compared the execution time of the inline and non-inline versions on a 4-socket Cascade Lake system using the NUMA stressor of stress-ng. I used this test command: $ stress-ng --numa 1500 -t 10m During the test, is_core_idle() was called ~200,000 times. To measure the execution time, I recorded the value of the TSC counter before and after calling is_core_idle() and calculated the difference. value. I arbitrarily removed outliers (defined as any delta larger than 5000 counts). This required removing ~40 samples. The table below summarizes the difference in execution time. All values are expressed in TSC counts, except for the standard deviation, expressed as a percentage of the average. Average Median Std(%) Mode TSCdelta inline 668.76 626 67.24 42 TSCdelta non-inline 677.64 624 67.67 46 The metrics show that both the inline and non-inline versions exhibit similar performance characteristics. --- Changes since v3: * None Changes since v2: * Brought back this previously dropped patch. * Profiled inline vs non-inline is_core_idle(). I found not major penalty. * Merged is_core_idle() and sched_smt_siblings_idle() into a single function. (Dietmar) Changes since v1: * Dropped this patch. --- include/linux/sched.h | 2 ++ kernel/sched/fair.c | 13 ++++++++++--- 2 files changed, 12 insertions(+), 3 deletions(-) diff --git a/include/linux/sched.h b/include/linux/sched.h index 719147460ca8..986344ebf2f6 100644 --- a/include/linux/sched.h +++ b/include/linux/sched.h @@ -2461,4 +2461,6 @@ static inline void sched_core_fork(struct task_struct *p) { } extern void sched_set_stop_task(int cpu, struct task_struct *stop); +extern bool sched_smt_siblings_idle(int cpu); + #endif diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c index da3e009eef42..a52589a6c561 100644 --- a/kernel/sched/fair.c +++ b/kernel/sched/fair.c @@ -1064,7 +1064,14 @@ update_stats_curr_start(struct cfs_rq *cfs_rq, struct sched_entity *se) * Scheduling class queueing methods: */ -static inline bool is_core_idle(int cpu) +/** + * sched_smt_siblings_idle - Check whether SMT siblings of a CPU are idle + * @cpu: The CPU to check + * + * Returns true if all the SMT siblings of @cpu are idle or @cpu does not have + * SMT siblings. The idle state of @cpu is not considered. + */ +bool sched_smt_siblings_idle(int cpu) { #ifdef CONFIG_SCHED_SMT int sibling; @@ -1767,7 +1774,7 @@ static inline int numa_idle_core(int idle_core, int cpu) * Prefer cores instead of packing HT siblings * and triggering future load balancing. */ - if (is_core_idle(cpu)) + if (sched_smt_siblings_idle(cpu)) idle_core = cpu; return idle_core; @@ -9652,7 +9659,7 @@ static bool sched_use_asym_prio(struct sched_domain *sd, int cpu) if (!sched_smt_active()) return true; - return sd->flags & SD_SHARE_CPUCAPACITY || is_core_idle(cpu); + return sd->flags & SD_SHARE_CPUCAPACITY || sched_smt_siblings_idle(cpu); } /** -- 2.25.1