Received: by 2002:a05:7412:98c1:b0:fa:551:50a7 with SMTP id kc1csp408981rdb; Fri, 5 Jan 2024 14:20:48 -0800 (PST) X-Google-Smtp-Source: AGHT+IEV15AM7riky8s3m0usA8KTbqUE5vsHd0xBmNpuvrTT/uujjlRU1CswdAE/4UXoR71gqJlP X-Received: by 2002:a05:620a:5301:b0:77f:3660:6035 with SMTP id oo1-20020a05620a530100b0077f36606035mr67221qkn.138.1704493248651; Fri, 05 Jan 2024 14:20:48 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1704493248; cv=none; d=google.com; s=arc-20160816; b=0lDGBN8wcJN0kZwNPwFWAfaQThSFLj9/5tXgiFKkzp3ZX90gr+9GrXKTTcgTKCu1Y7 6kzAKkP/l9v+mQZGDorEtQ+SIw1oe0GCFVxIGTfEg5V0ustq+7cQq5iYqXSeeoOJ8S4f iz036j3u1R79SexphxTwI0PbYT04Pw89zcKP1+5OR6dzLCGH0gJKx/RMP4VqdgGFhFLP MUgBDoJ6n3o6OusJKeuXM7dbgUydxAapKAHnSEDbR52u3VBH7zVPClod2PMPCJqyXxxX IO+Bj3PGy0IHPXlm8mVq83N/u28YSKNT8sgHgO6IlfGDjmsV3+TveyAgBKtNvoK6fyrU DRqw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=content-transfer-encoding:mime-version:list-unsubscribe :list-subscribe:list-id:precedence:references:in-reply-to:message-id :date:subject:cc:to:from:dkim-signature; bh=higS0mdLu/MLGgnWkn9tKbyrxY+yKhygKMHgrEHwEkU=; fh=LFahd4jziOLZv88JjnLsQAu+Hg6/pa5kYd3YC5wdhMk=; b=viylpbSkGhgeXoOl7bOjilvPPXbY+Cf866pTpAY06XrMgLjkMfmcfG4Xx/hs+KCjSZ t2iK9ZFVHDeoBmzj/zbgDgyqUp+kRLjd9ygx6xoDdxSEZKiEUDg0R8M86e1FXViil2ny M7rt5ZiYCNdS8FvcKTA8Ka18kpvm5EVCClw5cRazH5htGlxlpuZAxYcuTX/7FIJ/l/t3 u4UdRGq7VYAygz+vsULoIdlA/AQ2ayRKN8kwlGlVdk5yR/hop+DFcXR6Sq4sglzGaNcE Vnzi9nMSWFoZ2R9f8+JjaVECXyeZKOcA9Pdgs/5iXODopfhVaapCA+A/UWPLycrtG4zM 45eg== ARC-Authentication-Results: i=1; mx.google.com; dkim=pass header.i=@layalina-io.20230601.gappssmtp.com header.s=20230601 header.b=NmB9y3xC; spf=pass (google.com: domain of linux-kernel+bounces-18395-linux.lists.archive=gmail.com@vger.kernel.org designates 147.75.199.223 as permitted sender) smtp.mailfrom="linux-kernel+bounces-18395-linux.lists.archive=gmail.com@vger.kernel.org" Return-Path: Received: from ny.mirrors.kernel.org (ny.mirrors.kernel.org. [147.75.199.223]) by mx.google.com with ESMTPS id p13-20020a05620a22ed00b00783067773c2si2482261qki.545.2024.01.05.14.20.48 for (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Fri, 05 Jan 2024 14:20:48 -0800 (PST) Received-SPF: pass (google.com: domain of linux-kernel+bounces-18395-linux.lists.archive=gmail.com@vger.kernel.org designates 147.75.199.223 as permitted sender) client-ip=147.75.199.223; Authentication-Results: mx.google.com; dkim=pass header.i=@layalina-io.20230601.gappssmtp.com header.s=20230601 header.b=NmB9y3xC; spf=pass (google.com: domain of linux-kernel+bounces-18395-linux.lists.archive=gmail.com@vger.kernel.org designates 147.75.199.223 as permitted sender) smtp.mailfrom="linux-kernel+bounces-18395-linux.lists.archive=gmail.com@vger.kernel.org" Received: from smtp.subspace.kernel.org (wormhole.subspace.kernel.org [52.25.139.140]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by ny.mirrors.kernel.org (Postfix) with ESMTPS id 5A4D51C22EED for ; Fri, 5 Jan 2024 22:20:48 +0000 (UTC) Received: from localhost.localdomain (localhost.localdomain [127.0.0.1]) by smtp.subspace.kernel.org (Postfix) with ESMTP id 257D1360BB; Fri, 5 Jan 2024 22:20:30 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; dkim=pass (2048-bit key) header.d=layalina-io.20230601.gappssmtp.com header.i=@layalina-io.20230601.gappssmtp.com header.b="NmB9y3xC" X-Original-To: linux-kernel@vger.kernel.org Received: from mail-wm1-f45.google.com (mail-wm1-f45.google.com [209.85.128.45]) (using TLSv1.2 with cipher ECDHE-RSA-AES128-GCM-SHA256 (128/128 bits)) (No client certificate requested) by smtp.subspace.kernel.org (Postfix) with ESMTPS id 7A8F6358B3 for ; Fri, 5 Jan 2024 22:20:27 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; dmarc=none (p=none dis=none) header.from=layalina.io Authentication-Results: smtp.subspace.kernel.org; spf=pass smtp.mailfrom=layalina.io Received: by mail-wm1-f45.google.com with SMTP id 5b1f17b1804b1-40d4a7f0c4dso415215e9.1 for ; Fri, 05 Jan 2024 14:20:27 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=layalina-io.20230601.gappssmtp.com; s=20230601; t=1704493225; x=1705098025; darn=vger.kernel.org; h=content-transfer-encoding:mime-version:references:in-reply-to :message-id:date:subject:cc:to:from:from:to:cc:subject:date :message-id:reply-to; bh=higS0mdLu/MLGgnWkn9tKbyrxY+yKhygKMHgrEHwEkU=; b=NmB9y3xCKEGAzpsDhB9FJUddnndkWWZFWb9QI3f02tpua28nFfPzuF9JWetmVrFPmP BnS1YfeivSju5m3DqrTgahLoKsei6IjAoKo0AJPQdFxryKXT0F8z2ux1JCJhIl/P/VGr +x6QWJ1cNeHK618WqwzGtTLzNShEH7bpMAXTbAdERE5nuQNpSTqvARslbISVU3V0qpYg ymj/iMloLLg/XkckP3rfPDPytR5SJRQaHFNqrTMK6zrHicfvzk7A+u4F4PqBzKoZ4Nlt xt/buNfzYKrqYNS0ueiHY/umMZMXlSCdRA/ETGKAJaFO0y8uj+MRWLBrk5oc1224IR7F En2A== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20230601; t=1704493225; x=1705098025; h=content-transfer-encoding:mime-version:references:in-reply-to :message-id:date:subject:cc:to:from:x-gm-message-state:from:to:cc :subject:date:message-id:reply-to; bh=higS0mdLu/MLGgnWkn9tKbyrxY+yKhygKMHgrEHwEkU=; b=ohQ8fMh9ecihNqcGZdwK7K1BTnCWIg5oHB/s+Ldun2gS/Qd1ItmpUo3H8XYogsEF2y IGs7tA0KZ6g3zTnJ10XbCSsgzLVUuV9keRPr4hKPQaaNdACsop5e7u2PK0uXSAkvcZ7y AH1PEQYQUYsT4hzBw8M060aTZhi7ZWNCuClYYl/yf/qDKcqoXC1u6hlNbGeFKdVR9IVY i29HHN0zpAJVJC1vApNrL36iuocI4RQYVsMSIYjnZKwJJiauqP+jFKQImEnhDg7rn8dT 3Fh2UiJcnHzElVrQkyZD7t5TibK2q9NngpuI1HnF0kvp2zyi3DvLWWmzW6W6i9j8olyx DkXw== X-Gm-Message-State: AOJu0YxtPJAEgj0vnveh+bSHNqxxLzKGfAZRCmTqWiikgkRDxDLBRkQ5 ZiihdNqjJYjqlYpSJU06bQVl9E+izY8Ucg== X-Received: by 2002:a05:600c:a686:b0:40b:5e59:c570 with SMTP id ip6-20020a05600ca68600b0040b5e59c570mr79665wmb.154.1704493224711; Fri, 05 Jan 2024 14:20:24 -0800 (PST) Received: from airbuntu.. (host109-154-238-212.range109-154.btcentralplus.com. [109.154.238.212]) by smtp.gmail.com with ESMTPSA id q13-20020a05600c46cd00b0040d604dea3bsm2799022wmo.4.2024.01.05.14.20.24 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Fri, 05 Jan 2024 14:20:24 -0800 (PST) From: Qais Yousef To: Ingo Molnar , Peter Zijlstra , Vincent Guittot , Dietmar Eggemann Cc: linux-kernel@vger.kernel.org, "Pierre Gondois" , Qais Yousef Subject: [PATCH v4 1/2] sched/fair: Check a task has a fitting cpu when updating misfit Date: Fri, 5 Jan 2024 22:20:13 +0000 Message-Id: <20240105222014.1025040-2-qyousef@layalina.io> X-Mailer: git-send-email 2.34.1 In-Reply-To: <20240105222014.1025040-1-qyousef@layalina.io> References: <20240105222014.1025040-1-qyousef@layalina.io> Precedence: bulk X-Mailing-List: linux-kernel@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 Content-Transfer-Encoding: 8bit From: Qais Yousef If a misfit task is affined to a subset of the possible cpus, we need to verify that one of these cpus can fit it. Otherwise the load balancer code will continuously trigger needlessly leading the balance_interval to increase in return and eventually end up with a situation where real imbalances take a long time to address because of this impossible imbalance situation. This can happen in Android world where it's common for background tasks to be restricted to little cores. Similarly if we can't fit the biggest core, triggering misfit is pointless as it is the best we can ever get on this system. To be able to detect that; we use asym_cap_list to iterate through capacities in the system to see if the task is able to run at a higher capacity level based on its p->cpus_ptr. To do so safely, we convert the list to be RCU protected. To be able to iterate through capacity levels, export asym_cap_list to allow for fast traversal of all available capacity levels in the system. Test: ===== Add trace_printk("balance_interval = %lu\n", interval) in get_sd_balance_interval(). run if [ "$MASK" != "0" ]; then adb shell "taskset -a $MASK cat /dev/zero > /dev/null" fi sleep 10 // parse ftrace buffer counting the occurrence of each valaue Where MASK is either: * 0: no busy task running * 1: busy task is pinned to 1 cpu; handled today to not cause misfit * f: busy task pinned to little cores, simulates busy background task, demonstrates the problem to be fixed Results: ======== Note how occurrence of balance_interval = 128 overshoots for MASK = f. BEFORE ------ MASK=0 1 balance_interval = 175 120 balance_interval = 128 846 balance_interval = 64 55 balance_interval = 63 215 balance_interval = 32 2 balance_interval = 31 2 balance_interval = 16 4 balance_interval = 8 1870 balance_interval = 4 65 balance_interval = 2 MASK=1 27 balance_interval = 175 37 balance_interval = 127 840 balance_interval = 64 167 balance_interval = 63 449 balance_interval = 32 84 balance_interval = 31 304 balance_interval = 16 1156 balance_interval = 8 2781 balance_interval = 4 428 balance_interval = 2 MASK=f 1 balance_interval = 175 1328 balance_interval = 128 44 balance_interval = 64 101 balance_interval = 63 25 balance_interval = 32 5 balance_interval = 31 23 balance_interval = 16 23 balance_interval = 8 4306 balance_interval = 4 177 balance_interval = 2 AFTER ----- Note how the high values almost disappear for all MASK values. The system has background tasks that could trigger the problem without simulate it even with MASK=0. MASK=0 103 balance_interval = 63 19 balance_interval = 31 194 balance_interval = 8 4827 balance_interval = 4 179 balance_interval = 2 MASK=1 131 balance_interval = 63 1 balance_interval = 31 87 balance_interval = 8 3600 balance_interval = 4 7 balance_interval = 2 MASK=f 8 balance_interval = 127 182 balance_interval = 63 3 balance_interval = 31 9 balance_interval = 16 415 balance_interval = 8 3415 balance_interval = 4 21 balance_interval = 2 Signed-off-by: Qais Yousef Signed-off-by: Qais Yousef (Google) --- kernel/sched/fair.c | 65 ++++++++++++++++++++++++++++++++++------- kernel/sched/sched.h | 14 +++++++++ kernel/sched/topology.c | 29 ++++++++---------- 3 files changed, 81 insertions(+), 27 deletions(-) diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c index bcea3d55d95d..0830ceb7ca07 100644 --- a/kernel/sched/fair.c +++ b/kernel/sched/fair.c @@ -5065,17 +5065,61 @@ static inline int task_fits_cpu(struct task_struct *p, int cpu) static inline void update_misfit_status(struct task_struct *p, struct rq *rq) { + unsigned long uclamp_min, uclamp_max; + unsigned long util, cpu_cap; + int cpu = cpu_of(rq); + if (!sched_asym_cpucap_active()) return; - if (!p || p->nr_cpus_allowed == 1) { - rq->misfit_task_load = 0; - return; - } + if (!p || p->nr_cpus_allowed == 1) + goto out; - if (task_fits_cpu(p, cpu_of(rq))) { - rq->misfit_task_load = 0; - return; + cpu_cap = arch_scale_cpu_capacity(cpu); + + /* If we can't fit the biggest CPU, that's the best we can ever get. */ + if (cpu_cap == SCHED_CAPACITY_SCALE) + goto out; + + uclamp_min = uclamp_eff_value(p, UCLAMP_MIN); + uclamp_max = uclamp_eff_value(p, UCLAMP_MAX); + util = task_util_est(p); + + if (util_fits_cpu(util, uclamp_min, uclamp_max, cpu) > 0) + goto out; + + /* + * If the task affinity is not set to default, make sure it is not + * restricted to a subset where no CPU can ever fit it. Triggering + * misfit in this case is pointless as it has no where better to move + * to. And it can lead to balance_interval to grow too high as we'll + * continuously fail to move it anywhere. + */ + if (!cpumask_equal(p->cpus_ptr, cpu_possible_mask)) { + unsigned long clamped_util = clamp(util, uclamp_min, uclamp_max); + bool has_fitting_cpu = false; + struct asym_cap_data *entry; + + rcu_read_lock(); + list_for_each_entry_rcu(entry, &asym_cap_list, link) { + if (entry->capacity > cpu_cap) { + cpumask_t *cpumask; + + if (clamped_util > entry->capacity) + continue; + + cpumask = cpu_capacity_span(entry); + if (!cpumask_intersects(p->cpus_ptr, cpumask)) + continue; + + has_fitting_cpu = true; + break; + } + } + rcu_read_unlock(); + + if (!has_fitting_cpu) + goto out; } /* @@ -5083,6 +5127,9 @@ static inline void update_misfit_status(struct task_struct *p, struct rq *rq) * task_h_load() returns 0. */ rq->misfit_task_load = max_t(unsigned long, task_h_load(p), 1); + return; +out: + rq->misfit_task_load = 0; } #else /* CONFIG_SMP */ @@ -9583,9 +9630,7 @@ check_cpu_capacity(struct rq *rq, struct sched_domain *sd) */ static inline int check_misfit_status(struct rq *rq, struct sched_domain *sd) { - return rq->misfit_task_load && - (arch_scale_cpu_capacity(rq->cpu) < rq->rd->max_cpu_capacity || - check_cpu_capacity(rq, sd)); + return rq->misfit_task_load && check_cpu_capacity(rq, sd); } /* diff --git a/kernel/sched/sched.h b/kernel/sched/sched.h index e58a54bda77d..a653017a1b9b 100644 --- a/kernel/sched/sched.h +++ b/kernel/sched/sched.h @@ -109,6 +109,20 @@ extern int sysctl_sched_rt_period; extern int sysctl_sched_rt_runtime; extern int sched_rr_timeslice; +/* + * Asymmetric CPU capacity bits + */ +struct asym_cap_data { + struct list_head link; + struct rcu_head rcu; + unsigned long capacity; + unsigned long cpus[]; +}; + +extern struct list_head asym_cap_list; + +#define cpu_capacity_span(asym_data) to_cpumask((asym_data)->cpus) + /* * Helpers for converting nanosecond timing to jiffy resolution */ diff --git a/kernel/sched/topology.c b/kernel/sched/topology.c index 10d1391e7416..ba4a0b18ae25 100644 --- a/kernel/sched/topology.c +++ b/kernel/sched/topology.c @@ -1329,24 +1329,13 @@ static void init_sched_groups_capacity(int cpu, struct sched_domain *sd) update_group_capacity(sd, cpu); } -/* - * Asymmetric CPU capacity bits - */ -struct asym_cap_data { - struct list_head link; - unsigned long capacity; - unsigned long cpus[]; -}; - /* * Set of available CPUs grouped by their corresponding capacities * Each list entry contains a CPU mask reflecting CPUs that share the same * capacity. * The lifespan of data is unlimited. */ -static LIST_HEAD(asym_cap_list); - -#define cpu_capacity_span(asym_data) to_cpumask((asym_data)->cpus) +LIST_HEAD(asym_cap_list); /* * Verify whether there is any CPU capacity asymmetry in a given sched domain. @@ -1386,6 +1375,12 @@ asym_cpu_capacity_classify(const struct cpumask *sd_span, } +static void free_asym_cap_entry(struct rcu_head *head) +{ + struct asym_cap_data *entry = container_of(head, struct asym_cap_data, rcu); + kfree(entry); +} + static inline void asym_cpu_capacity_update_data(int cpu) { unsigned long capacity = arch_scale_cpu_capacity(cpu); @@ -1400,7 +1395,7 @@ static inline void asym_cpu_capacity_update_data(int cpu) if (WARN_ONCE(!entry, "Failed to allocate memory for asymmetry data\n")) return; entry->capacity = capacity; - list_add(&entry->link, &asym_cap_list); + list_add_rcu(&entry->link, &asym_cap_list); done: __cpumask_set_cpu(cpu, cpu_capacity_span(entry)); } @@ -1423,8 +1418,8 @@ static void asym_cpu_capacity_scan(void) list_for_each_entry_safe(entry, next, &asym_cap_list, link) { if (cpumask_empty(cpu_capacity_span(entry))) { - list_del(&entry->link); - kfree(entry); + list_del_rcu(&entry->link); + call_rcu(&entry->rcu, free_asym_cap_entry); } } @@ -1434,8 +1429,8 @@ static void asym_cpu_capacity_scan(void) */ if (list_is_singular(&asym_cap_list)) { entry = list_first_entry(&asym_cap_list, typeof(*entry), link); - list_del(&entry->link); - kfree(entry); + list_del_rcu(&entry->link); + call_rcu(&entry->rcu, free_asym_cap_entry); } } -- 2.34.1