Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by smtp.lore.kernel.org (Postfix) with ESMTP id 8B386C433F5 for ; Thu, 9 Dec 2021 14:23:46 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S238526AbhLIO1S (ORCPT ); Thu, 9 Dec 2021 09:27:18 -0500 Received: from foss.arm.com ([217.140.110.172]:57458 "EHLO foss.arm.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S232691AbhLIO1R (ORCPT ); Thu, 9 Dec 2021 09:27:17 -0500 Received: from usa-sjc-imap-foss1.foss.arm.com (unknown [10.121.207.14]) by usa-sjc-mx-foss1.foss.arm.com (Postfix) with ESMTP id D881D1FB; Thu, 9 Dec 2021 06:23:43 -0800 (PST) Received: from e113632-lin (e113632-lin.cambridge.arm.com [10.1.196.57]) by usa-sjc-imap-foss1.foss.arm.com (Postfix) with ESMTPSA id 8667B3F73B; Thu, 9 Dec 2021 06:23:42 -0800 (PST) From: Valentin Schneider To: Mel Gorman , Peter Zijlstra Cc: Ingo Molnar , Vincent Guittot , Aubrey Li , Barry Song , Mike Galbraith , Srikar Dronamraju , "Gautham R. Shenoy" , LKML Subject: Re: [PATCH 2/2] sched/fair: Adjust the allowed NUMA imbalance when SD_NUMA spans multiple LLCs In-Reply-To: <20211206151206.GH3366@techsingularity.net> References: <20211201151844.20488-1-mgorman@techsingularity.net> <20211201151844.20488-3-mgorman@techsingularity.net> <20211204104056.GR16608@worktop.programming.kicks-ass.net> <20211206151206.GH3366@techsingularity.net> Date: Thu, 09 Dec 2021 14:23:40 +0000 Message-ID: <87y24t97rn.mognet@arm.com> MIME-Version: 1.0 Content-Type: text/plain Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On 06/12/21 15:12, Mel Gorman wrote: > Gautham had similar reasoning to calculate the imbalance at each > higher-level domain instead of using a static value throughout and > it does make sense. For each level and splitting the imbalance between > two domains, this works out as > > > /* > * Calculate an allowed NUMA imbalance such that LLCs do not get > * imbalanced. > */ > for_each_cpu(i, cpu_map) { > for (sd = *per_cpu_ptr(d.sd, i); sd; sd = sd->parent) { > struct sched_domain *child = sd->child; > > if (!(sd->flags & SD_SHARE_PKG_RESOURCES) && child && > (child->flags & SD_SHARE_PKG_RESOURCES)) { > struct sched_domain *top = sd; > unsigned int llc_sq; > > /* > * nr_llcs = (top->span_weight / llc_weight); > * imb = (child_weight / nr_llcs) >> 1 > * > * is equivalent to > * > * imb = (llc_weight^2 / top->span_weight) >> 1 > * > */ > llc_sq = child->span_weight * child->span_weight; > while (top) { > top->imb_numa_nr = max(1U, > (llc_sq / top->span_weight) >> 1); > top = top->parent; > } > > break; > } > } > } > IIRC Peter suggested punting that logic to before domains get degenerated, but I don't see how that helps here. If you just want to grab the LLC domain (aka highest_flag_domain(cpu, SD_SHARE_PKG_RESOURCES)) and compare its span with that of its parents, that can happen after the degeneration, no? > I'll test this and should have results tomorrow. > > -- > Mel Gorman > SUSE Labs