Received: by 2002:ac2:464d:0:0:0:0:0 with SMTP id s13csp3304896lfo; Mon, 23 May 2022 01:12:18 -0700 (PDT) X-Google-Smtp-Source: ABdhPJw5D+dE8FNnQNVTy97KkmqZjzJoYG6oA3fv+gGvz42+yqv6voWZb9RA2azcvUhI90mdC2cc X-Received: by 2002:a17:903:25ce:b0:161:44a6:3346 with SMTP id jc14-20020a17090325ce00b0016144a63346mr21498473plb.146.1653293538671; Mon, 23 May 2022 01:12:18 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1653293538; cv=none; d=google.com; s=arc-20160816; b=gekrUo7V5MnmooHZucbSJo2GF54mSi3HXDwO5ZjUuaOMhwD+fDqN/jnDNyh2B4uOdB qvUB1HWEVkj3m2Mg/munFHRqKBeWtdR+eKRbVVVXYZTJ3kGaSzzUByDetaiBPgSC2Ed8 oase3ZkPKS0GPe7eQt4pi9IfiNTH65zHupPuW6VD1mnnyqY93Y8SFgvqg2OC+R9Hbb7u zIzIPN0FdImJbBhN3FIVB/AtDXZXaY7mbJPSnFAE5hmCJKaVBZLTvLtzZcKRNpuznKkC 86OYHCHuMH7t4wqeglx3dEqEl6B7Xai1qi2SsSIwKnCDtPItMZcz1/d3nOiTabzQb2cD M2SA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:in-reply-to:content-transfer-encoding :content-disposition:mime-version:references:message-id:subject:cc :to:from:date; bh=jYvVvxiyFwnbc5ZQvEX/+jW3n4piK0Ga3NuWYbOAZnA=; b=AdkuuDkAQijfm9CuTG49BMsL3qHxPvRfN8UAJ3Q52IbNzjZhEFhcZOVkzX/AD6Bx17 sfpixsxgXkDLToCJwJQArjmLoW4t13JjwCuzXrRdeMiArH4qzFHUK7KzeSQCVEf2aFaP 6QOtLCyvZcjfNNfRBi9pddSZosg22KkTzkwKTL0Zm1cDFkRfVprMnYeSQxeLQJbmgjLA TzY5CBmplubNwjpta1SIvk3KmLy1vLKEgDSvoZBkz47nokd3rjH77hqI9NMfc7z39Gs8 46mBkXmPk36ZvtT7RxSpErdC50AGcaE0m97AIhf3QG4NuaGRyAHUAN62/K0/Dw2F8rn2 /DwA== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=arm.com Return-Path: Received: from lindbergh.monkeyblade.net (lindbergh.monkeyblade.net. [2620:137:e000::1:18]) by mx.google.com with ESMTPS id x19-20020a17090aa39300b001dc7934acb3si11722415pjp.127.2022.05.23.01.12.18 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Mon, 23 May 2022 01:12:18 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:18 as permitted sender) client-ip=2620:137:e000::1:18; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=arm.com Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by lindbergh.monkeyblade.net (Postfix) with ESMTP id 2F653B7FC; Mon, 23 May 2022 00:06:46 -0700 (PDT) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1350853AbiETP1P (ORCPT + 99 others); Fri, 20 May 2022 11:27:15 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:38992 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1345085AbiETP1N (ORCPT ); Fri, 20 May 2022 11:27:13 -0400 Received: from foss.arm.com (foss.arm.com [217.140.110.172]) by lindbergh.monkeyblade.net (Postfix) with ESMTP id A95EE3C706 for ; Fri, 20 May 2022 08:27:11 -0700 (PDT) Received: from usa-sjc-imap-foss1.foss.arm.com (unknown [10.121.207.14]) by usa-sjc-mx-foss1.foss.arm.com (Postfix) with ESMTP id 5549B1477; Fri, 20 May 2022 08:27:11 -0700 (PDT) Received: from bogus (unknown [10.57.66.157]) by usa-sjc-imap-foss1.foss.arm.com (Postfix) with ESMTPSA id 1E43C3F73D; Fri, 20 May 2022 08:27:09 -0700 (PDT) Date: Fri, 20 May 2022 16:27:02 +0100 From: Sudeep Holla To: Ionela Voinescu Cc: Atish Patra , linux-kernel@vger.kernel.org, Atish Patra , Sudeep Holla , Vincent Guittot , Morten Rasmussen , Dietmar Eggemann , Qing Wang , linux-arm-kernel@lists.infradead.org, linux-riscv@lists.infradead.org, Rob Herring Subject: Re: [PATCH v2 3/8] arch_topology: Set cluster identifier in each core/thread from /cpu-map Message-ID: <20220520152702.ge3uxmizwljsg6mr@bogus> References: <20220518093325.2070336-1-sudeep.holla@arm.com> <20220518093325.2070336-4-sudeep.holla@arm.com> MIME-Version: 1.0 Content-Type: text/plain; charset=utf-8 Content-Disposition: inline Content-Transfer-Encoding: 8bit In-Reply-To: X-Spam-Status: No, score=-1.9 required=5.0 tests=BAYES_00, HEADER_FROM_DIFFERENT_DOMAINS,MAILING_LIST_MULTI,RDNS_NONE, SPF_HELO_NONE,T_SCC_BODY_TEXT_LINE autolearn=no autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Thu, May 19, 2022 at 05:55:30PM +0100, Ionela Voinescu wrote: > Hi, > > As said before, this creates trouble for CONFIG_SCHED_CLUSTER=y. > The output below is obtained from Juno. > > When cluster_id is populated, a new CLS level is created by the scheduler > topology code. In this case the clusters in DT determine that the cluster > siblings and llc siblings are the same so the MC scheduler domain will > be removed and, for Juno, only CLS and DIE will be kept. > Yes I have seen that. 1. Will that differ with ACPI on juno ? 2. Is that a problem ? I mean we are fixing those masks that are user visible with this series and if using them as is in sched domain is incorrect or not sufficient, we need to fix that. We can't change these masks. > root@debian-arm64-buster:/sys/kernel/debug/sched/domains/cpu1# grep . */* > domain0/busy_factor:16 > domain0/cache_nice_tries:1 > domain0/flags:SD_BALANCE_NEWIDLE SD_BALANCE_EXEC SD_BALANCE_FORK SD_WAKE_AFFINE SD_SHARE_PKG_RESOURCES SD_PREFER_SIBLING > domain0/imbalance_pct:117 > domain0/max_interval:4 > domain0/max_newidle_lb_cost:14907 > domain0/min_interval:2 > domain0/name:CLS > domain1/busy_factor:16 > domain1/cache_nice_tries:1 > domain1/flags:SD_BALANCE_NEWIDLE SD_BALANCE_EXEC SD_BALANCE_FORK SD_WAKE_AFFINE SD_ASYM_CPUCAPACITY SD_ASYM_CPUCAPACITY_FULL SD_PREFER_SIBLING > domain1/imbalance_pct:117 > domain1/max_interval:12 > domain1/max_newidle_lb_cost:11858 > domain1/min_interval:6 > domain1/name:DIE > > To be noted that we also get a new flag SD_PREFER_SIBLING for the CLS > level that is not appropriate. We usually remove it for the child of a > SD_ASYM_CPUCAPACITY domain, but we don't currently redo this after some > levels are degenerated. This is a fixable issue. > OK good. > But looking at the bigger picture, a good question is what is the best > thing to do when cluster domains and llc domains span the same CPUs? > Indeed, I will leave that to scheduler experts ????. My main goal is to get the topology masks that are user visible right and keeping current behavior intact. > Possibly it would be best to restrict clusters (which are almost an > arbitrary concept) to always span a subset of CPUs of the llc domain, > if llc siblings can be obtained? If those clusters are not properly set > up in DT to respect this condition, cluster_siblings would need to be > cleared (or set to the current CPU) so the CLS domain is not created at > all. > Yes, we already have all these complex conditions in cpu_coregroup_mask. Why not do something similar in cpu_clustergroup_mask ? > Additionally, should we use cluster information from DT (cluster_id) to > create an MC level if we don't have llc information, even if > CONFIG_SCHED_CLUSTER=n? > I don't know. We have all the topology and llc info now, we need to decide how to present them to scheduler. > I currently don't have a very clear picture of how cluster domains and > llc domains would "live" together in a variety of topologies. I'll try > other DT topologies to see if there are others that can lead to trouble. > Me neither. I avoided these changes for years because of such complex questions but that is wrong reason to, as the user visible topology has now diverged from ACPI once(people who complained about user space breakage cared only about ACPI) and we need to bring them in parity soon IMO. -- Regards, Sudeep