Received: by 2002:a6b:500f:0:0:0:0:0 with SMTP id e15csp4094521iob; Tue, 17 May 2022 13:55:14 -0700 (PDT) X-Google-Smtp-Source: ABdhPJyfLyWsij/d3Um3eDxzMoiBPk/Dzgf6HeCAKvp8GzVF56//NFKhZV4fU7HKh+SrPjrYu25s X-Received: by 2002:a17:907:1c24:b0:6f4:ff62:a393 with SMTP id nc36-20020a1709071c2400b006f4ff62a393mr21507796ejc.154.1652820914004; Tue, 17 May 2022 13:55:14 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1652820914; cv=none; d=google.com; s=arc-20160816; b=x8tI2GX3aBZInqLQ8REuAzvm2jhQRMb4TLIpY+yjoR4VxXtDT1gHAe4n9jjyM+ZYhQ xYRq84J8FcFStzI+EYRZuj793/Xlb0e40yz2ZUJijuxCOaJA2DtukhAhlY4THMJHRSmC 3XuhSUPwlKYIVl8H9wm+PafUZaAs2M3orwbAULfIP7t35tHfJm1TKHnaD5WwcnIF9Eh9 ufYFP0xaNYTOqNcANmAeLktPOKyOjtOiA2Aakr9xSwCkknLwpZTvNOtx9rUZ8WIW/rMn NWIq7oJ2y3QQpPWUH1YYFBOjkMjQ4aAa6fAYWsGJKEvJXlNcM9dQgyrRedqL3aOmnAWo txJw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:in-reply-to:content-disposition:mime-version :references:message-id:subject:cc:to:from:date; bh=MupB8MsBofBEA5mkwbhfRD09V86qTdugZgeeCPNVJhc=; b=rzuPXO2rsZ6hdWE7PjoRZqGj7drQgNo+oZiEP+BhaKhDlePFzus5LcTCJGqq0XMitT 15JYho/P+FGywUpOq7v38gADe9/2iVuCT4CyFEXscPvTQmI89eEHyluq5Q0+p4r91cZL dIZ1kP4yWd9ugO4md0TK6GwOKwbmiAWHInY35gO8IbdqYMWtYlwREKgqKQyoAM3aryAw ZQp9fWTjrcZIqvwyPOHtJyp8h+X4eWU4sEjOLRCr73Tn94SmpMw2LtXpiLq/xywbdK/4 9lQR/VXbmrDHzcqFBlNJR8uCPOUoncMVRPFBxOsv39kt0bblqw5+q5lib3mxjr54yHA0 R3Ow== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=arm.com Return-Path: Received: from out1.vger.email (out1.vger.email. [2620:137:e000::1:20]) by mx.google.com with ESMTP id w8-20020a05640234c800b004264545ad67si407790edc.18.2022.05.17.13.54.48; Tue, 17 May 2022 13:55:13 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) client-ip=2620:137:e000::1:20; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 2620:137:e000::1:20 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=arm.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1344652AbiEQK5p (ORCPT + 99 others); Tue, 17 May 2022 06:57:45 -0400 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:53658 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1344811AbiEQK5h (ORCPT ); Tue, 17 May 2022 06:57:37 -0400 Received: from foss.arm.com (foss.arm.com [217.140.110.172]) by lindbergh.monkeyblade.net (Postfix) with ESMTP id 12D2749FA0 for ; Tue, 17 May 2022 03:57:22 -0700 (PDT) Received: from usa-sjc-imap-foss1.foss.arm.com (unknown [10.121.207.14]) by usa-sjc-mx-foss1.foss.arm.com (Postfix) with ESMTP id D2AB91042; Tue, 17 May 2022 03:57:21 -0700 (PDT) Received: from bogus (e103737-lin.cambridge.arm.com [10.1.197.49]) by usa-sjc-imap-foss1.foss.arm.com (Postfix) with ESMTPSA id B32B73F66F; Tue, 17 May 2022 03:57:20 -0700 (PDT) Date: Tue, 17 May 2022 11:57:18 +0100 From: Sudeep Holla To: Dietmar Eggemann Cc: Qing Wang , Greg Kroah-Hartman , "Rafael J . Wysocki" , Sudeep Holla , Vincent Guittot , Morten Rasmussen , linux-kernel@vger.kernel.org Subject: Re: [RFC PATCH] arch_topology: Use llc_id instead of package_id Message-ID: <20220517105718.tvpmj2xxb2qj3bev@bogus> References: <20220513083400.343706-1-dietmar.eggemann@arm.com> <20220513090330.z25fwthekn4rjkwq@bogus> <20220513110312.wy6g5avs7ngkmhfu@bogus> <634a4b8c-84d2-51a9-ef54-33b81683c849@arm.com> <20220516103524.35swlxp2367baxx6@bogus> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: X-Spam-Status: No, score=-6.9 required=5.0 tests=BAYES_00,RCVD_IN_DNSWL_HI, SPF_HELO_NONE,SPF_PASS,T_SCC_BODY_TEXT_LINE autolearn=ham autolearn_force=no version=3.4.6 X-Spam-Checker-Version: SpamAssassin 3.4.6 (2021-04-09) on lindbergh.monkeyblade.net Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Hi Dietmar, Thanks for the detailed response. On Tue, May 17, 2022 at 11:14:44AM +0200, Dietmar Eggemann wrote: > On 16/05/2022 12:35, Sudeep Holla wrote: > > On Fri, May 13, 2022 at 02:04:29PM +0200, Dietmar Eggemann wrote: > >> On 13/05/2022 13:03, Sudeep Holla wrote: > >>> On Fri, May 13, 2022 at 12:42:00PM +0200, Dietmar Eggemann wrote: > >>>> On 13/05/2022 11:03, Sudeep Holla wrote: > >>>>> On Fri, May 13, 2022 at 10:34:00AM +0200, Dietmar Eggemann wrote: > > [...] > > >> cat /sys/kernel/debug/sched/domains/cpu0/domain*/name > >> CLS > >> MC > >> ... I skip the NUMA levels > >> > >> # cat /proc/schedstat | awk '{print $1 " " $2 }' | grep ^[cd] | head -5 > >> cpu0 0 > >> domain0 00000000,00000000,0000000f <-- 4 CPUs <-- cluster_id > >> domain1 00000000,00000000,00ffffff <-- 24 CPUs > >> > >> If you use cluster_id for 1. level cluster then you cover MC's 24 CPUs > > > > OK, the way I am looking at this from CPU topology perspective seem to be > > different from what you are expecting here w.r.t sched_domains. > > > > IMO, these cpumasks on its own must represent real CPU topology and how it > > is used via cpu_*_mask by the scheduler domains is different. > > I'm afraid that in case we want to change the mapping of scheduler > (topology) flags (like SD_SHARE_PKG_RESOURCES) via `*_flags()` of > default_topology[] we would have to consider all ACPI corner cases (see > cpu_coregroup_mask()) as well. > > See (1) further down. > Understood. > [...] > > > I see on Juno with SCHED_CLUSTER and cluster masks set, I see CLS and MC > > domains. > > Right but that's not really correct from the scheduler's POV. > OK > Juno has LLC on cpumasks [0,3-5] and [1-2], not on [0-5]. So the most > important information is the highest Sched Domain with the > SD_SHARE_PKG_RESOURCES flag. This is the MC layer (cpu_core_flags() in > default_topology[]). So the scheduler would think that [0-5] are sharing > LLC. > Ah OK, but if LLC sibling masks are updated, cpu_coregroup_mask() takes care of it IIUC, right ? > You have LLC at: > > cat /sys/devices/system/cpu/cpu0/cache/index2/shared_cpu_list > ^^^^^^ > 0,3-5 > > but the scheduler sees the highest SD_SHARE_PKG_RESOURCES on MC: > > cat /sys/kernel/debug/sched/domains/cpu0/domain1/flags > ^^^^^^^ > ... SD_SHARE_PKG_RESOURCES ... > > [...] > > >> For one level (MC) yes, but not for 2 (MC and CLS). And cluster_id was > >> introduces for the 2. level. > >> > > > > That sounds wrong and not what ACPI PPTT code says. My series just aligns > > with what is done with ACPI PPTT IIUC. I need to check that again if my > > understand differs from what is being done. But the example of Kunpeng920 > > aligns with my understanding. > > (1) IMHO, as long as we don't consider cache (llc) information in DT we > can't have the same coverage as ACPI. > Agreed. But we are not changing any topology masks as per sched domain requirements as they get exposed to the userspace as is. > Let's take an ACPI !CONFIG_NUMA Kunpeng920 as an example. > > # cat /sys/kernel/debug/sched/domains/cpu0/domain*/name > CLS > MC > DIE > > # cat /sys/kernel/debug/sched/domains/cpu0/domain*/flags > SD_BALANCE_NEWIDLE SD_BALANCE_EXEC SD_BALANCE_FORK SD_WAKE_AFFINE > SD_SHARE_PKG_RESOURCES SD_PREFER_SIBLING > ^^^^^^^^^^^^^^^^^^^^^^ > SD_BALANCE_NEWIDLE SD_BALANCE_EXEC SD_BALANCE_FORK SD_WAKE_AFFINE > SD_SHARE_PKG_RESOURCES SD_PREFER_SIBLING > ^^^^^^^^^^^^^^^^^^^^^^ > SD_BALANCE_NEWIDLE SD_BALANCE_EXEC SD_BALANCE_FORK SD_WAKE_AFFINE > SD_PREFER_SIBLING > > cat /proc/schedstat | awk '{print $1 " " $2 }' | grep ^[cd] | head -4 > cpu0 0 > domain0 00000000,00000000,0000000f > domain1 00000000,00000000,00ffffff <-- (2) > domain2 ffffffff,ffffffff,ffffffff > > cat /sys/devices/system/cpu/cpu0/topology/thread_siblings_list > 0 > cat /sys/devices/system/cpu/cpu0/topology/core_cpus_list > 0 > cat /sys/devices/system/cpu/cpu0/topology/cluster_cpus_list > 0-3 > cat /sys/devices/system/cpu/cpu0/topology/core_siblings_list > 0-47 > cat /sys/devices/system/cpu/cpu0/topology/package_cpus_list > 0-47 > > The MC mask 00ffffff is not derived from any topology mask but from the > llc (index3) mask: > > cat /sys/devices/system/cpu/cpu0/cache/index3/shared_cpu_list > ^^^^^^ > 0-23 <-- (2) > Understood and on Juno if we get llc_siblings right, the sched domains must be sorted correctly ? > > Coming back to the original request (the support of Armv9 L2 complexes > in Android) from Qing on systems like QC SM8450: > > .---------------. > CPU |0 1 2 3 4 5 6 7| > +---------------+ > uarch |l l l l m m m b| (so called tri-gear: little, medium, big) > +---------------+ > L2 | | | | | | | > +---------------+ > L3 |<-- -->| > +---------------+ > |<-- cluster -->| > +---------------+ > |<-- DSU -->| > '---------------' > > This still wouldn't be possible. We know that Phantom Domain (grouping > after uarch) is not supported in mainline but heavily used in Android > (legacy deps). > Correct, unless you convince to get a suitable notion of *whatever* phantom domains represent into the DT bindings, they don't exist. If someone wants to support this config, they must first represent that in the firmware so that OS can infer information from the same. > If we say we only support L2 sharing (asymmetric here since only CPU0-3 > have it !!!) and we don't support Phantom then your approach will work > for such systems. Thanks, as I said what is *Phantom* domain ;) ? Can you point me to the DT or ACPI binding for the same ? Just kidding, I know they don't exist. Anyways, I understand your concern that llc_sibling must go with my set of topology changes which I agree. Is that the only concern ? -- Regards, Sudeep