Received: by 2002:a05:6a10:9848:0:0:0:0 with SMTP id x8csp1463227pxf; Fri, 12 Mar 2021 10:03:08 -0800 (PST) X-Google-Smtp-Source: ABdhPJw81f2j6AQHDc15ngcO7yY+cS4ysqeq5xf8PUM0zeOR0UzVhIBg/9lftpTkJ/BIAoEve1tT X-Received: by 2002:aa7:c7c3:: with SMTP id o3mr15905540eds.8.1615572188350; Fri, 12 Mar 2021 10:03:08 -0800 (PST) ARC-Seal: i=1; a=rsa-sha256; t=1615572188; cv=none; d=google.com; s=arc-20160816; b=Jw6B9b6Z/P6xe+bsZG6EXJfu3yOUZ0DkiHcWKK5z4ikSatvihcN2u7PW+QInryh9Hg WXwA4WKZCSbjodVP0Bi8YZHPpISks0QX46zds0j73WCLKb4CjGfCq9BAK/1aVShDJ1sn ginp+ZRIrf2To0/RfJKcUD/2tYlEeQ5+xhCiZYgayXlOFuIYMehpG0h1A85RAf/HuvNx fJ34SkoqfOznvsPZ4qidkoH8LG5x7hgqiem84voAIbNtp8XdJVBMM+4HSrozw7u0jfgx jHxdUxlX99M1eNWWf1iwA7P26O1am8HO3+/u3Mp4CouHUaZEqWa5HUCDjiSMcNugF+2O N+ng== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:content-transfer-encoding:mime-version :references:in-reply-to:message-id:date:subject:cc:to:from; bh=0e+xUR7KP2wAlHLL5TUalTnpSYPGjeRgtaDoyfp5Yj0=; b=xOT42bNFlLVTjc/uFiV2SheGG+VRT038jzSj45ovtAUl3crBrlHvG/frA61pQlQ31q mp6HaG4YyqisjgvQpJr5uPHI0uq+PumUwO/WGzquuyM+zczG/IxKMVqG/G46rX4bzUev kqqQwqBNMWIZgrGE+sjbEt7Ar0K9s3fP3grYcEuMkDp2+cFFoAKGC/Xg2fixDNmOm8st yaHKWJYh8ydmIskK2mgKiRcnNxT3uwA93CQAYNqbsZkc43bhl84ztp/i22f1bcFKFYB2 FQ48WP8g62H9SIBaTiB1ZHbnC0t0/lkeEG6i/+WCbiBeuzabqXDpnGkZTWQ5SIB44KLH mvAQ== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=arm.com Return-Path: Received: from vger.kernel.org (vger.kernel.org. [23.128.96.18]) by mx.google.com with ESMTP id f24si4670345eds.18.2021.03.12.10.02.45; Fri, 12 Mar 2021 10:03:08 -0800 (PST) Received-SPF: pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) client-ip=23.128.96.18; Authentication-Results: mx.google.com; spf=pass (google.com: domain of linux-kernel-owner@vger.kernel.org designates 23.128.96.18 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org; dmarc=fail (p=NONE sp=NONE dis=NONE) header.from=arm.com Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S233129AbhCLSBM (ORCPT + 99 others); Fri, 12 Mar 2021 13:01:12 -0500 Received: from foss.arm.com ([217.140.110.172]:58894 "EHLO foss.arm.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S233110AbhCLSAl (ORCPT ); Fri, 12 Mar 2021 13:00:41 -0500 Received: from usa-sjc-imap-foss1.foss.arm.com (unknown [10.121.207.14]) by usa-sjc-mx-foss1.foss.arm.com (Postfix) with ESMTP id A7DA6ED1; Fri, 12 Mar 2021 10:00:40 -0800 (PST) Received: from merodach.members.linode.com (unknown [172.31.20.19]) by usa-sjc-imap-foss1.foss.arm.com (Postfix) with ESMTPSA id D50F33F7D7; Fri, 12 Mar 2021 10:00:38 -0800 (PST) From: James Morse To: x86@kernel.org, linux-kernel@vger.kernel.org Cc: Fenghua Yu , Reinette Chatre , Thomas Gleixner , Ingo Molnar , Borislav Petkov , H Peter Anvin , Babu Moger , James Morse , shameerali.kolothum.thodi@huawei.com, Jamie Iles , D Scott Phillips OS Subject: [PATCH v2 23/24] x86/resctrl: Remove rdt_cdp_peer_get() Date: Fri, 12 Mar 2021 17:58:48 +0000 Message-Id: <20210312175849.8327-24-james.morse@arm.com> X-Mailer: git-send-email 2.20.1 In-Reply-To: <20210312175849.8327-1-james.morse@arm.com> References: <20210312175849.8327-1-james.morse@arm.com> MIME-Version: 1.0 Content-Transfer-Encoding: 8bit Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org Now that the configuration can be read from either CDP peer, rdt_cdp_peer_get() is not needed to to map the resource and search for the corresponding domain. As __rdtgroup_cbm_overlaps() takes the configuration type from the schema, this can be replaced with a second call for the other configuration type if cdp is enabled. Reviewed-by: Jamie Iles Signed-off-by: James Morse Changes since v1: * Expanded commit mesasge. --- arch/x86/kernel/cpu/resctrl/rdtgroup.c | 99 ++++---------------------- 1 file changed, 14 insertions(+), 85 deletions(-) diff --git a/arch/x86/kernel/cpu/resctrl/rdtgroup.c b/arch/x86/kernel/cpu/resctrl/rdtgroup.c index 801bff59db06..36e2905f4da6 100644 --- a/arch/x86/kernel/cpu/resctrl/rdtgroup.c +++ b/arch/x86/kernel/cpu/resctrl/rdtgroup.c @@ -1092,82 +1092,17 @@ static int rdtgroup_mode_show(struct kernfs_open_file *of, return 0; } -/** - * rdt_cdp_peer_get - Retrieve CDP peer if it exists - * @r: RDT resource to which RDT domain @d belongs - * @d: Cache instance for which a CDP peer is requested - * @r_cdp: RDT resource that shares hardware with @r (RDT resource peer) - * Used to return the result. - * @d_cdp: RDT domain that shares hardware with @d (RDT domain peer) - * Used to return the result. - * @peer_type: The CDP configuration type of the peer resource. - * - * RDT resources are managed independently and by extension the RDT domains - * (RDT resource instances) are managed independently also. The Code and - * Data Prioritization (CDP) RDT resources, while managed independently, - * could refer to the same underlying hardware. For example, - * RDT_RESOURCE_L2CODE and RDT_RESOURCE_L2DATA both refer to the L2 cache. - * - * When provided with an RDT resource @r and an instance of that RDT - * resource @d rdt_cdp_peer_get() will return if there is a peer RDT - * resource and the exact instance that shares the same hardware. - * - * Return: 0 if a CDP peer was found, <0 on error or if no CDP peer exists. - * If a CDP peer was found, @r_cdp will point to the peer RDT resource - * and @d_cdp will point to the peer RDT domain. - */ -static int rdt_cdp_peer_get(struct rdt_resource *r, struct rdt_domain *d, - struct rdt_resource **r_cdp, - struct rdt_domain **d_cdp, - enum resctrl_conf_type *peer_type) +static enum resctrl_conf_type resctrl_peer_type(enum resctrl_conf_type my_type) { - struct rdt_resource *_r_cdp = NULL; - struct rdt_domain *_d_cdp = NULL; - int ret = 0; - - switch (r->rid) { - case RDT_RESOURCE_L3DATA: - _r_cdp = &rdt_resources_all[RDT_RESOURCE_L3CODE].resctrl; - *peer_type = CDP_CODE; - break; - case RDT_RESOURCE_L3CODE: - _r_cdp = &rdt_resources_all[RDT_RESOURCE_L3DATA].resctrl; - *peer_type = CDP_DATA; - break; - case RDT_RESOURCE_L2DATA: - _r_cdp = &rdt_resources_all[RDT_RESOURCE_L2CODE].resctrl; - *peer_type = CDP_CODE; - break; - case RDT_RESOURCE_L2CODE: - _r_cdp = &rdt_resources_all[RDT_RESOURCE_L2DATA].resctrl; - *peer_type = CDP_DATA; - break; + switch (my_type) { + case CDP_CODE: + return CDP_DATA; + case CDP_DATA: + return CDP_CODE; default: - ret = -ENOENT; - goto out; - } - - /* - * When a new CPU comes online and CDP is enabled then the new - * RDT domains (if any) associated with both CDP RDT resources - * are added in the same CPU online routine while the - * rdtgroup_mutex is held. It should thus not happen for one - * RDT domain to exist and be associated with its RDT CDP - * resource but there is no RDT domain associated with the - * peer RDT CDP resource. Hence the WARN. - */ - _d_cdp = rdt_find_domain(_r_cdp, d->id, NULL); - if (WARN_ON(IS_ERR_OR_NULL(_d_cdp))) { - _r_cdp = NULL; - _d_cdp = NULL; - ret = -EINVAL; + case CDP_BOTH: + return CDP_BOTH; } - -out: - *r_cdp = _r_cdp; - *d_cdp = _d_cdp; - - return ret; } /** @@ -1248,19 +1183,16 @@ static bool __rdtgroup_cbm_overlaps(struct rdt_resource *r, struct rdt_domain *d bool rdtgroup_cbm_overlaps(struct resctrl_schema *s, struct rdt_domain *d, unsigned long cbm, int closid, bool exclusive) { - enum resctrl_conf_type peer_type; + enum resctrl_conf_type peer_type = resctrl_peer_type(s->conf_type); struct rdt_resource *r = s->res; - struct rdt_resource *r_cdp; - struct rdt_domain *d_cdp; if (__rdtgroup_cbm_overlaps(r, d, cbm, closid, s->conf_type, exclusive)) return true; - if (rdt_cdp_peer_get(r, d, &r_cdp, &d_cdp, &peer_type) < 0) + if (!resctrl_arch_get_cdp_enabled(r->rid)) return false; - - return __rdtgroup_cbm_overlaps(r_cdp, d_cdp, cbm, closid, peer_type, exclusive); + return __rdtgroup_cbm_overlaps(r, d, cbm, closid, peer_type, exclusive); } /** @@ -2746,11 +2678,9 @@ static u32 cbm_ensure_valid(u32 _val, struct rdt_resource *r) static int __init_one_rdt_domain(struct rdt_domain *d, struct resctrl_schema *s, u32 closid) { + enum resctrl_conf_type peer_type = resctrl_peer_type(s->conf_type); enum resctrl_conf_type t = s->conf_type; - struct rdt_resource *r_cdp = NULL; struct resctrl_staged_config *cfg; - enum resctrl_conf_type peer_type; - struct rdt_domain *d_cdp = NULL; struct rdt_resource *r = s->res; u32 used_b = 0, unused_b = 0; unsigned long tmp_cbm; @@ -2758,7 +2688,6 @@ static int __init_one_rdt_domain(struct rdt_domain *d, struct resctrl_schema *s, u32 peer_ctl, ctrl_val; int i; - rdt_cdp_peer_get(r, d, &r_cdp, &d_cdp, &peer_type); cfg = &d->staged_config[t]; cfg->have_new_ctrl = false; cfg->new_ctrl = r->cache.shareable_bits; @@ -2778,8 +2707,8 @@ static int __init_one_rdt_domain(struct rdt_domain *d, struct resctrl_schema *s, * usage to ensure there is no overlap * with an exclusive group. */ - if (d_cdp) - resctrl_arch_get_config(r_cdp, d_cdp, i, peer_type, &peer_ctl); + if (resctrl_arch_get_cdp_enabled(r->rid)) + resctrl_arch_get_config(r, d, i, peer_type, &peer_ctl); else peer_ctl = 0; resctrl_arch_get_config(r, d, i, s->conf_type, &ctrl_val); -- 2.30.0