Received: by 2002:a25:31c3:0:0:0:0:0 with SMTP id x186csp656494ybx; Wed, 30 Oct 2019 02:58:55 -0700 (PDT) X-Google-Smtp-Source: APXvYqx+6nbXHBDbdv4yoLrnzIxGyVVs6fINo4+pfrFpwxAzCinI54b97YKie5gdgRf5fq06qT6h X-Received: by 2002:a17:906:9618:: with SMTP id s24mr8064165ejx.185.1572429535141; Wed, 30 Oct 2019 02:58:55 -0700 (PDT) ARC-Seal: i=1; a=rsa-sha256; t=1572429535; cv=none; d=google.com; s=arc-20160816; b=LBDBfIn29YWNabx7ybkPeqRh7c4Kwq/E8H7jB8JVUsQ7oHeDGaT/nbkCbMywZSAwE9 1JIU3assqoYMWnj4eDR3f+/TYv0+WTa3VAdXu3KAI5ZdAzvcynHKStg/fT6+2tQSAIZl DR42VQsjgVRt+C071qKigYpXeY6wZEy4mdxJlI8E6BRWA9eVNdqLAQI9GC1uKviG8M6v raRtwu1PTevwVxzUcCMYhH7ss/jg37D9p3vmo2ccWg/z99v449CzKXmcMlNKpLzI94gs 0j+xBt3FBzj328qOvcRpPV2k5qphTMwrruVY+65oUQWIzOdn7LBUdmed10/JoXPQmbCF xCPA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=list-id:precedence:sender:user-agent:in-reply-to :content-transfer-encoding:content-disposition:mime-version :references:message-id:subject:cc:to:from:date; bh=4Y8QX2rvA/6YQZD0PS7vM1s8tQ85537Q4Cf9kfdfZSg=; b=rQnqQH8XC7XxIA/cLM/qSTNWEpd+4GBlLlfk3hZYG/HpyWmS+KxTJowbVOD/GmVH6u /mFv+NFeahfplsJWT9ET2nJ82neqAvGLOU9e7VCjn/ZwzFDcsWP8j3cgIvbOUQ2auECK 9ctC5IxygJ71HniwYhYDGhplbwRgTgM8IxjUOEBCR9qjqsrDz+aVVNwK34ONay6vyaBC 09svfslQ8Tw2UcgcG/BgVkNZLTfKNmpRurarNRw6LKtEQrwCd4P9CZ6/0WoRKK4ql4N3 JoNH8keyARAK8hcOIvW0tROJBDziP8WQTEqYCosYgGUC/hpR49StY0qq79/YM+XkquO7 JtEg== ARC-Authentication-Results: i=1; mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Return-Path: Received: from vger.kernel.org (vger.kernel.org. [209.132.180.67]) by mx.google.com with ESMTP id l27si893735ejg.124.2019.10.30.02.58.31; Wed, 30 Oct 2019 02:58:55 -0700 (PDT) Received-SPF: pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) client-ip=209.132.180.67; Authentication-Results: mx.google.com; spf=pass (google.com: best guess record for domain of linux-kernel-owner@vger.kernel.org designates 209.132.180.67 as permitted sender) smtp.mailfrom=linux-kernel-owner@vger.kernel.org Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1726634AbfJ3JzL (ORCPT + 99 others); Wed, 30 Oct 2019 05:55:11 -0400 Received: from mx2.suse.de ([195.135.220.15]:38522 "EHLO mx1.suse.de" rhost-flags-OK-OK-OK-FAIL) by vger.kernel.org with ESMTP id S1726175AbfJ3JzL (ORCPT ); Wed, 30 Oct 2019 05:55:11 -0400 X-Virus-Scanned: by amavisd-new at test-mx.suse.de Received: from relay2.suse.de (unknown [195.135.220.254]) by mx1.suse.de (Postfix) with ESMTP id DCBFDAF3E; Wed, 30 Oct 2019 09:55:07 +0000 (UTC) Date: Wed, 30 Oct 2019 09:55:05 +0000 From: Mel Gorman To: ?????? Cc: Ingo Molnar , Peter Zijlstra , Juri Lelli , Vincent Guittot , Dietmar Eggemann , Steven Rostedt , Ben Segall , linux-kernel@vger.kernel.org Subject: Re: [PATCH] sched/numa: advanced per-cgroup numa statistic Message-ID: <20191030095505.GF28938@suse.de> References: <46b0fd25-7b73-aa80-372a-9fcd025154cb@linux.alibaba.com> MIME-Version: 1.0 Content-Type: text/plain; charset=iso-8859-15 Content-Disposition: inline Content-Transfer-Encoding: 8bit In-Reply-To: <46b0fd25-7b73-aa80-372a-9fcd025154cb@linux.alibaba.com> User-Agent: Mutt/1.10.1 (2018-07-13) Sender: linux-kernel-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-kernel@vger.kernel.org On Thu, Oct 24, 2019 at 11:08:01AM +0800, ?????? wrote: > Currently there are no good approach to monitoring the per-cgroup > numa efficiency, this could be a trouble especially when groups > are sharing CPUs, it's impossible to tell which one caused the > remote-memory access by reading hardware counter since multiple > workloads could sharing the same CPU, which make it painful when > one want to find out the root cause and fix the issue. > > In order to address this, we introduced new per-cgroup statistic > for numa: > * the numa locality to imply the numa balancing efficiency That is clear to some extent with the obvious caveat that a cgroup bound to a memory node will report this as being nearly 100% with shared pages being a possible exception. Locality might be fine but there could be large latencies due to reclaim if the cgroup limits are exceeded or the NUMA node is too small. It can give an artifical sense of benefit. > * the numa execution time on each node > This is less obvious because it does not define what NUMA execution time is and it's not documented by the patch. > The task locality is the local page accessing ratio traced on numa > balancing PF, and the group locality is the topology of task execution > time, sectioned by the locality into 8 regions. > This is another important limitation. Disabling NUMA balancing will also disable the stats which may be surprising to some users. It's not a show-stopper but arguably the file should be non-existant or always zeros if NUMA balancing is disabled. > For example the new entry 'cpu.numa_stat' show: > locality 15393 21259 13023 44461 21247 17012 28496 145402 > exectime 311900 407166 > > Here we know the workloads executed 311900ms on node_0 and 407166ms > on node_1, tasks with locality around 0~12% executed for 15393 ms, and > tasks with locality around 88~100% executed for 145402 ms, which imply > most of the memory access is local access, for the workloads of this > group. > This needs documentation because it's not obvious at all that the locality values are roughly percentiles. It's also not obvious what a sysadmin would *do* with this information. > By monitoring the new statistic, we will be able to know the numa > efficiency of each per-cgroup workloads on machine, whatever they > sharing the CPUs or not, we will be able to find out which one > introduced the remote access mostly. > > Besides, per-node memory topology from 'memory.numa_stat' become > more useful when we have the per-node execution time, workloads > always executing on node_0 while it's memory is all on node_1 is > usually a bad case. > > Cc: Peter Zijlstra > Cc: Michal Koutn? > Signed-off-by: Michael Wang So, superficially, I'm struggling to see what action a sysadmin would take, if any, with this information. It makes sense to track what individual tasks are doing but this can be done with ftrace if required. > --- > include/linux/sched.h | 8 ++++++- > kernel/sched/core.c | 59 +++++++++++++++++++++++++++++++++++++++++++++++++++ > kernel/sched/debug.c | 7 ++++++ > kernel/sched/fair.c | 51 ++++++++++++++++++++++++++++++++++++++++++++ > kernel/sched/sched.h | 29 +++++++++++++++++++++++++ > 5 files changed, 153 insertions(+), 1 deletion(-) > > diff --git a/include/linux/sched.h b/include/linux/sched.h > index 263cf089d1b3..46995be622c1 100644 > --- a/include/linux/sched.h > +++ b/include/linux/sched.h > @@ -1114,8 +1114,14 @@ struct task_struct { > * scan window were remote/local or failed to migrate. The task scan > * period is adapted based on the locality of the faults with different > * weights depending on whether they were shared or private faults > + * > + * 0 -- remote faults > + * 1 -- local faults > + * 2 -- page migration failure > + * 3 -- remote page accessing > + * 4 -- local page accessing > */ > - unsigned long numa_faults_locality[3]; > + unsigned long numa_faults_locality[5]; > Not clear from the comment what the difference between a remote fault and a remote page access is. Superficially, they're the same thing. > unsigned long numa_pages_migrated; > #endif /* CONFIG_NUMA_BALANCING */ > diff --git a/kernel/sched/core.c b/kernel/sched/core.c > index eb42b71faab9..4364da279f04 100644 > --- a/kernel/sched/core.c > +++ b/kernel/sched/core.c > @@ -6544,6 +6544,10 @@ static struct kmem_cache *task_group_cache __read_mostly; > DECLARE_PER_CPU(cpumask_var_t, load_balance_mask); > DECLARE_PER_CPU(cpumask_var_t, select_idle_mask); > > +#ifdef CONFIG_NUMA_BALANCING > +DECLARE_PER_CPU(struct numa_stat, root_numa_stat); > +#endif > + > void __init sched_init(void) > { > unsigned long ptr = 0; > @@ -6593,6 +6597,10 @@ void __init sched_init(void) > init_defrootdomain(); > #endif > > +#ifdef CONFIG_NUMA_BALANCING > + root_task_group.numa_stat = &root_numa_stat; > +#endif > + > #ifdef CONFIG_RT_GROUP_SCHED > init_rt_bandwidth(&root_task_group.rt_bandwidth, > global_rt_period(), global_rt_runtime()); This is an indication that there is a universal hit to collect this data whether cgroups are enabled or not. That hits the same problem I had with PSI when it was first introduced. For users that care, the information is useful, particularly as they generally have an application consuming the data. > @@ -6918,6 +6926,7 @@ static inline void alloc_uclamp_sched_group(struct task_group *tg, > > static void sched_free_group(struct task_group *tg) > { > + free_tg_numa_stat(tg); > free_fair_sched_group(tg); > free_rt_sched_group(tg); > autogroup_free(tg); > @@ -6933,6 +6942,9 @@ struct task_group *sched_create_group(struct task_group *parent) > if (!tg) > return ERR_PTR(-ENOMEM); > > + if (!alloc_tg_numa_stat(tg)) > + goto err; > + > if (!alloc_fair_sched_group(tg, parent)) > goto err; > While this is very unlikely to fail, I find it odd to think that an entire cgroup could fail to be created simply because stats cannot be collected, particularly when no userspace component may care at all. > @@ -7638,6 +7650,40 @@ static u64 cpu_rt_period_read_uint(struct cgroup_subsys_state *css, > } > #endif /* CONFIG_RT_GROUP_SCHED */ > > +#ifdef CONFIG_NUMA_BALANCING > +static int cpu_numa_stat_show(struct seq_file *sf, void *v) > +{ > + int nr; > + struct task_group *tg = css_tg(seq_css(sf)); > + > + seq_puts(sf, "locality"); > + for (nr = 0; nr < NR_NL_INTERVAL; nr++) { > + int cpu; > + u64 sum = 0; > + > + for_each_possible_cpu(cpu) > + sum += per_cpu(tg->numa_stat->locality[nr], cpu); > + > + seq_printf(sf, " %u", jiffies_to_msecs(sum)); > + } > + seq_putc(sf, '\n'); > + > + seq_puts(sf, "exectime"); > + for_each_online_node(nr) { > + int cpu; > + u64 sum = 0; > + > + for_each_cpu(cpu, cpumask_of_node(nr)) > + sum += per_cpu(tg->numa_stat->jiffies, cpu); > + > + seq_printf(sf, " %u", jiffies_to_msecs(sum)); > + } > + seq_putc(sf, '\n'); > + > + return 0; > +} > +#endif > + Ok, expensive on large machines but only hit with the proc file is read so that's ok. > static struct cftype cpu_legacy_files[] = { > #ifdef CONFIG_FAIR_GROUP_SCHED > { > @@ -7687,6 +7733,12 @@ static struct cftype cpu_legacy_files[] = { > .seq_show = cpu_uclamp_max_show, > .write = cpu_uclamp_max_write, > }, > +#endif > +#ifdef CONFIG_NUMA_BALANCING > + { > + .name = "numa_stat", > + .seq_show = cpu_numa_stat_show, > + }, > #endif > { } /* Terminate */ > }; > @@ -7868,6 +7920,13 @@ static struct cftype cpu_files[] = { > .seq_show = cpu_uclamp_max_show, > .write = cpu_uclamp_max_write, > }, > +#endif > +#ifdef CONFIG_NUMA_BALANCING > + { > + .name = "numa_stat", > + .flags = CFTYPE_NOT_ON_ROOT, > + .seq_show = cpu_numa_stat_show, > + }, > #endif > { } /* terminate */ > }; > diff --git a/kernel/sched/debug.c b/kernel/sched/debug.c > index f7e4579e746c..a22b2a62aee2 100644 > --- a/kernel/sched/debug.c > +++ b/kernel/sched/debug.c > @@ -848,6 +848,13 @@ static void sched_show_numa(struct task_struct *p, struct seq_file *m) > P(total_numa_faults); > SEQ_printf(m, "current_node=%d, numa_group_id=%d\n", > task_node(p), task_numa_group_id(p)); > + SEQ_printf(m, "faults_locality local=%lu remote=%lu failed=%lu ", > + p->numa_faults_locality[1], > + p->numa_faults_locality[0], > + p->numa_faults_locality[2]); This should be a separate patch. "failed=" does not tell much. It should be at least migfailed to give some indication it's about failed migrations. It might still be misleading because the failure could be due to lack of memory or because the page is pinned. However, I would not worry about that in this case. What *does* make this dangerous is that numa_faults_locality is often cleared. A user could easily think that this data somehow accumulates over time but it does not. This exposes implementation details as numa_faults_locality could change its behaviour in the future and tools should not rely on the contents being stable. While I recognise that some numa balancing information is already exposed in that file, it's relatively harmless with the possible exception of numa_scan_seq but at least that value is almost worthless (other than detecting if NUMA balancing is completely broken) and very unlikely to change behaviour. > + SEQ_printf(m, "lhit=%lu rhit=%lu\n", > + p->numa_faults_locality[4], > + p->numa_faults_locality[3]); > show_numa_stats(p, m); > mpol_put(pol); > #endif > diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c > index a81c36472822..4ba3a41cdca3 100644 > --- a/kernel/sched/fair.c > +++ b/kernel/sched/fair.c > @@ -2466,6 +2466,12 @@ void task_numa_fault(int last_cpupid, int mem_node, int pages, int flags) > p->numa_faults[task_faults_idx(NUMA_MEMBUF, mem_node, priv)] += pages; > p->numa_faults[task_faults_idx(NUMA_CPUBUF, cpu_node, priv)] += pages; > p->numa_faults_locality[local] += pages; > + /* > + * We want to have the real local/remote page access statistic > + * here, so use 'mem_node' which is the real residential node of > + * page after migrate_misplaced_page(). > + */ > + p->numa_faults_locality[3 + !!(mem_node == numa_node_id())] += pages; > } > This may be misleading if a task is using a preferred node policy that happens to be remote. It'll report bad locality but it's how the task waqs configured. Similarly, shared pages that are interleaves will show as remote accesses which is not necessarily bad. It goes back to "what does a sysadmin do with this information?" > static void reset_ptenuma_scan(struct task_struct *p) > @@ -2672,6 +2678,49 @@ void init_numa_balancing(unsigned long clone_flags, struct task_struct *p) > } > } > > +DEFINE_PER_CPU(struct numa_stat, root_numa_stat); > + > +int alloc_tg_numa_stat(struct task_group *tg) > +{ > + tg->numa_stat = alloc_percpu(struct numa_stat); > + if (!tg->numa_stat) > + return 0; > + > + return 1; > +} > + > +void free_tg_numa_stat(struct task_group *tg) > +{ > + free_percpu(tg->numa_stat); > +} > + > +static void update_tg_numa_stat(struct task_struct *p) > +{ > + struct task_group *tg; > + unsigned long remote = p->numa_faults_locality[3]; > + unsigned long local = p->numa_faults_locality[4]; > + int idx = -1; > + > + /* Tobe scaled? */ > + if (remote || local) > + idx = NR_NL_INTERVAL * local / (remote + local + 1); > + > + rcu_read_lock(); > + > + tg = task_group(p); > + while (tg) { > + /* skip account when there are no faults records */ > + if (idx != -1) > + this_cpu_inc(tg->numa_stat->locality[idx]); > + > + this_cpu_inc(tg->numa_stat->jiffies); > + > + tg = tg->parent; > + } > + > + rcu_read_unlock(); > +} > + This is potentially a long walk to do in the task tick context if there are a lot of groups. Also, if no faults are recorded because the scanner is running slowly, what does that mean for detecting locality? The information could be way off if a lot of accesses are remote and simply not caught by faults because of recent changes. So, overall, I see the general intent that you want to identify cgroups that have bad locality but it incurs a universal cost regardless of whether the user wants it or not. The documentation is non-existant but the biggest kicker is that it's unclear how a sysadmin would consume this information. The problem is compounded by the fact that interpreting the data accurately and making a decision requires knowledge of the implementation and that severely limits who can benefit. In my mind, it makes more sense to track the locality and NUMA behaviour of individual tasks which can be already done with ftrace. If cgroup tracking was required, a userspace application using ftrace could correlate task activity with what cgroup it belongs to and collect the data from userspace -- this is not necessarily easy and tracepoints might need updating, but it makes more sense to track this in userspace instead of in the kernel which never consumes the data. -- Mel Gorman SUSE Labs