Attaching a worker to a css_set isn't enough for all controllers to
throttle it. In particular, the memory controller currently bypasses
accounting for kernel threads.
Support memcg accounting for cgroup-aware workqueue workers so that
they're appropriately throttled.
Another, probably better way to do this is to have kernel threads, or
even specifically cgroup-aware workqueue workers, call
memalloc_use_memcg and memalloc_unuse_memcg during cgroup migration
(memcg attach callback maybe).
Signed-off-by: Daniel Jordan <[email protected]>
---
kernel/workqueue.c | 26 ++++++++++++++++++++++++++
kernel/workqueue_internal.h | 5 +++++
mm/memcontrol.c | 26 ++++++++++++++++++++++++--
3 files changed, 55 insertions(+), 2 deletions(-)
diff --git a/kernel/workqueue.c b/kernel/workqueue.c
index 89b90899bc09..c8cc69e296c0 100644
--- a/kernel/workqueue.c
+++ b/kernel/workqueue.c
@@ -50,6 +50,8 @@
#include <linux/sched/isolation.h>
#include <linux/nmi.h>
#include <linux/cgroup.h>
+#include <linux/memcontrol.h>
+#include <linux/sched/mm.h>
#include "workqueue_internal.h"
@@ -1829,6 +1831,28 @@ static inline bool worker_in_child_cgroup(struct worker *worker)
return (worker->flags & WORKER_CGROUP) && cgroup_parent(worker->cgroup);
}
+/* XXX Put this in the memory controller's attach callback. */
+#ifdef CONFIG_MEMCG
+static void worker_unuse_memcg(struct worker *worker)
+{
+ if (worker->task->active_memcg) {
+ struct mem_cgroup *memcg = worker->task->active_memcg;
+
+ memalloc_unuse_memcg();
+ css_put(&memcg->css);
+ }
+}
+
+static void worker_use_memcg(struct worker *worker)
+{
+ struct mem_cgroup *memcg;
+
+ worker_unuse_memcg(worker);
+ memcg = mem_cgroup_from_css(task_get_css(worker->task, memory_cgrp_id));
+ memalloc_use_memcg(memcg);
+}
+#endif /* CONFIG_MEMCG */
+
static void attach_worker_to_dfl_root(struct worker *worker)
{
int ret;
@@ -1841,6 +1865,7 @@ static void attach_worker_to_dfl_root(struct worker *worker)
rcu_read_lock();
worker->cgroup = task_dfl_cgroup(worker->task);
rcu_read_unlock();
+ worker_unuse_memcg(worker);
} else {
/*
* TODO Modify the cgroup migration path to guarantee that a
@@ -1880,6 +1905,7 @@ static void attach_worker_to_cgroup(struct worker *worker,
if (cgroup_attach_kthread(cgroup) == 0) {
worker->cgroup = cgroup;
+ worker_use_memcg(worker);
} else {
/*
* Attach failed, so attach to the default root so the
diff --git a/kernel/workqueue_internal.h b/kernel/workqueue_internal.h
index 3ad5861258ca..f254b93edc2c 100644
--- a/kernel/workqueue_internal.h
+++ b/kernel/workqueue_internal.h
@@ -79,6 +79,11 @@ work_func_t wq_worker_last_func(struct task_struct *task);
#ifdef CONFIG_CGROUPS
+#ifndef CONFIG_MEMCG
+static inline void worker_use_memcg(struct worker *worker) {}
+static inline void worker_unuse_memcg(struct worker *worker) {}
+#endif /* CONFIG_MEMCG */
+
/*
* A barrier work running in a cgroup-aware worker pool needs to specify a
* cgroup. For simplicity, WQ_BARRIER_CGROUP makes the worker stay in its
diff --git a/mm/memcontrol.c b/mm/memcontrol.c
index 81a0d3914ec9..1a80931b124a 100644
--- a/mm/memcontrol.c
+++ b/mm/memcontrol.c
@@ -2513,9 +2513,31 @@ static void memcg_schedule_kmem_cache_create(struct mem_cgroup *memcg,
static inline bool memcg_kmem_bypass(void)
{
- if (in_interrupt() || !current->mm || (current->flags & PF_KTHREAD))
+ if (in_interrupt())
return true;
- return false;
+
+ if (unlikely(current->flags & PF_WQ_WORKER)) {
+ struct cgroup *parent;
+
+ /*
+ * memcg should throttle cgroup-aware workers. Infer the
+ * worker is cgroup-aware by its presence in a non-root cgroup.
+ *
+ * This test won't detect a cgroup-aware worker attached to the
+ * default root, but in that case memcg doesn't need to
+ * throttle it anyway.
+ *
+ * XXX One alternative to this awkward block is adding a
+ * cgroup-aware-worker bit to task_struct.
+ */
+ rcu_read_lock();
+ parent = cgroup_parent(task_dfl_cgroup(current));
+ rcu_read_unlock();
+
+ return !parent;
+ }
+
+ return !current->mm || (current->flags & PF_KTHREAD);
}
/**
--
2.21.0