A previous change added a series of kfuncs for storing struct
task_struct objects as referenced kptrs. This patch adds a new
task_kfunc test suite for validating their expected behavior.
Signed-off-by: David Vernet <[email protected]>
---
tools/testing/selftests/bpf/DENYLIST.s390x | 1 +
.../selftests/bpf/prog_tests/task_kfunc.c | 159 +++++++++++
.../selftests/bpf/progs/task_kfunc_common.h | 71 +++++
.../selftests/bpf/progs/task_kfunc_failure.c | 259 ++++++++++++++++++
.../selftests/bpf/progs/task_kfunc_success.c | 149 ++++++++++
5 files changed, 639 insertions(+)
create mode 100644 tools/testing/selftests/bpf/prog_tests/task_kfunc.c
create mode 100644 tools/testing/selftests/bpf/progs/task_kfunc_common.h
create mode 100644 tools/testing/selftests/bpf/progs/task_kfunc_failure.c
create mode 100644 tools/testing/selftests/bpf/progs/task_kfunc_success.c
diff --git a/tools/testing/selftests/bpf/DENYLIST.s390x b/tools/testing/selftests/bpf/DENYLIST.s390x
index be4e3d47ea3e..97e522828d6c 100644
--- a/tools/testing/selftests/bpf/DENYLIST.s390x
+++ b/tools/testing/selftests/bpf/DENYLIST.s390x
@@ -53,6 +53,7 @@ skc_to_unix_sock # could not attach BPF object unexpecte
socket_cookie # prog_attach unexpected error: -524 (trampoline)
stacktrace_build_id # compare_map_keys stackid_hmap vs. stackmap err -2 errno 2 (?)
tailcalls # tail_calls are not allowed in non-JITed programs with bpf-to-bpf calls (?)
+task_kfunc # JIT does not support calling kernel function
task_local_storage # failed to auto-attach program 'trace_exit_creds': -524 (trampoline)
test_bpffs # bpffs test failed 255 (iterator)
test_bprm_opts # failed to auto-attach program 'secure_exec': -524 (trampoline)
diff --git a/tools/testing/selftests/bpf/prog_tests/task_kfunc.c b/tools/testing/selftests/bpf/prog_tests/task_kfunc.c
new file mode 100644
index 000000000000..a3aa36a4beb3
--- /dev/null
+++ b/tools/testing/selftests/bpf/prog_tests/task_kfunc.c
@@ -0,0 +1,159 @@
+// SPDX-License-Identifier: GPL-2.0
+/* Copyright (c) 2022 Meta Platforms, Inc. and affiliates. */
+
+#define _GNU_SOURCE
+#include <sys/wait.h>
+#include <test_progs.h>
+#include <unistd.h>
+
+#include "task_kfunc_failure.skel.h"
+#include "task_kfunc_success.skel.h"
+
+static size_t log_buf_sz = 1 << 20; /* 1 MB */
+static char obj_log_buf[1048576];
+
+static struct task_kfunc_success *open_load_task_kfunc_skel(void)
+{
+ struct task_kfunc_success *skel;
+ int err;
+
+ skel = task_kfunc_success__open();
+ if (!ASSERT_OK_PTR(skel, "skel_open"))
+ return NULL;
+
+ skel->bss->pid = getpid();
+
+ err = task_kfunc_success__load(skel);
+ if (!ASSERT_OK(err, "skel_load"))
+ goto cleanup;
+
+ return skel;
+
+cleanup:
+ task_kfunc_success__destroy(skel);
+ return NULL;
+}
+
+static void run_success_test(const char *prog_name)
+{
+ struct task_kfunc_success *skel;
+ int status;
+ pid_t child_pid;
+ struct bpf_program *prog;
+ struct bpf_link *link = NULL;
+
+ skel = open_load_task_kfunc_skel();
+ if (!ASSERT_OK_PTR(skel, "open_load_skel"))
+ return;
+
+ if (!ASSERT_OK(skel->bss->err, "pre_spawn_err"))
+ goto cleanup;
+
+ prog = bpf_object__find_program_by_name(skel->obj, prog_name);
+ if (!ASSERT_OK_PTR(prog, "bpf_object__find_program_by_name"))
+ goto cleanup;
+
+ link = bpf_program__attach(prog);
+ if (!ASSERT_OK_PTR(link, "attached_link"))
+ goto cleanup;
+
+ child_pid = fork();
+ if (!ASSERT_GT(child_pid, -1, "child_pid"))
+ goto cleanup;
+ if (child_pid == 0)
+ _exit(0);
+ waitpid(child_pid, &status, 0);
+
+ ASSERT_OK(skel->bss->err, "post_wait_err");
+
+cleanup:
+ bpf_link__destroy(link);
+ task_kfunc_success__destroy(skel);
+}
+
+static const char * const success_tests[] = {
+ "test_task_acquire_release_argument",
+ "test_task_acquire_release_current",
+ "test_task_acquire_leave_in_map",
+ "test_task_xchg_release",
+ "test_task_get_release",
+ "test_task_current_acquire_release",
+};
+
+static struct {
+ const char *prog_name;
+ const char *expected_err_msg;
+} failure_tests[] = {
+ {"task_kfunc_acquire_untrusted", "R1 must be referenced or trusted"},
+ {"task_kfunc_acquire_fp", "arg#0 pointer type STRUCT task_struct must point"},
+ {"task_kfunc_acquire_unsafe_kretprobe", "reg type unsupported for arg#0 function"},
+ {"task_kfunc_acquire_trusted_walked", "R1 must be referenced or trusted"},
+ {"task_kfunc_acquire_null", "arg#0 pointer type STRUCT task_struct must point"},
+ {"task_kfunc_acquire_unreleased", "Unreleased reference"},
+ {"task_kfunc_get_non_kptr_param", "arg#0 expected pointer to map value"},
+ {"task_kfunc_get_non_kptr_acquired", "arg#0 expected pointer to map value"},
+ {"task_kfunc_get_null", "arg#0 expected pointer to map value"},
+ {"task_kfunc_xchg_unreleased", "Unreleased reference"},
+ {"task_kfunc_get_unreleased", "Unreleased reference"},
+ {"task_kfunc_release_untrusted", "bpf_task_release arg#0 pointer had unexpected modifiers"},
+ {"task_kfunc_release_fp", "arg#0 pointer type STRUCT task_struct must point"},
+ {"task_kfunc_release_null", "bpf_task_release arg#0 pointer had unexpected modifiers"},
+ {"task_kfunc_release_unacquired", "release kernel function bpf_task_release expects"},
+};
+
+static void verify_fail(const char *prog_name, const char *expected_err_msg)
+{
+ LIBBPF_OPTS(bpf_object_open_opts, opts);
+ struct task_kfunc_failure *skel;
+ int err, i;
+
+ opts.kernel_log_buf = obj_log_buf;
+ opts.kernel_log_size = log_buf_sz;
+ opts.kernel_log_level = 1;
+
+ skel = task_kfunc_failure__open_opts(&opts);
+ if (!ASSERT_OK_PTR(skel, "task_kfunc_failure__open_opts"))
+ goto cleanup;
+
+ for (i = 0; i < ARRAY_SIZE(failure_tests); i++) {
+ struct bpf_program *prog;
+ const char *curr_name = failure_tests[i].prog_name;
+
+ prog = bpf_object__find_program_by_name(skel->obj, curr_name);
+ if (!ASSERT_OK_PTR(prog, "bpf_object__find_program_by_name"))
+ goto cleanup;
+
+ bpf_program__set_autoload(prog, !strcmp(curr_name, prog_name));
+ }
+
+ err = task_kfunc_failure__load(skel);
+ if (!ASSERT_ERR(err, "unexpected load success"))
+ goto cleanup;
+
+ if (!ASSERT_OK_PTR(strstr(obj_log_buf, expected_err_msg), "expected_err_msg")) {
+ fprintf(stderr, "Expected err_msg: %s\n", expected_err_msg);
+ fprintf(stderr, "Verifier output: %s\n", obj_log_buf);
+ }
+
+cleanup:
+ task_kfunc_failure__destroy(skel);
+}
+
+void test_task_kfunc(void)
+{
+ int i;
+
+ for (i = 0; i < ARRAY_SIZE(success_tests); i++) {
+ if (!test__start_subtest(success_tests[i]))
+ continue;
+
+ run_success_test(success_tests[i]);
+ }
+
+ for (i = 0; i < ARRAY_SIZE(failure_tests); i++) {
+ if (!test__start_subtest(failure_tests[i].prog_name))
+ continue;
+
+ verify_fail(failure_tests[i].prog_name, failure_tests[i].expected_err_msg);
+ }
+}
diff --git a/tools/testing/selftests/bpf/progs/task_kfunc_common.h b/tools/testing/selftests/bpf/progs/task_kfunc_common.h
new file mode 100644
index 000000000000..8489b952dc49
--- /dev/null
+++ b/tools/testing/selftests/bpf/progs/task_kfunc_common.h
@@ -0,0 +1,71 @@
+/* SPDX-License-Identifier: GPL-2.0 */
+/* Copyright (c) 2022 Meta Platforms, Inc. and affiliates. */
+
+#ifndef _TASK_KFUNC_COMMON_H
+#define _TASK_KFUNC_COMMON_H
+
+#include <errno.h>
+#include <vmlinux.h>
+#include <bpf/bpf_helpers.h>
+#include <bpf/bpf_tracing.h>
+
+struct __tasks_kfunc_map_value {
+ struct task_struct __kptr_ref * task;
+};
+
+struct hash_map {
+ __uint(type, BPF_MAP_TYPE_HASH);
+ __type(key, int);
+ __type(value, struct __tasks_kfunc_map_value);
+ __uint(max_entries, 1);
+} __tasks_kfunc_map SEC(".maps");
+
+struct task_struct *bpf_task_acquire(struct task_struct *p) __ksym;
+struct task_struct *bpf_task_kptr_get(struct task_struct **pp) __ksym;
+void bpf_task_release(struct task_struct *p) __ksym;
+
+static inline struct __tasks_kfunc_map_value *tasks_kfunc_map_value_lookup(struct task_struct *p)
+{
+ s32 pid;
+ long status;
+
+ status = bpf_probe_read_kernel(&pid, sizeof(pid), &p->pid);
+ if (status)
+ return NULL;
+
+ return bpf_map_lookup_elem(&__tasks_kfunc_map, &pid);
+}
+
+static inline int tasks_kfunc_map_insert(struct task_struct *p)
+{
+ struct __tasks_kfunc_map_value local, *v;
+ long status;
+ struct task_struct *acquired, *old;
+ s32 pid;
+
+ status = bpf_probe_read_kernel(&pid, sizeof(pid), &p->pid);
+ if (status)
+ return status;
+
+ local.task = NULL;
+ status = bpf_map_update_elem(&__tasks_kfunc_map, &pid, &local, BPF_NOEXIST);
+ if (status)
+ return status;
+
+ v = bpf_map_lookup_elem(&__tasks_kfunc_map, &pid);
+ if (!v) {
+ bpf_map_delete_elem(&__tasks_kfunc_map, &pid);
+ return status;
+ }
+
+ acquired = bpf_task_acquire(p);
+ old = bpf_kptr_xchg(&v->task, acquired);
+ if (old) {
+ bpf_task_release(old);
+ return -EEXIST;
+ }
+
+ return 0;
+}
+
+#endif /* _TASK_KFUNC_COMMON_H */
diff --git a/tools/testing/selftests/bpf/progs/task_kfunc_failure.c b/tools/testing/selftests/bpf/progs/task_kfunc_failure.c
new file mode 100644
index 000000000000..4c2e8a8f3544
--- /dev/null
+++ b/tools/testing/selftests/bpf/progs/task_kfunc_failure.c
@@ -0,0 +1,259 @@
+// SPDX-License-Identifier: GPL-2.0
+/* Copyright (c) 2022 Meta Platforms, Inc. and affiliates. */
+
+#include <vmlinux.h>
+#include <bpf/bpf_tracing.h>
+#include <bpf/bpf_helpers.h>
+
+#include "task_kfunc_common.h"
+
+char _license[] SEC("license") = "GPL";
+
+/* Prototype for all of the program trace events below:
+ *
+ * TRACE_EVENT(task_newtask,
+ * TP_PROTO(struct task_struct *p, u64 clone_flags)
+ */
+
+static struct __tasks_kfunc_map_value *insert_lookup_task(struct task_struct *task)
+{
+ int status;
+
+ status = tasks_kfunc_map_insert(task);
+ if (status)
+ return NULL;
+
+ return tasks_kfunc_map_value_lookup(task);
+}
+
+SEC("tp_btf/task_newtask")
+int BPF_PROG(task_kfunc_acquire_untrusted, struct task_struct *task, u64 clone_flags)
+{
+ struct task_struct *acquired;
+ struct __tasks_kfunc_map_value *v;
+
+ v = insert_lookup_task(task);
+ if (!v)
+ return 0;
+
+ /* Can't invoke bpf_task_acquire() on an untrusted pointer. */
+ acquired = bpf_task_acquire(v->task);
+ bpf_task_release(acquired);
+
+ return 0;
+}
+
+SEC("tp_btf/task_newtask")
+int BPF_PROG(task_kfunc_acquire_fp, struct task_struct *task, u64 clone_flags)
+{
+ struct task_struct *acquired, *stack_task = (struct task_struct *)&clone_flags;
+
+ /* Can't invoke bpf_task_acquire() on a random frame pointer. */
+ acquired = bpf_task_acquire((struct task_struct *)&stack_task);
+ bpf_task_release(acquired);
+
+ return 0;
+}
+
+SEC("kretprobe/free_task")
+int BPF_PROG(task_kfunc_acquire_unsafe_kretprobe, struct task_struct *task, u64 clone_flags)
+{
+ struct task_struct *acquired;
+
+ acquired = bpf_task_acquire(task);
+ /* Can't release a bpf_task_acquire()'d task without a NULL check. */
+ bpf_task_release(acquired);
+
+ return 0;
+}
+
+SEC("tp_btf/task_newtask")
+int BPF_PROG(task_kfunc_acquire_trusted_walked, struct task_struct *task, u64 clone_flags)
+{
+ struct task_struct *acquired;
+
+ /* Can't invoke bpf_task_acquire() on a trusted pointer obtained from walking a struct. */
+ acquired = bpf_task_acquire(task->last_wakee);
+ bpf_task_release(acquired);
+
+ return 0;
+}
+
+
+SEC("tp_btf/task_newtask")
+int BPF_PROG(task_kfunc_acquire_null, struct task_struct *task, u64 clone_flags)
+{
+ struct task_struct *acquired;
+
+ /* Can't invoke bpf_task_acquire() on a NULL pointer. */
+ acquired = bpf_task_acquire(NULL);
+ if (!acquired)
+ return 0;
+ bpf_task_release(acquired);
+
+ return 0;
+}
+
+SEC("tp_btf/task_newtask")
+int BPF_PROG(task_kfunc_acquire_unreleased, struct task_struct *task, u64 clone_flags)
+{
+ struct task_struct *acquired;
+
+ acquired = bpf_task_acquire(task);
+
+ /* Acquired task is never released. */
+
+ return 0;
+}
+
+SEC("tp_btf/task_newtask")
+int BPF_PROG(task_kfunc_get_non_kptr_param, struct task_struct *task, u64 clone_flags)
+{
+ struct task_struct *kptr;
+
+ /* Cannot use bpf_task_kptr_get() on a non-kptr, even on a valid task. */
+ kptr = bpf_task_kptr_get(&task);
+ if (!kptr)
+ return 0;
+
+ bpf_task_release(kptr);
+
+ return 0;
+}
+
+SEC("tp_btf/task_newtask")
+int BPF_PROG(task_kfunc_get_non_kptr_acquired, struct task_struct *task, u64 clone_flags)
+{
+ struct task_struct *kptr, *acquired;
+
+ acquired = bpf_task_acquire(task);
+
+ /* Cannot use bpf_task_kptr_get() on a non-kptr, even if it was acquired. */
+ kptr = bpf_task_kptr_get(&acquired);
+ bpf_task_release(acquired);
+ if (!kptr)
+ return 0;
+
+ bpf_task_release(kptr);
+
+ return 0;
+}
+
+SEC("tp_btf/task_newtask")
+int BPF_PROG(task_kfunc_get_null, struct task_struct *task, u64 clone_flags)
+{
+ struct task_struct *kptr;
+
+ /* Cannot use bpf_task_kptr_get() on a NULL pointer. */
+ kptr = bpf_task_kptr_get(NULL);
+ if (!kptr)
+ return 0;
+
+ bpf_task_release(kptr);
+
+ return 0;
+}
+
+SEC("tp_btf/task_newtask")
+int BPF_PROG(task_kfunc_xchg_unreleased, struct task_struct *task, u64 clone_flags)
+{
+ struct task_struct *kptr;
+ struct __tasks_kfunc_map_value *v;
+
+ v = insert_lookup_task(task);
+ if (!v)
+ return 0;
+
+ kptr = bpf_kptr_xchg(&v->task, NULL);
+ if (!kptr)
+ return 0;
+
+ /* Kptr retrieved from map is never released. */
+
+ return 0;
+}
+
+SEC("tp_btf/task_newtask")
+int BPF_PROG(task_kfunc_get_unreleased, struct task_struct *task, u64 clone_flags)
+{
+ struct task_struct *kptr;
+ struct __tasks_kfunc_map_value *v;
+
+ v = insert_lookup_task(task);
+ if (!v)
+ return 0;
+
+ kptr = bpf_task_kptr_get(&v->task);
+ if (!kptr)
+ return 0;
+
+ /* Kptr acquired above is never released. */
+
+ return 0;
+}
+
+SEC("tp_btf/task_newtask")
+int BPF_PROG(task_kfunc_release_untrusted, struct task_struct *task, u64 clone_flags)
+{
+ struct __tasks_kfunc_map_value *v;
+
+ v = insert_lookup_task(task);
+ if (!v)
+ return 0;
+
+ /* Can't invoke bpf_task_release() on an untrusted pointer. */
+ bpf_task_release(v->task);
+
+ return 0;
+}
+
+SEC("tp_btf/task_newtask")
+int BPF_PROG(task_kfunc_release_fp, struct task_struct *task, u64 clone_flags)
+{
+ struct task_struct *acquired = (struct task_struct *)&clone_flags;
+
+ /* Cannot release random frame pointer. */
+ bpf_task_release(acquired);
+
+ return 0;
+}
+
+SEC("tp_btf/task_newtask")
+int BPF_PROG(task_kfunc_release_null, struct task_struct *task, u64 clone_flags)
+{
+ struct __tasks_kfunc_map_value local, *v;
+ long status;
+ struct task_struct *acquired, *old;
+ s32 pid;
+
+ status = bpf_probe_read_kernel(&pid, sizeof(pid), &task->pid);
+ if (status)
+ return 0;
+
+ local.task = NULL;
+ status = bpf_map_update_elem(&__tasks_kfunc_map, &pid, &local, BPF_NOEXIST);
+ if (status)
+ return status;
+
+ v = bpf_map_lookup_elem(&__tasks_kfunc_map, &pid);
+ if (!v)
+ return status;
+
+ acquired = bpf_task_acquire(task);
+
+ old = bpf_kptr_xchg(&v->task, acquired);
+
+ /* old cannot be passed to bpf_task_release() without a NULL check. */
+ bpf_task_release(old);
+
+ return 0;
+}
+
+SEC("tp_btf/task_newtask")
+int BPF_PROG(task_kfunc_release_unacquired, struct task_struct *task, u64 clone_flags)
+{
+ /* Cannot release trusted task pointer which was not acquired. */
+ bpf_task_release(task);
+
+ return 0;
+}
diff --git a/tools/testing/selftests/bpf/progs/task_kfunc_success.c b/tools/testing/selftests/bpf/progs/task_kfunc_success.c
new file mode 100644
index 000000000000..be4534b5ba2e
--- /dev/null
+++ b/tools/testing/selftests/bpf/progs/task_kfunc_success.c
@@ -0,0 +1,149 @@
+// SPDX-License-Identifier: GPL-2.0
+/* Copyright (c) 2022 Meta Platforms, Inc. and affiliates. */
+
+#include <vmlinux.h>
+#include <bpf/bpf_tracing.h>
+#include <bpf/bpf_helpers.h>
+
+#include "task_kfunc_common.h"
+
+char _license[] SEC("license") = "GPL";
+
+int err, pid;
+
+/* Prototype for all of the program trace events below:
+ *
+ * TRACE_EVENT(task_newtask,
+ * TP_PROTO(struct task_struct *p, u64 clone_flags)
+ */
+
+static bool is_test_kfunc_task(void)
+{
+ int cur_pid = bpf_get_current_pid_tgid() >> 32;
+
+ return pid == cur_pid;
+}
+
+static int test_acquire_release(struct task_struct *task)
+{
+ struct task_struct *acquired;
+
+ acquired = bpf_task_acquire(task);
+ bpf_task_release(acquired);
+
+ return 0;
+}
+
+SEC("tp_btf/task_newtask")
+int BPF_PROG(test_task_acquire_release_argument, struct task_struct *task, u64 clone_flags)
+{
+ if (!is_test_kfunc_task())
+ return 0;
+
+ return test_acquire_release(task);
+}
+
+SEC("tp_btf/task_newtask")
+int BPF_PROG(test_task_acquire_release_current, struct task_struct *task, u64 clone_flags)
+{
+ if (!is_test_kfunc_task())
+ return 0;
+
+ return test_acquire_release(bpf_get_current_task_btf());
+}
+
+SEC("tp_btf/task_newtask")
+int BPF_PROG(test_task_acquire_leave_in_map, struct task_struct *task, u64 clone_flags)
+{
+ long status;
+
+ if (!is_test_kfunc_task())
+ return 0;
+
+ status = tasks_kfunc_map_insert(task);
+ if (status)
+ err = 1;
+
+ return 0;
+}
+
+SEC("tp_btf/task_newtask")
+int BPF_PROG(test_task_xchg_release, struct task_struct *task, u64 clone_flags)
+{
+ struct task_struct *kptr;
+ struct __tasks_kfunc_map_value *v;
+ long status;
+
+ if (!is_test_kfunc_task())
+ return 0;
+
+ status = tasks_kfunc_map_insert(task);
+ if (status) {
+ err = 1;
+ return 0;
+ }
+
+ v = tasks_kfunc_map_value_lookup(task);
+ if (!v) {
+ err = 2;
+ return 0;
+ }
+
+ kptr = bpf_kptr_xchg(&v->task, NULL);
+ if (!kptr) {
+ err = 3;
+ return 0;
+ }
+
+ bpf_task_release(kptr);
+
+ return 0;
+}
+
+SEC("tp_btf/task_newtask")
+int BPF_PROG(test_task_get_release, struct task_struct *task, u64 clone_flags)
+{
+ struct task_struct *kptr;
+ struct __tasks_kfunc_map_value *v;
+ long status;
+
+ if (!is_test_kfunc_task())
+ return 0;
+
+ status = tasks_kfunc_map_insert(task);
+ if (status) {
+ err = 1;
+ return 0;
+ }
+
+ v = tasks_kfunc_map_value_lookup(task);
+ if (!v) {
+ err = 2;
+ return 0;
+ }
+
+ kptr = bpf_task_kptr_get(&v->task);
+ if (!kptr) {
+ err = 3;
+ return 0;
+ }
+
+ bpf_task_release(kptr);
+
+ return 0;
+}
+
+SEC("tp_btf/task_newtask")
+int BPF_PROG(test_task_current_acquire_release, struct task_struct *task, u64 clone_flags)
+{
+ struct task_struct *current, *acquired;
+
+ if (!is_test_kfunc_task())
+ return 0;
+
+ current = bpf_get_current_task_btf();
+ acquired = bpf_task_acquire(current);
+ bpf_task_release(acquired);
+
+ return 0;
+}
--
2.38.1
On Wed, Nov 16, 2022 at 09:24:02PM -0600, David Vernet wrote:
> A previous change added a series of kfuncs for storing struct
> task_struct objects as referenced kptrs. This patch adds a new
> task_kfunc test suite for validating their expected behavior.
>
> Signed-off-by: David Vernet <[email protected]>
> ---
> tools/testing/selftests/bpf/DENYLIST.s390x | 1 +
> .../selftests/bpf/prog_tests/task_kfunc.c | 159 +++++++++++
> .../selftests/bpf/progs/task_kfunc_common.h | 71 +++++
> .../selftests/bpf/progs/task_kfunc_failure.c | 259 ++++++++++++++++++
> .../selftests/bpf/progs/task_kfunc_success.c | 149 ++++++++++
> 5 files changed, 639 insertions(+)
> create mode 100644 tools/testing/selftests/bpf/prog_tests/task_kfunc.c
> create mode 100644 tools/testing/selftests/bpf/progs/task_kfunc_common.h
> create mode 100644 tools/testing/selftests/bpf/progs/task_kfunc_failure.c
> create mode 100644 tools/testing/selftests/bpf/progs/task_kfunc_success.c
>
> diff --git a/tools/testing/selftests/bpf/DENYLIST.s390x b/tools/testing/selftests/bpf/DENYLIST.s390x
> index be4e3d47ea3e..97e522828d6c 100644
> --- a/tools/testing/selftests/bpf/DENYLIST.s390x
> +++ b/tools/testing/selftests/bpf/DENYLIST.s390x
> @@ -53,6 +53,7 @@ skc_to_unix_sock # could not attach BPF object unexpecte
> socket_cookie # prog_attach unexpected error: -524 (trampoline)
> stacktrace_build_id # compare_map_keys stackid_hmap vs. stackmap err -2 errno 2 (?)
> tailcalls # tail_calls are not allowed in non-JITed programs with bpf-to-bpf calls (?)
> +task_kfunc # JIT does not support calling kernel function
> task_local_storage # failed to auto-attach program 'trace_exit_creds': -524 (trampoline)
> test_bpffs # bpffs test failed 255 (iterator)
> test_bprm_opts # failed to auto-attach program 'secure_exec': -524 (trampoline)
> diff --git a/tools/testing/selftests/bpf/prog_tests/task_kfunc.c b/tools/testing/selftests/bpf/prog_tests/task_kfunc.c
> new file mode 100644
> index 000000000000..a3aa36a4beb3
> --- /dev/null
> +++ b/tools/testing/selftests/bpf/prog_tests/task_kfunc.c
> @@ -0,0 +1,159 @@
> +// SPDX-License-Identifier: GPL-2.0
> +/* Copyright (c) 2022 Meta Platforms, Inc. and affiliates. */
> +
> +#define _GNU_SOURCE
> +#include <sys/wait.h>
> +#include <test_progs.h>
> +#include <unistd.h>
> +
> +#include "task_kfunc_failure.skel.h"
> +#include "task_kfunc_success.skel.h"
> +
> +static size_t log_buf_sz = 1 << 20; /* 1 MB */
> +static char obj_log_buf[1048576];
> +
> +static struct task_kfunc_success *open_load_task_kfunc_skel(void)
> +{
> + struct task_kfunc_success *skel;
> + int err;
> +
> + skel = task_kfunc_success__open();
> + if (!ASSERT_OK_PTR(skel, "skel_open"))
> + return NULL;
> +
> + skel->bss->pid = getpid();
> +
> + err = task_kfunc_success__load(skel);
> + if (!ASSERT_OK(err, "skel_load"))
> + goto cleanup;
> +
> + return skel;
> +
> +cleanup:
> + task_kfunc_success__destroy(skel);
> + return NULL;
> +}
> +
> +static void run_success_test(const char *prog_name)
> +{
> + struct task_kfunc_success *skel;
> + int status;
> + pid_t child_pid;
> + struct bpf_program *prog;
> + struct bpf_link *link = NULL;
> +
> + skel = open_load_task_kfunc_skel();
> + if (!ASSERT_OK_PTR(skel, "open_load_skel"))
> + return;
> +
> + if (!ASSERT_OK(skel->bss->err, "pre_spawn_err"))
> + goto cleanup;
> +
> + prog = bpf_object__find_program_by_name(skel->obj, prog_name);
> + if (!ASSERT_OK_PTR(prog, "bpf_object__find_program_by_name"))
> + goto cleanup;
> +
> + link = bpf_program__attach(prog);
> + if (!ASSERT_OK_PTR(link, "attached_link"))
> + goto cleanup;
> +
> + child_pid = fork();
> + if (!ASSERT_GT(child_pid, -1, "child_pid"))
> + goto cleanup;
> + if (child_pid == 0)
> + _exit(0);
> + waitpid(child_pid, &status, 0);
> +
> + ASSERT_OK(skel->bss->err, "post_wait_err");
> +
> +cleanup:
> + bpf_link__destroy(link);
> + task_kfunc_success__destroy(skel);
> +}
> +
> +static const char * const success_tests[] = {
> + "test_task_acquire_release_argument",
> + "test_task_acquire_release_current",
> + "test_task_acquire_leave_in_map",
> + "test_task_xchg_release",
> + "test_task_get_release",
> + "test_task_current_acquire_release",
> +};
> +
> +static struct {
> + const char *prog_name;
> + const char *expected_err_msg;
> +} failure_tests[] = {
> + {"task_kfunc_acquire_untrusted", "R1 must be referenced or trusted"},
> + {"task_kfunc_acquire_fp", "arg#0 pointer type STRUCT task_struct must point"},
> + {"task_kfunc_acquire_unsafe_kretprobe", "reg type unsupported for arg#0 function"},
> + {"task_kfunc_acquire_trusted_walked", "R1 must be referenced or trusted"},
> + {"task_kfunc_acquire_null", "arg#0 pointer type STRUCT task_struct must point"},
> + {"task_kfunc_acquire_unreleased", "Unreleased reference"},
> + {"task_kfunc_get_non_kptr_param", "arg#0 expected pointer to map value"},
> + {"task_kfunc_get_non_kptr_acquired", "arg#0 expected pointer to map value"},
> + {"task_kfunc_get_null", "arg#0 expected pointer to map value"},
> + {"task_kfunc_xchg_unreleased", "Unreleased reference"},
> + {"task_kfunc_get_unreleased", "Unreleased reference"},
> + {"task_kfunc_release_untrusted", "bpf_task_release arg#0 pointer had unexpected modifiers"},
> + {"task_kfunc_release_fp", "arg#0 pointer type STRUCT task_struct must point"},
> + {"task_kfunc_release_null", "bpf_task_release arg#0 pointer had unexpected modifiers"},
> + {"task_kfunc_release_unacquired", "release kernel function bpf_task_release expects"},
> +};
> +
> +static void verify_fail(const char *prog_name, const char *expected_err_msg)
> +{
> + LIBBPF_OPTS(bpf_object_open_opts, opts);
> + struct task_kfunc_failure *skel;
> + int err, i;
> +
> + opts.kernel_log_buf = obj_log_buf;
> + opts.kernel_log_size = log_buf_sz;
> + opts.kernel_log_level = 1;
> +
> + skel = task_kfunc_failure__open_opts(&opts);
> + if (!ASSERT_OK_PTR(skel, "task_kfunc_failure__open_opts"))
> + goto cleanup;
> +
> + for (i = 0; i < ARRAY_SIZE(failure_tests); i++) {
> + struct bpf_program *prog;
> + const char *curr_name = failure_tests[i].prog_name;
> +
> + prog = bpf_object__find_program_by_name(skel->obj, curr_name);
> + if (!ASSERT_OK_PTR(prog, "bpf_object__find_program_by_name"))
> + goto cleanup;
> +
> + bpf_program__set_autoload(prog, !strcmp(curr_name, prog_name));
> + }
> +
> + err = task_kfunc_failure__load(skel);
> + if (!ASSERT_ERR(err, "unexpected load success"))
> + goto cleanup;
> +
> + if (!ASSERT_OK_PTR(strstr(obj_log_buf, expected_err_msg), "expected_err_msg")) {
> + fprintf(stderr, "Expected err_msg: %s\n", expected_err_msg);
> + fprintf(stderr, "Verifier output: %s\n", obj_log_buf);
> + }
> +
> +cleanup:
> + task_kfunc_failure__destroy(skel);
> +}
> +
> +void test_task_kfunc(void)
> +{
> + int i;
> +
> + for (i = 0; i < ARRAY_SIZE(success_tests); i++) {
> + if (!test__start_subtest(success_tests[i]))
> + continue;
> +
> + run_success_test(success_tests[i]);
> + }
> +
> + for (i = 0; i < ARRAY_SIZE(failure_tests); i++) {
> + if (!test__start_subtest(failure_tests[i].prog_name))
> + continue;
> +
> + verify_fail(failure_tests[i].prog_name, failure_tests[i].expected_err_msg);
> + }
> +}
> diff --git a/tools/testing/selftests/bpf/progs/task_kfunc_common.h b/tools/testing/selftests/bpf/progs/task_kfunc_common.h
> new file mode 100644
> index 000000000000..8489b952dc49
> --- /dev/null
> +++ b/tools/testing/selftests/bpf/progs/task_kfunc_common.h
> @@ -0,0 +1,71 @@
> +/* SPDX-License-Identifier: GPL-2.0 */
> +/* Copyright (c) 2022 Meta Platforms, Inc. and affiliates. */
> +
> +#ifndef _TASK_KFUNC_COMMON_H
> +#define _TASK_KFUNC_COMMON_H
> +
> +#include <errno.h>
> +#include <vmlinux.h>
> +#include <bpf/bpf_helpers.h>
> +#include <bpf/bpf_tracing.h>
> +
> +struct __tasks_kfunc_map_value {
> + struct task_struct __kptr_ref * task;
> +};
> +
> +struct hash_map {
> + __uint(type, BPF_MAP_TYPE_HASH);
> + __type(key, int);
> + __type(value, struct __tasks_kfunc_map_value);
> + __uint(max_entries, 1);
> +} __tasks_kfunc_map SEC(".maps");
> +
> +struct task_struct *bpf_task_acquire(struct task_struct *p) __ksym;
> +struct task_struct *bpf_task_kptr_get(struct task_struct **pp) __ksym;
> +void bpf_task_release(struct task_struct *p) __ksym;
> +
> +static inline struct __tasks_kfunc_map_value *tasks_kfunc_map_value_lookup(struct task_struct *p)
> +{
> + s32 pid;
> + long status;
> +
> + status = bpf_probe_read_kernel(&pid, sizeof(pid), &p->pid);
> + if (status)
> + return NULL;
> +
> + return bpf_map_lookup_elem(&__tasks_kfunc_map, &pid);
> +}
> +
> +static inline int tasks_kfunc_map_insert(struct task_struct *p)
> +{
> + struct __tasks_kfunc_map_value local, *v;
> + long status;
> + struct task_struct *acquired, *old;
> + s32 pid;
> +
> + status = bpf_probe_read_kernel(&pid, sizeof(pid), &p->pid);
> + if (status)
> + return status;
> +
> + local.task = NULL;
> + status = bpf_map_update_elem(&__tasks_kfunc_map, &pid, &local, BPF_NOEXIST);
> + if (status)
> + return status;
> +
> + v = bpf_map_lookup_elem(&__tasks_kfunc_map, &pid);
> + if (!v) {
> + bpf_map_delete_elem(&__tasks_kfunc_map, &pid);
> + return status;
here it will return 0, but probably should be returning error?
> + }
> +
> + acquired = bpf_task_acquire(p);
> + old = bpf_kptr_xchg(&v->task, acquired);
> + if (old) {
> + bpf_task_release(old);
> + return -EEXIST;
> + }
> +
> + return 0;
> +}
> +
> +#endif /* _TASK_KFUNC_COMMON_H */
> diff --git a/tools/testing/selftests/bpf/progs/task_kfunc_failure.c b/tools/testing/selftests/bpf/progs/task_kfunc_failure.c
> new file mode 100644
> index 000000000000..4c2e8a8f3544
> --- /dev/null
> +++ b/tools/testing/selftests/bpf/progs/task_kfunc_failure.c
> @@ -0,0 +1,259 @@
> +// SPDX-License-Identifier: GPL-2.0
> +/* Copyright (c) 2022 Meta Platforms, Inc. and affiliates. */
> +
> +#include <vmlinux.h>
> +#include <bpf/bpf_tracing.h>
> +#include <bpf/bpf_helpers.h>
> +
> +#include "task_kfunc_common.h"
> +
> +char _license[] SEC("license") = "GPL";
> +
> +/* Prototype for all of the program trace events below:
> + *
> + * TRACE_EVENT(task_newtask,
> + * TP_PROTO(struct task_struct *p, u64 clone_flags)
> + */
> +
> +static struct __tasks_kfunc_map_value *insert_lookup_task(struct task_struct *task)
> +{
> + int status;
> +
> + status = tasks_kfunc_map_insert(task);
> + if (status)
> + return NULL;
> +
> + return tasks_kfunc_map_value_lookup(task);
> +}
> +
> +SEC("tp_btf/task_newtask")
> +int BPF_PROG(task_kfunc_acquire_untrusted, struct task_struct *task, u64 clone_flags)
> +{
> + struct task_struct *acquired;
> + struct __tasks_kfunc_map_value *v;
> +
> + v = insert_lookup_task(task);
> + if (!v)
> + return 0;
> +
> + /* Can't invoke bpf_task_acquire() on an untrusted pointer. */
> + acquired = bpf_task_acquire(v->task);
> + bpf_task_release(acquired);
> +
> + return 0;
> +}
> +
> +SEC("tp_btf/task_newtask")
> +int BPF_PROG(task_kfunc_acquire_fp, struct task_struct *task, u64 clone_flags)
> +{
> + struct task_struct *acquired, *stack_task = (struct task_struct *)&clone_flags;
> +
> + /* Can't invoke bpf_task_acquire() on a random frame pointer. */
> + acquired = bpf_task_acquire((struct task_struct *)&stack_task);
> + bpf_task_release(acquired);
> +
> + return 0;
> +}
> +
> +SEC("kretprobe/free_task")
> +int BPF_PROG(task_kfunc_acquire_unsafe_kretprobe, struct task_struct *task, u64 clone_flags)
> +{
> + struct task_struct *acquired;
> +
> + acquired = bpf_task_acquire(task);
> + /* Can't release a bpf_task_acquire()'d task without a NULL check. */
> + bpf_task_release(acquired);
> +
> + return 0;
> +}
> +
> +SEC("tp_btf/task_newtask")
> +int BPF_PROG(task_kfunc_acquire_trusted_walked, struct task_struct *task, u64 clone_flags)
> +{
> + struct task_struct *acquired;
> +
> + /* Can't invoke bpf_task_acquire() on a trusted pointer obtained from walking a struct. */
> + acquired = bpf_task_acquire(task->last_wakee);
> + bpf_task_release(acquired);
> +
> + return 0;
> +}
> +
> +
> +SEC("tp_btf/task_newtask")
> +int BPF_PROG(task_kfunc_acquire_null, struct task_struct *task, u64 clone_flags)
> +{
> + struct task_struct *acquired;
> +
> + /* Can't invoke bpf_task_acquire() on a NULL pointer. */
> + acquired = bpf_task_acquire(NULL);
> + if (!acquired)
> + return 0;
> + bpf_task_release(acquired);
> +
> + return 0;
> +}
> +
> +SEC("tp_btf/task_newtask")
> +int BPF_PROG(task_kfunc_acquire_unreleased, struct task_struct *task, u64 clone_flags)
> +{
> + struct task_struct *acquired;
> +
> + acquired = bpf_task_acquire(task);
> +
> + /* Acquired task is never released. */
> +
> + return 0;
> +}
> +
> +SEC("tp_btf/task_newtask")
> +int BPF_PROG(task_kfunc_get_non_kptr_param, struct task_struct *task, u64 clone_flags)
> +{
> + struct task_struct *kptr;
> +
> + /* Cannot use bpf_task_kptr_get() on a non-kptr, even on a valid task. */
> + kptr = bpf_task_kptr_get(&task);
> + if (!kptr)
> + return 0;
> +
> + bpf_task_release(kptr);
> +
> + return 0;
> +}
> +
> +SEC("tp_btf/task_newtask")
> +int BPF_PROG(task_kfunc_get_non_kptr_acquired, struct task_struct *task, u64 clone_flags)
> +{
> + struct task_struct *kptr, *acquired;
> +
> + acquired = bpf_task_acquire(task);
> +
> + /* Cannot use bpf_task_kptr_get() on a non-kptr, even if it was acquired. */
> + kptr = bpf_task_kptr_get(&acquired);
> + bpf_task_release(acquired);
> + if (!kptr)
> + return 0;
> +
> + bpf_task_release(kptr);
> +
> + return 0;
> +}
> +
> +SEC("tp_btf/task_newtask")
> +int BPF_PROG(task_kfunc_get_null, struct task_struct *task, u64 clone_flags)
> +{
> + struct task_struct *kptr;
> +
> + /* Cannot use bpf_task_kptr_get() on a NULL pointer. */
> + kptr = bpf_task_kptr_get(NULL);
> + if (!kptr)
> + return 0;
> +
> + bpf_task_release(kptr);
> +
> + return 0;
> +}
> +
> +SEC("tp_btf/task_newtask")
> +int BPF_PROG(task_kfunc_xchg_unreleased, struct task_struct *task, u64 clone_flags)
> +{
> + struct task_struct *kptr;
> + struct __tasks_kfunc_map_value *v;
> +
> + v = insert_lookup_task(task);
> + if (!v)
> + return 0;
> +
> + kptr = bpf_kptr_xchg(&v->task, NULL);
> + if (!kptr)
> + return 0;
> +
> + /* Kptr retrieved from map is never released. */
> +
> + return 0;
> +}
> +
> +SEC("tp_btf/task_newtask")
> +int BPF_PROG(task_kfunc_get_unreleased, struct task_struct *task, u64 clone_flags)
> +{
> + struct task_struct *kptr;
> + struct __tasks_kfunc_map_value *v;
> +
> + v = insert_lookup_task(task);
> + if (!v)
> + return 0;
> +
> + kptr = bpf_task_kptr_get(&v->task);
> + if (!kptr)
> + return 0;
> +
> + /* Kptr acquired above is never released. */
> +
> + return 0;
> +}
> +
> +SEC("tp_btf/task_newtask")
> +int BPF_PROG(task_kfunc_release_untrusted, struct task_struct *task, u64 clone_flags)
> +{
> + struct __tasks_kfunc_map_value *v;
> +
> + v = insert_lookup_task(task);
> + if (!v)
> + return 0;
> +
> + /* Can't invoke bpf_task_release() on an untrusted pointer. */
> + bpf_task_release(v->task);
> +
> + return 0;
> +}
> +
> +SEC("tp_btf/task_newtask")
> +int BPF_PROG(task_kfunc_release_fp, struct task_struct *task, u64 clone_flags)
> +{
> + struct task_struct *acquired = (struct task_struct *)&clone_flags;
> +
> + /* Cannot release random frame pointer. */
> + bpf_task_release(acquired);
> +
> + return 0;
> +}
> +
> +SEC("tp_btf/task_newtask")
> +int BPF_PROG(task_kfunc_release_null, struct task_struct *task, u64 clone_flags)
> +{
> + struct __tasks_kfunc_map_value local, *v;
> + long status;
> + struct task_struct *acquired, *old;
> + s32 pid;
> +
> + status = bpf_probe_read_kernel(&pid, sizeof(pid), &task->pid);
> + if (status)
> + return 0;
> +
> + local.task = NULL;
> + status = bpf_map_update_elem(&__tasks_kfunc_map, &pid, &local, BPF_NOEXIST);
> + if (status)
> + return status;
> +
> + v = bpf_map_lookup_elem(&__tasks_kfunc_map, &pid);
> + if (!v)
> + return status;
should be return error instead?
> +
> + acquired = bpf_task_acquire(task);
> +
> + old = bpf_kptr_xchg(&v->task, acquired);
> +
> + /* old cannot be passed to bpf_task_release() without a NULL check. */
> + bpf_task_release(old);
> +
> + return 0;
> +}
> +
> +SEC("tp_btf/task_newtask")
> +int BPF_PROG(task_kfunc_release_unacquired, struct task_struct *task, u64 clone_flags)
> +{
> + /* Cannot release trusted task pointer which was not acquired. */
> + bpf_task_release(task);
> +
> + return 0;
> +}
> diff --git a/tools/testing/selftests/bpf/progs/task_kfunc_success.c b/tools/testing/selftests/bpf/progs/task_kfunc_success.c
> new file mode 100644
> index 000000000000..be4534b5ba2e
> --- /dev/null
> +++ b/tools/testing/selftests/bpf/progs/task_kfunc_success.c
> @@ -0,0 +1,149 @@
> +// SPDX-License-Identifier: GPL-2.0
> +/* Copyright (c) 2022 Meta Platforms, Inc. and affiliates. */
> +
> +#include <vmlinux.h>
> +#include <bpf/bpf_tracing.h>
> +#include <bpf/bpf_helpers.h>
> +
> +#include "task_kfunc_common.h"
> +
> +char _license[] SEC("license") = "GPL";
> +
> +int err, pid;
> +
> +/* Prototype for all of the program trace events below:
> + *
> + * TRACE_EVENT(task_newtask,
> + * TP_PROTO(struct task_struct *p, u64 clone_flags)
> + */
> +
> +static bool is_test_kfunc_task(void)
> +{
> + int cur_pid = bpf_get_current_pid_tgid() >> 32;
> +
> + return pid == cur_pid;
> +}
> +
> +static int test_acquire_release(struct task_struct *task)
> +{
> + struct task_struct *acquired;
> +
> + acquired = bpf_task_acquire(task);
> + bpf_task_release(acquired);
> +
> + return 0;
> +}
> +
> +SEC("tp_btf/task_newtask")
> +int BPF_PROG(test_task_acquire_release_argument, struct task_struct *task, u64 clone_flags)
> +{
> + if (!is_test_kfunc_task())
> + return 0;
> +
> + return test_acquire_release(task);
> +}
> +
> +SEC("tp_btf/task_newtask")
> +int BPF_PROG(test_task_acquire_release_current, struct task_struct *task, u64 clone_flags)
> +{
> + if (!is_test_kfunc_task())
> + return 0;
> +
> + return test_acquire_release(bpf_get_current_task_btf());
> +}
> +
> +SEC("tp_btf/task_newtask")
> +int BPF_PROG(test_task_acquire_leave_in_map, struct task_struct *task, u64 clone_flags)
> +{
> + long status;
> +
> + if (!is_test_kfunc_task())
> + return 0;
> +
> + status = tasks_kfunc_map_insert(task);
> + if (status)
> + err = 1;
> +
> + return 0;
> +}
> +
> +SEC("tp_btf/task_newtask")
> +int BPF_PROG(test_task_xchg_release, struct task_struct *task, u64 clone_flags)
> +{
> + struct task_struct *kptr;
> + struct __tasks_kfunc_map_value *v;
> + long status;
> +
> + if (!is_test_kfunc_task())
> + return 0;
> +
> + status = tasks_kfunc_map_insert(task);
> + if (status) {
> + err = 1;
> + return 0;
> + }
> +
> + v = tasks_kfunc_map_value_lookup(task);
> + if (!v) {
> + err = 2;
> + return 0;
> + }
> +
> + kptr = bpf_kptr_xchg(&v->task, NULL);
> + if (!kptr) {
> + err = 3;
> + return 0;
> + }
> +
> + bpf_task_release(kptr);
> +
> + return 0;
> +}
> +
> +SEC("tp_btf/task_newtask")
> +int BPF_PROG(test_task_get_release, struct task_struct *task, u64 clone_flags)
> +{
> + struct task_struct *kptr;
> + struct __tasks_kfunc_map_value *v;
> + long status;
> +
> + if (!is_test_kfunc_task())
> + return 0;
> +
> + status = tasks_kfunc_map_insert(task);
> + if (status) {
> + err = 1;
> + return 0;
> + }
> +
> + v = tasks_kfunc_map_value_lookup(task);
> + if (!v) {
> + err = 2;
> + return 0;
> + }
> +
> + kptr = bpf_task_kptr_get(&v->task);
> + if (!kptr) {
> + err = 3;
> + return 0;
> + }
> +
> + bpf_task_release(kptr);
> +
> + return 0;
> +}
> +
> +SEC("tp_btf/task_newtask")
> +int BPF_PROG(test_task_current_acquire_release, struct task_struct *task, u64 clone_flags)
> +{
> + struct task_struct *current, *acquired;
> +
> + if (!is_test_kfunc_task())
> + return 0;
> +
> + current = bpf_get_current_task_btf();
> + acquired = bpf_task_acquire(current);
> + bpf_task_release(acquired);
> +
> + return 0;
> +}
> --
> 2.38.1
>
On Thu, Nov 17, 2022 at 06:21:00PM -0800, Alexei Starovoitov wrote:
[...]
> > +static inline int tasks_kfunc_map_insert(struct task_struct *p)
> > +{
> > + struct __tasks_kfunc_map_value local, *v;
> > + long status;
> > + struct task_struct *acquired, *old;
> > + s32 pid;
> > +
> > + status = bpf_probe_read_kernel(&pid, sizeof(pid), &p->pid);
> > + if (status)
> > + return status;
> > +
> > + local.task = NULL;
> > + status = bpf_map_update_elem(&__tasks_kfunc_map, &pid, &local, BPF_NOEXIST);
> > + if (status)
> > + return status;
> > +
> > + v = bpf_map_lookup_elem(&__tasks_kfunc_map, &pid);
> > + if (!v) {
> > + bpf_map_delete_elem(&__tasks_kfunc_map, &pid);
> > + return status;
>
> here it will return 0, but probably should be returning error?
Ah, yes this should be returning -ENOENT. Thanks for catching this.
[...]
> > +SEC("tp_btf/task_newtask")
> > +int BPF_PROG(task_kfunc_release_null, struct task_struct *task, u64 clone_flags)
> > +{
> > + struct __tasks_kfunc_map_value local, *v;
> > + long status;
> > + struct task_struct *acquired, *old;
> > + s32 pid;
> > +
> > + status = bpf_probe_read_kernel(&pid, sizeof(pid), &task->pid);
> > + if (status)
> > + return 0;
> > +
> > + local.task = NULL;
> > + status = bpf_map_update_elem(&__tasks_kfunc_map, &pid, &local, BPF_NOEXIST);
> > + if (status)
> > + return status;
> > +
> > + v = bpf_map_lookup_elem(&__tasks_kfunc_map, &pid);
> > + if (!v)
> > + return status;
>
> should be return error instead?
Yep, here as well.
I'll fix both of these in v8.
[...]