Received: by 2002:a05:6500:1b41:b0:1fb:d597:ff75 with SMTP id cz1csp62760lqb; Tue, 4 Jun 2024 05:19:08 -0700 (PDT) X-Forwarded-Encrypted: i=3; AJvYcCXPfhZ3wDWL+8LYmn++F59AKdj1HrM9+M84mKxLygn+SnGskAXEh7Ikeb5Mr+1WTltQA9v8yOQGvs+FpKdepklN4lzgqERjV1nsZMrkPQ== X-Google-Smtp-Source: AGHT+IGmxrtSwSReuOMMla/w2rL8dXs0cy+lms3/N3ncQ1jtS3Par1M5pkcuSxUSPvDSKAa30Del X-Received: by 2002:a05:6a20:3d94:b0:1ad:7e4f:9564 with SMTP id adf61e73a8af0-1b2a2bda03cmr3694735637.9.1717503548478; Tue, 04 Jun 2024 05:19:08 -0700 (PDT) ARC-Seal: i=2; a=rsa-sha256; t=1717503548; cv=pass; d=google.com; s=arc-20160816; b=LNp6bBEhI1M6q8610x5W/c1R4c5lpE+MX6PNtYKgMJ0Lr3LVPbsntqi4ODJ5fNvTJT df9q/5N40iluJT2/o+tBftnBigbB2Csk43VjW8MRKC5JJhzOduoIMN4mxjGXktC6jTiH jVGZI8gtEG3sjKXo0lNVqrO+SzdoHs29apDq+f1TAps0soWHLxng28ockWjLZ/MSyon3 E+cjS3JZ54OmOHb/IybKfAdWBZqEsujkbH947w+lQORlqZyOYLPOq5rZTAjOJD+YpwOk Mt/zF938TmapDhdI710oRDtInsxa48OGKN228CY/+KAHC7GWv1z/fsXBEMJz1GbnWbJU 3uEA== ARC-Message-Signature: i=2; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=arc-20160816; h=content-transfer-encoding:mime-version:list-unsubscribe :list-subscribe:list-id:precedence:references:in-reply-to:message-id :subject:cc:to:from:date; bh=A6RuIRmxkYTPt6Hi1Y7GO4WBpi97SVAhjeIyK4fIsGY=; fh=lX5vvqxvF/oxfWNSA0bBkJteKCrzqbAvGXbxYf2FJd4=; b=Sj0vYOkYx8VcPcZe1rhgtSjICim3uglSokEJUXY4jb62jjEBGrTAVHncMMZmxRxTGL hxQbITTnKoQdfu6+1okLW8jpV8aC7zhqF8B7suJd8wjGNS9e6C/R8QVF8bErhqh+0kdt FXqmYmgPK8NgTBvUYLOw7llMOUsfhMmnJKG5brDADaljhSEHBovA6tL/QqmKywjSU901 ea1baeMue5K45elx/sxAQ+n+ofVnRx5Ao5P5Z1crQVU66MQZJswW0oHwpmwNXj15XKcL nWbfUBxxKdtSXQh9bcaMYhdrnDqiSnDtXyR3rMyGZsQPXkUnKf4GJ+D/EqJFv/yWU5m4 jzWA==; dara=google.com ARC-Authentication-Results: i=2; mx.google.com; arc=pass (i=1); spf=pass (google.com: domain of linux-kernel+bounces-200576-linux.lists.archive=gmail.com@vger.kernel.org designates 139.178.88.99 as permitted sender) smtp.mailfrom="linux-kernel+bounces-200576-linux.lists.archive=gmail.com@vger.kernel.org" Return-Path: Received: from sv.mirrors.kernel.org (sv.mirrors.kernel.org. [139.178.88.99]) by mx.google.com with ESMTPS id 41be03b00d2f7-6c35b40c4d9si8037209a12.531.2024.06.04.05.19.08 for (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Tue, 04 Jun 2024 05:19:08 -0700 (PDT) Received-SPF: pass (google.com: domain of linux-kernel+bounces-200576-linux.lists.archive=gmail.com@vger.kernel.org designates 139.178.88.99 as permitted sender) client-ip=139.178.88.99; Authentication-Results: mx.google.com; arc=pass (i=1); spf=pass (google.com: domain of linux-kernel+bounces-200576-linux.lists.archive=gmail.com@vger.kernel.org designates 139.178.88.99 as permitted sender) smtp.mailfrom="linux-kernel+bounces-200576-linux.lists.archive=gmail.com@vger.kernel.org" Received: from smtp.subspace.kernel.org (wormhole.subspace.kernel.org [52.25.139.140]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by sv.mirrors.kernel.org (Postfix) with ESMTPS id 14E9A283DE3 for ; Tue, 4 Jun 2024 12:19:08 +0000 (UTC) Received: from localhost.localdomain (localhost.localdomain [127.0.0.1]) by smtp.subspace.kernel.org (Postfix) with ESMTP id 7D3B6145B3B; Tue, 4 Jun 2024 12:18:51 +0000 (UTC) Received: from smtp.kernel.org (aws-us-west-2-korg-mail-1.web.codeaurora.org [10.30.226.201]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by smtp.subspace.kernel.org (Postfix) with ESMTPS id 78FFA145FE5; Tue, 4 Jun 2024 12:18:50 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=none smtp.client-ip=10.30.226.201 ARC-Seal:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1717503530; cv=none; b=UxbsecDFyrzX++yPAE0++JjOwU9OD6XBJHHoCcxoDrx9+gjhIAbdc5iYJa8R2nsRJM3Dsw8KiGy+gOw7Yvz0122p9pUn+MH4rhcikklPzGJpcZuv9MX2kBwHX3W/2roSH2TMW01dW7qbRCEy/zfWOrFKAs3ULRnjcVsubDYLECQ= ARC-Message-Signature:i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1717503530; c=relaxed/simple; bh=yKLV9ebaPYeEfpDgUNiv/3ezmojjTquRFzNXbNDybMI=; h=Date:From:To:Cc:Subject:Message-ID:In-Reply-To:References: MIME-Version:Content-Type; b=sdEIJNf/I2dJ9UT5tF3F02X8IoOvd8lowCpgdbcFsOoK2b4uLDQIFbjKCFe/otEanuD86mwfkZu6j9Nalhu+/8ZSHz2rsMuOvwIXQWsTIJbtClheyR6aL3hYkq5CqokKU1TrRvdfrGSYWiuS1EFGnloAB173jJ2f5HwahuzGtak= ARC-Authentication-Results:i=1; smtp.subspace.kernel.org; arc=none smtp.client-ip=10.30.226.201 Received: by smtp.kernel.org (Postfix) with ESMTPSA id 75478C4AF09; Tue, 4 Jun 2024 12:18:46 +0000 (UTC) Date: Tue, 4 Jun 2024 08:18:50 -0400 From: Steven Rostedt To: Masami Hiramatsu Cc: linux-kernel@vger.kernel.org, linux-trace-kernel@vger.kernel.org, Mark Rutland , Mathieu Desnoyers , Andrew Morton , Alexei Starovoitov , Florent Revest , Martin KaFai Lau , bpf , Sven Schnelle , Alexei Starovoitov , Jiri Olsa , Arnaldo Carvalho de Melo , Daniel Borkmann , Alan Maguire , Peter Zijlstra , Thomas Gleixner , Guo Ren Subject: Re: [PATCH v3 00/27] function_graph: Allow multiple users for function graph tracing Message-ID: <20240604081850.59267aa9@rorschach.local.home> In-Reply-To: <20240603190704.663840775@goodmis.org> References: <20240603190704.663840775@goodmis.org> X-Mailer: Claws Mail 3.17.8 (GTK+ 2.24.33; x86_64-pc-linux-gnu) Precedence: bulk X-Mailing-List: linux-kernel@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 Content-Type: text/plain; charset=US-ASCII Content-Transfer-Encoding: 7bit Masami, This series passed all my tests, are you comfortable with me pushing them to linux-next? -- Steve On Mon, 03 Jun 2024 15:07:04 -0400 Steven Rostedt wrote: > This is a continuation of the function graph multi user code. > I wrote a proof of concept back in 2019 of this code[1] and > Masami started cleaning it up. I started from Masami's work v10 > that can be found here: > > https://lore.kernel.org/linux-trace-kernel/171509088006.162236.7227326999861366050.stgit@devnote2/ > > This is *only* the code that allows multiple users of function > graph tracing. This is not the fprobe work that Masami is working > to add on top of it. As Masami took my proof of concept, there > was still several things I disliked about that code. Instead of > having Masami clean it up even more, I decided to take over on just > my code and change it up a bit. > > Changes since v2: https://lore.kernel.org/linux-trace-kernel/20240602033744.563858532@goodmis.org > > - Added comments describing which hashes the append and intersect > functions were used for. > > - Replaced checks of (NULL or EMPTY_HASH) with ftrace_hash_empty() > helper function. > > - Added check at the end of intersect_hash() to convert the hash > to EMPTY hash if it doesn't have any functions. > > - Renamed compare_ops() to ops_equal() and return boolean (inversed return > value). > > - Broke out __ftrace_hash_move_and_update_ops() to use in both > ftrace_hash_move_and_update_ops() and ftrace_hash_move_and_update_subops(). > > Diff between last version at end of this email. > > Masami Hiramatsu (Google) (3): > function_graph: Handle tail calls for stack unwinding > function_graph: Use a simple LRU for fgraph_array index number > ftrace: Add multiple fgraph storage selftest > > Steven Rostedt (Google) (9): > ftrace: Add subops logic to allow one ops to manage many > ftrace: Allow subops filtering to be modified > function_graph: Add pid tracing back to function graph tracer > function_graph: Use for_each_set_bit() in __ftrace_return_to_handler() > function_graph: Use bitmask to loop on fgraph entry > function_graph: Use static_call and branch to optimize entry function > function_graph: Use static_call and branch to optimize return function > selftests/ftrace: Add function_graph tracer to func-filter-pid test > selftests/ftrace: Add fgraph-multi.tc test > > Steven Rostedt (VMware) (15): > function_graph: Convert ret_stack to a series of longs > fgraph: Use BUILD_BUG_ON() to make sure we have structures divisible by long > function_graph: Add an array structure that will allow multiple callbacks > function_graph: Allow multiple users to attach to function graph > function_graph: Remove logic around ftrace_graph_entry and return > ftrace/function_graph: Pass fgraph_ops to function graph callbacks > ftrace: Allow function_graph tracer to be enabled in instances > ftrace: Allow ftrace startup flags to exist without dynamic ftrace > function_graph: Have the instances use their own ftrace_ops for filtering > function_graph: Add "task variables" per task for fgraph_ops > function_graph: Move set_graph_function tests to shadow stack global var > function_graph: Move graph depth stored data to shadow stack global var > function_graph: Move graph notrace bit to shadow stack global var > function_graph: Implement fgraph_reserve_data() and fgraph_retrieve_data() > function_graph: Add selftest for passing local variables > > ---- > include/linux/ftrace.h | 43 +- > include/linux/sched.h | 2 +- > include/linux/trace_recursion.h | 39 - > kernel/trace/fgraph.c | 1044 ++++++++++++++++---- > kernel/trace/ftrace.c | 522 +++++++++- > kernel/trace/ftrace_internal.h | 5 +- > kernel/trace/trace.h | 94 +- > kernel/trace/trace_functions.c | 8 + > kernel/trace/trace_functions_graph.c | 96 +- > kernel/trace/trace_irqsoff.c | 10 +- > kernel/trace/trace_sched_wakeup.c | 10 +- > kernel/trace/trace_selftest.c | 259 ++++- > .../selftests/ftrace/test.d/ftrace/fgraph-multi.tc | 103 ++ > .../ftrace/test.d/ftrace/func-filter-pid.tc | 27 +- > 14 files changed, 1945 insertions(+), 317 deletions(-) > create mode 100644 tools/testing/selftests/ftrace/test.d/ftrace/fgraph-multi.tc > > > diff --git a/kernel/trace/ftrace.c b/kernel/trace/ftrace.c > index 41fabc6d30e4..da7e6abf48b4 100644 > --- a/kernel/trace/ftrace.c > +++ b/kernel/trace/ftrace.c > @@ -3170,7 +3170,7 @@ int ftrace_shutdown(struct ftrace_ops *ops, int command) > /* Simply make a copy of @src and return it */ > static struct ftrace_hash *copy_hash(struct ftrace_hash *src) > { > - if (!src || src == EMPTY_HASH) > + if (ftrace_hash_empty(src)) > return EMPTY_HASH; > > return alloc_and_copy_ftrace_hash(src->size_bits, src); > @@ -3187,6 +3187,9 @@ static struct ftrace_hash *copy_hash(struct ftrace_hash *src) > * > * Otherwise, go through all of @new_hash and add anything that @hash > * doesn't already have, to @hash. > + * > + * The filter_hash updates uses just the append_hash() function > + * and the notrace_hash does not. > */ > static int append_hash(struct ftrace_hash **hash, struct ftrace_hash *new_hash) > { > @@ -3195,11 +3198,11 @@ static int append_hash(struct ftrace_hash **hash, struct ftrace_hash *new_hash) > int i; > > /* An empty hash does everything */ > - if (!*hash || *hash == EMPTY_HASH) > + if (ftrace_hash_empty(*hash)) > return 0; > > /* If new_hash has everything make hash have everything */ > - if (!new_hash || new_hash == EMPTY_HASH) { > + if (ftrace_hash_empty(new_hash)) { > free_ftrace_hash(*hash); > *hash = EMPTY_HASH; > return 0; > @@ -3217,7 +3220,12 @@ static int append_hash(struct ftrace_hash **hash, struct ftrace_hash *new_hash) > return 0; > } > > -/* Add to @hash only those that are in both @new_hash1 and @new_hash2 */ > +/* > + * Add to @hash only those that are in both @new_hash1 and @new_hash2 > + * > + * The notrace_hash updates uses just the intersect_hash() function > + * and the filter_hash does not. > + */ > static int intersect_hash(struct ftrace_hash **hash, struct ftrace_hash *new_hash1, > struct ftrace_hash *new_hash2) > { > @@ -3229,8 +3237,7 @@ static int intersect_hash(struct ftrace_hash **hash, struct ftrace_hash *new_has > * If new_hash1 or new_hash2 is the EMPTY_HASH then make the hash > * empty as well as empty for notrace means none are notraced. > */ > - if (!new_hash1 || new_hash1 == EMPTY_HASH || > - !new_hash2 || new_hash2 == EMPTY_HASH) { > + if (ftrace_hash_empty(new_hash1) || ftrace_hash_empty(new_hash2)) { > free_ftrace_hash(*hash); > *hash = EMPTY_HASH; > return 0; > @@ -3245,6 +3252,11 @@ static int intersect_hash(struct ftrace_hash **hash, struct ftrace_hash *new_has > return -ENOMEM; > } > } > + /* If nothing intersects, make it the empty set */ > + if (ftrace_hash_empty(*hash)) { > + free_ftrace_hash(*hash); > + *hash = EMPTY_HASH; > + } > return 0; > } > > @@ -3266,7 +3278,7 @@ static struct ftrace_hash *append_hashes(struct ftrace_ops *ops) > return NULL; > } > /* Nothing more to do if new_hash is empty */ > - if (new_hash == EMPTY_HASH) > + if (ftrace_hash_empty(new_hash)) > break; > } > return new_hash; > @@ -3300,59 +3312,76 @@ static struct ftrace_hash *intersect_hashes(struct ftrace_ops *ops) > return NULL; > } > /* Nothing more to do if new_hash is empty */ > - if (new_hash == EMPTY_HASH) > + if (ftrace_hash_empty(new_hash)) > break; > } > return new_hash; > } > > -/* Returns 0 on equal or non-zero on non-equal */ > -static int compare_ops(struct ftrace_hash *A, struct ftrace_hash *B) > +static bool ops_equal(struct ftrace_hash *A, struct ftrace_hash *B) > { > struct ftrace_func_entry *entry; > int size; > int i; > > - if (!A || A == EMPTY_HASH) > - return !(!B || B == EMPTY_HASH); > + if (ftrace_hash_empty(A)) > + return ftrace_hash_empty(B); > > - if (!B || B == EMPTY_HASH) > - return !(!A || A == EMPTY_HASH); > + if (ftrace_hash_empty(B)) > + return ftrace_hash_empty(A); > > if (A->count != B->count) > - return 1; > + return false; > > size = 1 << A->size_bits; > for (i = 0; i < size; i++) { > hlist_for_each_entry(entry, &A->buckets[i], hlist) { > if (!__ftrace_lookup_ip(B, entry->ip)) > - return 1; > + return false; > } > } > > - return 0; > + return true; > } > > -static int ftrace_hash_move_and_update_ops(struct ftrace_ops *ops, > - struct ftrace_hash **orig_hash, > - struct ftrace_hash *hash, > - int enable); > +static void ftrace_ops_update_code(struct ftrace_ops *ops, > + struct ftrace_ops_hash *old_hash); > + > +static int __ftrace_hash_move_and_update_ops(struct ftrace_ops *ops, > + struct ftrace_hash **orig_hash, > + struct ftrace_hash *hash, > + int enable) > +{ > + struct ftrace_ops_hash old_hash_ops; > + struct ftrace_hash *old_hash; > + int ret; > + > + old_hash = *orig_hash; > + old_hash_ops.filter_hash = ops->func_hash->filter_hash; > + old_hash_ops.notrace_hash = ops->func_hash->notrace_hash; > + ret = ftrace_hash_move(ops, enable, orig_hash, hash); > + if (!ret) { > + ftrace_ops_update_code(ops, &old_hash_ops); > + free_ftrace_hash_rcu(old_hash); > + } > + return ret; > +} > > static int ftrace_update_ops(struct ftrace_ops *ops, struct ftrace_hash *filter_hash, > struct ftrace_hash *notrace_hash) > { > int ret; > > - if (compare_ops(filter_hash, ops->func_hash->filter_hash)) { > - ret = ftrace_hash_move_and_update_ops(ops, &ops->func_hash->filter_hash, > - filter_hash, 1); > + if (!ops_equal(filter_hash, ops->func_hash->filter_hash)) { > + ret = __ftrace_hash_move_and_update_ops(ops, &ops->func_hash->filter_hash, > + filter_hash, 1); > if (ret < 0) > return ret; > } > > - if (compare_ops(notrace_hash, ops->func_hash->notrace_hash)) { > - ret = ftrace_hash_move_and_update_ops(ops, &ops->func_hash->notrace_hash, > - notrace_hash, 0); > + if (!ops_equal(notrace_hash, ops->func_hash->notrace_hash)) { > + ret = __ftrace_hash_move_and_update_ops(ops, &ops->func_hash->notrace_hash, > + notrace_hash, 0); > if (ret < 0) > return ret; > } > @@ -3438,8 +3467,8 @@ int ftrace_startup_subops(struct ftrace_ops *ops, struct ftrace_ops *subops, int > * o If either notrace_hash is empty then the final stays empty > * o Otherwise, the final is an intersection between the hashes > */ > - if (ops->func_hash->filter_hash == EMPTY_HASH || > - subops->func_hash->filter_hash == EMPTY_HASH) { > + if (ftrace_hash_empty(ops->func_hash->filter_hash) || > + ftrace_hash_empty(subops->func_hash->filter_hash)) { > filter_hash = EMPTY_HASH; > } else { > size_bits = max(ops->func_hash->filter_hash->size_bits, > @@ -3454,8 +3483,8 @@ int ftrace_startup_subops(struct ftrace_ops *ops, struct ftrace_ops *subops, int > } > } > > - if (ops->func_hash->notrace_hash == EMPTY_HASH || > - subops->func_hash->notrace_hash == EMPTY_HASH) { > + if (ftrace_hash_empty(ops->func_hash->notrace_hash) || > + ftrace_hash_empty(subops->func_hash->notrace_hash)) { > notrace_hash = EMPTY_HASH; > } else { > size_bits = max(ops->func_hash->filter_hash->size_bits, > @@ -3591,7 +3620,7 @@ static int ftrace_hash_move_and_update_subops(struct ftrace_ops *subops, > } > > /* Move the hash over to the new hash */ > - ret = ftrace_hash_move_and_update_ops(ops, orig_hash, new_hash, enable); > + ret = __ftrace_hash_move_and_update_ops(ops, orig_hash, new_hash, enable); > > free_ftrace_hash(new_hash); > > @@ -4822,11 +4851,6 @@ static int ftrace_hash_move_and_update_ops(struct ftrace_ops *ops, > struct ftrace_hash *hash, > int enable) > { > - struct ftrace_ops_hash old_hash_ops; > - struct ftrace_hash *old_hash; > - struct ftrace_ops *op; > - int ret; > - > if (ops->flags & FTRACE_OPS_FL_SUBOP) > return ftrace_hash_move_and_update_subops(ops, orig_hash, hash, enable); > > @@ -4838,6 +4862,8 @@ static int ftrace_hash_move_and_update_ops(struct ftrace_ops *ops, > * it will not affect subops that share it. > */ > if (!(ops->flags & FTRACE_OPS_FL_ENABLED)) { > + struct ftrace_ops *op; > + > /* Check if any other manager subops maps to this hash */ > do_for_each_ftrace_op(op, ftrace_ops_list) { > struct ftrace_ops *subops; > @@ -4851,15 +4877,7 @@ static int ftrace_hash_move_and_update_ops(struct ftrace_ops *ops, > } while_for_each_ftrace_op(op); > } > > - old_hash = *orig_hash; > - old_hash_ops.filter_hash = ops->func_hash->filter_hash; > - old_hash_ops.notrace_hash = ops->func_hash->notrace_hash; > - ret = ftrace_hash_move(ops, enable, orig_hash, hash); > - if (!ret) { > - ftrace_ops_update_code(ops, &old_hash_ops); > - free_ftrace_hash_rcu(old_hash); > - } > - return ret; > + return __ftrace_hash_move_and_update_ops(ops, orig_hash, hash, enable); > } > > static bool module_exists(const char *module)