All of lore.kernel.org
 help / color / mirror / Atom feed
* [PATCH v9 0/4] trace: Introduce objtrace trigger to trace the kernel object
@ 2022-02-04  3:56 Jeff Xie
  2022-02-04  3:56 ` [PATCH v9 1/4] trace: Add trace any " Jeff Xie
                   ` (4 more replies)
  0 siblings, 5 replies; 18+ messages in thread
From: Jeff Xie @ 2022-02-04  3:56 UTC (permalink / raw)
  To: rostedt; +Cc: mhiramat, mingo, zanussi, linux-kernel, Jeff Xie

Introduce a method based on function tracer to trace any object and get
the value of the object dynamically. the object can be obtained from the
dynamic event (kprobe_event/uprobe_event) or the static event(tracepoint).

Usage:
When using the kprobe event, only need to set the objtrace(a new trigger),
we can get the value of the object. The object is from the setting of the 
kprobe event.

For example:
For the function bio_add_page():

int bio_add_page(struct bio *bio, struct page *page,
	unsigned int len, unsigned int offset)

Firstly, we can set the base of the object, thus the first string "arg1"
stands for the value of the first parameter of this function bio_add_gage(),

# echo 'p bio_add_page arg1=$arg1' > ./kprobe_events

Secondly, we can get the value dynamically based on above object. 

find the offset of the bi_size in struct bio:
$ gdb vmlinux
(gdb) p &(((struct bio *)0)->bi_iter.bi_size)
$1 = (unsigned int *) 0x28

# echo 'objtrace:add:arg1,0x28:u32:1 if comm == "cat"' > ./events/kprobes/ \
	p_bio_add_page_0/trigger

# cd /sys/kernel/debug/tracing/
# echo 'p bio_add_page arg1=$arg1' > ./kprobe_events
# echo 'objtrace:add:arg1,0x28:u32:1 if comm == "cat"' > ./events/kprobes/p_bio_add_page_0/trigger

# du -sh /test.txt
12.0K   /test.txt

# cat  /test.txt > /dev/null
# cat ./trace
# tracer: nop
#
# entries-in-buffer/entries-written: 128/128   #P:4
#
#                                _-----=> irqs-off/BH-disabled
#                               / _----=> need-resched
#                              | / _---=> hardirq/softirq
#                              || / _--=> preempt-depth
#                              ||| / _-=> migrate-disable
#                              |||| /     delay
#           TASK-PID     CPU#  |||||  TIMESTAMP  FUNCTION
#              | |         |   |||||     |         |
             cat-117     [002] ...1.     1.602243: __bio_try_merge_page <-bio_add_page object:0xffff88811bee4000 value:0x0
             cat-117     [002] ...1.     1.602244: __bio_add_page <-bio_add_page object:0xffff88811bee4000 value:0x0
             cat-117     [002] ...2.     1.602244: bio_add_page <-ext4_mpage_readpages object:0xffff88811bee4000 value:0x1000
             cat-117     [002] ...1.     1.602245: __bio_try_merge_page <-bio_add_page object:0xffff88811bee4000 value:0x1000
             cat-117     [002] ...1.     1.602245: __bio_add_page <-bio_add_page object:0xffff88811bee4000 value:0x1000
             cat-117     [002] ...2.     1.602245: bio_add_page <-ext4_mpage_readpages object:0xffff88811bee4000 value:0x2000
             cat-117     [002] ...1.     1.602245: __bio_try_merge_page <-bio_add_page object:0xffff88811bee4000 value:0x2000
             cat-117     [002] ...1.     1.602245: __bio_add_page <-bio_add_page object:0xffff88811bee4000 value:0x2000
             cat-117     [002] ...1.     1.602245: submit_bio <-ext4_mpage_readpages object:0xffff88811bee4000 value:0x3000
             cat-117     [002] ...1.     1.602245: submit_bio_noacct <-ext4_mpage_readpages object:0xffff88811bee4000 value:0x3000
             cat-117     [002] ...1.     1.602246: __submit_bio <-submit_bio_noacct object:0xffff88811bee4000 value:0x3000
             cat-117     [002] ...1.     1.602246: submit_bio_checks <-__submit_bio object:0xffff88811bee4000 value:0x3000
             cat-117     [002] ...1.     1.602246: __cond_resched <-submit_bio_checks object:0xffff88811bee4000 value:0x3000
             cat-117     [002] ...1.     1.602246: should_fail_bio <-submit_bio_checks object:0xffff88811bee4000 value:0x3000
             cat-117     [002] ...1.     1.602246: blk_mq_submit_bio <-submit_bio_noacct object:0xffff88811bee4000 value:0x3000
             cat-117     [002] ...1.     1.602246: blk_attempt_plug_merge <-blk_mq_submit_bio object:0xffff88811bee4000 value:0x3000
             cat-117     [002] ...1.     1.602246: blk_mq_sched_bio_merge <-blk_mq_submit_bio object:0xffff88811bee4000 value:0x3000
             cat-117     [002] ...1.     1.602247: __rcu_read_lock <-blk_mq_submit_bio object:0xffff88811bee4000 value:0x3000
             cat-117     [002] ...1.     1.602247: __rcu_read_unlock <-blk_mq_submit_bio object:0xffff88811bee4000 value:0x3000
             cat-117     [002] ...1.     1.602247: __blk_mq_alloc_requests <-blk_mq_submit_bio object:0xffff88811bee4000 value:0x3000
          <idle>-0       [002] d..3.     1.602298: bio_endio <-blk_update_request object:0xffff88811bee4000 value:0x0
          <idle>-0       [002] d..3.     1.602298: mpage_end_io <-blk_update_request object:0xffff88811bee4000 value:0x0
          <idle>-0       [002] d..3.     1.602298: __read_end_io <-blk_update_request object:0xffff88811bee4000 value:0x0
          <idle>-0       [002] d..3.     1.602300: bio_put <-blk_update_request object:0xffff88811bee4000 value:0x0
          <idle>-0       [002] d..3.     1.602300: bio_free <-blk_update_request object:0xffff88811bee4000 value:0x0
          <idle>-0       [002] d..3.     1.602300: mempool_free <-blk_update_request object:0xffff88811bee4000 value:0x0
          <idle>-0       [002] d..3.     1.602300: mempool_free_slab <-blk_update_request object:0xffff88811bee4000 value:0x0
          <idle>-0       [002] d..3.     1.602300: kmem_cache_free <-blk_update_request object:0xffff88811bee4000 value:0x0
	  ...

Almost all changelogs were suggested by Masami(mhiramat@kernel.org)
and steve(rostedt@goodmis.org), thank you all so much.

v9:
- fix objtrace trigger output was incomplete
- fix the objtrace trigger was removed when using the existed parameter on
  event.
- add testcase for the second fix above.

v8:
- revert to use per-cpu recursion for the function trace_object_events_call
- recover the filter when getting the value of the object
- simplify the implementation for the function get_object_value
- fix the build error

v7:
- use fixed-size array for object pool instead of list structure
- use ftrace_test_recursion_trylock for function trace hook function
- fix trace_object_ref reference count in the init_trace_object
- invoke exit_trace_object no matter whether data->ops->free is null 
  in the unregister_object_trigger
- release private_data of event_trigger_data in the trace_object_trigger_free
- remove [RFC] tag

v6:
- change the objtrace trigger syntax.
- add patchset description
- add <tracefs>/README

v5:
- add testcasts
- add check the field->size
- add lockless to search object
- describe the object trace more clearly in Kconfig

v4:
- please ignore the v4 which is the same as v3

v3:
- change the objfilter to objtrace
- add a command to the objfilter syntax
- change to get the value of the object
- use trace_find_event_field to find the field instead of using argN
- get data from @rec in the event trigger callback funciton

v2:
- adding a "objfilter" trigger to update object

Jeff Xie (4):
  trace: Add trace any kernel object
  trace/objtrace: get the value of the object
  trace/objtrace: Add testcases for objtrace
  trace/objtrace: Add documentation for objtrace

 Documentation/trace/events.rst                |  83 +++
 include/linux/trace_events.h                  |   1 +
 kernel/trace/Kconfig                          |  10 +
 kernel/trace/Makefile                         |   1 +
 kernel/trace/trace.c                          |   3 +
 kernel/trace/trace.h                          |   8 +
 kernel/trace/trace_entries.h                  |  18 +
 kernel/trace/trace_events_trigger.c           |   1 +
 kernel/trace/trace_object.c                   | 676 ++++++++++++++++++
 kernel/trace/trace_output.c                   |  40 ++
 .../ftrace/test.d/trigger/trigger-objtrace.tc |  41 ++
 11 files changed, 882 insertions(+)
 create mode 100644 kernel/trace/trace_object.c
 create mode 100644 tools/testing/selftests/ftrace/test.d/trigger/trigger-objtrace.tc

-- 
2.25.1


^ permalink raw reply	[flat|nested] 18+ messages in thread

* [PATCH v9 1/4] trace: Add trace any kernel object
  2022-02-04  3:56 [PATCH v9 0/4] trace: Introduce objtrace trigger to trace the kernel object Jeff Xie
@ 2022-02-04  3:56 ` Jeff Xie
  2022-04-19  2:14   ` Steven Rostedt
  2022-04-27 15:50   ` Jeff Xie
  2022-02-04  3:56 ` [PATCH v9 2/4] trace/objtrace: get the value of the object Jeff Xie
                   ` (3 subsequent siblings)
  4 siblings, 2 replies; 18+ messages in thread
From: Jeff Xie @ 2022-02-04  3:56 UTC (permalink / raw)
  To: rostedt; +Cc: mhiramat, mingo, zanussi, linux-kernel, Jeff Xie

Introduce objtrace trigger to get the call flow by tracing any kernel
object in the function parameter.

The objtrace trigger makes a list of the target object address from
the given event parameter, and records all kernel function calls
which has the object address in its parameter.

Syntax:
	objtrace:add:obj[:count][if <filter>]

Usage:
	# echo 'p bio_add_page arg1=$arg1' > kprobe_events
	# cd events/kprobes/p_bio_add_page_0
	# echo 'objtrace:add:arg1:1 if comm == "cat"' > ./trigger
	# cat /test.txt

Signed-off-by: Jeff Xie <xiehuan09@gmail.com>
---
 include/linux/trace_events.h        |   1 +
 kernel/trace/Kconfig                |  10 +
 kernel/trace/Makefile               |   1 +
 kernel/trace/trace.c                |   3 +
 kernel/trace/trace.h                |   8 +
 kernel/trace/trace_entries.h        |  17 +
 kernel/trace/trace_events_trigger.c |   1 +
 kernel/trace/trace_object.c         | 515 ++++++++++++++++++++++++++++
 kernel/trace/trace_output.c         |  40 +++
 9 files changed, 596 insertions(+)
 create mode 100644 kernel/trace/trace_object.c

diff --git a/include/linux/trace_events.h b/include/linux/trace_events.h
index 70c069aef02c..3ccdbc1d25dd 100644
--- a/include/linux/trace_events.h
+++ b/include/linux/trace_events.h
@@ -685,6 +685,7 @@ enum event_trigger_type {
 	ETT_EVENT_HIST		= (1 << 4),
 	ETT_HIST_ENABLE		= (1 << 5),
 	ETT_EVENT_EPROBE	= (1 << 6),
+	ETT_TRACE_OBJECT	= (1 << 7),
 };
 
 extern int filter_match_preds(struct event_filter *filter, void *rec);
diff --git a/kernel/trace/Kconfig b/kernel/trace/Kconfig
index a5eb5e7fd624..c51b7eb1508d 100644
--- a/kernel/trace/Kconfig
+++ b/kernel/trace/Kconfig
@@ -250,6 +250,16 @@ config FUNCTION_PROFILER
 
 	  If in doubt, say N.
 
+config TRACE_OBJECT
+	bool "Trace kernel object in function parameter"
+	depends on FUNCTION_TRACER
+	depends on HAVE_FUNCTION_ARG_ACCESS_API
+	select TRACING
+	default y
+	help
+	 You can trace the kernel object in the kernel function parameter.
+	 The kernel object is dynamically specified via event trigger.
+
 config STACK_TRACER
 	bool "Trace max stack"
 	depends on HAVE_FUNCTION_TRACER
diff --git a/kernel/trace/Makefile b/kernel/trace/Makefile
index bedc5caceec7..b924b8e55922 100644
--- a/kernel/trace/Makefile
+++ b/kernel/trace/Makefile
@@ -67,6 +67,7 @@ obj-$(CONFIG_FUNCTION_GRAPH_TRACER) += trace_functions_graph.o
 obj-$(CONFIG_TRACE_BRANCH_PROFILING) += trace_branch.o
 obj-$(CONFIG_BLK_DEV_IO_TRACE) += blktrace.o
 obj-$(CONFIG_FUNCTION_GRAPH_TRACER) += fgraph.o
+obj-$(CONFIG_TRACE_OBJECT) += trace_object.o
 ifeq ($(CONFIG_BLOCK),y)
 obj-$(CONFIG_EVENT_TRACING) += blktrace.o
 endif
diff --git a/kernel/trace/trace.c b/kernel/trace/trace.c
index a734d5ae34c8..b4513c2bbd52 100644
--- a/kernel/trace/trace.c
+++ b/kernel/trace/trace.c
@@ -5605,6 +5605,9 @@ static const char readme_msg[] =
 	"\t            enable_hist:<system>:<event>\n"
 	"\t            disable_hist:<system>:<event>\n"
 #endif
+#ifdef CONFIG_TRACE_OBJECT
+	"\t            objtrace:add:obj[:count][if <filter>]\n"
+#endif
 #ifdef CONFIG_STACKTRACE
 	"\t\t    stacktrace\n"
 #endif
diff --git a/kernel/trace/trace.h b/kernel/trace/trace.h
index 0f5e22238cd2..8b66515a36d5 100644
--- a/kernel/trace/trace.h
+++ b/kernel/trace/trace.h
@@ -54,6 +54,7 @@ enum trace_type {
 	TRACE_TIMERLAT,
 	TRACE_RAW_DATA,
 	TRACE_FUNC_REPEATS,
+	TRACE_OBJECT,
 
 	__TRACE_LAST_TYPE,
 };
@@ -472,6 +473,7 @@ extern void __ftrace_bad_type(void);
 			  TRACE_GRAPH_RET);		\
 		IF_ASSIGN(var, ent, struct func_repeats_entry,		\
 			  TRACE_FUNC_REPEATS);				\
+		IF_ASSIGN(var, ent, struct trace_object_entry, TRACE_OBJECT);\
 		__ftrace_bad_type();					\
 	} while (0)
 
@@ -1537,6 +1539,12 @@ static inline int register_trigger_hist_cmd(void) { return 0; }
 static inline int register_trigger_hist_enable_disable_cmds(void) { return 0; }
 #endif
 
+#ifdef CONFIG_TRACE_OBJECT
+extern int register_trigger_object_cmd(void);
+#else
+static inline int register_trigger_object_cmd(void) { return 0; }
+#endif
+
 extern int register_trigger_cmds(void);
 extern void clear_event_triggers(struct trace_array *tr);
 
diff --git a/kernel/trace/trace_entries.h b/kernel/trace/trace_entries.h
index cd41e863b51c..bb120d9498a9 100644
--- a/kernel/trace/trace_entries.h
+++ b/kernel/trace/trace_entries.h
@@ -401,3 +401,20 @@ FTRACE_ENTRY(timerlat, timerlat_entry,
 		 __entry->context,
 		 __entry->timer_latency)
 );
+
+/*
+ * trace object entry:
+ */
+FTRACE_ENTRY(object, trace_object_entry,
+
+	TRACE_OBJECT,
+
+	F_STRUCT(
+		__field(	unsigned long,		ip		)
+		__field(	unsigned long,		parent_ip	)
+		__field(	unsigned long,		object		)
+	),
+
+	F_printk(" %ps <-- %ps object:%lx\n",
+		 (void *)__entry->ip, (void *)__entry->parent_ip, __entry->object)
+);
diff --git a/kernel/trace/trace_events_trigger.c b/kernel/trace/trace_events_trigger.c
index d00fee705f9c..c3371a6902af 100644
--- a/kernel/trace/trace_events_trigger.c
+++ b/kernel/trace/trace_events_trigger.c
@@ -2025,6 +2025,7 @@ __init int register_trigger_cmds(void)
 	register_trigger_enable_disable_cmds();
 	register_trigger_hist_enable_disable_cmds();
 	register_trigger_hist_cmd();
+	register_trigger_object_cmd();
 
 	return 0;
 }
diff --git a/kernel/trace/trace_object.c b/kernel/trace/trace_object.c
new file mode 100644
index 000000000000..540e387c613a
--- /dev/null
+++ b/kernel/trace/trace_object.c
@@ -0,0 +1,515 @@
+// SPDX-License-Identifier: GPL-2.0
+/*
+ * trace the kernel object in the kernel function parameter
+ * Copyright (C) 2021 Jeff Xie <xiehuan09@gmail.com>
+ */
+
+#define pr_fmt(fmt) "trace_object: " fmt
+
+#include "trace_output.h"
+
+#define MAX_TRACED_OBJECT 5
+#define OBJTRACE_CMD_LEN  10
+static DEFINE_PER_CPU(unsigned int, trace_object_event_disable);
+static DEFINE_RAW_SPINLOCK(trace_obj_lock);
+static struct trace_event_file event_trace_file;
+static const int max_args_num = 6;
+static atomic_t trace_object_ref;
+static atomic_t num_traced_obj;
+static int exit_trace_object(void);
+static int init_trace_object(void);
+
+static struct object_instance {
+	void *obj;
+} traced_obj[MAX_TRACED_OBJECT];
+
+/* objtrace private data */
+struct objtrace_trigger_data {
+	struct ftrace_event_field *field;
+	char objtrace_cmd[OBJTRACE_CMD_LEN];
+};
+
+static bool object_exist(void *obj)
+{
+	int i, max;
+
+	max = atomic_read(&num_traced_obj);
+	smp_rmb();
+	for (i = 0; i < max; i++) {
+		if (traced_obj[i].obj == obj)
+			return true;
+	}
+	return false;
+}
+
+static bool object_empty(void)
+{
+	return !atomic_read(&num_traced_obj);
+}
+
+static void set_trace_object(void *obj)
+{
+	unsigned long flags;
+
+	if (in_nmi())
+		return;
+
+	if (!obj)
+		return;
+
+	if (object_exist(obj))
+		return;
+
+	/* only this place has write operations */
+	raw_spin_lock_irqsave(&trace_obj_lock, flags);
+	if (atomic_read(&num_traced_obj) == MAX_TRACED_OBJECT) {
+		trace_printk("object_pool is full, can't trace object:0x%px\n", obj);
+		goto out;
+	}
+	traced_obj[atomic_read(&num_traced_obj)].obj = obj;
+	/* make sure the num_traced_obj update always appears after traced_obj update */
+	smp_wmb();
+	atomic_inc(&num_traced_obj);
+out:
+	raw_spin_unlock_irqrestore(&trace_obj_lock, flags);
+}
+
+static void submit_trace_object(unsigned long ip, unsigned long parent_ip,
+				 unsigned long object)
+{
+
+	struct trace_buffer *buffer;
+	struct ring_buffer_event *event;
+	struct trace_object_entry *entry;
+	int pc;
+
+	pc = preempt_count();
+	event = trace_event_buffer_lock_reserve(&buffer, &event_trace_file,
+			TRACE_OBJECT, sizeof(*entry), pc);
+	if (!event)
+		return;
+	entry   = ring_buffer_event_data(event);
+	entry->ip                       = ip;
+	entry->parent_ip                = parent_ip;
+	entry->object			= object;
+
+	event_trigger_unlock_commit(&event_trace_file, buffer, event,
+		entry, pc);
+}
+
+static void
+trace_object_events_call(unsigned long ip, unsigned long parent_ip,
+		struct ftrace_ops *op, struct ftrace_regs *fregs)
+{
+	struct pt_regs *pt_regs = ftrace_get_regs(fregs);
+	unsigned long obj;
+	unsigned int disabled;
+	int n;
+
+	preempt_disable_notrace();
+
+	disabled = this_cpu_inc_return(trace_object_event_disable);
+	if (disabled != 1)
+		goto out;
+
+	if (object_empty())
+		goto out;
+
+	for (n = 0; n < max_args_num; n++) {
+		obj = regs_get_kernel_argument(pt_regs, n);
+		if (object_exist((void *)obj))
+			submit_trace_object(ip, parent_ip, obj);
+	/* The parameters of a function may match multiple objects */
+	}
+out:
+	this_cpu_dec(trace_object_event_disable);
+	preempt_enable_notrace();
+}
+
+static struct ftrace_ops trace_ops = {
+	.func  = trace_object_events_call,
+	.flags = FTRACE_OPS_FL_SAVE_REGS,
+};
+
+static void
+trace_object_trigger(struct event_trigger_data *data,
+		   struct trace_buffer *buffer,  void *rec,
+		   struct ring_buffer_event *event)
+{
+	struct objtrace_trigger_data *obj_data = data->private_data;
+	struct ftrace_event_field *field;
+	void *obj = NULL;
+
+	field = obj_data->field;
+	memcpy(&obj, rec + field->offset, sizeof(obj));
+	set_trace_object(obj);
+}
+
+static void
+trace_object_trigger_free(struct event_trigger_ops *ops,
+		   struct event_trigger_data *data)
+{
+	if (WARN_ON_ONCE(data->ref <= 0))
+		return;
+
+	data->ref--;
+	if (!data->ref) {
+		kfree(data->private_data);
+		trigger_data_free(data);
+	}
+}
+
+static void
+trace_object_count_trigger(struct event_trigger_data *data,
+			 struct trace_buffer *buffer, void *rec,
+			 struct ring_buffer_event *event)
+{
+	if (!data->count)
+		return;
+
+	if (data->count != -1)
+		(data->count)--;
+
+	trace_object_trigger(data, buffer, rec, event);
+}
+
+static int event_object_trigger_init(struct event_trigger_ops *ops,
+		       struct event_trigger_data *data)
+{
+	data->ref++;
+	return 0;
+}
+
+static int
+event_trigger_print(const char *name, struct seq_file *m,
+		void *data, char *filter_str, void *objtrace_data)
+{
+	long count = (long)data;
+	struct objtrace_trigger_data *obj_data = objtrace_data;
+
+	seq_puts(m, name);
+
+	seq_printf(m, ":%s", obj_data->objtrace_cmd);
+	seq_printf(m, ":%s", obj_data->field->name);
+
+	if (count == -1)
+		seq_puts(m, ":unlimited");
+	else
+		seq_printf(m, ":count=%ld", count);
+
+	if (filter_str)
+		seq_printf(m, " if %s\n", filter_str);
+	else
+		seq_putc(m, '\n');
+
+	return 0;
+}
+
+static int
+trace_object_trigger_print(struct seq_file *m, struct event_trigger_ops *ops,
+			 struct event_trigger_data *data)
+{
+	return event_trigger_print("objtrace", m, (void *)data->count,
+				   data->filter_str, data->private_data);
+}
+
+static struct event_trigger_ops objecttrace_trigger_ops = {
+	.trigger		= trace_object_trigger,
+	.print			= trace_object_trigger_print,
+	.init			= event_object_trigger_init,
+	.free			= trace_object_trigger_free,
+};
+
+static struct event_trigger_ops objecttrace_count_trigger_ops = {
+	.trigger		= trace_object_count_trigger,
+	.print			= trace_object_trigger_print,
+	.init			= event_object_trigger_init,
+	.free			= trace_object_trigger_free,
+};
+
+static struct event_trigger_ops *
+objecttrace_get_trigger_ops(char *cmd, char *param)
+{
+	return param ? &objecttrace_count_trigger_ops : &objecttrace_trigger_ops;
+}
+
+static int register_object_trigger(char *glob,
+			    struct event_trigger_data *data,
+			    struct trace_event_file *file)
+{
+	struct event_trigger_data *test;
+	int ret = 0;
+
+	lockdep_assert_held(&event_mutex);
+
+	list_for_each_entry(test, &file->triggers, list) {
+		if (test->cmd_ops->trigger_type == data->cmd_ops->trigger_type) {
+			ret = -EEXIST;
+			goto out;
+		}
+	}
+
+	if (data->ops->init) {
+		ret = data->ops->init(data->ops, data);
+		if (ret < 0)
+			goto out;
+	}
+
+	list_add_rcu(&data->list, &file->triggers);
+	ret++;
+
+	update_cond_flag(file);
+	if (trace_event_trigger_enable_disable(file, 1) < 0) {
+		list_del_rcu(&data->list);
+		update_cond_flag(file);
+		ret--;
+	}
+	init_trace_object();
+out:
+	return ret;
+}
+
+static void unregister_object_trigger(char *glob,
+			       struct event_trigger_data *test,
+			       struct trace_event_file *file)
+{
+	struct event_trigger_data *data;
+	bool unregistered = false;
+
+	lockdep_assert_held(&event_mutex);
+
+	list_for_each_entry(data, &file->triggers, list) {
+		if (data->cmd_ops->trigger_type == test->cmd_ops->trigger_type) {
+			unregistered = true;
+			list_del_rcu(&data->list);
+			trace_event_trigger_enable_disable(file, 0);
+			update_cond_flag(file);
+			break;
+		}
+	}
+
+	if (unregistered) {
+		if (data->ops->free)
+			data->ops->free(data->ops, data);
+		exit_trace_object();
+	}
+}
+
+static bool field_exist(struct trace_event_file *file,
+			struct event_command *cmd_ops,
+			const char *field_name)
+{
+	struct event_trigger_data *data;
+	struct objtrace_trigger_data *obj_data;
+
+	lockdep_assert_held(&event_mutex);
+
+	list_for_each_entry(data, &file->triggers, list) {
+		if (data->cmd_ops->trigger_type == cmd_ops->trigger_type) {
+			obj_data = data->private_data;
+			if (!strcmp(obj_data->field->name, field_name))
+				return true;
+		}
+	}
+
+	return false;
+}
+
+static int
+event_object_trigger_parse(struct event_command *cmd_ops,
+		       struct trace_event_file *file,
+		       char *glob, char *cmd, char *param)
+{
+	struct event_trigger_data *trigger_data;
+	struct event_trigger_ops *trigger_ops;
+	struct objtrace_trigger_data *obj_data;
+	struct trace_event_call *call;
+	struct ftrace_event_field *field;
+	char *objtrace_cmd;
+	char *trigger = NULL;
+	char *arg;
+	char *number;
+	int ret;
+	bool remove = false;
+
+	ret = -EINVAL;
+	if (!param)
+		goto out;
+
+	/* separate the trigger from the filter (c:a:n [if filter]) */
+	trigger = strsep(&param, " \t");
+	if (!trigger)
+		goto out;
+	if (param) {
+		param = skip_spaces(param);
+		if (!*param)
+			param = NULL;
+	}
+
+	objtrace_cmd = strsep(&trigger, ":");
+	if (!objtrace_cmd || strcmp(objtrace_cmd, "add"))
+		goto out;
+
+	arg = strsep(&trigger, ":");
+	if (!arg)
+		goto out;
+	call = file->event_call;
+	field = trace_find_event_field(call, arg);
+	if (!field)
+		goto out;
+
+	if (field->size != sizeof(void *))
+		goto out;
+
+	if (glob[0] == '!')
+		remove = true;
+
+	if (remove && !field_exist(file, cmd_ops, field->name))
+	goto out;
+	trigger_ops = cmd_ops->get_trigger_ops(cmd, trigger);
+	ret = -ENOMEM;
+	obj_data = kzalloc(sizeof(*obj_data), GFP_KERNEL);
+	if (!obj_data)
+		goto out;
+	obj_data->field = field;
+	snprintf(obj_data->objtrace_cmd, OBJTRACE_CMD_LEN, objtrace_cmd);
+
+	trigger_data = kzalloc(sizeof(*trigger_data), GFP_KERNEL);
+	if (!trigger_data) {
+		kfree(obj_data);
+		goto out;
+	}
+
+	trigger_data->count = -1;
+	trigger_data->ops = trigger_ops;
+	trigger_data->cmd_ops = cmd_ops;
+	trigger_data->private_data = obj_data;
+	INIT_LIST_HEAD(&trigger_data->list);
+	INIT_LIST_HEAD(&trigger_data->named_list);
+
+	if (remove) {
+		cmd_ops->unreg(glob+1, trigger_data, file);
+		kfree(obj_data);
+		kfree(trigger_data);
+		ret = 0;
+		goto out;
+	}
+
+	if (trigger) {
+		number = strsep(&trigger, ":");
+
+		ret = -EINVAL;
+		if (!strlen(number))
+			goto out_free;
+
+		/*
+		 * We use the callback data field (which is a pointer)
+		 * as our counter.
+		 */
+		ret = kstrtoul(number, 0, &trigger_data->count);
+		if (ret)
+			goto out_free;
+	}
+
+	if (!param) /* if param is non-empty, it's supposed to be a filter */
+		goto out_reg;
+
+	if (!cmd_ops->set_filter)
+		goto out_reg;
+
+	ret = cmd_ops->set_filter(param, trigger_data, file);
+	if (ret < 0)
+		goto out_free;
+
+ out_reg:
+	/* Up the trigger_data count to make sure reg doesn't free it on failure */
+	event_object_trigger_init(trigger_ops, trigger_data);
+	ret = cmd_ops->reg(glob, trigger_data, file);
+	/*
+	 * The above returns on success the # of functions enabled,
+	 * but if it didn't find any functions it returns zero.
+	 * Consider no functions a failure too.
+	 */
+	if (!ret) {
+		cmd_ops->unreg(glob, trigger_data, file);
+		ret = -ENOENT;
+	} else if (ret > 0)
+		ret = 0;
+
+	/* Down the counter of trigger_data or free it if not used anymore */
+	trace_object_trigger_free(trigger_ops, trigger_data);
+ out:
+	return ret;
+
+ out_free:
+	if (cmd_ops->set_filter)
+		cmd_ops->set_filter(NULL, trigger_data, NULL);
+	kfree(obj_data);
+	kfree(trigger_data);
+	goto out;
+}
+
+static struct event_command trigger_object_cmd = {
+	.name			= "objtrace",
+	.trigger_type		= ETT_TRACE_OBJECT,
+	.flags			= EVENT_CMD_FL_NEEDS_REC,
+	.parse			= event_object_trigger_parse,
+	.reg			= register_object_trigger,
+	.unreg			= unregister_object_trigger,
+	.get_trigger_ops	= objecttrace_get_trigger_ops,
+	.set_filter		= set_trigger_filter,
+};
+
+__init int register_trigger_object_cmd(void)
+{
+	int ret;
+
+	ret = register_event_command(&trigger_object_cmd);
+	WARN_ON(ret < 0);
+
+	return ret;
+}
+
+static int init_trace_object(void)
+{
+	int ret;
+
+	if (atomic_inc_return(&trace_object_ref) != 1) {
+		ret = 0;
+		goto out;
+	}
+
+	event_trace_file.tr = top_trace_array();
+	if (WARN_ON(!event_trace_file.tr)) {
+		ret = -1;
+		atomic_dec(&trace_object_ref);
+		goto out;
+	}
+	ret = register_ftrace_function(&trace_ops);
+out:
+	return ret;
+}
+
+static int exit_trace_object(void)
+{
+	int ret;
+
+	if (WARN_ON_ONCE(atomic_read(&trace_object_ref) <= 0)) {
+		ret = -1;
+		goto out;
+	}
+
+	if (atomic_dec_return(&trace_object_ref) != 0) {
+		ret = 0;
+		goto out;
+	}
+
+	ret = unregister_ftrace_function(&trace_ops);
+	if (ret) {
+		pr_err("can't unregister ftrace for trace object\n");
+		goto out;
+	}
+	atomic_set(&num_traced_obj, 0);
+out:
+	return ret;
+}
diff --git a/kernel/trace/trace_output.c b/kernel/trace/trace_output.c
index 8aa493d25c73..265428154638 100644
--- a/kernel/trace/trace_output.c
+++ b/kernel/trace/trace_output.c
@@ -1547,6 +1547,45 @@ static struct trace_event trace_func_repeats_event = {
 	.funcs		= &trace_func_repeats_funcs,
 };
 
+/* TRACE_OBJECT */
+static enum print_line_t trace_object_print(struct trace_iterator *iter, int flags,
+					struct trace_event *event)
+{
+	struct trace_object_entry *field;
+	struct trace_seq *s = &iter->seq;
+
+	trace_assign_type(field, iter->ent);
+	print_fn_trace(s, field->ip, field->parent_ip, flags);
+	trace_seq_printf(s, " object:0x%lx", field->object);
+	trace_seq_putc(s, '\n');
+
+	return trace_handle_return(s);
+}
+
+static enum print_line_t trace_object_raw(struct trace_iterator *iter, int flags,
+				      struct trace_event *event)
+{
+	struct trace_object_entry *field;
+
+	trace_assign_type(field, iter->ent);
+
+	trace_seq_printf(&iter->seq, "%lx %lx\n",
+			 field->ip,
+			 field->parent_ip);
+
+	return trace_handle_return(&iter->seq);
+}
+
+static struct trace_event_functions trace_object_funcs = {
+	.trace		= trace_object_print,
+	.raw		= trace_object_raw,
+};
+
+static struct trace_event trace_object_event = {
+	.type		= TRACE_OBJECT,
+	.funcs		= &trace_object_funcs,
+};
+
 static struct trace_event *events[] __initdata = {
 	&trace_fn_event,
 	&trace_ctx_event,
@@ -1561,6 +1600,7 @@ static struct trace_event *events[] __initdata = {
 	&trace_timerlat_event,
 	&trace_raw_data_event,
 	&trace_func_repeats_event,
+	&trace_object_event,
 	NULL
 };
 
-- 
2.25.1


^ permalink raw reply related	[flat|nested] 18+ messages in thread

* [PATCH v9 2/4] trace/objtrace: get the value of the object
  2022-02-04  3:56 [PATCH v9 0/4] trace: Introduce objtrace trigger to trace the kernel object Jeff Xie
  2022-02-04  3:56 ` [PATCH v9 1/4] trace: Add trace any " Jeff Xie
@ 2022-02-04  3:56 ` Jeff Xie
  2022-02-04  3:56 ` [PATCH v9 3/4] trace/objtrace: Add testcases for objtrace Jeff Xie
                   ` (2 subsequent siblings)
  4 siblings, 0 replies; 18+ messages in thread
From: Jeff Xie @ 2022-02-04  3:56 UTC (permalink / raw)
  To: rostedt; +Cc: mhiramat, mingo, zanussi, linux-kernel, Jeff Xie

Using objtrace trigger to get the value of the object which from the kernel
function parameter.

Syntax:
	objtrace:add:obj[,offset][:type][:count][if <filter>]

Usage:
	# echo 'p bio_add_page arg1=$arg1' > ./kprobe_events
	# gdb vmlinux
	(gdb) p &(((struct bio *)0)->bi_iter.bi_size)
	$1 = (unsigned int *) 0x28
	# echo 'objtrace:add:arg1,0x28:u32:1 if comm == "cat"' > ./events/kprobes/ \
		 p_bio_add_page_0/trigger
	# cat /test.txt

Signed-off-by: Jeff Xie <xiehuan09@gmail.com>
---
 kernel/trace/trace.c         |   2 +-
 kernel/trace/trace_entries.h |   5 +-
 kernel/trace/trace_object.c  | 231 +++++++++++++++++++++++++++++------
 kernel/trace/trace_output.c  |   6 +-
 4 files changed, 203 insertions(+), 41 deletions(-)

diff --git a/kernel/trace/trace.c b/kernel/trace/trace.c
index b4513c2bbd52..5f145837d2ff 100644
--- a/kernel/trace/trace.c
+++ b/kernel/trace/trace.c
@@ -5606,7 +5606,7 @@ static const char readme_msg[] =
 	"\t            disable_hist:<system>:<event>\n"
 #endif
 #ifdef CONFIG_TRACE_OBJECT
-	"\t            objtrace:add:obj[:count][if <filter>]\n"
+	"\t            objtrace:add:obj[,offset][:type][:count][if <filter>]\n"
 #endif
 #ifdef CONFIG_STACKTRACE
 	"\t\t    stacktrace\n"
diff --git a/kernel/trace/trace_entries.h b/kernel/trace/trace_entries.h
index bb120d9498a9..2407c45a568c 100644
--- a/kernel/trace/trace_entries.h
+++ b/kernel/trace/trace_entries.h
@@ -413,8 +413,9 @@ FTRACE_ENTRY(object, trace_object_entry,
 		__field(	unsigned long,		ip		)
 		__field(	unsigned long,		parent_ip	)
 		__field(	unsigned long,		object		)
+		__field(	unsigned long,		value		)
 	),
 
-	F_printk(" %ps <-- %ps object:%lx\n",
-		 (void *)__entry->ip, (void *)__entry->parent_ip, __entry->object)
+	F_printk(" %ps <-- %ps object:%lx value:%lx\n", (void *)__entry->ip,
+	       (void *)__entry->parent_ip, __entry->object, __entry->value)
 );
diff --git a/kernel/trace/trace_object.c b/kernel/trace/trace_object.c
index 540e387c613a..8eb2d61c277a 100644
--- a/kernel/trace/trace_object.c
+++ b/kernel/trace/trace_object.c
@@ -19,14 +19,34 @@ static atomic_t num_traced_obj;
 static int exit_trace_object(void);
 static int init_trace_object(void);
 
+/*
+ * get the offset from the special object and
+ * the type size of the value
+ */
 static struct object_instance {
 	void *obj;
+	int obj_offset;
+	int obj_value_type_size;
 } traced_obj[MAX_TRACED_OBJECT];
 
 /* objtrace private data */
 struct objtrace_trigger_data {
 	struct ftrace_event_field *field;
 	char objtrace_cmd[OBJTRACE_CMD_LEN];
+	int obj_offset;
+	int obj_value_type_size;
+};
+
+/* get the type size for the special object */
+struct objtrace_fetch_type {
+	char *name;
+	int type_size;
+};
+
+enum objattr {
+	OBJ_OFFSET,
+	OBJ_VAL_TYPE_SIZE,
+	MAX_OBJ_ATTR
 };
 
 static bool object_exist(void *obj)
@@ -42,13 +62,37 @@ static bool object_exist(void *obj)
 	return false;
 }
 
+static int get_object_attr(void *obj, int objattr, int *result)
+{
+	int i, max;
+
+	max = atomic_read(&num_traced_obj);
+	smp_rmb();
+	for (i = 0; i < max; i++) {
+		if (traced_obj[i].obj == obj) {
+			switch (objattr) {
+			case OBJ_OFFSET:
+				*result =  traced_obj[i].obj_offset;
+				return 0;
+			case OBJ_VAL_TYPE_SIZE:
+				*result = traced_obj[i].obj_value_type_size;
+				return 0;
+			default:
+				return -EINVAL;
+			}
+		}
+	}
+	return -EINVAL;
+}
+
 static bool object_empty(void)
 {
 	return !atomic_read(&num_traced_obj);
 }
 
-static void set_trace_object(void *obj)
+static void set_trace_object(void *obj, int obj_offset, int obj_value_type_size)
 {
+	unsigned int traced_objs;
 	unsigned long flags;
 
 	if (in_nmi())
@@ -62,11 +106,14 @@ static void set_trace_object(void *obj)
 
 	/* only this place has write operations */
 	raw_spin_lock_irqsave(&trace_obj_lock, flags);
-	if (atomic_read(&num_traced_obj) == MAX_TRACED_OBJECT) {
+	traced_objs = atomic_read(&num_traced_obj);
+	if (traced_objs == MAX_TRACED_OBJECT) {
 		trace_printk("object_pool is full, can't trace object:0x%px\n", obj);
 		goto out;
 	}
-	traced_obj[atomic_read(&num_traced_obj)].obj = obj;
+	traced_obj[traced_objs].obj = obj;
+	traced_obj[traced_objs].obj_value_type_size = obj_value_type_size;
+	traced_obj[traced_objs].obj_offset = obj_offset;
 	/* make sure the num_traced_obj update always appears after traced_obj update */
 	smp_wmb();
 	atomic_inc(&num_traced_obj);
@@ -75,7 +122,7 @@ static void set_trace_object(void *obj)
 }
 
 static void submit_trace_object(unsigned long ip, unsigned long parent_ip,
-				 unsigned long object)
+				 unsigned long object, unsigned long value)
 {
 
 	struct trace_buffer *buffer;
@@ -92,19 +139,52 @@ static void submit_trace_object(unsigned long ip, unsigned long parent_ip,
 	entry->ip                       = ip;
 	entry->parent_ip                = parent_ip;
 	entry->object			= object;
+	entry->value			= value;
 
 	event_trigger_unlock_commit(&event_trace_file, buffer, event,
 		entry, pc);
 }
 
+static inline long get_object_value(unsigned long *val, void *obj, int type_size)
+{
+	char tmp[sizeof(u64)];
+	long ret = 0;
+
+	ret = copy_from_kernel_nofault(tmp, obj, sizeof(tmp));
+	if (ret)
+		return ret;
+	switch (type_size) {
+	case 1: {
+		*val = (unsigned long)*(u8 *)tmp;
+		break;
+	}
+	case 2: {
+		*val = (unsigned long)*(u16 *)tmp;
+		break;
+	}
+	case 4: {
+		*val = (unsigned long)*(u32 *)tmp;
+		break;
+	}
+	case 8: {
+		*val = (unsigned long)*(u64 *)tmp;
+		break;
+	}
+	default:
+		return -EINVAL;
+	}
+
+	return 0;
+}
+
 static void
 trace_object_events_call(unsigned long ip, unsigned long parent_ip,
 		struct ftrace_ops *op, struct ftrace_regs *fregs)
 {
 	struct pt_regs *pt_regs = ftrace_get_regs(fregs);
-	unsigned long obj;
+	int n, ret, val_type_size, obj_offset;
+	unsigned long obj, val;
 	unsigned int disabled;
-	int n;
 
 	preempt_disable_notrace();
 
@@ -117,8 +197,19 @@ trace_object_events_call(unsigned long ip, unsigned long parent_ip,
 
 	for (n = 0; n < max_args_num; n++) {
 		obj = regs_get_kernel_argument(pt_regs, n);
-		if (object_exist((void *)obj))
-			submit_trace_object(ip, parent_ip, obj);
+		if (object_exist((void *)obj)) {
+			ret =  get_object_attr((void *)obj, OBJ_OFFSET, &obj_offset);
+			if (unlikely(ret) < 0)
+				goto out;
+
+			get_object_attr((void *)obj, OBJ_VAL_TYPE_SIZE, &val_type_size);
+			if (unlikely(ret) < 0)
+				goto out;
+
+			if (get_object_value(&val, (void *)(obj + obj_offset), val_type_size))
+				continue;
+			submit_trace_object(ip, parent_ip, obj, val);
+		}
 	/* The parameters of a function may match multiple objects */
 	}
 out:
@@ -138,11 +229,12 @@ trace_object_trigger(struct event_trigger_data *data,
 {
 	struct objtrace_trigger_data *obj_data = data->private_data;
 	struct ftrace_event_field *field;
-	void *obj = NULL;
+	void *obj;
 
 	field = obj_data->field;
 	memcpy(&obj, rec + field->offset, sizeof(obj));
-	set_trace_object(obj);
+	/* set the offset from the special object and the type size of the value*/
+	set_trace_object(obj, obj_data->obj_offset, obj_data->obj_value_type_size);
 }
 
 static void
@@ -180,18 +272,45 @@ static int event_object_trigger_init(struct event_trigger_ops *ops,
 	return 0;
 }
 
+static const struct objtrace_fetch_type objtrace_fetch_types[] = {
+	{"u8", 1},
+	{"s8", 1},
+	{"x8", 1},
+	{"u16", 2},
+	{"s16", 2},
+	{"x16", 2},
+	{"u32", 4},
+	{"s32", 4},
+	{"x32", 4},
+	{"u64", 8},
+	{"s64", 8},
+	{"x64", 8},
+	{}
+};
+
 static int
 event_trigger_print(const char *name, struct seq_file *m,
-		void *data, char *filter_str, void *objtrace_data)
+		    void *data, char *filter_str, void *objtrace_data)
 {
+	int i;
 	long count = (long)data;
 	struct objtrace_trigger_data *obj_data = objtrace_data;
+	const char *value_type_name;
 
 	seq_puts(m, name);
 
 	seq_printf(m, ":%s", obj_data->objtrace_cmd);
 	seq_printf(m, ":%s", obj_data->field->name);
+	if (obj_data->obj_offset)
+		seq_printf(m, ",0x%x", obj_data->obj_offset);
 
+	for (i = 0; objtrace_fetch_types[i].name; i++) {
+		if (objtrace_fetch_types[i].type_size == obj_data->obj_value_type_size) {
+			value_type_name = objtrace_fetch_types[i].name;
+			break;
+		}
+	}
+	seq_printf(m, ":%s", value_type_name);
 	if (count == -1)
 		seq_puts(m, ":unlimited");
 	else
@@ -325,18 +444,20 @@ event_object_trigger_parse(struct event_command *cmd_ops,
 	struct objtrace_trigger_data *obj_data;
 	struct trace_event_call *call;
 	struct ftrace_event_field *field;
-	char *objtrace_cmd;
-	char *trigger = NULL;
-	char *arg;
-	char *number;
-	int ret;
+	char *type, *tr, *obj, *tmp, *trigger = NULL;
+	char *number, *objtrace_cmd;
+	int ret, i, def_type_size, obj_value_type_size = 0;
+	long offset = 0;
 	bool remove = false;
 
 	ret = -EINVAL;
 	if (!param)
 		goto out;
 
-	/* separate the trigger from the filter (c:a:n [if filter]) */
+	/*
+	 * separate the trigger from the filter:
+	 * objtrace:add:OBJ[,OFFS][:TYPE][:COUNT] [if filter]
+	 */
 	trigger = strsep(&param, " \t");
 	if (!trigger)
 		goto out;
@@ -345,16 +466,27 @@ event_object_trigger_parse(struct event_command *cmd_ops,
 		if (!*param)
 			param = NULL;
 	}
-
 	objtrace_cmd = strsep(&trigger, ":");
 	if (!objtrace_cmd || strcmp(objtrace_cmd, "add"))
 		goto out;
 
-	arg = strsep(&trigger, ":");
-	if (!arg)
+	obj = strsep(&trigger, ":");
+	if (!obj)
 		goto out;
+
+	tr = strchr(obj, ',');
+	if (!tr)
+		offset = 0;
+	else {
+		*tr++ = '\0';
+		ret = kstrtol(tr, 0, &offset);
+		if (ret)
+			goto out;
+	}
+
+	ret = -EINVAL;
 	call = file->event_call;
-	field = trace_find_event_field(call, arg);
+	field = trace_find_event_field(call, obj);
 	if (!field)
 		goto out;
 
@@ -365,36 +497,65 @@ event_object_trigger_parse(struct event_command *cmd_ops,
 		remove = true;
 
 	if (remove && !field_exist(file, cmd_ops, field->name))
-	goto out;
-	trigger_ops = cmd_ops->get_trigger_ops(cmd, trigger);
-	ret = -ENOMEM;
-	obj_data = kzalloc(sizeof(*obj_data), GFP_KERNEL);
-	if (!obj_data)
 		goto out;
-	obj_data->field = field;
-	snprintf(obj_data->objtrace_cmd, OBJTRACE_CMD_LEN, objtrace_cmd);
-
+	ret = -ENOMEM;
 	trigger_data = kzalloc(sizeof(*trigger_data), GFP_KERNEL);
-	if (!trigger_data) {
-		kfree(obj_data);
+	if (!trigger_data)
 		goto out;
-	}
 
 	trigger_data->count = -1;
-	trigger_data->ops = trigger_ops;
 	trigger_data->cmd_ops = cmd_ops;
-	trigger_data->private_data = obj_data;
 	INIT_LIST_HEAD(&trigger_data->list);
 	INIT_LIST_HEAD(&trigger_data->named_list);
 
 	if (remove) {
 		cmd_ops->unreg(glob+1, trigger_data, file);
-		kfree(obj_data);
 		kfree(trigger_data);
 		ret = 0;
 		goto out;
 	}
 
+	ret = -EINVAL;
+	def_type_size = sizeof(void *);
+	if (!trigger) {
+		obj_value_type_size = def_type_size;
+		goto skip_get_type;
+	}
+
+	tmp = trigger;
+	type = strsep(&trigger, ":");
+	if (!type)
+		obj_value_type_size = def_type_size;
+	else if (isdigit(type[0])) {
+		obj_value_type_size = def_type_size;
+		trigger = tmp;
+	} else {
+		for (i = 0; objtrace_fetch_types[i].name; i++) {
+			if (strcmp(objtrace_fetch_types[i].name, type) == 0) {
+				obj_value_type_size = objtrace_fetch_types[i].type_size;
+				break;
+			}
+		}
+	}
+	if (!obj_value_type_size)
+		goto out;
+skip_get_type:
+	trigger_ops = cmd_ops->get_trigger_ops(cmd, trigger);
+	trigger_data->ops = trigger_ops;
+
+	ret = -ENOMEM;
+	obj_data = kzalloc(sizeof(*obj_data), GFP_KERNEL);
+	if (!obj_data) {
+		kfree(trigger_data);
+		goto out;
+	}
+
+	obj_data->field = field;
+	obj_data->obj_offset = offset;
+	obj_data->obj_value_type_size = obj_value_type_size;
+	snprintf(obj_data->objtrace_cmd, OBJTRACE_CMD_LEN, objtrace_cmd);
+	trigger_data->private_data = obj_data;
+
 	if (trigger) {
 		number = strsep(&trigger, ":");
 
diff --git a/kernel/trace/trace_output.c b/kernel/trace/trace_output.c
index 265428154638..04804f53cadd 100644
--- a/kernel/trace/trace_output.c
+++ b/kernel/trace/trace_output.c
@@ -1557,6 +1557,7 @@ static enum print_line_t trace_object_print(struct trace_iterator *iter, int fla
 	trace_assign_type(field, iter->ent);
 	print_fn_trace(s, field->ip, field->parent_ip, flags);
 	trace_seq_printf(s, " object:0x%lx", field->object);
+	trace_seq_printf(s, " value:0x%lx", field->value);
 	trace_seq_putc(s, '\n');
 
 	return trace_handle_return(s);
@@ -1569,9 +1570,8 @@ static enum print_line_t trace_object_raw(struct trace_iterator *iter, int flags
 
 	trace_assign_type(field, iter->ent);
 
-	trace_seq_printf(&iter->seq, "%lx %lx\n",
-			 field->ip,
-			 field->parent_ip);
+	trace_seq_printf(&iter->seq, "%lx %lx %lx %lx\n", field->ip,
+			field->parent_ip, field->object, field->value);
 
 	return trace_handle_return(&iter->seq);
 }
-- 
2.25.1


^ permalink raw reply related	[flat|nested] 18+ messages in thread

* [PATCH v9 3/4] trace/objtrace: Add testcases for objtrace
  2022-02-04  3:56 [PATCH v9 0/4] trace: Introduce objtrace trigger to trace the kernel object Jeff Xie
  2022-02-04  3:56 ` [PATCH v9 1/4] trace: Add trace any " Jeff Xie
  2022-02-04  3:56 ` [PATCH v9 2/4] trace/objtrace: get the value of the object Jeff Xie
@ 2022-02-04  3:56 ` Jeff Xie
  2022-02-04  3:56 ` [PATCH v9 4/4] trace/objtrace: Add documentation " Jeff Xie
  2022-02-08 14:08 ` [PATCH v9 0/4] trace: Introduce objtrace trigger to trace the kernel object Masami Hiramatsu
  4 siblings, 0 replies; 18+ messages in thread
From: Jeff Xie @ 2022-02-04  3:56 UTC (permalink / raw)
  To: rostedt; +Cc: mhiramat, mingo, zanussi, linux-kernel, Jeff Xie

Add a series of testcases to illustrate correct and incorrect usage of
objtrace trigger.

Signed-off-by: Jeff Xie <xiehuan09@gmail.com>
---
 .../ftrace/test.d/trigger/trigger-objtrace.tc | 41 +++++++++++++++++++
 1 file changed, 41 insertions(+)
 create mode 100644 tools/testing/selftests/ftrace/test.d/trigger/trigger-objtrace.tc

diff --git a/tools/testing/selftests/ftrace/test.d/trigger/trigger-objtrace.tc b/tools/testing/selftests/ftrace/test.d/trigger/trigger-objtrace.tc
new file mode 100644
index 000000000000..d894442b6a30
--- /dev/null
+++ b/tools/testing/selftests/ftrace/test.d/trigger/trigger-objtrace.tc
@@ -0,0 +1,41 @@
+#!/bin/sh
+# SPDX-License-Identifier: GPL-2.0
+# description: event trigger - test objtrace-trigger
+# requires: kprobe_events "objtrace":README
+
+fail() { #msg
+    echo $1
+    exit_fail
+}
+
+echo 'p bio_add_page arg1=$arg1 arg2=$arg2' > kprobe_events
+
+FEATURE=`grep objtrace events/kprobes/p_bio_add_page_0/trigger`
+if [ -z "$FEATURE" ]; then
+    echo "objtrace trigger is not supported"
+    exit_unsupported
+fi
+
+echo "Test objtrace trigger"
+echo 'objtrace:add:arg1,0x28:u32:1 if comm == "cat"' > \
+	events/kprobes/p_bio_add_page_0/trigger
+if [ -z $? ]; then
+	fail "objtrace trigger syntax error"
+fi
+
+echo "Test objtrace semantic errors"
+
+# Being lack of objtrace command
+! echo 'objtrace:arg1,0x28:u32:1' > events/kprobes/p_bio_add_page_0/trigger
+# Bad parameter name
+! echo 'objtrace:add:argx:u32:1' > events/kprobes/p_bio_add_page_0/trigger
+# The parameter existed on event
+! echo 'objtrace:add:arg2:u32:1' > events/kprobes/p_bio_add_page_0/trigger
+
+echo "reset objtrace trigger"
+
+echo '!objtrace:add:arg1,0x28:u32' > \
+	events/kprobes/p_bio_add_page_0/trigger
+echo '-:p_bio_add_page_0' >> ./kprobe_events
+
+exit 0
-- 
2.25.1


^ permalink raw reply related	[flat|nested] 18+ messages in thread

* [PATCH v9 4/4] trace/objtrace: Add documentation for objtrace
  2022-02-04  3:56 [PATCH v9 0/4] trace: Introduce objtrace trigger to trace the kernel object Jeff Xie
                   ` (2 preceding siblings ...)
  2022-02-04  3:56 ` [PATCH v9 3/4] trace/objtrace: Add testcases for objtrace Jeff Xie
@ 2022-02-04  3:56 ` Jeff Xie
  2022-02-08 14:08 ` [PATCH v9 0/4] trace: Introduce objtrace trigger to trace the kernel object Masami Hiramatsu
  4 siblings, 0 replies; 18+ messages in thread
From: Jeff Xie @ 2022-02-04  3:56 UTC (permalink / raw)
  To: rostedt; +Cc: mhiramat, mingo, zanussi, linux-kernel, Jeff Xie

Added documentation explaining how to use objtrace trigger to get the value
of the object.

Signed-off-by: Jeff Xie <xiehuan09@gmail.com>
---
 Documentation/trace/events.rst | 83 ++++++++++++++++++++++++++++++++++
 1 file changed, 83 insertions(+)

diff --git a/Documentation/trace/events.rst b/Documentation/trace/events.rst
index c47f381d0c00..0dc475160133 100644
--- a/Documentation/trace/events.rst
+++ b/Documentation/trace/events.rst
@@ -546,6 +546,89 @@ The following commands are supported:
 
   See Documentation/trace/histogram.rst for details and examples.
 
+- objtrace
+
+  This command provides a way to get the value of any object, The object
+  can be obtained from the dynamic event(kprobe_event/uprobe_event) or the
+  static event(tracepoint).
+
+  Usage:
+  When using the kprobe event, only need to set the objtrace(a new trigger),
+  we can get the value of the object. The object is from the setting of the
+  kprobe event.
+
+  For example:
+  For the function bio_add_page():
+
+  int bio_add_page(struct bio *bio, struct page *page,
+	unsigned int len, unsigned int offset)
+
+  Firstly, we can set the base of the object, thus the first string "arg1"
+  stands for the value of the first parameter of this function bio_add_gage(),
+
+  # echo 'p bio_add_page arg1=$arg1' > ./kprobe_events
+
+  Secondly, we can get the value dynamically based on above object.
+
+  find the offset of the bi_size in struct bio:
+  $ gdb vmlinux
+  (gdb) p &(((struct bio *)0)->bi_iter.bi_size)
+  $1 = (unsigned int *) 0x28
+
+  # echo 'objtrace:add:arg1,0x28:u32:1 if comm == "cat"' > ./events/kprobes/ \
+	p_bio_add_page_0/trigger
+
+  # cd /sys/kernel/debug/tracing/
+  # echo 'p bio_add_page arg1=$arg1' > ./kprobe_events
+  # echo 'objtrace:add:arg1,0x28:u32:1 if comm == "cat"' > ./events/kprobes/p_bio_add_page_0/trigger
+
+  # du -sh /test.txt
+  12.0K   /test.txt
+
+  # cat  /test.txt > /dev/null
+  # cat ./trace
+  # tracer: nop
+  #
+  # entries-in-buffer/entries-written: 128/128   #P:4
+  #
+  #                                _-----=> irqs-off/BH-disabled
+  #                               / _----=> need-resched
+  #                              | / _---=> hardirq/softirq
+  #                              || / _--=> preempt-depth
+  #                              ||| / _-=> migrate-disable
+  #                              |||| /     delay
+  #           TASK-PID     CPU#  |||||  TIMESTAMP  FUNCTION
+  #              | |         |   |||||     |         |
+               cat-117     [002] ...1.     1.602243: __bio_try_merge_page <-bio_add_page object:0xffff88811bee4000 value:0x0
+               cat-117     [002] ...1.     1.602244: __bio_add_page <-bio_add_page object:0xffff88811bee4000 value:0x0
+               cat-117     [002] ...2.     1.602244: bio_add_page <-ext4_mpage_readpages object:0xffff88811bee4000 value:0x1000
+               cat-117     [002] ...1.     1.602245: __bio_try_merge_page <-bio_add_page object:0xffff88811bee4000 value:0x1000
+               cat-117     [002] ...1.     1.602245: __bio_add_page <-bio_add_page object:0xffff88811bee4000 value:0x1000
+               cat-117     [002] ...2.     1.602245: bio_add_page <-ext4_mpage_readpages object:0xffff88811bee4000 value:0x2000
+               cat-117     [002] ...1.     1.602245: __bio_try_merge_page <-bio_add_page object:0xffff88811bee4000 value:0x2000
+               cat-117     [002] ...1.     1.602245: __bio_add_page <-bio_add_page object:0xffff88811bee4000 value:0x2000
+               cat-117     [002] ...1.     1.602245: submit_bio <-ext4_mpage_readpages object:0xffff88811bee4000 value:0x3000
+               cat-117     [002] ...1.     1.602245: submit_bio_noacct <-ext4_mpage_readpages object:0xffff88811bee4000 value:0x3000
+               cat-117     [002] ...1.     1.602246: __submit_bio <-submit_bio_noacct object:0xffff88811bee4000 value:0x3000
+               cat-117     [002] ...1.     1.602246: submit_bio_checks <-__submit_bio object:0xffff88811bee4000 value:0x3000
+               cat-117     [002] ...1.     1.602246: __cond_resched <-submit_bio_checks object:0xffff88811bee4000 value:0x3000
+               cat-117     [002] ...1.     1.602246: should_fail_bio <-submit_bio_checks object:0xffff88811bee4000 value:0x3000
+               cat-117     [002] ...1.     1.602246: blk_mq_submit_bio <-submit_bio_noacct object:0xffff88811bee4000 value:0x3000
+               cat-117     [002] ...1.     1.602246: blk_attempt_plug_merge <-blk_mq_submit_bio object:0xffff88811bee4000 value:0x3000
+               cat-117     [002] ...1.     1.602246: blk_mq_sched_bio_merge <-blk_mq_submit_bio object:0xffff88811bee4000 value:0x3000
+               cat-117     [002] ...1.     1.602247: __rcu_read_lock <-blk_mq_submit_bio object:0xffff88811bee4000 value:0x3000
+               cat-117     [002] ...1.     1.602247: __rcu_read_unlock <-blk_mq_submit_bio object:0xffff88811bee4000 value:0x3000
+               cat-117     [002] ...1.     1.602247: __blk_mq_alloc_requests <-blk_mq_submit_bio object:0xffff88811bee4000 value:0x3000
+            <idle>-0       [002] d..3.     1.602298: bio_endio <-blk_update_request object:0xffff88811bee4000 value:0x0
+            <idle>-0       [002] d..3.     1.602298: mpage_end_io <-blk_update_request object:0xffff88811bee4000 value:0x0
+            <idle>-0       [002] d..3.     1.602298: __read_end_io <-blk_update_request object:0xffff88811bee4000 value:0x0
+            <idle>-0       [002] d..3.     1.602300: bio_put <-blk_update_request object:0xffff88811bee4000 value:0x0
+            <idle>-0       [002] d..3.     1.602300: bio_free <-blk_update_request object:0xffff88811bee4000 value:0x0
+            <idle>-0       [002] d..3.     1.602300: mempool_free <-blk_update_request object:0xffff88811bee4000 value:0x0
+            <idle>-0       [002] d..3.     1.602300: mempool_free_slab <-blk_update_request object:0xffff88811bee4000 value:0x0
+            <idle>-0       [002] d..3.     1.602300: kmem_cache_free <-blk_update_request object:0xffff88811bee4000 value:0x0
+             ...
+
 7. In-kernel trace event API
 ============================
 
-- 
2.25.1


^ permalink raw reply related	[flat|nested] 18+ messages in thread

* Re: [PATCH v9 0/4] trace: Introduce objtrace trigger to trace the kernel object
  2022-02-04  3:56 [PATCH v9 0/4] trace: Introduce objtrace trigger to trace the kernel object Jeff Xie
                   ` (3 preceding siblings ...)
  2022-02-04  3:56 ` [PATCH v9 4/4] trace/objtrace: Add documentation " Jeff Xie
@ 2022-02-08 14:08 ` Masami Hiramatsu
  2022-02-08 15:48   ` Steven Rostedt
  4 siblings, 1 reply; 18+ messages in thread
From: Masami Hiramatsu @ 2022-02-08 14:08 UTC (permalink / raw)
  To: Jeff Xie; +Cc: rostedt, mhiramat, mingo, zanussi, linux-kernel

On Fri,  4 Feb 2022 11:56:40 +0800
Jeff Xie <xiehuan09@gmail.com> wrote:

> Introduce a method based on function tracer to trace any object and get
> the value of the object dynamically. the object can be obtained from the
> dynamic event (kprobe_event/uprobe_event) or the static event(tracepoint).
> 
> Usage:
> When using the kprobe event, only need to set the objtrace(a new trigger),
> we can get the value of the object. The object is from the setting of the 
> kprobe event.
> 
> For example:
> For the function bio_add_page():
> 
> int bio_add_page(struct bio *bio, struct page *page,
> 	unsigned int len, unsigned int offset)
> 
> Firstly, we can set the base of the object, thus the first string "arg1"
> stands for the value of the first parameter of this function bio_add_gage(),
> 
> # echo 'p bio_add_page arg1=$arg1' > ./kprobe_events
> 
> Secondly, we can get the value dynamically based on above object. 
> 
> find the offset of the bi_size in struct bio:
> $ gdb vmlinux
> (gdb) p &(((struct bio *)0)->bi_iter.bi_size)
> $1 = (unsigned int *) 0x28
> 
> # echo 'objtrace:add:arg1,0x28:u32:1 if comm == "cat"' > ./events/kprobes/ \
> 	p_bio_add_page_0/trigger
> 
> # cd /sys/kernel/debug/tracing/
> # echo 'p bio_add_page arg1=$arg1' > ./kprobe_events
> # echo 'objtrace:add:arg1,0x28:u32:1 if comm == "cat"' > ./events/kprobes/p_bio_add_page_0/trigger
> 
> # du -sh /test.txt
> 12.0K   /test.txt
> 
> # cat  /test.txt > /dev/null
> # cat ./trace
> # tracer: nop
> #
> # entries-in-buffer/entries-written: 128/128   #P:4
> #
> #                                _-----=> irqs-off/BH-disabled
> #                               / _----=> need-resched
> #                              | / _---=> hardirq/softirq
> #                              || / _--=> preempt-depth
> #                              ||| / _-=> migrate-disable
> #                              |||| /     delay
> #           TASK-PID     CPU#  |||||  TIMESTAMP  FUNCTION
> #              | |         |   |||||     |         |
>              cat-117     [002] ...1.     1.602243: __bio_try_merge_page <-bio_add_page object:0xffff88811bee4000 value:0x0
>              cat-117     [002] ...1.     1.602244: __bio_add_page <-bio_add_page object:0xffff88811bee4000 value:0x0
>              cat-117     [002] ...2.     1.602244: bio_add_page <-ext4_mpage_readpages object:0xffff88811bee4000 value:0x1000
>              cat-117     [002] ...1.     1.602245: __bio_try_merge_page <-bio_add_page object:0xffff88811bee4000 value:0x1000
>              cat-117     [002] ...1.     1.602245: __bio_add_page <-bio_add_page object:0xffff88811bee4000 value:0x1000
>              cat-117     [002] ...2.     1.602245: bio_add_page <-ext4_mpage_readpages object:0xffff88811bee4000 value:0x2000
>              cat-117     [002] ...1.     1.602245: __bio_try_merge_page <-bio_add_page object:0xffff88811bee4000 value:0x2000
>              cat-117     [002] ...1.     1.602245: __bio_add_page <-bio_add_page object:0xffff88811bee4000 value:0x2000
>              cat-117     [002] ...1.     1.602245: submit_bio <-ext4_mpage_readpages object:0xffff88811bee4000 value:0x3000
>              cat-117     [002] ...1.     1.602245: submit_bio_noacct <-ext4_mpage_readpages object:0xffff88811bee4000 value:0x3000
>              cat-117     [002] ...1.     1.602246: __submit_bio <-submit_bio_noacct object:0xffff88811bee4000 value:0x3000
>              cat-117     [002] ...1.     1.602246: submit_bio_checks <-__submit_bio object:0xffff88811bee4000 value:0x3000
>              cat-117     [002] ...1.     1.602246: __cond_resched <-submit_bio_checks object:0xffff88811bee4000 value:0x3000
>              cat-117     [002] ...1.     1.602246: should_fail_bio <-submit_bio_checks object:0xffff88811bee4000 value:0x3000
>              cat-117     [002] ...1.     1.602246: blk_mq_submit_bio <-submit_bio_noacct object:0xffff88811bee4000 value:0x3000
>              cat-117     [002] ...1.     1.602246: blk_attempt_plug_merge <-blk_mq_submit_bio object:0xffff88811bee4000 value:0x3000
>              cat-117     [002] ...1.     1.602246: blk_mq_sched_bio_merge <-blk_mq_submit_bio object:0xffff88811bee4000 value:0x3000
>              cat-117     [002] ...1.     1.602247: __rcu_read_lock <-blk_mq_submit_bio object:0xffff88811bee4000 value:0x3000
>              cat-117     [002] ...1.     1.602247: __rcu_read_unlock <-blk_mq_submit_bio object:0xffff88811bee4000 value:0x3000
>              cat-117     [002] ...1.     1.602247: __blk_mq_alloc_requests <-blk_mq_submit_bio object:0xffff88811bee4000 value:0x3000
>           <idle>-0       [002] d..3.     1.602298: bio_endio <-blk_update_request object:0xffff88811bee4000 value:0x0
>           <idle>-0       [002] d..3.     1.602298: mpage_end_io <-blk_update_request object:0xffff88811bee4000 value:0x0
>           <idle>-0       [002] d..3.     1.602298: __read_end_io <-blk_update_request object:0xffff88811bee4000 value:0x0
>           <idle>-0       [002] d..3.     1.602300: bio_put <-blk_update_request object:0xffff88811bee4000 value:0x0
>           <idle>-0       [002] d..3.     1.602300: bio_free <-blk_update_request object:0xffff88811bee4000 value:0x0
>           <idle>-0       [002] d..3.     1.602300: mempool_free <-blk_update_request object:0xffff88811bee4000 value:0x0
>           <idle>-0       [002] d..3.     1.602300: mempool_free_slab <-blk_update_request object:0xffff88811bee4000 value:0x0
>           <idle>-0       [002] d..3.     1.602300: kmem_cache_free <-blk_update_request object:0xffff88811bee4000 value:0x0
> 	  ...
> 
> Almost all changelogs were suggested by Masami(mhiramat@kernel.org)
> and steve(rostedt@goodmis.org), thank you all so much.

Thanks for updating the series. I've tested the patches and confirmed
the objtrace is working :)

Tested-by: Masami Hiramatsu <mhiramat@kernel.org>
Reviewed-by: Masami Hiramatsu <mhiramat@kernel.org>

for the series.

Thank you!



> 
> v9:
> - fix objtrace trigger output was incomplete
> - fix the objtrace trigger was removed when using the existed parameter on
>   event.
> - add testcase for the second fix above.
> 
> v8:
> - revert to use per-cpu recursion for the function trace_object_events_call
> - recover the filter when getting the value of the object
> - simplify the implementation for the function get_object_value
> - fix the build error
> 
> v7:
> - use fixed-size array for object pool instead of list structure
> - use ftrace_test_recursion_trylock for function trace hook function
> - fix trace_object_ref reference count in the init_trace_object
> - invoke exit_trace_object no matter whether data->ops->free is null 
>   in the unregister_object_trigger
> - release private_data of event_trigger_data in the trace_object_trigger_free
> - remove [RFC] tag
> 
> v6:
> - change the objtrace trigger syntax.
> - add patchset description
> - add <tracefs>/README
> 
> v5:
> - add testcasts
> - add check the field->size
> - add lockless to search object
> - describe the object trace more clearly in Kconfig
> 
> v4:
> - please ignore the v4 which is the same as v3
> 
> v3:
> - change the objfilter to objtrace
> - add a command to the objfilter syntax
> - change to get the value of the object
> - use trace_find_event_field to find the field instead of using argN
> - get data from @rec in the event trigger callback funciton
> 
> v2:
> - adding a "objfilter" trigger to update object
> 
> Jeff Xie (4):
>   trace: Add trace any kernel object
>   trace/objtrace: get the value of the object
>   trace/objtrace: Add testcases for objtrace
>   trace/objtrace: Add documentation for objtrace
> 
>  Documentation/trace/events.rst                |  83 +++
>  include/linux/trace_events.h                  |   1 +
>  kernel/trace/Kconfig                          |  10 +
>  kernel/trace/Makefile                         |   1 +
>  kernel/trace/trace.c                          |   3 +
>  kernel/trace/trace.h                          |   8 +
>  kernel/trace/trace_entries.h                  |  18 +
>  kernel/trace/trace_events_trigger.c           |   1 +
>  kernel/trace/trace_object.c                   | 676 ++++++++++++++++++
>  kernel/trace/trace_output.c                   |  40 ++
>  .../ftrace/test.d/trigger/trigger-objtrace.tc |  41 ++
>  11 files changed, 882 insertions(+)
>  create mode 100644 kernel/trace/trace_object.c
>  create mode 100644 tools/testing/selftests/ftrace/test.d/trigger/trigger-objtrace.tc
> 
> -- 
> 2.25.1
> 


-- 
Masami Hiramatsu <mhiramat@kernel.org>

^ permalink raw reply	[flat|nested] 18+ messages in thread

* Re: [PATCH v9 0/4] trace: Introduce objtrace trigger to trace the kernel object
  2022-02-08 14:08 ` [PATCH v9 0/4] trace: Introduce objtrace trigger to trace the kernel object Masami Hiramatsu
@ 2022-02-08 15:48   ` Steven Rostedt
  2022-02-26 16:01     ` Jeff Xie
  0 siblings, 1 reply; 18+ messages in thread
From: Steven Rostedt @ 2022-02-08 15:48 UTC (permalink / raw)
  To: Masami Hiramatsu; +Cc: Jeff Xie, mingo, zanussi, linux-kernel

On Tue, 8 Feb 2022 23:08:30 +0900
Masami Hiramatsu <mhiramat@kernel.org> wrote:

> Thanks for updating the series. I've tested the patches and confirmed
> the objtrace is working :)
> 
> Tested-by: Masami Hiramatsu <mhiramat@kernel.org>
> Reviewed-by: Masami Hiramatsu <mhiramat@kernel.org>
> 
> for the series.

Thanks. Just to give you an update on my own workload.

I'm finishing my last week of on-boarding, and hopefully, I'll have more
time next week to look at more things upstream. And hopefully, I'll be able
to get to looking at this series then.

Cheers,

-- Steve

^ permalink raw reply	[flat|nested] 18+ messages in thread

* Re: [PATCH v9 0/4] trace: Introduce objtrace trigger to trace the kernel object
  2022-02-08 15:48   ` Steven Rostedt
@ 2022-02-26 16:01     ` Jeff Xie
  2022-02-28 16:08       ` Steven Rostedt
  0 siblings, 1 reply; 18+ messages in thread
From: Jeff Xie @ 2022-02-26 16:01 UTC (permalink / raw)
  To: Steven Rostedt; +Cc: Masami Hiramatsu, mingo, Tom Zanussi, linux-kernel

Hi Steve and Masami,

On Tue, Feb 8, 2022 at 11:48 PM Steven Rostedt <rostedt@goodmis.org> wrote:
>
> On Tue, 8 Feb 2022 23:08:30 +0900
> Masami Hiramatsu <mhiramat@kernel.org> wrote:
>
> > Thanks for updating the series. I've tested the patches and confirmed
> > the objtrace is working :)
> >
> > Tested-by: Masami Hiramatsu <mhiramat@kernel.org>
> > Reviewed-by: Masami Hiramatsu <mhiramat@kernel.org>
> >
> > for the series.

Thanks, It's very difficult to get your double "by" and I'm very excited ;-)

>
> Thanks. Just to give you an update on my own workload.
>
> I'm finishing my last week of on-boarding, and hopefully, I'll have more
> time next week to look at more things upstream. And hopefully, I'll be able
> to get to looking at this series then.
>

Congratulations on joining google.  Just check out this series when
you are free.

Please don't get me wrong, I'm not pushing anyone.
It just doesn't feel good that I haven't responded to emails for a long time ;-)

> Cheers,
>
> -- Steve
---
JeffXie

^ permalink raw reply	[flat|nested] 18+ messages in thread

* Re: [PATCH v9 0/4] trace: Introduce objtrace trigger to trace the kernel object
  2022-02-26 16:01     ` Jeff Xie
@ 2022-02-28 16:08       ` Steven Rostedt
  2022-03-22 17:20         ` Jeff Xie
  0 siblings, 1 reply; 18+ messages in thread
From: Steven Rostedt @ 2022-02-28 16:08 UTC (permalink / raw)
  To: Jeff Xie; +Cc: Masami Hiramatsu, mingo, Tom Zanussi, linux-kernel

On Sun, 27 Feb 2022 00:01:06 +0800
Jeff Xie <xiehuan09@gmail.com> wrote:

> Congratulations on joining google.  Just check out this series when
> you are free.
> 
> Please don't get me wrong, I'm not pushing anyone.
> It just doesn't feel good that I haven't responded to emails for a long time ;-)

And keep responding ;-) I want to look at this series, and your emails do
remind me (it's still in my patchwork queue, so it wont be forgotten, but
it is getting crowded in that queue of "todo"s).

Yeah, I'm hoping to start being able to do more upstream, but I'm still a
bit in the flux of figuring out what I'm suppose to be doing at work ;-)

-- Steve

^ permalink raw reply	[flat|nested] 18+ messages in thread

* Re: [PATCH v9 0/4] trace: Introduce objtrace trigger to trace the kernel object
  2022-02-28 16:08       ` Steven Rostedt
@ 2022-03-22 17:20         ` Jeff Xie
  2022-04-11 15:47           ` Jeff Xie
  0 siblings, 1 reply; 18+ messages in thread
From: Jeff Xie @ 2022-03-22 17:20 UTC (permalink / raw)
  To: Steven Rostedt; +Cc: Masami Hiramatsu, mingo, Tom Zanussi, linux-kernel

Hi steve,

On Tue, Mar 1, 2022 at 12:08 AM Steven Rostedt <rostedt@goodmis.org> wrote:
>
> On Sun, 27 Feb 2022 00:01:06 +0800
> Jeff Xie <xiehuan09@gmail.com> wrote:
>
> > Congratulations on joining google.  Just check out this series when
> > you are free.
> >
> > Please don't get me wrong, I'm not pushing anyone.
> > It just doesn't feel good that I haven't responded to emails for a long time ;-)
>
> And keep responding ;-) I want to look at this series, and your emails do
> remind me (it's still in my patchwork queue, so it wont be forgotten, but
> it is getting crowded in that queue of "todo"s).
>
> Yeah, I'm hoping to start being able to do more upstream, but I'm still a
> bit in the flux of figuring out what I'm suppose to be doing at work ;-)
>
> -- Steve

Hope this series is less crowded in your todo queue ;-)
---
JeffXie

^ permalink raw reply	[flat|nested] 18+ messages in thread

* Re: [PATCH v9 0/4] trace: Introduce objtrace trigger to trace the kernel object
  2022-03-22 17:20         ` Jeff Xie
@ 2022-04-11 15:47           ` Jeff Xie
  2022-04-14 16:52             ` Steven Rostedt
  0 siblings, 1 reply; 18+ messages in thread
From: Jeff Xie @ 2022-04-11 15:47 UTC (permalink / raw)
  To: Steven Rostedt; +Cc: Masami Hiramatsu, mingo, Tom Zanussi, linux-kernel

On Wed, Mar 23, 2022 at 1:20 AM Jeff Xie <xiehuan09@gmail.com> wrote:
>
> Hi steve,
>
> On Tue, Mar 1, 2022 at 12:08 AM Steven Rostedt <rostedt@goodmis.org> wrote:
> >
> > On Sun, 27 Feb 2022 00:01:06 +0800
> > Jeff Xie <xiehuan09@gmail.com> wrote:
> >
> > > Congratulations on joining google.  Just check out this series when
> > > you are free.
> > >
> > > Please don't get me wrong, I'm not pushing anyone.
> > > It just doesn't feel good that I haven't responded to emails for a long time ;-)
> >
> > And keep responding ;-) I want to look at this series, and your emails do
> > remind me (it's still in my patchwork queue, so it wont be forgotten, but
> > it is getting crowded in that queue of "todo"s).
> >
> > Yeah, I'm hoping to start being able to do more upstream, but I'm still a
> > bit in the flux of figuring out what I'm suppose to be doing at work ;-)
> >
> > -- Steve
>
> Hope this series is less crowded in your todo queue ;-)
> ---
> JeffXie

Hi Steve,

Just don't want you to forget this patch series ;-)

---
JeffXie

^ permalink raw reply	[flat|nested] 18+ messages in thread

* Re: [PATCH v9 0/4] trace: Introduce objtrace trigger to trace the kernel object
  2022-04-11 15:47           ` Jeff Xie
@ 2022-04-14 16:52             ` Steven Rostedt
  0 siblings, 0 replies; 18+ messages in thread
From: Steven Rostedt @ 2022-04-14 16:52 UTC (permalink / raw)
  To: Jeff Xie; +Cc: Masami Hiramatsu, mingo, Tom Zanussi, linux-kernel

On Mon, 11 Apr 2022 23:47:23 +0800
Jeff Xie <xiehuan09@gmail.com> wrote:

> Just don't want you to forget this patch series ;-)

I won't. I've just been super busy (on my way back home from Denver).
I'll try to get some time tomorrow to review them.

Thanks for being patient with me.

-- Steve

^ permalink raw reply	[flat|nested] 18+ messages in thread

* Re: [PATCH v9 1/4] trace: Add trace any kernel object
  2022-02-04  3:56 ` [PATCH v9 1/4] trace: Add trace any " Jeff Xie
@ 2022-04-19  2:14   ` Steven Rostedt
  2022-04-19  2:26     ` Steven Rostedt
  2022-04-19 16:26     ` Jeff Xie
  2022-04-27 15:50   ` Jeff Xie
  1 sibling, 2 replies; 18+ messages in thread
From: Steven Rostedt @ 2022-04-19  2:14 UTC (permalink / raw)
  To: Jeff Xie; +Cc: mhiramat, mingo, zanussi, linux-kernel

On Fri,  4 Feb 2022 11:56:41 +0800
Jeff Xie <xiehuan09@gmail.com> wrote:

> Introduce objtrace trigger to get the call flow by tracing any kernel
> object in the function parameter.
> 
> The objtrace trigger makes a list of the target object address from
> the given event parameter, and records all kernel function calls
> which has the object address in its parameter.
> 
> Syntax:
> 	objtrace:add:obj[:count][if <filter>]
> 
> Usage:
> 	# echo 'p bio_add_page arg1=$arg1' > kprobe_events
> 	# cd events/kprobes/p_bio_add_page_0
> 	# echo 'objtrace:add:arg1:1 if comm == "cat"' > ./trigger
> 	# cat /test.txt

OK, so patch one only does the object and not the more useful "value"
field. That comes in patch 2.

> 
> Signed-off-by: Jeff Xie <xiehuan09@gmail.com>
> ---
>  include/linux/trace_events.h        |   1 +
>  kernel/trace/Kconfig                |  10 +
>  kernel/trace/Makefile               |   1 +
>  kernel/trace/trace.c                |   3 +
>  kernel/trace/trace.h                |   8 +
>  kernel/trace/trace_entries.h        |  17 +
>  kernel/trace/trace_events_trigger.c |   1 +
>  kernel/trace/trace_object.c         | 515 ++++++++++++++++++++++++++++
>  kernel/trace/trace_output.c         |  40 +++
>  9 files changed, 596 insertions(+)
>  create mode 100644 kernel/trace/trace_object.c
> 
> diff --git a/include/linux/trace_events.h b/include/linux/trace_events.h
> index 70c069aef02c..3ccdbc1d25dd 100644
> --- a/include/linux/trace_events.h
> +++ b/include/linux/trace_events.h
> @@ -685,6 +685,7 @@ enum event_trigger_type {
>  	ETT_EVENT_HIST		= (1 << 4),
>  	ETT_HIST_ENABLE		= (1 << 5),
>  	ETT_EVENT_EPROBE	= (1 << 6),
> +	ETT_TRACE_OBJECT	= (1 << 7),
>  };
>  
>  extern int filter_match_preds(struct event_filter *filter, void *rec);
> diff --git a/kernel/trace/Kconfig b/kernel/trace/Kconfig
> index a5eb5e7fd624..c51b7eb1508d 100644
> --- a/kernel/trace/Kconfig
> +++ b/kernel/trace/Kconfig
> @@ -250,6 +250,16 @@ config FUNCTION_PROFILER
>  
>  	  If in doubt, say N.
>  
> +config TRACE_OBJECT
> +	bool "Trace kernel object in function parameter"
> +	depends on FUNCTION_TRACER
> +	depends on HAVE_FUNCTION_ARG_ACCESS_API
> +	select TRACING
> +	default y
> +	help
> +	 You can trace the kernel object in the kernel function parameter.
> +	 The kernel object is dynamically specified via event trigger.
> +
>  config STACK_TRACER
>  	bool "Trace max stack"
>  	depends on HAVE_FUNCTION_TRACER
> diff --git a/kernel/trace/Makefile b/kernel/trace/Makefile
> index bedc5caceec7..b924b8e55922 100644
> --- a/kernel/trace/Makefile
> +++ b/kernel/trace/Makefile
> @@ -67,6 +67,7 @@ obj-$(CONFIG_FUNCTION_GRAPH_TRACER) += trace_functions_graph.o
>  obj-$(CONFIG_TRACE_BRANCH_PROFILING) += trace_branch.o
>  obj-$(CONFIG_BLK_DEV_IO_TRACE) += blktrace.o
>  obj-$(CONFIG_FUNCTION_GRAPH_TRACER) += fgraph.o
> +obj-$(CONFIG_TRACE_OBJECT) += trace_object.o
>  ifeq ($(CONFIG_BLOCK),y)
>  obj-$(CONFIG_EVENT_TRACING) += blktrace.o
>  endif
> diff --git a/kernel/trace/trace.c b/kernel/trace/trace.c
> index a734d5ae34c8..b4513c2bbd52 100644
> --- a/kernel/trace/trace.c
> +++ b/kernel/trace/trace.c
> @@ -5605,6 +5605,9 @@ static const char readme_msg[] =
>  	"\t            enable_hist:<system>:<event>\n"
>  	"\t            disable_hist:<system>:<event>\n"
>  #endif
> +#ifdef CONFIG_TRACE_OBJECT
> +	"\t            objtrace:add:obj[:count][if <filter>]\n"
> +#endif
>  #ifdef CONFIG_STACKTRACE
>  	"\t\t    stacktrace\n"
>  #endif
> diff --git a/kernel/trace/trace.h b/kernel/trace/trace.h
> index 0f5e22238cd2..8b66515a36d5 100644
> --- a/kernel/trace/trace.h
> +++ b/kernel/trace/trace.h
> @@ -54,6 +54,7 @@ enum trace_type {
>  	TRACE_TIMERLAT,
>  	TRACE_RAW_DATA,
>  	TRACE_FUNC_REPEATS,
> +	TRACE_OBJECT,
>  
>  	__TRACE_LAST_TYPE,
>  };
> @@ -472,6 +473,7 @@ extern void __ftrace_bad_type(void);
>  			  TRACE_GRAPH_RET);		\
>  		IF_ASSIGN(var, ent, struct func_repeats_entry,		\
>  			  TRACE_FUNC_REPEATS);				\
> +		IF_ASSIGN(var, ent, struct trace_object_entry, TRACE_OBJECT);\
>  		__ftrace_bad_type();					\
>  	} while (0)
>  
> @@ -1537,6 +1539,12 @@ static inline int register_trigger_hist_cmd(void) { return 0; }
>  static inline int register_trigger_hist_enable_disable_cmds(void) { return 0; }
>  #endif
>  
> +#ifdef CONFIG_TRACE_OBJECT
> +extern int register_trigger_object_cmd(void);
> +#else
> +static inline int register_trigger_object_cmd(void) { return 0; }
> +#endif
> +
>  extern int register_trigger_cmds(void);
>  extern void clear_event_triggers(struct trace_array *tr);
>  
> diff --git a/kernel/trace/trace_entries.h b/kernel/trace/trace_entries.h
> index cd41e863b51c..bb120d9498a9 100644
> --- a/kernel/trace/trace_entries.h
> +++ b/kernel/trace/trace_entries.h
> @@ -401,3 +401,20 @@ FTRACE_ENTRY(timerlat, timerlat_entry,
>  		 __entry->context,
>  		 __entry->timer_latency)
>  );
> +
> +/*
> + * trace object entry:
> + */
> +FTRACE_ENTRY(object, trace_object_entry,
> +
> +	TRACE_OBJECT,
> +
> +	F_STRUCT(
> +		__field(	unsigned long,		ip		)
> +		__field(	unsigned long,		parent_ip	)
> +		__field(	unsigned long,		object		)
> +	),
> +
> +	F_printk(" %ps <-- %ps object:%lx\n",
> +		 (void *)__entry->ip, (void *)__entry->parent_ip, __entry->object)
> +);
> diff --git a/kernel/trace/trace_events_trigger.c b/kernel/trace/trace_events_trigger.c
> index d00fee705f9c..c3371a6902af 100644
> --- a/kernel/trace/trace_events_trigger.c
> +++ b/kernel/trace/trace_events_trigger.c
> @@ -2025,6 +2025,7 @@ __init int register_trigger_cmds(void)
>  	register_trigger_enable_disable_cmds();
>  	register_trigger_hist_enable_disable_cmds();
>  	register_trigger_hist_cmd();
> +	register_trigger_object_cmd();
>  
>  	return 0;
>  }
> diff --git a/kernel/trace/trace_object.c b/kernel/trace/trace_object.c
> new file mode 100644
> index 000000000000..540e387c613a
> --- /dev/null
> +++ b/kernel/trace/trace_object.c
> @@ -0,0 +1,515 @@
> +// SPDX-License-Identifier: GPL-2.0
> +/*
> + * trace the kernel object in the kernel function parameter
> + * Copyright (C) 2021 Jeff Xie <xiehuan09@gmail.com>
> + */
> +
> +#define pr_fmt(fmt) "trace_object: " fmt
> +
> +#include "trace_output.h"
> +
> +#define MAX_TRACED_OBJECT 5
> +#define OBJTRACE_CMD_LEN  10

The len should at least be a multiple of word size. 16?

> +static DEFINE_PER_CPU(unsigned int, trace_object_event_disable);
> +static DEFINE_RAW_SPINLOCK(trace_obj_lock);
> +static struct trace_event_file event_trace_file;

You do not want to use a global "event_trace_file". And the writes
should be done in the instance (trace_array) that the trigger was added
to. That is, if someone does:

  # mkdir instances/foo
  # echo 'objtrace:add:arg1:1 if comm == "cat"' > instances/foo/events/kprobes/p_bio_add_page_0/trigger

Then the tracing should happen in the instance/foo/trace and not at the
toplevel trace file.

> +static const int max_args_num = 6;
> +static atomic_t trace_object_ref;
> +static atomic_t num_traced_obj;
> +static int exit_trace_object(void);
> +static int init_trace_object(void);
> +
> +static struct object_instance {
> +	void *obj;

Either have a pointer to the trace_array that the object is for here,
or we can have each trace array have a max of 5 objects and add it to
the trace_array when it is created.

> +} traced_obj[MAX_TRACED_OBJECT];
> +
> +/* objtrace private data */
> +struct objtrace_trigger_data {
> +	struct ftrace_event_field *field;
> +	char objtrace_cmd[OBJTRACE_CMD_LEN];
> +};
> +
> +static bool object_exist(void *obj)
> +{
> +	int i, max;
> +
> +	max = atomic_read(&num_traced_obj);
> +	smp_rmb();
> +	for (i = 0; i < max; i++) {
> +		if (traced_obj[i].obj == obj)
> +			return true;
> +	}
> +	return false;
> +}
> +
> +static bool object_empty(void)
> +{
> +	return !atomic_read(&num_traced_obj);
> +}
> +
> +static void set_trace_object(void *obj)
> +{
> +	unsigned long flags;
> +
> +	if (in_nmi())
> +		return;
> +
> +	if (!obj)
> +		return;
> +
> +	if (object_exist(obj))
> +		return;
> +
> +	/* only this place has write operations */
> +	raw_spin_lock_irqsave(&trace_obj_lock, flags);
> +	if (atomic_read(&num_traced_obj) == MAX_TRACED_OBJECT) {
> +		trace_printk("object_pool is full, can't trace object:0x%px\n", obj);

You can't use trace_printk(). That will trigger a nasty message in
dmesg. You can use trace_array_printk_buf() instead.

> +		goto out;
> +	}
> +	traced_obj[atomic_read(&num_traced_obj)].obj = obj;
> +	/* make sure the num_traced_obj update always appears after traced_obj update */
> +	smp_wmb();
> +	atomic_inc(&num_traced_obj);
> +out:
> +	raw_spin_unlock_irqrestore(&trace_obj_lock, flags);
> +}
> +
> +static void submit_trace_object(unsigned long ip, unsigned long parent_ip,
> +				 unsigned long object)
> +{
> +
> +	struct trace_buffer *buffer;
> +	struct ring_buffer_event *event;
> +	struct trace_object_entry *entry;
> +	int pc;
> +
> +	pc = preempt_count();
> +	event = trace_event_buffer_lock_reserve(&buffer, &event_trace_file,
> +			TRACE_OBJECT, sizeof(*entry), pc);

We should not be using trace_event_buffer_lock_reserver() here, as it
is expected to be done for trace events, not unique data like this.
This is more like the function tracer, and use
trace_buffer_lock_reserve() instead.

> +	if (!event)
> +		return;
> +	entry   = ring_buffer_event_data(event);
> +	entry->ip                       = ip;
> +	entry->parent_ip                = parent_ip;
> +	entry->object			= object;
> +
> +	event_trigger_unlock_commit(&event_trace_file, buffer, event,
> +		entry, pc);
> +}
> +
> +static void
> +trace_object_events_call(unsigned long ip, unsigned long parent_ip,
> +		struct ftrace_ops *op, struct ftrace_regs *fregs)
> +{

So the op can hold data, which we should allocate it so that there is
one ftrace_ops per trace_array (like function tracing does). May need
callbacks from mkdir and rmdir in trace.c. But the trace_array (tr)
will hold the buffer you want to write to. Or at the very least, the
object can hold it.


> +	struct pt_regs *pt_regs = ftrace_get_regs(fregs);
> +	unsigned long obj;
> +	unsigned int disabled;
> +	int n;
> +
> +	preempt_disable_notrace();
> +
> +	disabled = this_cpu_inc_return(trace_object_event_disable);

If we make this list per instance, we will only want to disable the
instance.

> +	if (disabled != 1)
> +		goto out;
> +
> +	if (object_empty())
> +		goto out;
> +
> +	for (n = 0; n < max_args_num; n++) {
> +		obj = regs_get_kernel_argument(pt_regs, n);
> +		if (object_exist((void *)obj))
> +			submit_trace_object(ip, parent_ip, obj);
> +	/* The parameters of a function may match multiple objects */
> +	}
> +out:
> +	this_cpu_dec(trace_object_event_disable);
> +	preempt_enable_notrace();
> +}
> +
> +static struct ftrace_ops trace_ops = {
> +	.func  = trace_object_events_call,
> +	.flags = FTRACE_OPS_FL_SAVE_REGS,
> +};
> +
> +static void
> +trace_object_trigger(struct event_trigger_data *data,
> +		   struct trace_buffer *buffer,  void *rec,
> +		   struct ring_buffer_event *event)
> +{
> +	struct objtrace_trigger_data *obj_data = data->private_data;
> +	struct ftrace_event_field *field;
> +	void *obj = NULL;
> +
> +	field = obj_data->field;
> +	memcpy(&obj, rec + field->offset, sizeof(obj));
> +	set_trace_object(obj);
> +}
> +
> +static void
> +trace_object_trigger_free(struct event_trigger_ops *ops,
> +		   struct event_trigger_data *data)
> +{
> +	if (WARN_ON_ONCE(data->ref <= 0))
> +		return;
> +
> +	data->ref--;
> +	if (!data->ref) {
> +		kfree(data->private_data);
> +		trigger_data_free(data);
> +	}
> +}
> +
> +static void
> +trace_object_count_trigger(struct event_trigger_data *data,
> +			 struct trace_buffer *buffer, void *rec,
> +			 struct ring_buffer_event *event)
> +{
> +	if (!data->count)
> +		return;
> +
> +	if (data->count != -1)
> +		(data->count)--;
> +
> +	trace_object_trigger(data, buffer, rec, event);
> +}
> +
> +static int event_object_trigger_init(struct event_trigger_ops *ops,
> +		       struct event_trigger_data *data)
> +{
> +	data->ref++;
> +	return 0;
> +}
> +
> +static int
> +event_trigger_print(const char *name, struct seq_file *m,
> +		void *data, char *filter_str, void *objtrace_data)
> +{
> +	long count = (long)data;
> +	struct objtrace_trigger_data *obj_data = objtrace_data;
> +
> +	seq_puts(m, name);
> +
> +	seq_printf(m, ":%s", obj_data->objtrace_cmd);
> +	seq_printf(m, ":%s", obj_data->field->name);
> +
> +	if (count == -1)
> +		seq_puts(m, ":unlimited");
> +	else
> +		seq_printf(m, ":count=%ld", count);
> +
> +	if (filter_str)
> +		seq_printf(m, " if %s\n", filter_str);
> +	else
> +		seq_putc(m, '\n');
> +
> +	return 0;
> +}
> +
> +static int
> +trace_object_trigger_print(struct seq_file *m, struct event_trigger_ops *ops,
> +			 struct event_trigger_data *data)
> +{
> +	return event_trigger_print("objtrace", m, (void *)data->count,
> +				   data->filter_str, data->private_data);
> +}
> +
> +static struct event_trigger_ops objecttrace_trigger_ops = {
> +	.trigger		= trace_object_trigger,
> +	.print			= trace_object_trigger_print,
> +	.init			= event_object_trigger_init,
> +	.free			= trace_object_trigger_free,
> +};
> +
> +static struct event_trigger_ops objecttrace_count_trigger_ops = {
> +	.trigger		= trace_object_count_trigger,
> +	.print			= trace_object_trigger_print,
> +	.init			= event_object_trigger_init,
> +	.free			= trace_object_trigger_free,
> +};
> +
> +static struct event_trigger_ops *
> +objecttrace_get_trigger_ops(char *cmd, char *param)
> +{
> +	return param ? &objecttrace_count_trigger_ops : &objecttrace_trigger_ops;
> +}
> +
> +static int register_object_trigger(char *glob,
> +			    struct event_trigger_data *data,
> +			    struct trace_event_file *file)
> +{
> +	struct event_trigger_data *test;
> +	int ret = 0;
> +
> +	lockdep_assert_held(&event_mutex);
> +
> +	list_for_each_entry(test, &file->triggers, list) {
> +		if (test->cmd_ops->trigger_type == data->cmd_ops->trigger_type) {
> +			ret = -EEXIST;

This should just be "return -EEXIST", but you probably just copied this
from register_trigger() which does it like this as well. I'll need to
go fix that.

> +			goto out;
> +		}
> +	}
> +
> +	if (data->ops->init) {
> +		ret = data->ops->init(data->ops, data);
> +		if (ret < 0)

			return ret;

works too, but again, this is just a copy of an existing function that
has the useless goto as well.

> +			goto out;
> +	}
> +
> +	list_add_rcu(&data->list, &file->triggers);
> +	ret++;
> +
> +	update_cond_flag(file);
> +	if (trace_event_trigger_enable_disable(file, 1) < 0) {
> +		list_del_rcu(&data->list);
> +		update_cond_flag(file);
> +		ret--;
> +	}

OK, so this is pretty much identical to "register_tigger()" in
trace_events_trigger() except for the init below. We should probably
just make the register_trigger() global and reuse it here.

> +	init_trace_object();
> +out:
> +	return ret;
> +}
> +
> +static void unregister_object_trigger(char *glob,
> +			       struct event_trigger_data *test,
> +			       struct trace_event_file *file)
> +{
> +	struct event_trigger_data *data;
> +	bool unregistered = false;
> +
> +	lockdep_assert_held(&event_mutex);
> +
> +	list_for_each_entry(data, &file->triggers, list) {
> +		if (data->cmd_ops->trigger_type == test->cmd_ops->trigger_type) {
> +			unregistered = true;
> +			list_del_rcu(&data->list);
> +			trace_event_trigger_enable_disable(file, 0);
> +			update_cond_flag(file);
> +			break;
> +		}
> +	}
> +
> +	if (unregistered) {
> +		if (data->ops->free)
> +			data->ops->free(data->ops, data);
> +		exit_trace_object();
> +	}
> +}
> +
> +static bool field_exist(struct trace_event_file *file,
> +			struct event_command *cmd_ops,
> +			const char *field_name)
> +{
> +	struct event_trigger_data *data;
> +	struct objtrace_trigger_data *obj_data;
> +
> +	lockdep_assert_held(&event_mutex);
> +
> +	list_for_each_entry(data, &file->triggers, list) {
> +		if (data->cmd_ops->trigger_type == cmd_ops->trigger_type) {
> +			obj_data = data->private_data;
> +			if (!strcmp(obj_data->field->name, field_name))
> +				return true;
> +		}
> +	}
> +
> +	return false;
> +}
> +
> +static int
> +event_object_trigger_parse(struct event_command *cmd_ops,
> +		       struct trace_event_file *file,
> +		       char *glob, char *cmd, char *param)
> +{
> +	struct event_trigger_data *trigger_data;
> +	struct event_trigger_ops *trigger_ops;
> +	struct objtrace_trigger_data *obj_data;
> +	struct trace_event_call *call;
> +	struct ftrace_event_field *field;
> +	char *objtrace_cmd;
> +	char *trigger = NULL;
> +	char *arg;
> +	char *number;
> +	int ret;
> +	bool remove = false;
> +
> +	ret = -EINVAL;
> +	if (!param)
> +		goto out;

So this is also copied mostly from event_tigger_parse, and has the
unneeded gotos as well.

We need to consolidate the code a bit, and add helper functions to
share between them, instead of just cut and pasting and tweaking.
That's a maintenance nightmare.

> +
> +	/* separate the trigger from the filter (c:a:n [if filter]) */
> +	trigger = strsep(&param, " \t");
> +	if (!trigger)
> +		goto out;
> +	if (param) {
> +		param = skip_spaces(param);
> +		if (!*param)
> +			param = NULL;
> +	}
> +
> +	objtrace_cmd = strsep(&trigger, ":");
> +	if (!objtrace_cmd || strcmp(objtrace_cmd, "add"))
> +		goto out;
> +
> +	arg = strsep(&trigger, ":");
> +	if (!arg)
> +		goto out;
> +	call = file->event_call;
> +	field = trace_find_event_field(call, arg);
> +	if (!field)
> +		goto out;
> +
> +	if (field->size != sizeof(void *))
> +		goto out;

We need to add error messages to write to the error log when things
like this fail. Users will have no idea what happened.

> +
> +	if (glob[0] == '!')
> +		remove = true;
> +
> +	if (remove && !field_exist(file, cmd_ops, field->name))
> +	goto out;

Bad indentation.

> +	trigger_ops = cmd_ops->get_trigger_ops(cmd, trigger);
> +	ret = -ENOMEM;
> +	obj_data = kzalloc(sizeof(*obj_data), GFP_KERNEL);
> +	if (!obj_data)
> +		goto out;
> +	obj_data->field = field;
> +	snprintf(obj_data->objtrace_cmd, OBJTRACE_CMD_LEN, objtrace_cmd);
> +
> +	trigger_data = kzalloc(sizeof(*trigger_data), GFP_KERNEL);
> +	if (!trigger_data) {
> +		kfree(obj_data);
> +		goto out;
> +	}
> +
> +	trigger_data->count = -1;
> +	trigger_data->ops = trigger_ops;
> +	trigger_data->cmd_ops = cmd_ops;
> +	trigger_data->private_data = obj_data;
> +	INIT_LIST_HEAD(&trigger_data->list);
> +	INIT_LIST_HEAD(&trigger_data->named_list);
> +
> +	if (remove) {
> +		cmd_ops->unreg(glob+1, trigger_data, file);
> +		kfree(obj_data);
> +		kfree(trigger_data);
> +		ret = 0;
> +		goto out;
> +	}
> +
> +	if (trigger) {
> +		number = strsep(&trigger, ":");
> +
> +		ret = -EINVAL;
> +		if (!strlen(number))
> +			goto out_free;
> +
> +		/*
> +		 * We use the callback data field (which is a pointer)
> +		 * as our counter.
> +		 */
> +		ret = kstrtoul(number, 0, &trigger_data->count);
> +		if (ret)
> +			goto out_free;
> +	}
> +
> +	if (!param) /* if param is non-empty, it's supposed to be a filter */
> +		goto out_reg;
> +
> +	if (!cmd_ops->set_filter)
> +		goto out_reg;
> +
> +	ret = cmd_ops->set_filter(param, trigger_data, file);
> +	if (ret < 0)
> +		goto out_free;
> +
> + out_reg:
> +	/* Up the trigger_data count to make sure reg doesn't free it on failure */
> +	event_object_trigger_init(trigger_ops, trigger_data);
> +	ret = cmd_ops->reg(glob, trigger_data, file);
> +	/*
> +	 * The above returns on success the # of functions enabled,
> +	 * but if it didn't find any functions it returns zero.
> +	 * Consider no functions a failure too.
> +	 */
> +	if (!ret) {
> +		cmd_ops->unreg(glob, trigger_data, file);
> +		ret = -ENOENT;
> +	} else if (ret > 0)
> +		ret = 0;
> +
> +	/* Down the counter of trigger_data or free it if not used anymore */
> +	trace_object_trigger_free(trigger_ops, trigger_data);
> + out:
> +	return ret;
> +
> + out_free:
> +	if (cmd_ops->set_filter)
> +		cmd_ops->set_filter(NULL, trigger_data, NULL);
> +	kfree(obj_data);
> +	kfree(trigger_data);
> +	goto out;
> +}
> +
> +static struct event_command trigger_object_cmd = {
> +	.name			= "objtrace",
> +	.trigger_type		= ETT_TRACE_OBJECT,
> +	.flags			= EVENT_CMD_FL_NEEDS_REC,
> +	.parse			= event_object_trigger_parse,
> +	.reg			= register_object_trigger,
> +	.unreg			= unregister_object_trigger,
> +	.get_trigger_ops	= objecttrace_get_trigger_ops,
> +	.set_filter		= set_trigger_filter,
> +};
> +
> +__init int register_trigger_object_cmd(void)
> +{
> +	int ret;
> +
> +	ret = register_event_command(&trigger_object_cmd);
> +	WARN_ON(ret < 0);
> +
> +	return ret;
> +}
> +
> +static int init_trace_object(void)
> +{
> +	int ret;
> +
> +	if (atomic_inc_return(&trace_object_ref) != 1) {
> +		ret = 0;
> +		goto out;
> +	}
> +
> +	event_trace_file.tr = top_trace_array();
> +	if (WARN_ON(!event_trace_file.tr)) {
> +		ret = -1;
> +		atomic_dec(&trace_object_ref);
> +		goto out;
> +	}
> +	ret = register_ftrace_function(&trace_ops);
> +out:
> +	return ret;
> +}
> +
> +static int exit_trace_object(void)
> +{
> +	int ret;
> +
> +	if (WARN_ON_ONCE(atomic_read(&trace_object_ref) <= 0)) {

No reason for the goto out. This shuold be:

	if (WARN_ON_ONCE(atomic_read(&trace_object_ref) <= 0))
		return -1;

> +		ret = -1;
> +		goto out;
> +	}
> +
> +	if (atomic_dec_return(&trace_object_ref) != 0) {

And this:

	if (atomic_dec_return(&trace_object_ref) != 0)
		return 0;

goto's should only be used if there's something to be "undone". But I
see you copied code that had this done incorrectly as well.

So that's my first pass at reviewing the code. It looks promising, but
I think there's still a lot more to do to get it up to par with what is
expected for the kernel.

-- Steve


> +		ret = 0;
> +		goto out;
> +	}
> +
> +	ret = unregister_ftrace_function(&trace_ops);
> +	if (ret) {
> +		pr_err("can't unregister ftrace for trace object\n");
> +		goto out;
> +	}
> +	atomic_set(&num_traced_obj, 0);
> +out:
> +	return ret;
> +}
> diff --git a/kernel/trace/trace_output.c b/kernel/trace/trace_output.c
> index 8aa493d25c73..265428154638 100644
> --- a/kernel/trace/trace_output.c
> +++ b/kernel/trace/trace_output.c
> @@ -1547,6 +1547,45 @@ static struct trace_event trace_func_repeats_event = {
>  	.funcs		= &trace_func_repeats_funcs,
>  };
>  
> +/* TRACE_OBJECT */
> +static enum print_line_t trace_object_print(struct trace_iterator *iter, int flags,
> +					struct trace_event *event)
> +{
> +	struct trace_object_entry *field;
> +	struct trace_seq *s = &iter->seq;
> +
> +	trace_assign_type(field, iter->ent);
> +	print_fn_trace(s, field->ip, field->parent_ip, flags);
> +	trace_seq_printf(s, " object:0x%lx", field->object);
> +	trace_seq_putc(s, '\n');
> +
> +	return trace_handle_return(s);
> +}
> +
> +static enum print_line_t trace_object_raw(struct trace_iterator *iter, int flags,
> +				      struct trace_event *event)
> +{
> +	struct trace_object_entry *field;
> +
> +	trace_assign_type(field, iter->ent);
> +
> +	trace_seq_printf(&iter->seq, "%lx %lx\n",
> +			 field->ip,
> +			 field->parent_ip);
> +
> +	return trace_handle_return(&iter->seq);
> +}
> +
> +static struct trace_event_functions trace_object_funcs = {
> +	.trace		= trace_object_print,
> +	.raw		= trace_object_raw,
> +};
> +
> +static struct trace_event trace_object_event = {
> +	.type		= TRACE_OBJECT,
> +	.funcs		= &trace_object_funcs,
> +};
> +
>  static struct trace_event *events[] __initdata = {
>  	&trace_fn_event,
>  	&trace_ctx_event,
> @@ -1561,6 +1600,7 @@ static struct trace_event *events[] __initdata = {
>  	&trace_timerlat_event,
>  	&trace_raw_data_event,
>  	&trace_func_repeats_event,
> +	&trace_object_event,
>  	NULL
>  };
>  


^ permalink raw reply	[flat|nested] 18+ messages in thread

* Re: [PATCH v9 1/4] trace: Add trace any kernel object
  2022-04-19  2:14   ` Steven Rostedt
@ 2022-04-19  2:26     ` Steven Rostedt
  2022-04-19 16:26     ` Jeff Xie
  1 sibling, 0 replies; 18+ messages in thread
From: Steven Rostedt @ 2022-04-19  2:26 UTC (permalink / raw)
  To: Jeff Xie; +Cc: mhiramat, mingo, zanussi, linux-kernel

On Mon, 18 Apr 2022 22:14:31 -0400
Steven Rostedt <rostedt@goodmis.org> wrote:

> We need to consolidate the code a bit, and add helper functions to
> share between them, instead of just cut and pasting and tweaking.
> That's a maintenance nightmare.

And I forgot that Tom started to do this and I haven't applied his
patches yet. I'll start applying them now.

  https://lore.kernel.org/all/cover.1644010575.git.zanussi@kernel.org/

So when that is done, you can look at rebasing on top of it. Which will
hopefully make the code a bit simpler.

-- Steve

^ permalink raw reply	[flat|nested] 18+ messages in thread

* Re: [PATCH v9 1/4] trace: Add trace any kernel object
  2022-04-19  2:14   ` Steven Rostedt
  2022-04-19  2:26     ` Steven Rostedt
@ 2022-04-19 16:26     ` Jeff Xie
  1 sibling, 0 replies; 18+ messages in thread
From: Jeff Xie @ 2022-04-19 16:26 UTC (permalink / raw)
  To: Steven Rostedt; +Cc: Masami Hiramatsu, mingo, Tom Zanussi, linux-kernel

Hi steve,

Thank you for your patient review.

On Tue, Apr 19, 2022 at 10:14 AM Steven Rostedt <rostedt@goodmis.org> wrote:
>
> On Fri,  4 Feb 2022 11:56:41 +0800
> Jeff Xie <xiehuan09@gmail.com> wrote:
>
> > Introduce objtrace trigger to get the call flow by tracing any kernel
> > object in the function parameter.
> >
> > The objtrace trigger makes a list of the target object address from
> > the given event parameter, and records all kernel function calls
> > which has the object address in its parameter.
> >
> > Syntax:
> >       objtrace:add:obj[:count][if <filter>]
> >
> > Usage:
> >       # echo 'p bio_add_page arg1=$arg1' > kprobe_events
> >       # cd events/kprobes/p_bio_add_page_0
> >       # echo 'objtrace:add:arg1:1 if comm == "cat"' > ./trigger
> >       # cat /test.txt
>
> OK, so patch one only does the object and not the more useful "value"
> field. That comes in patch 2.

Yes, the  "value" is included in patch 2.

>
> >
> > Signed-off-by: Jeff Xie <xiehuan09@gmail.com>
> > ---
> >  include/linux/trace_events.h        |   1 +
> >  kernel/trace/Kconfig                |  10 +
> >  kernel/trace/Makefile               |   1 +
> >  kernel/trace/trace.c                |   3 +
> >  kernel/trace/trace.h                |   8 +
> >  kernel/trace/trace_entries.h        |  17 +
> >  kernel/trace/trace_events_trigger.c |   1 +
> >  kernel/trace/trace_object.c         | 515 ++++++++++++++++++++++++++++
> >  kernel/trace/trace_output.c         |  40 +++
> >  9 files changed, 596 insertions(+)
> >  create mode 100644 kernel/trace/trace_object.c
> >
> > diff --git a/include/linux/trace_events.h b/include/linux/trace_events.h
> > index 70c069aef02c..3ccdbc1d25dd 100644
> > --- a/include/linux/trace_events.h
> > +++ b/include/linux/trace_events.h
> > @@ -685,6 +685,7 @@ enum event_trigger_type {
> >       ETT_EVENT_HIST          = (1 << 4),
> >       ETT_HIST_ENABLE         = (1 << 5),
> >       ETT_EVENT_EPROBE        = (1 << 6),
> > +     ETT_TRACE_OBJECT        = (1 << 7),
> >  };
> >
> >  extern int filter_match_preds(struct event_filter *filter, void *rec);
> > diff --git a/kernel/trace/Kconfig b/kernel/trace/Kconfig
> > index a5eb5e7fd624..c51b7eb1508d 100644
> > --- a/kernel/trace/Kconfig
> > +++ b/kernel/trace/Kconfig
> > @@ -250,6 +250,16 @@ config FUNCTION_PROFILER
> >
> >         If in doubt, say N.
> >
> > +config TRACE_OBJECT
> > +     bool "Trace kernel object in function parameter"
> > +     depends on FUNCTION_TRACER
> > +     depends on HAVE_FUNCTION_ARG_ACCESS_API
> > +     select TRACING
> > +     default y
> > +     help
> > +      You can trace the kernel object in the kernel function parameter.
> > +      The kernel object is dynamically specified via event trigger.
> > +
> >  config STACK_TRACER
> >       bool "Trace max stack"
> >       depends on HAVE_FUNCTION_TRACER
> > diff --git a/kernel/trace/Makefile b/kernel/trace/Makefile
> > index bedc5caceec7..b924b8e55922 100644
> > --- a/kernel/trace/Makefile
> > +++ b/kernel/trace/Makefile
> > @@ -67,6 +67,7 @@ obj-$(CONFIG_FUNCTION_GRAPH_TRACER) += trace_functions_graph.o
> >  obj-$(CONFIG_TRACE_BRANCH_PROFILING) += trace_branch.o
> >  obj-$(CONFIG_BLK_DEV_IO_TRACE) += blktrace.o
> >  obj-$(CONFIG_FUNCTION_GRAPH_TRACER) += fgraph.o
> > +obj-$(CONFIG_TRACE_OBJECT) += trace_object.o
> >  ifeq ($(CONFIG_BLOCK),y)
> >  obj-$(CONFIG_EVENT_TRACING) += blktrace.o
> >  endif
> > diff --git a/kernel/trace/trace.c b/kernel/trace/trace.c
> > index a734d5ae34c8..b4513c2bbd52 100644
> > --- a/kernel/trace/trace.c
> > +++ b/kernel/trace/trace.c
> > @@ -5605,6 +5605,9 @@ static const char readme_msg[] =
> >       "\t            enable_hist:<system>:<event>\n"
> >       "\t            disable_hist:<system>:<event>\n"
> >  #endif
> > +#ifdef CONFIG_TRACE_OBJECT
> > +     "\t            objtrace:add:obj[:count][if <filter>]\n"
> > +#endif
> >  #ifdef CONFIG_STACKTRACE
> >       "\t\t    stacktrace\n"
> >  #endif
> > diff --git a/kernel/trace/trace.h b/kernel/trace/trace.h
> > index 0f5e22238cd2..8b66515a36d5 100644
> > --- a/kernel/trace/trace.h
> > +++ b/kernel/trace/trace.h
> > @@ -54,6 +54,7 @@ enum trace_type {
> >       TRACE_TIMERLAT,
> >       TRACE_RAW_DATA,
> >       TRACE_FUNC_REPEATS,
> > +     TRACE_OBJECT,
> >
> >       __TRACE_LAST_TYPE,
> >  };
> > @@ -472,6 +473,7 @@ extern void __ftrace_bad_type(void);
> >                         TRACE_GRAPH_RET);             \
> >               IF_ASSIGN(var, ent, struct func_repeats_entry,          \
> >                         TRACE_FUNC_REPEATS);                          \
> > +             IF_ASSIGN(var, ent, struct trace_object_entry, TRACE_OBJECT);\
> >               __ftrace_bad_type();                                    \
> >       } while (0)
> >
> > @@ -1537,6 +1539,12 @@ static inline int register_trigger_hist_cmd(void) { return 0; }
> >  static inline int register_trigger_hist_enable_disable_cmds(void) { return 0; }
> >  #endif
> >
> > +#ifdef CONFIG_TRACE_OBJECT
> > +extern int register_trigger_object_cmd(void);
> > +#else
> > +static inline int register_trigger_object_cmd(void) { return 0; }
> > +#endif
> > +
> >  extern int register_trigger_cmds(void);
> >  extern void clear_event_triggers(struct trace_array *tr);
> >
> > diff --git a/kernel/trace/trace_entries.h b/kernel/trace/trace_entries.h
> > index cd41e863b51c..bb120d9498a9 100644
> > --- a/kernel/trace/trace_entries.h
> > +++ b/kernel/trace/trace_entries.h
> > @@ -401,3 +401,20 @@ FTRACE_ENTRY(timerlat, timerlat_entry,
> >                __entry->context,
> >                __entry->timer_latency)
> >  );
> > +
> > +/*
> > + * trace object entry:
> > + */
> > +FTRACE_ENTRY(object, trace_object_entry,
> > +
> > +     TRACE_OBJECT,
> > +
> > +     F_STRUCT(
> > +             __field(        unsigned long,          ip              )
> > +             __field(        unsigned long,          parent_ip       )
> > +             __field(        unsigned long,          object          )
> > +     ),
> > +
> > +     F_printk(" %ps <-- %ps object:%lx\n",
> > +              (void *)__entry->ip, (void *)__entry->parent_ip, __entry->object)
> > +);
> > diff --git a/kernel/trace/trace_events_trigger.c b/kernel/trace/trace_events_trigger.c
> > index d00fee705f9c..c3371a6902af 100644
> > --- a/kernel/trace/trace_events_trigger.c
> > +++ b/kernel/trace/trace_events_trigger.c
> > @@ -2025,6 +2025,7 @@ __init int register_trigger_cmds(void)
> >       register_trigger_enable_disable_cmds();
> >       register_trigger_hist_enable_disable_cmds();
> >       register_trigger_hist_cmd();
> > +     register_trigger_object_cmd();
> >
> >       return 0;
> >  }
> > diff --git a/kernel/trace/trace_object.c b/kernel/trace/trace_object.c
> > new file mode 100644
> > index 000000000000..540e387c613a
> > --- /dev/null
> > +++ b/kernel/trace/trace_object.c
> > @@ -0,0 +1,515 @@
> > +// SPDX-License-Identifier: GPL-2.0
> > +/*
> > + * trace the kernel object in the kernel function parameter
> > + * Copyright (C) 2021 Jeff Xie <xiehuan09@gmail.com>
> > + */
> > +
> > +#define pr_fmt(fmt) "trace_object: " fmt
> > +
> > +#include "trace_output.h"
> > +
> > +#define MAX_TRACED_OBJECT 5
> > +#define OBJTRACE_CMD_LEN  10
>
> The len should at least be a multiple of word size. 16?

Thanks for your suggestion, maybe 16 would be better, I will redefine this size.

>
> > +static DEFINE_PER_CPU(unsigned int, trace_object_event_disable);
> > +static DEFINE_RAW_SPINLOCK(trace_obj_lock);
> > +static struct trace_event_file event_trace_file;
>
> You do not want to use a global "event_trace_file". And the writes
> should be done in the instance (trace_array) that the trigger was added
> to. That is, if someone does:
>
>   # mkdir instances/foo
>   # echo 'objtrace:add:arg1:1 if comm == "cat"' > instances/foo/events/kprobes/p_bio_add_page_0/trigger
>
> Then the tracing should happen in the instance/foo/trace and not at the
> toplevel trace file.

I will rethink how to use the event_trace_file, I ignored the use of
"instance" ;-)

> > +static const int max_args_num = 6;
> > +static atomic_t trace_object_ref;
> > +static atomic_t num_traced_obj;
> > +static int exit_trace_object(void);
> > +static int init_trace_object(void);
> > +
> > +static struct object_instance {
> > +     void *obj;
>
> Either have a pointer to the trace_array that the object is for here,
> or we can have each trace array have a max of 5 objects and add it to
> the trace_array when it is created.

Thanks. I will fix it.

> > +} traced_obj[MAX_TRACED_OBJECT];
> > +
> > +/* objtrace private data */
> > +struct objtrace_trigger_data {
> > +     struct ftrace_event_field *field;
> > +     char objtrace_cmd[OBJTRACE_CMD_LEN];
> > +};
> > +
> > +static bool object_exist(void *obj)
> > +{
> > +     int i, max;
> > +
> > +     max = atomic_read(&num_traced_obj);
> > +     smp_rmb();
> > +     for (i = 0; i < max; i++) {
> > +             if (traced_obj[i].obj == obj)
> > +                     return true;
> > +     }
> > +     return false;
> > +}
> > +
> > +static bool object_empty(void)
> > +{
> > +     return !atomic_read(&num_traced_obj);
> > +}
> > +
> > +static void set_trace_object(void *obj)
> > +{
> > +     unsigned long flags;
> > +
> > +     if (in_nmi())
> > +             return;
> > +
> > +     if (!obj)
> > +             return;
> > +
> > +     if (object_exist(obj))
> > +             return;
> > +
> > +     /* only this place has write operations */
> > +     raw_spin_lock_irqsave(&trace_obj_lock, flags);
> > +     if (atomic_read(&num_traced_obj) == MAX_TRACED_OBJECT) {
> > +             trace_printk("object_pool is full, can't trace object:0x%px\n", obj);
>
> You can't use trace_printk(). That will trigger a nasty message in
> dmesg. You can use trace_array_printk_buf() instead.

Thanks for your guidance.

>
> > +             goto out;
> > +     }
> > +     traced_obj[atomic_read(&num_traced_obj)].obj = obj;
> > +     /* make sure the num_traced_obj update always appears after traced_obj update */
> > +     smp_wmb();
> > +     atomic_inc(&num_traced_obj);
> > +out:
> > +     raw_spin_unlock_irqrestore(&trace_obj_lock, flags);
> > +}
> > +
> > +static void submit_trace_object(unsigned long ip, unsigned long parent_ip,
> > +                              unsigned long object)
> > +{
> > +
> > +     struct trace_buffer *buffer;
> > +     struct ring_buffer_event *event;
> > +     struct trace_object_entry *entry;
> > +     int pc;
> > +
> > +     pc = preempt_count();
> > +     event = trace_event_buffer_lock_reserve(&buffer, &event_trace_file,
> > +                     TRACE_OBJECT, sizeof(*entry), pc);
>
> We should not be using trace_event_buffer_lock_reserver() here, as it
> is expected to be done for trace events, not unique data like this.
> This is more like the function tracer, and use
> trace_buffer_lock_reserve() instead.

Yes, It looks like this trace_buffer_lock_reserve() should indeed be
used in this place.

>
> > +     if (!event)
> > +             return;
> > +     entry   = ring_buffer_event_data(event);
> > +     entry->ip                       = ip;
> > +     entry->parent_ip                = parent_ip;
> > +     entry->object                   = object;
> > +
> > +     event_trigger_unlock_commit(&event_trace_file, buffer, event,
> > +             entry, pc);
> > +}
> > +
> > +static void
> > +trace_object_events_call(unsigned long ip, unsigned long parent_ip,
> > +             struct ftrace_ops *op, struct ftrace_regs *fregs)
> > +{
>
> So the op can hold data, which we should allocate it so that there is
> one ftrace_ops per trace_array (like function tracing does). May need
> callbacks from mkdir and rmdir in trace.c. But the trace_array (tr)
> will hold the buffer you want to write to. Or at the very least, the
> object can hold it.

Thanks, I will  dig into this detail ;-)

>
> > +     struct pt_regs *pt_regs = ftrace_get_regs(fregs);
> > +     unsigned long obj;
> > +     unsigned int disabled;
> > +     int n;
> > +
> > +     preempt_disable_notrace();
> > +
> > +     disabled = this_cpu_inc_return(trace_object_event_disable);
>
> If we make this list per instance, we will only want to disable the
> instance.

Thanks, I will fix this.

> > +     if (disabled != 1)
> > +             goto out;
> > +
> > +     if (object_empty())
> > +             goto out;
> > +
> > +     for (n = 0; n < max_args_num; n++) {
> > +             obj = regs_get_kernel_argument(pt_regs, n);
> > +             if (object_exist((void *)obj))
> > +                     submit_trace_object(ip, parent_ip, obj);
> > +     /* The parameters of a function may match multiple objects */
> > +     }
> > +out:
> > +     this_cpu_dec(trace_object_event_disable);
> > +     preempt_enable_notrace();
> > +}
> > +
> > +static struct ftrace_ops trace_ops = {
> > +     .func  = trace_object_events_call,
> > +     .flags = FTRACE_OPS_FL_SAVE_REGS,
> > +};
> > +
> > +static void
> > +trace_object_trigger(struct event_trigger_data *data,
> > +                struct trace_buffer *buffer,  void *rec,
> > +                struct ring_buffer_event *event)
> > +{
> > +     struct objtrace_trigger_data *obj_data = data->private_data;
> > +     struct ftrace_event_field *field;
> > +     void *obj = NULL;
> > +
> > +     field = obj_data->field;
> > +     memcpy(&obj, rec + field->offset, sizeof(obj));
> > +     set_trace_object(obj);
> > +}
> > +
> > +static void
> > +trace_object_trigger_free(struct event_trigger_ops *ops,
> > +                struct event_trigger_data *data)
> > +{
> > +     if (WARN_ON_ONCE(data->ref <= 0))
> > +             return;
> > +
> > +     data->ref--;
> > +     if (!data->ref) {
> > +             kfree(data->private_data);
> > +             trigger_data_free(data);
> > +     }
> > +}
> > +
> > +static void
> > +trace_object_count_trigger(struct event_trigger_data *data,
> > +                      struct trace_buffer *buffer, void *rec,
> > +                      struct ring_buffer_event *event)
> > +{
> > +     if (!data->count)
> > +             return;
> > +
> > +     if (data->count != -1)
> > +             (data->count)--;
> > +
> > +     trace_object_trigger(data, buffer, rec, event);
> > +}
> > +
> > +static int event_object_trigger_init(struct event_trigger_ops *ops,
> > +                    struct event_trigger_data *data)
> > +{
> > +     data->ref++;
> > +     return 0;
> > +}
> > +
> > +static int
> > +event_trigger_print(const char *name, struct seq_file *m,
> > +             void *data, char *filter_str, void *objtrace_data)
> > +{
> > +     long count = (long)data;
> > +     struct objtrace_trigger_data *obj_data = objtrace_data;
> > +
> > +     seq_puts(m, name);
> > +
> > +     seq_printf(m, ":%s", obj_data->objtrace_cmd);
> > +     seq_printf(m, ":%s", obj_data->field->name);
> > +
> > +     if (count == -1)
> > +             seq_puts(m, ":unlimited");
> > +     else
> > +             seq_printf(m, ":count=%ld", count);
> > +
> > +     if (filter_str)
> > +             seq_printf(m, " if %s\n", filter_str);
> > +     else
> > +             seq_putc(m, '\n');
> > +
> > +     return 0;
> > +}
> > +
> > +static int
> > +trace_object_trigger_print(struct seq_file *m, struct event_trigger_ops *ops,
> > +                      struct event_trigger_data *data)
> > +{
> > +     return event_trigger_print("objtrace", m, (void *)data->count,
> > +                                data->filter_str, data->private_data);
> > +}
> > +
> > +static struct event_trigger_ops objecttrace_trigger_ops = {
> > +     .trigger                = trace_object_trigger,
> > +     .print                  = trace_object_trigger_print,
> > +     .init                   = event_object_trigger_init,
> > +     .free                   = trace_object_trigger_free,
> > +};
> > +
> > +static struct event_trigger_ops objecttrace_count_trigger_ops = {
> > +     .trigger                = trace_object_count_trigger,
> > +     .print                  = trace_object_trigger_print,
> > +     .init                   = event_object_trigger_init,
> > +     .free                   = trace_object_trigger_free,
> > +};
> > +
> > +static struct event_trigger_ops *
> > +objecttrace_get_trigger_ops(char *cmd, char *param)
> > +{
> > +     return param ? &objecttrace_count_trigger_ops : &objecttrace_trigger_ops;
> > +}
> > +
> > +static int register_object_trigger(char *glob,
> > +                         struct event_trigger_data *data,
> > +                         struct trace_event_file *file)
> > +{
> > +     struct event_trigger_data *test;
> > +     int ret = 0;
> > +
> > +     lockdep_assert_held(&event_mutex);
> > +
> > +     list_for_each_entry(test, &file->triggers, list) {
> > +             if (test->cmd_ops->trigger_type == data->cmd_ops->trigger_type) {
> > +                     ret = -EEXIST;
>
> This should just be "return -EEXIST", but you probably just copied this
> from register_trigger() which does it like this as well. I'll need to
> go fix that.
>
> > +                     goto out;
> > +             }
> > +     }
> > +
> > +     if (data->ops->init) {
> > +             ret = data->ops->init(data->ops, data);
> > +             if (ret < 0)
>
>                         return ret;
>
> works too, but again, this is just a copy of an existing function that
> has the useless goto as well.
>
> > +                     goto out;
> > +     }
> > +
> > +     list_add_rcu(&data->list, &file->triggers);
> > +     ret++;
> > +
> > +     update_cond_flag(file);
> > +     if (trace_event_trigger_enable_disable(file, 1) < 0) {
> > +             list_del_rcu(&data->list);
> > +             update_cond_flag(file);
> > +             ret--;
> > +     }
>
> OK, so this is pretty much identical to "register_tigger()" in
> trace_events_trigger() except for the init below. We should probably
> just make the register_trigger() global and reuse it here.
>
> > +     init_trace_object();
> > +out:
> > +     return ret;
> > +}
> > +
> > +static void unregister_object_trigger(char *glob,
> > +                            struct event_trigger_data *test,
> > +                            struct trace_event_file *file)
> > +{
> > +     struct event_trigger_data *data;
> > +     bool unregistered = false;
> > +
> > +     lockdep_assert_held(&event_mutex);
> > +
> > +     list_for_each_entry(data, &file->triggers, list) {
> > +             if (data->cmd_ops->trigger_type == test->cmd_ops->trigger_type) {
> > +                     unregistered = true;
> > +                     list_del_rcu(&data->list);
> > +                     trace_event_trigger_enable_disable(file, 0);
> > +                     update_cond_flag(file);
> > +                     break;
> > +             }
> > +     }
> > +
> > +     if (unregistered) {
> > +             if (data->ops->free)
> > +                     data->ops->free(data->ops, data);
> > +             exit_trace_object();
> > +     }
> > +}
> > +
> > +static bool field_exist(struct trace_event_file *file,
> > +                     struct event_command *cmd_ops,
> > +                     const char *field_name)
> > +{
> > +     struct event_trigger_data *data;
> > +     struct objtrace_trigger_data *obj_data;
> > +
> > +     lockdep_assert_held(&event_mutex);
> > +
> > +     list_for_each_entry(data, &file->triggers, list) {
> > +             if (data->cmd_ops->trigger_type == cmd_ops->trigger_type) {
> > +                     obj_data = data->private_data;
> > +                     if (!strcmp(obj_data->field->name, field_name))
> > +                             return true;
> > +             }
> > +     }
> > +
> > +     return false;
> > +}
> > +
> > +static int
> > +event_object_trigger_parse(struct event_command *cmd_ops,
> > +                    struct trace_event_file *file,
> > +                    char *glob, char *cmd, char *param)
> > +{
> > +     struct event_trigger_data *trigger_data;
> > +     struct event_trigger_ops *trigger_ops;
> > +     struct objtrace_trigger_data *obj_data;
> > +     struct trace_event_call *call;
> > +     struct ftrace_event_field *field;
> > +     char *objtrace_cmd;
> > +     char *trigger = NULL;
> > +     char *arg;
> > +     char *number;
> > +     int ret;
> > +     bool remove = false;
> > +
> > +     ret = -EINVAL;
> > +     if (!param)
> > +             goto out;
>
> So this is also copied mostly from event_tigger_parse, and has the
> unneeded gotos as well.
>
> We need to consolidate the code a bit, and add helper functions to
> share between them, instead of just cut and pasting and tweaking.
> That's a maintenance nightmare.
>
> > +
> > +     /* separate the trigger from the filter (c:a:n [if filter]) */
> > +     trigger = strsep(&param, " \t");
> > +     if (!trigger)
> > +             goto out;
> > +     if (param) {
> > +             param = skip_spaces(param);
> > +             if (!*param)
> > +                     param = NULL;
> > +     }
> > +
> > +     objtrace_cmd = strsep(&trigger, ":");
> > +     if (!objtrace_cmd || strcmp(objtrace_cmd, "add"))
> > +             goto out;
> > +
> > +     arg = strsep(&trigger, ":");
> > +     if (!arg)
> > +             goto out;
> > +     call = file->event_call;
> > +     field = trace_find_event_field(call, arg);
> > +     if (!field)
> > +             goto out;
> > +
> > +     if (field->size != sizeof(void *))
> > +             goto out;
>
> We need to add error messages to write to the error log when things
> like this fail. Users will have no idea what happened.

Yes, I will add some error logs.

>
> > +
> > +     if (glob[0] == '!')
> > +             remove = true;
> > +
> > +     if (remove && !field_exist(file, cmd_ops, field->name))
> > +     goto out;
>
> Bad indentation.
>
> > +     trigger_ops = cmd_ops->get_trigger_ops(cmd, trigger);
> > +     ret = -ENOMEM;
> > +     obj_data = kzalloc(sizeof(*obj_data), GFP_KERNEL);
> > +     if (!obj_data)
> > +             goto out;
> > +     obj_data->field = field;
> > +     snprintf(obj_data->objtrace_cmd, OBJTRACE_CMD_LEN, objtrace_cmd);
> > +
> > +     trigger_data = kzalloc(sizeof(*trigger_data), GFP_KERNEL);
> > +     if (!trigger_data) {
> > +             kfree(obj_data);
> > +             goto out;
> > +     }
> > +
> > +     trigger_data->count = -1;
> > +     trigger_data->ops = trigger_ops;
> > +     trigger_data->cmd_ops = cmd_ops;
> > +     trigger_data->private_data = obj_data;
> > +     INIT_LIST_HEAD(&trigger_data->list);
> > +     INIT_LIST_HEAD(&trigger_data->named_list);
> > +
> > +     if (remove) {
> > +             cmd_ops->unreg(glob+1, trigger_data, file);
> > +             kfree(obj_data);
> > +             kfree(trigger_data);
> > +             ret = 0;
> > +             goto out;
> > +     }
> > +
> > +     if (trigger) {
> > +             number = strsep(&trigger, ":");
> > +
> > +             ret = -EINVAL;
> > +             if (!strlen(number))
> > +                     goto out_free;
> > +
> > +             /*
> > +              * We use the callback data field (which is a pointer)
> > +              * as our counter.
> > +              */
> > +             ret = kstrtoul(number, 0, &trigger_data->count);
> > +             if (ret)
> > +                     goto out_free;
> > +     }
> > +
> > +     if (!param) /* if param is non-empty, it's supposed to be a filter */
> > +             goto out_reg;
> > +
> > +     if (!cmd_ops->set_filter)
> > +             goto out_reg;
> > +
> > +     ret = cmd_ops->set_filter(param, trigger_data, file);
> > +     if (ret < 0)
> > +             goto out_free;
> > +
> > + out_reg:
> > +     /* Up the trigger_data count to make sure reg doesn't free it on failure */
> > +     event_object_trigger_init(trigger_ops, trigger_data);
> > +     ret = cmd_ops->reg(glob, trigger_data, file);
> > +     /*
> > +      * The above returns on success the # of functions enabled,
> > +      * but if it didn't find any functions it returns zero.
> > +      * Consider no functions a failure too.
> > +      */
> > +     if (!ret) {
> > +             cmd_ops->unreg(glob, trigger_data, file);
> > +             ret = -ENOENT;
> > +     } else if (ret > 0)
> > +             ret = 0;
> > +
> > +     /* Down the counter of trigger_data or free it if not used anymore */
> > +     trace_object_trigger_free(trigger_ops, trigger_data);
> > + out:
> > +     return ret;
> > +
> > + out_free:
> > +     if (cmd_ops->set_filter)
> > +             cmd_ops->set_filter(NULL, trigger_data, NULL);
> > +     kfree(obj_data);
> > +     kfree(trigger_data);
> > +     goto out;
> > +}
> > +
> > +static struct event_command trigger_object_cmd = {
> > +     .name                   = "objtrace",
> > +     .trigger_type           = ETT_TRACE_OBJECT,
> > +     .flags                  = EVENT_CMD_FL_NEEDS_REC,
> > +     .parse                  = event_object_trigger_parse,
> > +     .reg                    = register_object_trigger,
> > +     .unreg                  = unregister_object_trigger,
> > +     .get_trigger_ops        = objecttrace_get_trigger_ops,
> > +     .set_filter             = set_trigger_filter,
> > +};
> > +
> > +__init int register_trigger_object_cmd(void)
> > +{
> > +     int ret;
> > +
> > +     ret = register_event_command(&trigger_object_cmd);
> > +     WARN_ON(ret < 0);
> > +
> > +     return ret;
> > +}
> > +
> > +static int init_trace_object(void)
> > +{
> > +     int ret;
> > +
> > +     if (atomic_inc_return(&trace_object_ref) != 1) {
> > +             ret = 0;
> > +             goto out;
> > +     }
> > +
> > +     event_trace_file.tr = top_trace_array();
> > +     if (WARN_ON(!event_trace_file.tr)) {
> > +             ret = -1;
> > +             atomic_dec(&trace_object_ref);
> > +             goto out;
> > +     }
> > +     ret = register_ftrace_function(&trace_ops);
> > +out:
> > +     return ret;
> > +}
> > +
> > +static int exit_trace_object(void)
> > +{
> > +     int ret;
> > +
> > +     if (WARN_ON_ONCE(atomic_read(&trace_object_ref) <= 0)) {
>
> No reason for the goto out. This shuold be:
>
>         if (WARN_ON_ONCE(atomic_read(&trace_object_ref) <= 0))
>                 return -1;
>
> > +             ret = -1;
> > +             goto out;
> > +     }
> > +
> > +     if (atomic_dec_return(&trace_object_ref) != 0) {
>
> And this:
>
>         if (atomic_dec_return(&trace_object_ref) != 0)
>                 return 0;
>
> goto's should only be used if there's something to be "undone". But I
> see you copied code that had this done incorrectly as well.

 Thanks for the detailed explanation, after I use Tom's code, I will
fix these goto.

> So that's my first pass at reviewing the code. It looks promising, but
> I think there's still a lot more to do to get it up to par with what is
> expected for the kernel.

Thank you for your very patient review, the kernel code is indeed very
high quality,
 I will try my best to complete this work well.

> -- Steve
>
>
> > +             ret = 0;
> > +             goto out;
> > +     }
> > +
> > +     ret = unregister_ftrace_function(&trace_ops);
> > +     if (ret) {
> > +             pr_err("can't unregister ftrace for trace object\n");
> > +             goto out;
> > +     }
> > +     atomic_set(&num_traced_obj, 0);
> > +out:
> > +     return ret;
> > +}
> > diff --git a/kernel/trace/trace_output.c b/kernel/trace/trace_output.c
> > index 8aa493d25c73..265428154638 100644
> > --- a/kernel/trace/trace_output.c
> > +++ b/kernel/trace/trace_output.c
> > @@ -1547,6 +1547,45 @@ static struct trace_event trace_func_repeats_event = {
> >       .funcs          = &trace_func_repeats_funcs,
> >  };
> >
> > +/* TRACE_OBJECT */
> > +static enum print_line_t trace_object_print(struct trace_iterator *iter, int flags,
> > +                                     struct trace_event *event)
> > +{
> > +     struct trace_object_entry *field;
> > +     struct trace_seq *s = &iter->seq;
> > +
> > +     trace_assign_type(field, iter->ent);
> > +     print_fn_trace(s, field->ip, field->parent_ip, flags);
> > +     trace_seq_printf(s, " object:0x%lx", field->object);
> > +     trace_seq_putc(s, '\n');
> > +
> > +     return trace_handle_return(s);
> > +}
> > +
> > +static enum print_line_t trace_object_raw(struct trace_iterator *iter, int flags,
> > +                                   struct trace_event *event)
> > +{
> > +     struct trace_object_entry *field;
> > +
> > +     trace_assign_type(field, iter->ent);
> > +
> > +     trace_seq_printf(&iter->seq, "%lx %lx\n",
> > +                      field->ip,
> > +                      field->parent_ip);
> > +
> > +     return trace_handle_return(&iter->seq);
> > +}
> > +
> > +static struct trace_event_functions trace_object_funcs = {
> > +     .trace          = trace_object_print,
> > +     .raw            = trace_object_raw,
> > +};
> > +
> > +static struct trace_event trace_object_event = {
> > +     .type           = TRACE_OBJECT,
> > +     .funcs          = &trace_object_funcs,
> > +};
> > +
> >  static struct trace_event *events[] __initdata = {
> >       &trace_fn_event,
> >       &trace_ctx_event,
> > @@ -1561,6 +1600,7 @@ static struct trace_event *events[] __initdata = {
> >       &trace_timerlat_event,
> >       &trace_raw_data_event,
> >       &trace_func_repeats_event,
> > +     &trace_object_event,
> >       NULL
> >  };
> >
>
---
JeffXie

^ permalink raw reply	[flat|nested] 18+ messages in thread

* Re: [PATCH v9 1/4] trace: Add trace any kernel object
  2022-02-04  3:56 ` [PATCH v9 1/4] trace: Add trace any " Jeff Xie
  2022-04-19  2:14   ` Steven Rostedt
@ 2022-04-27 15:50   ` Jeff Xie
  2022-05-02 17:42     ` Tom Zanussi
  1 sibling, 1 reply; 18+ messages in thread
From: Jeff Xie @ 2022-04-27 15:50 UTC (permalink / raw)
  To: Tom Zanussi; +Cc: Steven Rostedt, Masami Hiramatsu, mingo, linux-kernel

Hi Tom,

I have checked your patches:

https://lore.kernel.org/all/cover.1644010575.git.zanussi@kernel.org/

But I found that I can't use the generic function you implemented,
I don't know if I understand correctly, when you are free, hope you
can help to check it ;-) , thanks.

On Fri, Feb 4, 2022 at 11:57 AM Jeff Xie <xiehuan09@gmail.com> wrote:
>
> Introduce objtrace trigger to get the call flow by tracing any kernel
> object in the function parameter.
>
> The objtrace trigger makes a list of the target object address from
> the given event parameter, and records all kernel function calls
> which has the object address in its parameter.
>
> Syntax:
>         objtrace:add:obj[:count][if <filter>]
>
> Usage:
>         # echo 'p bio_add_page arg1=$arg1' > kprobe_events
>         # cd events/kprobes/p_bio_add_page_0
>         # echo 'objtrace:add:arg1:1 if comm == "cat"' > ./trigger
>         # cat /test.txt
>
> Signed-off-by: Jeff Xie <xiehuan09@gmail.com>
> ---
>  include/linux/trace_events.h        |   1 +
>  kernel/trace/Kconfig                |  10 +
>  kernel/trace/Makefile               |   1 +
>  kernel/trace/trace.c                |   3 +
>  kernel/trace/trace.h                |   8 +
>  kernel/trace/trace_entries.h        |  17 +
>  kernel/trace/trace_events_trigger.c |   1 +
>  kernel/trace/trace_object.c         | 515 ++++++++++++++++++++++++++++
>  kernel/trace/trace_output.c         |  40 +++
>  9 files changed, 596 insertions(+)
>  create mode 100644 kernel/trace/trace_object.c
>
> diff --git a/include/linux/trace_events.h b/include/linux/trace_events.h
> index 70c069aef02c..3ccdbc1d25dd 100644
> --- a/include/linux/trace_events.h
> +++ b/include/linux/trace_events.h
> @@ -685,6 +685,7 @@ enum event_trigger_type {
>         ETT_EVENT_HIST          = (1 << 4),
>         ETT_HIST_ENABLE         = (1 << 5),
>         ETT_EVENT_EPROBE        = (1 << 6),
> +       ETT_TRACE_OBJECT        = (1 << 7),
>  };
>
>  extern int filter_match_preds(struct event_filter *filter, void *rec);
> diff --git a/kernel/trace/Kconfig b/kernel/trace/Kconfig
> index a5eb5e7fd624..c51b7eb1508d 100644
> --- a/kernel/trace/Kconfig
> +++ b/kernel/trace/Kconfig
> @@ -250,6 +250,16 @@ config FUNCTION_PROFILER
>
>           If in doubt, say N.
>
> +config TRACE_OBJECT
> +       bool "Trace kernel object in function parameter"
> +       depends on FUNCTION_TRACER
> +       depends on HAVE_FUNCTION_ARG_ACCESS_API
> +       select TRACING
> +       default y
> +       help
> +        You can trace the kernel object in the kernel function parameter.
> +        The kernel object is dynamically specified via event trigger.
> +
>  config STACK_TRACER
>         bool "Trace max stack"
>         depends on HAVE_FUNCTION_TRACER
> diff --git a/kernel/trace/Makefile b/kernel/trace/Makefile
> index bedc5caceec7..b924b8e55922 100644
> --- a/kernel/trace/Makefile
> +++ b/kernel/trace/Makefile
> @@ -67,6 +67,7 @@ obj-$(CONFIG_FUNCTION_GRAPH_TRACER) += trace_functions_graph.o
>  obj-$(CONFIG_TRACE_BRANCH_PROFILING) += trace_branch.o
>  obj-$(CONFIG_BLK_DEV_IO_TRACE) += blktrace.o
>  obj-$(CONFIG_FUNCTION_GRAPH_TRACER) += fgraph.o
> +obj-$(CONFIG_TRACE_OBJECT) += trace_object.o
>  ifeq ($(CONFIG_BLOCK),y)
>  obj-$(CONFIG_EVENT_TRACING) += blktrace.o
>  endif
> diff --git a/kernel/trace/trace.c b/kernel/trace/trace.c
> index a734d5ae34c8..b4513c2bbd52 100644
> --- a/kernel/trace/trace.c
> +++ b/kernel/trace/trace.c
> @@ -5605,6 +5605,9 @@ static const char readme_msg[] =
>         "\t            enable_hist:<system>:<event>\n"
>         "\t            disable_hist:<system>:<event>\n"
>  #endif
> +#ifdef CONFIG_TRACE_OBJECT
> +       "\t            objtrace:add:obj[:count][if <filter>]\n"
> +#endif
>  #ifdef CONFIG_STACKTRACE
>         "\t\t    stacktrace\n"
>  #endif
> diff --git a/kernel/trace/trace.h b/kernel/trace/trace.h
> index 0f5e22238cd2..8b66515a36d5 100644
> --- a/kernel/trace/trace.h
> +++ b/kernel/trace/trace.h
> @@ -54,6 +54,7 @@ enum trace_type {
>         TRACE_TIMERLAT,
>         TRACE_RAW_DATA,
>         TRACE_FUNC_REPEATS,
> +       TRACE_OBJECT,
>
>         __TRACE_LAST_TYPE,
>  };
> @@ -472,6 +473,7 @@ extern void __ftrace_bad_type(void);
>                           TRACE_GRAPH_RET);             \
>                 IF_ASSIGN(var, ent, struct func_repeats_entry,          \
>                           TRACE_FUNC_REPEATS);                          \
> +               IF_ASSIGN(var, ent, struct trace_object_entry, TRACE_OBJECT);\
>                 __ftrace_bad_type();                                    \
>         } while (0)
>
> @@ -1537,6 +1539,12 @@ static inline int register_trigger_hist_cmd(void) { return 0; }
>  static inline int register_trigger_hist_enable_disable_cmds(void) { return 0; }
>  #endif
>
> +#ifdef CONFIG_TRACE_OBJECT
> +extern int register_trigger_object_cmd(void);
> +#else
> +static inline int register_trigger_object_cmd(void) { return 0; }
> +#endif
> +
>  extern int register_trigger_cmds(void);
>  extern void clear_event_triggers(struct trace_array *tr);
>
> diff --git a/kernel/trace/trace_entries.h b/kernel/trace/trace_entries.h
> index cd41e863b51c..bb120d9498a9 100644
> --- a/kernel/trace/trace_entries.h
> +++ b/kernel/trace/trace_entries.h
> @@ -401,3 +401,20 @@ FTRACE_ENTRY(timerlat, timerlat_entry,
>                  __entry->context,
>                  __entry->timer_latency)
>  );
> +
> +/*
> + * trace object entry:
> + */
> +FTRACE_ENTRY(object, trace_object_entry,
> +
> +       TRACE_OBJECT,
> +
> +       F_STRUCT(
> +               __field(        unsigned long,          ip              )
> +               __field(        unsigned long,          parent_ip       )
> +               __field(        unsigned long,          object          )
> +       ),
> +
> +       F_printk(" %ps <-- %ps object:%lx\n",
> +                (void *)__entry->ip, (void *)__entry->parent_ip, __entry->object)
> +);
> diff --git a/kernel/trace/trace_events_trigger.c b/kernel/trace/trace_events_trigger.c
> index d00fee705f9c..c3371a6902af 100644
> --- a/kernel/trace/trace_events_trigger.c
> +++ b/kernel/trace/trace_events_trigger.c
> @@ -2025,6 +2025,7 @@ __init int register_trigger_cmds(void)
>         register_trigger_enable_disable_cmds();
>         register_trigger_hist_enable_disable_cmds();
>         register_trigger_hist_cmd();
> +       register_trigger_object_cmd();
>
>         return 0;
>  }
> diff --git a/kernel/trace/trace_object.c b/kernel/trace/trace_object.c
> new file mode 100644
> index 000000000000..540e387c613a
> --- /dev/null
> +++ b/kernel/trace/trace_object.c
> @@ -0,0 +1,515 @@
> +// SPDX-License-Identifier: GPL-2.0
> +/*
> + * trace the kernel object in the kernel function parameter
> + * Copyright (C) 2021 Jeff Xie <xiehuan09@gmail.com>
> + */
> +
> +#define pr_fmt(fmt) "trace_object: " fmt
> +
> +#include "trace_output.h"
> +
> +#define MAX_TRACED_OBJECT 5
> +#define OBJTRACE_CMD_LEN  10
> +static DEFINE_PER_CPU(unsigned int, trace_object_event_disable);
> +static DEFINE_RAW_SPINLOCK(trace_obj_lock);
> +static struct trace_event_file event_trace_file;
> +static const int max_args_num = 6;
> +static atomic_t trace_object_ref;
> +static atomic_t num_traced_obj;
> +static int exit_trace_object(void);
> +static int init_trace_object(void);
> +
> +static struct object_instance {
> +       void *obj;
> +} traced_obj[MAX_TRACED_OBJECT];
> +
> +/* objtrace private data */
> +struct objtrace_trigger_data {
> +       struct ftrace_event_field *field;
> +       char objtrace_cmd[OBJTRACE_CMD_LEN];
> +};
> +
> +static bool object_exist(void *obj)
> +{
> +       int i, max;
> +
> +       max = atomic_read(&num_traced_obj);
> +       smp_rmb();
> +       for (i = 0; i < max; i++) {
> +               if (traced_obj[i].obj == obj)
> +                       return true;
> +       }
> +       return false;
> +}
> +
> +static bool object_empty(void)
> +{
> +       return !atomic_read(&num_traced_obj);
> +}
> +
> +static void set_trace_object(void *obj)
> +{
> +       unsigned long flags;
> +
> +       if (in_nmi())
> +               return;
> +
> +       if (!obj)
> +               return;
> +
> +       if (object_exist(obj))
> +               return;
> +
> +       /* only this place has write operations */
> +       raw_spin_lock_irqsave(&trace_obj_lock, flags);
> +       if (atomic_read(&num_traced_obj) == MAX_TRACED_OBJECT) {
> +               trace_printk("object_pool is full, can't trace object:0x%px\n", obj);
> +               goto out;
> +       }
> +       traced_obj[atomic_read(&num_traced_obj)].obj = obj;
> +       /* make sure the num_traced_obj update always appears after traced_obj update */
> +       smp_wmb();
> +       atomic_inc(&num_traced_obj);
> +out:
> +       raw_spin_unlock_irqrestore(&trace_obj_lock, flags);
> +}
> +
> +static void submit_trace_object(unsigned long ip, unsigned long parent_ip,
> +                                unsigned long object)
> +{
> +
> +       struct trace_buffer *buffer;
> +       struct ring_buffer_event *event;
> +       struct trace_object_entry *entry;
> +       int pc;
> +
> +       pc = preempt_count();
> +       event = trace_event_buffer_lock_reserve(&buffer, &event_trace_file,
> +                       TRACE_OBJECT, sizeof(*entry), pc);
> +       if (!event)
> +               return;
> +       entry   = ring_buffer_event_data(event);
> +       entry->ip                       = ip;
> +       entry->parent_ip                = parent_ip;
> +       entry->object                   = object;
> +
> +       event_trigger_unlock_commit(&event_trace_file, buffer, event,
> +               entry, pc);
> +}
> +
> +static void
> +trace_object_events_call(unsigned long ip, unsigned long parent_ip,
> +               struct ftrace_ops *op, struct ftrace_regs *fregs)
> +{
> +       struct pt_regs *pt_regs = ftrace_get_regs(fregs);
> +       unsigned long obj;
> +       unsigned int disabled;
> +       int n;
> +
> +       preempt_disable_notrace();
> +
> +       disabled = this_cpu_inc_return(trace_object_event_disable);
> +       if (disabled != 1)
> +               goto out;
> +
> +       if (object_empty())
> +               goto out;
> +
> +       for (n = 0; n < max_args_num; n++) {
> +               obj = regs_get_kernel_argument(pt_regs, n);
> +               if (object_exist((void *)obj))
> +                       submit_trace_object(ip, parent_ip, obj);
> +       /* The parameters of a function may match multiple objects */
> +       }
> +out:
> +       this_cpu_dec(trace_object_event_disable);
> +       preempt_enable_notrace();
> +}
> +
> +static struct ftrace_ops trace_ops = {
> +       .func  = trace_object_events_call,
> +       .flags = FTRACE_OPS_FL_SAVE_REGS,
> +};
> +
> +static void
> +trace_object_trigger(struct event_trigger_data *data,
> +                  struct trace_buffer *buffer,  void *rec,
> +                  struct ring_buffer_event *event)
> +{
> +       struct objtrace_trigger_data *obj_data = data->private_data;
> +       struct ftrace_event_field *field;
> +       void *obj = NULL;
> +
> +       field = obj_data->field;
> +       memcpy(&obj, rec + field->offset, sizeof(obj));
> +       set_trace_object(obj);
> +}
> +
> +static void
> +trace_object_trigger_free(struct event_trigger_ops *ops,
> +                  struct event_trigger_data *data)
> +{
> +       if (WARN_ON_ONCE(data->ref <= 0))
> +               return;
> +
> +       data->ref--;
> +       if (!data->ref) {
> +               kfree(data->private_data);
> +               trigger_data_free(data);
> +       }
> +}
> +
> +static void
> +trace_object_count_trigger(struct event_trigger_data *data,
> +                        struct trace_buffer *buffer, void *rec,
> +                        struct ring_buffer_event *event)
> +{
> +       if (!data->count)
> +               return;
> +
> +       if (data->count != -1)
> +               (data->count)--;
> +
> +       trace_object_trigger(data, buffer, rec, event);
> +}
> +
> +static int event_object_trigger_init(struct event_trigger_ops *ops,
> +                      struct event_trigger_data *data)
> +{
> +       data->ref++;
> +       return 0;
> +}
> +
> +static int
> +event_trigger_print(const char *name, struct seq_file *m,
> +               void *data, char *filter_str, void *objtrace_data)
> +{
> +       long count = (long)data;
> +       struct objtrace_trigger_data *obj_data = objtrace_data;
> +
> +       seq_puts(m, name);
> +
> +       seq_printf(m, ":%s", obj_data->objtrace_cmd);
> +       seq_printf(m, ":%s", obj_data->field->name);
> +
> +       if (count == -1)
> +               seq_puts(m, ":unlimited");
> +       else
> +               seq_printf(m, ":count=%ld", count);
> +
> +       if (filter_str)
> +               seq_printf(m, " if %s\n", filter_str);
> +       else
> +               seq_putc(m, '\n');
> +
> +       return 0;
> +}
> +
> +static int
> +trace_object_trigger_print(struct seq_file *m, struct event_trigger_ops *ops,
> +                        struct event_trigger_data *data)
> +{
> +       return event_trigger_print("objtrace", m, (void *)data->count,
> +                                  data->filter_str, data->private_data);
> +}
> +
> +static struct event_trigger_ops objecttrace_trigger_ops = {
> +       .trigger                = trace_object_trigger,
> +       .print                  = trace_object_trigger_print,
> +       .init                   = event_object_trigger_init,
> +       .free                   = trace_object_trigger_free,
> +};
> +
> +static struct event_trigger_ops objecttrace_count_trigger_ops = {
> +       .trigger                = trace_object_count_trigger,
> +       .print                  = trace_object_trigger_print,
> +       .init                   = event_object_trigger_init,
> +       .free                   = trace_object_trigger_free,
> +};
> +
> +static struct event_trigger_ops *
> +objecttrace_get_trigger_ops(char *cmd, char *param)
> +{
> +       return param ? &objecttrace_count_trigger_ops : &objecttrace_trigger_ops;
> +}
> +
> +static int register_object_trigger(char *glob,
> +                           struct event_trigger_data *data,
> +                           struct trace_event_file *file)
> +{
> +       struct event_trigger_data *test;
> +       int ret = 0;
> +
> +       lockdep_assert_held(&event_mutex);
> +
> +       list_for_each_entry(test, &file->triggers, list) {
> +               if (test->cmd_ops->trigger_type == data->cmd_ops->trigger_type) {
> +                       ret = -EEXIST;
> +                       goto out;
> +               }
> +       }
> +
> +       if (data->ops->init) {
> +               ret = data->ops->init(data->ops, data);
> +               if (ret < 0)
> +                       goto out;
> +       }
> +
> +       list_add_rcu(&data->list, &file->triggers);
> +       ret++;
> +
> +       update_cond_flag(file);
> +       if (trace_event_trigger_enable_disable(file, 1) < 0) {
> +               list_del_rcu(&data->list);
> +               update_cond_flag(file);
> +               ret--;
> +       }
> +       init_trace_object();
> +out:
> +       return ret;
> +}

Can't replace the register_object_trigger() with register_trigger(),
as there will be no place to invoke the init_trace_object(), maybe It
will be better to add a callback in struct event_command,
for example, cmd_ops->init()

> +static void unregister_object_trigger(char *glob,
> +                              struct event_trigger_data *test,
> +                              struct trace_event_file *file)
> +{
> +       struct event_trigger_data *data;
> +       bool unregistered = false;
> +
> +       lockdep_assert_held(&event_mutex);
> +
> +       list_for_each_entry(data, &file->triggers, list) {
> +               if (data->cmd_ops->trigger_type == test->cmd_ops->trigger_type) {
> +                       unregistered = true;
> +                       list_del_rcu(&data->list);
> +                       trace_event_trigger_enable_disable(file, 0);
> +                       update_cond_flag(file);
> +                       break;
> +               }
> +       }
> +
> +       if (unregistered) {
> +               if (data->ops->free)
> +                       data->ops->free(data->ops, data);
> +               exit_trace_object();
> +       }
> +}
> +

Can't replace the unregister_object_trigger() with unregister_trigger(),
as there will be no place to invoke the exit_trace_object().Maybe need
to add a callback, for example cmd_ops->free().

> +static bool field_exist(struct trace_event_file *file,
> +                       struct event_command *cmd_ops,
> +                       const char *field_name)
> +{
> +       struct event_trigger_data *data;
> +       struct objtrace_trigger_data *obj_data;
> +
> +       lockdep_assert_held(&event_mutex);
> +
> +       list_for_each_entry(data, &file->triggers, list) {
> +               if (data->cmd_ops->trigger_type == cmd_ops->trigger_type) {
> +                       obj_data = data->private_data;
> +                       if (!strcmp(obj_data->field->name, field_name))
> +                               return true;
> +               }
> +       }
> +
> +       return false;
> +}
> +
> +static int
> +event_object_trigger_parse(struct event_command *cmd_ops,
> +                      struct trace_event_file *file,
> +                      char *glob, char *cmd, char *param)
> +{
> +       struct event_trigger_data *trigger_data;
> +       struct event_trigger_ops *trigger_ops;
> +       struct objtrace_trigger_data *obj_data;
> +       struct trace_event_call *call;
> +       struct ftrace_event_field *field;
> +       char *objtrace_cmd;
> +       char *trigger = NULL;
> +       char *arg;
> +       char *number;
> +       int ret;
> +       bool remove = false;
> +
> +       ret = -EINVAL;
> +       if (!param)
> +               goto out;
> +
> +       /* separate the trigger from the filter (c:a:n [if filter]) */
> +       trigger = strsep(&param, " \t");
> +       if (!trigger)
> +               goto out;
> +       if (param) {
> +               param = skip_spaces(param);
> +               if (!*param)
> +                       param = NULL;
> +       }
> +
> +       objtrace_cmd = strsep(&trigger, ":");
> +       if (!objtrace_cmd || strcmp(objtrace_cmd, "add"))
> +               goto out;
> +
> +       arg = strsep(&trigger, ":");
> +       if (!arg)
> +               goto out;
> +       call = file->event_call;
> +       field = trace_find_event_field(call, arg);
> +       if (!field)
> +               goto out;
> +
> +       if (field->size != sizeof(void *))
> +               goto out;
> +
> +       if (glob[0] == '!')
> +               remove = true;
> +
> +       if (remove && !field_exist(file, cmd_ops, field->name))
> +       goto out;
> +       trigger_ops = cmd_ops->get_trigger_ops(cmd, trigger);
> +       ret = -ENOMEM;
> +       obj_data = kzalloc(sizeof(*obj_data), GFP_KERNEL);
> +       if (!obj_data)
> +               goto out;
> +       obj_data->field = field;
> +       snprintf(obj_data->objtrace_cmd, OBJTRACE_CMD_LEN, objtrace_cmd);
> +
> +       trigger_data = kzalloc(sizeof(*trigger_data), GFP_KERNEL);
> +       if (!trigger_data) {
> +               kfree(obj_data);
> +               goto out;
> +       }
> +
> +       trigger_data->count = -1;
> +       trigger_data->ops = trigger_ops;
> +       trigger_data->cmd_ops = cmd_ops;
> +       trigger_data->private_data = obj_data;
> +       INIT_LIST_HEAD(&trigger_data->list);
> +       INIT_LIST_HEAD(&trigger_data->named_list);
> +
> +       if (remove) {
> +               cmd_ops->unreg(glob+1, trigger_data, file);
> +               kfree(obj_data);
> +               kfree(trigger_data);
> +               ret = 0;
> +               goto out;
> +       }
> +
> +       if (trigger) {
> +               number = strsep(&trigger, ":");
> +
> +               ret = -EINVAL;
> +               if (!strlen(number))
> +                       goto out_free;
> +
> +               /*
> +                * We use the callback data field (which is a pointer)
> +                * as our counter.
> +                */
> +               ret = kstrtoul(number, 0, &trigger_data->count);
> +               if (ret)
> +                       goto out_free;
> +       }
> +
> +       if (!param) /* if param is non-empty, it's supposed to be a filter */
> +               goto out_reg;
> +
> +       if (!cmd_ops->set_filter)
> +               goto out_reg;
> +
> +       ret = cmd_ops->set_filter(param, trigger_data, file);
> +       if (ret < 0)
> +               goto out_free;
> +
> + out_reg:
> +       /* Up the trigger_data count to make sure reg doesn't free it on failure */
> +       event_object_trigger_init(trigger_ops, trigger_data);
> +       ret = cmd_ops->reg(glob, trigger_data, file);
> +       /*
> +        * The above returns on success the # of functions enabled,
> +        * but if it didn't find any functions it returns zero.
> +        * Consider no functions a failure too.
> +        */
> +       if (!ret) {
> +               cmd_ops->unreg(glob, trigger_data, file);
> +               ret = -ENOENT;
> +       } else if (ret > 0)
> +               ret = 0;
> +
> +       /* Down the counter of trigger_data or free it if not used anymore */
> +       trace_object_trigger_free(trigger_ops, trigger_data);
> + out:
> +       return ret;
> +
> + out_free:
> +       if (cmd_ops->set_filter)
> +               cmd_ops->set_filter(NULL, trigger_data, NULL);
> +       kfree(obj_data);
> +       kfree(trigger_data);
> +       goto out;
> +}

Can't replace the event_object_trigger_parse() with event_trigger_parse().
Because the event_object_trigger_parse() needs to parse more fields,
for example need to parse objtrace_cmd("add" etc.,)
and need to use other special private_data. Maybe It will have to use
the event_object_trigger_parse().

> +static struct event_command trigger_object_cmd = {
> +       .name                   = "objtrace",
> +       .trigger_type           = ETT_TRACE_OBJECT,
> +       .flags                  = EVENT_CMD_FL_NEEDS_REC,
> +       .parse                  = event_object_trigger_parse,
> +       .reg                    = register_object_trigger,
> +       .unreg                  = unregister_object_trigger,
> +       .get_trigger_ops        = objecttrace_get_trigger_ops,
> +       .set_filter             = set_trigger_filter,
> +};
> +
> +__init int register_trigger_object_cmd(void)
> +{
> +       int ret;
> +
> +       ret = register_event_command(&trigger_object_cmd);
> +       WARN_ON(ret < 0);
> +
> +       return ret;
> +}
> +
> +static int init_trace_object(void)
> +{
> +       int ret;
> +
> +       if (atomic_inc_return(&trace_object_ref) != 1) {
> +               ret = 0;
> +               goto out;
> +       }
> +
> +       event_trace_file.tr = top_trace_array();
> +       if (WARN_ON(!event_trace_file.tr)) {
> +               ret = -1;
> +               atomic_dec(&trace_object_ref);
> +               goto out;
> +       }
> +       ret = register_ftrace_function(&trace_ops);
> +out:
> +       return ret;
> +}
> +
> +static int exit_trace_object(void)
> +{
> +       int ret;
> +
> +       if (WARN_ON_ONCE(atomic_read(&trace_object_ref) <= 0)) {
> +               ret = -1;
> +               goto out;
> +       }
> +
> +       if (atomic_dec_return(&trace_object_ref) != 0) {
> +               ret = 0;
> +               goto out;
> +       }
> +
> +       ret = unregister_ftrace_function(&trace_ops);
> +       if (ret) {
> +               pr_err("can't unregister ftrace for trace object\n");
> +               goto out;
> +       }
> +       atomic_set(&num_traced_obj, 0);
> +out:
> +       return ret;
> +}
> diff --git a/kernel/trace/trace_output.c b/kernel/trace/trace_output.c
> index 8aa493d25c73..265428154638 100644
> --- a/kernel/trace/trace_output.c
> +++ b/kernel/trace/trace_output.c
> @@ -1547,6 +1547,45 @@ static struct trace_event trace_func_repeats_event = {
>         .funcs          = &trace_func_repeats_funcs,
>  };
>
> +/* TRACE_OBJECT */
> +static enum print_line_t trace_object_print(struct trace_iterator *iter, int flags,
> +                                       struct trace_event *event)
> +{
> +       struct trace_object_entry *field;
> +       struct trace_seq *s = &iter->seq;
> +
> +       trace_assign_type(field, iter->ent);
> +       print_fn_trace(s, field->ip, field->parent_ip, flags);
> +       trace_seq_printf(s, " object:0x%lx", field->object);
> +       trace_seq_putc(s, '\n');
> +
> +       return trace_handle_return(s);
> +}
> +
> +static enum print_line_t trace_object_raw(struct trace_iterator *iter, int flags,
> +                                     struct trace_event *event)
> +{
> +       struct trace_object_entry *field;
> +
> +       trace_assign_type(field, iter->ent);
> +
> +       trace_seq_printf(&iter->seq, "%lx %lx\n",
> +                        field->ip,
> +                        field->parent_ip);
> +
> +       return trace_handle_return(&iter->seq);
> +}
> +
> +static struct trace_event_functions trace_object_funcs = {
> +       .trace          = trace_object_print,
> +       .raw            = trace_object_raw,
> +};
> +
> +static struct trace_event trace_object_event = {
> +       .type           = TRACE_OBJECT,
> +       .funcs          = &trace_object_funcs,
> +};
> +
>  static struct trace_event *events[] __initdata = {
>         &trace_fn_event,
>         &trace_ctx_event,
> @@ -1561,6 +1600,7 @@ static struct trace_event *events[] __initdata = {
>         &trace_timerlat_event,
>         &trace_raw_data_event,
>         &trace_func_repeats_event,
> +       &trace_object_event,
>         NULL
>  };
>
> --
> 2.25.1
>

Thanks,
JeffXie

^ permalink raw reply	[flat|nested] 18+ messages in thread

* Re: [PATCH v9 1/4] trace: Add trace any kernel object
  2022-04-27 15:50   ` Jeff Xie
@ 2022-05-02 17:42     ` Tom Zanussi
  2022-05-04  7:42       ` Jeff Xie
  0 siblings, 1 reply; 18+ messages in thread
From: Tom Zanussi @ 2022-05-02 17:42 UTC (permalink / raw)
  To: Jeff Xie; +Cc: Steven Rostedt, Masami Hiramatsu, mingo, linux-kernel

Hi Jeff,

On Wed, 2022-04-27 at 23:50 +0800, Jeff Xie wrote:
> Hi Tom,
> 
> I have checked your patches:
> 
> https://lore.kernel.org/all/cover.1644010575.git.zanussi@kernel.org/
> 
> But I found that I can't use the generic function you implemented,
> I don't know if I understand correctly, when you are free, hope you
> can help to check it ;-) , thanks.
> 
> On Fri, Feb 4, 2022 at 11:57 AM Jeff Xie <xiehuan09@gmail.com> wrote:
> > 
> > Introduce objtrace trigger to get the call flow by tracing any
> > kernel
> > object in the function parameter.
> > 
> > The objtrace trigger makes a list of the target object address from
> > the given event parameter, and records all kernel function calls
> > which has the object address in its parameter.
> > 
> > Syntax:
> >         objtrace:add:obj[:count][if <filter>]
> > 
> > Usage:
> >         # echo 'p bio_add_page arg1=$arg1' > kprobe_events
> >         # cd events/kprobes/p_bio_add_page_0
> >         # echo 'objtrace:add:arg1:1 if comm == "cat"' > ./trigger
> >         # cat /test.txt
> > 
> > Signed-off-by: Jeff Xie <xiehuan09@gmail.com>
> > ---
> >  include/linux/trace_events.h        |   1 +
> >  kernel/trace/Kconfig                |  10 +
> >  kernel/trace/Makefile               |   1 +
> >  kernel/trace/trace.c                |   3 +
> >  kernel/trace/trace.h                |   8 +
> >  kernel/trace/trace_entries.h        |  17 +
> >  kernel/trace/trace_events_trigger.c |   1 +
> >  kernel/trace/trace_object.c         | 515
> > ++++++++++++++++++++++++++++
> >  kernel/trace/trace_output.c         |  40 +++
> >  9 files changed, 596 insertions(+)
> >  create mode 100644 kernel/trace/trace_object.c
> > 
> > diff --git a/include/linux/trace_events.h
> > b/include/linux/trace_events.h
> > index 70c069aef02c..3ccdbc1d25dd 100644
> > --- a/include/linux/trace_events.h
> > +++ b/include/linux/trace_events.h
> > @@ -685,6 +685,7 @@ enum event_trigger_type {
> >         ETT_EVENT_HIST          = (1 << 4),
> >         ETT_HIST_ENABLE         = (1 << 5),
> >         ETT_EVENT_EPROBE        = (1 << 6),
> > +       ETT_TRACE_OBJECT        = (1 << 7),
> >  };
> > 
> >  extern int filter_match_preds(struct event_filter *filter, void
> > *rec);
> > diff --git a/kernel/trace/Kconfig b/kernel/trace/Kconfig
> > index a5eb5e7fd624..c51b7eb1508d 100644
> > --- a/kernel/trace/Kconfig
> > +++ b/kernel/trace/Kconfig
> > @@ -250,6 +250,16 @@ config FUNCTION_PROFILER
> > 
> >           If in doubt, say N.
> > 
> > +config TRACE_OBJECT
> > +       bool "Trace kernel object in function parameter"
> > +       depends on FUNCTION_TRACER
> > +       depends on HAVE_FUNCTION_ARG_ACCESS_API
> > +       select TRACING
> > +       default y
> > +       help
> > +        You can trace the kernel object in the kernel function
> > parameter.
> > +        The kernel object is dynamically specified via event
> > trigger.
> > +
> >  config STACK_TRACER
> >         bool "Trace max stack"
> >         depends on HAVE_FUNCTION_TRACER
> > diff --git a/kernel/trace/Makefile b/kernel/trace/Makefile
> > index bedc5caceec7..b924b8e55922 100644
> > --- a/kernel/trace/Makefile
> > +++ b/kernel/trace/Makefile
> > @@ -67,6 +67,7 @@ obj-$(CONFIG_FUNCTION_GRAPH_TRACER) +=
> > trace_functions_graph.o
> >  obj-$(CONFIG_TRACE_BRANCH_PROFILING) += trace_branch.o
> >  obj-$(CONFIG_BLK_DEV_IO_TRACE) += blktrace.o
> >  obj-$(CONFIG_FUNCTION_GRAPH_TRACER) += fgraph.o
> > +obj-$(CONFIG_TRACE_OBJECT) += trace_object.o
> >  ifeq ($(CONFIG_BLOCK),y)
> >  obj-$(CONFIG_EVENT_TRACING) += blktrace.o
> >  endif
> > diff --git a/kernel/trace/trace.c b/kernel/trace/trace.c
> > index a734d5ae34c8..b4513c2bbd52 100644
> > --- a/kernel/trace/trace.c
> > +++ b/kernel/trace/trace.c
> > @@ -5605,6 +5605,9 @@ static const char readme_msg[] =
> >         "\t            enable_hist:<system>:<event>\n"
> >         "\t            disable_hist:<system>:<event>\n"
> >  #endif
> > +#ifdef CONFIG_TRACE_OBJECT
> > +       "\t            objtrace:add:obj[:count][if <filter>]\n"
> > +#endif
> >  #ifdef CONFIG_STACKTRACE
> >         "\t\t    stacktrace\n"
> >  #endif
> > diff --git a/kernel/trace/trace.h b/kernel/trace/trace.h
> > index 0f5e22238cd2..8b66515a36d5 100644
> > --- a/kernel/trace/trace.h
> > +++ b/kernel/trace/trace.h
> > @@ -54,6 +54,7 @@ enum trace_type {
> >         TRACE_TIMERLAT,
> >         TRACE_RAW_DATA,
> >         TRACE_FUNC_REPEATS,
> > +       TRACE_OBJECT,
> > 
> >         __TRACE_LAST_TYPE,
> >  };
> > @@ -472,6 +473,7 @@ extern void __ftrace_bad_type(void);
> >                           TRACE_GRAPH_RET);             \
> >                 IF_ASSIGN(var, ent, struct
> > func_repeats_entry,          \
> >                           TRACE_FUNC_REPEATS);                     
> >      \
> > +               IF_ASSIGN(var, ent, struct trace_object_entry,
> > TRACE_OBJECT);\
> >                 __ftrace_bad_type();                               
> >      \
> >         } while (0)
> > 
> > @@ -1537,6 +1539,12 @@ static inline int
> > register_trigger_hist_cmd(void) { return 0; }
> >  static inline int register_trigger_hist_enable_disable_cmds(void)
> > { return 0; }
> >  #endif
> > 
> > +#ifdef CONFIG_TRACE_OBJECT
> > +extern int register_trigger_object_cmd(void);
> > +#else
> > +static inline int register_trigger_object_cmd(void) { return 0; }
> > +#endif
> > +
> >  extern int register_trigger_cmds(void);
> >  extern void clear_event_triggers(struct trace_array *tr);
> > 
> > diff --git a/kernel/trace/trace_entries.h
> > b/kernel/trace/trace_entries.h
> > index cd41e863b51c..bb120d9498a9 100644
> > --- a/kernel/trace/trace_entries.h
> > +++ b/kernel/trace/trace_entries.h
> > @@ -401,3 +401,20 @@ FTRACE_ENTRY(timerlat, timerlat_entry,
> >                  __entry->context,
> >                  __entry->timer_latency)
> >  );
> > +
> > +/*
> > + * trace object entry:
> > + */
> > +FTRACE_ENTRY(object, trace_object_entry,
> > +
> > +       TRACE_OBJECT,
> > +
> > +       F_STRUCT(
> > +               __field(        unsigned
> > long,          ip              )
> > +               __field(        unsigned
> > long,          parent_ip       )
> > +               __field(        unsigned
> > long,          object          )
> > +       ),
> > +
> > +       F_printk(" %ps <-- %ps object:%lx\n",
> > +                (void *)__entry->ip, (void *)__entry->parent_ip,
> > __entry->object)
> > +);
> > diff --git a/kernel/trace/trace_events_trigger.c
> > b/kernel/trace/trace_events_trigger.c
> > index d00fee705f9c..c3371a6902af 100644
> > --- a/kernel/trace/trace_events_trigger.c
> > +++ b/kernel/trace/trace_events_trigger.c
> > @@ -2025,6 +2025,7 @@ __init int register_trigger_cmds(void)
> >         register_trigger_enable_disable_cmds();
> >         register_trigger_hist_enable_disable_cmds();
> >         register_trigger_hist_cmd();
> > +       register_trigger_object_cmd();
> > 
> >         return 0;
> >  }
> > diff --git a/kernel/trace/trace_object.c
> > b/kernel/trace/trace_object.c
> > new file mode 100644
> > index 000000000000..540e387c613a
> > --- /dev/null
> > +++ b/kernel/trace/trace_object.c
> > @@ -0,0 +1,515 @@
> > +// SPDX-License-Identifier: GPL-2.0
> > +/*
> > + * trace the kernel object in the kernel function parameter
> > + * Copyright (C) 2021 Jeff Xie <xiehuan09@gmail.com>
> > + */
> > +
> > +#define pr_fmt(fmt) "trace_object: " fmt
> > +
> > +#include "trace_output.h"
> > +
> > +#define MAX_TRACED_OBJECT 5
> > +#define OBJTRACE_CMD_LEN  10
> > +static DEFINE_PER_CPU(unsigned int, trace_object_event_disable);
> > +static DEFINE_RAW_SPINLOCK(trace_obj_lock);
> > +static struct trace_event_file event_trace_file;
> > +static const int max_args_num = 6;
> > +static atomic_t trace_object_ref;
> > +static atomic_t num_traced_obj;
> > +static int exit_trace_object(void);
> > +static int init_trace_object(void);
> > +
> > +static struct object_instance {
> > +       void *obj;
> > +} traced_obj[MAX_TRACED_OBJECT];
> > +
> > +/* objtrace private data */
> > +struct objtrace_trigger_data {
> > +       struct ftrace_event_field *field;
> > +       char objtrace_cmd[OBJTRACE_CMD_LEN];
> > +};
> > +
> > +static bool object_exist(void *obj)
> > +{
> > +       int i, max;
> > +
> > +       max = atomic_read(&num_traced_obj);
> > +       smp_rmb();
> > +       for (i = 0; i < max; i++) {
> > +               if (traced_obj[i].obj == obj)
> > +                       return true;
> > +       }
> > +       return false;
> > +}
> > +
> > +static bool object_empty(void)
> > +{
> > +       return !atomic_read(&num_traced_obj);
> > +}
> > +
> > +static void set_trace_object(void *obj)
> > +{
> > +       unsigned long flags;
> > +
> > +       if (in_nmi())
> > +               return;
> > +
> > +       if (!obj)
> > +               return;
> > +
> > +       if (object_exist(obj))
> > +               return;
> > +
> > +       /* only this place has write operations */
> > +       raw_spin_lock_irqsave(&trace_obj_lock, flags);
> > +       if (atomic_read(&num_traced_obj) == MAX_TRACED_OBJECT) {
> > +               trace_printk("object_pool is full, can't trace
> > object:0x%px\n", obj);
> > +               goto out;
> > +       }
> > +       traced_obj[atomic_read(&num_traced_obj)].obj = obj;
> > +       /* make sure the num_traced_obj update always appears after
> > traced_obj update */
> > +       smp_wmb();
> > +       atomic_inc(&num_traced_obj);
> > +out:
> > +       raw_spin_unlock_irqrestore(&trace_obj_lock, flags);
> > +}
> > +
> > +static void submit_trace_object(unsigned long ip, unsigned long
> > parent_ip,
> > +                                unsigned long object)
> > +{
> > +
> > +       struct trace_buffer *buffer;
> > +       struct ring_buffer_event *event;
> > +       struct trace_object_entry *entry;
> > +       int pc;
> > +
> > +       pc = preempt_count();
> > +       event = trace_event_buffer_lock_reserve(&buffer,
> > &event_trace_file,
> > +                       TRACE_OBJECT, sizeof(*entry), pc);
> > +       if (!event)
> > +               return;
> > +       entry   = ring_buffer_event_data(event);
> > +       entry->ip                       = ip;
> > +       entry->parent_ip                = parent_ip;
> > +       entry->object                   = object;
> > +
> > +       event_trigger_unlock_commit(&event_trace_file, buffer,
> > event,
> > +               entry, pc);
> > +}
> > +
> > +static void
> > +trace_object_events_call(unsigned long ip, unsigned long
> > parent_ip,
> > +               struct ftrace_ops *op, struct ftrace_regs *fregs)
> > +{
> > +       struct pt_regs *pt_regs = ftrace_get_regs(fregs);
> > +       unsigned long obj;
> > +       unsigned int disabled;
> > +       int n;
> > +
> > +       preempt_disable_notrace();
> > +
> > +       disabled = this_cpu_inc_return(trace_object_event_disable);
> > +       if (disabled != 1)
> > +               goto out;
> > +
> > +       if (object_empty())
> > +               goto out;
> > +
> > +       for (n = 0; n < max_args_num; n++) {
> > +               obj = regs_get_kernel_argument(pt_regs, n);
> > +               if (object_exist((void *)obj))
> > +                       submit_trace_object(ip, parent_ip, obj);
> > +       /* The parameters of a function may match multiple objects
> > */
> > +       }
> > +out:
> > +       this_cpu_dec(trace_object_event_disable);
> > +       preempt_enable_notrace();
> > +}
> > +
> > +static struct ftrace_ops trace_ops = {
> > +       .func  = trace_object_events_call,
> > +       .flags = FTRACE_OPS_FL_SAVE_REGS,
> > +};
> > +
> > +static void
> > +trace_object_trigger(struct event_trigger_data *data,
> > +                  struct trace_buffer *buffer,  void *rec,
> > +                  struct ring_buffer_event *event)
> > +{
> > +       struct objtrace_trigger_data *obj_data = data-
> > >private_data;
> > +       struct ftrace_event_field *field;
> > +       void *obj = NULL;
> > +
> > +       field = obj_data->field;
> > +       memcpy(&obj, rec + field->offset, sizeof(obj));
> > +       set_trace_object(obj);
> > +}
> > +
> > +static void
> > +trace_object_trigger_free(struct event_trigger_ops *ops,
> > +                  struct event_trigger_data *data)
> > +{
> > +       if (WARN_ON_ONCE(data->ref <= 0))
> > +               return;
> > +
> > +       data->ref--;
> > +       if (!data->ref) {
> > +               kfree(data->private_data);
> > +               trigger_data_free(data);
> > +       }
> > +}
> > +
> > +static void
> > +trace_object_count_trigger(struct event_trigger_data *data,
> > +                        struct trace_buffer *buffer, void *rec,
> > +                        struct ring_buffer_event *event)
> > +{
> > +       if (!data->count)
> > +               return;
> > +
> > +       if (data->count != -1)
> > +               (data->count)--;
> > +
> > +       trace_object_trigger(data, buffer, rec, event);
> > +}
> > +
> > +static int event_object_trigger_init(struct event_trigger_ops
> > *ops,
> > +                      struct event_trigger_data *data)
> > +{
> > +       data->ref++;
> > +       return 0;
> > +}
> > +
> > +static int
> > +event_trigger_print(const char *name, struct seq_file *m,
> > +               void *data, char *filter_str, void *objtrace_data)
> > +{
> > +       long count = (long)data;
> > +       struct objtrace_trigger_data *obj_data = objtrace_data;
> > +
> > +       seq_puts(m, name);
> > +
> > +       seq_printf(m, ":%s", obj_data->objtrace_cmd);
> > +       seq_printf(m, ":%s", obj_data->field->name);
> > +
> > +       if (count == -1)
> > +               seq_puts(m, ":unlimited");
> > +       else
> > +               seq_printf(m, ":count=%ld", count);
> > +
> > +       if (filter_str)
> > +               seq_printf(m, " if %s\n", filter_str);
> > +       else
> > +               seq_putc(m, '\n');
> > +
> > +       return 0;
> > +}
> > +
> > +static int
> > +trace_object_trigger_print(struct seq_file *m, struct
> > event_trigger_ops *ops,
> > +                        struct event_trigger_data *data)
> > +{
> > +       return event_trigger_print("objtrace", m, (void *)data-
> > >count,
> > +                                  data->filter_str, data-
> > >private_data);
> > +}
> > +
> > +static struct event_trigger_ops objecttrace_trigger_ops = {
> > +       .trigger                = trace_object_trigger,
> > +       .print                  = trace_object_trigger_print,
> > +       .init                   = event_object_trigger_init,
> > +       .free                   = trace_object_trigger_free,
> > +};
> > +
> > +static struct event_trigger_ops objecttrace_count_trigger_ops = {
> > +       .trigger                = trace_object_count_trigger,
> > +       .print                  = trace_object_trigger_print,
> > +       .init                   = event_object_trigger_init,
> > +       .free                   = trace_object_trigger_free,
> > +};
> > +
> > +static struct event_trigger_ops *
> > +objecttrace_get_trigger_ops(char *cmd, char *param)
> > +{
> > +       return param ? &objecttrace_count_trigger_ops :
> > &objecttrace_trigger_ops;
> > +}
> > +
> > +static int register_object_trigger(char *glob,
> > +                           struct event_trigger_data *data,
> > +                           struct trace_event_file *file)
> > +{
> > +       struct event_trigger_data *test;
> > +       int ret = 0;
> > +
> > +       lockdep_assert_held(&event_mutex);
> > +
> > +       list_for_each_entry(test, &file->triggers, list) {
> > +               if (test->cmd_ops->trigger_type == data->cmd_ops-
> > >trigger_type) {
> > +                       ret = -EEXIST;
> > +                       goto out;
> > +               }
> > +       }
> > +
> > +       if (data->ops->init) {
> > +               ret = data->ops->init(data->ops, data);
> > +               if (ret < 0)
> > +                       goto out;
> > +       }
> > +
> > +       list_add_rcu(&data->list, &file->triggers);
> > +       ret++;
> > +
> > +       update_cond_flag(file);
> > +       if (trace_event_trigger_enable_disable(file, 1) < 0) {
> > +               list_del_rcu(&data->list);
> > +               update_cond_flag(file);
> > +               ret--;
> > +       }
> > +       init_trace_object();
> > +out:
> > +       return ret;
> > +}
> 
> Can't replace the register_object_trigger() with register_trigger(),
> as there will be no place to invoke the init_trace_object(), maybe It
> will be better to add a callback in struct event_command,
> for example, cmd_ops->init()

register_trigger() calls event_trigger->ops->init(), can you call
init_trace_object() from there?

> 
> > +static void unregister_object_trigger(char *glob,
> > +                              struct event_trigger_data *test,
> > +                              struct trace_event_file *file)
> > +{
> > +       struct event_trigger_data *data;
> > +       bool unregistered = false;
> > +
> > +       lockdep_assert_held(&event_mutex);
> > +
> > +       list_for_each_entry(data, &file->triggers, list) {
> > +               if (data->cmd_ops->trigger_type == test->cmd_ops-
> > >trigger_type) {
> > +                       unregistered = true;
> > +                       list_del_rcu(&data->list);
> > +                       trace_event_trigger_enable_disable(file,
> > 0);
> > +                       update_cond_flag(file);
> > +                       break;
> > +               }
> > +       }
> > +
> > +       if (unregistered) {
> > +               if (data->ops->free)
> > +                       data->ops->free(data->ops, data);
> > +               exit_trace_object();
> > +       }
> > +}
> > +
> 
> Can't replace the unregister_object_trigger() with
> unregister_trigger(),
> as there will be no place to invoke the exit_trace_object().Maybe
> need
> to add a callback, for example cmd_ops->free().

Similarly, unregister_trigger() calls event_trigger->ops->free().

> 
> > +static bool field_exist(struct trace_event_file *file,
> > +                       struct event_command *cmd_ops,
> > +                       const char *field_name)
> > +{
> > +       struct event_trigger_data *data;
> > +       struct objtrace_trigger_data *obj_data;
> > +
> > +       lockdep_assert_held(&event_mutex);
> > +
> > +       list_for_each_entry(data, &file->triggers, list) {
> > +               if (data->cmd_ops->trigger_type == cmd_ops-
> > >trigger_type) {
> > +                       obj_data = data->private_data;
> > +                       if (!strcmp(obj_data->field->name,
> > field_name))
> > +                               return true;
> > +               }
> > +       }
> > +
> > +       return false;
> > +}
> > +
> > +static int
> > +event_object_trigger_parse(struct event_command *cmd_ops,
> > +                      struct trace_event_file *file,
> > +                      char *glob, char *cmd, char *param)
> > +{
> > +       struct event_trigger_data *trigger_data;
> > +       struct event_trigger_ops *trigger_ops;
> > +       struct objtrace_trigger_data *obj_data;
> > +       struct trace_event_call *call;
> > +       struct ftrace_event_field *field;
> > +       char *objtrace_cmd;
> > +       char *trigger = NULL;
> > +       char *arg;
> > +       char *number;
> > +       int ret;
> > +       bool remove = false;
> > +
> > +       ret = -EINVAL;
> > +       if (!param)
> > +               goto out;
> > +
> > +       /* separate the trigger from the filter (c:a:n [if filter])
> > */
> > +       trigger = strsep(&param, " \t");
> > +       if (!trigger)
> > +               goto out;
> > +       if (param) {
> > +               param = skip_spaces(param);
> > +               if (!*param)
> > +                       param = NULL;
> > +       }
> > +
> > +       objtrace_cmd = strsep(&trigger, ":");
> > +       if (!objtrace_cmd || strcmp(objtrace_cmd, "add"))
> > +               goto out;
> > +
> > +       arg = strsep(&trigger, ":");
> > +       if (!arg)
> > +               goto out;
> > +       call = file->event_call;
> > +       field = trace_find_event_field(call, arg);
> > +       if (!field)
> > +               goto out;
> > +
> > +       if (field->size != sizeof(void *))
> > +               goto out;
> > +
> > +       if (glob[0] == '!')
> > +               remove = true;
> > +
> > +       if (remove && !field_exist(file, cmd_ops, field->name))
> > +       goto out;
> > +       trigger_ops = cmd_ops->get_trigger_ops(cmd, trigger);
> > +       ret = -ENOMEM;
> > +       obj_data = kzalloc(sizeof(*obj_data), GFP_KERNEL);
> > +       if (!obj_data)
> > +               goto out;
> > +       obj_data->field = field;
> > +       snprintf(obj_data->objtrace_cmd, OBJTRACE_CMD_LEN,
> > objtrace_cmd);
> > +
> > +       trigger_data = kzalloc(sizeof(*trigger_data), GFP_KERNEL);
> > +       if (!trigger_data) {
> > +               kfree(obj_data);
> > +               goto out;
> > +       }
> > +
> > +       trigger_data->count = -1;
> > +       trigger_data->ops = trigger_ops;
> > +       trigger_data->cmd_ops = cmd_ops;
> > +       trigger_data->private_data = obj_data;
> > +       INIT_LIST_HEAD(&trigger_data->list);
> > +       INIT_LIST_HEAD(&trigger_data->named_list);

In the new code, the above can now be done by event_trigger_alloc().

> > +
> > +       if (remove) {
> > +               cmd_ops->unreg(glob+1, trigger_data, file);
> > +               kfree(obj_data);
> > +               kfree(trigger_data);
> > +               ret = 0;
> > +               goto out;
> > +       }
> > +
> > +       if (trigger) {
> > +               number = strsep(&trigger, ":");
> > +
> > +               ret = -EINVAL;
> > +               if (!strlen(number))
> > +                       goto out_free;
> > +
> > +               /*
> > +                * We use the callback data field (which is a
> > pointer)
> > +                * as our counter.
> > +                */
> > +               ret = kstrtoul(number, 0, &trigger_data->count);
> > +               if (ret)
> > +                       goto out_free;
> > +       }

Similarly, event_trigger_parse_num() should cover this.

> > +
> > +       if (!param) /* if param is non-empty, it's supposed to be a
> > filter */
> > +               goto out_reg;
> > +
> > +       if (!cmd_ops->set_filter)
> > +               goto out_reg;
> > +
> > +       ret = cmd_ops->set_filter(param, trigger_data, file);
> > +       if (ret < 0)
> > +               goto out_free;

Similarly, event_trigger_set_filter()...

> > +
> > + out_reg:
> > +       /* Up the trigger_data count to make sure reg doesn't free
> > it on failure */
> > +       event_object_trigger_init(trigger_ops, trigger_data);
> > +       ret = cmd_ops->reg(glob, trigger_data, file);
> > +       /*
> > +        * The above returns on success the # of functions enabled,
> > +        * but if it didn't find any functions it returns zero.
> > +        * Consider no functions a failure too.
> > +        */
> > +       if (!ret) {
> > +               cmd_ops->unreg(glob, trigger_data, file);
> > +               ret = -ENOENT;
> > +       } else if (ret > 0)
> > +               ret = 0;
> > +
> > +       /* Down the counter of trigger_data or free it if not used
> > anymore */
> > +       trace_object_trigger_free(trigger_ops, trigger_data);
> > + out:
> > +       return ret;
> > +
> > + out_free:
> > +       if (cmd_ops->set_filter)
> > +               cmd_ops->set_filter(NULL, trigger_data, NULL);
> > +       kfree(obj_data);
> > +       kfree(trigger_data);
> > +       goto out;
> > +}
> 
> Can't replace the event_object_trigger_parse() with
> event_trigger_parse().
> Because the event_object_trigger_parse() needs to parse more fields,
> for example need to parse objtrace_cmd("add" etc.,)
> and need to use other special private_data. Maybe It will have to use
> the event_object_trigger_parse().

Yeah, for this one, you'll need to customize, since new commands like
this need their own parsing.  The functions I noted above (and others)
attempt to simplify the overall parsing by putting the boilerplate in
separate functions, but you still need to do whatever extras your
parsing function needs.

You'll probably end up with some kind of combination of
event_hist_trigger_parse() and event_trigger_parse() - I'd take a look
at those or other event_command parse() functions for other commands.

hth,

Tom

> 
> > +static struct event_command trigger_object_cmd = {
> > +       .name                   = "objtrace",
> > +       .trigger_type           = ETT_TRACE_OBJECT,
> > +       .flags                  = EVENT_CMD_FL_NEEDS_REC,
> > +       .parse                  = event_object_trigger_parse,
> > +       .reg                    = register_object_trigger,
> > +       .unreg                  = unregister_object_trigger,
> > +       .get_trigger_ops        = objecttrace_get_trigger_ops,
> > +       .set_filter             = set_trigger_filter,
> > +};
> > +
> > +__init int register_trigger_object_cmd(void)
> > +{
> > +       int ret;
> > +
> > +       ret = register_event_command(&trigger_object_cmd);
> > +       WARN_ON(ret < 0);
> > +
> > +       return ret;
> > +}
> > +
> > +static int init_trace_object(void)
> > +{
> > +       int ret;
> > +
> > +       if (atomic_inc_return(&trace_object_ref) != 1) {
> > +               ret = 0;
> > +               goto out;
> > +       }
> > +
> > +       event_trace_file.tr = top_trace_array();
> > +       if (WARN_ON(!event_trace_file.tr)) {
> > +               ret = -1;
> > +               atomic_dec(&trace_object_ref);
> > +               goto out;
> > +       }
> > +       ret = register_ftrace_function(&trace_ops);
> > +out:
> > +       return ret;
> > +}
> > +
> > +static int exit_trace_object(void)
> > +{
> > +       int ret;
> > +
> > +       if (WARN_ON_ONCE(atomic_read(&trace_object_ref) <= 0)) {
> > +               ret = -1;
> > +               goto out;
> > +       }
> > +
> > +       if (atomic_dec_return(&trace_object_ref) != 0) {
> > +               ret = 0;
> > +               goto out;
> > +       }
> > +
> > +       ret = unregister_ftrace_function(&trace_ops);
> > +       if (ret) {
> > +               pr_err("can't unregister ftrace for trace
> > object\n");
> > +               goto out;
> > +       }
> > +       atomic_set(&num_traced_obj, 0);
> > +out:
> > +       return ret;
> > +}
> > diff --git a/kernel/trace/trace_output.c
> > b/kernel/trace/trace_output.c
> > index 8aa493d25c73..265428154638 100644
> > --- a/kernel/trace/trace_output.c
> > +++ b/kernel/trace/trace_output.c
> > @@ -1547,6 +1547,45 @@ static struct trace_event
> > trace_func_repeats_event = {
> >         .funcs          = &trace_func_repeats_funcs,
> >  };
> > 
> > +/* TRACE_OBJECT */
> > +static enum print_line_t trace_object_print(struct trace_iterator
> > *iter, int flags,
> > +                                       struct trace_event *event)
> > +{
> > +       struct trace_object_entry *field;
> > +       struct trace_seq *s = &iter->seq;
> > +
> > +       trace_assign_type(field, iter->ent);
> > +       print_fn_trace(s, field->ip, field->parent_ip, flags);
> > +       trace_seq_printf(s, " object:0x%lx", field->object);
> > +       trace_seq_putc(s, '\n');
> > +
> > +       return trace_handle_return(s);
> > +}
> > +
> > +static enum print_line_t trace_object_raw(struct trace_iterator
> > *iter, int flags,
> > +                                     struct trace_event *event)
> > +{
> > +       struct trace_object_entry *field;
> > +
> > +       trace_assign_type(field, iter->ent);
> > +
> > +       trace_seq_printf(&iter->seq, "%lx %lx\n",
> > +                        field->ip,
> > +                        field->parent_ip);
> > +
> > +       return trace_handle_return(&iter->seq);
> > +}
> > +
> > +static struct trace_event_functions trace_object_funcs = {
> > +       .trace          = trace_object_print,
> > +       .raw            = trace_object_raw,
> > +};
> > +
> > +static struct trace_event trace_object_event = {
> > +       .type           = TRACE_OBJECT,
> > +       .funcs          = &trace_object_funcs,
> > +};
> > +
> >  static struct trace_event *events[] __initdata = {
> >         &trace_fn_event,
> >         &trace_ctx_event,
> > @@ -1561,6 +1600,7 @@ static struct trace_event *events[]
> > __initdata = {
> >         &trace_timerlat_event,
> >         &trace_raw_data_event,
> >         &trace_func_repeats_event,
> > +       &trace_object_event,
> >         NULL
> >  };
> > 
> > --
> > 2.25.1
> > 
> 
> Thanks,
> JeffXie


^ permalink raw reply	[flat|nested] 18+ messages in thread

* Re: [PATCH v9 1/4] trace: Add trace any kernel object
  2022-05-02 17:42     ` Tom Zanussi
@ 2022-05-04  7:42       ` Jeff Xie
  0 siblings, 0 replies; 18+ messages in thread
From: Jeff Xie @ 2022-05-04  7:42 UTC (permalink / raw)
  To: Tom Zanussi; +Cc: Steven Rostedt, Masami Hiramatsu, mingo, linux-kernel

Hi Tom,

Thanks for the detailed explanation.

On Tue, May 3, 2022 at 1:42 AM Tom Zanussi <zanussi@kernel.org> wrote:
>
> Hi Jeff,
>
> On Wed, 2022-04-27 at 23:50 +0800, Jeff Xie wrote:
> > Hi Tom,
> >
> > I have checked your patches:
> >
> > https://lore.kernel.org/all/cover.1644010575.git.zanussi@kernel.org/
> >
> > But I found that I can't use the generic function you implemented,
> > I don't know if I understand correctly, when you are free, hope you
> > can help to check it ;-) , thanks.
> >
> > On Fri, Feb 4, 2022 at 11:57 AM Jeff Xie <xiehuan09@gmail.com> wrote:
> > >
> > > Introduce objtrace trigger to get the call flow by tracing any
> > > kernel
> > > object in the function parameter.
> > >
> > > The objtrace trigger makes a list of the target object address from
> > > the given event parameter, and records all kernel function calls
> > > which has the object address in its parameter.
> > >
> > > Syntax:
> > >         objtrace:add:obj[:count][if <filter>]
> > >
> > > Usage:
> > >         # echo 'p bio_add_page arg1=$arg1' > kprobe_events
> > >         # cd events/kprobes/p_bio_add_page_0
> > >         # echo 'objtrace:add:arg1:1 if comm == "cat"' > ./trigger
> > >         # cat /test.txt
> > >
> > > Signed-off-by: Jeff Xie <xiehuan09@gmail.com>
> > > ---
> > >  include/linux/trace_events.h        |   1 +
> > >  kernel/trace/Kconfig                |  10 +
> > >  kernel/trace/Makefile               |   1 +
> > >  kernel/trace/trace.c                |   3 +
> > >  kernel/trace/trace.h                |   8 +
> > >  kernel/trace/trace_entries.h        |  17 +
> > >  kernel/trace/trace_events_trigger.c |   1 +
> > >  kernel/trace/trace_object.c         | 515
> > > ++++++++++++++++++++++++++++
> > >  kernel/trace/trace_output.c         |  40 +++
> > >  9 files changed, 596 insertions(+)
> > >  create mode 100644 kernel/trace/trace_object.c
> > >
> > > diff --git a/include/linux/trace_events.h
> > > b/include/linux/trace_events.h
> > > index 70c069aef02c..3ccdbc1d25dd 100644
> > > --- a/include/linux/trace_events.h
> > > +++ b/include/linux/trace_events.h
> > > @@ -685,6 +685,7 @@ enum event_trigger_type {
> > >         ETT_EVENT_HIST          = (1 << 4),
> > >         ETT_HIST_ENABLE         = (1 << 5),
> > >         ETT_EVENT_EPROBE        = (1 << 6),
> > > +       ETT_TRACE_OBJECT        = (1 << 7),
> > >  };
> > >
> > >  extern int filter_match_preds(struct event_filter *filter, void
> > > *rec);
> > > diff --git a/kernel/trace/Kconfig b/kernel/trace/Kconfig
> > > index a5eb5e7fd624..c51b7eb1508d 100644
> > > --- a/kernel/trace/Kconfig
> > > +++ b/kernel/trace/Kconfig
> > > @@ -250,6 +250,16 @@ config FUNCTION_PROFILER
> > >
> > >           If in doubt, say N.
> > >
> > > +config TRACE_OBJECT
> > > +       bool "Trace kernel object in function parameter"
> > > +       depends on FUNCTION_TRACER
> > > +       depends on HAVE_FUNCTION_ARG_ACCESS_API
> > > +       select TRACING
> > > +       default y
> > > +       help
> > > +        You can trace the kernel object in the kernel function
> > > parameter.
> > > +        The kernel object is dynamically specified via event
> > > trigger.
> > > +
> > >  config STACK_TRACER
> > >         bool "Trace max stack"
> > >         depends on HAVE_FUNCTION_TRACER
> > > diff --git a/kernel/trace/Makefile b/kernel/trace/Makefile
> > > index bedc5caceec7..b924b8e55922 100644
> > > --- a/kernel/trace/Makefile
> > > +++ b/kernel/trace/Makefile
> > > @@ -67,6 +67,7 @@ obj-$(CONFIG_FUNCTION_GRAPH_TRACER) +=
> > > trace_functions_graph.o
> > >  obj-$(CONFIG_TRACE_BRANCH_PROFILING) += trace_branch.o
> > >  obj-$(CONFIG_BLK_DEV_IO_TRACE) += blktrace.o
> > >  obj-$(CONFIG_FUNCTION_GRAPH_TRACER) += fgraph.o
> > > +obj-$(CONFIG_TRACE_OBJECT) += trace_object.o
> > >  ifeq ($(CONFIG_BLOCK),y)
> > >  obj-$(CONFIG_EVENT_TRACING) += blktrace.o
> > >  endif
> > > diff --git a/kernel/trace/trace.c b/kernel/trace/trace.c
> > > index a734d5ae34c8..b4513c2bbd52 100644
> > > --- a/kernel/trace/trace.c
> > > +++ b/kernel/trace/trace.c
> > > @@ -5605,6 +5605,9 @@ static const char readme_msg[] =
> > >         "\t            enable_hist:<system>:<event>\n"
> > >         "\t            disable_hist:<system>:<event>\n"
> > >  #endif
> > > +#ifdef CONFIG_TRACE_OBJECT
> > > +       "\t            objtrace:add:obj[:count][if <filter>]\n"
> > > +#endif
> > >  #ifdef CONFIG_STACKTRACE
> > >         "\t\t    stacktrace\n"
> > >  #endif
> > > diff --git a/kernel/trace/trace.h b/kernel/trace/trace.h
> > > index 0f5e22238cd2..8b66515a36d5 100644
> > > --- a/kernel/trace/trace.h
> > > +++ b/kernel/trace/trace.h
> > > @@ -54,6 +54,7 @@ enum trace_type {
> > >         TRACE_TIMERLAT,
> > >         TRACE_RAW_DATA,
> > >         TRACE_FUNC_REPEATS,
> > > +       TRACE_OBJECT,
> > >
> > >         __TRACE_LAST_TYPE,
> > >  };
> > > @@ -472,6 +473,7 @@ extern void __ftrace_bad_type(void);
> > >                           TRACE_GRAPH_RET);             \
> > >                 IF_ASSIGN(var, ent, struct
> > > func_repeats_entry,          \
> > >                           TRACE_FUNC_REPEATS);
> > >      \
> > > +               IF_ASSIGN(var, ent, struct trace_object_entry,
> > > TRACE_OBJECT);\
> > >                 __ftrace_bad_type();
> > >      \
> > >         } while (0)
> > >
> > > @@ -1537,6 +1539,12 @@ static inline int
> > > register_trigger_hist_cmd(void) { return 0; }
> > >  static inline int register_trigger_hist_enable_disable_cmds(void)
> > > { return 0; }
> > >  #endif
> > >
> > > +#ifdef CONFIG_TRACE_OBJECT
> > > +extern int register_trigger_object_cmd(void);
> > > +#else
> > > +static inline int register_trigger_object_cmd(void) { return 0; }
> > > +#endif
> > > +
> > >  extern int register_trigger_cmds(void);
> > >  extern void clear_event_triggers(struct trace_array *tr);
> > >
> > > diff --git a/kernel/trace/trace_entries.h
> > > b/kernel/trace/trace_entries.h
> > > index cd41e863b51c..bb120d9498a9 100644
> > > --- a/kernel/trace/trace_entries.h
> > > +++ b/kernel/trace/trace_entries.h
> > > @@ -401,3 +401,20 @@ FTRACE_ENTRY(timerlat, timerlat_entry,
> > >                  __entry->context,
> > >                  __entry->timer_latency)
> > >  );
> > > +
> > > +/*
> > > + * trace object entry:
> > > + */
> > > +FTRACE_ENTRY(object, trace_object_entry,
> > > +
> > > +       TRACE_OBJECT,
> > > +
> > > +       F_STRUCT(
> > > +               __field(        unsigned
> > > long,          ip              )
> > > +               __field(        unsigned
> > > long,          parent_ip       )
> > > +               __field(        unsigned
> > > long,          object          )
> > > +       ),
> > > +
> > > +       F_printk(" %ps <-- %ps object:%lx\n",
> > > +                (void *)__entry->ip, (void *)__entry->parent_ip,
> > > __entry->object)
> > > +);
> > > diff --git a/kernel/trace/trace_events_trigger.c
> > > b/kernel/trace/trace_events_trigger.c
> > > index d00fee705f9c..c3371a6902af 100644
> > > --- a/kernel/trace/trace_events_trigger.c
> > > +++ b/kernel/trace/trace_events_trigger.c
> > > @@ -2025,6 +2025,7 @@ __init int register_trigger_cmds(void)
> > >         register_trigger_enable_disable_cmds();
> > >         register_trigger_hist_enable_disable_cmds();
> > >         register_trigger_hist_cmd();
> > > +       register_trigger_object_cmd();
> > >
> > >         return 0;
> > >  }
> > > diff --git a/kernel/trace/trace_object.c
> > > b/kernel/trace/trace_object.c
> > > new file mode 100644
> > > index 000000000000..540e387c613a
> > > --- /dev/null
> > > +++ b/kernel/trace/trace_object.c
> > > @@ -0,0 +1,515 @@
> > > +// SPDX-License-Identifier: GPL-2.0
> > > +/*
> > > + * trace the kernel object in the kernel function parameter
> > > + * Copyright (C) 2021 Jeff Xie <xiehuan09@gmail.com>
> > > + */
> > > +
> > > +#define pr_fmt(fmt) "trace_object: " fmt
> > > +
> > > +#include "trace_output.h"
> > > +
> > > +#define MAX_TRACED_OBJECT 5
> > > +#define OBJTRACE_CMD_LEN  10
> > > +static DEFINE_PER_CPU(unsigned int, trace_object_event_disable);
> > > +static DEFINE_RAW_SPINLOCK(trace_obj_lock);
> > > +static struct trace_event_file event_trace_file;
> > > +static const int max_args_num = 6;
> > > +static atomic_t trace_object_ref;
> > > +static atomic_t num_traced_obj;
> > > +static int exit_trace_object(void);
> > > +static int init_trace_object(void);
> > > +
> > > +static struct object_instance {
> > > +       void *obj;
> > > +} traced_obj[MAX_TRACED_OBJECT];
> > > +
> > > +/* objtrace private data */
> > > +struct objtrace_trigger_data {
> > > +       struct ftrace_event_field *field;
> > > +       char objtrace_cmd[OBJTRACE_CMD_LEN];
> > > +};
> > > +
> > > +static bool object_exist(void *obj)
> > > +{
> > > +       int i, max;
> > > +
> > > +       max = atomic_read(&num_traced_obj);
> > > +       smp_rmb();
> > > +       for (i = 0; i < max; i++) {
> > > +               if (traced_obj[i].obj == obj)
> > > +                       return true;
> > > +       }
> > > +       return false;
> > > +}
> > > +
> > > +static bool object_empty(void)
> > > +{
> > > +       return !atomic_read(&num_traced_obj);
> > > +}
> > > +
> > > +static void set_trace_object(void *obj)
> > > +{
> > > +       unsigned long flags;
> > > +
> > > +       if (in_nmi())
> > > +               return;
> > > +
> > > +       if (!obj)
> > > +               return;
> > > +
> > > +       if (object_exist(obj))
> > > +               return;
> > > +
> > > +       /* only this place has write operations */
> > > +       raw_spin_lock_irqsave(&trace_obj_lock, flags);
> > > +       if (atomic_read(&num_traced_obj) == MAX_TRACED_OBJECT) {
> > > +               trace_printk("object_pool is full, can't trace
> > > object:0x%px\n", obj);
> > > +               goto out;
> > > +       }
> > > +       traced_obj[atomic_read(&num_traced_obj)].obj = obj;
> > > +       /* make sure the num_traced_obj update always appears after
> > > traced_obj update */
> > > +       smp_wmb();
> > > +       atomic_inc(&num_traced_obj);
> > > +out:
> > > +       raw_spin_unlock_irqrestore(&trace_obj_lock, flags);
> > > +}
> > > +
> > > +static void submit_trace_object(unsigned long ip, unsigned long
> > > parent_ip,
> > > +                                unsigned long object)
> > > +{
> > > +
> > > +       struct trace_buffer *buffer;
> > > +       struct ring_buffer_event *event;
> > > +       struct trace_object_entry *entry;
> > > +       int pc;
> > > +
> > > +       pc = preempt_count();
> > > +       event = trace_event_buffer_lock_reserve(&buffer,
> > > &event_trace_file,
> > > +                       TRACE_OBJECT, sizeof(*entry), pc);
> > > +       if (!event)
> > > +               return;
> > > +       entry   = ring_buffer_event_data(event);
> > > +       entry->ip                       = ip;
> > > +       entry->parent_ip                = parent_ip;
> > > +       entry->object                   = object;
> > > +
> > > +       event_trigger_unlock_commit(&event_trace_file, buffer,
> > > event,
> > > +               entry, pc);
> > > +}
> > > +
> > > +static void
> > > +trace_object_events_call(unsigned long ip, unsigned long
> > > parent_ip,
> > > +               struct ftrace_ops *op, struct ftrace_regs *fregs)
> > > +{
> > > +       struct pt_regs *pt_regs = ftrace_get_regs(fregs);
> > > +       unsigned long obj;
> > > +       unsigned int disabled;
> > > +       int n;
> > > +
> > > +       preempt_disable_notrace();
> > > +
> > > +       disabled = this_cpu_inc_return(trace_object_event_disable);
> > > +       if (disabled != 1)
> > > +               goto out;
> > > +
> > > +       if (object_empty())
> > > +               goto out;
> > > +
> > > +       for (n = 0; n < max_args_num; n++) {
> > > +               obj = regs_get_kernel_argument(pt_regs, n);
> > > +               if (object_exist((void *)obj))
> > > +                       submit_trace_object(ip, parent_ip, obj);
> > > +       /* The parameters of a function may match multiple objects
> > > */
> > > +       }
> > > +out:
> > > +       this_cpu_dec(trace_object_event_disable);
> > > +       preempt_enable_notrace();
> > > +}
> > > +
> > > +static struct ftrace_ops trace_ops = {
> > > +       .func  = trace_object_events_call,
> > > +       .flags = FTRACE_OPS_FL_SAVE_REGS,
> > > +};
> > > +
> > > +static void
> > > +trace_object_trigger(struct event_trigger_data *data,
> > > +                  struct trace_buffer *buffer,  void *rec,
> > > +                  struct ring_buffer_event *event)
> > > +{
> > > +       struct objtrace_trigger_data *obj_data = data-
> > > >private_data;
> > > +       struct ftrace_event_field *field;
> > > +       void *obj = NULL;
> > > +
> > > +       field = obj_data->field;
> > > +       memcpy(&obj, rec + field->offset, sizeof(obj));
> > > +       set_trace_object(obj);
> > > +}
> > > +
> > > +static void
> > > +trace_object_trigger_free(struct event_trigger_ops *ops,
> > > +                  struct event_trigger_data *data)
> > > +{
> > > +       if (WARN_ON_ONCE(data->ref <= 0))
> > > +               return;
> > > +
> > > +       data->ref--;
> > > +       if (!data->ref) {
> > > +               kfree(data->private_data);
> > > +               trigger_data_free(data);
> > > +       }
> > > +}
> > > +
> > > +static void
> > > +trace_object_count_trigger(struct event_trigger_data *data,
> > > +                        struct trace_buffer *buffer, void *rec,
> > > +                        struct ring_buffer_event *event)
> > > +{
> > > +       if (!data->count)
> > > +               return;
> > > +
> > > +       if (data->count != -1)
> > > +               (data->count)--;
> > > +
> > > +       trace_object_trigger(data, buffer, rec, event);
> > > +}
> > > +
> > > +static int event_object_trigger_init(struct event_trigger_ops
> > > *ops,
> > > +                      struct event_trigger_data *data)
> > > +{
> > > +       data->ref++;
> > > +       return 0;
> > > +}
> > > +
> > > +static int
> > > +event_trigger_print(const char *name, struct seq_file *m,
> > > +               void *data, char *filter_str, void *objtrace_data)
> > > +{
> > > +       long count = (long)data;
> > > +       struct objtrace_trigger_data *obj_data = objtrace_data;
> > > +
> > > +       seq_puts(m, name);
> > > +
> > > +       seq_printf(m, ":%s", obj_data->objtrace_cmd);
> > > +       seq_printf(m, ":%s", obj_data->field->name);
> > > +
> > > +       if (count == -1)
> > > +               seq_puts(m, ":unlimited");
> > > +       else
> > > +               seq_printf(m, ":count=%ld", count);
> > > +
> > > +       if (filter_str)
> > > +               seq_printf(m, " if %s\n", filter_str);
> > > +       else
> > > +               seq_putc(m, '\n');
> > > +
> > > +       return 0;
> > > +}
> > > +
> > > +static int
> > > +trace_object_trigger_print(struct seq_file *m, struct
> > > event_trigger_ops *ops,
> > > +                        struct event_trigger_data *data)
> > > +{
> > > +       return event_trigger_print("objtrace", m, (void *)data-
> > > >count,
> > > +                                  data->filter_str, data-
> > > >private_data);
> > > +}
> > > +
> > > +static struct event_trigger_ops objecttrace_trigger_ops = {
> > > +       .trigger                = trace_object_trigger,
> > > +       .print                  = trace_object_trigger_print,
> > > +       .init                   = event_object_trigger_init,
> > > +       .free                   = trace_object_trigger_free,
> > > +};
> > > +
> > > +static struct event_trigger_ops objecttrace_count_trigger_ops = {
> > > +       .trigger                = trace_object_count_trigger,
> > > +       .print                  = trace_object_trigger_print,
> > > +       .init                   = event_object_trigger_init,
> > > +       .free                   = trace_object_trigger_free,
> > > +};
> > > +
> > > +static struct event_trigger_ops *
> > > +objecttrace_get_trigger_ops(char *cmd, char *param)
> > > +{
> > > +       return param ? &objecttrace_count_trigger_ops :
> > > &objecttrace_trigger_ops;
> > > +}
> > > +
> > > +static int register_object_trigger(char *glob,
> > > +                           struct event_trigger_data *data,
> > > +                           struct trace_event_file *file)
> > > +{
> > > +       struct event_trigger_data *test;
> > > +       int ret = 0;
> > > +
> > > +       lockdep_assert_held(&event_mutex);
> > > +
> > > +       list_for_each_entry(test, &file->triggers, list) {
> > > +               if (test->cmd_ops->trigger_type == data->cmd_ops-
> > > >trigger_type) {
> > > +                       ret = -EEXIST;
> > > +                       goto out;
> > > +               }
> > > +       }
> > > +
> > > +       if (data->ops->init) {
> > > +               ret = data->ops->init(data->ops, data);
> > > +               if (ret < 0)
> > > +                       goto out;
> > > +       }
> > > +
> > > +       list_add_rcu(&data->list, &file->triggers);
> > > +       ret++;
> > > +
> > > +       update_cond_flag(file);
> > > +       if (trace_event_trigger_enable_disable(file, 1) < 0) {
> > > +               list_del_rcu(&data->list);
> > > +               update_cond_flag(file);
> > > +               ret--;
> > > +       }
> > > +       init_trace_object();
> > > +out:
> > > +       return ret;
> > > +}
> >
> > Can't replace the register_object_trigger() with register_trigger(),
> > as there will be no place to invoke the init_trace_object(), maybe It
> > will be better to add a callback in struct event_command,
> > for example, cmd_ops->init()
>
> register_trigger() calls event_trigger->ops->init(), can you call
> init_trace_object() from there?

Thanks, I can call init_trace_object() using event_trigger->ops->init()  now,
when I refereed  the  new code in event_trigger_parse().

> >
> > > +static void unregister_object_trigger(char *glob,
> > > +                              struct event_trigger_data *test,
> > > +                              struct trace_event_file *file)
> > > +{
> > > +       struct event_trigger_data *data;
> > > +       bool unregistered = false;
> > > +
> > > +       lockdep_assert_held(&event_mutex);
> > > +
> > > +       list_for_each_entry(data, &file->triggers, list) {
> > > +               if (data->cmd_ops->trigger_type == test->cmd_ops-
> > > >trigger_type) {
> > > +                       unregistered = true;
> > > +                       list_del_rcu(&data->list);
> > > +                       trace_event_trigger_enable_disable(file,
> > > 0);
> > > +                       update_cond_flag(file);
> > > +                       break;
> > > +               }
> > > +       }
> > > +
> > > +       if (unregistered) {
> > > +               if (data->ops->free)
> > > +                       data->ops->free(data->ops, data);
> > > +               exit_trace_object();
> > > +       }
> > > +}
> > > +
> >
> > Can't replace the unregister_object_trigger() with
> > unregister_trigger(),
> > as there will be no place to invoke the exit_trace_object().Maybe
> > need
> > to add a callback, for example cmd_ops->free().
>
> Similarly, unregister_trigger() calls event_trigger->ops->free().

Similarly, I can also use this interface.

> >
> > > +static bool field_exist(struct trace_event_file *file,
> > > +                       struct event_command *cmd_ops,
> > > +                       const char *field_name)
> > > +{
> > > +       struct event_trigger_data *data;
> > > +       struct objtrace_trigger_data *obj_data;
> > > +
> > > +       lockdep_assert_held(&event_mutex);
> > > +
> > > +       list_for_each_entry(data, &file->triggers, list) {
> > > +               if (data->cmd_ops->trigger_type == cmd_ops-
> > > >trigger_type) {
> > > +                       obj_data = data->private_data;
> > > +                       if (!strcmp(obj_data->field->name,
> > > field_name))
> > > +                               return true;
> > > +               }
> > > +       }
> > > +
> > > +       return false;
> > > +}
> > > +
> > > +static int
> > > +event_object_trigger_parse(struct event_command *cmd_ops,
> > > +                      struct trace_event_file *file,
> > > +                      char *glob, char *cmd, char *param)
> > > +{
> > > +       struct event_trigger_data *trigger_data;
> > > +       struct event_trigger_ops *trigger_ops;
> > > +       struct objtrace_trigger_data *obj_data;
> > > +       struct trace_event_call *call;
> > > +       struct ftrace_event_field *field;
> > > +       char *objtrace_cmd;
> > > +       char *trigger = NULL;
> > > +       char *arg;
> > > +       char *number;
> > > +       int ret;
> > > +       bool remove = false;
> > > +
> > > +       ret = -EINVAL;
> > > +       if (!param)
> > > +               goto out;
> > > +
> > > +       /* separate the trigger from the filter (c:a:n [if filter])
> > > */
> > > +       trigger = strsep(&param, " \t");
> > > +       if (!trigger)
> > > +               goto out;
> > > +       if (param) {
> > > +               param = skip_spaces(param);
> > > +               if (!*param)
> > > +                       param = NULL;
> > > +       }
> > > +
> > > +       objtrace_cmd = strsep(&trigger, ":");
> > > +       if (!objtrace_cmd || strcmp(objtrace_cmd, "add"))
> > > +               goto out;
> > > +
> > > +       arg = strsep(&trigger, ":");
> > > +       if (!arg)
> > > +               goto out;
> > > +       call = file->event_call;
> > > +       field = trace_find_event_field(call, arg);
> > > +       if (!field)
> > > +               goto out;
> > > +
> > > +       if (field->size != sizeof(void *))
> > > +               goto out;
> > > +
> > > +       if (glob[0] == '!')
> > > +               remove = true;
> > > +
> > > +       if (remove && !field_exist(file, cmd_ops, field->name))
> > > +       goto out;
> > > +       trigger_ops = cmd_ops->get_trigger_ops(cmd, trigger);
> > > +       ret = -ENOMEM;
> > > +       obj_data = kzalloc(sizeof(*obj_data), GFP_KERNEL);
> > > +       if (!obj_data)
> > > +               goto out;
> > > +       obj_data->field = field;
> > > +       snprintf(obj_data->objtrace_cmd, OBJTRACE_CMD_LEN,
> > > objtrace_cmd);
> > > +
> > > +       trigger_data = kzalloc(sizeof(*trigger_data), GFP_KERNEL);
> > > +       if (!trigger_data) {
> > > +               kfree(obj_data);
> > > +               goto out;
> > > +       }
> > > +
> > > +       trigger_data->count = -1;
> > > +       trigger_data->ops = trigger_ops;
> > > +       trigger_data->cmd_ops = cmd_ops;
> > > +       trigger_data->private_data = obj_data;
> > > +       INIT_LIST_HEAD(&trigger_data->list);
> > > +       INIT_LIST_HEAD(&trigger_data->named_list);
>
> In the new code, the above can now be done by event_trigger_alloc().

Thanks for the reminder.

> > > +
> > > +       if (remove) {
> > > +               cmd_ops->unreg(glob+1, trigger_data, file);
> > > +               kfree(obj_data);
> > > +               kfree(trigger_data);
> > > +               ret = 0;
> > > +               goto out;
> > > +       }
> > > +
> > > +       if (trigger) {
> > > +               number = strsep(&trigger, ":");
> > > +
> > > +               ret = -EINVAL;
> > > +               if (!strlen(number))
> > > +                       goto out_free;
> > > +
> > > +               /*
> > > +                * We use the callback data field (which is a
> > > pointer)
> > > +                * as our counter.
> > > +                */
> > > +               ret = kstrtoul(number, 0, &trigger_data->count);
> > > +               if (ret)
> > > +                       goto out_free;
> > > +       }
>
> Similarly, event_trigger_parse_num() should cover this.

Yes, thanks for the reminder.

> > > +
> > > +       if (!param) /* if param is non-empty, it's supposed to be a
> > > filter */
> > > +               goto out_reg;
> > > +
> > > +       if (!cmd_ops->set_filter)
> > > +               goto out_reg;
> > > +
> > > +       ret = cmd_ops->set_filter(param, trigger_data, file);
> > > +       if (ret < 0)
> > > +               goto out_free;
>
> Similarly, event_trigger_set_filter()...
>
> > > +
> > > + out_reg:
> > > +       /* Up the trigger_data count to make sure reg doesn't free
> > > it on failure */
> > > +       event_object_trigger_init(trigger_ops, trigger_data);
> > > +       ret = cmd_ops->reg(glob, trigger_data, file);
> > > +       /*
> > > +        * The above returns on success the # of functions enabled,
> > > +        * but if it didn't find any functions it returns zero.
> > > +        * Consider no functions a failure too.
> > > +        */
> > > +       if (!ret) {
> > > +               cmd_ops->unreg(glob, trigger_data, file);
> > > +               ret = -ENOENT;
> > > +       } else if (ret > 0)
> > > +               ret = 0;
> > > +
> > > +       /* Down the counter of trigger_data or free it if not used
> > > anymore */
> > > +       trace_object_trigger_free(trigger_ops, trigger_data);
> > > + out:
> > > +       return ret;
> > > +
> > > + out_free:
> > > +       if (cmd_ops->set_filter)
> > > +               cmd_ops->set_filter(NULL, trigger_data, NULL);
> > > +       kfree(obj_data);
> > > +       kfree(trigger_data);
> > > +       goto out;
> > > +}
> >
> > Can't replace the event_object_trigger_parse() with
> > event_trigger_parse().
> > Because the event_object_trigger_parse() needs to parse more fields,
> > for example need to parse objtrace_cmd("add" etc.,)
> > and need to use other special private_data. Maybe It will have to use
> > the event_object_trigger_parse().
>
> Yeah, for this one, you'll need to customize, since new commands like
> this need their own parsing.  The functions I noted above (and others)
> attempt to simplify the overall parsing by putting the boilerplate in
> separate functions, but you still need to do whatever extras your
> parsing function needs.
>
> You'll probably end up with some kind of combination of
> event_hist_trigger_parse() and event_trigger_parse() - I'd take a look
> at those or other event_command parse() functions for other commands.

The event_hist_trigger_parse() and event_trigger_parse() really
simplify a lot of repetitive logic with new code.
and I will refer to them.

> hth,
>
> Tom
>
> >
> > > +static struct event_command trigger_object_cmd = {
> > > +       .name                   = "objtrace",
> > > +       .trigger_type           = ETT_TRACE_OBJECT,
> > > +       .flags                  = EVENT_CMD_FL_NEEDS_REC,
> > > +       .parse                  = event_object_trigger_parse,
> > > +       .reg                    = register_object_trigger,
> > > +       .unreg                  = unregister_object_trigger,
> > > +       .get_trigger_ops        = objecttrace_get_trigger_ops,
> > > +       .set_filter             = set_trigger_filter,
> > > +};
> > > +
> > > +__init int register_trigger_object_cmd(void)
> > > +{
> > > +       int ret;
> > > +
> > > +       ret = register_event_command(&trigger_object_cmd);
> > > +       WARN_ON(ret < 0);
> > > +
> > > +       return ret;
> > > +}
> > > +
> > > +static int init_trace_object(void)
> > > +{
> > > +       int ret;
> > > +
> > > +       if (atomic_inc_return(&trace_object_ref) != 1) {
> > > +               ret = 0;
> > > +               goto out;
> > > +       }
> > > +
> > > +       event_trace_file.tr = top_trace_array();
> > > +       if (WARN_ON(!event_trace_file.tr)) {
> > > +               ret = -1;
> > > +               atomic_dec(&trace_object_ref);
> > > +               goto out;
> > > +       }
> > > +       ret = register_ftrace_function(&trace_ops);
> > > +out:
> > > +       return ret;
> > > +}
> > > +
> > > +static int exit_trace_object(void)
> > > +{
> > > +       int ret;
> > > +
> > > +       if (WARN_ON_ONCE(atomic_read(&trace_object_ref) <= 0)) {
> > > +               ret = -1;
> > > +               goto out;
> > > +       }
> > > +
> > > +       if (atomic_dec_return(&trace_object_ref) != 0) {
> > > +               ret = 0;
> > > +               goto out;
> > > +       }
> > > +
> > > +       ret = unregister_ftrace_function(&trace_ops);
> > > +       if (ret) {
> > > +               pr_err("can't unregister ftrace for trace
> > > object\n");
> > > +               goto out;
> > > +       }
> > > +       atomic_set(&num_traced_obj, 0);
> > > +out:
> > > +       return ret;
> > > +}
> > > diff --git a/kernel/trace/trace_output.c
> > > b/kernel/trace/trace_output.c
> > > index 8aa493d25c73..265428154638 100644
> > > --- a/kernel/trace/trace_output.c
> > > +++ b/kernel/trace/trace_output.c
> > > @@ -1547,6 +1547,45 @@ static struct trace_event
> > > trace_func_repeats_event = {
> > >         .funcs          = &trace_func_repeats_funcs,
> > >  };
> > >
> > > +/* TRACE_OBJECT */
> > > +static enum print_line_t trace_object_print(struct trace_iterator
> > > *iter, int flags,
> > > +                                       struct trace_event *event)
> > > +{
> > > +       struct trace_object_entry *field;
> > > +       struct trace_seq *s = &iter->seq;
> > > +
> > > +       trace_assign_type(field, iter->ent);
> > > +       print_fn_trace(s, field->ip, field->parent_ip, flags);
> > > +       trace_seq_printf(s, " object:0x%lx", field->object);
> > > +       trace_seq_putc(s, '\n');
> > > +
> > > +       return trace_handle_return(s);
> > > +}
> > > +
> > > +static enum print_line_t trace_object_raw(struct trace_iterator
> > > *iter, int flags,
> > > +                                     struct trace_event *event)
> > > +{
> > > +       struct trace_object_entry *field;
> > > +
> > > +       trace_assign_type(field, iter->ent);
> > > +
> > > +       trace_seq_printf(&iter->seq, "%lx %lx\n",
> > > +                        field->ip,
> > > +                        field->parent_ip);
> > > +
> > > +       return trace_handle_return(&iter->seq);
> > > +}
> > > +
> > > +static struct trace_event_functions trace_object_funcs = {
> > > +       .trace          = trace_object_print,
> > > +       .raw            = trace_object_raw,
> > > +};
> > > +
> > > +static struct trace_event trace_object_event = {
> > > +       .type           = TRACE_OBJECT,
> > > +       .funcs          = &trace_object_funcs,
> > > +};
> > > +
> > >  static struct trace_event *events[] __initdata = {
> > >         &trace_fn_event,
> > >         &trace_ctx_event,
> > > @@ -1561,6 +1600,7 @@ static struct trace_event *events[]
> > > __initdata = {
> > >         &trace_timerlat_event,
> > >         &trace_raw_data_event,
> > >         &trace_func_repeats_event,
> > > +       &trace_object_event,
> > >         NULL
> > >  };
> > >
> > > --
> > > 2.25.1
> > >
> >
> > Thanks,
> > JeffXie
>

Thanks,
JeffXie

^ permalink raw reply	[flat|nested] 18+ messages in thread

end of thread, other threads:[~2022-05-04  7:43 UTC | newest]

Thread overview: 18+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2022-02-04  3:56 [PATCH v9 0/4] trace: Introduce objtrace trigger to trace the kernel object Jeff Xie
2022-02-04  3:56 ` [PATCH v9 1/4] trace: Add trace any " Jeff Xie
2022-04-19  2:14   ` Steven Rostedt
2022-04-19  2:26     ` Steven Rostedt
2022-04-19 16:26     ` Jeff Xie
2022-04-27 15:50   ` Jeff Xie
2022-05-02 17:42     ` Tom Zanussi
2022-05-04  7:42       ` Jeff Xie
2022-02-04  3:56 ` [PATCH v9 2/4] trace/objtrace: get the value of the object Jeff Xie
2022-02-04  3:56 ` [PATCH v9 3/4] trace/objtrace: Add testcases for objtrace Jeff Xie
2022-02-04  3:56 ` [PATCH v9 4/4] trace/objtrace: Add documentation " Jeff Xie
2022-02-08 14:08 ` [PATCH v9 0/4] trace: Introduce objtrace trigger to trace the kernel object Masami Hiramatsu
2022-02-08 15:48   ` Steven Rostedt
2022-02-26 16:01     ` Jeff Xie
2022-02-28 16:08       ` Steven Rostedt
2022-03-22 17:20         ` Jeff Xie
2022-04-11 15:47           ` Jeff Xie
2022-04-14 16:52             ` Steven Rostedt

This is an external index of several public inboxes,
see mirroring instructions on how to clone and mirror
all data and code used by this external index.