From: Mathieu Desnoyers <mathieu.desnoyers@polymtl.ca>
To: Steven Rostedt <rostedt@goodmis.org>
Cc: LKML <linux-kernel@vger.kernel.org>, Ingo Molnar <mingo@elte.hu>,
Linus Torvalds <torvalds@linux-foundation.org>,
Andrew Morton <akpm@linux-foundation.org>,
Peter Zijlstra <a.p.zijlstra@chello.nl>,
Christoph Hellwig <hch@infradead.org>,
Gregory Haskins <ghaskins@novell.com>,
Arnaldo Carvalho de Melo <acme@ghostprotocols.net>,
Thomas Gleixner <tglx@linutronix.de>,
Tim Bird <tim.bird@am.sony.com>, Sam Ravnborg <sam@ravnborg.org>,
"Frank Ch. Eigler" <fche@redhat.com>,
Jan Kiszka <jan.kiszka@siemens.com>,
Steven Rostedt <srostedt@redhat.com>
Subject: Re: [RFC PATCH 28/30 v3] Generic command line storage
Date: Tue, 15 Jan 2008 16:30:44 -0500 [thread overview]
Message-ID: <20080115213044.GC17439@Krystal> (raw)
In-Reply-To: <20080115205026.559348520@goodmis.org>
* Steven Rostedt (rostedt@goodmis.org) wrote:
> Saving the comm of tasks for each trace is very expensive.
> This patch includes in the context switch hook, a way to
> store the last 100 command lines of tasks. This table is
> examined when a trace is to be printed.
>
Instead of saving the comm at context switch, could we save them when a
process exits ? (actually, if we want to do this right, we would also
have to trace the changes of "comm" upon exec(), so we deal correctly
with processes that do multiple execs).
This way, we would not duplicate the comm of processes still active in
the system, and would trigger on a much lower event-rate trace point.
And the idea would be to save only the comm of processes present in your
summary (top X processes); why do you save the last 100 ? (if the system
has more than 100 active tasks, this code will always be executed, which
seems bad)
Also, if we do as I propose, we would have to keep process exit/exec
tracing active until the summary is cleared, even though the rest of
"high event rate" tracing could be stopped sooner.
Mathieu
> Note: The comm may be destroyed if other traces are performed.
> Later (TBD) patches may simply store this information in the trace
> itself.
>
> Signed-off-by: Steven Rostedt <srostedt@redhat.com>
> ---
> lib/tracing/Kconfig | 1
> lib/tracing/trace_function.c | 2
> lib/tracing/trace_irqsoff.c | 3 +
> lib/tracing/trace_sched_switch.c | 14 +++-
> lib/tracing/tracer.c | 112 +++++++++++++++++++++++++++++++++++++--
> lib/tracing/tracer.h | 5 +
> 6 files changed, 128 insertions(+), 9 deletions(-)
>
> Index: linux-compile.git/lib/tracing/Kconfig
> ===================================================================
> --- linux-compile.git.orig/lib/tracing/Kconfig 2008-01-15 10:34:22.000000000 -0500
> +++ linux-compile.git/lib/tracing/Kconfig 2008-01-15 10:41:28.000000000 -0500
> @@ -19,6 +19,7 @@ config FUNCTION_TRACER
> default n
> select MCOUNT
> select TRACING
> + select CONTEXT_SWITCH_TRACER
> help
> Use profiler instrumentation, adding -pg to CFLAGS. This will
> insert a call to an architecture specific __mcount routine,
> Index: linux-compile.git/lib/tracing/trace_function.c
> ===================================================================
> --- linux-compile.git.orig/lib/tracing/trace_function.c 2008-01-15 10:26:28.000000000 -0500
> +++ linux-compile.git/lib/tracing/trace_function.c 2008-01-15 10:41:28.000000000 -0500
> @@ -70,9 +70,11 @@ static void function_trace_ctrl_update(s
> if (tr->ctrl ^ val) {
> if (val) {
> trace_enabled = 1;
> + atomic_inc(&trace_record_cmdline);
> register_mcount_function(&trace_ops);
> } else {
> trace_enabled = 0;
> + atomic_dec(&trace_record_cmdline);
> unregister_mcount_function(&trace_ops);
> }
> tr->ctrl = val;
> Index: linux-compile.git/lib/tracing/trace_irqsoff.c
> ===================================================================
> --- linux-compile.git.orig/lib/tracing/trace_irqsoff.c 2008-01-15 10:27:32.000000000 -0500
> +++ linux-compile.git/lib/tracing/trace_irqsoff.c 2008-01-15 10:41:28.000000000 -0500
> @@ -93,6 +93,9 @@ static void update_max_tr(struct tracing
> save->policy = current->policy;
> save->rt_priority = current->rt_priority;
>
> + /* record this tasks comm */
> + tracing_record_cmdline(current);
> +
> /* from memcpy above: save->trace = data->trace */
> data->trace = max_buffer;
> max_buffer = save->trace;
> Index: linux-compile.git/lib/tracing/trace_sched_switch.c
> ===================================================================
> --- linux-compile.git.orig/lib/tracing/trace_sched_switch.c 2008-01-15 10:37:11.000000000 -0500
> +++ linux-compile.git/lib/tracing/trace_sched_switch.c 2008-01-15 10:41:28.000000000 -0500
> @@ -31,7 +31,7 @@ static notrace void sched_switch_callbac
> va_list ap;
> int cpu;
>
> - if (!trace_enabled)
> + if (!atomic_read(&trace_record_cmdline))
> return;
>
> va_start(ap, format);
> @@ -49,6 +49,11 @@ static notrace void sched_switch_callbac
> /* Ignore prev_state, since we get that from prev itself */
> va_end(ap);
>
> + tracing_record_cmdline(prev);
> +
> + if (!trace_enabled)
> + return;
> +
> raw_local_irq_save(flags);
> cpu = raw_smp_processor_id();
> data = tr->data[cpu];
> @@ -82,10 +87,13 @@ static void sched_switch_trace_ctrl_upda
> sched_switch_reset(tr);
>
> if (tr->ctrl ^ val) {
> - if (val)
> + if (val) {
> + atomic_inc(&trace_record_cmdline);
> trace_enabled = 1;
> - else
> + } else {
> + atomic_dec(&trace_record_cmdline);
> trace_enabled = 0;
> + }
> tr->ctrl = val;
> }
> }
> Index: linux-compile.git/lib/tracing/tracer.c
> ===================================================================
> --- linux-compile.git.orig/lib/tracing/tracer.c 2008-01-15 10:37:38.000000000 -0500
> +++ linux-compile.git/lib/tracing/tracer.c 2008-01-15 10:42:46.000000000 -0500
> @@ -49,6 +49,88 @@ void notrace tracing_reset(struct tracin
> atomic_set(&data->underrun, 0);
> }
>
> +#define SAVED_CMDLINES 128
> +static unsigned map_pid_to_cmdline[PID_MAX_DEFAULT+1];
> +static unsigned map_cmdline_to_pid[SAVED_CMDLINES];
> +static char saved_cmdlines[SAVED_CMDLINES][TASK_COMM_LEN];
> +static int cmdline_idx;
> +static DEFINE_SPINLOCK(trace_cmdline_lock);
> +atomic_t trace_record_cmdline;
> +atomic_t trace_record_cmdline_disabled;
> +
> +static void trace_init_cmdlines(void)
> +{
> + memset(&map_pid_to_cmdline, -1, sizeof(map_pid_to_cmdline));
> + memset(&map_cmdline_to_pid, -1, sizeof(map_cmdline_to_pid));
> + cmdline_idx = 0;
> +}
> +
> +notrace void trace_stop_cmdline_recording(void);
> +
> +static void notrace trace_save_cmdline(struct task_struct *tsk)
> +{
> + unsigned map;
> + unsigned idx;
> +
> + if (!tsk->pid || unlikely(tsk->pid > PID_MAX_DEFAULT))
> + return;
> +
> + /*
> + * It's not the end of the world if we don't get
> + * the lock, but we also don't want to spin
> + * nor do we want to disable interrupts,
> + * so if we miss here, then better luck next time.
> + */
> + if (!spin_trylock(&trace_cmdline_lock))
> + return;
> +
> + idx = map_pid_to_cmdline[tsk->pid];
> + if (idx >= SAVED_CMDLINES) {
> + idx = (cmdline_idx + 1) % SAVED_CMDLINES;
> +
> + map = map_cmdline_to_pid[idx];
> + if (map <= PID_MAX_DEFAULT)
> + map_pid_to_cmdline[map] = (unsigned)-1;
> +
> + map_pid_to_cmdline[tsk->pid] = idx;
> +
> + cmdline_idx = idx;
> + }
> +
> + memcpy(&saved_cmdlines[idx], tsk->comm, TASK_COMM_LEN);
> +
> + spin_unlock(&trace_cmdline_lock);
> +}
> +
> +static notrace char *trace_find_cmdline(int pid)
> +{
> + char *cmdline = "<...>";
> + unsigned map;
> +
> + if (!pid)
> + return "<idle>";
> +
> + if (pid > PID_MAX_DEFAULT)
> + goto out;
> +
> + map = map_pid_to_cmdline[pid];
> + if (map >= SAVED_CMDLINES)
> + goto out;
> +
> + cmdline = saved_cmdlines[map];
> +
> + out:
> + return cmdline;
> +}
> +
> +void tracing_record_cmdline(struct task_struct *tsk)
> +{
> + if (atomic_read(&trace_record_cmdline_disabled))
> + return;
> +
> + trace_save_cmdline(tsk);
> +}
> +
> static inline notrace struct tracing_entry *
> tracing_get_trace_entry(struct tracing_trace *tr,
> struct tracing_trace_cpu *data)
> @@ -90,7 +172,6 @@ tracing_generic_entry_update(struct trac
> ((pc & HARDIRQ_MASK) ? TRACE_FLAG_HARDIRQ : 0) |
> ((pc & SOFTIRQ_MASK) ? TRACE_FLAG_SOFTIRQ : 0) |
> (need_resched() ? TRACE_FLAG_NEED_RESCHED : 0);
> - memcpy(entry->comm, tsk->comm, TASK_COMM_LEN);
> }
>
> notrace void tracing_function_trace(struct tracing_trace *tr,
> @@ -242,6 +323,8 @@ static void *s_start(struct seq_file *m,
> loff_t l = 0;
> int i;
>
> + atomic_inc(&trace_record_cmdline_disabled);
> +
> /* let the tracer grab locks here if needed */
> if (iter->tr->start)
> iter->tr->start(iter);
> @@ -269,6 +352,8 @@ static void s_stop(struct seq_file *m, v
> {
> struct tracing_iterator *iter = m->private;
>
> + atomic_dec(&trace_record_cmdline_disabled);
> +
> /* let the tracer release locks here if needed */
> if (iter->tr->stop)
> iter->tr->stop(iter);
> @@ -390,8 +475,11 @@ static void notrace
> lat_print_generic(struct seq_file *m, struct tracing_entry *entry, int cpu)
> {
> int hardirq, softirq;
> + char *comm;
>
> - seq_printf(m, "%8.8s-%-5d ", entry->comm, entry->pid);
> + comm = trace_find_cmdline(entry->pid);
> +
> + seq_printf(m, "%8.8s-%-5d ", comm, entry->pid);
> seq_printf(m, "%d", cpu);
> seq_printf(m, "%c%c",
> (entry->flags & TRACE_FLAG_IRQS_OFF) ? 'd' : '.',
> @@ -453,9 +541,12 @@ print_lat_fmt(struct seq_file *m, struct
> abs_usecs = cycles_to_usecs(entry->t - iter->tr->time_start);
>
> if (verbose) {
> + char *comm;
> +
> + comm = trace_find_cmdline(entry->pid);
> seq_printf(m, "%16s %5d %d %d %08x %08x [%08lx]"
> " %ld.%03ldms (+%ld.%03ldms): ",
> - entry->comm,
> + comm,
> entry->pid, cpu, entry->flags,
> entry->preempt_count, trace_idx,
> cycles_to_usecs(entry->t),
> @@ -491,6 +582,9 @@ static void notrace print_trace_fmt(stru
> unsigned long secs;
> int sym_only = !!(trace_flags & TRACE_ITER_SYM_ONLY);
> unsigned long long t;
> + char *comm;
> +
> + comm = trace_find_cmdline(iter->ent->pid);
>
> t = cycles_to_usecs(iter->ent->t);
> usec_rem = do_div(t, 1000000ULL);
> @@ -498,7 +592,7 @@ static void notrace print_trace_fmt(stru
>
> seq_printf(m, "[%5lu.%06lu] ", secs, usec_rem);
> seq_printf(m, "CPU %d: ", iter->cpu);
> - seq_printf(m, "%s:%d ", iter->ent->comm,
> + seq_printf(m, "%s:%d ", comm,
> iter->ent->pid);
> switch (iter->ent->type) {
> case TRACE_FN:
> @@ -812,6 +906,14 @@ static __init int trace_init_debugfs(voi
> return 0;
> }
>
> -device_initcall(trace_init_debugfs);
> +static __init int trace_init(void)
> +{
> + trace_init_cmdlines();
> +
> + return trace_init_debugfs();
> +
> +}
> +
> +device_initcall(trace_init);
>
> #endif /* CONFIG_DEBUG_FS */
> Index: linux-compile.git/lib/tracing/tracer.h
> ===================================================================
> --- linux-compile.git.orig/lib/tracing/tracer.h 2008-01-15 10:34:22.000000000 -0500
> +++ linux-compile.git/lib/tracing/tracer.h 2008-01-15 10:41:28.000000000 -0500
> @@ -25,7 +25,6 @@ struct tracing_entry {
> char preempt_count; /* assumes PREEMPT_MASK is 8 bits or less */
> int pid;
> cycle_t t;
> - char comm[TASK_COMM_LEN];
> union {
> struct tracing_function fn;
> struct tracing_sched_switch ctx;
> @@ -92,11 +91,15 @@ void tracing_sched_switch_trace(struct t
> struct task_struct *prev,
> int next_pid,
> unsigned long flags);
> +void tracing_record_cmdline(struct task_struct *tsk);
>
> extern struct file_operations tracing_fops;
> extern struct file_operations tracing_lt_fops;
> extern struct file_operations tracing_ctrl_fops;
>
> +extern atomic_t trace_record_cmdline;
> +extern atomic_t trace_record_cmdline_disabled;
> +
> static inline notrace cycle_t now(void)
> {
> return get_monotonic_cycles();
>
> --
--
Mathieu Desnoyers
Computer Engineering Ph.D. Student, Ecole Polytechnique de Montreal
OpenPGP key fingerprint: 8CD5 52C3 8E3C 4140 715F BA06 3F25 A8FE 3BAE 9A68
next prev parent reply other threads:[~2008-01-15 21:30 UTC|newest]
Thread overview: 35+ messages / expand[flat|nested] mbox.gz Atom feed top
2008-01-15 20:49 [RFC PATCH 00/30 v3] mcount and latency tracing utility -v3 Steven Rostedt
2008-01-15 20:49 ` [RFC PATCH 01/30 v3] Add basic support for gcc profiler instrumentation Steven Rostedt
2008-01-15 20:49 ` [RFC PATCH 02/30 v3] Annotate core code that should not be traced Steven Rostedt
2008-01-15 20:49 ` [RFC PATCH 03/30 v3] x86_64: notrace annotations Steven Rostedt
2008-01-15 20:49 ` [RFC PATCH 04/30 v3] add notrace annotations to vsyscall Steven Rostedt
2008-01-15 20:49 ` [RFC PATCH 05/30 v3] add notrace annotations for NMI routines Steven Rostedt
2008-01-15 20:49 ` [RFC PATCH 06/30 v3] mcount based trace in the form of a header file library Steven Rostedt
2008-01-15 20:49 ` [RFC PATCH 07/30 v3] tracer add debugfs interface Steven Rostedt
2008-01-15 20:49 ` [RFC PATCH 08/30 v3] mcount tracer output file Steven Rostedt
2008-01-15 20:49 ` [RFC PATCH 09/30 v3] mcount tracer show task comm and pid Steven Rostedt
2008-01-15 20:49 ` [RFC PATCH 10/30 v3] Add a symbol only trace output Steven Rostedt
2008-01-15 20:49 ` [RFC PATCH 11/30 v3] Reset the tracer when started Steven Rostedt
2008-01-15 20:49 ` [RFC PATCH 12/30 v3] separate out the percpu date into a percpu struct Steven Rostedt
2008-01-15 20:49 ` [RFC PATCH 13/30 v3] handle accurate time keeping over long delays Steven Rostedt
2008-01-15 20:49 ` [RFC PATCH 14/30 v3] ppc clock accumulate fix Steven Rostedt
2008-01-15 20:49 ` [RFC PATCH 15/30 v3] Fixup merge between xtime_cache and timkkeeping starvation fix Steven Rostedt
2008-01-15 20:49 ` [RFC PATCH 16/30 v3] time keeping add cycle_raw for actual incrementation Steven Rostedt
2008-01-15 20:49 ` [RFC PATCH 17/30 v3] initialize the clock source to jiffies clock Steven Rostedt
2008-01-15 21:14 ` Mathieu Desnoyers
2008-01-15 21:27 ` Steven Rostedt
2008-01-15 20:49 ` [RFC PATCH 18/30 v3] add get_monotonic_cycles Steven Rostedt
2008-01-15 20:49 ` [RFC PATCH 19/30 v3] add notrace annotations to timing events Steven Rostedt
2008-01-15 20:49 ` [RFC PATCH 20/30 v3] Add timestamps to tracer Steven Rostedt
2008-01-15 20:49 ` [RFC PATCH 21/30 v3] Sort trace by timestamp Steven Rostedt
2008-01-15 20:49 ` [RFC PATCH 22/30 v3] speed up the output of the tracer Steven Rostedt
2008-01-15 20:49 ` [RFC PATCH 23/30 v3] Add latency_trace format tor tracer Steven Rostedt
2008-01-15 20:49 ` [RFC PATCH 24/30 v3] Split out specific tracing functions Steven Rostedt
2008-01-15 20:49 ` [RFC PATCH 25/30 v3] Trace irq disabled critical timings Steven Rostedt
2008-01-15 20:49 ` [RFC PATCH 26/30 v3] Add context switch marker to sched.c Steven Rostedt
2008-01-15 20:49 ` [RFC PATCH 27/30 v3] Add tracing of context switches Steven Rostedt
2008-01-15 20:49 ` [RFC PATCH 28/30 v3] Generic command line storage Steven Rostedt
2008-01-15 21:30 ` Mathieu Desnoyers [this message]
2008-01-15 22:15 ` Steven Rostedt
2008-01-15 20:49 ` [RFC PATCH 29/30 v3] make varaible size buffers for traces Steven Rostedt
2008-01-15 20:49 ` [RFC PATCH 30/30 v3] trace preempt off critical timings Steven Rostedt
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20080115213044.GC17439@Krystal \
--to=mathieu.desnoyers@polymtl.ca \
--cc=a.p.zijlstra@chello.nl \
--cc=acme@ghostprotocols.net \
--cc=akpm@linux-foundation.org \
--cc=fche@redhat.com \
--cc=ghaskins@novell.com \
--cc=hch@infradead.org \
--cc=jan.kiszka@siemens.com \
--cc=linux-kernel@vger.kernel.org \
--cc=mingo@elte.hu \
--cc=rostedt@goodmis.org \
--cc=sam@ravnborg.org \
--cc=srostedt@redhat.com \
--cc=tglx@linutronix.de \
--cc=tim.bird@am.sony.com \
--cc=torvalds@linux-foundation.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox