* [PATCH 1/2] tracing/ftrace: handle more than one stat file per tracer
@ 2008-12-31 15:43 Frederic Weisbecker
2009-01-05 20:53 ` Steven Rostedt
0 siblings, 1 reply; 4+ messages in thread
From: Frederic Weisbecker @ 2008-12-31 15:43 UTC (permalink / raw)
To: Ingo Molnar
Cc: Linux Kernel Mailing List, Steven Rostedt, Pekka Enberg,
Eduard - Gabriel Munteanu, fweisbec
Impact: new API for tracers
Make the stat tracing API reentrant. And also provide the new directory
/debugfs/tracing/trace_stat which will contain all the stat files for the
current active tracer.
Now a tracer will, if desired, want to provide a zero terminated array of
tracer_stat structures.
Each one contains the callbacks necessary for one stat file.
It have to provide at least a name for its stat file, an iterator with
stat_start/start_next callback and an output callback for one stat entry.
Signed-off-by: Frederic Weisbecker <fweisbec@gmail.com>
---
diff --git a/kernel/trace/trace.h b/kernel/trace/trace.h
index 46677ea..b093f36 100644
--- a/kernel/trace/trace.h
+++ b/kernel/trace/trace.h
@@ -335,6 +335,25 @@ struct tracer_flags {
#define TRACER_OPT(s, b) .name = #s, .bit = b
/*
+ * If you want to provide a stat file (one-shot statistics), fill
+ * an iterator with stat_start/stat_next and a stat_show callbacks.
+ * The others callbacks are optional.
+ */
+struct tracer_stat {
+ /* The name of your stat file */
+ const char *name;
+ /* Iteration over statistic entries */
+ void *(*stat_start)(void);
+ void *(*stat_next)(void *prev, int idx);
+ /* Compare two entries for sorting (optional) for stats */
+ int (*stat_cmp)(void *p1, void *p2);
+ /* Print a stat entry */
+ int (*stat_show)(struct seq_file *s, void *p);
+ /* Print the headers of your stat entries */
+ int (*stat_headers)(struct seq_file *s);
+};
+
+/*
* A specific tracer, represented by methods that operate on a trace array:
*/
struct tracer {
@@ -361,21 +380,7 @@ struct tracer {
struct tracer *next;
int print_max;
struct tracer_flags *flags;
-
- /*
- * If you change one of the following on tracing runtime, recall
- * init_tracer_stat()
- */
-
- /* Iteration over statistic entries */
- void *(*stat_start)(void);
- void *(*stat_next)(void *prev, int idx);
- /* Compare two entries for sorting (optional) for stats */
- int (*stat_cmp)(void *p1, void *p2);
- /* Print a stat entry */
- int (*stat_show)(struct seq_file *s, void *p);
- /* Print the headers of your stat entries */
- int (*stat_headers)(struct seq_file *s);
+ struct tracer_stat *stats;
};
struct trace_seq {
diff --git a/kernel/trace/trace_stat.c b/kernel/trace/trace_stat.c
index 6f194a3..293be1c 100644
--- a/kernel/trace/trace_stat.c
+++ b/kernel/trace/trace_stat.c
@@ -21,31 +21,29 @@ struct trace_stat_list {
void *stat;
};
-static struct trace_stat_list stat_list;
+/* A stat session is the stats output in one file */
+struct tracer_stat_session {
+ struct tracer_stat *ts;
+ struct trace_stat_list stat_list;
+ struct mutex stat_mutex;
+};
-/*
- * This is a copy of the current tracer to avoid racy
- * and dangerous output while the current tracer is
- * switched.
- */
-static struct tracer current_tracer;
-
-/*
- * Protect both the current tracer and the global
- * stat list.
- */
-static DEFINE_MUTEX(stat_list_mutex);
+/* All of the sessions currently in use. Each stat file embeed one session */
+static struct tracer_stat_session **all_stat_sessions;
+static int nb_sessions;
+static struct dentry *stat_dir, **stat_files;
-static void reset_stat_list(void)
+static void reset_stat_session(struct tracer_stat_session *session)
{
+ struct trace_stat_list *root = &session->stat_list;
struct trace_stat_list *node;
struct list_head *next;
- if (list_empty(&stat_list.list))
+ if (list_empty(&root->list))
return;
- node = list_entry(stat_list.list.next, struct trace_stat_list, list);
+ node = list_entry(root->list.next, struct trace_stat_list, list);
next = node->list.next;
while (&node->list != next) {
@@ -54,14 +52,65 @@ static void reset_stat_list(void)
}
kfree(node);
- INIT_LIST_HEAD(&stat_list.list);
+ INIT_LIST_HEAD(&root->list);
}
-void init_tracer_stat(struct tracer *trace)
+/* Called when a tracer is initialized */
+static int init_all_sessions(int nb, struct tracer_stat *ts)
+{
+ int i, j;
+ struct tracer_stat_session *session;
+
+ if (all_stat_sessions) {
+ for (i = 0; i < nb_sessions; i++) {
+ session = all_stat_sessions[i];
+ reset_stat_session(session);
+ mutex_destroy(&session->stat_mutex);
+ kfree(session);
+ }
+ }
+ all_stat_sessions = kmalloc(sizeof(struct tracer_stat_session *) * nb,
+ GFP_KERNEL);
+ if (!all_stat_sessions)
+ return -ENOMEM;
+
+ for (i = 0; i < nb; i++) {
+ session = kmalloc(sizeof(struct tracer_stat_session) * nb,
+ GFP_KERNEL);
+ if (!session)
+ goto free_sessions;
+
+ INIT_LIST_HEAD(&session->stat_list.list);
+ mutex_init(&session->stat_mutex);
+ session->ts = &ts[i];
+ all_stat_sessions[i] = session;
+ }
+ nb_sessions = nb;
+ return 0;
+
+free_sessions:
+
+ for (j = 0; j < i; j++)
+ kfree(all_stat_sessions[i]);
+
+ kfree(all_stat_sessions);
+ all_stat_sessions = NULL;
+
+ return -ENOMEM;
+}
+
+static int basic_tracer_stat_checks(struct tracer_stat *ts)
{
- mutex_lock(&stat_list_mutex);
- current_tracer = *trace;
- mutex_unlock(&stat_list_mutex);
+ int i;
+
+ if (!ts)
+ return 0;
+
+ for (i = 0; ts[i].name; i++) {
+ if (!ts[i].stat_start || !ts[i].stat_next || !ts[i].stat_show)
+ return -EBUSY;
+ }
+ return i;
}
/*
@@ -79,22 +128,19 @@ static int dummy_cmp(void *p1, void *p2)
* All of these copies and sorting are required on all opening
* since the stats could have changed between two file sessions.
*/
-static int stat_seq_init(void)
+static int stat_seq_init(struct tracer_stat_session *session)
{
- struct trace_stat_list *iter_entry, *new_entry;
+ struct trace_stat_list *iter_entry, *new_entry, *root;
+ struct tracer_stat *ts = session->ts;
void *prev_stat;
int ret = 0;
int i;
- mutex_lock(&stat_list_mutex);
- reset_stat_list();
-
- if (!current_tracer.stat_start || !current_tracer.stat_next ||
- !current_tracer.stat_show)
- goto exit;
+ mutex_lock(&session->stat_mutex);
+ reset_stat_session(session);
- if (!current_tracer.stat_cmp)
- current_tracer.stat_cmp = dummy_cmp;
+ if (!ts->stat_cmp)
+ ts->stat_cmp = dummy_cmp;
/*
* The first entry. Actually this is the second, but the first
@@ -107,8 +153,9 @@ static int stat_seq_init(void)
}
INIT_LIST_HEAD(&new_entry->list);
- list_add(&new_entry->list, &stat_list.list);
- new_entry->stat = current_tracer.stat_start();
+ root = &session->stat_list;
+ list_add(&new_entry->list, &root->list);
+ new_entry->stat = ts->stat_start();
prev_stat = new_entry->stat;
@@ -124,15 +171,15 @@ static int stat_seq_init(void)
}
INIT_LIST_HEAD(&new_entry->list);
- new_entry->stat = current_tracer.stat_next(prev_stat, i);
+ new_entry->stat = ts->stat_next(prev_stat, i);
/* End of insertion */
if (!new_entry->stat)
break;
- list_for_each_entry(iter_entry, &stat_list.list, list) {
+ list_for_each_entry(iter_entry, &root->list, list) {
/* Insertion with a descendent sorting */
- if (current_tracer.stat_cmp(new_entry->stat,
+ if (ts->stat_cmp(new_entry->stat,
iter_entry->stat) > 0) {
list_add_tail(&new_entry->list,
@@ -141,7 +188,7 @@ static int stat_seq_init(void)
/* The current smaller value */
} else if (list_is_last(&iter_entry->list,
- &stat_list.list)) {
+ &root->list)) {
list_add(&new_entry->list, &iter_entry->list);
break;
}
@@ -150,46 +197,49 @@ static int stat_seq_init(void)
prev_stat = new_entry->stat;
}
exit:
- mutex_unlock(&stat_list_mutex);
+ mutex_unlock(&session->stat_mutex);
return ret;
exit_free_list:
- reset_stat_list();
- mutex_unlock(&stat_list_mutex);
+ reset_stat_session(session);
+ mutex_unlock(&session->stat_mutex);
return ret;
}
static void *stat_seq_start(struct seq_file *s, loff_t *pos)
{
- struct trace_stat_list *l = (struct trace_stat_list *)s->private;
+ struct tracer_stat_session *session = s->private;
/* Prevent from tracer switch or stat_list modification */
- mutex_lock(&stat_list_mutex);
+ mutex_lock(&session->stat_mutex);
/* If we are in the beginning of the file, print the headers */
- if (!*pos && current_tracer.stat_headers)
- current_tracer.stat_headers(s);
+ if (!*pos && session->ts->stat_headers)
+ session->ts->stat_headers(s);
- return seq_list_start(&l->list, *pos);
+ return seq_list_start(&session->stat_list.list, *pos);
}
static void *stat_seq_next(struct seq_file *s, void *p, loff_t *pos)
{
- struct trace_stat_list *l = (struct trace_stat_list *)s->private;
+ struct tracer_stat_session *session = s->private;
- return seq_list_next(p, &l->list, pos);
+ return seq_list_next(p, &session->stat_list.list, pos);
}
-static void stat_seq_stop(struct seq_file *m, void *p)
+static void stat_seq_stop(struct seq_file *s, void *p)
{
- mutex_unlock(&stat_list_mutex);
+ struct tracer_stat_session *session = s->private;
+ mutex_unlock(&session->stat_mutex);
}
static int stat_seq_show(struct seq_file *s, void *v)
{
+ struct tracer_stat_session *session = s->private;
struct trace_stat_list *l = list_entry(v, struct trace_stat_list, list);
- return current_tracer.stat_show(s, l->stat);
+
+ return session->ts->stat_show(s, l->stat);
}
static const struct seq_operations trace_stat_seq_ops = {
@@ -199,15 +249,18 @@ static const struct seq_operations trace_stat_seq_ops = {
.show = stat_seq_show
};
+/* The session stat is refilled and resorted at each stat file opening */
static int tracing_stat_open(struct inode *inode, struct file *file)
{
int ret;
+ struct tracer_stat_session *session = inode->i_private;
+
ret = seq_open(file, &trace_stat_seq_ops);
if (!ret) {
struct seq_file *m = file->private_data;
- m->private = &stat_list;
- ret = stat_seq_init();
+ m->private = session;
+ ret = stat_seq_init(session);
}
return ret;
@@ -219,9 +272,11 @@ static int tracing_stat_open(struct inode *inode, struct file *file)
*/
static int tracing_stat_release(struct inode *i, struct file *f)
{
- mutex_lock(&stat_list_mutex);
- reset_stat_list();
- mutex_unlock(&stat_list_mutex);
+ struct tracer_stat_session *session = i->i_private;
+
+ mutex_lock(&session->stat_mutex);
+ reset_stat_session(session);
+ mutex_unlock(&session->stat_mutex);
return 0;
}
@@ -232,18 +287,70 @@ static const struct file_operations tracing_stat_fops = {
.release = tracing_stat_release
};
+
+static void destroy_trace_stat_files(void)
+{
+ int i;
+
+ if (stat_files) {
+ for (i = 0; i < nb_sessions; i++)
+ debugfs_remove(stat_files[i]);
+ kfree(stat_files);
+ stat_files = NULL;
+ }
+}
+
+static void init_trace_stat_files(void)
+{
+ int i;
+
+ if (!stat_dir || !nb_sessions)
+ return;
+
+ stat_files = kmalloc(sizeof(struct dentry *) * nb_sessions, GFP_KERNEL);
+
+ if (!stat_files) {
+ pr_warning("trace stat: not enough memory\n");
+ return;
+ }
+
+ for (i = 0; i < nb_sessions; i++) {
+ struct tracer_stat_session *session = all_stat_sessions[i];
+ stat_files[i] = debugfs_create_file(session->ts->name, 0644,
+ stat_dir,
+ session, &tracing_stat_fops);
+ if (!stat_files[i])
+ pr_warning("cannot create %s entry\n",
+ session->ts->name);
+ }
+}
+
+void init_tracer_stat(struct tracer *trace)
+{
+ int nb = basic_tracer_stat_checks(trace->stats);
+
+ destroy_trace_stat_files();
+
+ if (nb < 0) {
+ pr_warning("stat tracing: missing stat callback on %s\n",
+ trace->name);
+ return;
+ }
+ if (!nb)
+ return;
+
+ init_all_sessions(nb, trace->stats);
+ init_trace_stat_files();
+}
+
static int __init tracing_stat_init(void)
{
struct dentry *d_tracing;
- struct dentry *entry;
- INIT_LIST_HEAD(&stat_list.list);
d_tracing = tracing_init_dentry();
- entry = debugfs_create_file("trace_stat", 0444, d_tracing,
- NULL,
- &tracing_stat_fops);
- if (!entry)
+ stat_dir = debugfs_create_dir("trace_stat", d_tracing);
+ if (!stat_dir)
pr_warning("Could not create debugfs "
"'trace_stat' entry\n");
return 0;
--
1.6.0.4
^ permalink raw reply related [flat|nested] 4+ messages in thread
* Re: [PATCH 1/2] tracing/ftrace: handle more than one stat file per tracer
2008-12-31 15:43 [PATCH 1/2] tracing/ftrace: handle more than one stat file per tracer Frederic Weisbecker
@ 2009-01-05 20:53 ` Steven Rostedt
2009-01-06 0:30 ` Frederic Weisbecker
0 siblings, 1 reply; 4+ messages in thread
From: Steven Rostedt @ 2009-01-05 20:53 UTC (permalink / raw)
To: Frederic Weisbecker
Cc: Ingo Molnar, Linux Kernel Mailing List, Pekka Enberg,
Eduard - Gabriel Munteanu
On Wed, 31 Dec 2008, Frederic Weisbecker wrote:
> Impact: new API for tracers
>
> Make the stat tracing API reentrant. And also provide the new directory
> /debugfs/tracing/trace_stat which will contain all the stat files for the
> current active tracer.
>
> Now a tracer will, if desired, want to provide a zero terminated array of
> tracer_stat structures.
> Each one contains the callbacks necessary for one stat file.
> It have to provide at least a name for its stat file, an iterator with
> stat_start/start_next callback and an output callback for one stat entry.
>
> Signed-off-by: Frederic Weisbecker <fweisbec@gmail.com>
> ---
> diff --git a/kernel/trace/trace.h b/kernel/trace/trace.h
> index 46677ea..b093f36 100644
> --- a/kernel/trace/trace.h
> +++ b/kernel/trace/trace.h
> @@ -335,6 +335,25 @@ struct tracer_flags {
> #define TRACER_OPT(s, b) .name = #s, .bit = b
>
> /*
> + * If you want to provide a stat file (one-shot statistics), fill
> + * an iterator with stat_start/stat_next and a stat_show callbacks.
> + * The others callbacks are optional.
> + */
> +struct tracer_stat {
> + /* The name of your stat file */
> + const char *name;
> + /* Iteration over statistic entries */
> + void *(*stat_start)(void);
> + void *(*stat_next)(void *prev, int idx);
> + /* Compare two entries for sorting (optional) for stats */
> + int (*stat_cmp)(void *p1, void *p2);
> + /* Print a stat entry */
> + int (*stat_show)(struct seq_file *s, void *p);
> + /* Print the headers of your stat entries */
> + int (*stat_headers)(struct seq_file *s);
> +};
> +
> +/*
> * A specific tracer, represented by methods that operate on a trace array:
> */
> struct tracer {
> @@ -361,21 +380,7 @@ struct tracer {
> struct tracer *next;
> int print_max;
> struct tracer_flags *flags;
> -
> - /*
> - * If you change one of the following on tracing runtime, recall
> - * init_tracer_stat()
> - */
> -
> - /* Iteration over statistic entries */
> - void *(*stat_start)(void);
> - void *(*stat_next)(void *prev, int idx);
> - /* Compare two entries for sorting (optional) for stats */
> - int (*stat_cmp)(void *p1, void *p2);
> - /* Print a stat entry */
> - int (*stat_show)(struct seq_file *s, void *p);
> - /* Print the headers of your stat entries */
> - int (*stat_headers)(struct seq_file *s);
> + struct tracer_stat *stats;
> };
>
> struct trace_seq {
> diff --git a/kernel/trace/trace_stat.c b/kernel/trace/trace_stat.c
> index 6f194a3..293be1c 100644
> --- a/kernel/trace/trace_stat.c
> +++ b/kernel/trace/trace_stat.c
> @@ -21,31 +21,29 @@ struct trace_stat_list {
> void *stat;
> };
>
> -static struct trace_stat_list stat_list;
> +/* A stat session is the stats output in one file */
> +struct tracer_stat_session {
> + struct tracer_stat *ts;
> + struct trace_stat_list stat_list;
> + struct mutex stat_mutex;
> +};
>
> -/*
> - * This is a copy of the current tracer to avoid racy
> - * and dangerous output while the current tracer is
> - * switched.
> - */
> -static struct tracer current_tracer;
> -
> -/*
> - * Protect both the current tracer and the global
> - * stat list.
> - */
> -static DEFINE_MUTEX(stat_list_mutex);
> +/* All of the sessions currently in use. Each stat file embeed one session */
> +static struct tracer_stat_session **all_stat_sessions;
> +static int nb_sessions;
> +static struct dentry *stat_dir, **stat_files;
>
>
> -static void reset_stat_list(void)
> +static void reset_stat_session(struct tracer_stat_session *session)
> {
> + struct trace_stat_list *root = &session->stat_list;
> struct trace_stat_list *node;
> struct list_head *next;
>
> - if (list_empty(&stat_list.list))
> + if (list_empty(&root->list))
> return;
>
> - node = list_entry(stat_list.list.next, struct trace_stat_list, list);
> + node = list_entry(root->list.next, struct trace_stat_list, list);
> next = node->list.next;
>
> while (&node->list != next) {
> @@ -54,14 +52,65 @@ static void reset_stat_list(void)
> }
> kfree(node);
>
> - INIT_LIST_HEAD(&stat_list.list);
> + INIT_LIST_HEAD(&root->list);
> }
>
> -void init_tracer_stat(struct tracer *trace)
> +/* Called when a tracer is initialized */
> +static int init_all_sessions(int nb, struct tracer_stat *ts)
> +{
> + int i, j;
> + struct tracer_stat_session *session;
> +
> + if (all_stat_sessions) {
> + for (i = 0; i < nb_sessions; i++) {
> + session = all_stat_sessions[i];
> + reset_stat_session(session);
> + mutex_destroy(&session->stat_mutex);
> + kfree(session);
> + }
> + }
> + all_stat_sessions = kmalloc(sizeof(struct tracer_stat_session *) * nb,
> + GFP_KERNEL);
You want this to be kzalloc.
> + if (!all_stat_sessions)
> + return -ENOMEM;
> +
> + for (i = 0; i < nb; i++) {
> + session = kmalloc(sizeof(struct tracer_stat_session) * nb,
> + GFP_KERNEL);
Not so important, but you probably want to do it here too.
> + if (!session)
> + goto free_sessions;
> +
> + INIT_LIST_HEAD(&session->stat_list.list);
> + mutex_init(&session->stat_mutex);
> + session->ts = &ts[i];
> + all_stat_sessions[i] = session;
> + }
> + nb_sessions = nb;
> + return 0;
> +
> +free_sessions:
> +
> + for (j = 0; j < i; j++)
> + kfree(all_stat_sessions[i]);
With out the kzalloc above, this may crash the kernel, because
all_stat_sessions[i] might not be NULL when they were not initialized.
> +
> + kfree(all_stat_sessions);
> + all_stat_sessions = NULL;
> +
> + return -ENOMEM;
> +}
> +
> +static int basic_tracer_stat_checks(struct tracer_stat *ts)
> {
> - mutex_lock(&stat_list_mutex);
> - current_tracer = *trace;
> - mutex_unlock(&stat_list_mutex);
> + int i;
> +
> + if (!ts)
> + return 0;
> +
> + for (i = 0; ts[i].name; i++) {
> + if (!ts[i].stat_start || !ts[i].stat_next || !ts[i].stat_show)
> + return -EBUSY;
> + }
> + return i;
> }
>
> /*
> @@ -79,22 +128,19 @@ static int dummy_cmp(void *p1, void *p2)
> * All of these copies and sorting are required on all opening
> * since the stats could have changed between two file sessions.
> */
> -static int stat_seq_init(void)
> +static int stat_seq_init(struct tracer_stat_session *session)
> {
> - struct trace_stat_list *iter_entry, *new_entry;
> + struct trace_stat_list *iter_entry, *new_entry, *root;
> + struct tracer_stat *ts = session->ts;
> void *prev_stat;
> int ret = 0;
> int i;
>
> - mutex_lock(&stat_list_mutex);
> - reset_stat_list();
> -
> - if (!current_tracer.stat_start || !current_tracer.stat_next ||
> - !current_tracer.stat_show)
> - goto exit;
> + mutex_lock(&session->stat_mutex);
> + reset_stat_session(session);
>
> - if (!current_tracer.stat_cmp)
> - current_tracer.stat_cmp = dummy_cmp;
> + if (!ts->stat_cmp)
> + ts->stat_cmp = dummy_cmp;
>
> /*
> * The first entry. Actually this is the second, but the first
> @@ -107,8 +153,9 @@ static int stat_seq_init(void)
> }
>
> INIT_LIST_HEAD(&new_entry->list);
> - list_add(&new_entry->list, &stat_list.list);
> - new_entry->stat = current_tracer.stat_start();
> + root = &session->stat_list;
> + list_add(&new_entry->list, &root->list);
> + new_entry->stat = ts->stat_start();
>
> prev_stat = new_entry->stat;
>
> @@ -124,15 +171,15 @@ static int stat_seq_init(void)
> }
>
> INIT_LIST_HEAD(&new_entry->list);
> - new_entry->stat = current_tracer.stat_next(prev_stat, i);
> + new_entry->stat = ts->stat_next(prev_stat, i);
>
> /* End of insertion */
> if (!new_entry->stat)
> break;
>
> - list_for_each_entry(iter_entry, &stat_list.list, list) {
> + list_for_each_entry(iter_entry, &root->list, list) {
> /* Insertion with a descendent sorting */
> - if (current_tracer.stat_cmp(new_entry->stat,
> + if (ts->stat_cmp(new_entry->stat,
> iter_entry->stat) > 0) {
>
> list_add_tail(&new_entry->list,
> @@ -141,7 +188,7 @@ static int stat_seq_init(void)
>
> /* The current smaller value */
> } else if (list_is_last(&iter_entry->list,
> - &stat_list.list)) {
> + &root->list)) {
> list_add(&new_entry->list, &iter_entry->list);
> break;
> }
> @@ -150,46 +197,49 @@ static int stat_seq_init(void)
> prev_stat = new_entry->stat;
> }
> exit:
> - mutex_unlock(&stat_list_mutex);
> + mutex_unlock(&session->stat_mutex);
> return ret;
>
> exit_free_list:
> - reset_stat_list();
> - mutex_unlock(&stat_list_mutex);
> + reset_stat_session(session);
> + mutex_unlock(&session->stat_mutex);
> return ret;
> }
>
>
> static void *stat_seq_start(struct seq_file *s, loff_t *pos)
> {
> - struct trace_stat_list *l = (struct trace_stat_list *)s->private;
> + struct tracer_stat_session *session = s->private;
>
> /* Prevent from tracer switch or stat_list modification */
> - mutex_lock(&stat_list_mutex);
> + mutex_lock(&session->stat_mutex);
>
> /* If we are in the beginning of the file, print the headers */
> - if (!*pos && current_tracer.stat_headers)
> - current_tracer.stat_headers(s);
> + if (!*pos && session->ts->stat_headers)
Hmm, looks like you do need the kzalloc for both allocs above.
-- Steve
> + session->ts->stat_headers(s);
>
> - return seq_list_start(&l->list, *pos);
> + return seq_list_start(&session->stat_list.list, *pos);
> }
>
> static void *stat_seq_next(struct seq_file *s, void *p, loff_t *pos)
> {
> - struct trace_stat_list *l = (struct trace_stat_list *)s->private;
> + struct tracer_stat_session *session = s->private;
>
> - return seq_list_next(p, &l->list, pos);
> + return seq_list_next(p, &session->stat_list.list, pos);
> }
>
> -static void stat_seq_stop(struct seq_file *m, void *p)
> +static void stat_seq_stop(struct seq_file *s, void *p)
> {
> - mutex_unlock(&stat_list_mutex);
> + struct tracer_stat_session *session = s->private;
> + mutex_unlock(&session->stat_mutex);
> }
>
> static int stat_seq_show(struct seq_file *s, void *v)
> {
> + struct tracer_stat_session *session = s->private;
> struct trace_stat_list *l = list_entry(v, struct trace_stat_list, list);
> - return current_tracer.stat_show(s, l->stat);
> +
> + return session->ts->stat_show(s, l->stat);
> }
>
> static const struct seq_operations trace_stat_seq_ops = {
> @@ -199,15 +249,18 @@ static const struct seq_operations trace_stat_seq_ops = {
> .show = stat_seq_show
> };
>
> +/* The session stat is refilled and resorted at each stat file opening */
> static int tracing_stat_open(struct inode *inode, struct file *file)
> {
> int ret;
>
> + struct tracer_stat_session *session = inode->i_private;
> +
> ret = seq_open(file, &trace_stat_seq_ops);
> if (!ret) {
> struct seq_file *m = file->private_data;
> - m->private = &stat_list;
> - ret = stat_seq_init();
> + m->private = session;
> + ret = stat_seq_init(session);
> }
>
> return ret;
> @@ -219,9 +272,11 @@ static int tracing_stat_open(struct inode *inode, struct file *file)
> */
> static int tracing_stat_release(struct inode *i, struct file *f)
> {
> - mutex_lock(&stat_list_mutex);
> - reset_stat_list();
> - mutex_unlock(&stat_list_mutex);
> + struct tracer_stat_session *session = i->i_private;
> +
> + mutex_lock(&session->stat_mutex);
> + reset_stat_session(session);
> + mutex_unlock(&session->stat_mutex);
> return 0;
> }
>
> @@ -232,18 +287,70 @@ static const struct file_operations tracing_stat_fops = {
> .release = tracing_stat_release
> };
>
> +
> +static void destroy_trace_stat_files(void)
> +{
> + int i;
> +
> + if (stat_files) {
> + for (i = 0; i < nb_sessions; i++)
> + debugfs_remove(stat_files[i]);
> + kfree(stat_files);
> + stat_files = NULL;
> + }
> +}
> +
> +static void init_trace_stat_files(void)
> +{
> + int i;
> +
> + if (!stat_dir || !nb_sessions)
> + return;
> +
> + stat_files = kmalloc(sizeof(struct dentry *) * nb_sessions, GFP_KERNEL);
> +
> + if (!stat_files) {
> + pr_warning("trace stat: not enough memory\n");
> + return;
> + }
> +
> + for (i = 0; i < nb_sessions; i++) {
> + struct tracer_stat_session *session = all_stat_sessions[i];
> + stat_files[i] = debugfs_create_file(session->ts->name, 0644,
> + stat_dir,
> + session, &tracing_stat_fops);
> + if (!stat_files[i])
> + pr_warning("cannot create %s entry\n",
> + session->ts->name);
> + }
> +}
> +
> +void init_tracer_stat(struct tracer *trace)
> +{
> + int nb = basic_tracer_stat_checks(trace->stats);
> +
> + destroy_trace_stat_files();
> +
> + if (nb < 0) {
> + pr_warning("stat tracing: missing stat callback on %s\n",
> + trace->name);
> + return;
> + }
> + if (!nb)
> + return;
> +
> + init_all_sessions(nb, trace->stats);
> + init_trace_stat_files();
> +}
> +
> static int __init tracing_stat_init(void)
> {
> struct dentry *d_tracing;
> - struct dentry *entry;
>
> - INIT_LIST_HEAD(&stat_list.list);
> d_tracing = tracing_init_dentry();
>
> - entry = debugfs_create_file("trace_stat", 0444, d_tracing,
> - NULL,
> - &tracing_stat_fops);
> - if (!entry)
> + stat_dir = debugfs_create_dir("trace_stat", d_tracing);
> + if (!stat_dir)
> pr_warning("Could not create debugfs "
> "'trace_stat' entry\n");
> return 0;
> --
> 1.6.0.4
>
>
>
>
^ permalink raw reply [flat|nested] 4+ messages in thread
* Re: [PATCH 1/2] tracing/ftrace: handle more than one stat file per tracer
2009-01-05 20:53 ` Steven Rostedt
@ 2009-01-06 0:30 ` Frederic Weisbecker
2009-01-06 17:28 ` Steven Rostedt
0 siblings, 1 reply; 4+ messages in thread
From: Frederic Weisbecker @ 2009-01-06 0:30 UTC (permalink / raw)
To: Steven Rostedt
Cc: Ingo Molnar, Linux Kernel Mailing List, Pekka Enberg,
Eduard - Gabriel Munteanu
On Mon, Jan 05, 2009 at 03:53:38PM -0500, Steven Rostedt wrote:
>
> On Wed, 31 Dec 2008, Frederic Weisbecker wrote:
>
> > Impact: new API for tracers
> >
> > Make the stat tracing API reentrant. And also provide the new directory
> > /debugfs/tracing/trace_stat which will contain all the stat files for the
> > current active tracer.
> >
> > Now a tracer will, if desired, want to provide a zero terminated array of
> > tracer_stat structures.
> > Each one contains the callbacks necessary for one stat file.
> > It have to provide at least a name for its stat file, an iterator with
> > stat_start/start_next callback and an output callback for one stat entry.
> >
> > Signed-off-by: Frederic Weisbecker <fweisbec@gmail.com>
> > ---
> > diff --git a/kernel/trace/trace.h b/kernel/trace/trace.h
> > index 46677ea..b093f36 100644
> > --- a/kernel/trace/trace.h
> > +++ b/kernel/trace/trace.h
> > @@ -335,6 +335,25 @@ struct tracer_flags {
> > #define TRACER_OPT(s, b) .name = #s, .bit = b
> >
> > /*
> > + * If you want to provide a stat file (one-shot statistics), fill
> > + * an iterator with stat_start/stat_next and a stat_show callbacks.
> > + * The others callbacks are optional.
> > + */
> > +struct tracer_stat {
> > + /* The name of your stat file */
> > + const char *name;
> > + /* Iteration over statistic entries */
> > + void *(*stat_start)(void);
> > + void *(*stat_next)(void *prev, int idx);
> > + /* Compare two entries for sorting (optional) for stats */
> > + int (*stat_cmp)(void *p1, void *p2);
> > + /* Print a stat entry */
> > + int (*stat_show)(struct seq_file *s, void *p);
> > + /* Print the headers of your stat entries */
> > + int (*stat_headers)(struct seq_file *s);
> > +};
> > +
> > +/*
> > * A specific tracer, represented by methods that operate on a trace array:
> > */
> > struct tracer {
> > @@ -361,21 +380,7 @@ struct tracer {
> > struct tracer *next;
> > int print_max;
> > struct tracer_flags *flags;
> > -
> > - /*
> > - * If you change one of the following on tracing runtime, recall
> > - * init_tracer_stat()
> > - */
> > -
> > - /* Iteration over statistic entries */
> > - void *(*stat_start)(void);
> > - void *(*stat_next)(void *prev, int idx);
> > - /* Compare two entries for sorting (optional) for stats */
> > - int (*stat_cmp)(void *p1, void *p2);
> > - /* Print a stat entry */
> > - int (*stat_show)(struct seq_file *s, void *p);
> > - /* Print the headers of your stat entries */
> > - int (*stat_headers)(struct seq_file *s);
> > + struct tracer_stat *stats;
> > };
> >
> > struct trace_seq {
> > diff --git a/kernel/trace/trace_stat.c b/kernel/trace/trace_stat.c
> > index 6f194a3..293be1c 100644
> > --- a/kernel/trace/trace_stat.c
> > +++ b/kernel/trace/trace_stat.c
> > @@ -21,31 +21,29 @@ struct trace_stat_list {
> > void *stat;
> > };
> >
> > -static struct trace_stat_list stat_list;
> > +/* A stat session is the stats output in one file */
> > +struct tracer_stat_session {
> > + struct tracer_stat *ts;
> > + struct trace_stat_list stat_list;
> > + struct mutex stat_mutex;
> > +};
> >
> > -/*
> > - * This is a copy of the current tracer to avoid racy
> > - * and dangerous output while the current tracer is
> > - * switched.
> > - */
> > -static struct tracer current_tracer;
> > -
> > -/*
> > - * Protect both the current tracer and the global
> > - * stat list.
> > - */
> > -static DEFINE_MUTEX(stat_list_mutex);
> > +/* All of the sessions currently in use. Each stat file embeed one session */
> > +static struct tracer_stat_session **all_stat_sessions;
> > +static int nb_sessions;
> > +static struct dentry *stat_dir, **stat_files;
> >
> >
> > -static void reset_stat_list(void)
> > +static void reset_stat_session(struct tracer_stat_session *session)
> > {
> > + struct trace_stat_list *root = &session->stat_list;
> > struct trace_stat_list *node;
> > struct list_head *next;
> >
> > - if (list_empty(&stat_list.list))
> > + if (list_empty(&root->list))
> > return;
> >
> > - node = list_entry(stat_list.list.next, struct trace_stat_list, list);
> > + node = list_entry(root->list.next, struct trace_stat_list, list);
> > next = node->list.next;
> >
> > while (&node->list != next) {
> > @@ -54,14 +52,65 @@ static void reset_stat_list(void)
> > }
> > kfree(node);
> >
> > - INIT_LIST_HEAD(&stat_list.list);
> > + INIT_LIST_HEAD(&root->list);
> > }
> >
> > -void init_tracer_stat(struct tracer *trace)
> > +/* Called when a tracer is initialized */
> > +static int init_all_sessions(int nb, struct tracer_stat *ts)
> > +{
> > + int i, j;
> > + struct tracer_stat_session *session;
> > +
> > + if (all_stat_sessions) {
> > + for (i = 0; i < nb_sessions; i++) {
> > + session = all_stat_sessions[i];
> > + reset_stat_session(session);
> > + mutex_destroy(&session->stat_mutex);
> > + kfree(session);
> > + }
> > + }
> > + all_stat_sessions = kmalloc(sizeof(struct tracer_stat_session *) * nb,
> > + GFP_KERNEL);
>
> You want this to be kzalloc.
>
> > + if (!all_stat_sessions)
> > + return -ENOMEM;
> > +
> > + for (i = 0; i < nb; i++) {
> > + session = kmalloc(sizeof(struct tracer_stat_session) * nb,
> > + GFP_KERNEL);
>
> Not so important, but you probably want to do it here too.
>
> > + if (!session)
> > + goto free_sessions;
> > +
> > + INIT_LIST_HEAD(&session->stat_list.list);
> > + mutex_init(&session->stat_mutex);
> > + session->ts = &ts[i];
> > + all_stat_sessions[i] = session;
> > + }
> > + nb_sessions = nb;
> > + return 0;
> > +
> > +free_sessions:
> > +
> > + for (j = 0; j < i; j++)
> > + kfree(all_stat_sessions[i]);
>
> With out the kzalloc above, this may crash the kernel, because
> all_stat_sessions[i] might not be NULL when they were not initialized.
I'm not sure. j will not go through indexes of rooms that weren't allocated.
> > +
> > + kfree(all_stat_sessions);
> > + all_stat_sessions = NULL;
> > +
> > + return -ENOMEM;
> > +}
> > +
> > +static int basic_tracer_stat_checks(struct tracer_stat *ts)
> > {
> > - mutex_lock(&stat_list_mutex);
> > - current_tracer = *trace;
> > - mutex_unlock(&stat_list_mutex);
> > + int i;
> > +
> > + if (!ts)
> > + return 0;
> > +
> > + for (i = 0; ts[i].name; i++) {
> > + if (!ts[i].stat_start || !ts[i].stat_next || !ts[i].stat_show)
> > + return -EBUSY;
> > + }
> > + return i;
> > }
> >
> > /*
> > @@ -79,22 +128,19 @@ static int dummy_cmp(void *p1, void *p2)
> > * All of these copies and sorting are required on all opening
> > * since the stats could have changed between two file sessions.
> > */
> > -static int stat_seq_init(void)
> > +static int stat_seq_init(struct tracer_stat_session *session)
> > {
> > - struct trace_stat_list *iter_entry, *new_entry;
> > + struct trace_stat_list *iter_entry, *new_entry, *root;
> > + struct tracer_stat *ts = session->ts;
> > void *prev_stat;
> > int ret = 0;
> > int i;
> >
> > - mutex_lock(&stat_list_mutex);
> > - reset_stat_list();
> > -
> > - if (!current_tracer.stat_start || !current_tracer.stat_next ||
> > - !current_tracer.stat_show)
> > - goto exit;
> > + mutex_lock(&session->stat_mutex);
> > + reset_stat_session(session);
> >
> > - if (!current_tracer.stat_cmp)
> > - current_tracer.stat_cmp = dummy_cmp;
> > + if (!ts->stat_cmp)
> > + ts->stat_cmp = dummy_cmp;
> >
> > /*
> > * The first entry. Actually this is the second, but the first
> > @@ -107,8 +153,9 @@ static int stat_seq_init(void)
> > }
> >
> > INIT_LIST_HEAD(&new_entry->list);
> > - list_add(&new_entry->list, &stat_list.list);
> > - new_entry->stat = current_tracer.stat_start();
> > + root = &session->stat_list;
> > + list_add(&new_entry->list, &root->list);
> > + new_entry->stat = ts->stat_start();
> >
> > prev_stat = new_entry->stat;
> >
> > @@ -124,15 +171,15 @@ static int stat_seq_init(void)
> > }
> >
> > INIT_LIST_HEAD(&new_entry->list);
> > - new_entry->stat = current_tracer.stat_next(prev_stat, i);
> > + new_entry->stat = ts->stat_next(prev_stat, i);
> >
> > /* End of insertion */
> > if (!new_entry->stat)
> > break;
> >
> > - list_for_each_entry(iter_entry, &stat_list.list, list) {
> > + list_for_each_entry(iter_entry, &root->list, list) {
> > /* Insertion with a descendent sorting */
> > - if (current_tracer.stat_cmp(new_entry->stat,
> > + if (ts->stat_cmp(new_entry->stat,
> > iter_entry->stat) > 0) {
> >
> > list_add_tail(&new_entry->list,
> > @@ -141,7 +188,7 @@ static int stat_seq_init(void)
> >
> > /* The current smaller value */
> > } else if (list_is_last(&iter_entry->list,
> > - &stat_list.list)) {
> > + &root->list)) {
> > list_add(&new_entry->list, &iter_entry->list);
> > break;
> > }
> > @@ -150,46 +197,49 @@ static int stat_seq_init(void)
> > prev_stat = new_entry->stat;
> > }
> > exit:
> > - mutex_unlock(&stat_list_mutex);
> > + mutex_unlock(&session->stat_mutex);
> > return ret;
> >
> > exit_free_list:
> > - reset_stat_list();
> > - mutex_unlock(&stat_list_mutex);
> > + reset_stat_session(session);
> > + mutex_unlock(&session->stat_mutex);
> > return ret;
> > }
> >
> >
> > static void *stat_seq_start(struct seq_file *s, loff_t *pos)
> > {
> > - struct trace_stat_list *l = (struct trace_stat_list *)s->private;
> > + struct tracer_stat_session *session = s->private;
> >
> > /* Prevent from tracer switch or stat_list modification */
> > - mutex_lock(&stat_list_mutex);
> > + mutex_lock(&session->stat_mutex);
> >
> > /* If we are in the beginning of the file, print the headers */
> > - if (!*pos && current_tracer.stat_headers)
> > - current_tracer.stat_headers(s);
> > + if (!*pos && session->ts->stat_headers)
>
> Hmm, looks like you do need the kzalloc for both allocs above.
>
> -- Steve
There is no check about -ENOMEM before creating the debugfs file.
So yes, I should check it before dereferencing a session, or more likely
not create the debugfs files in this case. But I don't know why kzalloc...
>
> > + session->ts->stat_headers(s);
> >
> > - return seq_list_start(&l->list, *pos);
> > + return seq_list_start(&session->stat_list.list, *pos);
> > }
> >
> > static void *stat_seq_next(struct seq_file *s, void *p, loff_t *pos)
> > {
> > - struct trace_stat_list *l = (struct trace_stat_list *)s->private;
> > + struct tracer_stat_session *session = s->private;
> >
> > - return seq_list_next(p, &l->list, pos);
> > + return seq_list_next(p, &session->stat_list.list, pos);
> > }
> >
> > -static void stat_seq_stop(struct seq_file *m, void *p)
> > +static void stat_seq_stop(struct seq_file *s, void *p)
> > {
> > - mutex_unlock(&stat_list_mutex);
> > + struct tracer_stat_session *session = s->private;
> > + mutex_unlock(&session->stat_mutex);
> > }
> >
> > static int stat_seq_show(struct seq_file *s, void *v)
> > {
> > + struct tracer_stat_session *session = s->private;
> > struct trace_stat_list *l = list_entry(v, struct trace_stat_list, list);
> > - return current_tracer.stat_show(s, l->stat);
> > +
> > + return session->ts->stat_show(s, l->stat);
> > }
> >
> > static const struct seq_operations trace_stat_seq_ops = {
> > @@ -199,15 +249,18 @@ static const struct seq_operations trace_stat_seq_ops = {
> > .show = stat_seq_show
> > };
> >
> > +/* The session stat is refilled and resorted at each stat file opening */
> > static int tracing_stat_open(struct inode *inode, struct file *file)
> > {
> > int ret;
> >
> > + struct tracer_stat_session *session = inode->i_private;
> > +
> > ret = seq_open(file, &trace_stat_seq_ops);
> > if (!ret) {
> > struct seq_file *m = file->private_data;
> > - m->private = &stat_list;
> > - ret = stat_seq_init();
> > + m->private = session;
> > + ret = stat_seq_init(session);
> > }
> >
> > return ret;
> > @@ -219,9 +272,11 @@ static int tracing_stat_open(struct inode *inode, struct file *file)
> > */
> > static int tracing_stat_release(struct inode *i, struct file *f)
> > {
> > - mutex_lock(&stat_list_mutex);
> > - reset_stat_list();
> > - mutex_unlock(&stat_list_mutex);
> > + struct tracer_stat_session *session = i->i_private;
> > +
> > + mutex_lock(&session->stat_mutex);
> > + reset_stat_session(session);
> > + mutex_unlock(&session->stat_mutex);
> > return 0;
> > }
> >
> > @@ -232,18 +287,70 @@ static const struct file_operations tracing_stat_fops = {
> > .release = tracing_stat_release
> > };
> >
> > +
> > +static void destroy_trace_stat_files(void)
> > +{
> > + int i;
> > +
> > + if (stat_files) {
> > + for (i = 0; i < nb_sessions; i++)
> > + debugfs_remove(stat_files[i]);
> > + kfree(stat_files);
> > + stat_files = NULL;
> > + }
> > +}
> > +
> > +static void init_trace_stat_files(void)
> > +{
> > + int i;
> > +
> > + if (!stat_dir || !nb_sessions)
> > + return;
> > +
> > + stat_files = kmalloc(sizeof(struct dentry *) * nb_sessions, GFP_KERNEL);
> > +
> > + if (!stat_files) {
> > + pr_warning("trace stat: not enough memory\n");
> > + return;
> > + }
> > +
> > + for (i = 0; i < nb_sessions; i++) {
> > + struct tracer_stat_session *session = all_stat_sessions[i];
> > + stat_files[i] = debugfs_create_file(session->ts->name, 0644,
> > + stat_dir,
> > + session, &tracing_stat_fops);
> > + if (!stat_files[i])
> > + pr_warning("cannot create %s entry\n",
> > + session->ts->name);
> > + }
> > +}
> > +
> > +void init_tracer_stat(struct tracer *trace)
> > +{
> > + int nb = basic_tracer_stat_checks(trace->stats);
> > +
> > + destroy_trace_stat_files();
> > +
> > + if (nb < 0) {
> > + pr_warning("stat tracing: missing stat callback on %s\n",
> > + trace->name);
> > + return;
> > + }
> > + if (!nb)
> > + return;
> > +
> > + init_all_sessions(nb, trace->stats);
> > + init_trace_stat_files();
> > +}
> > +
> > static int __init tracing_stat_init(void)
> > {
> > struct dentry *d_tracing;
> > - struct dentry *entry;
> >
> > - INIT_LIST_HEAD(&stat_list.list);
> > d_tracing = tracing_init_dentry();
> >
> > - entry = debugfs_create_file("trace_stat", 0444, d_tracing,
> > - NULL,
> > - &tracing_stat_fops);
> > - if (!entry)
> > + stat_dir = debugfs_create_dir("trace_stat", d_tracing);
> > + if (!stat_dir)
> > pr_warning("Could not create debugfs "
> > "'trace_stat' entry\n");
> > return 0;
> > --
> > 1.6.0.4
> >
> >
> >
> >
^ permalink raw reply [flat|nested] 4+ messages in thread
* Re: [PATCH 1/2] tracing/ftrace: handle more than one stat file per tracer
2009-01-06 0:30 ` Frederic Weisbecker
@ 2009-01-06 17:28 ` Steven Rostedt
0 siblings, 0 replies; 4+ messages in thread
From: Steven Rostedt @ 2009-01-06 17:28 UTC (permalink / raw)
To: Frederic Weisbecker
Cc: Ingo Molnar, Linux Kernel Mailing List, Pekka Enberg,
Eduard - Gabriel Munteanu
On Tue, 6 Jan 2009, Frederic Weisbecker wrote:
> > > +
> > > + if (all_stat_sessions) {
> > > + for (i = 0; i < nb_sessions; i++) {
> > > + session = all_stat_sessions[i];
> > > + reset_stat_session(session);
> > > + mutex_destroy(&session->stat_mutex);
> > > + kfree(session);
> > > + }
> > > + }
> > > + all_stat_sessions = kmalloc(sizeof(struct tracer_stat_session *) * nb,
> > > + GFP_KERNEL);
> >
> > You want this to be kzalloc.
> >
> > > + if (!all_stat_sessions)
> > > + return -ENOMEM;
> > > +
> > > + for (i = 0; i < nb; i++) {
> > > + session = kmalloc(sizeof(struct tracer_stat_session) * nb,
> > > + GFP_KERNEL);
> >
> > Not so important, but you probably want to do it here too.
> >
> > > + if (!session)
> > > + goto free_sessions;
> > > +
> > > + INIT_LIST_HEAD(&session->stat_list.list);
> > > + mutex_init(&session->stat_mutex);
> > > + session->ts = &ts[i];
> > > + all_stat_sessions[i] = session;
> > > + }
> > > + nb_sessions = nb;
> > > + return 0;
> > > +
> > > +free_sessions:
> > > +
> > > + for (j = 0; j < i; j++)
> > > + kfree(all_stat_sessions[i]);
> >
> > With out the kzalloc above, this may crash the kernel, because
> > all_stat_sessions[i] might not be NULL when they were not initialized.
>
>
> I'm not sure. j will not go through indexes of rooms that weren't allocated.
Ah, OK, my eyes confused i's and j's so I did not notice the loop to last
failed. Which is strange because I write similar code :-/.
-- Steve
^ permalink raw reply [flat|nested] 4+ messages in thread
end of thread, other threads:[~2009-01-06 17:28 UTC | newest]
Thread overview: 4+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
2008-12-31 15:43 [PATCH 1/2] tracing/ftrace: handle more than one stat file per tracer Frederic Weisbecker
2009-01-05 20:53 ` Steven Rostedt
2009-01-06 0:30 ` Frederic Weisbecker
2009-01-06 17:28 ` Steven Rostedt
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox