From: Steven Rostedt <rostedt@goodmis.org>
To: "Yordan Karadzhov (VMware)" <y.karadz@gmail.com>
Cc: linux-trace-devel@vger.kernel.org
Subject: Re: [PATCH 2/3] kernel-shark-qt: Consolidate load_data_entries and load_data_records
Date: Thu, 5 Jul 2018 12:35:09 -0400 [thread overview]
Message-ID: <20180705123509.26f1c0c9@gandalf.local.home> (raw)
In-Reply-To: <993cf66c-316a-1b7a-59a6-a5f607476667@gmail.com>
On Wed, 4 Jul 2018 14:39:01 +0300
"Yordan Karadzhov (VMware)" <y.karadz@gmail.com> wrote:
Your comment actually had me re-examine the code, and I found a bug.
> >
> > +struct rec_list {
> > + struct pevent_record *rec;
> > + struct rec_list *next;
> > +};
> > +
> > +static void free_rec_list(struct rec_list **rec_list, int n_cpus)
> > +{
> > + struct rec_list *temp_rec;
> > + int cpu;
> > +
> > + for (cpu = 0; cpu < n_cpus; ++cpu) {
> > + while (rec_list[cpu]) {
> > + temp_rec = rec_list[cpu];
> > + rec_list[cpu] = temp_rec->next;
I need a:
free_record(temp_rec->rec);
> > + free(temp_rec);
> > + }
> > + }
> > + free(rec_list);
> > +}
> > +
> > @@ -521,9 +587,11 @@ ssize_t kshark_load_data_entries(struct kshark_context *kshark_ctx,
> > struct kshark_entry ***data_rows)
> > {
> > struct event_filter *adv_filter = kshark_ctx->advanced_event_filter;
> > - struct kshark_entry **cpu_list, **rows;
> > - struct kshark_entry *entry, **next;
> > struct kshark_task_list *task;
> > + struct kshark_entry **rows;
> > + struct kshark_entry *entry;
> > + struct rec_list **rec_list;
> > + struct rec_list *temp_rec;
> > struct pevent_record *rec;
> > int cpu, n_cpus, next_cpu;
> > size_t count, total = 0;
> > @@ -533,24 +601,41 @@ ssize_t kshark_load_data_entries(struct kshark_context *kshark_ctx,
> > if (*data_rows)
> > free(*data_rows);
> >
> > + total = get_records(kshark_ctx, &rec_list);
> > + if (total < 0)
> > + goto fail;
> > +
> > + rows = calloc(total, sizeof(struct kshark_entry *));
> > + if(!rows)
> > + goto fail;
> > +
> > n_cpus = tracecmd_cpus(kshark_ctx->handle);
> > - cpu_list = calloc(n_cpus, sizeof(struct kshark_entry *));
> >
> > - for (cpu = 0; cpu < n_cpus; ++cpu) {
> > - count = 0;
> > - cpu_list[cpu] = NULL;
> > - next = &cpu_list[cpu];
> > + for (count = 0; count < total; count++) {
> > + ts = 0;
> > + next_cpu = -1;
> > + for (cpu = 0; cpu < n_cpus; ++cpu) {
> > + if (!rec_list[cpu])
> > + continue;
> >
> > - rec = tracecmd_read_cpu_first(kshark_ctx->handle, cpu);
> > - while (rec) {
> > - *next = entry = malloc(sizeof(struct kshark_entry));
> > + if (!ts || rec_list[cpu]->rec->ts < ts) {
> > + ts = rec_list[cpu]->rec->ts;
> > + next_cpu = cpu;
> > + }
> > + }
> > +
> > + if (next_cpu >= 0) {
> > + entry = malloc(sizeof(struct kshark_entry));
> > if (!entry)
> > - goto fail;
> > + goto fail_free;
> > +
> > + rec = rec_list[next_cpu]->rec;
> > + rows[count] = entry;
> >
> > kshark_set_entry_values(kshark_ctx, rec, entry);
> > task = kshark_add_task(kshark_ctx, entry->pid);
> > if (!task)
> > - goto fail;
> > + goto fail_free;
> >
> > /* Apply event filtering. */
> > ret = FILTER_NONE;
> > @@ -567,47 +652,26 @@ ssize_t kshark_load_data_entries(struct kshark_context *kshark_ctx,
> > entry->visible &= ~kshark_ctx->filter_mask;
> > }
> >
> > - entry->next = NULL;
> > - next = &entry->next;
> > + temp_rec = rec_list[next_cpu];
> > + rec_list[next_cpu] = rec_list[next_cpu]->next;
> > + free(temp_rec);
I should add a comment here:
/* The record will no longer be referenced */
> > free_record(rec);
> > -
> > - ++count;
> > - rec = tracecmd_read_data(kshark_ctx->handle, cpu);
> > - }
> > -
> > - total += count;
> > - }
> > -
> > - rows = calloc(total, sizeof(struct kshark_entry *));
> > - if (!rows)
> > - goto fail;
> > -
> > - count = 0;
> > - while (count < total) {
> > - ts = 0;
> > - next_cpu = -1;
> > - for (cpu = 0; cpu < n_cpus; ++cpu) {
> > - if (!cpu_list[cpu])
> > - continue;
> > -
> > - if (!ts || cpu_list[cpu]->ts < ts) {
> > - ts = cpu_list[cpu]->ts;
> > - next_cpu = cpu;
> > - }
> > - }
> > -
> > - if (next_cpu >= 0) {
> > - rows[count] = cpu_list[next_cpu];
> > - cpu_list[next_cpu] = cpu_list[next_cpu]->next;
> > }
> > - ++count;
> > }
> >
> > - free(cpu_list);
> > + free_rec_list(rec_list, n_cpus);
> > *data_rows = rows;
> > return total;
> >
> > -fail:
> > + fail_free:
> > + free_rec_list(rec_list, n_cpus);
> > + for (count = 0; count < total; count++) {
> > + if (!rows[count])
> > + break;
> > + free(rows[count]);
> > + }
> > + free(rows);
> > + fail:
> > fprintf(stderr, "Failed to allocate memory during data loading.\n");
> > return -ENOMEM;
> > }
> > @@ -625,82 +689,60 @@ fail:
> > ssize_t kshark_load_data_records(struct kshark_context *kshark_ctx,
> > struct pevent_record ***data_rows)
> > {
[..]
> > rows = calloc(total, sizeof(struct pevent_record *));
> > - if (!rows)
> > + if(!rows)
> > goto fail;
> >
> > - count = 0;
> > - while (count < total) {
> > + n_cpus = tracecmd_cpus(kshark_ctx->handle);
> > +
> > + for (count = 0; count < total; count++) {
> > ts = 0;
> > next_cpu = -1;
> > for (cpu = 0; cpu < n_cpus; ++cpu) {
> > - if (!cpu_list[cpu])
> > + if (!rec_list[cpu])
> > continue;
> >
> > - if (!ts || cpu_list[cpu]->rec->ts < ts) {
> > - ts = cpu_list[cpu]->rec->ts;
> > + if (!ts || rec_list[cpu]->rec->ts < ts) {
> > + ts = rec_list[cpu]->rec->ts;
> > next_cpu = cpu;
> > }
> > }
> >
> > if (next_cpu >= 0) {
> > - rows[count] = cpu_list[next_cpu]->rec;
> > - temp_rec = cpu_list[next_cpu];
> > - cpu_list[next_cpu] = cpu_list[next_cpu]->next;
> > - free (temp_rec);
> > - }
> > + rec = rec_list[next_cpu]->rec;
> > + rows[count] = rec;
> >
> > - ++count;
> > + pid = pevent_data_pid(kshark_ctx->pevent, rec);
> > + task = kshark_add_task(kshark_ctx, pid);
> > + if (!task)
> > + goto fail;
> > +
> > + temp_rec = rec_list[next_cpu];
> > + rec_list[next_cpu] = rec_list[next_cpu]->next;
> > + free(temp_rec);
> > + free_record(rec);
>
> You have to free only the element of the list.
> The user is responsible for freeing the record.
Right! And I'll replace that with a comment:
/* The record is still referenced in rows */
Thanks for the review.
-- Steve
>
> > + }
> > }
> >
> > - free(cpu_list);
> > + free_rec_list(rec_list, n_cpus);
> > *data_rows = rows;
> > return total;
> >
> > -fail:
> > + fail:
> > fprintf(stderr, "Failed to allocate memory during data loading.\n");
> > return -ENOMEM;
> > }
> >
next prev parent reply other threads:[~2018-07-05 16:35 UTC|newest]
Thread overview: 18+ messages / expand[flat|nested] mbox.gz Atom feed top
2018-07-03 16:21 [PATCH 2/3] kernel-shark-qt: Consolidate load_data_entries and load_data_records Steven Rostedt
2018-07-04 11:39 ` Yordan Karadzhov (VMware)
2018-07-05 14:47 ` Steven Rostedt
2018-07-05 14:53 ` Yordan Karadzhov (VMware)
2018-07-05 15:48 ` Steven Rostedt
2018-07-05 16:08 ` Yordan Karadzhov (VMware)
2018-07-05 16:25 ` Steven Rostedt
2018-07-05 16:35 ` Steven Rostedt [this message]
2018-07-05 16:38 ` Steven Rostedt
2018-07-05 22:05 ` Steven Rostedt
2018-07-06 10:51 ` Yordan Karadzhov (VMware)
2018-07-06 11:36 ` Yordan Karadzhov (VMware)
2018-07-06 12:46 ` Yordan Karadzhov (VMware)
2018-07-06 18:21 ` Steven Rostedt
2018-07-06 18:43 ` Steven Rostedt
2018-07-09 15:54 ` Yordan Karadzhov (VMware)
2018-07-06 18:24 ` Steven Rostedt
2018-07-09 15:54 ` Yordan Karadzhov (VMware)
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20180705123509.26f1c0c9@gandalf.local.home \
--to=rostedt@goodmis.org \
--cc=linux-trace-devel@vger.kernel.org \
--cc=y.karadz@gmail.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox;
as well as URLs for NNTP newsgroup(s).