* Re: [for-next][PATCH 2/2] tracing: Use temp buffer when filtering events [not found] ` <20160504135241.308454993@goodmis.org> @ 2020-08-28 7:53 ` Wen Gong 2020-08-28 22:49 ` Steven Rostedt 0 siblings, 1 reply; 5+ messages in thread From: Wen Gong @ 2020-08-28 7:53 UTC (permalink / raw) To: Steven Rostedt Cc: linux-kernel-owner, tientzu, abhishekpandit, drinkcat, Alexei Starovoitov, ath10k, linux-kernel, ath11k, briannorris, Andrew Morton, Ingo Molnar On 2016-05-04 21:52, Steven Rostedt wrote: > From: "Steven Rostedt (Red Hat)" <rostedt@goodmis.org> > > Filtering of events requires the data to be written to the ring buffer > before it can be decided to filter or not. This is because the > parameters of > the filter are based on the result that is written to the ring buffer > and > not on the parameters that are passed into the trace functions. > > The ftrace ring buffer is optimized for writing into the ring buffer > and > committing. The discard procedure used when filtering decides the event > should be discarded is much more heavy weight. Thus, using a temporary > filter when filtering events can speed things up drastically. > > Without a temp buffer we have: > > # trace-cmd start -p nop > # perf stat -r 10 hackbench 50 > 0.790706626 seconds time elapsed ( +- 0.71% ) > > # trace-cmd start -e all > # perf stat -r 10 hackbench 50 > 1.566904059 seconds time elapsed ( +- 0.27% ) > > # trace-cmd start -e all -f 'common_preempt_count==20' > # perf stat -r 10 hackbench 50 > 1.690598511 seconds time elapsed ( +- 0.19% ) > > # trace-cmd start -e all -f 'common_preempt_count!=20' > # perf stat -r 10 hackbench 50 > 1.707486364 seconds time elapsed ( +- 0.30% ) > > The first run above is without any tracing, just to get a based figure. > hackbench takes ~0.79 seconds to run on the system. > > The second run enables tracing all events where nothing is filtered. > This > increases the time by 100% and hackbench takes 1.57 seconds to run. > > The third run filters all events where the preempt count will equal > "20" > (this should never happen) thus all events are discarded. This takes > 1.69 > seconds to run. This is 10% slower than just committing the events! > > The last run enables all events and filters where the filter will > commit all > events, and this takes 1.70 seconds to run. The filtering overhead is > approximately 10%. Thus, the discard and commit of an event from the > ring > buffer may be about the same time. > > With this patch, the numbers change: > > # trace-cmd start -p nop > # perf stat -r 10 hackbench 50 > 0.778233033 seconds time elapsed ( +- 0.38% ) > > # trace-cmd start -e all > # perf stat -r 10 hackbench 50 > 1.582102692 seconds time elapsed ( +- 0.28% ) > > # trace-cmd start -e all -f 'common_preempt_count==20' > # perf stat -r 10 hackbench 50 > 1.309230710 seconds time elapsed ( +- 0.22% ) > > # trace-cmd start -e all -f 'common_preempt_count!=20' > # perf stat -r 10 hackbench 50 > 1.786001924 seconds time elapsed ( +- 0.20% ) > > The first run is again the base with no tracing. > > The second run is all tracing with no filtering. It is a little slower, > but > that may be well within the noise. > > The third run shows that discarding all events only took 1.3 seconds. > This > is a speed up of 23%! The discard is much faster than even the commit. > > The one downside is shown in the last run. Events that are not > discarded by > the filter will take longer to add, this is due to the extra copy of > the > event. > > Cc: Alexei Starovoitov <ast@kernel.org> > Signed-off-by: Steven Rostedt <rostedt@goodmis.org> > --- > kernel/trace/trace.c | 154 > +++++++++++++++++++++++++++++++++++-- > kernel/trace/trace.h | 19 ++++- > kernel/trace/trace_events.c | 10 +++ > kernel/trace/trace_events_filter.c | 10 +++ > 4 files changed, 185 insertions(+), 8 deletions(-) > ... > @@ -1695,8 +1820,23 @@ trace_event_buffer_lock_reserve(struct > ring_buffer **current_rb, > unsigned long flags, int pc) > { > struct ring_buffer_event *entry; > + int val; > > *current_rb = trace_file->tr->trace_buffer.buffer; > + > + if ((trace_file->flags & > + (EVENT_FILE_FL_SOFT_DISABLED | EVENT_FILE_FL_FILTERED)) && > + (entry = this_cpu_read(trace_buffered_event))) { > + /* Try to use the per cpu buffer first */ > + val = this_cpu_inc_return(trace_buffered_event_cnt); > + if (val == 1) { > + trace_event_setup(entry, type, flags, pc); > + entry->array[0] = len; > + return entry; > + } > + this_cpu_dec(trace_buffered_event_cnt); > + } > + > entry = trace_buffer_lock_reserve(*current_rb, > type, len, flags, pc); > /* this patch commit id is : 0fc1b09ff1ff404ddf753f5ffa5cd0adc8fdcdc9 which has upstream. how much size is the per cpu buffer? seems it is initilized in trace_buffered_event_enable, it is only 1 page size as below: void trace_buffered_event_enable(void) { ... for_each_tracing_cpu(cpu) { page = alloc_pages_node(cpu_to_node(cpu), GFP_KERNEL | __GFP_NORETRY, 0); If the size of buffer to trace is more than 1 page, such as 46680, then it trigger kernel crash/panic in my case while run trace-cmd. After debugging, the trace_file->flags in trace_event_buffer_lock_reserve is 0x40b while run trace-cmd, and it is 0x403 while collecting ftrace log. Is it have any operation to disable this patch dynamically? -- ath11k mailing list ath11k@lists.infradead.org http://lists.infradead.org/mailman/listinfo/ath11k ^ permalink raw reply [flat|nested] 5+ messages in thread
* Re: [for-next][PATCH 2/2] tracing: Use temp buffer when filtering events 2020-08-28 7:53 ` [for-next][PATCH 2/2] tracing: Use temp buffer when filtering events Wen Gong @ 2020-08-28 22:49 ` Steven Rostedt 2020-08-28 22:54 ` Steven Rostedt 0 siblings, 1 reply; 5+ messages in thread From: Steven Rostedt @ 2020-08-28 22:49 UTC (permalink / raw) To: Wen Gong Cc: linux-kernel-owner, tientzu, abhishekpandit, drinkcat, Alexei Starovoitov, ath10k, linux-kernel, ath11k, briannorris, Andrew Morton, Ingo Molnar On Fri, 28 Aug 2020 15:53:06 +0800 Wen Gong <wgong@codeaurora.org> wrote: > this patch commit id is : 0fc1b09ff1ff404ddf753f5ffa5cd0adc8fdcdc9 which > has upstream. > > how much size is the per cpu buffer? > seems it is initilized in trace_buffered_event_enable, > it is only 1 page size as below: > void trace_buffered_event_enable(void) > { > ... > for_each_tracing_cpu(cpu) { > page = alloc_pages_node(cpu_to_node(cpu), > GFP_KERNEL | __GFP_NORETRY, 0); > If the size of buffer to trace is more than 1 page, such as 46680, then > it trigger kernel crash/panic in my case while run trace-cmd. > After debugging, the trace_file->flags in > trace_event_buffer_lock_reserve is 0x40b while run trace-cmd, and it is > 0x403 while collecting ftrace log. > > Is it have any operation to disable this patch dynamically? It shouldn't be disabled, this is a bug that needs to be fixed. Also, if an event is more than a page, it wont be saved in the ftrace ring buffer, as events are limited by page size minus the headers. -- Steve -- ath11k mailing list ath11k@lists.infradead.org http://lists.infradead.org/mailman/listinfo/ath11k ^ permalink raw reply [flat|nested] 5+ messages in thread
* Re: [for-next][PATCH 2/2] tracing: Use temp buffer when filtering events 2020-08-28 22:49 ` Steven Rostedt @ 2020-08-28 22:54 ` Steven Rostedt 2020-08-28 22:55 ` Steven Rostedt 0 siblings, 1 reply; 5+ messages in thread From: Steven Rostedt @ 2020-08-28 22:54 UTC (permalink / raw) To: Wen Gong Cc: linux-kernel-owner, tientzu, abhishekpandit, drinkcat, Alexei Starovoitov, ath10k, linux-kernel, ath11k, briannorris, Andrew Morton, Ingo Molnar On Fri, 28 Aug 2020 18:49:55 -0400 Steven Rostedt <rostedt@goodmis.org> wrote: > On Fri, 28 Aug 2020 15:53:06 +0800 > Wen Gong <wgong@codeaurora.org> wrote: > > > this patch commit id is : 0fc1b09ff1ff404ddf753f5ffa5cd0adc8fdcdc9 which > > has upstream. > > > > how much size is the per cpu buffer? > > seems it is initilized in trace_buffered_event_enable, > > it is only 1 page size as below: > > void trace_buffered_event_enable(void) > > { > > ... > > for_each_tracing_cpu(cpu) { > > page = alloc_pages_node(cpu_to_node(cpu), > > GFP_KERNEL | __GFP_NORETRY, 0); > > If the size of buffer to trace is more than 1 page, such as 46680, then > > it trigger kernel crash/panic in my case while run trace-cmd. > > After debugging, the trace_file->flags in > > trace_event_buffer_lock_reserve is 0x40b while run trace-cmd, and it is > > 0x403 while collecting ftrace log. > > > > Is it have any operation to disable this patch dynamically? > > It shouldn't be disabled, this is a bug that needs to be fixed. > > Also, if an event is more than a page, it wont be saved in the ftrace > ring buffer, as events are limited by page size minus the headers. > Untested (not even compiled, as I'm now on PTO) but does this patch work for you? -- Steve diff --git a/kernel/trace/trace.c b/kernel/trace/trace.c index f40d850ebabc..3a9b4422e7fc 100644 --- a/kernel/trace/trace.c +++ b/kernel/trace/trace.c @@ -2598,7 +2598,7 @@ trace_event_buffer_lock_reserve(struct trace_buffer **current_rb, (entry = this_cpu_read(trace_buffered_event))) { /* Try to use the per cpu buffer first */ val = this_cpu_inc_return(trace_buffered_event_cnt); - if (val == 1) { + if (val == 1 || (len > (PAGE_SIZE - 8))) { trace_event_setup(entry, type, flags, pc); entry->array[0] = len; return entry; -- ath11k mailing list ath11k@lists.infradead.org http://lists.infradead.org/mailman/listinfo/ath11k ^ permalink raw reply related [flat|nested] 5+ messages in thread
* Re: [for-next][PATCH 2/2] tracing: Use temp buffer when filtering events 2020-08-28 22:54 ` Steven Rostedt @ 2020-08-28 22:55 ` Steven Rostedt 2020-08-29 15:52 ` Wen Gong 0 siblings, 1 reply; 5+ messages in thread From: Steven Rostedt @ 2020-08-28 22:55 UTC (permalink / raw) To: Wen Gong Cc: linux-kernel-owner, tientzu, abhishekpandit, drinkcat, Alexei Starovoitov, ath10k, linux-kernel, ath11k, briannorris, Andrew Morton, Ingo Molnar On Fri, 28 Aug 2020 18:54:50 -0400 Steven Rostedt <rostedt@goodmis.org> wrote: > On Fri, 28 Aug 2020 18:49:55 -0400 > Steven Rostedt <rostedt@goodmis.org> wrote: > > > On Fri, 28 Aug 2020 15:53:06 +0800 > > Wen Gong <wgong@codeaurora.org> wrote: > > > > > this patch commit id is : 0fc1b09ff1ff404ddf753f5ffa5cd0adc8fdcdc9 which > > > has upstream. > > > > > > how much size is the per cpu buffer? > > > seems it is initilized in trace_buffered_event_enable, > > > it is only 1 page size as below: > > > void trace_buffered_event_enable(void) > > > { > > > ... > > > for_each_tracing_cpu(cpu) { > > > page = alloc_pages_node(cpu_to_node(cpu), > > > GFP_KERNEL | __GFP_NORETRY, 0); > > > If the size of buffer to trace is more than 1 page, such as 46680, then > > > it trigger kernel crash/panic in my case while run trace-cmd. > > > After debugging, the trace_file->flags in > > > trace_event_buffer_lock_reserve is 0x40b while run trace-cmd, and it is > > > 0x403 while collecting ftrace log. > > > > > > Is it have any operation to disable this patch dynamically? > > > > It shouldn't be disabled, this is a bug that needs to be fixed. > > > > Also, if an event is more than a page, it wont be saved in the ftrace > > ring buffer, as events are limited by page size minus the headers. > > > > Untested (not even compiled, as I'm now on PTO) but does this patch > work for you? > > -- Steve > > diff --git a/kernel/trace/trace.c b/kernel/trace/trace.c > index f40d850ebabc..3a9b4422e7fc 100644 > --- a/kernel/trace/trace.c > +++ b/kernel/trace/trace.c > @@ -2598,7 +2598,7 @@ trace_event_buffer_lock_reserve(struct trace_buffer **current_rb, > (entry = this_cpu_read(trace_buffered_event))) { > /* Try to use the per cpu buffer first */ > val = this_cpu_inc_return(trace_buffered_event_cnt); > - if (val == 1) { > + if (val == 1 || (len > (PAGE_SIZE - 8))) { That was suppose to be: if (val == 1 && (len < (PAGE_SIZE - 8))) { -- Steve > trace_event_setup(entry, type, flags, pc); > entry->array[0] = len; > return entry; -- ath11k mailing list ath11k@lists.infradead.org http://lists.infradead.org/mailman/listinfo/ath11k ^ permalink raw reply [flat|nested] 5+ messages in thread
* Re: [for-next][PATCH 2/2] tracing: Use temp buffer when filtering events 2020-08-28 22:55 ` Steven Rostedt @ 2020-08-29 15:52 ` Wen Gong 0 siblings, 0 replies; 5+ messages in thread From: Wen Gong @ 2020-08-29 15:52 UTC (permalink / raw) To: Steven Rostedt Cc: linux-kernel-owner, tientzu, abhishekpandit, drinkcat, Alexei Starovoitov, ath10k, linux-kernel, ath11k, briannorris, Andrew Morton, Ingo Molnar On 2020-08-29 06:55, Steven Rostedt wrote: > On Fri, 28 Aug 2020 18:54:50 -0400 > Steven Rostedt <rostedt@goodmis.org> wrote: > >> On Fri, 28 Aug 2020 18:49:55 -0400 >> Steven Rostedt <rostedt@goodmis.org> wrote: >> >> > On Fri, 28 Aug 2020 15:53:06 +0800 >> > Wen Gong <wgong@codeaurora.org> wrote: >> > >> > > this patch commit id is : 0fc1b09ff1ff404ddf753f5ffa5cd0adc8fdcdc9 which >> > > has upstream. >> > > >> > > how much size is the per cpu buffer? >> > > seems it is initilized in trace_buffered_event_enable, >> > > it is only 1 page size as below: >> > > void trace_buffered_event_enable(void) >> > > { >> > > ... >> > > for_each_tracing_cpu(cpu) { >> > > page = alloc_pages_node(cpu_to_node(cpu), >> > > GFP_KERNEL | __GFP_NORETRY, 0); >> > > If the size of buffer to trace is more than 1 page, such as 46680, then >> > > it trigger kernel crash/panic in my case while run trace-cmd. >> > > After debugging, the trace_file->flags in >> > > trace_event_buffer_lock_reserve is 0x40b while run trace-cmd, and it is >> > > 0x403 while collecting ftrace log. >> > > >> > > Is it have any operation to disable this patch dynamically? >> > >> > It shouldn't be disabled, this is a bug that needs to be fixed. >> > >> > Also, if an event is more than a page, it wont be saved in the ftrace >> > ring buffer, as events are limited by page size minus the headers. >> > >> >> Untested (not even compiled, as I'm now on PTO) but does this patch >> work for you? >> >> -- Steve >> >> diff --git a/kernel/trace/trace.c b/kernel/trace/trace.c >> index f40d850ebabc..3a9b4422e7fc 100644 >> --- a/kernel/trace/trace.c >> +++ b/kernel/trace/trace.c >> @@ -2598,7 +2598,7 @@ trace_event_buffer_lock_reserve(struct >> trace_buffer **current_rb, >> (entry = this_cpu_read(trace_buffered_event))) { >> /* Try to use the per cpu buffer first */ >> val = this_cpu_inc_return(trace_buffered_event_cnt); >> - if (val == 1) { >> + if (val == 1 || (len > (PAGE_SIZE - 8))) { > > That was suppose to be: > > if (val == 1 && (len < (PAGE_SIZE - 8))) { > > -- Steve Thanks Steve! If change like this, I think it will fix my issue of crash. Will you commit this change? > >> trace_event_setup(entry, type, flags, pc); >> entry->array[0] = len; >> return entry; -- ath11k mailing list ath11k@lists.infradead.org http://lists.infradead.org/mailman/listinfo/ath11k ^ permalink raw reply [flat|nested] 5+ messages in thread
end of thread, other threads:[~2020-08-29 15:52 UTC | newest]
Thread overview: 5+ messages (download: mbox.gz follow: Atom feed
-- links below jump to the message on this page --
[not found] <20160504135202.422290539@goodmis.org>
[not found] ` <20160504135241.308454993@goodmis.org>
2020-08-28 7:53 ` [for-next][PATCH 2/2] tracing: Use temp buffer when filtering events Wen Gong
2020-08-28 22:49 ` Steven Rostedt
2020-08-28 22:54 ` Steven Rostedt
2020-08-28 22:55 ` Steven Rostedt
2020-08-29 15:52 ` Wen Gong
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox