* [PATCH 3/3] arm64: ftrace: add cond_resched() to func ftrace_make_(call|nop) @ 2018-11-30 15:09 Anders Roxell 2018-11-30 15:23 ` Steven Rostedt 2018-12-03 19:22 ` Will Deacon 0 siblings, 2 replies; 9+ messages in thread From: Anders Roxell @ 2018-11-30 15:09 UTC (permalink / raw) To: rostedt, mingo, catalin.marinas, will.deacon Cc: Arnd Bergmann, Anders Roxell, keescook, linux-arm-kernel, linux-kernel Both of those functions end up calling ftrace_modify_code(), which is expensive because it changes the page tables and flush caches. Microseconds add up because this is called in a loop for each dyn_ftrace record, and this triggers the softlockup watchdog unless we let it sleep occasionally. Rework so that we call cond_resched() before going into the ftrace_modify_code() function. Co-developed-by: Arnd Bergmann <arnd@arndb.de> Signed-off-by: Arnd Bergmann <arnd@arndb.de> Signed-off-by: Anders Roxell <anders.roxell@linaro.org> --- arch/arm64/kernel/ftrace.c | 10 ++++++++++ 1 file changed, 10 insertions(+) diff --git a/arch/arm64/kernel/ftrace.c b/arch/arm64/kernel/ftrace.c index de1a397d2d3f..9da38da58df7 100644 --- a/arch/arm64/kernel/ftrace.c +++ b/arch/arm64/kernel/ftrace.c @@ -130,6 +130,11 @@ int ftrace_make_call(struct dyn_ftrace *rec, unsigned long addr) old = aarch64_insn_gen_nop(); new = aarch64_insn_gen_branch_imm(pc, addr, AARCH64_INSN_BRANCH_LINK); + /* This function can take a long time when sanitizers are enabled, so + * lets make sure we allow RCU processing. + */ + cond_resched(); + return ftrace_modify_code(pc, old, new, true); } @@ -188,6 +193,11 @@ int ftrace_make_nop(struct module *mod, struct dyn_ftrace *rec, new = aarch64_insn_gen_nop(); + /* This function can take a long time when sanitizers are enabled, so + * lets make sure we allow RCU processing. + */ + cond_resched(); + return ftrace_modify_code(pc, old, new, validate); } -- 2.19.2 _______________________________________________ linux-arm-kernel mailing list linux-arm-kernel@lists.infradead.org http://lists.infradead.org/mailman/listinfo/linux-arm-kernel ^ permalink raw reply related [flat|nested] 9+ messages in thread
* Re: [PATCH 3/3] arm64: ftrace: add cond_resched() to func ftrace_make_(call|nop) 2018-11-30 15:09 [PATCH 3/3] arm64: ftrace: add cond_resched() to func ftrace_make_(call|nop) Anders Roxell @ 2018-11-30 15:23 ` Steven Rostedt 2018-12-03 19:22 ` Will Deacon 1 sibling, 0 replies; 9+ messages in thread From: Steven Rostedt @ 2018-11-30 15:23 UTC (permalink / raw) To: Anders Roxell Cc: keescook, Arnd Bergmann, catalin.marinas, will.deacon, linux-kernel, mingo, linux-arm-kernel On Fri, 30 Nov 2018 16:09:56 +0100 Anders Roxell <anders.roxell@linaro.org> wrote: > Both of those functions end up calling ftrace_modify_code(), which is > expensive because it changes the page tables and flush caches. > Microseconds add up because this is called in a loop for each dyn_ftrace > record, and this triggers the softlockup watchdog unless we let it sleep > occasionally. > Rework so that we call cond_resched() before going into the > ftrace_modify_code() function. > > Co-developed-by: Arnd Bergmann <arnd@arndb.de> > Signed-off-by: Arnd Bergmann <arnd@arndb.de> > Signed-off-by: Anders Roxell <anders.roxell@linaro.org> I'm fine with this patch, but I'm not placing an ack on this patch just because I don't know the repercussions of such a change. I'll let you folks take full responsibility ;-) -- Steve > --- > arch/arm64/kernel/ftrace.c | 10 ++++++++++ > 1 file changed, 10 insertions(+) > > diff --git a/arch/arm64/kernel/ftrace.c b/arch/arm64/kernel/ftrace.c > index de1a397d2d3f..9da38da58df7 100644 > --- a/arch/arm64/kernel/ftrace.c > +++ b/arch/arm64/kernel/ftrace.c > @@ -130,6 +130,11 @@ int ftrace_make_call(struct dyn_ftrace *rec, unsigned long addr) > old = aarch64_insn_gen_nop(); > new = aarch64_insn_gen_branch_imm(pc, addr, AARCH64_INSN_BRANCH_LINK); > > + /* This function can take a long time when sanitizers are enabled, so > + * lets make sure we allow RCU processing. > + */ > + cond_resched(); > + > return ftrace_modify_code(pc, old, new, true); > } > > @@ -188,6 +193,11 @@ int ftrace_make_nop(struct module *mod, struct dyn_ftrace *rec, > > new = aarch64_insn_gen_nop(); > > + /* This function can take a long time when sanitizers are enabled, so > + * lets make sure we allow RCU processing. > + */ > + cond_resched(); > + > return ftrace_modify_code(pc, old, new, validate); > } > _______________________________________________ linux-arm-kernel mailing list linux-arm-kernel@lists.infradead.org http://lists.infradead.org/mailman/listinfo/linux-arm-kernel ^ permalink raw reply [flat|nested] 9+ messages in thread
* Re: [PATCH 3/3] arm64: ftrace: add cond_resched() to func ftrace_make_(call|nop) 2018-11-30 15:09 [PATCH 3/3] arm64: ftrace: add cond_resched() to func ftrace_make_(call|nop) Anders Roxell 2018-11-30 15:23 ` Steven Rostedt @ 2018-12-03 19:22 ` Will Deacon 2018-12-03 21:51 ` Arnd Bergmann 1 sibling, 1 reply; 9+ messages in thread From: Will Deacon @ 2018-12-03 19:22 UTC (permalink / raw) To: Anders Roxell Cc: keescook, Arnd Bergmann, catalin.marinas, linux-kernel, rostedt, mingo, linux-arm-kernel Hi Anders, On Fri, Nov 30, 2018 at 04:09:56PM +0100, Anders Roxell wrote: > Both of those functions end up calling ftrace_modify_code(), which is > expensive because it changes the page tables and flush caches. > Microseconds add up because this is called in a loop for each dyn_ftrace > record, and this triggers the softlockup watchdog unless we let it sleep > occasionally. > Rework so that we call cond_resched() before going into the > ftrace_modify_code() function. > > Co-developed-by: Arnd Bergmann <arnd@arndb.de> > Signed-off-by: Arnd Bergmann <arnd@arndb.de> > Signed-off-by: Anders Roxell <anders.roxell@linaro.org> > --- > arch/arm64/kernel/ftrace.c | 10 ++++++++++ > 1 file changed, 10 insertions(+) It sounds like you're running into issues with the existing code, but I'd like to understand a bit more about exactly what you're seeing. Which part of the ftrace patching is proving to be expensive? The page table manipulation only happens once per module when using PLTs, and the cache maintenance is just a single line per patch site without an IPI. Is it the loop in ftrace_replace_code() that is causing the hassle? Will _______________________________________________ linux-arm-kernel mailing list linux-arm-kernel@lists.infradead.org http://lists.infradead.org/mailman/listinfo/linux-arm-kernel ^ permalink raw reply [flat|nested] 9+ messages in thread
* Re: [PATCH 3/3] arm64: ftrace: add cond_resched() to func ftrace_make_(call|nop) 2018-12-03 19:22 ` Will Deacon @ 2018-12-03 21:51 ` Arnd Bergmann 2018-12-04 5:50 ` Steven Rostedt 0 siblings, 1 reply; 9+ messages in thread From: Arnd Bergmann @ 2018-12-03 21:51 UTC (permalink / raw) To: Will Deacon Cc: Anders Roxell, Kees Cook, Catalin Marinas, Linux Kernel Mailing List, Steven Rostedt, Ingo Molnar, Linux ARM On Mon, Dec 3, 2018 at 8:22 PM Will Deacon <will.deacon@arm.com> wrote: > > Hi Anders, > > On Fri, Nov 30, 2018 at 04:09:56PM +0100, Anders Roxell wrote: > > Both of those functions end up calling ftrace_modify_code(), which is > > expensive because it changes the page tables and flush caches. > > Microseconds add up because this is called in a loop for each dyn_ftrace > > record, and this triggers the softlockup watchdog unless we let it sleep > > occasionally. > > Rework so that we call cond_resched() before going into the > > ftrace_modify_code() function. > > > > Co-developed-by: Arnd Bergmann <arnd@arndb.de> > > Signed-off-by: Arnd Bergmann <arnd@arndb.de> > > Signed-off-by: Anders Roxell <anders.roxell@linaro.org> > > --- > > arch/arm64/kernel/ftrace.c | 10 ++++++++++ > > 1 file changed, 10 insertions(+) > > It sounds like you're running into issues with the existing code, but I'd > like to understand a bit more about exactly what you're seeing. Which part > of the ftrace patching is proving to be expensive? > > The page table manipulation only happens once per module when using PLTs, > and the cache maintenance is just a single line per patch site without an > IPI. > > Is it the loop in ftrace_replace_code() that is causing the hassle? Yes: with an allmodconfig kernel, the ftrace selftest calls ftrace_replace_code to look >40000 through ftrace_make_call/ftrace_make_nop, and these end up calling static int __kprobes __aarch64_insn_write(void *addr, __le32 insn) { void *waddr = addr; unsigned long flags = 0; int ret; raw_spin_lock_irqsave(&patch_lock, flags); waddr = patch_map(addr, FIX_TEXT_POKE0); ret = probe_kernel_write(waddr, &insn, AARCH64_INSN_SIZE); patch_unmap(FIX_TEXT_POKE0); raw_spin_unlock_irqrestore(&patch_lock, flags); return ret; } int __kprobes aarch64_insn_patch_text_nosync(void *addr, u32 insn) { u32 *tp = addr; int ret; /* A64 instructions must be word aligned */ if ((uintptr_t)tp & 0x3) return -EINVAL; ret = aarch64_insn_write(tp, insn); if (ret == 0) __flush_icache_range((uintptr_t)tp, (uintptr_t)tp + AARCH64_INSN_SIZE); return ret; } which seems to be where the main cost is. This is with inside of qemu, and with lots of debugging options (in particular kcov and ubsan) enabled, that make each function call more expensive. Arnd _______________________________________________ linux-arm-kernel mailing list linux-arm-kernel@lists.infradead.org http://lists.infradead.org/mailman/listinfo/linux-arm-kernel ^ permalink raw reply [flat|nested] 9+ messages in thread
* Re: [PATCH 3/3] arm64: ftrace: add cond_resched() to func ftrace_make_(call|nop) 2018-12-03 21:51 ` Arnd Bergmann @ 2018-12-04 5:50 ` Steven Rostedt 2018-12-04 11:12 ` Will Deacon 0 siblings, 1 reply; 9+ messages in thread From: Steven Rostedt @ 2018-12-04 5:50 UTC (permalink / raw) To: Arnd Bergmann Cc: Anders Roxell, Kees Cook, Catalin Marinas, Will Deacon, Linux Kernel Mailing List, Ingo Molnar, Linux ARM On Mon, 3 Dec 2018 22:51:52 +0100 Arnd Bergmann <arnd@arndb.de> wrote: > On Mon, Dec 3, 2018 at 8:22 PM Will Deacon <will.deacon@arm.com> wrote: > > > > Hi Anders, > > > > On Fri, Nov 30, 2018 at 04:09:56PM +0100, Anders Roxell wrote: > > > Both of those functions end up calling ftrace_modify_code(), which is > > > expensive because it changes the page tables and flush caches. > > > Microseconds add up because this is called in a loop for each dyn_ftrace > > > record, and this triggers the softlockup watchdog unless we let it sleep > > > occasionally. > > > Rework so that we call cond_resched() before going into the > > > ftrace_modify_code() function. > > > > > > Co-developed-by: Arnd Bergmann <arnd@arndb.de> > > > Signed-off-by: Arnd Bergmann <arnd@arndb.de> > > > Signed-off-by: Anders Roxell <anders.roxell@linaro.org> > > > --- > > > arch/arm64/kernel/ftrace.c | 10 ++++++++++ > > > 1 file changed, 10 insertions(+) > > > > It sounds like you're running into issues with the existing code, but I'd > > like to understand a bit more about exactly what you're seeing. Which part > > of the ftrace patching is proving to be expensive? > > > > The page table manipulation only happens once per module when using PLTs, > > and the cache maintenance is just a single line per patch site without an > > IPI. > > > > Is it the loop in ftrace_replace_code() that is causing the hassle? > > Yes: with an allmodconfig kernel, the ftrace selftest calls ftrace_replace_code > to look >40000 through ftrace_make_call/ftrace_make_nop, and these > end up calling > > static int __kprobes __aarch64_insn_write(void *addr, __le32 insn) > { > void *waddr = addr; > unsigned long flags = 0; > int ret; > > raw_spin_lock_irqsave(&patch_lock, flags); > waddr = patch_map(addr, FIX_TEXT_POKE0); > > ret = probe_kernel_write(waddr, &insn, AARCH64_INSN_SIZE); > > patch_unmap(FIX_TEXT_POKE0); > raw_spin_unlock_irqrestore(&patch_lock, flags); > > return ret; > } > int __kprobes aarch64_insn_patch_text_nosync(void *addr, u32 insn) > { > u32 *tp = addr; > int ret; > > /* A64 instructions must be word aligned */ > if ((uintptr_t)tp & 0x3) > return -EINVAL; > > ret = aarch64_insn_write(tp, insn); > if (ret == 0) > __flush_icache_range((uintptr_t)tp, > (uintptr_t)tp + AARCH64_INSN_SIZE); > > return ret; > } > > which seems to be where the main cost is. This is with inside of > qemu, and with lots of debugging options (in particular > kcov and ubsan) enabled, that make each function call > more expensive. I was thinking more about this. Would something like this work? -- Steve diff --git a/kernel/trace/ftrace.c b/kernel/trace/ftrace.c index 8ef9fc226037..42e89397778b 100644 --- a/kernel/trace/ftrace.c +++ b/kernel/trace/ftrace.c @@ -2393,11 +2393,14 @@ void __weak ftrace_replace_code(int enable) { struct dyn_ftrace *rec; struct ftrace_page *pg; + bool schedulable; int failed; if (unlikely(ftrace_disabled)) return; + schedulable = !irqs_disabled() & !preempt_count(); + do_for_each_ftrace_rec(pg, rec) { if (rec->flags & FTRACE_FL_DISABLED) @@ -2409,6 +2412,8 @@ void __weak ftrace_replace_code(int enable) /* Stop processing */ return; } + if (schedulable) + cond_resched(); } while_for_each_ftrace_rec(); } _______________________________________________ linux-arm-kernel mailing list linux-arm-kernel@lists.infradead.org http://lists.infradead.org/mailman/listinfo/linux-arm-kernel ^ permalink raw reply related [flat|nested] 9+ messages in thread
* Re: [PATCH 3/3] arm64: ftrace: add cond_resched() to func ftrace_make_(call|nop) 2018-12-04 5:50 ` Steven Rostedt @ 2018-12-04 11:12 ` Will Deacon 2018-12-04 17:26 ` Steven Rostedt 0 siblings, 1 reply; 9+ messages in thread From: Will Deacon @ 2018-12-04 11:12 UTC (permalink / raw) To: Steven Rostedt Cc: Anders Roxell, Kees Cook, Arnd Bergmann, Catalin Marinas, Linux Kernel Mailing List, Ingo Molnar, Linux ARM Hi Steve, Arnd, On Tue, Dec 04, 2018 at 12:50:12AM -0500, Steven Rostedt wrote: > On Mon, 3 Dec 2018 22:51:52 +0100 > Arnd Bergmann <arnd@arndb.de> wrote: > > On Mon, Dec 3, 2018 at 8:22 PM Will Deacon <will.deacon@arm.com> wrote: > > > On Fri, Nov 30, 2018 at 04:09:56PM +0100, Anders Roxell wrote: > > > > Both of those functions end up calling ftrace_modify_code(), which is > > > > expensive because it changes the page tables and flush caches. > > > > Microseconds add up because this is called in a loop for each dyn_ftrace > > > > record, and this triggers the softlockup watchdog unless we let it sleep > > > > occasionally. > > > > Rework so that we call cond_resched() before going into the > > > > ftrace_modify_code() function. > > > > > > > > Co-developed-by: Arnd Bergmann <arnd@arndb.de> > > > > Signed-off-by: Arnd Bergmann <arnd@arndb.de> > > > > Signed-off-by: Anders Roxell <anders.roxell@linaro.org> > > > > --- > > > > arch/arm64/kernel/ftrace.c | 10 ++++++++++ > > > > 1 file changed, 10 insertions(+) > > > > > > It sounds like you're running into issues with the existing code, but I'd > > > like to understand a bit more about exactly what you're seeing. Which part > > > of the ftrace patching is proving to be expensive? > > > > > > The page table manipulation only happens once per module when using PLTs, > > > and the cache maintenance is just a single line per patch site without an > > > IPI. > > > > > > Is it the loop in ftrace_replace_code() that is causing the hassle? > > > > Yes: with an allmodconfig kernel, the ftrace selftest calls ftrace_replace_code > > to look >40000 through ftrace_make_call/ftrace_make_nop, and these > > end up calling Ok, 40000 invocations would do it! > > static int __kprobes __aarch64_insn_write(void *addr, __le32 insn) > > { > > void *waddr = addr; > > unsigned long flags = 0; > > int ret; > > > > raw_spin_lock_irqsave(&patch_lock, flags); > > waddr = patch_map(addr, FIX_TEXT_POKE0); > > > > ret = probe_kernel_write(waddr, &insn, AARCH64_INSN_SIZE); > > > > patch_unmap(FIX_TEXT_POKE0); > > raw_spin_unlock_irqrestore(&patch_lock, flags); > > > > return ret; > > } > > int __kprobes aarch64_insn_patch_text_nosync(void *addr, u32 insn) > > { > > u32 *tp = addr; > > int ret; > > > > /* A64 instructions must be word aligned */ > > if ((uintptr_t)tp & 0x3) > > return -EINVAL; > > > > ret = aarch64_insn_write(tp, insn); > > if (ret == 0) > > __flush_icache_range((uintptr_t)tp, > > (uintptr_t)tp + AARCH64_INSN_SIZE); > > > > return ret; > > } > > > > which seems to be where the main cost is. This is with inside of > > qemu, and with lots of debugging options (in particular > > kcov and ubsan) enabled, that make each function call > > more expensive. > > I was thinking more about this. Would something like this work? > > -- Steve > > diff --git a/kernel/trace/ftrace.c b/kernel/trace/ftrace.c > index 8ef9fc226037..42e89397778b 100644 > --- a/kernel/trace/ftrace.c > +++ b/kernel/trace/ftrace.c > @@ -2393,11 +2393,14 @@ void __weak ftrace_replace_code(int enable) > { > struct dyn_ftrace *rec; > struct ftrace_page *pg; > + bool schedulable; > int failed; > > if (unlikely(ftrace_disabled)) > return; > > + schedulable = !irqs_disabled() & !preempt_count(); Looks suspiciously like a bitwise preemptible() to me! > + > do_for_each_ftrace_rec(pg, rec) { > > if (rec->flags & FTRACE_FL_DISABLED) > @@ -2409,6 +2412,8 @@ void __weak ftrace_replace_code(int enable) > /* Stop processing */ > return; > } > + if (schedulable) > + cond_resched(); > } while_for_each_ftrace_rec(); > } If this solves the problem in core code, them I'm all for it. Otherwise, I was thinking of rolling our own ftrace_replace_code() for arm64, but that's going to involve a fair amount of duplication. Will _______________________________________________ linux-arm-kernel mailing list linux-arm-kernel@lists.infradead.org http://lists.infradead.org/mailman/listinfo/linux-arm-kernel ^ permalink raw reply [flat|nested] 9+ messages in thread
* Re: [PATCH 3/3] arm64: ftrace: add cond_resched() to func ftrace_make_(call|nop) 2018-12-04 11:12 ` Will Deacon @ 2018-12-04 17:26 ` Steven Rostedt 2018-12-04 18:07 ` Anders Roxell 0 siblings, 1 reply; 9+ messages in thread From: Steven Rostedt @ 2018-12-04 17:26 UTC (permalink / raw) To: Will Deacon Cc: Anders Roxell, Kees Cook, Arnd Bergmann, Catalin Marinas, Linux Kernel Mailing List, Ingo Molnar, Linux ARM On Tue, 4 Dec 2018 11:12:43 +0000 Will Deacon <will.deacon@arm.com> wrote: > > diff --git a/kernel/trace/ftrace.c b/kernel/trace/ftrace.c > > index 8ef9fc226037..42e89397778b 100644 > > --- a/kernel/trace/ftrace.c > > +++ b/kernel/trace/ftrace.c > > @@ -2393,11 +2393,14 @@ void __weak ftrace_replace_code(int enable) > > { > > struct dyn_ftrace *rec; > > struct ftrace_page *pg; > > + bool schedulable; > > int failed; > > > > if (unlikely(ftrace_disabled)) > > return; > > > > + schedulable = !irqs_disabled() & !preempt_count(); > > Looks suspiciously like a bitwise preemptible() to me! Ah, thanks. Yeah, that should have been &&. But what did you expect. I didn't even compile this ;-) > > > + > > do_for_each_ftrace_rec(pg, rec) { > > > > if (rec->flags & FTRACE_FL_DISABLED) > > @@ -2409,6 +2412,8 @@ void __weak ftrace_replace_code(int enable) > > /* Stop processing */ > > return; > > } > > + if (schedulable) > > + cond_resched(); > > } while_for_each_ftrace_rec(); > > } > > If this solves the problem in core code, them I'm all for it. Otherwise, I > was thinking of rolling our own ftrace_replace_code() for arm64, but that's > going to involve a fair amount of duplication. > If it does, then I'll add it. Or take a patch for it ;-) My main concern is that this can be called from non schedulable context. -- Steve _______________________________________________ linux-arm-kernel mailing list linux-arm-kernel@lists.infradead.org http://lists.infradead.org/mailman/listinfo/linux-arm-kernel ^ permalink raw reply [flat|nested] 9+ messages in thread
* Re: [PATCH 3/3] arm64: ftrace: add cond_resched() to func ftrace_make_(call|nop) 2018-12-04 17:26 ` Steven Rostedt @ 2018-12-04 18:07 ` Anders Roxell 2018-12-04 18:23 ` Steven Rostedt 0 siblings, 1 reply; 9+ messages in thread From: Anders Roxell @ 2018-12-04 18:07 UTC (permalink / raw) To: rostedt Cc: Kees Cook, Arnd Bergmann, Catalin Marinas, Will Deacon, Linux Kernel Mailing List, mingo, Linux ARM On Tue, 4 Dec 2018 at 18:26, Steven Rostedt <rostedt@goodmis.org> wrote: > > On Tue, 4 Dec 2018 11:12:43 +0000 > Will Deacon <will.deacon@arm.com> wrote: > > > > diff --git a/kernel/trace/ftrace.c b/kernel/trace/ftrace.c > > > index 8ef9fc226037..42e89397778b 100644 > > > --- a/kernel/trace/ftrace.c > > > +++ b/kernel/trace/ftrace.c > > > @@ -2393,11 +2393,14 @@ void __weak ftrace_replace_code(int enable) > > > { > > > struct dyn_ftrace *rec; > > > struct ftrace_page *pg; > > > + bool schedulable; > > > int failed; > > > > > > if (unlikely(ftrace_disabled)) > > > return; > > > > > > + schedulable = !irqs_disabled() & !preempt_count(); > > > > Looks suspiciously like a bitwise preemptible() to me! > > Ah, thanks. Yeah, that should have been &&. But what did you expect. > I didn't even compile this ;-) > > > > > > + > > > do_for_each_ftrace_rec(pg, rec) { > > > > > > if (rec->flags & FTRACE_FL_DISABLED) > > > @@ -2409,6 +2412,8 @@ void __weak ftrace_replace_code(int enable) > > > /* Stop processing */ > > > return; > > > } > > > + if (schedulable) > > > + cond_resched(); > > > } while_for_each_ftrace_rec(); > > > } > > > > If this solves the problem in core code, them I'm all for it. Otherwise, I > > was thinking of rolling our own ftrace_replace_code() for arm64, but that's > > going to involve a fair amount of duplication. > > > > If it does, then I'll add it. Or take a patch for it ;-) I tested your patch. it worked. I'll send a patch shortly. Cheers, Anders _______________________________________________ linux-arm-kernel mailing list linux-arm-kernel@lists.infradead.org http://lists.infradead.org/mailman/listinfo/linux-arm-kernel ^ permalink raw reply [flat|nested] 9+ messages in thread
* Re: [PATCH 3/3] arm64: ftrace: add cond_resched() to func ftrace_make_(call|nop) 2018-12-04 18:07 ` Anders Roxell @ 2018-12-04 18:23 ` Steven Rostedt 0 siblings, 0 replies; 9+ messages in thread From: Steven Rostedt @ 2018-12-04 18:23 UTC (permalink / raw) To: Anders Roxell Cc: Kees Cook, Arnd Bergmann, Catalin Marinas, Will Deacon, Linux Kernel Mailing List, mingo, Linux ARM On Tue, 4 Dec 2018 19:07:16 +0100 Anders Roxell <anders.roxell@linaro.org> wrote: > > > > + schedulable = !irqs_disabled() & !preempt_count(); > > > > > > Looks suspiciously like a bitwise preemptible() to me! > > > > Ah, thanks. Yeah, that should have been &&. But what did you expect. > > I didn't even compile this ;-) > > > > If it does, then I'll add it. Or take a patch for it ;-) > > I tested your patch. it worked. > > I'll send a patch shortly. > Thanks. Please add a comment above the schedulable test stating that some archs call this with interrupts or preemption disabled, but other archs don't and this can cause a tremendous unneeded latency. -- Steve _______________________________________________ linux-arm-kernel mailing list linux-arm-kernel@lists.infradead.org http://lists.infradead.org/mailman/listinfo/linux-arm-kernel ^ permalink raw reply [flat|nested] 9+ messages in thread
end of thread, other threads:[~2018-12-04 18:24 UTC | newest] Thread overview: 9+ messages (download: mbox.gz follow: Atom feed -- links below jump to the message on this page -- 2018-11-30 15:09 [PATCH 3/3] arm64: ftrace: add cond_resched() to func ftrace_make_(call|nop) Anders Roxell 2018-11-30 15:23 ` Steven Rostedt 2018-12-03 19:22 ` Will Deacon 2018-12-03 21:51 ` Arnd Bergmann 2018-12-04 5:50 ` Steven Rostedt 2018-12-04 11:12 ` Will Deacon 2018-12-04 17:26 ` Steven Rostedt 2018-12-04 18:07 ` Anders Roxell 2018-12-04 18:23 ` Steven Rostedt
This is a public inbox, see mirroring instructions for how to clone and mirror all data and code used for this inbox; as well as URLs for NNTP newsgroup(s).