On Fri, 7 Aug 2020 at 19:06, Paul E. McKenney <paul...@kernel.org> wrote: > On Fri, Aug 07, 2020 at 11:00:31AM +0200, Marco Elver wrote: > > Since KCSAN instrumentation is everywhere, we need to treat the hooks > > NMI-like for interrupt tracing. In order to present an as 'normal' as > > possible context to the code called by KCSAN when reporting errors, we > > need to update the IRQ-tracing state. > > > > Tested: Several runs through kcsan-test with different configuration > > (PROVE_LOCKING on/off), as well as hours of syzbot testing with the > > original config that caught the problem (without CONFIG_PARAVIRT=y, > > which appears to cause IRQ state tracking inconsistencies even when > > KCSAN remains off, see Link). > > > > Link: https://lkml.kernel.org/r/0000000000007d3b2d05ac1c3...@google.com > > Fixes: 248591f5d257 ("kcsan: Make KCSAN compatible with new IRQ state > > tracking") > > Reported-by: syzbot+8db9e1ecde74e590a...@syzkaller.appspotmail.com > > Co-developed-by: Peter Zijlstra (Intel) <pet...@infradead.org>
Peter, if you're fine with it, I think we'll require your Signed-off-by (since Co-developed-by). > > Signed-off-by: Marco Elver <el...@google.com> > > --- > > Patch Note: This patch applies to latest mainline. While current > > mainline suffers from the above problem, the configs required to hit the > > issue are likely not enabled too often (of course with PROVE_LOCKING on; > > we hit it on syzbot though). It'll probably be wise to queue this as > > normal on -rcu, just in case something is still off, given the > > non-trivial nature of the issue. (If it should instead go to mainline > > right now as a fix, I'd like some more test time on syzbot.) > > The usual, please let me know when/if you would like me to apply > to -rcu. And have a great weekend! I think we need to wait until you have rebased -rcu to 5.9-rc1 some time next week. I will send a reminder after, and if it doesn't apply cleanly, I'll send a rebased patch. Thank you! -- Marco > Thanx, Paul > > > --- > > kernel/kcsan/core.c | 79 ++++++++++++++++++++++++++++++++++---------- > > kernel/kcsan/kcsan.h | 3 +- > > 2 files changed, 62 insertions(+), 20 deletions(-) > > > > diff --git a/kernel/kcsan/core.c b/kernel/kcsan/core.c > > index 9147ff6a12e5..6202a645f1e2 100644 > > --- a/kernel/kcsan/core.c > > +++ b/kernel/kcsan/core.c > > @@ -291,13 +291,28 @@ static inline unsigned int get_delay(void) > > 0); > > } > > > > -void kcsan_save_irqtrace(struct task_struct *task) > > -{ > > +/* > > + * KCSAN instrumentation is everywhere, which means we must treat the hooks > > + * NMI-like for interrupt tracing. In order to present a 'normal' as > > possible > > + * context to the code called by KCSAN when reporting errors we need to > > update > > + * the IRQ-tracing state. > > + * > > + * Save and restore the IRQ state trace touched by KCSAN, since KCSAN's > > + * runtime is entered for every memory access, and potentially useful > > + * information is lost if dirtied by KCSAN. > > + */ > > + > > +struct kcsan_irq_state { > > + unsigned long flags; > > #ifdef CONFIG_TRACE_IRQFLAGS > > - task->kcsan_save_irqtrace = task->irqtrace; > > + int hardirqs_enabled; > > #endif > > -} > > +}; > > > > +/* > > + * This is also called by the reporting task for the other task, to > > generate the > > + * right report with CONFIG_KCSAN_VERBOSE. No harm in restoring more than > > once. > > + */ > > void kcsan_restore_irqtrace(struct task_struct *task) > > { > > #ifdef CONFIG_TRACE_IRQFLAGS > > @@ -305,6 +320,41 @@ void kcsan_restore_irqtrace(struct task_struct *task) > > #endif > > } > > > > +/* > > + * Saves/restores IRQ state (see comment above). Need noinline to work > > around > > + * unfortunate code-gen upon inlining, resulting in objtool getting > > confused as > > + * well as losing stack trace information. > > + */ > > +static noinline void kcsan_irq_save(struct kcsan_irq_state *irq_state) > > +{ > > +#ifdef CONFIG_TRACE_IRQFLAGS > > + current->kcsan_save_irqtrace = current->irqtrace; > > + irq_state->hardirqs_enabled = lockdep_hardirqs_enabled(); > > +#endif > > + if (!kcsan_interrupt_watcher) { > > + kcsan_disable_current(); /* Lockdep might WARN, etc. */ > > + raw_local_irq_save(irq_state->flags); > > + lockdep_hardirqs_off(_RET_IP_); > > + kcsan_enable_current(); > > + } > > +} > > + > > +static noinline void kcsan_irq_restore(struct kcsan_irq_state *irq_state) > > +{ > > + if (!kcsan_interrupt_watcher) { > > + kcsan_disable_current(); /* Lockdep might WARN, etc. */ > > +#ifdef CONFIG_TRACE_IRQFLAGS > > + if (irq_state->hardirqs_enabled) { > > + lockdep_hardirqs_on_prepare(_RET_IP_); > > + lockdep_hardirqs_on(_RET_IP_); > > + } > > +#endif > > + raw_local_irq_restore(irq_state->flags); > > + kcsan_enable_current(); > > + } > > + kcsan_restore_irqtrace(current); > > +} > > + > > /* > > * Pull everything together: check_access() below contains the performance > > * critical operations; the fast-path (including check_access) functions > > should > > @@ -350,11 +400,13 @@ static noinline void kcsan_found_watchpoint(const > > volatile void *ptr, > > flags = user_access_save(); > > > > if (consumed) { > > - kcsan_save_irqtrace(current); > > + struct kcsan_irq_state irqstate; > > + > > + kcsan_irq_save(&irqstate); > > kcsan_report(ptr, size, type, KCSAN_VALUE_CHANGE_MAYBE, > > KCSAN_REPORT_CONSUMED_WATCHPOINT, > > watchpoint - watchpoints); > > - kcsan_restore_irqtrace(current); > > + kcsan_irq_restore(&irqstate); > > } else { > > /* > > * The other thread may not print any diagnostics, as it has > > @@ -387,7 +439,7 @@ kcsan_setup_watchpoint(const volatile void *ptr, size_t > > size, int type) > > unsigned long access_mask; > > enum kcsan_value_change value_change = KCSAN_VALUE_CHANGE_MAYBE; > > unsigned long ua_flags = user_access_save(); > > - unsigned long irq_flags = 0; > > + struct kcsan_irq_state irqstate; > > > > /* > > * Always reset kcsan_skip counter in slow-path to avoid underflow; > > see > > @@ -412,14 +464,7 @@ kcsan_setup_watchpoint(const volatile void *ptr, > > size_t size, int type) > > goto out; > > } > > > > - /* > > - * Save and restore the IRQ state trace touched by KCSAN, since > > KCSAN's > > - * runtime is entered for every memory access, and potentially useful > > - * information is lost if dirtied by KCSAN. > > - */ > > - kcsan_save_irqtrace(current); > > - if (!kcsan_interrupt_watcher) > > - local_irq_save(irq_flags); > > + kcsan_irq_save(&irqstate); > > > > watchpoint = insert_watchpoint((unsigned long)ptr, size, is_write); > > if (watchpoint == NULL) { > > @@ -559,9 +604,7 @@ kcsan_setup_watchpoint(const volatile void *ptr, size_t > > size, int type) > > remove_watchpoint(watchpoint); > > kcsan_counter_dec(KCSAN_COUNTER_USED_WATCHPOINTS); > > out_unlock: > > - if (!kcsan_interrupt_watcher) > > - local_irq_restore(irq_flags); > > - kcsan_restore_irqtrace(current); > > + kcsan_irq_restore(&irqstate); > > out: > > user_access_restore(ua_flags); > > } > > diff --git a/kernel/kcsan/kcsan.h b/kernel/kcsan/kcsan.h > > index 29480010dc30..6eb35a9514d8 100644 > > --- a/kernel/kcsan/kcsan.h > > +++ b/kernel/kcsan/kcsan.h > > @@ -24,9 +24,8 @@ extern unsigned int kcsan_udelay_interrupt; > > extern bool kcsan_enabled; > > > > /* > > - * Save/restore IRQ flags state trace dirtied by KCSAN. > > + * Restore IRQ flags state trace dirtied by KCSAN. > > */ > > -void kcsan_save_irqtrace(struct task_struct *task); > > void kcsan_restore_irqtrace(struct task_struct *task); > > > > /* > > -- > > 2.28.0.236.gb10cc79966-goog > > > > -- > You received this message because you are subscribed to the Google Groups > "kasan-dev" group. > To unsubscribe from this group and stop receiving emails from it, send an > email to kasan-dev+unsubscr...@googlegroups.com. > To view this discussion on the web visit > https://groups.google.com/d/msgid/kasan-dev/20200807170618.GW4295%40paulmck-ThinkPad-P72.