Re: [PATCH] kcsan: Treat runtime as NMI-like with interrupt tracing

From: Marco Elver
Date: Mon Aug 10 2020 - 04:08:02 EST


On Fri, 7 Aug 2020 at 19:06, Paul E. McKenney <paulmck@xxxxxxxxxx> wrote:
> On Fri, Aug 07, 2020 at 11:00:31AM +0200, Marco Elver wrote:
> > Since KCSAN instrumentation is everywhere, we need to treat the hooks
> > NMI-like for interrupt tracing. In order to present an as 'normal' as
> > possible context to the code called by KCSAN when reporting errors, we
> > need to update the IRQ-tracing state.
> >
> > Tested: Several runs through kcsan-test with different configuration
> > (PROVE_LOCKING on/off), as well as hours of syzbot testing with the
> > original config that caught the problem (without CONFIG_PARAVIRT=y,
> > which appears to cause IRQ state tracking inconsistencies even when
> > KCSAN remains off, see Link).
> >
> > Link: https://lkml.kernel.org/r/0000000000007d3b2d05ac1c303e@xxxxxxxxxx
> > Fixes: 248591f5d257 ("kcsan: Make KCSAN compatible with new IRQ state tracking")
> > Reported-by: syzbot+8db9e1ecde74e590a657@xxxxxxxxxxxxxxxxxxxxxxxxx
> > Co-developed-by: Peter Zijlstra (Intel) <peterz@xxxxxxxxxxxxx>

Peter, if you're fine with it, I think we'll require your
Signed-off-by (since Co-developed-by).

> > Signed-off-by: Marco Elver <elver@xxxxxxxxxx>
> > ---
> > Patch Note: This patch applies to latest mainline. While current
> > mainline suffers from the above problem, the configs required to hit the
> > issue are likely not enabled too often (of course with PROVE_LOCKING on;
> > we hit it on syzbot though). It'll probably be wise to queue this as
> > normal on -rcu, just in case something is still off, given the
> > non-trivial nature of the issue. (If it should instead go to mainline
> > right now as a fix, I'd like some more test time on syzbot.)
>
> The usual, please let me know when/if you would like me to apply
> to -rcu. And have a great weekend!

I think we need to wait until you have rebased -rcu to 5.9-rc1 some
time next week. I will send a reminder after, and if it doesn't apply
cleanly, I'll send a rebased patch.

Thank you!

-- Marco

> Thanx, Paul
>
> > ---
> > kernel/kcsan/core.c | 79 ++++++++++++++++++++++++++++++++++----------
> > kernel/kcsan/kcsan.h | 3 +-
> > 2 files changed, 62 insertions(+), 20 deletions(-)
> >
> > diff --git a/kernel/kcsan/core.c b/kernel/kcsan/core.c
> > index 9147ff6a12e5..6202a645f1e2 100644
> > --- a/kernel/kcsan/core.c
> > +++ b/kernel/kcsan/core.c
> > @@ -291,13 +291,28 @@ static inline unsigned int get_delay(void)
> > 0);
> > }
> >
> > -void kcsan_save_irqtrace(struct task_struct *task)
> > -{
> > +/*
> > + * KCSAN instrumentation is everywhere, which means we must treat the hooks
> > + * NMI-like for interrupt tracing. In order to present a 'normal' as possible
> > + * context to the code called by KCSAN when reporting errors we need to update
> > + * the IRQ-tracing state.
> > + *
> > + * Save and restore the IRQ state trace touched by KCSAN, since KCSAN's
> > + * runtime is entered for every memory access, and potentially useful
> > + * information is lost if dirtied by KCSAN.
> > + */
> > +
> > +struct kcsan_irq_state {
> > + unsigned long flags;
> > #ifdef CONFIG_TRACE_IRQFLAGS
> > - task->kcsan_save_irqtrace = task->irqtrace;
> > + int hardirqs_enabled;
> > #endif
> > -}
> > +};
> >
> > +/*
> > + * This is also called by the reporting task for the other task, to generate the
> > + * right report with CONFIG_KCSAN_VERBOSE. No harm in restoring more than once.
> > + */
> > void kcsan_restore_irqtrace(struct task_struct *task)
> > {
> > #ifdef CONFIG_TRACE_IRQFLAGS
> > @@ -305,6 +320,41 @@ void kcsan_restore_irqtrace(struct task_struct *task)
> > #endif
> > }
> >
> > +/*
> > + * Saves/restores IRQ state (see comment above). Need noinline to work around
> > + * unfortunate code-gen upon inlining, resulting in objtool getting confused as
> > + * well as losing stack trace information.
> > + */
> > +static noinline void kcsan_irq_save(struct kcsan_irq_state *irq_state)
> > +{
> > +#ifdef CONFIG_TRACE_IRQFLAGS
> > + current->kcsan_save_irqtrace = current->irqtrace;
> > + irq_state->hardirqs_enabled = lockdep_hardirqs_enabled();
> > +#endif
> > + if (!kcsan_interrupt_watcher) {
> > + kcsan_disable_current(); /* Lockdep might WARN, etc. */
> > + raw_local_irq_save(irq_state->flags);
> > + lockdep_hardirqs_off(_RET_IP_);
> > + kcsan_enable_current();
> > + }
> > +}
> > +
> > +static noinline void kcsan_irq_restore(struct kcsan_irq_state *irq_state)
> > +{
> > + if (!kcsan_interrupt_watcher) {
> > + kcsan_disable_current(); /* Lockdep might WARN, etc. */
> > +#ifdef CONFIG_TRACE_IRQFLAGS
> > + if (irq_state->hardirqs_enabled) {
> > + lockdep_hardirqs_on_prepare(_RET_IP_);
> > + lockdep_hardirqs_on(_RET_IP_);
> > + }
> > +#endif
> > + raw_local_irq_restore(irq_state->flags);
> > + kcsan_enable_current();
> > + }
> > + kcsan_restore_irqtrace(current);
> > +}
> > +
> > /*
> > * Pull everything together: check_access() below contains the performance
> > * critical operations; the fast-path (including check_access) functions should
> > @@ -350,11 +400,13 @@ static noinline void kcsan_found_watchpoint(const volatile void *ptr,
> > flags = user_access_save();
> >
> > if (consumed) {
> > - kcsan_save_irqtrace(current);
> > + struct kcsan_irq_state irqstate;
> > +
> > + kcsan_irq_save(&irqstate);
> > kcsan_report(ptr, size, type, KCSAN_VALUE_CHANGE_MAYBE,
> > KCSAN_REPORT_CONSUMED_WATCHPOINT,
> > watchpoint - watchpoints);
> > - kcsan_restore_irqtrace(current);
> > + kcsan_irq_restore(&irqstate);
> > } else {
> > /*
> > * The other thread may not print any diagnostics, as it has
> > @@ -387,7 +439,7 @@ kcsan_setup_watchpoint(const volatile void *ptr, size_t size, int type)
> > unsigned long access_mask;
> > enum kcsan_value_change value_change = KCSAN_VALUE_CHANGE_MAYBE;
> > unsigned long ua_flags = user_access_save();
> > - unsigned long irq_flags = 0;
> > + struct kcsan_irq_state irqstate;
> >
> > /*
> > * Always reset kcsan_skip counter in slow-path to avoid underflow; see
> > @@ -412,14 +464,7 @@ kcsan_setup_watchpoint(const volatile void *ptr, size_t size, int type)
> > goto out;
> > }
> >
> > - /*
> > - * Save and restore the IRQ state trace touched by KCSAN, since KCSAN's
> > - * runtime is entered for every memory access, and potentially useful
> > - * information is lost if dirtied by KCSAN.
> > - */
> > - kcsan_save_irqtrace(current);
> > - if (!kcsan_interrupt_watcher)
> > - local_irq_save(irq_flags);
> > + kcsan_irq_save(&irqstate);
> >
> > watchpoint = insert_watchpoint((unsigned long)ptr, size, is_write);
> > if (watchpoint == NULL) {
> > @@ -559,9 +604,7 @@ kcsan_setup_watchpoint(const volatile void *ptr, size_t size, int type)
> > remove_watchpoint(watchpoint);
> > kcsan_counter_dec(KCSAN_COUNTER_USED_WATCHPOINTS);
> > out_unlock:
> > - if (!kcsan_interrupt_watcher)
> > - local_irq_restore(irq_flags);
> > - kcsan_restore_irqtrace(current);
> > + kcsan_irq_restore(&irqstate);
> > out:
> > user_access_restore(ua_flags);
> > }
> > diff --git a/kernel/kcsan/kcsan.h b/kernel/kcsan/kcsan.h
> > index 29480010dc30..6eb35a9514d8 100644
> > --- a/kernel/kcsan/kcsan.h
> > +++ b/kernel/kcsan/kcsan.h
> > @@ -24,9 +24,8 @@ extern unsigned int kcsan_udelay_interrupt;
> > extern bool kcsan_enabled;
> >
> > /*
> > - * Save/restore IRQ flags state trace dirtied by KCSAN.
> > + * Restore IRQ flags state trace dirtied by KCSAN.
> > */
> > -void kcsan_save_irqtrace(struct task_struct *task);
> > void kcsan_restore_irqtrace(struct task_struct *task);
> >
> > /*
> > --
> > 2.28.0.236.gb10cc79966-goog
> >
>
> --
> You received this message because you are subscribed to the Google Groups "kasan-dev" group.
> To unsubscribe from this group and stop receiving emails from it, send an email to kasan-dev+unsubscribe@xxxxxxxxxxxxxxxx.
> To view this discussion on the web visit https://groups.google.com/d/msgid/kasan-dev/20200807170618.GW4295%40paulmck-ThinkPad-P72.