kcsan: Use tracing-safe version of prandom

In the core runtime, we must minimize any calls to external library
functions to avoid any kind of recursion. This can happen even though
instrumentation is disabled for called functions, but tracing is
enabled.

Most recently, prandom_u32() added a tracepoint, which can cause
problems for KCSAN even if the rcuidle variant is used. For example:
	kcsan -> prandom_u32() -> trace_prandom_u32_rcuidle ->
	srcu_read_lock_notrace -> __srcu_read_lock -> kcsan ...

While we could disable KCSAN in kcsan_setup_watchpoint(), this does not
solve other unexpected behaviour we may get due recursing into functions
that may not be tolerant to such recursion:
	__srcu_read_lock -> kcsan -> ... -> __srcu_read_lock

Therefore, switch to using prandom_u32_state(), which is uninstrumented,
and does not have a tracepoint.

Link: https://lkml.kernel.org/r/20200821063043.1949509-1-elver@google.com
Link: https://lkml.kernel.org/r/20200820172046.GA177701@elver.google.com
Signed-off-by: Marco Elver <elver@google.com>
Signed-off-by: Paul E. McKenney <paulmck@kernel.org>
This commit is contained in:
Marco Elver 2020-08-21 14:31:26 +02:00 committed by Paul E. McKenney
parent 068df05363
commit cd290ec246

View File

@ -100,6 +100,9 @@ static atomic_long_t watchpoints[CONFIG_KCSAN_NUM_WATCHPOINTS + NUM_SLOTS-1];
*/ */
static DEFINE_PER_CPU(long, kcsan_skip); static DEFINE_PER_CPU(long, kcsan_skip);
/* For kcsan_prandom_u32_max(). */
static DEFINE_PER_CPU(struct rnd_state, kcsan_rand_state);
static __always_inline atomic_long_t *find_watchpoint(unsigned long addr, static __always_inline atomic_long_t *find_watchpoint(unsigned long addr,
size_t size, size_t size,
bool expect_write, bool expect_write,
@ -271,11 +274,28 @@ should_watch(const volatile void *ptr, size_t size, int type, struct kcsan_ctx *
return true; return true;
} }
/*
* Returns a pseudo-random number in interval [0, ep_ro). See prandom_u32_max()
* for more details.
*
* The open-coded version here is using only safe primitives for all contexts
* where we can have KCSAN instrumentation. In particular, we cannot use
* prandom_u32() directly, as its tracepoint could cause recursion.
*/
static u32 kcsan_prandom_u32_max(u32 ep_ro)
{
struct rnd_state *state = &get_cpu_var(kcsan_rand_state);
const u32 res = prandom_u32_state(state);
put_cpu_var(kcsan_rand_state);
return (u32)(((u64) res * ep_ro) >> 32);
}
static inline void reset_kcsan_skip(void) static inline void reset_kcsan_skip(void)
{ {
long skip_count = kcsan_skip_watch - long skip_count = kcsan_skip_watch -
(IS_ENABLED(CONFIG_KCSAN_SKIP_WATCH_RANDOMIZE) ? (IS_ENABLED(CONFIG_KCSAN_SKIP_WATCH_RANDOMIZE) ?
prandom_u32_max(kcsan_skip_watch) : kcsan_prandom_u32_max(kcsan_skip_watch) :
0); 0);
this_cpu_write(kcsan_skip, skip_count); this_cpu_write(kcsan_skip, skip_count);
} }
@ -285,16 +305,18 @@ static __always_inline bool kcsan_is_enabled(void)
return READ_ONCE(kcsan_enabled) && get_ctx()->disable_count == 0; return READ_ONCE(kcsan_enabled) && get_ctx()->disable_count == 0;
} }
static inline unsigned int get_delay(int type) /* Introduce delay depending on context and configuration. */
static void delay_access(int type)
{ {
unsigned int delay = in_task() ? kcsan_udelay_task : kcsan_udelay_interrupt; unsigned int delay = in_task() ? kcsan_udelay_task : kcsan_udelay_interrupt;
/* For certain access types, skew the random delay to be longer. */ /* For certain access types, skew the random delay to be longer. */
unsigned int skew_delay_order = unsigned int skew_delay_order =
(type & (KCSAN_ACCESS_COMPOUND | KCSAN_ACCESS_ASSERT)) ? 1 : 0; (type & (KCSAN_ACCESS_COMPOUND | KCSAN_ACCESS_ASSERT)) ? 1 : 0;
return delay - (IS_ENABLED(CONFIG_KCSAN_DELAY_RANDOMIZE) ? delay -= IS_ENABLED(CONFIG_KCSAN_DELAY_RANDOMIZE) ?
prandom_u32_max(delay >> skew_delay_order) : kcsan_prandom_u32_max(delay >> skew_delay_order) :
0); 0;
udelay(delay);
} }
void kcsan_save_irqtrace(struct task_struct *task) void kcsan_save_irqtrace(struct task_struct *task)
@ -476,7 +498,7 @@ kcsan_setup_watchpoint(const volatile void *ptr, size_t size, int type)
* Delay this thread, to increase probability of observing a racy * Delay this thread, to increase probability of observing a racy
* conflicting access. * conflicting access.
*/ */
udelay(get_delay(type)); delay_access(type);
/* /*
* Re-read value, and check if it is as expected; if not, we infer a * Re-read value, and check if it is as expected; if not, we infer a
@ -620,6 +642,7 @@ void __init kcsan_init(void)
BUG_ON(!in_task()); BUG_ON(!in_task());
kcsan_debugfs_init(); kcsan_debugfs_init();
prandom_seed_full_state(&kcsan_rand_state);
/* /*
* We are in the init task, and no other tasks should be running; * We are in the init task, and no other tasks should be running;