forked from luck/tmp_suning_uos_patched
kcsan: Use tracing-safe version of prandom
In the core runtime, we must minimize any calls to external library functions to avoid any kind of recursion. This can happen even though instrumentation is disabled for called functions, but tracing is enabled. Most recently, prandom_u32() added a tracepoint, which can cause problems for KCSAN even if the rcuidle variant is used. For example: kcsan -> prandom_u32() -> trace_prandom_u32_rcuidle -> srcu_read_lock_notrace -> __srcu_read_lock -> kcsan ... While we could disable KCSAN in kcsan_setup_watchpoint(), this does not solve other unexpected behaviour we may get due recursing into functions that may not be tolerant to such recursion: __srcu_read_lock -> kcsan -> ... -> __srcu_read_lock Therefore, switch to using prandom_u32_state(), which is uninstrumented, and does not have a tracepoint. Link: https://lkml.kernel.org/r/20200821063043.1949509-1-elver@google.com Link: https://lkml.kernel.org/r/20200820172046.GA177701@elver.google.com Signed-off-by: Marco Elver <elver@google.com> Signed-off-by: Paul E. McKenney <paulmck@kernel.org>
This commit is contained in:
parent
068df05363
commit
cd290ec246
|
@ -100,6 +100,9 @@ static atomic_long_t watchpoints[CONFIG_KCSAN_NUM_WATCHPOINTS + NUM_SLOTS-1];
|
||||||
*/
|
*/
|
||||||
static DEFINE_PER_CPU(long, kcsan_skip);
|
static DEFINE_PER_CPU(long, kcsan_skip);
|
||||||
|
|
||||||
|
/* For kcsan_prandom_u32_max(). */
|
||||||
|
static DEFINE_PER_CPU(struct rnd_state, kcsan_rand_state);
|
||||||
|
|
||||||
static __always_inline atomic_long_t *find_watchpoint(unsigned long addr,
|
static __always_inline atomic_long_t *find_watchpoint(unsigned long addr,
|
||||||
size_t size,
|
size_t size,
|
||||||
bool expect_write,
|
bool expect_write,
|
||||||
|
@ -271,11 +274,28 @@ should_watch(const volatile void *ptr, size_t size, int type, struct kcsan_ctx *
|
||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Returns a pseudo-random number in interval [0, ep_ro). See prandom_u32_max()
|
||||||
|
* for more details.
|
||||||
|
*
|
||||||
|
* The open-coded version here is using only safe primitives for all contexts
|
||||||
|
* where we can have KCSAN instrumentation. In particular, we cannot use
|
||||||
|
* prandom_u32() directly, as its tracepoint could cause recursion.
|
||||||
|
*/
|
||||||
|
static u32 kcsan_prandom_u32_max(u32 ep_ro)
|
||||||
|
{
|
||||||
|
struct rnd_state *state = &get_cpu_var(kcsan_rand_state);
|
||||||
|
const u32 res = prandom_u32_state(state);
|
||||||
|
|
||||||
|
put_cpu_var(kcsan_rand_state);
|
||||||
|
return (u32)(((u64) res * ep_ro) >> 32);
|
||||||
|
}
|
||||||
|
|
||||||
static inline void reset_kcsan_skip(void)
|
static inline void reset_kcsan_skip(void)
|
||||||
{
|
{
|
||||||
long skip_count = kcsan_skip_watch -
|
long skip_count = kcsan_skip_watch -
|
||||||
(IS_ENABLED(CONFIG_KCSAN_SKIP_WATCH_RANDOMIZE) ?
|
(IS_ENABLED(CONFIG_KCSAN_SKIP_WATCH_RANDOMIZE) ?
|
||||||
prandom_u32_max(kcsan_skip_watch) :
|
kcsan_prandom_u32_max(kcsan_skip_watch) :
|
||||||
0);
|
0);
|
||||||
this_cpu_write(kcsan_skip, skip_count);
|
this_cpu_write(kcsan_skip, skip_count);
|
||||||
}
|
}
|
||||||
|
@ -285,16 +305,18 @@ static __always_inline bool kcsan_is_enabled(void)
|
||||||
return READ_ONCE(kcsan_enabled) && get_ctx()->disable_count == 0;
|
return READ_ONCE(kcsan_enabled) && get_ctx()->disable_count == 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
static inline unsigned int get_delay(int type)
|
/* Introduce delay depending on context and configuration. */
|
||||||
|
static void delay_access(int type)
|
||||||
{
|
{
|
||||||
unsigned int delay = in_task() ? kcsan_udelay_task : kcsan_udelay_interrupt;
|
unsigned int delay = in_task() ? kcsan_udelay_task : kcsan_udelay_interrupt;
|
||||||
/* For certain access types, skew the random delay to be longer. */
|
/* For certain access types, skew the random delay to be longer. */
|
||||||
unsigned int skew_delay_order =
|
unsigned int skew_delay_order =
|
||||||
(type & (KCSAN_ACCESS_COMPOUND | KCSAN_ACCESS_ASSERT)) ? 1 : 0;
|
(type & (KCSAN_ACCESS_COMPOUND | KCSAN_ACCESS_ASSERT)) ? 1 : 0;
|
||||||
|
|
||||||
return delay - (IS_ENABLED(CONFIG_KCSAN_DELAY_RANDOMIZE) ?
|
delay -= IS_ENABLED(CONFIG_KCSAN_DELAY_RANDOMIZE) ?
|
||||||
prandom_u32_max(delay >> skew_delay_order) :
|
kcsan_prandom_u32_max(delay >> skew_delay_order) :
|
||||||
0);
|
0;
|
||||||
|
udelay(delay);
|
||||||
}
|
}
|
||||||
|
|
||||||
void kcsan_save_irqtrace(struct task_struct *task)
|
void kcsan_save_irqtrace(struct task_struct *task)
|
||||||
|
@ -476,7 +498,7 @@ kcsan_setup_watchpoint(const volatile void *ptr, size_t size, int type)
|
||||||
* Delay this thread, to increase probability of observing a racy
|
* Delay this thread, to increase probability of observing a racy
|
||||||
* conflicting access.
|
* conflicting access.
|
||||||
*/
|
*/
|
||||||
udelay(get_delay(type));
|
delay_access(type);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Re-read value, and check if it is as expected; if not, we infer a
|
* Re-read value, and check if it is as expected; if not, we infer a
|
||||||
|
@ -620,6 +642,7 @@ void __init kcsan_init(void)
|
||||||
BUG_ON(!in_task());
|
BUG_ON(!in_task());
|
||||||
|
|
||||||
kcsan_debugfs_init();
|
kcsan_debugfs_init();
|
||||||
|
prandom_seed_full_state(&kcsan_rand_state);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* We are in the init task, and no other tasks should be running;
|
* We are in the init task, and no other tasks should be running;
|
||||||
|
|
Loading…
Reference in New Issue
Block a user