forked from luck/tmp_suning_uos_patched
uaccess: add force_uaccess_{begin,end} helpers
Add helpers to wrap the get_fs/set_fs magic for undoing any damange done by set_fs(KERNEL_DS). There is no real functional benefit, but this documents the intent of these calls better, and will allow stubbing the functions out easily for kernels builds that do not allow address space overrides in the future. [hch@lst.de: drop two incorrect hunks, fix a commit log typo] Link: http://lkml.kernel.org/r/20200714105505.935079-6-hch@lst.de Signed-off-by: Christoph Hellwig <hch@lst.de> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Acked-by: Linus Torvalds <torvalds@linux-foundation.org> Acked-by: Mark Rutland <mark.rutland@arm.com> Acked-by: Greentime Hu <green.hu@gmail.com> Acked-by: Geert Uytterhoeven <geert@linux-m68k.org> Cc: Nick Hu <nickhu@andestech.com> Cc: Vincent Chen <deanbo422@gmail.com> Cc: Paul Walmsley <paul.walmsley@sifive.com> Cc: Palmer Dabbelt <palmer@dabbelt.com> Link: http://lkml.kernel.org/r/20200710135706.537715-6-hch@lst.de Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
This commit is contained in:
parent
428e2976a5
commit
3d13f313ce
|
@ -180,7 +180,7 @@ static __kprobes unsigned long _sdei_handler(struct pt_regs *regs,
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* We didn't take an exception to get here, set PAN. UAO will be cleared
|
* We didn't take an exception to get here, set PAN. UAO will be cleared
|
||||||
* by sdei_event_handler()s set_fs(USER_DS) call.
|
* by sdei_event_handler()s force_uaccess_begin() call.
|
||||||
*/
|
*/
|
||||||
__uaccess_enable_hw_pan();
|
__uaccess_enable_hw_pan();
|
||||||
|
|
||||||
|
|
|
@ -85,10 +85,10 @@ static inline void flush_tlb_mm(struct mm_struct *mm)
|
||||||
static inline void flush_tlb_page(struct vm_area_struct *vma, unsigned long addr)
|
static inline void flush_tlb_page(struct vm_area_struct *vma, unsigned long addr)
|
||||||
{
|
{
|
||||||
if (vma->vm_mm == current->active_mm) {
|
if (vma->vm_mm == current->active_mm) {
|
||||||
mm_segment_t old_fs = get_fs();
|
mm_segment_t old_fs = force_uaccess_begin();
|
||||||
set_fs(USER_DS);
|
|
||||||
__flush_tlb_one(addr);
|
__flush_tlb_one(addr);
|
||||||
set_fs(old_fs);
|
force_uaccess_end(old_fs);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -191,17 +191,16 @@ static void emulate_load_store_insn(struct pt_regs *regs,
|
||||||
* memory, so we need to "switch" the address limit to
|
* memory, so we need to "switch" the address limit to
|
||||||
* user space, so that address check can work properly.
|
* user space, so that address check can work properly.
|
||||||
*/
|
*/
|
||||||
seg = get_fs();
|
seg = force_uaccess_begin();
|
||||||
set_fs(USER_DS);
|
|
||||||
switch (insn.spec3_format.func) {
|
switch (insn.spec3_format.func) {
|
||||||
case lhe_op:
|
case lhe_op:
|
||||||
if (!access_ok(addr, 2)) {
|
if (!access_ok(addr, 2)) {
|
||||||
set_fs(seg);
|
force_uaccess_end(seg);
|
||||||
goto sigbus;
|
goto sigbus;
|
||||||
}
|
}
|
||||||
LoadHWE(addr, value, res);
|
LoadHWE(addr, value, res);
|
||||||
if (res) {
|
if (res) {
|
||||||
set_fs(seg);
|
force_uaccess_end(seg);
|
||||||
goto fault;
|
goto fault;
|
||||||
}
|
}
|
||||||
compute_return_epc(regs);
|
compute_return_epc(regs);
|
||||||
|
@ -209,12 +208,12 @@ static void emulate_load_store_insn(struct pt_regs *regs,
|
||||||
break;
|
break;
|
||||||
case lwe_op:
|
case lwe_op:
|
||||||
if (!access_ok(addr, 4)) {
|
if (!access_ok(addr, 4)) {
|
||||||
set_fs(seg);
|
force_uaccess_end(seg);
|
||||||
goto sigbus;
|
goto sigbus;
|
||||||
}
|
}
|
||||||
LoadWE(addr, value, res);
|
LoadWE(addr, value, res);
|
||||||
if (res) {
|
if (res) {
|
||||||
set_fs(seg);
|
force_uaccess_end(seg);
|
||||||
goto fault;
|
goto fault;
|
||||||
}
|
}
|
||||||
compute_return_epc(regs);
|
compute_return_epc(regs);
|
||||||
|
@ -222,12 +221,12 @@ static void emulate_load_store_insn(struct pt_regs *regs,
|
||||||
break;
|
break;
|
||||||
case lhue_op:
|
case lhue_op:
|
||||||
if (!access_ok(addr, 2)) {
|
if (!access_ok(addr, 2)) {
|
||||||
set_fs(seg);
|
force_uaccess_end(seg);
|
||||||
goto sigbus;
|
goto sigbus;
|
||||||
}
|
}
|
||||||
LoadHWUE(addr, value, res);
|
LoadHWUE(addr, value, res);
|
||||||
if (res) {
|
if (res) {
|
||||||
set_fs(seg);
|
force_uaccess_end(seg);
|
||||||
goto fault;
|
goto fault;
|
||||||
}
|
}
|
||||||
compute_return_epc(regs);
|
compute_return_epc(regs);
|
||||||
|
@ -235,35 +234,35 @@ static void emulate_load_store_insn(struct pt_regs *regs,
|
||||||
break;
|
break;
|
||||||
case she_op:
|
case she_op:
|
||||||
if (!access_ok(addr, 2)) {
|
if (!access_ok(addr, 2)) {
|
||||||
set_fs(seg);
|
force_uaccess_end(seg);
|
||||||
goto sigbus;
|
goto sigbus;
|
||||||
}
|
}
|
||||||
compute_return_epc(regs);
|
compute_return_epc(regs);
|
||||||
value = regs->regs[insn.spec3_format.rt];
|
value = regs->regs[insn.spec3_format.rt];
|
||||||
StoreHWE(addr, value, res);
|
StoreHWE(addr, value, res);
|
||||||
if (res) {
|
if (res) {
|
||||||
set_fs(seg);
|
force_uaccess_end(seg);
|
||||||
goto fault;
|
goto fault;
|
||||||
}
|
}
|
||||||
break;
|
break;
|
||||||
case swe_op:
|
case swe_op:
|
||||||
if (!access_ok(addr, 4)) {
|
if (!access_ok(addr, 4)) {
|
||||||
set_fs(seg);
|
force_uaccess_end(seg);
|
||||||
goto sigbus;
|
goto sigbus;
|
||||||
}
|
}
|
||||||
compute_return_epc(regs);
|
compute_return_epc(regs);
|
||||||
value = regs->regs[insn.spec3_format.rt];
|
value = regs->regs[insn.spec3_format.rt];
|
||||||
StoreWE(addr, value, res);
|
StoreWE(addr, value, res);
|
||||||
if (res) {
|
if (res) {
|
||||||
set_fs(seg);
|
force_uaccess_end(seg);
|
||||||
goto fault;
|
goto fault;
|
||||||
}
|
}
|
||||||
break;
|
break;
|
||||||
default:
|
default:
|
||||||
set_fs(seg);
|
force_uaccess_end(seg);
|
||||||
goto sigill;
|
goto sigill;
|
||||||
}
|
}
|
||||||
set_fs(seg);
|
force_uaccess_end(seg);
|
||||||
}
|
}
|
||||||
#endif
|
#endif
|
||||||
break;
|
break;
|
||||||
|
|
|
@ -512,7 +512,7 @@ int do_unaligned_access(unsigned long addr, struct pt_regs *regs)
|
||||||
{
|
{
|
||||||
unsigned long inst;
|
unsigned long inst;
|
||||||
int ret = -EFAULT;
|
int ret = -EFAULT;
|
||||||
mm_segment_t seg = get_fs();
|
mm_segment_t seg;
|
||||||
|
|
||||||
inst = get_inst(regs->ipc);
|
inst = get_inst(regs->ipc);
|
||||||
|
|
||||||
|
@ -520,13 +520,12 @@ int do_unaligned_access(unsigned long addr, struct pt_regs *regs)
|
||||||
"Faulting addr: 0x%08lx, pc: 0x%08lx [inst: 0x%08lx ]\n", addr,
|
"Faulting addr: 0x%08lx, pc: 0x%08lx [inst: 0x%08lx ]\n", addr,
|
||||||
regs->ipc, inst);
|
regs->ipc, inst);
|
||||||
|
|
||||||
set_fs(USER_DS);
|
seg = force_uaccess_begin();
|
||||||
|
|
||||||
if (inst & NDS32_16BIT_INSTRUCTION)
|
if (inst & NDS32_16BIT_INSTRUCTION)
|
||||||
ret = do_16((inst >> 16) & 0xffff, regs);
|
ret = do_16((inst >> 16) & 0xffff, regs);
|
||||||
else
|
else
|
||||||
ret = do_32(inst, regs);
|
ret = do_32(inst, regs);
|
||||||
set_fs(seg);
|
force_uaccess_end(seg);
|
||||||
|
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
|
@ -482,8 +482,6 @@ asmlinkage void do_address_error(struct pt_regs *regs,
|
||||||
error_code = lookup_exception_vector();
|
error_code = lookup_exception_vector();
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
oldfs = get_fs();
|
|
||||||
|
|
||||||
if (user_mode(regs)) {
|
if (user_mode(regs)) {
|
||||||
int si_code = BUS_ADRERR;
|
int si_code = BUS_ADRERR;
|
||||||
unsigned int user_action;
|
unsigned int user_action;
|
||||||
|
@ -491,13 +489,13 @@ asmlinkage void do_address_error(struct pt_regs *regs,
|
||||||
local_irq_enable();
|
local_irq_enable();
|
||||||
inc_unaligned_user_access();
|
inc_unaligned_user_access();
|
||||||
|
|
||||||
set_fs(USER_DS);
|
oldfs = force_uaccess_begin();
|
||||||
if (copy_from_user(&instruction, (insn_size_t *)(regs->pc & ~1),
|
if (copy_from_user(&instruction, (insn_size_t *)(regs->pc & ~1),
|
||||||
sizeof(instruction))) {
|
sizeof(instruction))) {
|
||||||
set_fs(oldfs);
|
force_uaccess_end(oldfs);
|
||||||
goto uspace_segv;
|
goto uspace_segv;
|
||||||
}
|
}
|
||||||
set_fs(oldfs);
|
force_uaccess_end(oldfs);
|
||||||
|
|
||||||
/* shout about userspace fixups */
|
/* shout about userspace fixups */
|
||||||
unaligned_fixups_notify(current, instruction, regs);
|
unaligned_fixups_notify(current, instruction, regs);
|
||||||
|
@ -520,11 +518,11 @@ asmlinkage void do_address_error(struct pt_regs *regs,
|
||||||
goto uspace_segv;
|
goto uspace_segv;
|
||||||
}
|
}
|
||||||
|
|
||||||
set_fs(USER_DS);
|
oldfs = force_uaccess_begin();
|
||||||
tmp = handle_unaligned_access(instruction, regs,
|
tmp = handle_unaligned_access(instruction, regs,
|
||||||
&user_mem_access, 0,
|
&user_mem_access, 0,
|
||||||
address);
|
address);
|
||||||
set_fs(oldfs);
|
force_uaccess_end(oldfs);
|
||||||
|
|
||||||
if (tmp == 0)
|
if (tmp == 0)
|
||||||
return; /* sorted */
|
return; /* sorted */
|
||||||
|
|
|
@ -1136,15 +1136,14 @@ int sdei_event_handler(struct pt_regs *regs,
|
||||||
* access kernel memory.
|
* access kernel memory.
|
||||||
* Do the same here because this doesn't come via the same entry code.
|
* Do the same here because this doesn't come via the same entry code.
|
||||||
*/
|
*/
|
||||||
orig_addr_limit = get_fs();
|
orig_addr_limit = force_uaccess_begin();
|
||||||
set_fs(USER_DS);
|
|
||||||
|
|
||||||
err = arg->callback(event_num, regs, arg->callback_arg);
|
err = arg->callback(event_num, regs, arg->callback_arg);
|
||||||
if (err)
|
if (err)
|
||||||
pr_err_ratelimited("event %u on CPU %u failed with error: %d\n",
|
pr_err_ratelimited("event %u on CPU %u failed with error: %d\n",
|
||||||
event_num, smp_processor_id(), err);
|
event_num, smp_processor_id(), err);
|
||||||
|
|
||||||
set_fs(orig_addr_limit);
|
force_uaccess_end(orig_addr_limit);
|
||||||
|
|
||||||
return err;
|
return err;
|
||||||
}
|
}
|
||||||
|
|
|
@ -8,6 +8,24 @@
|
||||||
|
|
||||||
#include <asm/uaccess.h>
|
#include <asm/uaccess.h>
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Force the uaccess routines to be wired up for actual userspace access,
|
||||||
|
* overriding any possible set_fs(KERNEL_DS) still lingering around. Undone
|
||||||
|
* using force_uaccess_end below.
|
||||||
|
*/
|
||||||
|
static inline mm_segment_t force_uaccess_begin(void)
|
||||||
|
{
|
||||||
|
mm_segment_t fs = get_fs();
|
||||||
|
|
||||||
|
set_fs(USER_DS);
|
||||||
|
return fs;
|
||||||
|
}
|
||||||
|
|
||||||
|
static inline void force_uaccess_end(mm_segment_t oldfs)
|
||||||
|
{
|
||||||
|
set_fs(oldfs);
|
||||||
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Architectures should provide two primitives (raw_copy_{to,from}_user())
|
* Architectures should provide two primitives (raw_copy_{to,from}_user())
|
||||||
* and get rid of their private instances of copy_{to,from}_user() and
|
* and get rid of their private instances of copy_{to,from}_user() and
|
||||||
|
|
|
@ -217,10 +217,9 @@ get_perf_callchain(struct pt_regs *regs, u32 init_nr, bool kernel, bool user,
|
||||||
if (add_mark)
|
if (add_mark)
|
||||||
perf_callchain_store_context(&ctx, PERF_CONTEXT_USER);
|
perf_callchain_store_context(&ctx, PERF_CONTEXT_USER);
|
||||||
|
|
||||||
fs = get_fs();
|
fs = force_uaccess_begin();
|
||||||
set_fs(USER_DS);
|
|
||||||
perf_callchain_user(&ctx, regs);
|
perf_callchain_user(&ctx, regs);
|
||||||
set_fs(fs);
|
force_uaccess_end(fs);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -6453,10 +6453,9 @@ perf_output_sample_ustack(struct perf_output_handle *handle, u64 dump_size,
|
||||||
|
|
||||||
/* Data. */
|
/* Data. */
|
||||||
sp = perf_user_stack_pointer(regs);
|
sp = perf_user_stack_pointer(regs);
|
||||||
fs = get_fs();
|
fs = force_uaccess_begin();
|
||||||
set_fs(USER_DS);
|
|
||||||
rem = __output_copy_user(handle, (void *) sp, dump_size);
|
rem = __output_copy_user(handle, (void *) sp, dump_size);
|
||||||
set_fs(fs);
|
force_uaccess_end(fs);
|
||||||
dyn_size = dump_size - rem;
|
dyn_size = dump_size - rem;
|
||||||
|
|
||||||
perf_output_skip(handle, rem);
|
perf_output_skip(handle, rem);
|
||||||
|
|
|
@ -1258,8 +1258,7 @@ void kthread_use_mm(struct mm_struct *mm)
|
||||||
if (active_mm != mm)
|
if (active_mm != mm)
|
||||||
mmdrop(active_mm);
|
mmdrop(active_mm);
|
||||||
|
|
||||||
to_kthread(tsk)->oldfs = get_fs();
|
to_kthread(tsk)->oldfs = force_uaccess_begin();
|
||||||
set_fs(USER_DS);
|
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL_GPL(kthread_use_mm);
|
EXPORT_SYMBOL_GPL(kthread_use_mm);
|
||||||
|
|
||||||
|
@ -1274,7 +1273,7 @@ void kthread_unuse_mm(struct mm_struct *mm)
|
||||||
WARN_ON_ONCE(!(tsk->flags & PF_KTHREAD));
|
WARN_ON_ONCE(!(tsk->flags & PF_KTHREAD));
|
||||||
WARN_ON_ONCE(!tsk->mm);
|
WARN_ON_ONCE(!tsk->mm);
|
||||||
|
|
||||||
set_fs(to_kthread(tsk)->oldfs);
|
force_uaccess_end(to_kthread(tsk)->oldfs);
|
||||||
|
|
||||||
task_lock(tsk);
|
task_lock(tsk);
|
||||||
sync_mm_rss(mm);
|
sync_mm_rss(mm);
|
||||||
|
|
|
@ -233,10 +233,9 @@ unsigned int stack_trace_save_user(unsigned long *store, unsigned int size)
|
||||||
if (current->flags & PF_KTHREAD)
|
if (current->flags & PF_KTHREAD)
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
fs = get_fs();
|
fs = force_uaccess_begin();
|
||||||
set_fs(USER_DS);
|
|
||||||
arch_stack_walk_user(consume_entry, &c, task_pt_regs(current));
|
arch_stack_walk_user(consume_entry, &c, task_pt_regs(current));
|
||||||
set_fs(fs);
|
force_uaccess_end(fs);
|
||||||
|
|
||||||
return c.len;
|
return c.len;
|
||||||
}
|
}
|
||||||
|
|
22
mm/maccess.c
22
mm/maccess.c
|
@ -205,15 +205,14 @@ long strncpy_from_kernel_nofault(char *dst, const void *unsafe_addr, long count)
|
||||||
long copy_from_user_nofault(void *dst, const void __user *src, size_t size)
|
long copy_from_user_nofault(void *dst, const void __user *src, size_t size)
|
||||||
{
|
{
|
||||||
long ret = -EFAULT;
|
long ret = -EFAULT;
|
||||||
mm_segment_t old_fs = get_fs();
|
mm_segment_t old_fs = force_uaccess_begin();
|
||||||
|
|
||||||
set_fs(USER_DS);
|
|
||||||
if (access_ok(src, size)) {
|
if (access_ok(src, size)) {
|
||||||
pagefault_disable();
|
pagefault_disable();
|
||||||
ret = __copy_from_user_inatomic(dst, src, size);
|
ret = __copy_from_user_inatomic(dst, src, size);
|
||||||
pagefault_enable();
|
pagefault_enable();
|
||||||
}
|
}
|
||||||
set_fs(old_fs);
|
force_uaccess_end(old_fs);
|
||||||
|
|
||||||
if (ret)
|
if (ret)
|
||||||
return -EFAULT;
|
return -EFAULT;
|
||||||
|
@ -233,15 +232,14 @@ EXPORT_SYMBOL_GPL(copy_from_user_nofault);
|
||||||
long copy_to_user_nofault(void __user *dst, const void *src, size_t size)
|
long copy_to_user_nofault(void __user *dst, const void *src, size_t size)
|
||||||
{
|
{
|
||||||
long ret = -EFAULT;
|
long ret = -EFAULT;
|
||||||
mm_segment_t old_fs = get_fs();
|
mm_segment_t old_fs = force_uaccess_begin();
|
||||||
|
|
||||||
set_fs(USER_DS);
|
|
||||||
if (access_ok(dst, size)) {
|
if (access_ok(dst, size)) {
|
||||||
pagefault_disable();
|
pagefault_disable();
|
||||||
ret = __copy_to_user_inatomic(dst, src, size);
|
ret = __copy_to_user_inatomic(dst, src, size);
|
||||||
pagefault_enable();
|
pagefault_enable();
|
||||||
}
|
}
|
||||||
set_fs(old_fs);
|
force_uaccess_end(old_fs);
|
||||||
|
|
||||||
if (ret)
|
if (ret)
|
||||||
return -EFAULT;
|
return -EFAULT;
|
||||||
|
@ -270,17 +268,17 @@ EXPORT_SYMBOL_GPL(copy_to_user_nofault);
|
||||||
long strncpy_from_user_nofault(char *dst, const void __user *unsafe_addr,
|
long strncpy_from_user_nofault(char *dst, const void __user *unsafe_addr,
|
||||||
long count)
|
long count)
|
||||||
{
|
{
|
||||||
mm_segment_t old_fs = get_fs();
|
mm_segment_t old_fs;
|
||||||
long ret;
|
long ret;
|
||||||
|
|
||||||
if (unlikely(count <= 0))
|
if (unlikely(count <= 0))
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
set_fs(USER_DS);
|
old_fs = force_uaccess_begin();
|
||||||
pagefault_disable();
|
pagefault_disable();
|
||||||
ret = strncpy_from_user(dst, unsafe_addr, count);
|
ret = strncpy_from_user(dst, unsafe_addr, count);
|
||||||
pagefault_enable();
|
pagefault_enable();
|
||||||
set_fs(old_fs);
|
force_uaccess_end(old_fs);
|
||||||
|
|
||||||
if (ret >= count) {
|
if (ret >= count) {
|
||||||
ret = count;
|
ret = count;
|
||||||
|
@ -310,14 +308,14 @@ long strncpy_from_user_nofault(char *dst, const void __user *unsafe_addr,
|
||||||
*/
|
*/
|
||||||
long strnlen_user_nofault(const void __user *unsafe_addr, long count)
|
long strnlen_user_nofault(const void __user *unsafe_addr, long count)
|
||||||
{
|
{
|
||||||
mm_segment_t old_fs = get_fs();
|
mm_segment_t old_fs;
|
||||||
int ret;
|
int ret;
|
||||||
|
|
||||||
set_fs(USER_DS);
|
old_fs = force_uaccess_begin();
|
||||||
pagefault_disable();
|
pagefault_disable();
|
||||||
ret = strnlen_user(unsafe_addr, count);
|
ret = strnlen_user(unsafe_addr, count);
|
||||||
pagefault_enable();
|
pagefault_enable();
|
||||||
set_fs(old_fs);
|
force_uaccess_end(old_fs);
|
||||||
|
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
Loading…
Reference in New Issue
Block a user