forked from luck/tmp_suning_uos_patched
4b21e8abf9
Use the lockless GVA helpers to implement the dynamic translation of guest instructions. This will allow it to handle asynchronous TLB flushes when they are implemented. Signed-off-by: James Hogan <james.hogan@imgtec.com> Cc: Paolo Bonzini <pbonzini@redhat.com> Cc: "Radim Krčmář" <rkrcmar@redhat.com> Cc: Ralf Baechle <ralf@linux-mips.org> Cc: linux-mips@linux-mips.org Cc: kvm@vger.kernel.org
144 lines
4.0 KiB
C
144 lines
4.0 KiB
C
/*
|
|
* This file is subject to the terms and conditions of the GNU General Public
|
|
* License. See the file "COPYING" in the main directory of this archive
|
|
* for more details.
|
|
*
|
|
* KVM/MIPS: Binary Patching for privileged instructions, reduces traps.
|
|
*
|
|
* Copyright (C) 2012 MIPS Technologies, Inc. All rights reserved.
|
|
* Authors: Sanjay Lal <sanjayl@kymasys.com>
|
|
*/
|
|
|
|
#include <linux/errno.h>
|
|
#include <linux/err.h>
|
|
#include <linux/highmem.h>
|
|
#include <linux/kvm_host.h>
|
|
#include <linux/uaccess.h>
|
|
#include <linux/vmalloc.h>
|
|
#include <linux/fs.h>
|
|
#include <linux/bootmem.h>
|
|
#include <asm/cacheflush.h>
|
|
|
|
#include "commpage.h"
|
|
|
|
/**
|
|
* kvm_mips_trans_replace() - Replace trapping instruction in guest memory.
|
|
* @vcpu: Virtual CPU.
|
|
* @opc: PC of instruction to replace.
|
|
* @replace: Instruction to write
|
|
*/
|
|
static int kvm_mips_trans_replace(struct kvm_vcpu *vcpu, u32 *opc,
|
|
union mips_instruction replace)
|
|
{
|
|
unsigned long vaddr = (unsigned long)opc;
|
|
int err;
|
|
|
|
retry:
|
|
/* The GVA page table is still active so use the Linux TLB handlers */
|
|
kvm_trap_emul_gva_lockless_begin(vcpu);
|
|
err = put_user(replace.word, opc);
|
|
kvm_trap_emul_gva_lockless_end(vcpu);
|
|
|
|
if (unlikely(err)) {
|
|
/*
|
|
* We write protect clean pages in GVA page table so normal
|
|
* Linux TLB mod handler doesn't silently dirty the page.
|
|
* Its also possible we raced with a GVA invalidation.
|
|
* Try to force the page to become dirty.
|
|
*/
|
|
err = kvm_trap_emul_gva_fault(vcpu, vaddr, true);
|
|
if (unlikely(err)) {
|
|
kvm_info("%s: Address unwriteable: %p\n",
|
|
__func__, opc);
|
|
return -EFAULT;
|
|
}
|
|
|
|
/*
|
|
* Try again. This will likely trigger a TLB refill, which will
|
|
* fetch the new dirty entry from the GVA page table, which
|
|
* should then succeed.
|
|
*/
|
|
goto retry;
|
|
}
|
|
__local_flush_icache_user_range(vaddr, vaddr + 4);
|
|
|
|
return 0;
|
|
}
|
|
|
|
int kvm_mips_trans_cache_index(union mips_instruction inst, u32 *opc,
|
|
struct kvm_vcpu *vcpu)
|
|
{
|
|
union mips_instruction nop_inst = { 0 };
|
|
|
|
/* Replace the CACHE instruction, with a NOP */
|
|
return kvm_mips_trans_replace(vcpu, opc, nop_inst);
|
|
}
|
|
|
|
/*
|
|
* Address based CACHE instructions are transformed into synci(s). A little
|
|
* heavy for just D-cache invalidates, but avoids an expensive trap
|
|
*/
|
|
int kvm_mips_trans_cache_va(union mips_instruction inst, u32 *opc,
|
|
struct kvm_vcpu *vcpu)
|
|
{
|
|
union mips_instruction synci_inst = { 0 };
|
|
|
|
synci_inst.i_format.opcode = bcond_op;
|
|
synci_inst.i_format.rs = inst.i_format.rs;
|
|
synci_inst.i_format.rt = synci_op;
|
|
if (cpu_has_mips_r6)
|
|
synci_inst.i_format.simmediate = inst.spec3_format.simmediate;
|
|
else
|
|
synci_inst.i_format.simmediate = inst.i_format.simmediate;
|
|
|
|
return kvm_mips_trans_replace(vcpu, opc, synci_inst);
|
|
}
|
|
|
|
int kvm_mips_trans_mfc0(union mips_instruction inst, u32 *opc,
|
|
struct kvm_vcpu *vcpu)
|
|
{
|
|
union mips_instruction mfc0_inst = { 0 };
|
|
u32 rd, sel;
|
|
|
|
rd = inst.c0r_format.rd;
|
|
sel = inst.c0r_format.sel;
|
|
|
|
if (rd == MIPS_CP0_ERRCTL && sel == 0) {
|
|
mfc0_inst.r_format.opcode = spec_op;
|
|
mfc0_inst.r_format.rd = inst.c0r_format.rt;
|
|
mfc0_inst.r_format.func = add_op;
|
|
} else {
|
|
mfc0_inst.i_format.opcode = lw_op;
|
|
mfc0_inst.i_format.rt = inst.c0r_format.rt;
|
|
mfc0_inst.i_format.simmediate = KVM_GUEST_COMMPAGE_ADDR |
|
|
offsetof(struct kvm_mips_commpage, cop0.reg[rd][sel]);
|
|
#ifdef CONFIG_CPU_BIG_ENDIAN
|
|
if (sizeof(vcpu->arch.cop0->reg[0][0]) == 8)
|
|
mfc0_inst.i_format.simmediate |= 4;
|
|
#endif
|
|
}
|
|
|
|
return kvm_mips_trans_replace(vcpu, opc, mfc0_inst);
|
|
}
|
|
|
|
int kvm_mips_trans_mtc0(union mips_instruction inst, u32 *opc,
|
|
struct kvm_vcpu *vcpu)
|
|
{
|
|
union mips_instruction mtc0_inst = { 0 };
|
|
u32 rd, sel;
|
|
|
|
rd = inst.c0r_format.rd;
|
|
sel = inst.c0r_format.sel;
|
|
|
|
mtc0_inst.i_format.opcode = sw_op;
|
|
mtc0_inst.i_format.rt = inst.c0r_format.rt;
|
|
mtc0_inst.i_format.simmediate = KVM_GUEST_COMMPAGE_ADDR |
|
|
offsetof(struct kvm_mips_commpage, cop0.reg[rd][sel]);
|
|
#ifdef CONFIG_CPU_BIG_ENDIAN
|
|
if (sizeof(vcpu->arch.cop0->reg[0][0]) == 8)
|
|
mtc0_inst.i_format.simmediate |= 4;
|
|
#endif
|
|
|
|
return kvm_mips_trans_replace(vcpu, opc, mtc0_inst);
|
|
}
|