forked from luck/tmp_suning_uos_patched
139ec7c416
It turns out that the most called ops, by several orders of magnitude, are the interrupt manipulation ops. These are obvious candidates for patching, so mark them up and create infrastructure for it. The method used is that the ops structure has a patch function, which is called for each place which needs to be patched: this returns a number of instructions (the rest are NOP-padded). Usually we can spare a register (%eax) for the binary patched code to use, but in a couple of critical places in entry.S we can't: we make the clobbers explicit at the call site, and manually clobber the allowed registers in debug mode as an extra check. And: Don't abuse CONFIG_DEBUG_KERNEL, add CONFIG_DEBUG_PARAVIRT. And: AK: Fix warnings in x86-64 alternative.c build And: AK: Fix compilation with defconfig And: ^From: Andrew Morton <akpm@osdl.org> Some binutlises still like to emit references to __stop_parainstructions and __start_parainstructions. And: AK: Fix warnings about unused variables when PARAVIRT is disabled. Signed-off-by: Rusty Russell <rusty@rustcorp.com.au> Signed-off-by: Jeremy Fitzhardinge <jeremy@xensource.com> Signed-off-by: Chris Wright <chrisw@sous-sol.org> Signed-off-by: Zachary Amsden <zach@vmware.com> Signed-off-by: Andi Kleen <ak@suse.de> Signed-off-by: Andrew Morton <akpm@osdl.org>
214 lines
5.9 KiB
C
214 lines
5.9 KiB
C
#ifndef __ARCH_DESC_H
|
|
#define __ARCH_DESC_H
|
|
|
|
#include <asm/ldt.h>
|
|
#include <asm/segment.h>
|
|
|
|
#ifndef __ASSEMBLY__
|
|
|
|
#include <linux/preempt.h>
|
|
#include <linux/smp.h>
|
|
#include <linux/percpu.h>
|
|
|
|
#include <asm/mmu.h>
|
|
|
|
extern struct desc_struct cpu_gdt_table[GDT_ENTRIES];
|
|
|
|
struct Xgt_desc_struct {
|
|
unsigned short size;
|
|
unsigned long address __attribute__((packed));
|
|
unsigned short pad;
|
|
} __attribute__ ((packed));
|
|
|
|
extern struct Xgt_desc_struct idt_descr;
|
|
DECLARE_PER_CPU(struct Xgt_desc_struct, cpu_gdt_descr);
|
|
|
|
|
|
static inline struct desc_struct *get_cpu_gdt_table(unsigned int cpu)
|
|
{
|
|
return (struct desc_struct *)per_cpu(cpu_gdt_descr, cpu).address;
|
|
}
|
|
|
|
extern struct desc_struct idt_table[];
|
|
extern void set_intr_gate(unsigned int irq, void * addr);
|
|
|
|
static inline void pack_descriptor(__u32 *a, __u32 *b,
|
|
unsigned long base, unsigned long limit, unsigned char type, unsigned char flags)
|
|
{
|
|
*a = ((base & 0xffff) << 16) | (limit & 0xffff);
|
|
*b = (base & 0xff000000) | ((base & 0xff0000) >> 16) |
|
|
(limit & 0x000f0000) | ((type & 0xff) << 8) | ((flags & 0xf) << 20);
|
|
}
|
|
|
|
static inline void pack_gate(__u32 *a, __u32 *b,
|
|
unsigned long base, unsigned short seg, unsigned char type, unsigned char flags)
|
|
{
|
|
*a = (seg << 16) | (base & 0xffff);
|
|
*b = (base & 0xffff0000) | ((type & 0xff) << 8) | (flags & 0xff);
|
|
}
|
|
|
|
#define DESCTYPE_LDT 0x82 /* present, system, DPL-0, LDT */
|
|
#define DESCTYPE_TSS 0x89 /* present, system, DPL-0, 32-bit TSS */
|
|
#define DESCTYPE_TASK 0x85 /* present, system, DPL-0, task gate */
|
|
#define DESCTYPE_INT 0x8e /* present, system, DPL-0, interrupt gate */
|
|
#define DESCTYPE_TRAP 0x8f /* present, system, DPL-0, trap gate */
|
|
#define DESCTYPE_DPL3 0x60 /* DPL-3 */
|
|
#define DESCTYPE_S 0x10 /* !system */
|
|
|
|
#ifdef CONFIG_PARAVIRT
|
|
#include <asm/paravirt.h>
|
|
#else
|
|
#define load_TR_desc() __asm__ __volatile__("ltr %w0"::"q" (GDT_ENTRY_TSS*8))
|
|
|
|
#define load_gdt(dtr) __asm__ __volatile("lgdt %0"::"m" (*dtr))
|
|
#define load_idt(dtr) __asm__ __volatile("lidt %0"::"m" (*dtr))
|
|
#define load_tr(tr) __asm__ __volatile("ltr %0"::"m" (tr))
|
|
#define load_ldt(ldt) __asm__ __volatile("lldt %0"::"m" (ldt))
|
|
|
|
#define store_gdt(dtr) __asm__ ("sgdt %0":"=m" (*dtr))
|
|
#define store_idt(dtr) __asm__ ("sidt %0":"=m" (*dtr))
|
|
#define store_tr(tr) __asm__ ("str %0":"=m" (tr))
|
|
#define store_ldt(ldt) __asm__ ("sldt %0":"=m" (ldt))
|
|
|
|
#if TLS_SIZE != 24
|
|
# error update this code.
|
|
#endif
|
|
|
|
static inline void load_TLS(struct thread_struct *t, unsigned int cpu)
|
|
{
|
|
#define C(i) get_cpu_gdt_table(cpu)[GDT_ENTRY_TLS_MIN + i] = t->tls_array[i]
|
|
C(0); C(1); C(2);
|
|
#undef C
|
|
}
|
|
|
|
#define write_ldt_entry(dt, entry, a, b) write_dt_entry(dt, entry, a, b)
|
|
#define write_gdt_entry(dt, entry, a, b) write_dt_entry(dt, entry, a, b)
|
|
#define write_idt_entry(dt, entry, a, b) write_dt_entry(dt, entry, a, b)
|
|
|
|
static inline void write_dt_entry(void *dt, int entry, __u32 entry_a, __u32 entry_b)
|
|
{
|
|
__u32 *lp = (__u32 *)((char *)dt + entry*8);
|
|
*lp = entry_a;
|
|
*(lp+1) = entry_b;
|
|
}
|
|
|
|
#define set_ldt native_set_ldt
|
|
#endif /* CONFIG_PARAVIRT */
|
|
|
|
static inline fastcall void native_set_ldt(const void *addr,
|
|
unsigned int entries)
|
|
{
|
|
if (likely(entries == 0))
|
|
__asm__ __volatile__("lldt %w0"::"q" (0));
|
|
else {
|
|
unsigned cpu = smp_processor_id();
|
|
__u32 a, b;
|
|
|
|
pack_descriptor(&a, &b, (unsigned long)addr,
|
|
entries * sizeof(struct desc_struct) - 1,
|
|
DESCTYPE_LDT, 0);
|
|
write_gdt_entry(get_cpu_gdt_table(cpu), GDT_ENTRY_LDT, a, b);
|
|
__asm__ __volatile__("lldt %w0"::"q" (GDT_ENTRY_LDT*8));
|
|
}
|
|
}
|
|
|
|
static inline void _set_gate(int gate, unsigned int type, void *addr, unsigned short seg)
|
|
{
|
|
__u32 a, b;
|
|
pack_gate(&a, &b, (unsigned long)addr, seg, type, 0);
|
|
write_idt_entry(idt_table, gate, a, b);
|
|
}
|
|
|
|
static inline void __set_tss_desc(unsigned int cpu, unsigned int entry, const void *addr)
|
|
{
|
|
__u32 a, b;
|
|
pack_descriptor(&a, &b, (unsigned long)addr,
|
|
offsetof(struct tss_struct, __cacheline_filler) - 1,
|
|
DESCTYPE_TSS, 0);
|
|
write_gdt_entry(get_cpu_gdt_table(cpu), entry, a, b);
|
|
}
|
|
|
|
|
|
#define set_tss_desc(cpu,addr) __set_tss_desc(cpu, GDT_ENTRY_TSS, addr)
|
|
|
|
#define LDT_entry_a(info) \
|
|
((((info)->base_addr & 0x0000ffff) << 16) | ((info)->limit & 0x0ffff))
|
|
|
|
#define LDT_entry_b(info) \
|
|
(((info)->base_addr & 0xff000000) | \
|
|
(((info)->base_addr & 0x00ff0000) >> 16) | \
|
|
((info)->limit & 0xf0000) | \
|
|
(((info)->read_exec_only ^ 1) << 9) | \
|
|
((info)->contents << 10) | \
|
|
(((info)->seg_not_present ^ 1) << 15) | \
|
|
((info)->seg_32bit << 22) | \
|
|
((info)->limit_in_pages << 23) | \
|
|
((info)->useable << 20) | \
|
|
0x7000)
|
|
|
|
#define LDT_empty(info) (\
|
|
(info)->base_addr == 0 && \
|
|
(info)->limit == 0 && \
|
|
(info)->contents == 0 && \
|
|
(info)->read_exec_only == 1 && \
|
|
(info)->seg_32bit == 0 && \
|
|
(info)->limit_in_pages == 0 && \
|
|
(info)->seg_not_present == 1 && \
|
|
(info)->useable == 0 )
|
|
|
|
static inline void clear_LDT(void)
|
|
{
|
|
set_ldt(NULL, 0);
|
|
}
|
|
|
|
/*
|
|
* load one particular LDT into the current CPU
|
|
*/
|
|
static inline void load_LDT_nolock(mm_context_t *pc)
|
|
{
|
|
set_ldt(pc->ldt, pc->size);
|
|
}
|
|
|
|
static inline void load_LDT(mm_context_t *pc)
|
|
{
|
|
preempt_disable();
|
|
load_LDT_nolock(pc);
|
|
preempt_enable();
|
|
}
|
|
|
|
static inline unsigned long get_desc_base(unsigned long *desc)
|
|
{
|
|
unsigned long base;
|
|
base = ((desc[0] >> 16) & 0x0000ffff) |
|
|
((desc[1] << 16) & 0x00ff0000) |
|
|
(desc[1] & 0xff000000);
|
|
return base;
|
|
}
|
|
|
|
#else /* __ASSEMBLY__ */
|
|
|
|
/*
|
|
* GET_DESC_BASE reads the descriptor base of the specified segment.
|
|
*
|
|
* Args:
|
|
* idx - descriptor index
|
|
* gdt - GDT pointer
|
|
* base - 32bit register to which the base will be written
|
|
* lo_w - lo word of the "base" register
|
|
* lo_b - lo byte of the "base" register
|
|
* hi_b - hi byte of the low word of the "base" register
|
|
*
|
|
* Example:
|
|
* GET_DESC_BASE(GDT_ENTRY_ESPFIX_SS, %ebx, %eax, %ax, %al, %ah)
|
|
* Will read the base address of GDT_ENTRY_ESPFIX_SS and put it into %eax.
|
|
*/
|
|
#define GET_DESC_BASE(idx, gdt, base, lo_w, lo_b, hi_b) \
|
|
movb idx*8+4(gdt), lo_b; \
|
|
movb idx*8+7(gdt), hi_b; \
|
|
shll $16, base; \
|
|
movw idx*8+2(gdt), lo_w;
|
|
|
|
#endif /* !__ASSEMBLY__ */
|
|
|
|
#endif
|