forked from luck/tmp_suning_uos_patched
a8ab26fe5b
This will allow hotplug CPU in the future and in general cleans up a lot of crufty code. It also should plug some races that the old hackish way introduces. Remove one old race workaround in NMI watchdog setup that is not needed anymore. I removed the old total sum of bogomips reporting code. The brag value of BogoMips has been greatly devalued in the last years on the open market. Real CPU hotplug will need some more work, but the infrastructure for it is there now. One drawback: the new TSC sync algorithm is less accurate than before. The old way of zeroing TSCs is too intrusive to do later. Instead the TSC of the BP is duplicated now, which is less accurate. akpm: - sync_tsc_bp_init seems to have the sense of `init' inverted. - SPIN_LOCK_UNLOCKED is deprecated - use DEFINE_SPINLOCK. Cc: <rusty@rustcorp.com.au> Cc: <mingo@elte.hu> Signed-off-by: Andi Kleen <ak@suse.de> Signed-off-by: Andrew Morton <akpm@osdl.org> Signed-off-by: Linus Torvalds <torvalds@osdl.org>
389 lines
11 KiB
C
389 lines
11 KiB
C
#ifndef X86_64_MSR_H
|
|
#define X86_64_MSR_H 1
|
|
|
|
#ifndef __ASSEMBLY__
|
|
/*
|
|
* Access to machine-specific registers (available on 586 and better only)
|
|
* Note: the rd* operations modify the parameters directly (without using
|
|
* pointer indirection), this allows gcc to optimize better
|
|
*/
|
|
|
|
#define rdmsr(msr,val1,val2) \
|
|
__asm__ __volatile__("rdmsr" \
|
|
: "=a" (val1), "=d" (val2) \
|
|
: "c" (msr))
|
|
|
|
|
|
#define rdmsrl(msr,val) do { unsigned long a__,b__; \
|
|
__asm__ __volatile__("rdmsr" \
|
|
: "=a" (a__), "=d" (b__) \
|
|
: "c" (msr)); \
|
|
val = a__ | (b__<<32); \
|
|
} while(0);
|
|
|
|
#define wrmsr(msr,val1,val2) \
|
|
__asm__ __volatile__("wrmsr" \
|
|
: /* no outputs */ \
|
|
: "c" (msr), "a" (val1), "d" (val2))
|
|
|
|
#define wrmsrl(msr,val) wrmsr(msr,(__u32)((__u64)(val)),((__u64)(val))>>32)
|
|
|
|
/* wrmsr with exception handling */
|
|
#define wrmsr_safe(msr,a,b) ({ int ret__; \
|
|
asm volatile("2: wrmsr ; xorl %0,%0\n" \
|
|
"1:\n\t" \
|
|
".section .fixup,\"ax\"\n\t" \
|
|
"3: movl %4,%0 ; jmp 1b\n\t" \
|
|
".previous\n\t" \
|
|
".section __ex_table,\"a\"\n" \
|
|
" .align 8\n\t" \
|
|
" .quad 2b,3b\n\t" \
|
|
".previous" \
|
|
: "=a" (ret__) \
|
|
: "c" (msr), "0" (a), "d" (b), "i" (-EFAULT));\
|
|
ret__; })
|
|
|
|
#define checking_wrmsrl(msr,val) wrmsr_safe(msr,(u32)(val),(u32)((val)>>32))
|
|
|
|
#define rdtsc(low,high) \
|
|
__asm__ __volatile__("rdtsc" : "=a" (low), "=d" (high))
|
|
|
|
#define rdtscl(low) \
|
|
__asm__ __volatile__ ("rdtsc" : "=a" (low) : : "edx")
|
|
|
|
#define rdtscll(val) do { \
|
|
unsigned int __a,__d; \
|
|
asm volatile("rdtsc" : "=a" (__a), "=d" (__d)); \
|
|
(val) = ((unsigned long)__a) | (((unsigned long)__d)<<32); \
|
|
} while(0)
|
|
|
|
#define rdpmc(counter,low,high) \
|
|
__asm__ __volatile__("rdpmc" \
|
|
: "=a" (low), "=d" (high) \
|
|
: "c" (counter))
|
|
|
|
#define write_tsc(val1,val2) wrmsr(0x10, val1, val2)
|
|
|
|
#define rdpmc(counter,low,high) \
|
|
__asm__ __volatile__("rdpmc" \
|
|
: "=a" (low), "=d" (high) \
|
|
: "c" (counter))
|
|
|
|
extern inline void cpuid(int op, unsigned int *eax, unsigned int *ebx,
|
|
unsigned int *ecx, unsigned int *edx)
|
|
{
|
|
__asm__("cpuid"
|
|
: "=a" (*eax),
|
|
"=b" (*ebx),
|
|
"=c" (*ecx),
|
|
"=d" (*edx)
|
|
: "0" (op));
|
|
}
|
|
|
|
/* Some CPUID calls want 'count' to be placed in ecx */
|
|
static inline void cpuid_count(int op, int count, int *eax, int *ebx, int *ecx,
|
|
int *edx)
|
|
{
|
|
__asm__("cpuid"
|
|
: "=a" (*eax),
|
|
"=b" (*ebx),
|
|
"=c" (*ecx),
|
|
"=d" (*edx)
|
|
: "0" (op), "c" (count));
|
|
}
|
|
|
|
/*
|
|
* CPUID functions returning a single datum
|
|
*/
|
|
extern inline unsigned int cpuid_eax(unsigned int op)
|
|
{
|
|
unsigned int eax;
|
|
|
|
__asm__("cpuid"
|
|
: "=a" (eax)
|
|
: "0" (op)
|
|
: "bx", "cx", "dx");
|
|
return eax;
|
|
}
|
|
extern inline unsigned int cpuid_ebx(unsigned int op)
|
|
{
|
|
unsigned int eax, ebx;
|
|
|
|
__asm__("cpuid"
|
|
: "=a" (eax), "=b" (ebx)
|
|
: "0" (op)
|
|
: "cx", "dx" );
|
|
return ebx;
|
|
}
|
|
extern inline unsigned int cpuid_ecx(unsigned int op)
|
|
{
|
|
unsigned int eax, ecx;
|
|
|
|
__asm__("cpuid"
|
|
: "=a" (eax), "=c" (ecx)
|
|
: "0" (op)
|
|
: "bx", "dx" );
|
|
return ecx;
|
|
}
|
|
extern inline unsigned int cpuid_edx(unsigned int op)
|
|
{
|
|
unsigned int eax, edx;
|
|
|
|
__asm__("cpuid"
|
|
: "=a" (eax), "=d" (edx)
|
|
: "0" (op)
|
|
: "bx", "cx");
|
|
return edx;
|
|
}
|
|
|
|
#define MSR_IA32_UCODE_WRITE 0x79
|
|
#define MSR_IA32_UCODE_REV 0x8b
|
|
|
|
|
|
#endif
|
|
|
|
/* AMD/K8 specific MSRs */
|
|
#define MSR_EFER 0xc0000080 /* extended feature register */
|
|
#define MSR_STAR 0xc0000081 /* legacy mode SYSCALL target */
|
|
#define MSR_LSTAR 0xc0000082 /* long mode SYSCALL target */
|
|
#define MSR_CSTAR 0xc0000083 /* compatibility mode SYSCALL target */
|
|
#define MSR_SYSCALL_MASK 0xc0000084 /* EFLAGS mask for syscall */
|
|
#define MSR_FS_BASE 0xc0000100 /* 64bit GS base */
|
|
#define MSR_GS_BASE 0xc0000101 /* 64bit FS base */
|
|
#define MSR_KERNEL_GS_BASE 0xc0000102 /* SwapGS GS shadow (or USER_GS from kernel) */
|
|
/* EFER bits: */
|
|
#define _EFER_SCE 0 /* SYSCALL/SYSRET */
|
|
#define _EFER_LME 8 /* Long mode enable */
|
|
#define _EFER_LMA 10 /* Long mode active (read-only) */
|
|
#define _EFER_NX 11 /* No execute enable */
|
|
|
|
#define EFER_SCE (1<<_EFER_SCE)
|
|
#define EFER_LME (1<<_EFER_LME)
|
|
#define EFER_LMA (1<<_EFER_LMA)
|
|
#define EFER_NX (1<<_EFER_NX)
|
|
|
|
/* Intel MSRs. Some also available on other CPUs */
|
|
#define MSR_IA32_TSC 0x10
|
|
#define MSR_IA32_PLATFORM_ID 0x17
|
|
|
|
#define MSR_IA32_PERFCTR0 0xc1
|
|
#define MSR_IA32_PERFCTR1 0xc2
|
|
|
|
#define MSR_MTRRcap 0x0fe
|
|
#define MSR_IA32_BBL_CR_CTL 0x119
|
|
|
|
#define MSR_IA32_SYSENTER_CS 0x174
|
|
#define MSR_IA32_SYSENTER_ESP 0x175
|
|
#define MSR_IA32_SYSENTER_EIP 0x176
|
|
|
|
#define MSR_IA32_MCG_CAP 0x179
|
|
#define MSR_IA32_MCG_STATUS 0x17a
|
|
#define MSR_IA32_MCG_CTL 0x17b
|
|
|
|
#define MSR_IA32_EVNTSEL0 0x186
|
|
#define MSR_IA32_EVNTSEL1 0x187
|
|
|
|
#define MSR_IA32_DEBUGCTLMSR 0x1d9
|
|
#define MSR_IA32_LASTBRANCHFROMIP 0x1db
|
|
#define MSR_IA32_LASTBRANCHTOIP 0x1dc
|
|
#define MSR_IA32_LASTINTFROMIP 0x1dd
|
|
#define MSR_IA32_LASTINTTOIP 0x1de
|
|
|
|
#define MSR_MTRRfix64K_00000 0x250
|
|
#define MSR_MTRRfix16K_80000 0x258
|
|
#define MSR_MTRRfix16K_A0000 0x259
|
|
#define MSR_MTRRfix4K_C0000 0x268
|
|
#define MSR_MTRRfix4K_C8000 0x269
|
|
#define MSR_MTRRfix4K_D0000 0x26a
|
|
#define MSR_MTRRfix4K_D8000 0x26b
|
|
#define MSR_MTRRfix4K_E0000 0x26c
|
|
#define MSR_MTRRfix4K_E8000 0x26d
|
|
#define MSR_MTRRfix4K_F0000 0x26e
|
|
#define MSR_MTRRfix4K_F8000 0x26f
|
|
#define MSR_MTRRdefType 0x2ff
|
|
|
|
#define MSR_IA32_MC0_CTL 0x400
|
|
#define MSR_IA32_MC0_STATUS 0x401
|
|
#define MSR_IA32_MC0_ADDR 0x402
|
|
#define MSR_IA32_MC0_MISC 0x403
|
|
|
|
#define MSR_P6_PERFCTR0 0xc1
|
|
#define MSR_P6_PERFCTR1 0xc2
|
|
#define MSR_P6_EVNTSEL0 0x186
|
|
#define MSR_P6_EVNTSEL1 0x187
|
|
|
|
/* K7/K8 MSRs. Not complete. See the architecture manual for a more complete list. */
|
|
#define MSR_K7_EVNTSEL0 0xC0010000
|
|
#define MSR_K7_PERFCTR0 0xC0010004
|
|
#define MSR_K7_EVNTSEL1 0xC0010001
|
|
#define MSR_K7_PERFCTR1 0xC0010005
|
|
#define MSR_K7_EVNTSEL2 0xC0010002
|
|
#define MSR_K7_PERFCTR2 0xC0010006
|
|
#define MSR_K7_EVNTSEL3 0xC0010003
|
|
#define MSR_K7_PERFCTR3 0xC0010007
|
|
#define MSR_K8_TOP_MEM1 0xC001001A
|
|
#define MSR_K8_TOP_MEM2 0xC001001D
|
|
#define MSR_K8_SYSCFG 0xC0000010
|
|
|
|
/* K6 MSRs */
|
|
#define MSR_K6_EFER 0xC0000080
|
|
#define MSR_K6_STAR 0xC0000081
|
|
#define MSR_K6_WHCR 0xC0000082
|
|
#define MSR_K6_UWCCR 0xC0000085
|
|
#define MSR_K6_PSOR 0xC0000087
|
|
#define MSR_K6_PFIR 0xC0000088
|
|
|
|
/* Centaur-Hauls/IDT defined MSRs. */
|
|
#define MSR_IDT_FCR1 0x107
|
|
#define MSR_IDT_FCR2 0x108
|
|
#define MSR_IDT_FCR3 0x109
|
|
#define MSR_IDT_FCR4 0x10a
|
|
|
|
#define MSR_IDT_MCR0 0x110
|
|
#define MSR_IDT_MCR1 0x111
|
|
#define MSR_IDT_MCR2 0x112
|
|
#define MSR_IDT_MCR3 0x113
|
|
#define MSR_IDT_MCR4 0x114
|
|
#define MSR_IDT_MCR5 0x115
|
|
#define MSR_IDT_MCR6 0x116
|
|
#define MSR_IDT_MCR7 0x117
|
|
#define MSR_IDT_MCR_CTRL 0x120
|
|
|
|
/* VIA Cyrix defined MSRs*/
|
|
#define MSR_VIA_FCR 0x1107
|
|
#define MSR_VIA_LONGHAUL 0x110a
|
|
#define MSR_VIA_RNG 0x110b
|
|
#define MSR_VIA_BCR2 0x1147
|
|
|
|
/* Intel defined MSRs. */
|
|
#define MSR_IA32_P5_MC_ADDR 0
|
|
#define MSR_IA32_P5_MC_TYPE 1
|
|
#define MSR_IA32_PLATFORM_ID 0x17
|
|
#define MSR_IA32_EBL_CR_POWERON 0x2a
|
|
|
|
#define MSR_IA32_APICBASE 0x1b
|
|
#define MSR_IA32_APICBASE_BSP (1<<8)
|
|
#define MSR_IA32_APICBASE_ENABLE (1<<11)
|
|
#define MSR_IA32_APICBASE_BASE (0xfffff<<12)
|
|
|
|
/* P4/Xeon+ specific */
|
|
#define MSR_IA32_MCG_EAX 0x180
|
|
#define MSR_IA32_MCG_EBX 0x181
|
|
#define MSR_IA32_MCG_ECX 0x182
|
|
#define MSR_IA32_MCG_EDX 0x183
|
|
#define MSR_IA32_MCG_ESI 0x184
|
|
#define MSR_IA32_MCG_EDI 0x185
|
|
#define MSR_IA32_MCG_EBP 0x186
|
|
#define MSR_IA32_MCG_ESP 0x187
|
|
#define MSR_IA32_MCG_EFLAGS 0x188
|
|
#define MSR_IA32_MCG_EIP 0x189
|
|
#define MSR_IA32_MCG_RESERVED 0x18A
|
|
|
|
#define MSR_P6_EVNTSEL0 0x186
|
|
#define MSR_P6_EVNTSEL1 0x187
|
|
|
|
#define MSR_IA32_PERF_STATUS 0x198
|
|
#define MSR_IA32_PERF_CTL 0x199
|
|
|
|
#define MSR_IA32_THERM_CONTROL 0x19a
|
|
#define MSR_IA32_THERM_INTERRUPT 0x19b
|
|
#define MSR_IA32_THERM_STATUS 0x19c
|
|
#define MSR_IA32_MISC_ENABLE 0x1a0
|
|
|
|
#define MSR_IA32_DEBUGCTLMSR 0x1d9
|
|
#define MSR_IA32_LASTBRANCHFROMIP 0x1db
|
|
#define MSR_IA32_LASTBRANCHTOIP 0x1dc
|
|
#define MSR_IA32_LASTINTFROMIP 0x1dd
|
|
#define MSR_IA32_LASTINTTOIP 0x1de
|
|
|
|
#define MSR_IA32_MC0_CTL 0x400
|
|
#define MSR_IA32_MC0_STATUS 0x401
|
|
#define MSR_IA32_MC0_ADDR 0x402
|
|
#define MSR_IA32_MC0_MISC 0x403
|
|
|
|
/* Pentium IV performance counter MSRs */
|
|
#define MSR_P4_BPU_PERFCTR0 0x300
|
|
#define MSR_P4_BPU_PERFCTR1 0x301
|
|
#define MSR_P4_BPU_PERFCTR2 0x302
|
|
#define MSR_P4_BPU_PERFCTR3 0x303
|
|
#define MSR_P4_MS_PERFCTR0 0x304
|
|
#define MSR_P4_MS_PERFCTR1 0x305
|
|
#define MSR_P4_MS_PERFCTR2 0x306
|
|
#define MSR_P4_MS_PERFCTR3 0x307
|
|
#define MSR_P4_FLAME_PERFCTR0 0x308
|
|
#define MSR_P4_FLAME_PERFCTR1 0x309
|
|
#define MSR_P4_FLAME_PERFCTR2 0x30a
|
|
#define MSR_P4_FLAME_PERFCTR3 0x30b
|
|
#define MSR_P4_IQ_PERFCTR0 0x30c
|
|
#define MSR_P4_IQ_PERFCTR1 0x30d
|
|
#define MSR_P4_IQ_PERFCTR2 0x30e
|
|
#define MSR_P4_IQ_PERFCTR3 0x30f
|
|
#define MSR_P4_IQ_PERFCTR4 0x310
|
|
#define MSR_P4_IQ_PERFCTR5 0x311
|
|
#define MSR_P4_BPU_CCCR0 0x360
|
|
#define MSR_P4_BPU_CCCR1 0x361
|
|
#define MSR_P4_BPU_CCCR2 0x362
|
|
#define MSR_P4_BPU_CCCR3 0x363
|
|
#define MSR_P4_MS_CCCR0 0x364
|
|
#define MSR_P4_MS_CCCR1 0x365
|
|
#define MSR_P4_MS_CCCR2 0x366
|
|
#define MSR_P4_MS_CCCR3 0x367
|
|
#define MSR_P4_FLAME_CCCR0 0x368
|
|
#define MSR_P4_FLAME_CCCR1 0x369
|
|
#define MSR_P4_FLAME_CCCR2 0x36a
|
|
#define MSR_P4_FLAME_CCCR3 0x36b
|
|
#define MSR_P4_IQ_CCCR0 0x36c
|
|
#define MSR_P4_IQ_CCCR1 0x36d
|
|
#define MSR_P4_IQ_CCCR2 0x36e
|
|
#define MSR_P4_IQ_CCCR3 0x36f
|
|
#define MSR_P4_IQ_CCCR4 0x370
|
|
#define MSR_P4_IQ_CCCR5 0x371
|
|
#define MSR_P4_ALF_ESCR0 0x3ca
|
|
#define MSR_P4_ALF_ESCR1 0x3cb
|
|
#define MSR_P4_BPU_ESCR0 0x3b2
|
|
#define MSR_P4_BPU_ESCR1 0x3b3
|
|
#define MSR_P4_BSU_ESCR0 0x3a0
|
|
#define MSR_P4_BSU_ESCR1 0x3a1
|
|
#define MSR_P4_CRU_ESCR0 0x3b8
|
|
#define MSR_P4_CRU_ESCR1 0x3b9
|
|
#define MSR_P4_CRU_ESCR2 0x3cc
|
|
#define MSR_P4_CRU_ESCR3 0x3cd
|
|
#define MSR_P4_CRU_ESCR4 0x3e0
|
|
#define MSR_P4_CRU_ESCR5 0x3e1
|
|
#define MSR_P4_DAC_ESCR0 0x3a8
|
|
#define MSR_P4_DAC_ESCR1 0x3a9
|
|
#define MSR_P4_FIRM_ESCR0 0x3a4
|
|
#define MSR_P4_FIRM_ESCR1 0x3a5
|
|
#define MSR_P4_FLAME_ESCR0 0x3a6
|
|
#define MSR_P4_FLAME_ESCR1 0x3a7
|
|
#define MSR_P4_FSB_ESCR0 0x3a2
|
|
#define MSR_P4_FSB_ESCR1 0x3a3
|
|
#define MSR_P4_IQ_ESCR0 0x3ba
|
|
#define MSR_P4_IQ_ESCR1 0x3bb
|
|
#define MSR_P4_IS_ESCR0 0x3b4
|
|
#define MSR_P4_IS_ESCR1 0x3b5
|
|
#define MSR_P4_ITLB_ESCR0 0x3b6
|
|
#define MSR_P4_ITLB_ESCR1 0x3b7
|
|
#define MSR_P4_IX_ESCR0 0x3c8
|
|
#define MSR_P4_IX_ESCR1 0x3c9
|
|
#define MSR_P4_MOB_ESCR0 0x3aa
|
|
#define MSR_P4_MOB_ESCR1 0x3ab
|
|
#define MSR_P4_MS_ESCR0 0x3c0
|
|
#define MSR_P4_MS_ESCR1 0x3c1
|
|
#define MSR_P4_PMH_ESCR0 0x3ac
|
|
#define MSR_P4_PMH_ESCR1 0x3ad
|
|
#define MSR_P4_RAT_ESCR0 0x3bc
|
|
#define MSR_P4_RAT_ESCR1 0x3bd
|
|
#define MSR_P4_SAAT_ESCR0 0x3ae
|
|
#define MSR_P4_SAAT_ESCR1 0x3af
|
|
#define MSR_P4_SSU_ESCR0 0x3be
|
|
#define MSR_P4_SSU_ESCR1 0x3bf /* guess: not defined in manual */
|
|
#define MSR_P4_TBPU_ESCR0 0x3c2
|
|
#define MSR_P4_TBPU_ESCR1 0x3c3
|
|
#define MSR_P4_TC_ESCR0 0x3c4
|
|
#define MSR_P4_TC_ESCR1 0x3c5
|
|
#define MSR_P4_U2L_ESCR0 0x3b0
|
|
#define MSR_P4_U2L_ESCR1 0x3b1
|
|
|
|
#endif
|