Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux
fork

Configure Feed

Select the types of activity you want to include in your feed.

at v4.17 1043 lines 34 kB view raw
1/* 2 * tools/testing/selftests/kvm/include/x86.h 3 * 4 * Copyright (C) 2018, Google LLC. 5 * 6 * This work is licensed under the terms of the GNU GPL, version 2. 7 * 8 */ 9 10#ifndef SELFTEST_KVM_X86_H 11#define SELFTEST_KVM_X86_H 12 13#include <assert.h> 14#include <stdint.h> 15 16#define X86_EFLAGS_FIXED (1u << 1) 17 18#define X86_CR4_VME (1ul << 0) 19#define X86_CR4_PVI (1ul << 1) 20#define X86_CR4_TSD (1ul << 2) 21#define X86_CR4_DE (1ul << 3) 22#define X86_CR4_PSE (1ul << 4) 23#define X86_CR4_PAE (1ul << 5) 24#define X86_CR4_MCE (1ul << 6) 25#define X86_CR4_PGE (1ul << 7) 26#define X86_CR4_PCE (1ul << 8) 27#define X86_CR4_OSFXSR (1ul << 9) 28#define X86_CR4_OSXMMEXCPT (1ul << 10) 29#define X86_CR4_UMIP (1ul << 11) 30#define X86_CR4_VMXE (1ul << 13) 31#define X86_CR4_SMXE (1ul << 14) 32#define X86_CR4_FSGSBASE (1ul << 16) 33#define X86_CR4_PCIDE (1ul << 17) 34#define X86_CR4_OSXSAVE (1ul << 18) 35#define X86_CR4_SMEP (1ul << 20) 36#define X86_CR4_SMAP (1ul << 21) 37#define X86_CR4_PKE (1ul << 22) 38 39/* The enum values match the intruction encoding of each register */ 40enum x86_register { 41 RAX = 0, 42 RCX, 43 RDX, 44 RBX, 45 RSP, 46 RBP, 47 RSI, 48 RDI, 49 R8, 50 R9, 51 R10, 52 R11, 53 R12, 54 R13, 55 R14, 56 R15, 57}; 58 59struct desc64 { 60 uint16_t limit0; 61 uint16_t base0; 62 unsigned base1:8, type:5, dpl:2, p:1; 63 unsigned limit1:4, zero0:3, g:1, base2:8; 64 uint32_t base3; 65 uint32_t zero1; 66} __attribute__((packed)); 67 68struct desc_ptr { 69 uint16_t size; 70 uint64_t address; 71} __attribute__((packed)); 72 73static inline uint64_t get_desc64_base(const struct desc64 *desc) 74{ 75 return ((uint64_t)desc->base3 << 32) | 76 (desc->base0 | ((desc->base1) << 16) | ((desc->base2) << 24)); 77} 78 79static inline uint64_t rdtsc(void) 80{ 81 uint32_t eax, edx; 82 83 /* 84 * The lfence is to wait (on Intel CPUs) until all previous 85 * instructions have been executed. 86 */ 87 __asm__ __volatile__("lfence; rdtsc" : "=a"(eax), "=d"(edx)); 88 return ((uint64_t)edx) << 32 | eax; 89} 90 91static inline uint64_t rdtscp(uint32_t *aux) 92{ 93 uint32_t eax, edx; 94 95 __asm__ __volatile__("rdtscp" : "=a"(eax), "=d"(edx), "=c"(*aux)); 96 return ((uint64_t)edx) << 32 | eax; 97} 98 99static inline uint64_t rdmsr(uint32_t msr) 100{ 101 uint32_t a, d; 102 103 __asm__ __volatile__("rdmsr" : "=a"(a), "=d"(d) : "c"(msr) : "memory"); 104 105 return a | ((uint64_t) d << 32); 106} 107 108static inline void wrmsr(uint32_t msr, uint64_t value) 109{ 110 uint32_t a = value; 111 uint32_t d = value >> 32; 112 113 __asm__ __volatile__("wrmsr" :: "a"(a), "d"(d), "c"(msr) : "memory"); 114} 115 116 117static inline uint16_t inw(uint16_t port) 118{ 119 uint16_t tmp; 120 121 __asm__ __volatile__("in %%dx, %%ax" 122 : /* output */ "=a" (tmp) 123 : /* input */ "d" (port)); 124 125 return tmp; 126} 127 128static inline uint16_t get_es(void) 129{ 130 uint16_t es; 131 132 __asm__ __volatile__("mov %%es, %[es]" 133 : /* output */ [es]"=rm"(es)); 134 return es; 135} 136 137static inline uint16_t get_cs(void) 138{ 139 uint16_t cs; 140 141 __asm__ __volatile__("mov %%cs, %[cs]" 142 : /* output */ [cs]"=rm"(cs)); 143 return cs; 144} 145 146static inline uint16_t get_ss(void) 147{ 148 uint16_t ss; 149 150 __asm__ __volatile__("mov %%ss, %[ss]" 151 : /* output */ [ss]"=rm"(ss)); 152 return ss; 153} 154 155static inline uint16_t get_ds(void) 156{ 157 uint16_t ds; 158 159 __asm__ __volatile__("mov %%ds, %[ds]" 160 : /* output */ [ds]"=rm"(ds)); 161 return ds; 162} 163 164static inline uint16_t get_fs(void) 165{ 166 uint16_t fs; 167 168 __asm__ __volatile__("mov %%fs, %[fs]" 169 : /* output */ [fs]"=rm"(fs)); 170 return fs; 171} 172 173static inline uint16_t get_gs(void) 174{ 175 uint16_t gs; 176 177 __asm__ __volatile__("mov %%gs, %[gs]" 178 : /* output */ [gs]"=rm"(gs)); 179 return gs; 180} 181 182static inline uint16_t get_tr(void) 183{ 184 uint16_t tr; 185 186 __asm__ __volatile__("str %[tr]" 187 : /* output */ [tr]"=rm"(tr)); 188 return tr; 189} 190 191static inline uint64_t get_cr0(void) 192{ 193 uint64_t cr0; 194 195 __asm__ __volatile__("mov %%cr0, %[cr0]" 196 : /* output */ [cr0]"=r"(cr0)); 197 return cr0; 198} 199 200static inline uint64_t get_cr3(void) 201{ 202 uint64_t cr3; 203 204 __asm__ __volatile__("mov %%cr3, %[cr3]" 205 : /* output */ [cr3]"=r"(cr3)); 206 return cr3; 207} 208 209static inline uint64_t get_cr4(void) 210{ 211 uint64_t cr4; 212 213 __asm__ __volatile__("mov %%cr4, %[cr4]" 214 : /* output */ [cr4]"=r"(cr4)); 215 return cr4; 216} 217 218static inline void set_cr4(uint64_t val) 219{ 220 __asm__ __volatile__("mov %0, %%cr4" : : "r" (val) : "memory"); 221} 222 223static inline uint64_t get_gdt_base(void) 224{ 225 struct desc_ptr gdt; 226 __asm__ __volatile__("sgdt %[gdt]" 227 : /* output */ [gdt]"=m"(gdt)); 228 return gdt.address; 229} 230 231static inline uint64_t get_idt_base(void) 232{ 233 struct desc_ptr idt; 234 __asm__ __volatile__("sidt %[idt]" 235 : /* output */ [idt]"=m"(idt)); 236 return idt.address; 237} 238 239#define SET_XMM(__var, __xmm) \ 240 asm volatile("movq %0, %%"#__xmm : : "r"(__var) : #__xmm) 241 242static inline void set_xmm(int n, unsigned long val) 243{ 244 switch (n) { 245 case 0: 246 SET_XMM(val, xmm0); 247 break; 248 case 1: 249 SET_XMM(val, xmm1); 250 break; 251 case 2: 252 SET_XMM(val, xmm2); 253 break; 254 case 3: 255 SET_XMM(val, xmm3); 256 break; 257 case 4: 258 SET_XMM(val, xmm4); 259 break; 260 case 5: 261 SET_XMM(val, xmm5); 262 break; 263 case 6: 264 SET_XMM(val, xmm6); 265 break; 266 case 7: 267 SET_XMM(val, xmm7); 268 break; 269 } 270} 271 272typedef unsigned long v1di __attribute__ ((vector_size (8))); 273static inline unsigned long get_xmm(int n) 274{ 275 assert(n >= 0 && n <= 7); 276 277 register v1di xmm0 __asm__("%xmm0"); 278 register v1di xmm1 __asm__("%xmm1"); 279 register v1di xmm2 __asm__("%xmm2"); 280 register v1di xmm3 __asm__("%xmm3"); 281 register v1di xmm4 __asm__("%xmm4"); 282 register v1di xmm5 __asm__("%xmm5"); 283 register v1di xmm6 __asm__("%xmm6"); 284 register v1di xmm7 __asm__("%xmm7"); 285 switch (n) { 286 case 0: 287 return (unsigned long)xmm0; 288 case 1: 289 return (unsigned long)xmm1; 290 case 2: 291 return (unsigned long)xmm2; 292 case 3: 293 return (unsigned long)xmm3; 294 case 4: 295 return (unsigned long)xmm4; 296 case 5: 297 return (unsigned long)xmm5; 298 case 6: 299 return (unsigned long)xmm6; 300 case 7: 301 return (unsigned long)xmm7; 302 } 303 return 0; 304} 305 306/* 307 * Basic CPU control in CR0 308 */ 309#define X86_CR0_PE (1UL<<0) /* Protection Enable */ 310#define X86_CR0_MP (1UL<<1) /* Monitor Coprocessor */ 311#define X86_CR0_EM (1UL<<2) /* Emulation */ 312#define X86_CR0_TS (1UL<<3) /* Task Switched */ 313#define X86_CR0_ET (1UL<<4) /* Extension Type */ 314#define X86_CR0_NE (1UL<<5) /* Numeric Error */ 315#define X86_CR0_WP (1UL<<16) /* Write Protect */ 316#define X86_CR0_AM (1UL<<18) /* Alignment Mask */ 317#define X86_CR0_NW (1UL<<29) /* Not Write-through */ 318#define X86_CR0_CD (1UL<<30) /* Cache Disable */ 319#define X86_CR0_PG (1UL<<31) /* Paging */ 320 321/* 322 * CPU model specific register (MSR) numbers. 323 */ 324 325/* x86-64 specific MSRs */ 326#define MSR_EFER 0xc0000080 /* extended feature register */ 327#define MSR_STAR 0xc0000081 /* legacy mode SYSCALL target */ 328#define MSR_LSTAR 0xc0000082 /* long mode SYSCALL target */ 329#define MSR_CSTAR 0xc0000083 /* compat mode SYSCALL target */ 330#define MSR_SYSCALL_MASK 0xc0000084 /* EFLAGS mask for syscall */ 331#define MSR_FS_BASE 0xc0000100 /* 64bit FS base */ 332#define MSR_GS_BASE 0xc0000101 /* 64bit GS base */ 333#define MSR_KERNEL_GS_BASE 0xc0000102 /* SwapGS GS shadow */ 334#define MSR_TSC_AUX 0xc0000103 /* Auxiliary TSC */ 335 336/* EFER bits: */ 337#define EFER_SCE (1<<0) /* SYSCALL/SYSRET */ 338#define EFER_LME (1<<8) /* Long mode enable */ 339#define EFER_LMA (1<<10) /* Long mode active (read-only) */ 340#define EFER_NX (1<<11) /* No execute enable */ 341#define EFER_SVME (1<<12) /* Enable virtualization */ 342#define EFER_LMSLE (1<<13) /* Long Mode Segment Limit Enable */ 343#define EFER_FFXSR (1<<14) /* Enable Fast FXSAVE/FXRSTOR */ 344 345/* Intel MSRs. Some also available on other CPUs */ 346 347#define MSR_PPIN_CTL 0x0000004e 348#define MSR_PPIN 0x0000004f 349 350#define MSR_IA32_PERFCTR0 0x000000c1 351#define MSR_IA32_PERFCTR1 0x000000c2 352#define MSR_FSB_FREQ 0x000000cd 353#define MSR_PLATFORM_INFO 0x000000ce 354#define MSR_PLATFORM_INFO_CPUID_FAULT_BIT 31 355#define MSR_PLATFORM_INFO_CPUID_FAULT BIT_ULL(MSR_PLATFORM_INFO_CPUID_FAULT_BIT) 356 357#define MSR_PKG_CST_CONFIG_CONTROL 0x000000e2 358#define NHM_C3_AUTO_DEMOTE (1UL << 25) 359#define NHM_C1_AUTO_DEMOTE (1UL << 26) 360#define ATM_LNC_C6_AUTO_DEMOTE (1UL << 25) 361#define SNB_C1_AUTO_UNDEMOTE (1UL << 27) 362#define SNB_C3_AUTO_UNDEMOTE (1UL << 28) 363 364#define MSR_MTRRcap 0x000000fe 365#define MSR_IA32_BBL_CR_CTL 0x00000119 366#define MSR_IA32_BBL_CR_CTL3 0x0000011e 367 368#define MSR_IA32_SYSENTER_CS 0x00000174 369#define MSR_IA32_SYSENTER_ESP 0x00000175 370#define MSR_IA32_SYSENTER_EIP 0x00000176 371 372#define MSR_IA32_MCG_CAP 0x00000179 373#define MSR_IA32_MCG_STATUS 0x0000017a 374#define MSR_IA32_MCG_CTL 0x0000017b 375#define MSR_IA32_MCG_EXT_CTL 0x000004d0 376 377#define MSR_OFFCORE_RSP_0 0x000001a6 378#define MSR_OFFCORE_RSP_1 0x000001a7 379#define MSR_TURBO_RATIO_LIMIT 0x000001ad 380#define MSR_TURBO_RATIO_LIMIT1 0x000001ae 381#define MSR_TURBO_RATIO_LIMIT2 0x000001af 382 383#define MSR_LBR_SELECT 0x000001c8 384#define MSR_LBR_TOS 0x000001c9 385#define MSR_LBR_NHM_FROM 0x00000680 386#define MSR_LBR_NHM_TO 0x000006c0 387#define MSR_LBR_CORE_FROM 0x00000040 388#define MSR_LBR_CORE_TO 0x00000060 389 390#define MSR_LBR_INFO_0 0x00000dc0 /* ... 0xddf for _31 */ 391#define LBR_INFO_MISPRED BIT_ULL(63) 392#define LBR_INFO_IN_TX BIT_ULL(62) 393#define LBR_INFO_ABORT BIT_ULL(61) 394#define LBR_INFO_CYCLES 0xffff 395 396#define MSR_IA32_PEBS_ENABLE 0x000003f1 397#define MSR_IA32_DS_AREA 0x00000600 398#define MSR_IA32_PERF_CAPABILITIES 0x00000345 399#define MSR_PEBS_LD_LAT_THRESHOLD 0x000003f6 400 401#define MSR_IA32_RTIT_CTL 0x00000570 402#define MSR_IA32_RTIT_STATUS 0x00000571 403#define MSR_IA32_RTIT_ADDR0_A 0x00000580 404#define MSR_IA32_RTIT_ADDR0_B 0x00000581 405#define MSR_IA32_RTIT_ADDR1_A 0x00000582 406#define MSR_IA32_RTIT_ADDR1_B 0x00000583 407#define MSR_IA32_RTIT_ADDR2_A 0x00000584 408#define MSR_IA32_RTIT_ADDR2_B 0x00000585 409#define MSR_IA32_RTIT_ADDR3_A 0x00000586 410#define MSR_IA32_RTIT_ADDR3_B 0x00000587 411#define MSR_IA32_RTIT_CR3_MATCH 0x00000572 412#define MSR_IA32_RTIT_OUTPUT_BASE 0x00000560 413#define MSR_IA32_RTIT_OUTPUT_MASK 0x00000561 414 415#define MSR_MTRRfix64K_00000 0x00000250 416#define MSR_MTRRfix16K_80000 0x00000258 417#define MSR_MTRRfix16K_A0000 0x00000259 418#define MSR_MTRRfix4K_C0000 0x00000268 419#define MSR_MTRRfix4K_C8000 0x00000269 420#define MSR_MTRRfix4K_D0000 0x0000026a 421#define MSR_MTRRfix4K_D8000 0x0000026b 422#define MSR_MTRRfix4K_E0000 0x0000026c 423#define MSR_MTRRfix4K_E8000 0x0000026d 424#define MSR_MTRRfix4K_F0000 0x0000026e 425#define MSR_MTRRfix4K_F8000 0x0000026f 426#define MSR_MTRRdefType 0x000002ff 427 428#define MSR_IA32_CR_PAT 0x00000277 429 430#define MSR_IA32_DEBUGCTLMSR 0x000001d9 431#define MSR_IA32_LASTBRANCHFROMIP 0x000001db 432#define MSR_IA32_LASTBRANCHTOIP 0x000001dc 433#define MSR_IA32_LASTINTFROMIP 0x000001dd 434#define MSR_IA32_LASTINTTOIP 0x000001de 435 436/* DEBUGCTLMSR bits (others vary by model): */ 437#define DEBUGCTLMSR_LBR (1UL << 0) /* last branch recording */ 438#define DEBUGCTLMSR_BTF_SHIFT 1 439#define DEBUGCTLMSR_BTF (1UL << 1) /* single-step on branches */ 440#define DEBUGCTLMSR_TR (1UL << 6) 441#define DEBUGCTLMSR_BTS (1UL << 7) 442#define DEBUGCTLMSR_BTINT (1UL << 8) 443#define DEBUGCTLMSR_BTS_OFF_OS (1UL << 9) 444#define DEBUGCTLMSR_BTS_OFF_USR (1UL << 10) 445#define DEBUGCTLMSR_FREEZE_LBRS_ON_PMI (1UL << 11) 446#define DEBUGCTLMSR_FREEZE_IN_SMM_BIT 14 447#define DEBUGCTLMSR_FREEZE_IN_SMM (1UL << DEBUGCTLMSR_FREEZE_IN_SMM_BIT) 448 449#define MSR_PEBS_FRONTEND 0x000003f7 450 451#define MSR_IA32_POWER_CTL 0x000001fc 452 453#define MSR_IA32_MC0_CTL 0x00000400 454#define MSR_IA32_MC0_STATUS 0x00000401 455#define MSR_IA32_MC0_ADDR 0x00000402 456#define MSR_IA32_MC0_MISC 0x00000403 457 458/* C-state Residency Counters */ 459#define MSR_PKG_C3_RESIDENCY 0x000003f8 460#define MSR_PKG_C6_RESIDENCY 0x000003f9 461#define MSR_ATOM_PKG_C6_RESIDENCY 0x000003fa 462#define MSR_PKG_C7_RESIDENCY 0x000003fa 463#define MSR_CORE_C3_RESIDENCY 0x000003fc 464#define MSR_CORE_C6_RESIDENCY 0x000003fd 465#define MSR_CORE_C7_RESIDENCY 0x000003fe 466#define MSR_KNL_CORE_C6_RESIDENCY 0x000003ff 467#define MSR_PKG_C2_RESIDENCY 0x0000060d 468#define MSR_PKG_C8_RESIDENCY 0x00000630 469#define MSR_PKG_C9_RESIDENCY 0x00000631 470#define MSR_PKG_C10_RESIDENCY 0x00000632 471 472/* Interrupt Response Limit */ 473#define MSR_PKGC3_IRTL 0x0000060a 474#define MSR_PKGC6_IRTL 0x0000060b 475#define MSR_PKGC7_IRTL 0x0000060c 476#define MSR_PKGC8_IRTL 0x00000633 477#define MSR_PKGC9_IRTL 0x00000634 478#define MSR_PKGC10_IRTL 0x00000635 479 480/* Run Time Average Power Limiting (RAPL) Interface */ 481 482#define MSR_RAPL_POWER_UNIT 0x00000606 483 484#define MSR_PKG_POWER_LIMIT 0x00000610 485#define MSR_PKG_ENERGY_STATUS 0x00000611 486#define MSR_PKG_PERF_STATUS 0x00000613 487#define MSR_PKG_POWER_INFO 0x00000614 488 489#define MSR_DRAM_POWER_LIMIT 0x00000618 490#define MSR_DRAM_ENERGY_STATUS 0x00000619 491#define MSR_DRAM_PERF_STATUS 0x0000061b 492#define MSR_DRAM_POWER_INFO 0x0000061c 493 494#define MSR_PP0_POWER_LIMIT 0x00000638 495#define MSR_PP0_ENERGY_STATUS 0x00000639 496#define MSR_PP0_POLICY 0x0000063a 497#define MSR_PP0_PERF_STATUS 0x0000063b 498 499#define MSR_PP1_POWER_LIMIT 0x00000640 500#define MSR_PP1_ENERGY_STATUS 0x00000641 501#define MSR_PP1_POLICY 0x00000642 502 503/* Config TDP MSRs */ 504#define MSR_CONFIG_TDP_NOMINAL 0x00000648 505#define MSR_CONFIG_TDP_LEVEL_1 0x00000649 506#define MSR_CONFIG_TDP_LEVEL_2 0x0000064A 507#define MSR_CONFIG_TDP_CONTROL 0x0000064B 508#define MSR_TURBO_ACTIVATION_RATIO 0x0000064C 509 510#define MSR_PLATFORM_ENERGY_STATUS 0x0000064D 511 512#define MSR_PKG_WEIGHTED_CORE_C0_RES 0x00000658 513#define MSR_PKG_ANY_CORE_C0_RES 0x00000659 514#define MSR_PKG_ANY_GFXE_C0_RES 0x0000065A 515#define MSR_PKG_BOTH_CORE_GFXE_C0_RES 0x0000065B 516 517#define MSR_CORE_C1_RES 0x00000660 518#define MSR_MODULE_C6_RES_MS 0x00000664 519 520#define MSR_CC6_DEMOTION_POLICY_CONFIG 0x00000668 521#define MSR_MC6_DEMOTION_POLICY_CONFIG 0x00000669 522 523#define MSR_ATOM_CORE_RATIOS 0x0000066a 524#define MSR_ATOM_CORE_VIDS 0x0000066b 525#define MSR_ATOM_CORE_TURBO_RATIOS 0x0000066c 526#define MSR_ATOM_CORE_TURBO_VIDS 0x0000066d 527 528 529#define MSR_CORE_PERF_LIMIT_REASONS 0x00000690 530#define MSR_GFX_PERF_LIMIT_REASONS 0x000006B0 531#define MSR_RING_PERF_LIMIT_REASONS 0x000006B1 532 533/* Hardware P state interface */ 534#define MSR_PPERF 0x0000064e 535#define MSR_PERF_LIMIT_REASONS 0x0000064f 536#define MSR_PM_ENABLE 0x00000770 537#define MSR_HWP_CAPABILITIES 0x00000771 538#define MSR_HWP_REQUEST_PKG 0x00000772 539#define MSR_HWP_INTERRUPT 0x00000773 540#define MSR_HWP_REQUEST 0x00000774 541#define MSR_HWP_STATUS 0x00000777 542 543/* CPUID.6.EAX */ 544#define HWP_BASE_BIT (1<<7) 545#define HWP_NOTIFICATIONS_BIT (1<<8) 546#define HWP_ACTIVITY_WINDOW_BIT (1<<9) 547#define HWP_ENERGY_PERF_PREFERENCE_BIT (1<<10) 548#define HWP_PACKAGE_LEVEL_REQUEST_BIT (1<<11) 549 550/* IA32_HWP_CAPABILITIES */ 551#define HWP_HIGHEST_PERF(x) (((x) >> 0) & 0xff) 552#define HWP_GUARANTEED_PERF(x) (((x) >> 8) & 0xff) 553#define HWP_MOSTEFFICIENT_PERF(x) (((x) >> 16) & 0xff) 554#define HWP_LOWEST_PERF(x) (((x) >> 24) & 0xff) 555 556/* IA32_HWP_REQUEST */ 557#define HWP_MIN_PERF(x) (x & 0xff) 558#define HWP_MAX_PERF(x) ((x & 0xff) << 8) 559#define HWP_DESIRED_PERF(x) ((x & 0xff) << 16) 560#define HWP_ENERGY_PERF_PREFERENCE(x) (((unsigned long long) x & 0xff) << 24) 561#define HWP_EPP_PERFORMANCE 0x00 562#define HWP_EPP_BALANCE_PERFORMANCE 0x80 563#define HWP_EPP_BALANCE_POWERSAVE 0xC0 564#define HWP_EPP_POWERSAVE 0xFF 565#define HWP_ACTIVITY_WINDOW(x) ((unsigned long long)(x & 0xff3) << 32) 566#define HWP_PACKAGE_CONTROL(x) ((unsigned long long)(x & 0x1) << 42) 567 568/* IA32_HWP_STATUS */ 569#define HWP_GUARANTEED_CHANGE(x) (x & 0x1) 570#define HWP_EXCURSION_TO_MINIMUM(x) (x & 0x4) 571 572/* IA32_HWP_INTERRUPT */ 573#define HWP_CHANGE_TO_GUARANTEED_INT(x) (x & 0x1) 574#define HWP_EXCURSION_TO_MINIMUM_INT(x) (x & 0x2) 575 576#define MSR_AMD64_MC0_MASK 0xc0010044 577 578#define MSR_IA32_MCx_CTL(x) (MSR_IA32_MC0_CTL + 4*(x)) 579#define MSR_IA32_MCx_STATUS(x) (MSR_IA32_MC0_STATUS + 4*(x)) 580#define MSR_IA32_MCx_ADDR(x) (MSR_IA32_MC0_ADDR + 4*(x)) 581#define MSR_IA32_MCx_MISC(x) (MSR_IA32_MC0_MISC + 4*(x)) 582 583#define MSR_AMD64_MCx_MASK(x) (MSR_AMD64_MC0_MASK + (x)) 584 585/* These are consecutive and not in the normal 4er MCE bank block */ 586#define MSR_IA32_MC0_CTL2 0x00000280 587#define MSR_IA32_MCx_CTL2(x) (MSR_IA32_MC0_CTL2 + (x)) 588 589#define MSR_P6_PERFCTR0 0x000000c1 590#define MSR_P6_PERFCTR1 0x000000c2 591#define MSR_P6_EVNTSEL0 0x00000186 592#define MSR_P6_EVNTSEL1 0x00000187 593 594#define MSR_KNC_PERFCTR0 0x00000020 595#define MSR_KNC_PERFCTR1 0x00000021 596#define MSR_KNC_EVNTSEL0 0x00000028 597#define MSR_KNC_EVNTSEL1 0x00000029 598 599/* Alternative perfctr range with full access. */ 600#define MSR_IA32_PMC0 0x000004c1 601 602/* AMD64 MSRs. Not complete. See the architecture manual for a more 603 complete list. */ 604 605#define MSR_AMD64_PATCH_LEVEL 0x0000008b 606#define MSR_AMD64_TSC_RATIO 0xc0000104 607#define MSR_AMD64_NB_CFG 0xc001001f 608#define MSR_AMD64_PATCH_LOADER 0xc0010020 609#define MSR_AMD64_OSVW_ID_LENGTH 0xc0010140 610#define MSR_AMD64_OSVW_STATUS 0xc0010141 611#define MSR_AMD64_LS_CFG 0xc0011020 612#define MSR_AMD64_DC_CFG 0xc0011022 613#define MSR_AMD64_BU_CFG2 0xc001102a 614#define MSR_AMD64_IBSFETCHCTL 0xc0011030 615#define MSR_AMD64_IBSFETCHLINAD 0xc0011031 616#define MSR_AMD64_IBSFETCHPHYSAD 0xc0011032 617#define MSR_AMD64_IBSFETCH_REG_COUNT 3 618#define MSR_AMD64_IBSFETCH_REG_MASK ((1UL<<MSR_AMD64_IBSFETCH_REG_COUNT)-1) 619#define MSR_AMD64_IBSOPCTL 0xc0011033 620#define MSR_AMD64_IBSOPRIP 0xc0011034 621#define MSR_AMD64_IBSOPDATA 0xc0011035 622#define MSR_AMD64_IBSOPDATA2 0xc0011036 623#define MSR_AMD64_IBSOPDATA3 0xc0011037 624#define MSR_AMD64_IBSDCLINAD 0xc0011038 625#define MSR_AMD64_IBSDCPHYSAD 0xc0011039 626#define MSR_AMD64_IBSOP_REG_COUNT 7 627#define MSR_AMD64_IBSOP_REG_MASK ((1UL<<MSR_AMD64_IBSOP_REG_COUNT)-1) 628#define MSR_AMD64_IBSCTL 0xc001103a 629#define MSR_AMD64_IBSBRTARGET 0xc001103b 630#define MSR_AMD64_IBSOPDATA4 0xc001103d 631#define MSR_AMD64_IBS_REG_COUNT_MAX 8 /* includes MSR_AMD64_IBSBRTARGET */ 632#define MSR_AMD64_SEV 0xc0010131 633#define MSR_AMD64_SEV_ENABLED_BIT 0 634#define MSR_AMD64_SEV_ENABLED BIT_ULL(MSR_AMD64_SEV_ENABLED_BIT) 635 636/* Fam 17h MSRs */ 637#define MSR_F17H_IRPERF 0xc00000e9 638 639/* Fam 16h MSRs */ 640#define MSR_F16H_L2I_PERF_CTL 0xc0010230 641#define MSR_F16H_L2I_PERF_CTR 0xc0010231 642#define MSR_F16H_DR1_ADDR_MASK 0xc0011019 643#define MSR_F16H_DR2_ADDR_MASK 0xc001101a 644#define MSR_F16H_DR3_ADDR_MASK 0xc001101b 645#define MSR_F16H_DR0_ADDR_MASK 0xc0011027 646 647/* Fam 15h MSRs */ 648#define MSR_F15H_PERF_CTL 0xc0010200 649#define MSR_F15H_PERF_CTR 0xc0010201 650#define MSR_F15H_NB_PERF_CTL 0xc0010240 651#define MSR_F15H_NB_PERF_CTR 0xc0010241 652#define MSR_F15H_PTSC 0xc0010280 653#define MSR_F15H_IC_CFG 0xc0011021 654 655/* Fam 10h MSRs */ 656#define MSR_FAM10H_MMIO_CONF_BASE 0xc0010058 657#define FAM10H_MMIO_CONF_ENABLE (1<<0) 658#define FAM10H_MMIO_CONF_BUSRANGE_MASK 0xf 659#define FAM10H_MMIO_CONF_BUSRANGE_SHIFT 2 660#define FAM10H_MMIO_CONF_BASE_MASK 0xfffffffULL 661#define FAM10H_MMIO_CONF_BASE_SHIFT 20 662#define MSR_FAM10H_NODE_ID 0xc001100c 663#define MSR_F10H_DECFG 0xc0011029 664#define MSR_F10H_DECFG_LFENCE_SERIALIZE_BIT 1 665#define MSR_F10H_DECFG_LFENCE_SERIALIZE BIT_ULL(MSR_F10H_DECFG_LFENCE_SERIALIZE_BIT) 666 667/* K8 MSRs */ 668#define MSR_K8_TOP_MEM1 0xc001001a 669#define MSR_K8_TOP_MEM2 0xc001001d 670#define MSR_K8_SYSCFG 0xc0010010 671#define MSR_K8_SYSCFG_MEM_ENCRYPT_BIT 23 672#define MSR_K8_SYSCFG_MEM_ENCRYPT BIT_ULL(MSR_K8_SYSCFG_MEM_ENCRYPT_BIT) 673#define MSR_K8_INT_PENDING_MSG 0xc0010055 674/* C1E active bits in int pending message */ 675#define K8_INTP_C1E_ACTIVE_MASK 0x18000000 676#define MSR_K8_TSEG_ADDR 0xc0010112 677#define MSR_K8_TSEG_MASK 0xc0010113 678#define K8_MTRRFIXRANGE_DRAM_ENABLE 0x00040000 /* MtrrFixDramEn bit */ 679#define K8_MTRRFIXRANGE_DRAM_MODIFY 0x00080000 /* MtrrFixDramModEn bit */ 680#define K8_MTRR_RDMEM_WRMEM_MASK 0x18181818 /* Mask: RdMem|WrMem */ 681 682/* K7 MSRs */ 683#define MSR_K7_EVNTSEL0 0xc0010000 684#define MSR_K7_PERFCTR0 0xc0010004 685#define MSR_K7_EVNTSEL1 0xc0010001 686#define MSR_K7_PERFCTR1 0xc0010005 687#define MSR_K7_EVNTSEL2 0xc0010002 688#define MSR_K7_PERFCTR2 0xc0010006 689#define MSR_K7_EVNTSEL3 0xc0010003 690#define MSR_K7_PERFCTR3 0xc0010007 691#define MSR_K7_CLK_CTL 0xc001001b 692#define MSR_K7_HWCR 0xc0010015 693#define MSR_K7_HWCR_SMMLOCK_BIT 0 694#define MSR_K7_HWCR_SMMLOCK BIT_ULL(MSR_K7_HWCR_SMMLOCK_BIT) 695#define MSR_K7_FID_VID_CTL 0xc0010041 696#define MSR_K7_FID_VID_STATUS 0xc0010042 697 698/* K6 MSRs */ 699#define MSR_K6_WHCR 0xc0000082 700#define MSR_K6_UWCCR 0xc0000085 701#define MSR_K6_EPMR 0xc0000086 702#define MSR_K6_PSOR 0xc0000087 703#define MSR_K6_PFIR 0xc0000088 704 705/* Centaur-Hauls/IDT defined MSRs. */ 706#define MSR_IDT_FCR1 0x00000107 707#define MSR_IDT_FCR2 0x00000108 708#define MSR_IDT_FCR3 0x00000109 709#define MSR_IDT_FCR4 0x0000010a 710 711#define MSR_IDT_MCR0 0x00000110 712#define MSR_IDT_MCR1 0x00000111 713#define MSR_IDT_MCR2 0x00000112 714#define MSR_IDT_MCR3 0x00000113 715#define MSR_IDT_MCR4 0x00000114 716#define MSR_IDT_MCR5 0x00000115 717#define MSR_IDT_MCR6 0x00000116 718#define MSR_IDT_MCR7 0x00000117 719#define MSR_IDT_MCR_CTRL 0x00000120 720 721/* VIA Cyrix defined MSRs*/ 722#define MSR_VIA_FCR 0x00001107 723#define MSR_VIA_LONGHAUL 0x0000110a 724#define MSR_VIA_RNG 0x0000110b 725#define MSR_VIA_BCR2 0x00001147 726 727/* Transmeta defined MSRs */ 728#define MSR_TMTA_LONGRUN_CTRL 0x80868010 729#define MSR_TMTA_LONGRUN_FLAGS 0x80868011 730#define MSR_TMTA_LRTI_READOUT 0x80868018 731#define MSR_TMTA_LRTI_VOLT_MHZ 0x8086801a 732 733/* Intel defined MSRs. */ 734#define MSR_IA32_P5_MC_ADDR 0x00000000 735#define MSR_IA32_P5_MC_TYPE 0x00000001 736#define MSR_IA32_TSC 0x00000010 737#define MSR_IA32_PLATFORM_ID 0x00000017 738#define MSR_IA32_EBL_CR_POWERON 0x0000002a 739#define MSR_EBC_FREQUENCY_ID 0x0000002c 740#define MSR_SMI_COUNT 0x00000034 741#define MSR_IA32_FEATURE_CONTROL 0x0000003a 742#define MSR_IA32_TSC_ADJUST 0x0000003b 743#define MSR_IA32_BNDCFGS 0x00000d90 744 745#define MSR_IA32_BNDCFGS_RSVD 0x00000ffc 746 747#define MSR_IA32_XSS 0x00000da0 748 749#define FEATURE_CONTROL_LOCKED (1<<0) 750#define FEATURE_CONTROL_VMXON_ENABLED_INSIDE_SMX (1<<1) 751#define FEATURE_CONTROL_VMXON_ENABLED_OUTSIDE_SMX (1<<2) 752#define FEATURE_CONTROL_LMCE (1<<20) 753 754#define MSR_IA32_APICBASE 0x0000001b 755#define MSR_IA32_APICBASE_BSP (1<<8) 756#define MSR_IA32_APICBASE_ENABLE (1<<11) 757#define MSR_IA32_APICBASE_BASE (0xfffff<<12) 758 759#define MSR_IA32_TSCDEADLINE 0x000006e0 760 761#define MSR_IA32_UCODE_WRITE 0x00000079 762#define MSR_IA32_UCODE_REV 0x0000008b 763 764#define MSR_IA32_SMM_MONITOR_CTL 0x0000009b 765#define MSR_IA32_SMBASE 0x0000009e 766 767#define MSR_IA32_PERF_STATUS 0x00000198 768#define MSR_IA32_PERF_CTL 0x00000199 769#define INTEL_PERF_CTL_MASK 0xffff 770#define MSR_AMD_PSTATE_DEF_BASE 0xc0010064 771#define MSR_AMD_PERF_STATUS 0xc0010063 772#define MSR_AMD_PERF_CTL 0xc0010062 773 774#define MSR_IA32_MPERF 0x000000e7 775#define MSR_IA32_APERF 0x000000e8 776 777#define MSR_IA32_THERM_CONTROL 0x0000019a 778#define MSR_IA32_THERM_INTERRUPT 0x0000019b 779 780#define THERM_INT_HIGH_ENABLE (1 << 0) 781#define THERM_INT_LOW_ENABLE (1 << 1) 782#define THERM_INT_PLN_ENABLE (1 << 24) 783 784#define MSR_IA32_THERM_STATUS 0x0000019c 785 786#define THERM_STATUS_PROCHOT (1 << 0) 787#define THERM_STATUS_POWER_LIMIT (1 << 10) 788 789#define MSR_THERM2_CTL 0x0000019d 790 791#define MSR_THERM2_CTL_TM_SELECT (1ULL << 16) 792 793#define MSR_IA32_MISC_ENABLE 0x000001a0 794 795#define MSR_IA32_TEMPERATURE_TARGET 0x000001a2 796 797#define MSR_MISC_FEATURE_CONTROL 0x000001a4 798#define MSR_MISC_PWR_MGMT 0x000001aa 799 800#define MSR_IA32_ENERGY_PERF_BIAS 0x000001b0 801#define ENERGY_PERF_BIAS_PERFORMANCE 0 802#define ENERGY_PERF_BIAS_BALANCE_PERFORMANCE 4 803#define ENERGY_PERF_BIAS_NORMAL 6 804#define ENERGY_PERF_BIAS_BALANCE_POWERSAVE 8 805#define ENERGY_PERF_BIAS_POWERSAVE 15 806 807#define MSR_IA32_PACKAGE_THERM_STATUS 0x000001b1 808 809#define PACKAGE_THERM_STATUS_PROCHOT (1 << 0) 810#define PACKAGE_THERM_STATUS_POWER_LIMIT (1 << 10) 811 812#define MSR_IA32_PACKAGE_THERM_INTERRUPT 0x000001b2 813 814#define PACKAGE_THERM_INT_HIGH_ENABLE (1 << 0) 815#define PACKAGE_THERM_INT_LOW_ENABLE (1 << 1) 816#define PACKAGE_THERM_INT_PLN_ENABLE (1 << 24) 817 818/* Thermal Thresholds Support */ 819#define THERM_INT_THRESHOLD0_ENABLE (1 << 15) 820#define THERM_SHIFT_THRESHOLD0 8 821#define THERM_MASK_THRESHOLD0 (0x7f << THERM_SHIFT_THRESHOLD0) 822#define THERM_INT_THRESHOLD1_ENABLE (1 << 23) 823#define THERM_SHIFT_THRESHOLD1 16 824#define THERM_MASK_THRESHOLD1 (0x7f << THERM_SHIFT_THRESHOLD1) 825#define THERM_STATUS_THRESHOLD0 (1 << 6) 826#define THERM_LOG_THRESHOLD0 (1 << 7) 827#define THERM_STATUS_THRESHOLD1 (1 << 8) 828#define THERM_LOG_THRESHOLD1 (1 << 9) 829 830/* MISC_ENABLE bits: architectural */ 831#define MSR_IA32_MISC_ENABLE_FAST_STRING_BIT 0 832#define MSR_IA32_MISC_ENABLE_FAST_STRING (1ULL << MSR_IA32_MISC_ENABLE_FAST_STRING_BIT) 833#define MSR_IA32_MISC_ENABLE_TCC_BIT 1 834#define MSR_IA32_MISC_ENABLE_TCC (1ULL << MSR_IA32_MISC_ENABLE_TCC_BIT) 835#define MSR_IA32_MISC_ENABLE_EMON_BIT 7 836#define MSR_IA32_MISC_ENABLE_EMON (1ULL << MSR_IA32_MISC_ENABLE_EMON_BIT) 837#define MSR_IA32_MISC_ENABLE_BTS_UNAVAIL_BIT 11 838#define MSR_IA32_MISC_ENABLE_BTS_UNAVAIL (1ULL << MSR_IA32_MISC_ENABLE_BTS_UNAVAIL_BIT) 839#define MSR_IA32_MISC_ENABLE_PEBS_UNAVAIL_BIT 12 840#define MSR_IA32_MISC_ENABLE_PEBS_UNAVAIL (1ULL << MSR_IA32_MISC_ENABLE_PEBS_UNAVAIL_BIT) 841#define MSR_IA32_MISC_ENABLE_ENHANCED_SPEEDSTEP_BIT 16 842#define MSR_IA32_MISC_ENABLE_ENHANCED_SPEEDSTEP (1ULL << MSR_IA32_MISC_ENABLE_ENHANCED_SPEEDSTEP_BIT) 843#define MSR_IA32_MISC_ENABLE_MWAIT_BIT 18 844#define MSR_IA32_MISC_ENABLE_MWAIT (1ULL << MSR_IA32_MISC_ENABLE_MWAIT_BIT) 845#define MSR_IA32_MISC_ENABLE_LIMIT_CPUID_BIT 22 846#define MSR_IA32_MISC_ENABLE_LIMIT_CPUID (1ULL << MSR_IA32_MISC_ENABLE_LIMIT_CPUID_BIT) 847#define MSR_IA32_MISC_ENABLE_XTPR_DISABLE_BIT 23 848#define MSR_IA32_MISC_ENABLE_XTPR_DISABLE (1ULL << MSR_IA32_MISC_ENABLE_XTPR_DISABLE_BIT) 849#define MSR_IA32_MISC_ENABLE_XD_DISABLE_BIT 34 850#define MSR_IA32_MISC_ENABLE_XD_DISABLE (1ULL << MSR_IA32_MISC_ENABLE_XD_DISABLE_BIT) 851 852/* MISC_ENABLE bits: model-specific, meaning may vary from core to core */ 853#define MSR_IA32_MISC_ENABLE_X87_COMPAT_BIT 2 854#define MSR_IA32_MISC_ENABLE_X87_COMPAT (1ULL << MSR_IA32_MISC_ENABLE_X87_COMPAT_BIT) 855#define MSR_IA32_MISC_ENABLE_TM1_BIT 3 856#define MSR_IA32_MISC_ENABLE_TM1 (1ULL << MSR_IA32_MISC_ENABLE_TM1_BIT) 857#define MSR_IA32_MISC_ENABLE_SPLIT_LOCK_DISABLE_BIT 4 858#define MSR_IA32_MISC_ENABLE_SPLIT_LOCK_DISABLE (1ULL << MSR_IA32_MISC_ENABLE_SPLIT_LOCK_DISABLE_BIT) 859#define MSR_IA32_MISC_ENABLE_L3CACHE_DISABLE_BIT 6 860#define MSR_IA32_MISC_ENABLE_L3CACHE_DISABLE (1ULL << MSR_IA32_MISC_ENABLE_L3CACHE_DISABLE_BIT) 861#define MSR_IA32_MISC_ENABLE_SUPPRESS_LOCK_BIT 8 862#define MSR_IA32_MISC_ENABLE_SUPPRESS_LOCK (1ULL << MSR_IA32_MISC_ENABLE_SUPPRESS_LOCK_BIT) 863#define MSR_IA32_MISC_ENABLE_PREFETCH_DISABLE_BIT 9 864#define MSR_IA32_MISC_ENABLE_PREFETCH_DISABLE (1ULL << MSR_IA32_MISC_ENABLE_PREFETCH_DISABLE_BIT) 865#define MSR_IA32_MISC_ENABLE_FERR_BIT 10 866#define MSR_IA32_MISC_ENABLE_FERR (1ULL << MSR_IA32_MISC_ENABLE_FERR_BIT) 867#define MSR_IA32_MISC_ENABLE_FERR_MULTIPLEX_BIT 10 868#define MSR_IA32_MISC_ENABLE_FERR_MULTIPLEX (1ULL << MSR_IA32_MISC_ENABLE_FERR_MULTIPLEX_BIT) 869#define MSR_IA32_MISC_ENABLE_TM2_BIT 13 870#define MSR_IA32_MISC_ENABLE_TM2 (1ULL << MSR_IA32_MISC_ENABLE_TM2_BIT) 871#define MSR_IA32_MISC_ENABLE_ADJ_PREF_DISABLE_BIT 19 872#define MSR_IA32_MISC_ENABLE_ADJ_PREF_DISABLE (1ULL << MSR_IA32_MISC_ENABLE_ADJ_PREF_DISABLE_BIT) 873#define MSR_IA32_MISC_ENABLE_SPEEDSTEP_LOCK_BIT 20 874#define MSR_IA32_MISC_ENABLE_SPEEDSTEP_LOCK (1ULL << MSR_IA32_MISC_ENABLE_SPEEDSTEP_LOCK_BIT) 875#define MSR_IA32_MISC_ENABLE_L1D_CONTEXT_BIT 24 876#define MSR_IA32_MISC_ENABLE_L1D_CONTEXT (1ULL << MSR_IA32_MISC_ENABLE_L1D_CONTEXT_BIT) 877#define MSR_IA32_MISC_ENABLE_DCU_PREF_DISABLE_BIT 37 878#define MSR_IA32_MISC_ENABLE_DCU_PREF_DISABLE (1ULL << MSR_IA32_MISC_ENABLE_DCU_PREF_DISABLE_BIT) 879#define MSR_IA32_MISC_ENABLE_TURBO_DISABLE_BIT 38 880#define MSR_IA32_MISC_ENABLE_TURBO_DISABLE (1ULL << MSR_IA32_MISC_ENABLE_TURBO_DISABLE_BIT) 881#define MSR_IA32_MISC_ENABLE_IP_PREF_DISABLE_BIT 39 882#define MSR_IA32_MISC_ENABLE_IP_PREF_DISABLE (1ULL << MSR_IA32_MISC_ENABLE_IP_PREF_DISABLE_BIT) 883 884/* MISC_FEATURES_ENABLES non-architectural features */ 885#define MSR_MISC_FEATURES_ENABLES 0x00000140 886 887#define MSR_MISC_FEATURES_ENABLES_CPUID_FAULT_BIT 0 888#define MSR_MISC_FEATURES_ENABLES_CPUID_FAULT BIT_ULL(MSR_MISC_FEATURES_ENABLES_CPUID_FAULT_BIT) 889#define MSR_MISC_FEATURES_ENABLES_RING3MWAIT_BIT 1 890 891#define MSR_IA32_TSC_DEADLINE 0x000006E0 892 893/* P4/Xeon+ specific */ 894#define MSR_IA32_MCG_EAX 0x00000180 895#define MSR_IA32_MCG_EBX 0x00000181 896#define MSR_IA32_MCG_ECX 0x00000182 897#define MSR_IA32_MCG_EDX 0x00000183 898#define MSR_IA32_MCG_ESI 0x00000184 899#define MSR_IA32_MCG_EDI 0x00000185 900#define MSR_IA32_MCG_EBP 0x00000186 901#define MSR_IA32_MCG_ESP 0x00000187 902#define MSR_IA32_MCG_EFLAGS 0x00000188 903#define MSR_IA32_MCG_EIP 0x00000189 904#define MSR_IA32_MCG_RESERVED 0x0000018a 905 906/* Pentium IV performance counter MSRs */ 907#define MSR_P4_BPU_PERFCTR0 0x00000300 908#define MSR_P4_BPU_PERFCTR1 0x00000301 909#define MSR_P4_BPU_PERFCTR2 0x00000302 910#define MSR_P4_BPU_PERFCTR3 0x00000303 911#define MSR_P4_MS_PERFCTR0 0x00000304 912#define MSR_P4_MS_PERFCTR1 0x00000305 913#define MSR_P4_MS_PERFCTR2 0x00000306 914#define MSR_P4_MS_PERFCTR3 0x00000307 915#define MSR_P4_FLAME_PERFCTR0 0x00000308 916#define MSR_P4_FLAME_PERFCTR1 0x00000309 917#define MSR_P4_FLAME_PERFCTR2 0x0000030a 918#define MSR_P4_FLAME_PERFCTR3 0x0000030b 919#define MSR_P4_IQ_PERFCTR0 0x0000030c 920#define MSR_P4_IQ_PERFCTR1 0x0000030d 921#define MSR_P4_IQ_PERFCTR2 0x0000030e 922#define MSR_P4_IQ_PERFCTR3 0x0000030f 923#define MSR_P4_IQ_PERFCTR4 0x00000310 924#define MSR_P4_IQ_PERFCTR5 0x00000311 925#define MSR_P4_BPU_CCCR0 0x00000360 926#define MSR_P4_BPU_CCCR1 0x00000361 927#define MSR_P4_BPU_CCCR2 0x00000362 928#define MSR_P4_BPU_CCCR3 0x00000363 929#define MSR_P4_MS_CCCR0 0x00000364 930#define MSR_P4_MS_CCCR1 0x00000365 931#define MSR_P4_MS_CCCR2 0x00000366 932#define MSR_P4_MS_CCCR3 0x00000367 933#define MSR_P4_FLAME_CCCR0 0x00000368 934#define MSR_P4_FLAME_CCCR1 0x00000369 935#define MSR_P4_FLAME_CCCR2 0x0000036a 936#define MSR_P4_FLAME_CCCR3 0x0000036b 937#define MSR_P4_IQ_CCCR0 0x0000036c 938#define MSR_P4_IQ_CCCR1 0x0000036d 939#define MSR_P4_IQ_CCCR2 0x0000036e 940#define MSR_P4_IQ_CCCR3 0x0000036f 941#define MSR_P4_IQ_CCCR4 0x00000370 942#define MSR_P4_IQ_CCCR5 0x00000371 943#define MSR_P4_ALF_ESCR0 0x000003ca 944#define MSR_P4_ALF_ESCR1 0x000003cb 945#define MSR_P4_BPU_ESCR0 0x000003b2 946#define MSR_P4_BPU_ESCR1 0x000003b3 947#define MSR_P4_BSU_ESCR0 0x000003a0 948#define MSR_P4_BSU_ESCR1 0x000003a1 949#define MSR_P4_CRU_ESCR0 0x000003b8 950#define MSR_P4_CRU_ESCR1 0x000003b9 951#define MSR_P4_CRU_ESCR2 0x000003cc 952#define MSR_P4_CRU_ESCR3 0x000003cd 953#define MSR_P4_CRU_ESCR4 0x000003e0 954#define MSR_P4_CRU_ESCR5 0x000003e1 955#define MSR_P4_DAC_ESCR0 0x000003a8 956#define MSR_P4_DAC_ESCR1 0x000003a9 957#define MSR_P4_FIRM_ESCR0 0x000003a4 958#define MSR_P4_FIRM_ESCR1 0x000003a5 959#define MSR_P4_FLAME_ESCR0 0x000003a6 960#define MSR_P4_FLAME_ESCR1 0x000003a7 961#define MSR_P4_FSB_ESCR0 0x000003a2 962#define MSR_P4_FSB_ESCR1 0x000003a3 963#define MSR_P4_IQ_ESCR0 0x000003ba 964#define MSR_P4_IQ_ESCR1 0x000003bb 965#define MSR_P4_IS_ESCR0 0x000003b4 966#define MSR_P4_IS_ESCR1 0x000003b5 967#define MSR_P4_ITLB_ESCR0 0x000003b6 968#define MSR_P4_ITLB_ESCR1 0x000003b7 969#define MSR_P4_IX_ESCR0 0x000003c8 970#define MSR_P4_IX_ESCR1 0x000003c9 971#define MSR_P4_MOB_ESCR0 0x000003aa 972#define MSR_P4_MOB_ESCR1 0x000003ab 973#define MSR_P4_MS_ESCR0 0x000003c0 974#define MSR_P4_MS_ESCR1 0x000003c1 975#define MSR_P4_PMH_ESCR0 0x000003ac 976#define MSR_P4_PMH_ESCR1 0x000003ad 977#define MSR_P4_RAT_ESCR0 0x000003bc 978#define MSR_P4_RAT_ESCR1 0x000003bd 979#define MSR_P4_SAAT_ESCR0 0x000003ae 980#define MSR_P4_SAAT_ESCR1 0x000003af 981#define MSR_P4_SSU_ESCR0 0x000003be 982#define MSR_P4_SSU_ESCR1 0x000003bf /* guess: not in manual */ 983 984#define MSR_P4_TBPU_ESCR0 0x000003c2 985#define MSR_P4_TBPU_ESCR1 0x000003c3 986#define MSR_P4_TC_ESCR0 0x000003c4 987#define MSR_P4_TC_ESCR1 0x000003c5 988#define MSR_P4_U2L_ESCR0 0x000003b0 989#define MSR_P4_U2L_ESCR1 0x000003b1 990 991#define MSR_P4_PEBS_MATRIX_VERT 0x000003f2 992 993/* Intel Core-based CPU performance counters */ 994#define MSR_CORE_PERF_FIXED_CTR0 0x00000309 995#define MSR_CORE_PERF_FIXED_CTR1 0x0000030a 996#define MSR_CORE_PERF_FIXED_CTR2 0x0000030b 997#define MSR_CORE_PERF_FIXED_CTR_CTRL 0x0000038d 998#define MSR_CORE_PERF_GLOBAL_STATUS 0x0000038e 999#define MSR_CORE_PERF_GLOBAL_CTRL 0x0000038f 1000#define MSR_CORE_PERF_GLOBAL_OVF_CTRL 0x00000390 1001 1002/* Geode defined MSRs */ 1003#define MSR_GEODE_BUSCONT_CONF0 0x00001900 1004 1005/* Intel VT MSRs */ 1006#define MSR_IA32_VMX_BASIC 0x00000480 1007#define MSR_IA32_VMX_PINBASED_CTLS 0x00000481 1008#define MSR_IA32_VMX_PROCBASED_CTLS 0x00000482 1009#define MSR_IA32_VMX_EXIT_CTLS 0x00000483 1010#define MSR_IA32_VMX_ENTRY_CTLS 0x00000484 1011#define MSR_IA32_VMX_MISC 0x00000485 1012#define MSR_IA32_VMX_CR0_FIXED0 0x00000486 1013#define MSR_IA32_VMX_CR0_FIXED1 0x00000487 1014#define MSR_IA32_VMX_CR4_FIXED0 0x00000488 1015#define MSR_IA32_VMX_CR4_FIXED1 0x00000489 1016#define MSR_IA32_VMX_VMCS_ENUM 0x0000048a 1017#define MSR_IA32_VMX_PROCBASED_CTLS2 0x0000048b 1018#define MSR_IA32_VMX_EPT_VPID_CAP 0x0000048c 1019#define MSR_IA32_VMX_TRUE_PINBASED_CTLS 0x0000048d 1020#define MSR_IA32_VMX_TRUE_PROCBASED_CTLS 0x0000048e 1021#define MSR_IA32_VMX_TRUE_EXIT_CTLS 0x0000048f 1022#define MSR_IA32_VMX_TRUE_ENTRY_CTLS 0x00000490 1023#define MSR_IA32_VMX_VMFUNC 0x00000491 1024 1025/* VMX_BASIC bits and bitmasks */ 1026#define VMX_BASIC_VMCS_SIZE_SHIFT 32 1027#define VMX_BASIC_TRUE_CTLS (1ULL << 55) 1028#define VMX_BASIC_64 0x0001000000000000LLU 1029#define VMX_BASIC_MEM_TYPE_SHIFT 50 1030#define VMX_BASIC_MEM_TYPE_MASK 0x003c000000000000LLU 1031#define VMX_BASIC_MEM_TYPE_WB 6LLU 1032#define VMX_BASIC_INOUT 0x0040000000000000LLU 1033 1034/* MSR_IA32_VMX_MISC bits */ 1035#define MSR_IA32_VMX_MISC_VMWRITE_SHADOW_RO_FIELDS (1ULL << 29) 1036#define MSR_IA32_VMX_MISC_PREEMPTION_TIMER_SCALE 0x1F 1037/* AMD-V MSRs */ 1038 1039#define MSR_VM_CR 0xc0010114 1040#define MSR_VM_IGNNE 0xc0010115 1041#define MSR_VM_HSAVE_PA 0xc0010117 1042 1043#endif /* !SELFTEST_KVM_X86_H */