Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux
1
fork

Configure Feed

Select the types of activity you want to include in your feed.

at v4.11-rc7 317 lines 9.5 kB view raw
1/* 2 * Copyright (C) 2012 - Virtual Open Systems and Columbia University 3 * Author: Christoffer Dall <c.dall@virtualopensystems.com> 4 * 5 * This program is free software; you can redistribute it and/or modify 6 * it under the terms of the GNU General Public License, version 2, as 7 * published by the Free Software Foundation. 8 * 9 * This program is distributed in the hope that it will be useful, 10 * but WITHOUT ANY WARRANTY; without even the implied warranty of 11 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the 12 * GNU General Public License for more details. 13 * 14 * You should have received a copy of the GNU General Public License 15 * along with this program; if not, write to the Free Software 16 * Foundation, 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA. 17 */ 18 19#ifndef __ARM_KVM_HOST_H__ 20#define __ARM_KVM_HOST_H__ 21 22#include <linux/types.h> 23#include <linux/kvm_types.h> 24#include <asm/kvm.h> 25#include <asm/kvm_asm.h> 26#include <asm/kvm_mmio.h> 27#include <asm/fpstate.h> 28#include <kvm/arm_arch_timer.h> 29 30#define __KVM_HAVE_ARCH_INTC_INITIALIZED 31 32#define KVM_USER_MEM_SLOTS 32 33#define KVM_COALESCED_MMIO_PAGE_OFFSET 1 34#define KVM_HAVE_ONE_REG 35#define KVM_HALT_POLL_NS_DEFAULT 500000 36 37#define KVM_VCPU_MAX_FEATURES 2 38 39#include <kvm/arm_vgic.h> 40 41 42#ifdef CONFIG_ARM_GIC_V3 43#define KVM_MAX_VCPUS VGIC_V3_MAX_CPUS 44#else 45#define KVM_MAX_VCPUS VGIC_V2_MAX_CPUS 46#endif 47 48#define KVM_REQ_VCPU_EXIT 8 49 50u32 *kvm_vcpu_reg(struct kvm_vcpu *vcpu, u8 reg_num, u32 mode); 51int __attribute_const__ kvm_target_cpu(void); 52int kvm_reset_vcpu(struct kvm_vcpu *vcpu); 53void kvm_reset_coprocs(struct kvm_vcpu *vcpu); 54 55struct kvm_arch { 56 /* VTTBR value associated with below pgd and vmid */ 57 u64 vttbr; 58 59 /* The last vcpu id that ran on each physical CPU */ 60 int __percpu *last_vcpu_ran; 61 62 /* 63 * Anything that is not used directly from assembly code goes 64 * here. 65 */ 66 67 /* The VMID generation used for the virt. memory system */ 68 u64 vmid_gen; 69 u32 vmid; 70 71 /* Stage-2 page table */ 72 pgd_t *pgd; 73 74 /* Interrupt controller */ 75 struct vgic_dist vgic; 76 int max_vcpus; 77}; 78 79#define KVM_NR_MEM_OBJS 40 80 81/* 82 * We don't want allocation failures within the mmu code, so we preallocate 83 * enough memory for a single page fault in a cache. 84 */ 85struct kvm_mmu_memory_cache { 86 int nobjs; 87 void *objects[KVM_NR_MEM_OBJS]; 88}; 89 90struct kvm_vcpu_fault_info { 91 u32 hsr; /* Hyp Syndrome Register */ 92 u32 hxfar; /* Hyp Data/Inst. Fault Address Register */ 93 u32 hpfar; /* Hyp IPA Fault Address Register */ 94}; 95 96/* 97 * 0 is reserved as an invalid value. 98 * Order should be kept in sync with the save/restore code. 99 */ 100enum vcpu_sysreg { 101 __INVALID_SYSREG__, 102 c0_MPIDR, /* MultiProcessor ID Register */ 103 c0_CSSELR, /* Cache Size Selection Register */ 104 c1_SCTLR, /* System Control Register */ 105 c1_ACTLR, /* Auxiliary Control Register */ 106 c1_CPACR, /* Coprocessor Access Control */ 107 c2_TTBR0, /* Translation Table Base Register 0 */ 108 c2_TTBR0_high, /* TTBR0 top 32 bits */ 109 c2_TTBR1, /* Translation Table Base Register 1 */ 110 c2_TTBR1_high, /* TTBR1 top 32 bits */ 111 c2_TTBCR, /* Translation Table Base Control R. */ 112 c3_DACR, /* Domain Access Control Register */ 113 c5_DFSR, /* Data Fault Status Register */ 114 c5_IFSR, /* Instruction Fault Status Register */ 115 c5_ADFSR, /* Auxilary Data Fault Status R */ 116 c5_AIFSR, /* Auxilary Instrunction Fault Status R */ 117 c6_DFAR, /* Data Fault Address Register */ 118 c6_IFAR, /* Instruction Fault Address Register */ 119 c7_PAR, /* Physical Address Register */ 120 c7_PAR_high, /* PAR top 32 bits */ 121 c9_L2CTLR, /* Cortex A15/A7 L2 Control Register */ 122 c10_PRRR, /* Primary Region Remap Register */ 123 c10_NMRR, /* Normal Memory Remap Register */ 124 c12_VBAR, /* Vector Base Address Register */ 125 c13_CID, /* Context ID Register */ 126 c13_TID_URW, /* Thread ID, User R/W */ 127 c13_TID_URO, /* Thread ID, User R/O */ 128 c13_TID_PRIV, /* Thread ID, Privileged */ 129 c14_CNTKCTL, /* Timer Control Register (PL1) */ 130 c10_AMAIR0, /* Auxilary Memory Attribute Indirection Reg0 */ 131 c10_AMAIR1, /* Auxilary Memory Attribute Indirection Reg1 */ 132 NR_CP15_REGS /* Number of regs (incl. invalid) */ 133}; 134 135struct kvm_cpu_context { 136 struct kvm_regs gp_regs; 137 struct vfp_hard_struct vfp; 138 u32 cp15[NR_CP15_REGS]; 139}; 140 141typedef struct kvm_cpu_context kvm_cpu_context_t; 142 143struct kvm_vcpu_arch { 144 struct kvm_cpu_context ctxt; 145 146 int target; /* Processor target */ 147 DECLARE_BITMAP(features, KVM_VCPU_MAX_FEATURES); 148 149 /* The CPU type we expose to the VM */ 150 u32 midr; 151 152 /* HYP trapping configuration */ 153 u32 hcr; 154 155 /* Interrupt related fields */ 156 u32 irq_lines; /* IRQ and FIQ levels */ 157 158 /* Exception Information */ 159 struct kvm_vcpu_fault_info fault; 160 161 /* Host FP context */ 162 kvm_cpu_context_t *host_cpu_context; 163 164 /* VGIC state */ 165 struct vgic_cpu vgic_cpu; 166 struct arch_timer_cpu timer_cpu; 167 168 /* 169 * Anything that is not used directly from assembly code goes 170 * here. 171 */ 172 173 /* vcpu power-off state */ 174 bool power_off; 175 176 /* Don't run the guest (internal implementation need) */ 177 bool pause; 178 179 /* IO related fields */ 180 struct kvm_decode mmio_decode; 181 182 /* Cache some mmu pages needed inside spinlock regions */ 183 struct kvm_mmu_memory_cache mmu_page_cache; 184 185 /* Detect first run of a vcpu */ 186 bool has_run_once; 187}; 188 189struct kvm_vm_stat { 190 ulong remote_tlb_flush; 191}; 192 193struct kvm_vcpu_stat { 194 u64 halt_successful_poll; 195 u64 halt_attempted_poll; 196 u64 halt_poll_invalid; 197 u64 halt_wakeup; 198 u64 hvc_exit_stat; 199 u64 wfe_exit_stat; 200 u64 wfi_exit_stat; 201 u64 mmio_exit_user; 202 u64 mmio_exit_kernel; 203 u64 exits; 204}; 205 206#define vcpu_cp15(v,r) (v)->arch.ctxt.cp15[r] 207 208int kvm_vcpu_preferred_target(struct kvm_vcpu_init *init); 209unsigned long kvm_arm_num_regs(struct kvm_vcpu *vcpu); 210int kvm_arm_copy_reg_indices(struct kvm_vcpu *vcpu, u64 __user *indices); 211int kvm_arm_get_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg); 212int kvm_arm_set_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *reg); 213unsigned long kvm_call_hyp(void *hypfn, ...); 214void force_vm_exit(const cpumask_t *mask); 215 216#define KVM_ARCH_WANT_MMU_NOTIFIER 217int kvm_unmap_hva(struct kvm *kvm, unsigned long hva); 218int kvm_unmap_hva_range(struct kvm *kvm, 219 unsigned long start, unsigned long end); 220void kvm_set_spte_hva(struct kvm *kvm, unsigned long hva, pte_t pte); 221 222unsigned long kvm_arm_num_regs(struct kvm_vcpu *vcpu); 223int kvm_arm_copy_reg_indices(struct kvm_vcpu *vcpu, u64 __user *indices); 224int kvm_age_hva(struct kvm *kvm, unsigned long start, unsigned long end); 225int kvm_test_age_hva(struct kvm *kvm, unsigned long hva); 226 227/* We do not have shadow page tables, hence the empty hooks */ 228static inline void kvm_arch_mmu_notifier_invalidate_page(struct kvm *kvm, 229 unsigned long address) 230{ 231} 232 233struct kvm_vcpu *kvm_arm_get_running_vcpu(void); 234struct kvm_vcpu __percpu **kvm_get_running_vcpus(void); 235void kvm_arm_halt_guest(struct kvm *kvm); 236void kvm_arm_resume_guest(struct kvm *kvm); 237void kvm_arm_halt_vcpu(struct kvm_vcpu *vcpu); 238void kvm_arm_resume_vcpu(struct kvm_vcpu *vcpu); 239 240int kvm_arm_copy_coproc_indices(struct kvm_vcpu *vcpu, u64 __user *uindices); 241unsigned long kvm_arm_num_coproc_regs(struct kvm_vcpu *vcpu); 242int kvm_arm_coproc_get_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *); 243int kvm_arm_coproc_set_reg(struct kvm_vcpu *vcpu, const struct kvm_one_reg *); 244 245int handle_exit(struct kvm_vcpu *vcpu, struct kvm_run *run, 246 int exception_index); 247 248static inline void __cpu_init_hyp_mode(phys_addr_t pgd_ptr, 249 unsigned long hyp_stack_ptr, 250 unsigned long vector_ptr) 251{ 252 /* 253 * Call initialization code, and switch to the full blown HYP 254 * code. The init code doesn't need to preserve these 255 * registers as r0-r3 are already callee saved according to 256 * the AAPCS. 257 * Note that we slightly misuse the prototype by casting the 258 * stack pointer to a void *. 259 260 * The PGDs are always passed as the third argument, in order 261 * to be passed into r2-r3 to the init code (yes, this is 262 * compliant with the PCS!). 263 */ 264 265 kvm_call_hyp((void*)hyp_stack_ptr, vector_ptr, pgd_ptr); 266} 267 268static inline void __cpu_init_stage2(void) 269{ 270 kvm_call_hyp(__init_stage2_translation); 271} 272 273static inline void __cpu_reset_hyp_mode(unsigned long vector_ptr, 274 phys_addr_t phys_idmap_start) 275{ 276 kvm_call_hyp((void *)virt_to_idmap(__kvm_hyp_reset), vector_ptr); 277} 278 279static inline int kvm_arch_dev_ioctl_check_extension(struct kvm *kvm, long ext) 280{ 281 return 0; 282} 283 284int kvm_perf_init(void); 285int kvm_perf_teardown(void); 286 287void kvm_mmu_wp_memory_region(struct kvm *kvm, int slot); 288 289struct kvm_vcpu *kvm_mpidr_to_vcpu(struct kvm *kvm, unsigned long mpidr); 290 291static inline void kvm_arch_hardware_unsetup(void) {} 292static inline void kvm_arch_sync_events(struct kvm *kvm) {} 293static inline void kvm_arch_vcpu_uninit(struct kvm_vcpu *vcpu) {} 294static inline void kvm_arch_sched_in(struct kvm_vcpu *vcpu, int cpu) {} 295static inline void kvm_arch_vcpu_block_finish(struct kvm_vcpu *vcpu) {} 296 297static inline void kvm_arm_init_debug(void) {} 298static inline void kvm_arm_setup_debug(struct kvm_vcpu *vcpu) {} 299static inline void kvm_arm_clear_debug(struct kvm_vcpu *vcpu) {} 300static inline void kvm_arm_reset_debug_ptr(struct kvm_vcpu *vcpu) {} 301static inline int kvm_arm_vcpu_arch_set_attr(struct kvm_vcpu *vcpu, 302 struct kvm_device_attr *attr) 303{ 304 return -ENXIO; 305} 306static inline int kvm_arm_vcpu_arch_get_attr(struct kvm_vcpu *vcpu, 307 struct kvm_device_attr *attr) 308{ 309 return -ENXIO; 310} 311static inline int kvm_arm_vcpu_arch_has_attr(struct kvm_vcpu *vcpu, 312 struct kvm_device_attr *attr) 313{ 314 return -ENXIO; 315} 316 317#endif /* __ARM_KVM_HOST_H__ */