Linux kernel mirror (for testing)
git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel
os
linux
1/* SPDX-License-Identifier: GPL-2.0-only */
2
3#ifndef __KVM_TYPES_H__
4#define __KVM_TYPES_H__
5
6struct kvm;
7struct kvm_async_pf;
8struct kvm_device_ops;
9struct kvm_interrupt;
10struct kvm_irq_routing_table;
11struct kvm_memory_slot;
12struct kvm_one_reg;
13struct kvm_run;
14struct kvm_userspace_memory_region;
15struct kvm_vcpu;
16struct kvm_vcpu_init;
17struct kvm_memslots;
18
19enum kvm_mr_change;
20
21#include <linux/bits.h>
22#include <linux/mutex.h>
23#include <linux/types.h>
24#include <linux/spinlock_types.h>
25
26#include <asm/kvm_types.h>
27
28/*
29 * Address types:
30 *
31 * gva - guest virtual address
32 * gpa - guest physical address
33 * gfn - guest frame number
34 * hva - host virtual address
35 * hpa - host physical address
36 * hfn - host frame number
37 */
38
39typedef unsigned long gva_t;
40typedef u64 gpa_t;
41typedef u64 gfn_t;
42
43#define GPA_INVALID (~(gpa_t)0)
44
45typedef unsigned long hva_t;
46typedef u64 hpa_t;
47typedef u64 hfn_t;
48
49typedef hfn_t kvm_pfn_t;
50
51enum pfn_cache_usage {
52 KVM_GUEST_USES_PFN = BIT(0),
53 KVM_HOST_USES_PFN = BIT(1),
54 KVM_GUEST_AND_HOST_USE_PFN = KVM_GUEST_USES_PFN | KVM_HOST_USES_PFN,
55};
56
57struct gfn_to_hva_cache {
58 u64 generation;
59 gpa_t gpa;
60 unsigned long hva;
61 unsigned long len;
62 struct kvm_memory_slot *memslot;
63};
64
65struct gfn_to_pfn_cache {
66 u64 generation;
67 gpa_t gpa;
68 unsigned long uhva;
69 struct kvm_memory_slot *memslot;
70 struct kvm_vcpu *vcpu;
71 struct list_head list;
72 rwlock_t lock;
73 struct mutex refresh_lock;
74 void *khva;
75 kvm_pfn_t pfn;
76 enum pfn_cache_usage usage;
77 bool active;
78 bool valid;
79};
80
81#ifdef KVM_ARCH_NR_OBJS_PER_MEMORY_CACHE
82/*
83 * Memory caches are used to preallocate memory ahead of various MMU flows,
84 * e.g. page fault handlers. Gracefully handling allocation failures deep in
85 * MMU flows is problematic, as is triggering reclaim, I/O, etc... while
86 * holding MMU locks. Note, these caches act more like prefetch buffers than
87 * classical caches, i.e. objects are not returned to the cache on being freed.
88 *
89 * The @capacity field and @objects array are lazily initialized when the cache
90 * is topped up (__kvm_mmu_topup_memory_cache()).
91 */
92struct kvm_mmu_memory_cache {
93 int nobjs;
94 gfp_t gfp_zero;
95 gfp_t gfp_custom;
96 struct kmem_cache *kmem_cache;
97 int capacity;
98 void **objects;
99};
100#endif
101
102#define HALT_POLL_HIST_COUNT 32
103
104struct kvm_vm_stat_generic {
105 u64 remote_tlb_flush;
106 u64 remote_tlb_flush_requests;
107};
108
109struct kvm_vcpu_stat_generic {
110 u64 halt_successful_poll;
111 u64 halt_attempted_poll;
112 u64 halt_poll_invalid;
113 u64 halt_wakeup;
114 u64 halt_poll_success_ns;
115 u64 halt_poll_fail_ns;
116 u64 halt_wait_ns;
117 u64 halt_poll_success_hist[HALT_POLL_HIST_COUNT];
118 u64 halt_poll_fail_hist[HALT_POLL_HIST_COUNT];
119 u64 halt_wait_hist[HALT_POLL_HIST_COUNT];
120 u64 blocking;
121};
122
123#define KVM_STATS_NAME_SIZE 48
124
125#endif /* __KVM_TYPES_H__ */