at master 264 lines 6.9 kB view raw
1// SPDX-License-Identifier: GPL-2.0 2#include <errno.h> 3#include <string.h> 4#include "../../../util/kvm-stat.h" 5#include "../../../util/evsel.h" 6#include "../../../util/env.h" 7#include <asm/svm.h> 8#include <asm/vmx.h> 9#include <asm/kvm.h> 10#include <subcmd/parse-options.h> 11 12define_exit_reasons_table(vmx_exit_reasons, VMX_EXIT_REASONS); 13define_exit_reasons_table(svm_exit_reasons, SVM_EXIT_REASONS); 14 15static struct kvm_events_ops exit_events = { 16 .is_begin_event = exit_event_begin, 17 .is_end_event = exit_event_end, 18 .decode_key = exit_event_decode_key, 19 .name = "VM-EXIT" 20}; 21 22const char *vcpu_id_str = "vcpu_id"; 23const char *kvm_exit_reason = "exit_reason"; 24const char *kvm_entry_trace = "kvm:kvm_entry"; 25const char *kvm_exit_trace = "kvm:kvm_exit"; 26 27/* 28 * For the mmio events, we treat: 29 * the time of MMIO write: kvm_mmio(KVM_TRACE_MMIO_WRITE...) -> kvm_entry 30 * the time of MMIO read: kvm_exit -> kvm_mmio(KVM_TRACE_MMIO_READ...). 31 */ 32static void mmio_event_get_key(struct evsel *evsel, struct perf_sample *sample, 33 struct event_key *key) 34{ 35 key->key = evsel__intval(evsel, sample, "gpa"); 36 key->info = evsel__intval(evsel, sample, "type"); 37} 38 39#define KVM_TRACE_MMIO_READ_UNSATISFIED 0 40#define KVM_TRACE_MMIO_READ 1 41#define KVM_TRACE_MMIO_WRITE 2 42 43static bool mmio_event_begin(struct evsel *evsel, 44 struct perf_sample *sample, struct event_key *key) 45{ 46 /* MMIO read begin event in kernel. */ 47 if (kvm_exit_event(evsel)) 48 return true; 49 50 /* MMIO write begin event in kernel. */ 51 if (evsel__name_is(evsel, "kvm:kvm_mmio") && 52 evsel__intval(evsel, sample, "type") == KVM_TRACE_MMIO_WRITE) { 53 mmio_event_get_key(evsel, sample, key); 54 return true; 55 } 56 57 return false; 58} 59 60static bool mmio_event_end(struct evsel *evsel, struct perf_sample *sample, 61 struct event_key *key) 62{ 63 /* MMIO write end event in kernel. */ 64 if (kvm_entry_event(evsel)) 65 return true; 66 67 /* MMIO read end event in kernel.*/ 68 if (evsel__name_is(evsel, "kvm:kvm_mmio") && 69 evsel__intval(evsel, sample, "type") == KVM_TRACE_MMIO_READ) { 70 mmio_event_get_key(evsel, sample, key); 71 return true; 72 } 73 74 return false; 75} 76 77static void mmio_event_decode_key(struct perf_kvm_stat *kvm __maybe_unused, 78 struct event_key *key, 79 char *decode) 80{ 81 scnprintf(decode, KVM_EVENT_NAME_LEN, "%#lx:%s", 82 (unsigned long)key->key, 83 key->info == KVM_TRACE_MMIO_WRITE ? "W" : "R"); 84} 85 86static struct kvm_events_ops mmio_events = { 87 .is_begin_event = mmio_event_begin, 88 .is_end_event = mmio_event_end, 89 .decode_key = mmio_event_decode_key, 90 .name = "MMIO Access" 91}; 92 93 /* The time of emulation pio access is from kvm_pio to kvm_entry. */ 94static void ioport_event_get_key(struct evsel *evsel, 95 struct perf_sample *sample, 96 struct event_key *key) 97{ 98 key->key = evsel__intval(evsel, sample, "port"); 99 key->info = evsel__intval(evsel, sample, "rw"); 100} 101 102static bool ioport_event_begin(struct evsel *evsel, 103 struct perf_sample *sample, 104 struct event_key *key) 105{ 106 if (evsel__name_is(evsel, "kvm:kvm_pio")) { 107 ioport_event_get_key(evsel, sample, key); 108 return true; 109 } 110 111 return false; 112} 113 114static bool ioport_event_end(struct evsel *evsel, 115 struct perf_sample *sample __maybe_unused, 116 struct event_key *key __maybe_unused) 117{ 118 return kvm_entry_event(evsel); 119} 120 121static void ioport_event_decode_key(struct perf_kvm_stat *kvm __maybe_unused, 122 struct event_key *key, 123 char *decode) 124{ 125 scnprintf(decode, KVM_EVENT_NAME_LEN, "%#llx:%s", 126 (unsigned long long)key->key, 127 key->info ? "POUT" : "PIN"); 128} 129 130static struct kvm_events_ops ioport_events = { 131 .is_begin_event = ioport_event_begin, 132 .is_end_event = ioport_event_end, 133 .decode_key = ioport_event_decode_key, 134 .name = "IO Port Access" 135}; 136 137 /* The time of emulation msr is from kvm_msr to kvm_entry. */ 138static void msr_event_get_key(struct evsel *evsel, 139 struct perf_sample *sample, 140 struct event_key *key) 141{ 142 key->key = evsel__intval(evsel, sample, "ecx"); 143 key->info = evsel__intval(evsel, sample, "write"); 144} 145 146static bool msr_event_begin(struct evsel *evsel, 147 struct perf_sample *sample, 148 struct event_key *key) 149{ 150 if (evsel__name_is(evsel, "kvm:kvm_msr")) { 151 msr_event_get_key(evsel, sample, key); 152 return true; 153 } 154 155 return false; 156} 157 158static bool msr_event_end(struct evsel *evsel, 159 struct perf_sample *sample __maybe_unused, 160 struct event_key *key __maybe_unused) 161{ 162 return kvm_entry_event(evsel); 163} 164 165static void msr_event_decode_key(struct perf_kvm_stat *kvm __maybe_unused, 166 struct event_key *key, 167 char *decode) 168{ 169 scnprintf(decode, KVM_EVENT_NAME_LEN, "%#llx:%s", 170 (unsigned long long)key->key, 171 key->info ? "W" : "R"); 172} 173 174static struct kvm_events_ops msr_events = { 175 .is_begin_event = msr_event_begin, 176 .is_end_event = msr_event_end, 177 .decode_key = msr_event_decode_key, 178 .name = "MSR Access" 179}; 180 181const char *kvm_events_tp[] = { 182 "kvm:kvm_entry", 183 "kvm:kvm_exit", 184 "kvm:kvm_mmio", 185 "kvm:kvm_pio", 186 "kvm:kvm_msr", 187 NULL, 188}; 189 190struct kvm_reg_events_ops kvm_reg_events_ops[] = { 191 { .name = "vmexit", .ops = &exit_events }, 192 { .name = "mmio", .ops = &mmio_events }, 193 { .name = "ioport", .ops = &ioport_events }, 194 { .name = "msr", .ops = &msr_events }, 195 { NULL, NULL }, 196}; 197 198const char * const kvm_skip_events[] = { 199 "HLT", 200 NULL, 201}; 202 203int cpu_isa_init(struct perf_kvm_stat *kvm, const char *cpuid) 204{ 205 if (strstr(cpuid, "Intel")) { 206 kvm->exit_reasons = vmx_exit_reasons; 207 kvm->exit_reasons_isa = "VMX"; 208 } else if (strstr(cpuid, "AMD") || strstr(cpuid, "Hygon")) { 209 kvm->exit_reasons = svm_exit_reasons; 210 kvm->exit_reasons_isa = "SVM"; 211 } else 212 return -ENOTSUP; 213 214 return 0; 215} 216 217/* 218 * After KVM supports PEBS for guest on Intel platforms 219 * (https://lore.kernel.org/all/20220411101946.20262-1-likexu@tencent.com/), 220 * host loses the capability to sample guest with PEBS since all PEBS related 221 * MSRs are switched to guest value after vm-entry, like IA32_DS_AREA MSR is 222 * switched to guest GVA at vm-entry. This would lead to "perf kvm record" 223 * fails to sample guest on Intel platforms since "cycles:P" event is used to 224 * sample guest by default. 225 * 226 * So, to avoid this issue explicitly use "cycles" instead of "cycles:P" event 227 * by default to sample guest on Intel platforms. 228 */ 229int kvm_add_default_arch_event(int *argc, const char **argv) 230{ 231 const char **tmp; 232 bool event = false; 233 int ret = 0, i, j = *argc; 234 235 const struct option event_options[] = { 236 OPT_BOOLEAN('e', "event", &event, NULL), 237 OPT_BOOLEAN(0, "pfm-events", &event, NULL), 238 OPT_END() 239 }; 240 241 if (!x86__is_intel_cpu()) 242 return 0; 243 244 tmp = calloc(j + 1, sizeof(char *)); 245 if (!tmp) 246 return -ENOMEM; 247 248 for (i = 0; i < j; i++) 249 tmp[i] = argv[i]; 250 251 parse_options(j, tmp, event_options, NULL, PARSE_OPT_KEEP_UNKNOWN); 252 if (!event) { 253 argv[j++] = STRDUP_FAIL_EXIT("-e"); 254 argv[j++] = STRDUP_FAIL_EXIT("cycles"); 255 *argc += 2; 256 } 257 258 free(tmp); 259 return 0; 260 261EXIT: 262 free(tmp); 263 return ret; 264}