Linux kernel mirror (for testing)
git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel
os
linux
1/* SPDX-License-Identifier: GPL-2.0 */
2#if !defined(_TRACE_KVM_H) || defined(TRACE_HEADER_MULTI_READ)
3#define _TRACE_KVM_H
4
5#include <linux/tracepoint.h>
6#include <asm/vmx.h>
7#include <asm/svm.h>
8#include <asm/clocksource.h>
9#include <asm/pvclock-abi.h>
10
11#undef TRACE_SYSTEM
12#define TRACE_SYSTEM kvm
13
14#ifdef CREATE_TRACE_POINTS
15#define tracing_kvm_rip_read(vcpu) ({ \
16 typeof(vcpu) __vcpu = vcpu; \
17 __vcpu->arch.guest_state_protected ? 0 : kvm_rip_read(__vcpu); \
18 })
19#endif
20
21/*
22 * Tracepoint for guest mode entry.
23 */
24TRACE_EVENT(kvm_entry,
25 TP_PROTO(struct kvm_vcpu *vcpu, bool force_immediate_exit),
26 TP_ARGS(vcpu, force_immediate_exit),
27
28 TP_STRUCT__entry(
29 __field( unsigned int, vcpu_id )
30 __field( unsigned long, rip )
31 __field( bool, immediate_exit )
32 __field( u32, intr_info )
33 __field( u32, error_code )
34 ),
35
36 TP_fast_assign(
37 __entry->vcpu_id = vcpu->vcpu_id;
38 __entry->rip = tracing_kvm_rip_read(vcpu);
39 __entry->immediate_exit = force_immediate_exit;
40
41 kvm_x86_call(get_entry_info)(vcpu, &__entry->intr_info,
42 &__entry->error_code);
43 ),
44
45 TP_printk("vcpu %u, rip 0x%lx intr_info 0x%08x error_code 0x%08x%s",
46 __entry->vcpu_id, __entry->rip,
47 __entry->intr_info, __entry->error_code,
48 __entry->immediate_exit ? "[immediate exit]" : "")
49);
50
51/*
52 * Tracepoint for hypercall.
53 */
54TRACE_EVENT(kvm_hypercall,
55 TP_PROTO(unsigned long nr, unsigned long a0, unsigned long a1,
56 unsigned long a2, unsigned long a3),
57 TP_ARGS(nr, a0, a1, a2, a3),
58
59 TP_STRUCT__entry(
60 __field( unsigned long, nr )
61 __field( unsigned long, a0 )
62 __field( unsigned long, a1 )
63 __field( unsigned long, a2 )
64 __field( unsigned long, a3 )
65 ),
66
67 TP_fast_assign(
68 __entry->nr = nr;
69 __entry->a0 = a0;
70 __entry->a1 = a1;
71 __entry->a2 = a2;
72 __entry->a3 = a3;
73 ),
74
75 TP_printk("nr 0x%lx a0 0x%lx a1 0x%lx a2 0x%lx a3 0x%lx",
76 __entry->nr, __entry->a0, __entry->a1, __entry->a2,
77 __entry->a3)
78);
79
80/*
81 * Tracepoint for hypercall.
82 */
83TRACE_EVENT(kvm_hv_hypercall,
84 TP_PROTO(__u16 code, bool fast, __u16 var_cnt, __u16 rep_cnt,
85 __u16 rep_idx, __u64 ingpa, __u64 outgpa),
86 TP_ARGS(code, fast, var_cnt, rep_cnt, rep_idx, ingpa, outgpa),
87
88 TP_STRUCT__entry(
89 __field( __u16, rep_cnt )
90 __field( __u16, rep_idx )
91 __field( __u64, ingpa )
92 __field( __u64, outgpa )
93 __field( __u16, code )
94 __field( __u16, var_cnt )
95 __field( bool, fast )
96 ),
97
98 TP_fast_assign(
99 __entry->rep_cnt = rep_cnt;
100 __entry->rep_idx = rep_idx;
101 __entry->ingpa = ingpa;
102 __entry->outgpa = outgpa;
103 __entry->code = code;
104 __entry->var_cnt = var_cnt;
105 __entry->fast = fast;
106 ),
107
108 TP_printk("code 0x%x %s var_cnt 0x%x rep_cnt 0x%x idx 0x%x in 0x%llx out 0x%llx",
109 __entry->code, __entry->fast ? "fast" : "slow",
110 __entry->var_cnt, __entry->rep_cnt, __entry->rep_idx,
111 __entry->ingpa, __entry->outgpa)
112);
113
114TRACE_EVENT(kvm_hv_hypercall_done,
115 TP_PROTO(u64 result),
116 TP_ARGS(result),
117
118 TP_STRUCT__entry(
119 __field(__u64, result)
120 ),
121
122 TP_fast_assign(
123 __entry->result = result;
124 ),
125
126 TP_printk("result 0x%llx", __entry->result)
127);
128
129/*
130 * Tracepoint for Xen hypercall.
131 */
132TRACE_EVENT(kvm_xen_hypercall,
133 TP_PROTO(u8 cpl, unsigned long nr,
134 unsigned long a0, unsigned long a1, unsigned long a2,
135 unsigned long a3, unsigned long a4, unsigned long a5),
136 TP_ARGS(cpl, nr, a0, a1, a2, a3, a4, a5),
137
138 TP_STRUCT__entry(
139 __field(u8, cpl)
140 __field(unsigned long, nr)
141 __field(unsigned long, a0)
142 __field(unsigned long, a1)
143 __field(unsigned long, a2)
144 __field(unsigned long, a3)
145 __field(unsigned long, a4)
146 __field(unsigned long, a5)
147 ),
148
149 TP_fast_assign(
150 __entry->cpl = cpl;
151 __entry->nr = nr;
152 __entry->a0 = a0;
153 __entry->a1 = a1;
154 __entry->a2 = a2;
155 __entry->a3 = a3;
156 __entry->a4 = a4;
157 __entry->a4 = a5;
158 ),
159
160 TP_printk("cpl %d nr 0x%lx a0 0x%lx a1 0x%lx a2 0x%lx a3 0x%lx a4 0x%lx a5 %lx",
161 __entry->cpl, __entry->nr,
162 __entry->a0, __entry->a1, __entry->a2,
163 __entry->a3, __entry->a4, __entry->a5)
164);
165
166
167
168/*
169 * Tracepoint for PIO.
170 */
171
172#define KVM_PIO_IN 0
173#define KVM_PIO_OUT 1
174
175TRACE_EVENT(kvm_pio,
176 TP_PROTO(unsigned int rw, unsigned int port, unsigned int size,
177 unsigned int count, const void *data),
178 TP_ARGS(rw, port, size, count, data),
179
180 TP_STRUCT__entry(
181 __field( unsigned int, rw )
182 __field( unsigned int, port )
183 __field( unsigned int, size )
184 __field( unsigned int, count )
185 __field( unsigned int, val )
186 ),
187
188 TP_fast_assign(
189 __entry->rw = rw;
190 __entry->port = port;
191 __entry->size = size;
192 __entry->count = count;
193 if (size == 1)
194 __entry->val = *(unsigned char *)data;
195 else if (size == 2)
196 __entry->val = *(unsigned short *)data;
197 else
198 __entry->val = *(unsigned int *)data;
199 ),
200
201 TP_printk("pio_%s at 0x%x size %d count %d val 0x%x %s",
202 __entry->rw ? "write" : "read",
203 __entry->port, __entry->size, __entry->count, __entry->val,
204 __entry->count > 1 ? "(...)" : "")
205);
206
207/*
208 * Tracepoint for fast mmio.
209 */
210TRACE_EVENT(kvm_fast_mmio,
211 TP_PROTO(u64 gpa),
212 TP_ARGS(gpa),
213
214 TP_STRUCT__entry(
215 __field(u64, gpa)
216 ),
217
218 TP_fast_assign(
219 __entry->gpa = gpa;
220 ),
221
222 TP_printk("fast mmio at gpa 0x%llx", __entry->gpa)
223);
224
225/*
226 * Tracepoint for cpuid.
227 */
228TRACE_EVENT(kvm_cpuid,
229 TP_PROTO(unsigned int function, unsigned int index, unsigned long rax,
230 unsigned long rbx, unsigned long rcx, unsigned long rdx,
231 bool found, bool used_max_basic),
232 TP_ARGS(function, index, rax, rbx, rcx, rdx, found, used_max_basic),
233
234 TP_STRUCT__entry(
235 __field( unsigned int, function )
236 __field( unsigned int, index )
237 __field( unsigned long, rax )
238 __field( unsigned long, rbx )
239 __field( unsigned long, rcx )
240 __field( unsigned long, rdx )
241 __field( bool, found )
242 __field( bool, used_max_basic )
243 ),
244
245 TP_fast_assign(
246 __entry->function = function;
247 __entry->index = index;
248 __entry->rax = rax;
249 __entry->rbx = rbx;
250 __entry->rcx = rcx;
251 __entry->rdx = rdx;
252 __entry->found = found;
253 __entry->used_max_basic = used_max_basic;
254 ),
255
256 TP_printk("func %x idx %x rax %lx rbx %lx rcx %lx rdx %lx, cpuid entry %s%s",
257 __entry->function, __entry->index, __entry->rax,
258 __entry->rbx, __entry->rcx, __entry->rdx,
259 __entry->found ? "found" : "not found",
260 __entry->used_max_basic ? ", used max basic" : "")
261);
262
263#define kvm_deliver_mode \
264 {0x0, "Fixed"}, \
265 {0x1, "LowPrio"}, \
266 {0x2, "SMI"}, \
267 {0x3, "Res3"}, \
268 {0x4, "NMI"}, \
269 {0x5, "INIT"}, \
270 {0x6, "SIPI"}, \
271 {0x7, "ExtINT"}
272
273#ifdef CONFIG_KVM_IOAPIC
274TRACE_EVENT(kvm_ioapic_set_irq,
275 TP_PROTO(__u64 e, int pin, bool coalesced),
276 TP_ARGS(e, pin, coalesced),
277
278 TP_STRUCT__entry(
279 __field( __u64, e )
280 __field( int, pin )
281 __field( bool, coalesced )
282 ),
283
284 TP_fast_assign(
285 __entry->e = e;
286 __entry->pin = pin;
287 __entry->coalesced = coalesced;
288 ),
289
290 TP_printk("pin %u dst %x vec %u (%s|%s|%s%s)%s",
291 __entry->pin, (u8)(__entry->e >> 56), (u8)__entry->e,
292 __print_symbolic((__entry->e >> 8 & 0x7), kvm_deliver_mode),
293 (__entry->e & (1<<11)) ? "logical" : "physical",
294 (__entry->e & (1<<15)) ? "level" : "edge",
295 (__entry->e & (1<<16)) ? "|masked" : "",
296 __entry->coalesced ? " (coalesced)" : "")
297);
298
299TRACE_EVENT(kvm_ioapic_delayed_eoi_inj,
300 TP_PROTO(__u64 e),
301 TP_ARGS(e),
302
303 TP_STRUCT__entry(
304 __field( __u64, e )
305 ),
306
307 TP_fast_assign(
308 __entry->e = e;
309 ),
310
311 TP_printk("dst %x vec %u (%s|%s|%s%s)",
312 (u8)(__entry->e >> 56), (u8)__entry->e,
313 __print_symbolic((__entry->e >> 8 & 0x7), kvm_deliver_mode),
314 (__entry->e & (1<<11)) ? "logical" : "physical",
315 (__entry->e & (1<<15)) ? "level" : "edge",
316 (__entry->e & (1<<16)) ? "|masked" : "")
317);
318#endif
319
320TRACE_EVENT(kvm_msi_set_irq,
321 TP_PROTO(__u64 address, __u64 data),
322 TP_ARGS(address, data),
323
324 TP_STRUCT__entry(
325 __field( __u64, address )
326 __field( __u64, data )
327 ),
328
329 TP_fast_assign(
330 __entry->address = address;
331 __entry->data = data;
332 ),
333
334 TP_printk("dst %llx vec %u (%s|%s|%s%s)",
335 (u8)(__entry->address >> 12) | ((__entry->address >> 32) & 0xffffff00),
336 (u8)__entry->data,
337 __print_symbolic((__entry->data >> 8 & 0x7), kvm_deliver_mode),
338 (__entry->address & (1<<2)) ? "logical" : "physical",
339 (__entry->data & (1<<15)) ? "level" : "edge",
340 (__entry->address & (1<<3)) ? "|rh" : "")
341);
342
343#define AREG(x) { APIC_##x, "APIC_" #x }
344
345#define kvm_trace_symbol_apic \
346 AREG(ID), AREG(LVR), AREG(TASKPRI), AREG(ARBPRI), AREG(PROCPRI), \
347 AREG(EOI), AREG(RRR), AREG(LDR), AREG(DFR), AREG(SPIV), AREG(ISR), \
348 AREG(TMR), AREG(IRR), AREG(ESR), AREG(ICR), AREG(ICR2), AREG(LVTT), \
349 AREG(LVTTHMR), AREG(LVTPC), AREG(LVT0), AREG(LVT1), AREG(LVTERR), \
350 AREG(TMICT), AREG(TMCCT), AREG(TDCR), AREG(SELF_IPI), AREG(EFEAT), \
351 AREG(ECTRL)
352/*
353 * Tracepoint for apic access.
354 */
355TRACE_EVENT(kvm_apic,
356 TP_PROTO(unsigned int rw, unsigned int reg, u64 val),
357 TP_ARGS(rw, reg, val),
358
359 TP_STRUCT__entry(
360 __field( unsigned int, rw )
361 __field( unsigned int, reg )
362 __field( u64, val )
363 ),
364
365 TP_fast_assign(
366 __entry->rw = rw;
367 __entry->reg = reg;
368 __entry->val = val;
369 ),
370
371 TP_printk("apic_%s %s = 0x%llx",
372 __entry->rw ? "write" : "read",
373 __print_symbolic(__entry->reg, kvm_trace_symbol_apic),
374 __entry->val)
375);
376
377#define trace_kvm_apic_read(reg, val) trace_kvm_apic(0, reg, val)
378#define trace_kvm_apic_write(reg, val) trace_kvm_apic(1, reg, val)
379
380#define KVM_ISA_VMX 1
381#define KVM_ISA_SVM 2
382
383#define kvm_print_exit_reason(exit_reason, isa) \
384 (isa == KVM_ISA_VMX) ? \
385 __print_symbolic(exit_reason & 0xffff, VMX_EXIT_REASONS) : \
386 __print_symbolic(exit_reason, SVM_EXIT_REASONS), \
387 (isa == KVM_ISA_VMX && exit_reason & ~0xffff) ? " " : "", \
388 (isa == KVM_ISA_VMX) ? \
389 __print_flags(exit_reason & ~0xffff, " ", VMX_EXIT_REASON_FLAGS) : ""
390
391#define TRACE_EVENT_KVM_EXIT(name) \
392TRACE_EVENT(name, \
393 TP_PROTO(struct kvm_vcpu *vcpu, u32 isa), \
394 TP_ARGS(vcpu, isa), \
395 \
396 TP_STRUCT__entry( \
397 __field( unsigned int, exit_reason ) \
398 __field( unsigned long, guest_rip ) \
399 __field( u32, isa ) \
400 __field( u64, info1 ) \
401 __field( u64, info2 ) \
402 __field( u32, intr_info ) \
403 __field( u32, error_code ) \
404 __field( unsigned int, vcpu_id ) \
405 __field( u64, requests ) \
406 ), \
407 \
408 TP_fast_assign( \
409 __entry->guest_rip = tracing_kvm_rip_read(vcpu); \
410 __entry->isa = isa; \
411 __entry->vcpu_id = vcpu->vcpu_id; \
412 __entry->requests = READ_ONCE(vcpu->requests); \
413 kvm_x86_call(get_exit_info)(vcpu, \
414 &__entry->exit_reason, \
415 &__entry->info1, \
416 &__entry->info2, \
417 &__entry->intr_info, \
418 &__entry->error_code); \
419 ), \
420 \
421 TP_printk("vcpu %u reason %s%s%s rip 0x%lx info1 0x%016llx " \
422 "info2 0x%016llx intr_info 0x%08x error_code 0x%08x " \
423 "requests 0x%016llx", \
424 __entry->vcpu_id, \
425 kvm_print_exit_reason(__entry->exit_reason, __entry->isa), \
426 __entry->guest_rip, __entry->info1, __entry->info2, \
427 __entry->intr_info, __entry->error_code, \
428 __entry->requests) \
429)
430
431/*
432 * Tracepoint for kvm guest exit:
433 */
434TRACE_EVENT_KVM_EXIT(kvm_exit);
435
436/*
437 * Tracepoint for kvm interrupt injection:
438 */
439TRACE_EVENT(kvm_inj_virq,
440 TP_PROTO(unsigned int vector, bool soft, bool reinjected),
441 TP_ARGS(vector, soft, reinjected),
442
443 TP_STRUCT__entry(
444 __field( unsigned int, vector )
445 __field( bool, soft )
446 __field( bool, reinjected )
447 ),
448
449 TP_fast_assign(
450 __entry->vector = vector;
451 __entry->soft = soft;
452 __entry->reinjected = reinjected;
453 ),
454
455 TP_printk("%s 0x%x%s",
456 __entry->soft ? "Soft/INTn" : "IRQ", __entry->vector,
457 __entry->reinjected ? " [reinjected]" : "")
458);
459
460#define EXS(x) { x##_VECTOR, "#" #x }
461
462#define kvm_trace_sym_exc \
463 EXS(DE), EXS(DB), EXS(BP), EXS(OF), EXS(BR), EXS(UD), EXS(NM), \
464 EXS(DF), EXS(TS), EXS(NP), EXS(SS), EXS(GP), EXS(PF), \
465 EXS(MF), EXS(AC), EXS(MC)
466
467/*
468 * Tracepoint for kvm interrupt injection:
469 */
470TRACE_EVENT(kvm_inj_exception,
471 TP_PROTO(unsigned exception, bool has_error, unsigned error_code,
472 bool reinjected),
473 TP_ARGS(exception, has_error, error_code, reinjected),
474
475 TP_STRUCT__entry(
476 __field( u8, exception )
477 __field( u8, has_error )
478 __field( u32, error_code )
479 __field( bool, reinjected )
480 ),
481
482 TP_fast_assign(
483 __entry->exception = exception;
484 __entry->has_error = has_error;
485 __entry->error_code = error_code;
486 __entry->reinjected = reinjected;
487 ),
488
489 TP_printk("%s%s%s%s%s",
490 __print_symbolic(__entry->exception, kvm_trace_sym_exc),
491 !__entry->has_error ? "" : " (",
492 !__entry->has_error ? "" : __print_symbolic(__entry->error_code, { }),
493 !__entry->has_error ? "" : ")",
494 __entry->reinjected ? " [reinjected]" : "")
495);
496
497/*
498 * Tracepoint for page fault.
499 */
500TRACE_EVENT(kvm_page_fault,
501 TP_PROTO(struct kvm_vcpu *vcpu, u64 fault_address, u64 error_code),
502 TP_ARGS(vcpu, fault_address, error_code),
503
504 TP_STRUCT__entry(
505 __field( unsigned int, vcpu_id )
506 __field( unsigned long, guest_rip )
507 __field( u64, fault_address )
508 __field( u64, error_code )
509 ),
510
511 TP_fast_assign(
512 __entry->vcpu_id = vcpu->vcpu_id;
513 __entry->guest_rip = tracing_kvm_rip_read(vcpu);
514 __entry->fault_address = fault_address;
515 __entry->error_code = error_code;
516 ),
517
518 TP_printk("vcpu %u rip 0x%lx address 0x%016llx error_code 0x%llx",
519 __entry->vcpu_id, __entry->guest_rip,
520 __entry->fault_address, __entry->error_code)
521);
522
523/*
524 * Tracepoint for guest MSR access.
525 */
526TRACE_EVENT(kvm_msr,
527 TP_PROTO(unsigned write, u32 ecx, u64 data, bool exception),
528 TP_ARGS(write, ecx, data, exception),
529
530 TP_STRUCT__entry(
531 __field( unsigned, write )
532 __field( u32, ecx )
533 __field( u64, data )
534 __field( u8, exception )
535 ),
536
537 TP_fast_assign(
538 __entry->write = write;
539 __entry->ecx = ecx;
540 __entry->data = data;
541 __entry->exception = exception;
542 ),
543
544 TP_printk("msr_%s %x = 0x%llx%s",
545 __entry->write ? "write" : "read",
546 __entry->ecx, __entry->data,
547 __entry->exception ? " (#GP)" : "")
548);
549
550#define trace_kvm_msr_read(ecx, data) trace_kvm_msr(0, ecx, data, false)
551#define trace_kvm_msr_write(ecx, data) trace_kvm_msr(1, ecx, data, false)
552#define trace_kvm_msr_read_ex(ecx) trace_kvm_msr(0, ecx, 0, true)
553#define trace_kvm_msr_write_ex(ecx, data) trace_kvm_msr(1, ecx, data, true)
554
555/*
556 * Tracepoint for guest CR access.
557 */
558TRACE_EVENT(kvm_cr,
559 TP_PROTO(unsigned int rw, unsigned int cr, unsigned long val),
560 TP_ARGS(rw, cr, val),
561
562 TP_STRUCT__entry(
563 __field( unsigned int, rw )
564 __field( unsigned int, cr )
565 __field( unsigned long, val )
566 ),
567
568 TP_fast_assign(
569 __entry->rw = rw;
570 __entry->cr = cr;
571 __entry->val = val;
572 ),
573
574 TP_printk("cr_%s %x = 0x%lx",
575 __entry->rw ? "write" : "read",
576 __entry->cr, __entry->val)
577);
578
579#define trace_kvm_cr_read(cr, val) trace_kvm_cr(0, cr, val)
580#define trace_kvm_cr_write(cr, val) trace_kvm_cr(1, cr, val)
581
582TRACE_EVENT(kvm_pic_set_irq,
583 TP_PROTO(__u8 chip, __u8 pin, __u8 elcr, __u8 imr, bool coalesced),
584 TP_ARGS(chip, pin, elcr, imr, coalesced),
585
586 TP_STRUCT__entry(
587 __field( __u8, chip )
588 __field( __u8, pin )
589 __field( __u8, elcr )
590 __field( __u8, imr )
591 __field( bool, coalesced )
592 ),
593
594 TP_fast_assign(
595 __entry->chip = chip;
596 __entry->pin = pin;
597 __entry->elcr = elcr;
598 __entry->imr = imr;
599 __entry->coalesced = coalesced;
600 ),
601
602 TP_printk("chip %u pin %u (%s%s)%s",
603 __entry->chip, __entry->pin,
604 (__entry->elcr & (1 << __entry->pin)) ? "level":"edge",
605 (__entry->imr & (1 << __entry->pin)) ? "|masked":"",
606 __entry->coalesced ? " (coalesced)" : "")
607);
608
609#define kvm_apic_dst_shorthand \
610 {0x0, "dst"}, \
611 {0x1, "self"}, \
612 {0x2, "all"}, \
613 {0x3, "all-but-self"}
614
615TRACE_EVENT(kvm_apic_ipi,
616 TP_PROTO(__u32 icr_low, __u32 dest_id),
617 TP_ARGS(icr_low, dest_id),
618
619 TP_STRUCT__entry(
620 __field( __u32, icr_low )
621 __field( __u32, dest_id )
622 ),
623
624 TP_fast_assign(
625 __entry->icr_low = icr_low;
626 __entry->dest_id = dest_id;
627 ),
628
629 TP_printk("dst %x vec %u (%s|%s|%s|%s|%s)",
630 __entry->dest_id, (u8)__entry->icr_low,
631 __print_symbolic((__entry->icr_low >> 8 & 0x7),
632 kvm_deliver_mode),
633 (__entry->icr_low & (1<<11)) ? "logical" : "physical",
634 (__entry->icr_low & (1<<14)) ? "assert" : "de-assert",
635 (__entry->icr_low & (1<<15)) ? "level" : "edge",
636 __print_symbolic((__entry->icr_low >> 18 & 0x3),
637 kvm_apic_dst_shorthand))
638);
639
640TRACE_EVENT(kvm_apic_accept_irq,
641 TP_PROTO(__u32 apicid, __u16 dm, __u16 tm, __u8 vec),
642 TP_ARGS(apicid, dm, tm, vec),
643
644 TP_STRUCT__entry(
645 __field( __u32, apicid )
646 __field( __u16, dm )
647 __field( __u16, tm )
648 __field( __u8, vec )
649 ),
650
651 TP_fast_assign(
652 __entry->apicid = apicid;
653 __entry->dm = dm;
654 __entry->tm = tm;
655 __entry->vec = vec;
656 ),
657
658 TP_printk("apicid %x vec %u (%s|%s)",
659 __entry->apicid, __entry->vec,
660 __print_symbolic((__entry->dm >> 8 & 0x7), kvm_deliver_mode),
661 __entry->tm ? "level" : "edge")
662);
663
664TRACE_EVENT(kvm_eoi,
665 TP_PROTO(struct kvm_lapic *apic, int vector),
666 TP_ARGS(apic, vector),
667
668 TP_STRUCT__entry(
669 __field( __u32, apicid )
670 __field( int, vector )
671 ),
672
673 TP_fast_assign(
674 __entry->apicid = apic->vcpu->vcpu_id;
675 __entry->vector = vector;
676 ),
677
678 TP_printk("apicid %x vector %d", __entry->apicid, __entry->vector)
679);
680
681TRACE_EVENT(kvm_pv_eoi,
682 TP_PROTO(struct kvm_lapic *apic, int vector),
683 TP_ARGS(apic, vector),
684
685 TP_STRUCT__entry(
686 __field( __u32, apicid )
687 __field( int, vector )
688 ),
689
690 TP_fast_assign(
691 __entry->apicid = apic->vcpu->vcpu_id;
692 __entry->vector = vector;
693 ),
694
695 TP_printk("apicid %x vector %d", __entry->apicid, __entry->vector)
696);
697
698/*
699 * Tracepoint for nested VMRUN
700 */
701TRACE_EVENT(kvm_nested_vmenter,
702 TP_PROTO(__u64 rip, __u64 vmcb, __u64 nested_rip, __u32 int_ctl,
703 __u32 event_inj, bool tdp_enabled, __u64 guest_tdp_pgd,
704 __u64 guest_cr3, __u32 isa),
705 TP_ARGS(rip, vmcb, nested_rip, int_ctl, event_inj, tdp_enabled,
706 guest_tdp_pgd, guest_cr3, isa),
707
708 TP_STRUCT__entry(
709 __field( __u64, rip )
710 __field( __u64, vmcb )
711 __field( __u64, nested_rip )
712 __field( __u32, int_ctl )
713 __field( __u32, event_inj )
714 __field( bool, tdp_enabled )
715 __field( __u64, guest_pgd )
716 __field( __u32, isa )
717 ),
718
719 TP_fast_assign(
720 __entry->rip = rip;
721 __entry->vmcb = vmcb;
722 __entry->nested_rip = nested_rip;
723 __entry->int_ctl = int_ctl;
724 __entry->event_inj = event_inj;
725 __entry->tdp_enabled = tdp_enabled;
726 __entry->guest_pgd = tdp_enabled ? guest_tdp_pgd : guest_cr3;
727 __entry->isa = isa;
728 ),
729
730 TP_printk("rip: 0x%016llx %s: 0x%016llx nested_rip: 0x%016llx "
731 "int_ctl: 0x%08x event_inj: 0x%08x nested_%s=%s %s: 0x%016llx",
732 __entry->rip,
733 __entry->isa == KVM_ISA_VMX ? "vmcs" : "vmcb",
734 __entry->vmcb,
735 __entry->nested_rip,
736 __entry->int_ctl,
737 __entry->event_inj,
738 __entry->isa == KVM_ISA_VMX ? "ept" : "npt",
739 __entry->tdp_enabled ? "y" : "n",
740 !__entry->tdp_enabled ? "guest_cr3" :
741 __entry->isa == KVM_ISA_VMX ? "nested_eptp" : "nested_cr3",
742 __entry->guest_pgd)
743);
744
745TRACE_EVENT(kvm_nested_intercepts,
746 TP_PROTO(__u16 cr_read, __u16 cr_write, __u32 exceptions,
747 __u32 intercept1, __u32 intercept2, __u32 intercept3),
748 TP_ARGS(cr_read, cr_write, exceptions, intercept1,
749 intercept2, intercept3),
750
751 TP_STRUCT__entry(
752 __field( __u16, cr_read )
753 __field( __u16, cr_write )
754 __field( __u32, exceptions )
755 __field( __u32, intercept1 )
756 __field( __u32, intercept2 )
757 __field( __u32, intercept3 )
758 ),
759
760 TP_fast_assign(
761 __entry->cr_read = cr_read;
762 __entry->cr_write = cr_write;
763 __entry->exceptions = exceptions;
764 __entry->intercept1 = intercept1;
765 __entry->intercept2 = intercept2;
766 __entry->intercept3 = intercept3;
767 ),
768
769 TP_printk("cr_read: %04x cr_write: %04x excp: %08x "
770 "intercepts: %08x %08x %08x",
771 __entry->cr_read, __entry->cr_write, __entry->exceptions,
772 __entry->intercept1, __entry->intercept2, __entry->intercept3)
773);
774/*
775 * Tracepoint for #VMEXIT while nested
776 */
777TRACE_EVENT_KVM_EXIT(kvm_nested_vmexit);
778
779/*
780 * Tracepoint for #VMEXIT reinjected to the guest
781 */
782TRACE_EVENT(kvm_nested_vmexit_inject,
783 TP_PROTO(__u32 exit_code,
784 __u64 exit_info1, __u64 exit_info2,
785 __u32 exit_int_info, __u32 exit_int_info_err, __u32 isa),
786 TP_ARGS(exit_code, exit_info1, exit_info2,
787 exit_int_info, exit_int_info_err, isa),
788
789 TP_STRUCT__entry(
790 __field( __u32, exit_code )
791 __field( __u64, exit_info1 )
792 __field( __u64, exit_info2 )
793 __field( __u32, exit_int_info )
794 __field( __u32, exit_int_info_err )
795 __field( __u32, isa )
796 ),
797
798 TP_fast_assign(
799 __entry->exit_code = exit_code;
800 __entry->exit_info1 = exit_info1;
801 __entry->exit_info2 = exit_info2;
802 __entry->exit_int_info = exit_int_info;
803 __entry->exit_int_info_err = exit_int_info_err;
804 __entry->isa = isa;
805 ),
806
807 TP_printk("reason: %s%s%s ext_inf1: 0x%016llx "
808 "ext_inf2: 0x%016llx ext_int: 0x%08x ext_int_err: 0x%08x",
809 kvm_print_exit_reason(__entry->exit_code, __entry->isa),
810 __entry->exit_info1, __entry->exit_info2,
811 __entry->exit_int_info, __entry->exit_int_info_err)
812);
813
814/*
815 * Tracepoint for nested #vmexit because of interrupt pending
816 */
817TRACE_EVENT(kvm_nested_intr_vmexit,
818 TP_PROTO(__u64 rip),
819 TP_ARGS(rip),
820
821 TP_STRUCT__entry(
822 __field( __u64, rip )
823 ),
824
825 TP_fast_assign(
826 __entry->rip = rip
827 ),
828
829 TP_printk("rip: 0x%016llx", __entry->rip)
830);
831
832/*
833 * Tracepoint for nested #vmexit because of interrupt pending
834 */
835TRACE_EVENT(kvm_invlpga,
836 TP_PROTO(__u64 rip, unsigned int asid, u64 address),
837 TP_ARGS(rip, asid, address),
838
839 TP_STRUCT__entry(
840 __field( __u64, rip )
841 __field( unsigned int, asid )
842 __field( __u64, address )
843 ),
844
845 TP_fast_assign(
846 __entry->rip = rip;
847 __entry->asid = asid;
848 __entry->address = address;
849 ),
850
851 TP_printk("rip: 0x%016llx asid: %u address: 0x%016llx",
852 __entry->rip, __entry->asid, __entry->address)
853);
854
855/*
856 * Tracepoint for nested #vmexit because of interrupt pending
857 */
858TRACE_EVENT(kvm_skinit,
859 TP_PROTO(__u64 rip, __u32 slb),
860 TP_ARGS(rip, slb),
861
862 TP_STRUCT__entry(
863 __field( __u64, rip )
864 __field( __u32, slb )
865 ),
866
867 TP_fast_assign(
868 __entry->rip = rip;
869 __entry->slb = slb;
870 ),
871
872 TP_printk("rip: 0x%016llx slb: 0x%08x",
873 __entry->rip, __entry->slb)
874);
875
876#define KVM_EMUL_INSN_F_CR0_PE (1 << 0)
877#define KVM_EMUL_INSN_F_EFL_VM (1 << 1)
878#define KVM_EMUL_INSN_F_CS_D (1 << 2)
879#define KVM_EMUL_INSN_F_CS_L (1 << 3)
880
881#define kvm_trace_symbol_emul_flags \
882 { 0, "real" }, \
883 { KVM_EMUL_INSN_F_CR0_PE \
884 | KVM_EMUL_INSN_F_EFL_VM, "vm16" }, \
885 { KVM_EMUL_INSN_F_CR0_PE, "prot16" }, \
886 { KVM_EMUL_INSN_F_CR0_PE \
887 | KVM_EMUL_INSN_F_CS_D, "prot32" }, \
888 { KVM_EMUL_INSN_F_CR0_PE \
889 | KVM_EMUL_INSN_F_CS_L, "prot64" }
890
891#define kei_decode_mode(mode) ({ \
892 u8 flags = 0xff; \
893 switch (mode) { \
894 case X86EMUL_MODE_REAL: \
895 flags = 0; \
896 break; \
897 case X86EMUL_MODE_VM86: \
898 flags = KVM_EMUL_INSN_F_EFL_VM; \
899 break; \
900 case X86EMUL_MODE_PROT16: \
901 flags = KVM_EMUL_INSN_F_CR0_PE; \
902 break; \
903 case X86EMUL_MODE_PROT32: \
904 flags = KVM_EMUL_INSN_F_CR0_PE \
905 | KVM_EMUL_INSN_F_CS_D; \
906 break; \
907 case X86EMUL_MODE_PROT64: \
908 flags = KVM_EMUL_INSN_F_CR0_PE \
909 | KVM_EMUL_INSN_F_CS_L; \
910 break; \
911 } \
912 flags; \
913 })
914
915TRACE_EVENT(kvm_emulate_insn,
916 TP_PROTO(struct kvm_vcpu *vcpu, __u8 failed),
917 TP_ARGS(vcpu, failed),
918
919 TP_STRUCT__entry(
920 __field( __u64, rip )
921 __field( __u32, csbase )
922 __field( __u8, len )
923 __array( __u8, insn, X86_MAX_INSTRUCTION_LENGTH )
924 __field( __u8, flags )
925 __field( __u8, failed )
926 ),
927
928 TP_fast_assign(
929 __entry->csbase = kvm_x86_call(get_segment_base)(vcpu,
930 VCPU_SREG_CS);
931 __entry->len = vcpu->arch.emulate_ctxt->fetch.ptr
932 - vcpu->arch.emulate_ctxt->fetch.data;
933 __entry->rip = vcpu->arch.emulate_ctxt->_eip - __entry->len;
934 memcpy(__entry->insn,
935 vcpu->arch.emulate_ctxt->fetch.data,
936 X86_MAX_INSTRUCTION_LENGTH);
937 __entry->flags = kei_decode_mode(vcpu->arch.emulate_ctxt->mode);
938 __entry->failed = failed;
939 ),
940
941 TP_printk("%x:%llx:%s (%s)%s",
942 __entry->csbase, __entry->rip,
943 __print_hex(__entry->insn, __entry->len),
944 __print_symbolic(__entry->flags,
945 kvm_trace_symbol_emul_flags),
946 __entry->failed ? " failed" : ""
947 )
948 );
949
950#define trace_kvm_emulate_insn_start(vcpu) trace_kvm_emulate_insn(vcpu, 0)
951#define trace_kvm_emulate_insn_failed(vcpu) trace_kvm_emulate_insn(vcpu, 1)
952
953TRACE_EVENT(
954 vcpu_match_mmio,
955 TP_PROTO(gva_t gva, gpa_t gpa, bool write, bool gpa_match),
956 TP_ARGS(gva, gpa, write, gpa_match),
957
958 TP_STRUCT__entry(
959 __field(gva_t, gva)
960 __field(gpa_t, gpa)
961 __field(bool, write)
962 __field(bool, gpa_match)
963 ),
964
965 TP_fast_assign(
966 __entry->gva = gva;
967 __entry->gpa = gpa;
968 __entry->write = write;
969 __entry->gpa_match = gpa_match
970 ),
971
972 TP_printk("gva %#lx gpa %#llx %s %s", __entry->gva, __entry->gpa,
973 __entry->write ? "Write" : "Read",
974 __entry->gpa_match ? "GPA" : "GVA")
975);
976
977TRACE_EVENT(kvm_write_tsc_offset,
978 TP_PROTO(unsigned int vcpu_id, __u64 previous_tsc_offset,
979 __u64 next_tsc_offset),
980 TP_ARGS(vcpu_id, previous_tsc_offset, next_tsc_offset),
981
982 TP_STRUCT__entry(
983 __field( unsigned int, vcpu_id )
984 __field( __u64, previous_tsc_offset )
985 __field( __u64, next_tsc_offset )
986 ),
987
988 TP_fast_assign(
989 __entry->vcpu_id = vcpu_id;
990 __entry->previous_tsc_offset = previous_tsc_offset;
991 __entry->next_tsc_offset = next_tsc_offset;
992 ),
993
994 TP_printk("vcpu=%u prev=%llu next=%llu", __entry->vcpu_id,
995 __entry->previous_tsc_offset, __entry->next_tsc_offset)
996);
997
998#ifdef CONFIG_X86_64
999
1000#define host_clocks \
1001 {VDSO_CLOCKMODE_NONE, "none"}, \
1002 {VDSO_CLOCKMODE_TSC, "tsc"} \
1003
1004TRACE_EVENT(kvm_update_master_clock,
1005 TP_PROTO(bool use_master_clock, unsigned int host_clock, bool offset_matched),
1006 TP_ARGS(use_master_clock, host_clock, offset_matched),
1007
1008 TP_STRUCT__entry(
1009 __field( bool, use_master_clock )
1010 __field( unsigned int, host_clock )
1011 __field( bool, offset_matched )
1012 ),
1013
1014 TP_fast_assign(
1015 __entry->use_master_clock = use_master_clock;
1016 __entry->host_clock = host_clock;
1017 __entry->offset_matched = offset_matched;
1018 ),
1019
1020 TP_printk("masterclock %d hostclock %s offsetmatched %u",
1021 __entry->use_master_clock,
1022 __print_symbolic(__entry->host_clock, host_clocks),
1023 __entry->offset_matched)
1024);
1025
1026TRACE_EVENT(kvm_track_tsc,
1027 TP_PROTO(unsigned int vcpu_id, unsigned int nr_matched,
1028 unsigned int online_vcpus, bool use_master_clock,
1029 unsigned int host_clock),
1030 TP_ARGS(vcpu_id, nr_matched, online_vcpus, use_master_clock,
1031 host_clock),
1032
1033 TP_STRUCT__entry(
1034 __field( unsigned int, vcpu_id )
1035 __field( unsigned int, nr_vcpus_matched_tsc )
1036 __field( unsigned int, online_vcpus )
1037 __field( bool, use_master_clock )
1038 __field( unsigned int, host_clock )
1039 ),
1040
1041 TP_fast_assign(
1042 __entry->vcpu_id = vcpu_id;
1043 __entry->nr_vcpus_matched_tsc = nr_matched;
1044 __entry->online_vcpus = online_vcpus;
1045 __entry->use_master_clock = use_master_clock;
1046 __entry->host_clock = host_clock;
1047 ),
1048
1049 TP_printk("vcpu_id %u masterclock %u offsetmatched %u nr_online %u"
1050 " hostclock %s",
1051 __entry->vcpu_id, __entry->use_master_clock,
1052 __entry->nr_vcpus_matched_tsc, __entry->online_vcpus,
1053 __print_symbolic(__entry->host_clock, host_clocks))
1054);
1055
1056#endif /* CONFIG_X86_64 */
1057
1058/*
1059 * Tracepoint for PML full VMEXIT.
1060 */
1061TRACE_EVENT(kvm_pml_full,
1062 TP_PROTO(unsigned int vcpu_id),
1063 TP_ARGS(vcpu_id),
1064
1065 TP_STRUCT__entry(
1066 __field( unsigned int, vcpu_id )
1067 ),
1068
1069 TP_fast_assign(
1070 __entry->vcpu_id = vcpu_id;
1071 ),
1072
1073 TP_printk("vcpu %d: PML full", __entry->vcpu_id)
1074);
1075
1076TRACE_EVENT(kvm_ple_window_update,
1077 TP_PROTO(unsigned int vcpu_id, unsigned int new, unsigned int old),
1078 TP_ARGS(vcpu_id, new, old),
1079
1080 TP_STRUCT__entry(
1081 __field( unsigned int, vcpu_id )
1082 __field( unsigned int, new )
1083 __field( unsigned int, old )
1084 ),
1085
1086 TP_fast_assign(
1087 __entry->vcpu_id = vcpu_id;
1088 __entry->new = new;
1089 __entry->old = old;
1090 ),
1091
1092 TP_printk("vcpu %u old %u new %u (%s)",
1093 __entry->vcpu_id, __entry->old, __entry->new,
1094 __entry->old < __entry->new ? "growed" : "shrinked")
1095);
1096
1097TRACE_EVENT(kvm_pvclock_update,
1098 TP_PROTO(unsigned int vcpu_id, struct pvclock_vcpu_time_info *pvclock),
1099 TP_ARGS(vcpu_id, pvclock),
1100
1101 TP_STRUCT__entry(
1102 __field( unsigned int, vcpu_id )
1103 __field( __u32, version )
1104 __field( __u64, tsc_timestamp )
1105 __field( __u64, system_time )
1106 __field( __u32, tsc_to_system_mul )
1107 __field( __s8, tsc_shift )
1108 __field( __u8, flags )
1109 ),
1110
1111 TP_fast_assign(
1112 __entry->vcpu_id = vcpu_id;
1113 __entry->version = pvclock->version;
1114 __entry->tsc_timestamp = pvclock->tsc_timestamp;
1115 __entry->system_time = pvclock->system_time;
1116 __entry->tsc_to_system_mul = pvclock->tsc_to_system_mul;
1117 __entry->tsc_shift = pvclock->tsc_shift;
1118 __entry->flags = pvclock->flags;
1119 ),
1120
1121 TP_printk("vcpu_id %u, pvclock { version %u, tsc_timestamp 0x%llx, "
1122 "system_time 0x%llx, tsc_to_system_mul 0x%x, tsc_shift %d, "
1123 "flags 0x%x }",
1124 __entry->vcpu_id,
1125 __entry->version,
1126 __entry->tsc_timestamp,
1127 __entry->system_time,
1128 __entry->tsc_to_system_mul,
1129 __entry->tsc_shift,
1130 __entry->flags)
1131);
1132
1133TRACE_EVENT(kvm_wait_lapic_expire,
1134 TP_PROTO(unsigned int vcpu_id, s64 delta),
1135 TP_ARGS(vcpu_id, delta),
1136
1137 TP_STRUCT__entry(
1138 __field( unsigned int, vcpu_id )
1139 __field( s64, delta )
1140 ),
1141
1142 TP_fast_assign(
1143 __entry->vcpu_id = vcpu_id;
1144 __entry->delta = delta;
1145 ),
1146
1147 TP_printk("vcpu %u: delta %lld (%s)",
1148 __entry->vcpu_id,
1149 __entry->delta,
1150 __entry->delta < 0 ? "early" : "late")
1151);
1152
1153TRACE_EVENT(kvm_smm_transition,
1154 TP_PROTO(unsigned int vcpu_id, u64 smbase, bool entering),
1155 TP_ARGS(vcpu_id, smbase, entering),
1156
1157 TP_STRUCT__entry(
1158 __field( unsigned int, vcpu_id )
1159 __field( u64, smbase )
1160 __field( bool, entering )
1161 ),
1162
1163 TP_fast_assign(
1164 __entry->vcpu_id = vcpu_id;
1165 __entry->smbase = smbase;
1166 __entry->entering = entering;
1167 ),
1168
1169 TP_printk("vcpu %u: %s SMM, smbase 0x%llx",
1170 __entry->vcpu_id,
1171 __entry->entering ? "entering" : "leaving",
1172 __entry->smbase)
1173);
1174
1175/*
1176 * Tracepoint for VT-d posted-interrupts and AMD-Vi Guest Virtual APIC.
1177 */
1178TRACE_EVENT(kvm_pi_irte_update,
1179 TP_PROTO(unsigned int host_irq, struct kvm_vcpu *vcpu,
1180 unsigned int gsi, unsigned int gvec, bool set),
1181 TP_ARGS(host_irq, vcpu, gsi, gvec, set),
1182
1183 TP_STRUCT__entry(
1184 __field( unsigned int, host_irq )
1185 __field( int, vcpu_id )
1186 __field( unsigned int, gsi )
1187 __field( unsigned int, gvec )
1188 __field( bool, set )
1189 ),
1190
1191 TP_fast_assign(
1192 __entry->host_irq = host_irq;
1193 __entry->vcpu_id = vcpu ? vcpu->vcpu_id : -1;
1194 __entry->gsi = gsi;
1195 __entry->gvec = gvec;
1196 __entry->set = set;
1197 ),
1198
1199 TP_printk("PI is %s for irq %u, vcpu %d, gsi: 0x%x, gvec: 0x%x",
1200 __entry->set ? "enabled and being updated" : "disabled",
1201 __entry->host_irq,
1202 __entry->vcpu_id,
1203 __entry->gsi,
1204 __entry->gvec)
1205);
1206
1207/*
1208 * Tracepoint for kvm_hv_notify_acked_sint.
1209 */
1210TRACE_EVENT(kvm_hv_notify_acked_sint,
1211 TP_PROTO(int vcpu_id, u32 sint),
1212 TP_ARGS(vcpu_id, sint),
1213
1214 TP_STRUCT__entry(
1215 __field(int, vcpu_id)
1216 __field(u32, sint)
1217 ),
1218
1219 TP_fast_assign(
1220 __entry->vcpu_id = vcpu_id;
1221 __entry->sint = sint;
1222 ),
1223
1224 TP_printk("vcpu_id %d sint %u", __entry->vcpu_id, __entry->sint)
1225);
1226
1227/*
1228 * Tracepoint for synic_set_irq.
1229 */
1230TRACE_EVENT(kvm_hv_synic_set_irq,
1231 TP_PROTO(int vcpu_id, u32 sint, int vector, int ret),
1232 TP_ARGS(vcpu_id, sint, vector, ret),
1233
1234 TP_STRUCT__entry(
1235 __field(int, vcpu_id)
1236 __field(u32, sint)
1237 __field(int, vector)
1238 __field(int, ret)
1239 ),
1240
1241 TP_fast_assign(
1242 __entry->vcpu_id = vcpu_id;
1243 __entry->sint = sint;
1244 __entry->vector = vector;
1245 __entry->ret = ret;
1246 ),
1247
1248 TP_printk("vcpu_id %d sint %u vector %d ret %d",
1249 __entry->vcpu_id, __entry->sint, __entry->vector,
1250 __entry->ret)
1251);
1252
1253/*
1254 * Tracepoint for kvm_hv_synic_send_eoi.
1255 */
1256TRACE_EVENT(kvm_hv_synic_send_eoi,
1257 TP_PROTO(int vcpu_id, int vector),
1258 TP_ARGS(vcpu_id, vector),
1259
1260 TP_STRUCT__entry(
1261 __field(int, vcpu_id)
1262 __field(u32, sint)
1263 __field(int, vector)
1264 __field(int, ret)
1265 ),
1266
1267 TP_fast_assign(
1268 __entry->vcpu_id = vcpu_id;
1269 __entry->vector = vector;
1270 ),
1271
1272 TP_printk("vcpu_id %d vector %d", __entry->vcpu_id, __entry->vector)
1273);
1274
1275/*
1276 * Tracepoint for synic_set_msr.
1277 */
1278TRACE_EVENT(kvm_hv_synic_set_msr,
1279 TP_PROTO(int vcpu_id, u32 msr, u64 data, bool host),
1280 TP_ARGS(vcpu_id, msr, data, host),
1281
1282 TP_STRUCT__entry(
1283 __field(int, vcpu_id)
1284 __field(u32, msr)
1285 __field(u64, data)
1286 __field(bool, host)
1287 ),
1288
1289 TP_fast_assign(
1290 __entry->vcpu_id = vcpu_id;
1291 __entry->msr = msr;
1292 __entry->data = data;
1293 __entry->host = host
1294 ),
1295
1296 TP_printk("vcpu_id %d msr 0x%x data 0x%llx host %d",
1297 __entry->vcpu_id, __entry->msr, __entry->data, __entry->host)
1298);
1299
1300/*
1301 * Tracepoint for stimer_set_config.
1302 */
1303TRACE_EVENT(kvm_hv_stimer_set_config,
1304 TP_PROTO(int vcpu_id, int timer_index, u64 config, bool host),
1305 TP_ARGS(vcpu_id, timer_index, config, host),
1306
1307 TP_STRUCT__entry(
1308 __field(int, vcpu_id)
1309 __field(int, timer_index)
1310 __field(u64, config)
1311 __field(bool, host)
1312 ),
1313
1314 TP_fast_assign(
1315 __entry->vcpu_id = vcpu_id;
1316 __entry->timer_index = timer_index;
1317 __entry->config = config;
1318 __entry->host = host;
1319 ),
1320
1321 TP_printk("vcpu_id %d timer %d config 0x%llx host %d",
1322 __entry->vcpu_id, __entry->timer_index, __entry->config,
1323 __entry->host)
1324);
1325
1326/*
1327 * Tracepoint for stimer_set_count.
1328 */
1329TRACE_EVENT(kvm_hv_stimer_set_count,
1330 TP_PROTO(int vcpu_id, int timer_index, u64 count, bool host),
1331 TP_ARGS(vcpu_id, timer_index, count, host),
1332
1333 TP_STRUCT__entry(
1334 __field(int, vcpu_id)
1335 __field(int, timer_index)
1336 __field(u64, count)
1337 __field(bool, host)
1338 ),
1339
1340 TP_fast_assign(
1341 __entry->vcpu_id = vcpu_id;
1342 __entry->timer_index = timer_index;
1343 __entry->count = count;
1344 __entry->host = host;
1345 ),
1346
1347 TP_printk("vcpu_id %d timer %d count %llu host %d",
1348 __entry->vcpu_id, __entry->timer_index, __entry->count,
1349 __entry->host)
1350);
1351
1352/*
1353 * Tracepoint for stimer_start(periodic timer case).
1354 */
1355TRACE_EVENT(kvm_hv_stimer_start_periodic,
1356 TP_PROTO(int vcpu_id, int timer_index, u64 time_now, u64 exp_time),
1357 TP_ARGS(vcpu_id, timer_index, time_now, exp_time),
1358
1359 TP_STRUCT__entry(
1360 __field(int, vcpu_id)
1361 __field(int, timer_index)
1362 __field(u64, time_now)
1363 __field(u64, exp_time)
1364 ),
1365
1366 TP_fast_assign(
1367 __entry->vcpu_id = vcpu_id;
1368 __entry->timer_index = timer_index;
1369 __entry->time_now = time_now;
1370 __entry->exp_time = exp_time;
1371 ),
1372
1373 TP_printk("vcpu_id %d timer %d time_now %llu exp_time %llu",
1374 __entry->vcpu_id, __entry->timer_index, __entry->time_now,
1375 __entry->exp_time)
1376);
1377
1378/*
1379 * Tracepoint for stimer_start(one-shot timer case).
1380 */
1381TRACE_EVENT(kvm_hv_stimer_start_one_shot,
1382 TP_PROTO(int vcpu_id, int timer_index, u64 time_now, u64 count),
1383 TP_ARGS(vcpu_id, timer_index, time_now, count),
1384
1385 TP_STRUCT__entry(
1386 __field(int, vcpu_id)
1387 __field(int, timer_index)
1388 __field(u64, time_now)
1389 __field(u64, count)
1390 ),
1391
1392 TP_fast_assign(
1393 __entry->vcpu_id = vcpu_id;
1394 __entry->timer_index = timer_index;
1395 __entry->time_now = time_now;
1396 __entry->count = count;
1397 ),
1398
1399 TP_printk("vcpu_id %d timer %d time_now %llu count %llu",
1400 __entry->vcpu_id, __entry->timer_index, __entry->time_now,
1401 __entry->count)
1402);
1403
1404/*
1405 * Tracepoint for stimer_timer_callback.
1406 */
1407TRACE_EVENT(kvm_hv_stimer_callback,
1408 TP_PROTO(int vcpu_id, int timer_index),
1409 TP_ARGS(vcpu_id, timer_index),
1410
1411 TP_STRUCT__entry(
1412 __field(int, vcpu_id)
1413 __field(int, timer_index)
1414 ),
1415
1416 TP_fast_assign(
1417 __entry->vcpu_id = vcpu_id;
1418 __entry->timer_index = timer_index;
1419 ),
1420
1421 TP_printk("vcpu_id %d timer %d",
1422 __entry->vcpu_id, __entry->timer_index)
1423);
1424
1425/*
1426 * Tracepoint for stimer_expiration.
1427 */
1428TRACE_EVENT(kvm_hv_stimer_expiration,
1429 TP_PROTO(int vcpu_id, int timer_index, int direct, int msg_send_result),
1430 TP_ARGS(vcpu_id, timer_index, direct, msg_send_result),
1431
1432 TP_STRUCT__entry(
1433 __field(int, vcpu_id)
1434 __field(int, timer_index)
1435 __field(int, direct)
1436 __field(int, msg_send_result)
1437 ),
1438
1439 TP_fast_assign(
1440 __entry->vcpu_id = vcpu_id;
1441 __entry->timer_index = timer_index;
1442 __entry->direct = direct;
1443 __entry->msg_send_result = msg_send_result;
1444 ),
1445
1446 TP_printk("vcpu_id %d timer %d direct %d send result %d",
1447 __entry->vcpu_id, __entry->timer_index,
1448 __entry->direct, __entry->msg_send_result)
1449);
1450
1451/*
1452 * Tracepoint for stimer_cleanup.
1453 */
1454TRACE_EVENT(kvm_hv_stimer_cleanup,
1455 TP_PROTO(int vcpu_id, int timer_index),
1456 TP_ARGS(vcpu_id, timer_index),
1457
1458 TP_STRUCT__entry(
1459 __field(int, vcpu_id)
1460 __field(int, timer_index)
1461 ),
1462
1463 TP_fast_assign(
1464 __entry->vcpu_id = vcpu_id;
1465 __entry->timer_index = timer_index;
1466 ),
1467
1468 TP_printk("vcpu_id %d timer %d",
1469 __entry->vcpu_id, __entry->timer_index)
1470);
1471
1472#define kvm_print_apicv_inhibit_reasons(inhibits) \
1473 (inhibits), (inhibits) ? " " : "", \
1474 (inhibits) ? __print_flags(inhibits, "|", APICV_INHIBIT_REASONS) : ""
1475
1476TRACE_EVENT(kvm_apicv_inhibit_changed,
1477 TP_PROTO(int reason, bool set, unsigned long inhibits),
1478 TP_ARGS(reason, set, inhibits),
1479
1480 TP_STRUCT__entry(
1481 __field(int, reason)
1482 __field(bool, set)
1483 __field(unsigned long, inhibits)
1484 ),
1485
1486 TP_fast_assign(
1487 __entry->reason = reason;
1488 __entry->set = set;
1489 __entry->inhibits = inhibits;
1490 ),
1491
1492 TP_printk("%s reason=%u, inhibits=0x%lx%s%s",
1493 __entry->set ? "set" : "cleared",
1494 __entry->reason,
1495 kvm_print_apicv_inhibit_reasons(__entry->inhibits))
1496);
1497
1498TRACE_EVENT(kvm_apicv_accept_irq,
1499 TP_PROTO(__u32 apicid, __u16 dm, __u16 tm, __u8 vec),
1500 TP_ARGS(apicid, dm, tm, vec),
1501
1502 TP_STRUCT__entry(
1503 __field( __u32, apicid )
1504 __field( __u16, dm )
1505 __field( __u16, tm )
1506 __field( __u8, vec )
1507 ),
1508
1509 TP_fast_assign(
1510 __entry->apicid = apicid;
1511 __entry->dm = dm;
1512 __entry->tm = tm;
1513 __entry->vec = vec;
1514 ),
1515
1516 TP_printk("apicid %x vec %u (%s|%s)",
1517 __entry->apicid, __entry->vec,
1518 __print_symbolic((__entry->dm >> 8 & 0x7), kvm_deliver_mode),
1519 __entry->tm ? "level" : "edge")
1520);
1521
1522/*
1523 * Tracepoint for AMD AVIC
1524 */
1525TRACE_EVENT(kvm_avic_incomplete_ipi,
1526 TP_PROTO(u32 vcpu, u32 icrh, u32 icrl, u32 id, u32 index),
1527 TP_ARGS(vcpu, icrh, icrl, id, index),
1528
1529 TP_STRUCT__entry(
1530 __field(u32, vcpu)
1531 __field(u32, icrh)
1532 __field(u32, icrl)
1533 __field(u32, id)
1534 __field(u32, index)
1535 ),
1536
1537 TP_fast_assign(
1538 __entry->vcpu = vcpu;
1539 __entry->icrh = icrh;
1540 __entry->icrl = icrl;
1541 __entry->id = id;
1542 __entry->index = index;
1543 ),
1544
1545 TP_printk("vcpu=%u, icrh:icrl=%#010x:%08x, id=%u, index=%u",
1546 __entry->vcpu, __entry->icrh, __entry->icrl,
1547 __entry->id, __entry->index)
1548);
1549
1550TRACE_EVENT(kvm_avic_unaccelerated_access,
1551 TP_PROTO(u32 vcpu, u32 offset, bool ft, bool rw, u32 vec),
1552 TP_ARGS(vcpu, offset, ft, rw, vec),
1553
1554 TP_STRUCT__entry(
1555 __field(u32, vcpu)
1556 __field(u32, offset)
1557 __field(bool, ft)
1558 __field(bool, rw)
1559 __field(u32, vec)
1560 ),
1561
1562 TP_fast_assign(
1563 __entry->vcpu = vcpu;
1564 __entry->offset = offset;
1565 __entry->ft = ft;
1566 __entry->rw = rw;
1567 __entry->vec = vec;
1568 ),
1569
1570 TP_printk("vcpu=%u, offset=%#x(%s), %s, %s, vec=%#x",
1571 __entry->vcpu,
1572 __entry->offset,
1573 __print_symbolic(__entry->offset, kvm_trace_symbol_apic),
1574 __entry->ft ? "trap" : "fault",
1575 __entry->rw ? "write" : "read",
1576 __entry->vec)
1577);
1578
1579TRACE_EVENT(kvm_avic_ga_log,
1580 TP_PROTO(u32 vmid, u32 vcpuid),
1581 TP_ARGS(vmid, vcpuid),
1582
1583 TP_STRUCT__entry(
1584 __field(u32, vmid)
1585 __field(u32, vcpuid)
1586 ),
1587
1588 TP_fast_assign(
1589 __entry->vmid = vmid;
1590 __entry->vcpuid = vcpuid;
1591 ),
1592
1593 TP_printk("vmid=%u, vcpuid=%u",
1594 __entry->vmid, __entry->vcpuid)
1595);
1596
1597TRACE_EVENT(kvm_avic_kick_vcpu_slowpath,
1598 TP_PROTO(u32 icrh, u32 icrl, u32 index),
1599 TP_ARGS(icrh, icrl, index),
1600
1601 TP_STRUCT__entry(
1602 __field(u32, icrh)
1603 __field(u32, icrl)
1604 __field(u32, index)
1605 ),
1606
1607 TP_fast_assign(
1608 __entry->icrh = icrh;
1609 __entry->icrl = icrl;
1610 __entry->index = index;
1611 ),
1612
1613 TP_printk("icrh:icrl=%#08x:%08x, index=%u",
1614 __entry->icrh, __entry->icrl, __entry->index)
1615);
1616
1617TRACE_EVENT(kvm_avic_doorbell,
1618 TP_PROTO(u32 vcpuid, u32 apicid),
1619 TP_ARGS(vcpuid, apicid),
1620
1621 TP_STRUCT__entry(
1622 __field(u32, vcpuid)
1623 __field(u32, apicid)
1624 ),
1625
1626 TP_fast_assign(
1627 __entry->vcpuid = vcpuid;
1628 __entry->apicid = apicid;
1629 ),
1630
1631 TP_printk("vcpuid=%u, apicid=%u",
1632 __entry->vcpuid, __entry->apicid)
1633);
1634
1635TRACE_EVENT(kvm_hv_timer_state,
1636 TP_PROTO(unsigned int vcpu_id, unsigned int hv_timer_in_use),
1637 TP_ARGS(vcpu_id, hv_timer_in_use),
1638 TP_STRUCT__entry(
1639 __field(unsigned int, vcpu_id)
1640 __field(unsigned int, hv_timer_in_use)
1641 ),
1642 TP_fast_assign(
1643 __entry->vcpu_id = vcpu_id;
1644 __entry->hv_timer_in_use = hv_timer_in_use;
1645 ),
1646 TP_printk("vcpu_id %x hv_timer %x",
1647 __entry->vcpu_id,
1648 __entry->hv_timer_in_use)
1649);
1650
1651/*
1652 * Tracepoint for kvm_hv_flush_tlb.
1653 */
1654TRACE_EVENT(kvm_hv_flush_tlb,
1655 TP_PROTO(u64 processor_mask, u64 address_space, u64 flags, bool guest_mode),
1656 TP_ARGS(processor_mask, address_space, flags, guest_mode),
1657
1658 TP_STRUCT__entry(
1659 __field(u64, processor_mask)
1660 __field(u64, address_space)
1661 __field(u64, flags)
1662 __field(bool, guest_mode)
1663 ),
1664
1665 TP_fast_assign(
1666 __entry->processor_mask = processor_mask;
1667 __entry->address_space = address_space;
1668 __entry->flags = flags;
1669 __entry->guest_mode = guest_mode;
1670 ),
1671
1672 TP_printk("processor_mask 0x%llx address_space 0x%llx flags 0x%llx %s",
1673 __entry->processor_mask, __entry->address_space,
1674 __entry->flags, __entry->guest_mode ? "(L2)" : "")
1675);
1676
1677/*
1678 * Tracepoint for kvm_hv_flush_tlb_ex.
1679 */
1680TRACE_EVENT(kvm_hv_flush_tlb_ex,
1681 TP_PROTO(u64 valid_bank_mask, u64 format, u64 address_space, u64 flags, bool guest_mode),
1682 TP_ARGS(valid_bank_mask, format, address_space, flags, guest_mode),
1683
1684 TP_STRUCT__entry(
1685 __field(u64, valid_bank_mask)
1686 __field(u64, format)
1687 __field(u64, address_space)
1688 __field(u64, flags)
1689 __field(bool, guest_mode)
1690 ),
1691
1692 TP_fast_assign(
1693 __entry->valid_bank_mask = valid_bank_mask;
1694 __entry->format = format;
1695 __entry->address_space = address_space;
1696 __entry->flags = flags;
1697 __entry->guest_mode = guest_mode;
1698 ),
1699
1700 TP_printk("valid_bank_mask 0x%llx format 0x%llx "
1701 "address_space 0x%llx flags 0x%llx %s",
1702 __entry->valid_bank_mask, __entry->format,
1703 __entry->address_space, __entry->flags,
1704 __entry->guest_mode ? "(L2)" : "")
1705);
1706
1707/*
1708 * Tracepoints for kvm_hv_send_ipi.
1709 */
1710TRACE_EVENT(kvm_hv_send_ipi,
1711 TP_PROTO(u32 vector, u64 processor_mask),
1712 TP_ARGS(vector, processor_mask),
1713
1714 TP_STRUCT__entry(
1715 __field(u32, vector)
1716 __field(u64, processor_mask)
1717 ),
1718
1719 TP_fast_assign(
1720 __entry->vector = vector;
1721 __entry->processor_mask = processor_mask;
1722 ),
1723
1724 TP_printk("vector %x processor_mask 0x%llx",
1725 __entry->vector, __entry->processor_mask)
1726);
1727
1728TRACE_EVENT(kvm_hv_send_ipi_ex,
1729 TP_PROTO(u32 vector, u64 format, u64 valid_bank_mask),
1730 TP_ARGS(vector, format, valid_bank_mask),
1731
1732 TP_STRUCT__entry(
1733 __field(u32, vector)
1734 __field(u64, format)
1735 __field(u64, valid_bank_mask)
1736 ),
1737
1738 TP_fast_assign(
1739 __entry->vector = vector;
1740 __entry->format = format;
1741 __entry->valid_bank_mask = valid_bank_mask;
1742 ),
1743
1744 TP_printk("vector %x format %llx valid_bank_mask 0x%llx",
1745 __entry->vector, __entry->format,
1746 __entry->valid_bank_mask)
1747);
1748
1749TRACE_EVENT(kvm_pv_tlb_flush,
1750 TP_PROTO(unsigned int vcpu_id, bool need_flush_tlb),
1751 TP_ARGS(vcpu_id, need_flush_tlb),
1752
1753 TP_STRUCT__entry(
1754 __field( unsigned int, vcpu_id )
1755 __field( bool, need_flush_tlb )
1756 ),
1757
1758 TP_fast_assign(
1759 __entry->vcpu_id = vcpu_id;
1760 __entry->need_flush_tlb = need_flush_tlb;
1761 ),
1762
1763 TP_printk("vcpu %u need_flush_tlb %s", __entry->vcpu_id,
1764 __entry->need_flush_tlb ? "true" : "false")
1765);
1766
1767/*
1768 * Tracepoint for failed nested VMX VM-Enter.
1769 */
1770TRACE_EVENT(kvm_nested_vmenter_failed,
1771 TP_PROTO(const char *msg, u32 err),
1772 TP_ARGS(msg, err),
1773
1774 TP_STRUCT__entry(
1775 __string(msg, msg)
1776 __field(u32, err)
1777 ),
1778
1779 TP_fast_assign(
1780 __assign_str(msg);
1781 __entry->err = err;
1782 ),
1783
1784 TP_printk("%s%s", __get_str(msg), !__entry->err ? "" :
1785 __print_symbolic(__entry->err, VMX_VMENTER_INSTRUCTION_ERRORS))
1786);
1787
1788/*
1789 * Tracepoint for syndbg_set_msr.
1790 */
1791TRACE_EVENT(kvm_hv_syndbg_set_msr,
1792 TP_PROTO(int vcpu_id, u32 vp_index, u32 msr, u64 data),
1793 TP_ARGS(vcpu_id, vp_index, msr, data),
1794
1795 TP_STRUCT__entry(
1796 __field(int, vcpu_id)
1797 __field(u32, vp_index)
1798 __field(u32, msr)
1799 __field(u64, data)
1800 ),
1801
1802 TP_fast_assign(
1803 __entry->vcpu_id = vcpu_id;
1804 __entry->vp_index = vp_index;
1805 __entry->msr = msr;
1806 __entry->data = data;
1807 ),
1808
1809 TP_printk("vcpu_id %d vp_index %u msr 0x%x data 0x%llx",
1810 __entry->vcpu_id, __entry->vp_index, __entry->msr,
1811 __entry->data)
1812);
1813
1814/*
1815 * Tracepoint for syndbg_get_msr.
1816 */
1817TRACE_EVENT(kvm_hv_syndbg_get_msr,
1818 TP_PROTO(int vcpu_id, u32 vp_index, u32 msr, u64 data),
1819 TP_ARGS(vcpu_id, vp_index, msr, data),
1820
1821 TP_STRUCT__entry(
1822 __field(int, vcpu_id)
1823 __field(u32, vp_index)
1824 __field(u32, msr)
1825 __field(u64, data)
1826 ),
1827
1828 TP_fast_assign(
1829 __entry->vcpu_id = vcpu_id;
1830 __entry->vp_index = vp_index;
1831 __entry->msr = msr;
1832 __entry->data = data;
1833 ),
1834
1835 TP_printk("vcpu_id %d vp_index %u msr 0x%x data 0x%llx",
1836 __entry->vcpu_id, __entry->vp_index, __entry->msr,
1837 __entry->data)
1838);
1839
1840/*
1841 * Tracepoint for the start of VMGEXIT processing
1842 */
1843TRACE_EVENT(kvm_vmgexit_enter,
1844 TP_PROTO(unsigned int vcpu_id, struct ghcb *ghcb),
1845 TP_ARGS(vcpu_id, ghcb),
1846
1847 TP_STRUCT__entry(
1848 __field(unsigned int, vcpu_id)
1849 __field(u64, exit_reason)
1850 __field(u64, info1)
1851 __field(u64, info2)
1852 ),
1853
1854 TP_fast_assign(
1855 __entry->vcpu_id = vcpu_id;
1856 __entry->exit_reason = ghcb->save.sw_exit_code;
1857 __entry->info1 = ghcb->save.sw_exit_info_1;
1858 __entry->info2 = ghcb->save.sw_exit_info_2;
1859 ),
1860
1861 TP_printk("vcpu %u, exit_reason %llx, exit_info1 %llx, exit_info2 %llx",
1862 __entry->vcpu_id, __entry->exit_reason,
1863 __entry->info1, __entry->info2)
1864);
1865
1866/*
1867 * Tracepoint for the end of VMGEXIT processing
1868 */
1869TRACE_EVENT(kvm_vmgexit_exit,
1870 TP_PROTO(unsigned int vcpu_id, struct ghcb *ghcb),
1871 TP_ARGS(vcpu_id, ghcb),
1872
1873 TP_STRUCT__entry(
1874 __field(unsigned int, vcpu_id)
1875 __field(u64, exit_reason)
1876 __field(u64, info1)
1877 __field(u64, info2)
1878 ),
1879
1880 TP_fast_assign(
1881 __entry->vcpu_id = vcpu_id;
1882 __entry->exit_reason = ghcb->save.sw_exit_code;
1883 __entry->info1 = ghcb->save.sw_exit_info_1;
1884 __entry->info2 = ghcb->save.sw_exit_info_2;
1885 ),
1886
1887 TP_printk("vcpu %u, exit_reason %llx, exit_info1 %llx, exit_info2 %llx",
1888 __entry->vcpu_id, __entry->exit_reason,
1889 __entry->info1, __entry->info2)
1890);
1891
1892/*
1893 * Tracepoint for the start of VMGEXIT MSR procotol processing
1894 */
1895TRACE_EVENT(kvm_vmgexit_msr_protocol_enter,
1896 TP_PROTO(unsigned int vcpu_id, u64 ghcb_gpa),
1897 TP_ARGS(vcpu_id, ghcb_gpa),
1898
1899 TP_STRUCT__entry(
1900 __field(unsigned int, vcpu_id)
1901 __field(u64, ghcb_gpa)
1902 ),
1903
1904 TP_fast_assign(
1905 __entry->vcpu_id = vcpu_id;
1906 __entry->ghcb_gpa = ghcb_gpa;
1907 ),
1908
1909 TP_printk("vcpu %u, ghcb_gpa %016llx",
1910 __entry->vcpu_id, __entry->ghcb_gpa)
1911);
1912
1913/*
1914 * Tracepoint for the end of VMGEXIT MSR procotol processing
1915 */
1916TRACE_EVENT(kvm_vmgexit_msr_protocol_exit,
1917 TP_PROTO(unsigned int vcpu_id, u64 ghcb_gpa, int result),
1918 TP_ARGS(vcpu_id, ghcb_gpa, result),
1919
1920 TP_STRUCT__entry(
1921 __field(unsigned int, vcpu_id)
1922 __field(u64, ghcb_gpa)
1923 __field(int, result)
1924 ),
1925
1926 TP_fast_assign(
1927 __entry->vcpu_id = vcpu_id;
1928 __entry->ghcb_gpa = ghcb_gpa;
1929 __entry->result = result;
1930 ),
1931
1932 TP_printk("vcpu %u, ghcb_gpa %016llx, result %d",
1933 __entry->vcpu_id, __entry->ghcb_gpa, __entry->result)
1934);
1935
1936/*
1937 * Tracepoint for #NPFs due to RMP faults.
1938 */
1939TRACE_EVENT(kvm_rmp_fault,
1940 TP_PROTO(struct kvm_vcpu *vcpu, u64 gpa, u64 pfn, u64 error_code,
1941 int rmp_level, int psmash_ret),
1942 TP_ARGS(vcpu, gpa, pfn, error_code, rmp_level, psmash_ret),
1943
1944 TP_STRUCT__entry(
1945 __field(unsigned int, vcpu_id)
1946 __field(u64, gpa)
1947 __field(u64, pfn)
1948 __field(u64, error_code)
1949 __field(int, rmp_level)
1950 __field(int, psmash_ret)
1951 ),
1952
1953 TP_fast_assign(
1954 __entry->vcpu_id = vcpu->vcpu_id;
1955 __entry->gpa = gpa;
1956 __entry->pfn = pfn;
1957 __entry->error_code = error_code;
1958 __entry->rmp_level = rmp_level;
1959 __entry->psmash_ret = psmash_ret;
1960 ),
1961
1962 TP_printk("vcpu %u gpa %016llx pfn 0x%llx error_code 0x%llx rmp_level %d psmash_ret %d",
1963 __entry->vcpu_id, __entry->gpa, __entry->pfn,
1964 __entry->error_code, __entry->rmp_level, __entry->psmash_ret)
1965);
1966
1967#endif /* _TRACE_KVM_H */
1968
1969#undef TRACE_INCLUDE_PATH
1970#define TRACE_INCLUDE_PATH ../../arch/x86/kvm
1971#undef TRACE_INCLUDE_FILE
1972#define TRACE_INCLUDE_FILE trace
1973
1974/* This part must be outside protection */
1975#include <trace/define_trace.h>