1 /* SPDX-License-Identifier: GPL-2.0 */ 2 #if !defined(_TRACE_KVM_H) || defined(TRACE_HEADER_MULTI_READ) 3 #define _TRACE_KVM_H 4 5 #include <linux/tracepoint.h> 6 #include <asm/vmx.h> 7 #include <asm/svm.h> 8 #include <asm/clocksource.h> 9 #include <asm/pvclock-abi.h> 10 11 #undef TRACE_SYSTEM 12 #define TRACE_SYSTEM kvm 13 14 /* 15 * Tracepoint for guest mode entry. 16 */ 17 TRACE_EVENT(kvm_entry, 18 TP_PROTO(struct kvm_vcpu *vcpu), 19 TP_ARGS(vcpu), 20 21 TP_STRUCT__entry( 22 __field( unsigned int, vcpu_id ) 23 __field( unsigned long, rip ) 24 ), 25 26 TP_fast_assign( 27 __entry->vcpu_id = vcpu->vcpu_id; 28 __entry->rip = kvm_rip_read(vcpu); 29 ), 30 31 TP_printk("vcpu %u, rip 0x%lx", __entry->vcpu_id, __entry->rip) 32 ); 33 34 /* 35 * Tracepoint for hypercall. 36 */ 37 TRACE_EVENT(kvm_hypercall, 38 TP_PROTO(unsigned long nr, unsigned long a0, unsigned long a1, 39 unsigned long a2, unsigned long a3), 40 TP_ARGS(nr, a0, a1, a2, a3), 41 42 TP_STRUCT__entry( 43 __field( unsigned long, nr ) 44 __field( unsigned long, a0 ) 45 __field( unsigned long, a1 ) 46 __field( unsigned long, a2 ) 47 __field( unsigned long, a3 ) 48 ), 49 50 TP_fast_assign( 51 __entry->nr = nr; 52 __entry->a0 = a0; 53 __entry->a1 = a1; 54 __entry->a2 = a2; 55 __entry->a3 = a3; 56 ), 57 58 TP_printk("nr 0x%lx a0 0x%lx a1 0x%lx a2 0x%lx a3 0x%lx", 59 __entry->nr, __entry->a0, __entry->a1, __entry->a2, 60 __entry->a3) 61 ); 62 63 /* 64 * Tracepoint for hypercall. 65 */ 66 TRACE_EVENT(kvm_hv_hypercall, 67 TP_PROTO(__u16 code, bool fast, __u16 rep_cnt, __u16 rep_idx, 68 __u64 ingpa, __u64 outgpa), 69 TP_ARGS(code, fast, rep_cnt, rep_idx, ingpa, outgpa), 70 71 TP_STRUCT__entry( 72 __field( __u16, rep_cnt ) 73 __field( __u16, rep_idx ) 74 __field( __u64, ingpa ) 75 __field( __u64, outgpa ) 76 __field( __u16, code ) 77 __field( bool, fast ) 78 ), 79 80 TP_fast_assign( 81 __entry->rep_cnt = rep_cnt; 82 __entry->rep_idx = rep_idx; 83 __entry->ingpa = ingpa; 84 __entry->outgpa = outgpa; 85 __entry->code = code; 86 __entry->fast = fast; 87 ), 88 89 TP_printk("code 0x%x %s cnt 0x%x idx 0x%x in 0x%llx out 0x%llx", 90 __entry->code, __entry->fast ? "fast" : "slow", 91 __entry->rep_cnt, __entry->rep_idx, __entry->ingpa, 92 __entry->outgpa) 93 ); 94 95 /* 96 * Tracepoint for PIO. 97 */ 98 99 #define KVM_PIO_IN 0 100 #define KVM_PIO_OUT 1 101 102 TRACE_EVENT(kvm_pio, 103 TP_PROTO(unsigned int rw, unsigned int port, unsigned int size, 104 unsigned int count, void *data), 105 TP_ARGS(rw, port, size, count, data), 106 107 TP_STRUCT__entry( 108 __field( unsigned int, rw ) 109 __field( unsigned int, port ) 110 __field( unsigned int, size ) 111 __field( unsigned int, count ) 112 __field( unsigned int, val ) 113 ), 114 115 TP_fast_assign( 116 __entry->rw = rw; 117 __entry->port = port; 118 __entry->size = size; 119 __entry->count = count; 120 if (size == 1) 121 __entry->val = *(unsigned char *)data; 122 else if (size == 2) 123 __entry->val = *(unsigned short *)data; 124 else 125 __entry->val = *(unsigned int *)data; 126 ), 127 128 TP_printk("pio_%s at 0x%x size %d count %d val 0x%x %s", 129 __entry->rw ? "write" : "read", 130 __entry->port, __entry->size, __entry->count, __entry->val, 131 __entry->count > 1 ? "(...)" : "") 132 ); 133 134 /* 135 * Tracepoint for fast mmio. 136 */ 137 TRACE_EVENT(kvm_fast_mmio, 138 TP_PROTO(u64 gpa), 139 TP_ARGS(gpa), 140 141 TP_STRUCT__entry( 142 __field(u64, gpa) 143 ), 144 145 TP_fast_assign( 146 __entry->gpa = gpa; 147 ), 148 149 TP_printk("fast mmio at gpa 0x%llx", __entry->gpa) 150 ); 151 152 /* 153 * Tracepoint for cpuid. 154 */ 155 TRACE_EVENT(kvm_cpuid, 156 TP_PROTO(unsigned int function, unsigned int index, unsigned long rax, 157 unsigned long rbx, unsigned long rcx, unsigned long rdx, 158 bool found, bool used_max_basic), 159 TP_ARGS(function, index, rax, rbx, rcx, rdx, found, used_max_basic), 160 161 TP_STRUCT__entry( 162 __field( unsigned int, function ) 163 __field( unsigned int, index ) 164 __field( unsigned long, rax ) 165 __field( unsigned long, rbx ) 166 __field( unsigned long, rcx ) 167 __field( unsigned long, rdx ) 168 __field( bool, found ) 169 __field( bool, used_max_basic ) 170 ), 171 172 TP_fast_assign( 173 __entry->function = function; 174 __entry->index = index; 175 __entry->rax = rax; 176 __entry->rbx = rbx; 177 __entry->rcx = rcx; 178 __entry->rdx = rdx; 179 __entry->found = found; 180 __entry->used_max_basic = used_max_basic; 181 ), 182 183 TP_printk("func %x idx %x rax %lx rbx %lx rcx %lx rdx %lx, cpuid entry %s%s", 184 __entry->function, __entry->index, __entry->rax, 185 __entry->rbx, __entry->rcx, __entry->rdx, 186 __entry->found ? "found" : "not found", 187 __entry->used_max_basic ? ", used max basic" : "") 188 ); 189 190 #define AREG(x) { APIC_##x, "APIC_" #x } 191 192 #define kvm_trace_symbol_apic \ 193 AREG(ID), AREG(LVR), AREG(TASKPRI), AREG(ARBPRI), AREG(PROCPRI), \ 194 AREG(EOI), AREG(RRR), AREG(LDR), AREG(DFR), AREG(SPIV), AREG(ISR), \ 195 AREG(TMR), AREG(IRR), AREG(ESR), AREG(ICR), AREG(ICR2), AREG(LVTT), \ 196 AREG(LVTTHMR), AREG(LVTPC), AREG(LVT0), AREG(LVT1), AREG(LVTERR), \ 197 AREG(TMICT), AREG(TMCCT), AREG(TDCR), AREG(SELF_IPI), AREG(EFEAT), \ 198 AREG(ECTRL) 199 /* 200 * Tracepoint for apic access. 201 */ 202 TRACE_EVENT(kvm_apic, 203 TP_PROTO(unsigned int rw, unsigned int reg, unsigned int val), 204 TP_ARGS(rw, reg, val), 205 206 TP_STRUCT__entry( 207 __field( unsigned int, rw ) 208 __field( unsigned int, reg ) 209 __field( unsigned int, val ) 210 ), 211 212 TP_fast_assign( 213 __entry->rw = rw; 214 __entry->reg = reg; 215 __entry->val = val; 216 ), 217 218 TP_printk("apic_%s %s = 0x%x", 219 __entry->rw ? "write" : "read", 220 __print_symbolic(__entry->reg, kvm_trace_symbol_apic), 221 __entry->val) 222 ); 223 224 #define trace_kvm_apic_read(reg, val) trace_kvm_apic(0, reg, val) 225 #define trace_kvm_apic_write(reg, val) trace_kvm_apic(1, reg, val) 226 227 #define KVM_ISA_VMX 1 228 #define KVM_ISA_SVM 2 229 230 #define kvm_print_exit_reason(exit_reason, isa) \ 231 (isa == KVM_ISA_VMX) ? \ 232 __print_symbolic(exit_reason & 0xffff, VMX_EXIT_REASONS) : \ 233 __print_symbolic(exit_reason, SVM_EXIT_REASONS), \ 234 (isa == KVM_ISA_VMX && exit_reason & ~0xffff) ? " " : "", \ 235 (isa == KVM_ISA_VMX) ? \ 236 __print_flags(exit_reason & ~0xffff, " ", VMX_EXIT_REASON_FLAGS) : "" 237 238 #define TRACE_EVENT_KVM_EXIT(name) \ 239 TRACE_EVENT(name, \ 240 TP_PROTO(unsigned int exit_reason, struct kvm_vcpu *vcpu, u32 isa), \ 241 TP_ARGS(exit_reason, vcpu, isa), \ 242 \ 243 TP_STRUCT__entry( \ 244 __field( unsigned int, exit_reason ) \ 245 __field( unsigned long, guest_rip ) \ 246 __field( u32, isa ) \ 247 __field( u64, info1 ) \ 248 __field( u64, info2 ) \ 249 __field( u32, intr_info ) \ 250 __field( u32, error_code ) \ 251 __field( unsigned int, vcpu_id ) \ 252 ), \ 253 \ 254 TP_fast_assign( \ 255 __entry->exit_reason = exit_reason; \ 256 __entry->guest_rip = kvm_rip_read(vcpu); \ 257 __entry->isa = isa; \ 258 __entry->vcpu_id = vcpu->vcpu_id; \ 259 kvm_x86_ops.get_exit_info(vcpu, &__entry->info1, \ 260 &__entry->info2, \ 261 &__entry->intr_info, \ 262 &__entry->error_code); \ 263 ), \ 264 \ 265 TP_printk("vcpu %u reason %s%s%s rip 0x%lx info1 0x%016llx " \ 266 "info2 0x%016llx intr_info 0x%08x error_code 0x%08x", \ 267 __entry->vcpu_id, \ 268 kvm_print_exit_reason(__entry->exit_reason, __entry->isa), \ 269 __entry->guest_rip, __entry->info1, __entry->info2, \ 270 __entry->intr_info, __entry->error_code) \ 271 ) 272 273 /* 274 * Tracepoint for kvm guest exit: 275 */ 276 TRACE_EVENT_KVM_EXIT(kvm_exit); 277 278 /* 279 * Tracepoint for kvm interrupt injection: 280 */ 281 TRACE_EVENT(kvm_inj_virq, 282 TP_PROTO(unsigned int irq), 283 TP_ARGS(irq), 284 285 TP_STRUCT__entry( 286 __field( unsigned int, irq ) 287 ), 288 289 TP_fast_assign( 290 __entry->irq = irq; 291 ), 292 293 TP_printk("irq %u", __entry->irq) 294 ); 295 296 #define EXS(x) { x##_VECTOR, "#" #x } 297 298 #define kvm_trace_sym_exc \ 299 EXS(DE), EXS(DB), EXS(BP), EXS(OF), EXS(BR), EXS(UD), EXS(NM), \ 300 EXS(DF), EXS(TS), EXS(NP), EXS(SS), EXS(GP), EXS(PF), \ 301 EXS(MF), EXS(AC), EXS(MC) 302 303 /* 304 * Tracepoint for kvm interrupt injection: 305 */ 306 TRACE_EVENT(kvm_inj_exception, 307 TP_PROTO(unsigned exception, bool has_error, unsigned error_code, 308 bool reinjected), 309 TP_ARGS(exception, has_error, error_code, reinjected), 310 311 TP_STRUCT__entry( 312 __field( u8, exception ) 313 __field( u8, has_error ) 314 __field( u32, error_code ) 315 __field( bool, reinjected ) 316 ), 317 318 TP_fast_assign( 319 __entry->exception = exception; 320 __entry->has_error = has_error; 321 __entry->error_code = error_code; 322 __entry->reinjected = reinjected; 323 ), 324 325 TP_printk("%s (0x%x)%s", 326 __print_symbolic(__entry->exception, kvm_trace_sym_exc), 327 /* FIXME: don't print error_code if not present */ 328 __entry->has_error ? __entry->error_code : 0, 329 __entry->reinjected ? " [reinjected]" : "") 330 ); 331 332 /* 333 * Tracepoint for page fault. 334 */ 335 TRACE_EVENT(kvm_page_fault, 336 TP_PROTO(unsigned long fault_address, unsigned int error_code), 337 TP_ARGS(fault_address, error_code), 338 339 TP_STRUCT__entry( 340 __field( unsigned long, fault_address ) 341 __field( unsigned int, error_code ) 342 ), 343 344 TP_fast_assign( 345 __entry->fault_address = fault_address; 346 __entry->error_code = error_code; 347 ), 348 349 TP_printk("address %lx error_code %x", 350 __entry->fault_address, __entry->error_code) 351 ); 352 353 /* 354 * Tracepoint for guest MSR access. 355 */ 356 TRACE_EVENT(kvm_msr, 357 TP_PROTO(unsigned write, u32 ecx, u64 data, bool exception), 358 TP_ARGS(write, ecx, data, exception), 359 360 TP_STRUCT__entry( 361 __field( unsigned, write ) 362 __field( u32, ecx ) 363 __field( u64, data ) 364 __field( u8, exception ) 365 ), 366 367 TP_fast_assign( 368 __entry->write = write; 369 __entry->ecx = ecx; 370 __entry->data = data; 371 __entry->exception = exception; 372 ), 373 374 TP_printk("msr_%s %x = 0x%llx%s", 375 __entry->write ? "write" : "read", 376 __entry->ecx, __entry->data, 377 __entry->exception ? " (#GP)" : "") 378 ); 379 380 #define trace_kvm_msr_read(ecx, data) trace_kvm_msr(0, ecx, data, false) 381 #define trace_kvm_msr_write(ecx, data) trace_kvm_msr(1, ecx, data, false) 382 #define trace_kvm_msr_read_ex(ecx) trace_kvm_msr(0, ecx, 0, true) 383 #define trace_kvm_msr_write_ex(ecx, data) trace_kvm_msr(1, ecx, data, true) 384 385 /* 386 * Tracepoint for guest CR access. 387 */ 388 TRACE_EVENT(kvm_cr, 389 TP_PROTO(unsigned int rw, unsigned int cr, unsigned long val), 390 TP_ARGS(rw, cr, val), 391 392 TP_STRUCT__entry( 393 __field( unsigned int, rw ) 394 __field( unsigned int, cr ) 395 __field( unsigned long, val ) 396 ), 397 398 TP_fast_assign( 399 __entry->rw = rw; 400 __entry->cr = cr; 401 __entry->val = val; 402 ), 403 404 TP_printk("cr_%s %x = 0x%lx", 405 __entry->rw ? "write" : "read", 406 __entry->cr, __entry->val) 407 ); 408 409 #define trace_kvm_cr_read(cr, val) trace_kvm_cr(0, cr, val) 410 #define trace_kvm_cr_write(cr, val) trace_kvm_cr(1, cr, val) 411 412 TRACE_EVENT(kvm_pic_set_irq, 413 TP_PROTO(__u8 chip, __u8 pin, __u8 elcr, __u8 imr, bool coalesced), 414 TP_ARGS(chip, pin, elcr, imr, coalesced), 415 416 TP_STRUCT__entry( 417 __field( __u8, chip ) 418 __field( __u8, pin ) 419 __field( __u8, elcr ) 420 __field( __u8, imr ) 421 __field( bool, coalesced ) 422 ), 423 424 TP_fast_assign( 425 __entry->chip = chip; 426 __entry->pin = pin; 427 __entry->elcr = elcr; 428 __entry->imr = imr; 429 __entry->coalesced = coalesced; 430 ), 431 432 TP_printk("chip %u pin %u (%s%s)%s", 433 __entry->chip, __entry->pin, 434 (__entry->elcr & (1 << __entry->pin)) ? "level":"edge", 435 (__entry->imr & (1 << __entry->pin)) ? "|masked":"", 436 __entry->coalesced ? " (coalesced)" : "") 437 ); 438 439 #define kvm_apic_dst_shorthand \ 440 {0x0, "dst"}, \ 441 {0x1, "self"}, \ 442 {0x2, "all"}, \ 443 {0x3, "all-but-self"} 444 445 TRACE_EVENT(kvm_apic_ipi, 446 TP_PROTO(__u32 icr_low, __u32 dest_id), 447 TP_ARGS(icr_low, dest_id), 448 449 TP_STRUCT__entry( 450 __field( __u32, icr_low ) 451 __field( __u32, dest_id ) 452 ), 453 454 TP_fast_assign( 455 __entry->icr_low = icr_low; 456 __entry->dest_id = dest_id; 457 ), 458 459 TP_printk("dst %x vec %u (%s|%s|%s|%s|%s)", 460 __entry->dest_id, (u8)__entry->icr_low, 461 __print_symbolic((__entry->icr_low >> 8 & 0x7), 462 kvm_deliver_mode), 463 (__entry->icr_low & (1<<11)) ? "logical" : "physical", 464 (__entry->icr_low & (1<<14)) ? "assert" : "de-assert", 465 (__entry->icr_low & (1<<15)) ? "level" : "edge", 466 __print_symbolic((__entry->icr_low >> 18 & 0x3), 467 kvm_apic_dst_shorthand)) 468 ); 469 470 TRACE_EVENT(kvm_apic_accept_irq, 471 TP_PROTO(__u32 apicid, __u16 dm, __u16 tm, __u8 vec), 472 TP_ARGS(apicid, dm, tm, vec), 473 474 TP_STRUCT__entry( 475 __field( __u32, apicid ) 476 __field( __u16, dm ) 477 __field( __u16, tm ) 478 __field( __u8, vec ) 479 ), 480 481 TP_fast_assign( 482 __entry->apicid = apicid; 483 __entry->dm = dm; 484 __entry->tm = tm; 485 __entry->vec = vec; 486 ), 487 488 TP_printk("apicid %x vec %u (%s|%s)", 489 __entry->apicid, __entry->vec, 490 __print_symbolic((__entry->dm >> 8 & 0x7), kvm_deliver_mode), 491 __entry->tm ? "level" : "edge") 492 ); 493 494 TRACE_EVENT(kvm_eoi, 495 TP_PROTO(struct kvm_lapic *apic, int vector), 496 TP_ARGS(apic, vector), 497 498 TP_STRUCT__entry( 499 __field( __u32, apicid ) 500 __field( int, vector ) 501 ), 502 503 TP_fast_assign( 504 __entry->apicid = apic->vcpu->vcpu_id; 505 __entry->vector = vector; 506 ), 507 508 TP_printk("apicid %x vector %d", __entry->apicid, __entry->vector) 509 ); 510 511 TRACE_EVENT(kvm_pv_eoi, 512 TP_PROTO(struct kvm_lapic *apic, int vector), 513 TP_ARGS(apic, vector), 514 515 TP_STRUCT__entry( 516 __field( __u32, apicid ) 517 __field( int, vector ) 518 ), 519 520 TP_fast_assign( 521 __entry->apicid = apic->vcpu->vcpu_id; 522 __entry->vector = vector; 523 ), 524 525 TP_printk("apicid %x vector %d", __entry->apicid, __entry->vector) 526 ); 527 528 /* 529 * Tracepoint for nested VMRUN 530 */ 531 TRACE_EVENT(kvm_nested_vmrun, 532 TP_PROTO(__u64 rip, __u64 vmcb, __u64 nested_rip, __u32 int_ctl, 533 __u32 event_inj, bool npt), 534 TP_ARGS(rip, vmcb, nested_rip, int_ctl, event_inj, npt), 535 536 TP_STRUCT__entry( 537 __field( __u64, rip ) 538 __field( __u64, vmcb ) 539 __field( __u64, nested_rip ) 540 __field( __u32, int_ctl ) 541 __field( __u32, event_inj ) 542 __field( bool, npt ) 543 ), 544 545 TP_fast_assign( 546 __entry->rip = rip; 547 __entry->vmcb = vmcb; 548 __entry->nested_rip = nested_rip; 549 __entry->int_ctl = int_ctl; 550 __entry->event_inj = event_inj; 551 __entry->npt = npt; 552 ), 553 554 TP_printk("rip: 0x%016llx vmcb: 0x%016llx nrip: 0x%016llx int_ctl: 0x%08x " 555 "event_inj: 0x%08x npt: %s", 556 __entry->rip, __entry->vmcb, __entry->nested_rip, 557 __entry->int_ctl, __entry->event_inj, 558 __entry->npt ? "on" : "off") 559 ); 560 561 TRACE_EVENT(kvm_nested_intercepts, 562 TP_PROTO(__u16 cr_read, __u16 cr_write, __u32 exceptions, 563 __u32 intercept1, __u32 intercept2, __u32 intercept3), 564 TP_ARGS(cr_read, cr_write, exceptions, intercept1, 565 intercept2, intercept3), 566 567 TP_STRUCT__entry( 568 __field( __u16, cr_read ) 569 __field( __u16, cr_write ) 570 __field( __u32, exceptions ) 571 __field( __u32, intercept1 ) 572 __field( __u32, intercept2 ) 573 __field( __u32, intercept3 ) 574 ), 575 576 TP_fast_assign( 577 __entry->cr_read = cr_read; 578 __entry->cr_write = cr_write; 579 __entry->exceptions = exceptions; 580 __entry->intercept1 = intercept1; 581 __entry->intercept2 = intercept2; 582 __entry->intercept3 = intercept3; 583 ), 584 585 TP_printk("cr_read: %04x cr_write: %04x excp: %08x " 586 "intercepts: %08x %08x %08x", 587 __entry->cr_read, __entry->cr_write, __entry->exceptions, 588 __entry->intercept1, __entry->intercept2, __entry->intercept3) 589 ); 590 /* 591 * Tracepoint for #VMEXIT while nested 592 */ 593 TRACE_EVENT_KVM_EXIT(kvm_nested_vmexit); 594 595 /* 596 * Tracepoint for #VMEXIT reinjected to the guest 597 */ 598 TRACE_EVENT(kvm_nested_vmexit_inject, 599 TP_PROTO(__u32 exit_code, 600 __u64 exit_info1, __u64 exit_info2, 601 __u32 exit_int_info, __u32 exit_int_info_err, __u32 isa), 602 TP_ARGS(exit_code, exit_info1, exit_info2, 603 exit_int_info, exit_int_info_err, isa), 604 605 TP_STRUCT__entry( 606 __field( __u32, exit_code ) 607 __field( __u64, exit_info1 ) 608 __field( __u64, exit_info2 ) 609 __field( __u32, exit_int_info ) 610 __field( __u32, exit_int_info_err ) 611 __field( __u32, isa ) 612 ), 613 614 TP_fast_assign( 615 __entry->exit_code = exit_code; 616 __entry->exit_info1 = exit_info1; 617 __entry->exit_info2 = exit_info2; 618 __entry->exit_int_info = exit_int_info; 619 __entry->exit_int_info_err = exit_int_info_err; 620 __entry->isa = isa; 621 ), 622 623 TP_printk("reason: %s%s%s ext_inf1: 0x%016llx " 624 "ext_inf2: 0x%016llx ext_int: 0x%08x ext_int_err: 0x%08x", 625 kvm_print_exit_reason(__entry->exit_code, __entry->isa), 626 __entry->exit_info1, __entry->exit_info2, 627 __entry->exit_int_info, __entry->exit_int_info_err) 628 ); 629 630 /* 631 * Tracepoint for nested #vmexit because of interrupt pending 632 */ 633 TRACE_EVENT(kvm_nested_intr_vmexit, 634 TP_PROTO(__u64 rip), 635 TP_ARGS(rip), 636 637 TP_STRUCT__entry( 638 __field( __u64, rip ) 639 ), 640 641 TP_fast_assign( 642 __entry->rip = rip 643 ), 644 645 TP_printk("rip: 0x%016llx", __entry->rip) 646 ); 647 648 /* 649 * Tracepoint for nested #vmexit because of interrupt pending 650 */ 651 TRACE_EVENT(kvm_invlpga, 652 TP_PROTO(__u64 rip, int asid, u64 address), 653 TP_ARGS(rip, asid, address), 654 655 TP_STRUCT__entry( 656 __field( __u64, rip ) 657 __field( int, asid ) 658 __field( __u64, address ) 659 ), 660 661 TP_fast_assign( 662 __entry->rip = rip; 663 __entry->asid = asid; 664 __entry->address = address; 665 ), 666 667 TP_printk("rip: 0x%016llx asid: %d address: 0x%016llx", 668 __entry->rip, __entry->asid, __entry->address) 669 ); 670 671 /* 672 * Tracepoint for nested #vmexit because of interrupt pending 673 */ 674 TRACE_EVENT(kvm_skinit, 675 TP_PROTO(__u64 rip, __u32 slb), 676 TP_ARGS(rip, slb), 677 678 TP_STRUCT__entry( 679 __field( __u64, rip ) 680 __field( __u32, slb ) 681 ), 682 683 TP_fast_assign( 684 __entry->rip = rip; 685 __entry->slb = slb; 686 ), 687 688 TP_printk("rip: 0x%016llx slb: 0x%08x", 689 __entry->rip, __entry->slb) 690 ); 691 692 #define KVM_EMUL_INSN_F_CR0_PE (1 << 0) 693 #define KVM_EMUL_INSN_F_EFL_VM (1 << 1) 694 #define KVM_EMUL_INSN_F_CS_D (1 << 2) 695 #define KVM_EMUL_INSN_F_CS_L (1 << 3) 696 697 #define kvm_trace_symbol_emul_flags \ 698 { 0, "real" }, \ 699 { KVM_EMUL_INSN_F_CR0_PE \ 700 | KVM_EMUL_INSN_F_EFL_VM, "vm16" }, \ 701 { KVM_EMUL_INSN_F_CR0_PE, "prot16" }, \ 702 { KVM_EMUL_INSN_F_CR0_PE \ 703 | KVM_EMUL_INSN_F_CS_D, "prot32" }, \ 704 { KVM_EMUL_INSN_F_CR0_PE \ 705 | KVM_EMUL_INSN_F_CS_L, "prot64" } 706 707 #define kei_decode_mode(mode) ({ \ 708 u8 flags = 0xff; \ 709 switch (mode) { \ 710 case X86EMUL_MODE_REAL: \ 711 flags = 0; \ 712 break; \ 713 case X86EMUL_MODE_VM86: \ 714 flags = KVM_EMUL_INSN_F_EFL_VM; \ 715 break; \ 716 case X86EMUL_MODE_PROT16: \ 717 flags = KVM_EMUL_INSN_F_CR0_PE; \ 718 break; \ 719 case X86EMUL_MODE_PROT32: \ 720 flags = KVM_EMUL_INSN_F_CR0_PE \ 721 | KVM_EMUL_INSN_F_CS_D; \ 722 break; \ 723 case X86EMUL_MODE_PROT64: \ 724 flags = KVM_EMUL_INSN_F_CR0_PE \ 725 | KVM_EMUL_INSN_F_CS_L; \ 726 break; \ 727 } \ 728 flags; \ 729 }) 730 731 TRACE_EVENT(kvm_emulate_insn, 732 TP_PROTO(struct kvm_vcpu *vcpu, __u8 failed), 733 TP_ARGS(vcpu, failed), 734 735 TP_STRUCT__entry( 736 __field( __u64, rip ) 737 __field( __u32, csbase ) 738 __field( __u8, len ) 739 __array( __u8, insn, 15 ) 740 __field( __u8, flags ) 741 __field( __u8, failed ) 742 ), 743 744 TP_fast_assign( 745 __entry->csbase = kvm_x86_ops.get_segment_base(vcpu, VCPU_SREG_CS); 746 __entry->len = vcpu->arch.emulate_ctxt->fetch.ptr 747 - vcpu->arch.emulate_ctxt->fetch.data; 748 __entry->rip = vcpu->arch.emulate_ctxt->_eip - __entry->len; 749 memcpy(__entry->insn, 750 vcpu->arch.emulate_ctxt->fetch.data, 751 15); 752 __entry->flags = kei_decode_mode(vcpu->arch.emulate_ctxt->mode); 753 __entry->failed = failed; 754 ), 755 756 TP_printk("%x:%llx:%s (%s)%s", 757 __entry->csbase, __entry->rip, 758 __print_hex(__entry->insn, __entry->len), 759 __print_symbolic(__entry->flags, 760 kvm_trace_symbol_emul_flags), 761 __entry->failed ? " failed" : "" 762 ) 763 ); 764 765 #define trace_kvm_emulate_insn_start(vcpu) trace_kvm_emulate_insn(vcpu, 0) 766 #define trace_kvm_emulate_insn_failed(vcpu) trace_kvm_emulate_insn(vcpu, 1) 767 768 TRACE_EVENT( 769 vcpu_match_mmio, 770 TP_PROTO(gva_t gva, gpa_t gpa, bool write, bool gpa_match), 771 TP_ARGS(gva, gpa, write, gpa_match), 772 773 TP_STRUCT__entry( 774 __field(gva_t, gva) 775 __field(gpa_t, gpa) 776 __field(bool, write) 777 __field(bool, gpa_match) 778 ), 779 780 TP_fast_assign( 781 __entry->gva = gva; 782 __entry->gpa = gpa; 783 __entry->write = write; 784 __entry->gpa_match = gpa_match 785 ), 786 787 TP_printk("gva %#lx gpa %#llx %s %s", __entry->gva, __entry->gpa, 788 __entry->write ? "Write" : "Read", 789 __entry->gpa_match ? "GPA" : "GVA") 790 ); 791 792 TRACE_EVENT(kvm_write_tsc_offset, 793 TP_PROTO(unsigned int vcpu_id, __u64 previous_tsc_offset, 794 __u64 next_tsc_offset), 795 TP_ARGS(vcpu_id, previous_tsc_offset, next_tsc_offset), 796 797 TP_STRUCT__entry( 798 __field( unsigned int, vcpu_id ) 799 __field( __u64, previous_tsc_offset ) 800 __field( __u64, next_tsc_offset ) 801 ), 802 803 TP_fast_assign( 804 __entry->vcpu_id = vcpu_id; 805 __entry->previous_tsc_offset = previous_tsc_offset; 806 __entry->next_tsc_offset = next_tsc_offset; 807 ), 808 809 TP_printk("vcpu=%u prev=%llu next=%llu", __entry->vcpu_id, 810 __entry->previous_tsc_offset, __entry->next_tsc_offset) 811 ); 812 813 #ifdef CONFIG_X86_64 814 815 #define host_clocks \ 816 {VDSO_CLOCKMODE_NONE, "none"}, \ 817 {VDSO_CLOCKMODE_TSC, "tsc"} \ 818 819 TRACE_EVENT(kvm_update_master_clock, 820 TP_PROTO(bool use_master_clock, unsigned int host_clock, bool offset_matched), 821 TP_ARGS(use_master_clock, host_clock, offset_matched), 822 823 TP_STRUCT__entry( 824 __field( bool, use_master_clock ) 825 __field( unsigned int, host_clock ) 826 __field( bool, offset_matched ) 827 ), 828 829 TP_fast_assign( 830 __entry->use_master_clock = use_master_clock; 831 __entry->host_clock = host_clock; 832 __entry->offset_matched = offset_matched; 833 ), 834 835 TP_printk("masterclock %d hostclock %s offsetmatched %u", 836 __entry->use_master_clock, 837 __print_symbolic(__entry->host_clock, host_clocks), 838 __entry->offset_matched) 839 ); 840 841 TRACE_EVENT(kvm_track_tsc, 842 TP_PROTO(unsigned int vcpu_id, unsigned int nr_matched, 843 unsigned int online_vcpus, bool use_master_clock, 844 unsigned int host_clock), 845 TP_ARGS(vcpu_id, nr_matched, online_vcpus, use_master_clock, 846 host_clock), 847 848 TP_STRUCT__entry( 849 __field( unsigned int, vcpu_id ) 850 __field( unsigned int, nr_vcpus_matched_tsc ) 851 __field( unsigned int, online_vcpus ) 852 __field( bool, use_master_clock ) 853 __field( unsigned int, host_clock ) 854 ), 855 856 TP_fast_assign( 857 __entry->vcpu_id = vcpu_id; 858 __entry->nr_vcpus_matched_tsc = nr_matched; 859 __entry->online_vcpus = online_vcpus; 860 __entry->use_master_clock = use_master_clock; 861 __entry->host_clock = host_clock; 862 ), 863 864 TP_printk("vcpu_id %u masterclock %u offsetmatched %u nr_online %u" 865 " hostclock %s", 866 __entry->vcpu_id, __entry->use_master_clock, 867 __entry->nr_vcpus_matched_tsc, __entry->online_vcpus, 868 __print_symbolic(__entry->host_clock, host_clocks)) 869 ); 870 871 #endif /* CONFIG_X86_64 */ 872 873 /* 874 * Tracepoint for PML full VMEXIT. 875 */ 876 TRACE_EVENT(kvm_pml_full, 877 TP_PROTO(unsigned int vcpu_id), 878 TP_ARGS(vcpu_id), 879 880 TP_STRUCT__entry( 881 __field( unsigned int, vcpu_id ) 882 ), 883 884 TP_fast_assign( 885 __entry->vcpu_id = vcpu_id; 886 ), 887 888 TP_printk("vcpu %d: PML full", __entry->vcpu_id) 889 ); 890 891 TRACE_EVENT(kvm_ple_window_update, 892 TP_PROTO(unsigned int vcpu_id, unsigned int new, unsigned int old), 893 TP_ARGS(vcpu_id, new, old), 894 895 TP_STRUCT__entry( 896 __field( unsigned int, vcpu_id ) 897 __field( unsigned int, new ) 898 __field( unsigned int, old ) 899 ), 900 901 TP_fast_assign( 902 __entry->vcpu_id = vcpu_id; 903 __entry->new = new; 904 __entry->old = old; 905 ), 906 907 TP_printk("vcpu %u old %u new %u (%s)", 908 __entry->vcpu_id, __entry->old, __entry->new, 909 __entry->old < __entry->new ? "growed" : "shrinked") 910 ); 911 912 TRACE_EVENT(kvm_pvclock_update, 913 TP_PROTO(unsigned int vcpu_id, struct pvclock_vcpu_time_info *pvclock), 914 TP_ARGS(vcpu_id, pvclock), 915 916 TP_STRUCT__entry( 917 __field( unsigned int, vcpu_id ) 918 __field( __u32, version ) 919 __field( __u64, tsc_timestamp ) 920 __field( __u64, system_time ) 921 __field( __u32, tsc_to_system_mul ) 922 __field( __s8, tsc_shift ) 923 __field( __u8, flags ) 924 ), 925 926 TP_fast_assign( 927 __entry->vcpu_id = vcpu_id; 928 __entry->version = pvclock->version; 929 __entry->tsc_timestamp = pvclock->tsc_timestamp; 930 __entry->system_time = pvclock->system_time; 931 __entry->tsc_to_system_mul = pvclock->tsc_to_system_mul; 932 __entry->tsc_shift = pvclock->tsc_shift; 933 __entry->flags = pvclock->flags; 934 ), 935 936 TP_printk("vcpu_id %u, pvclock { version %u, tsc_timestamp 0x%llx, " 937 "system_time 0x%llx, tsc_to_system_mul 0x%x, tsc_shift %d, " 938 "flags 0x%x }", 939 __entry->vcpu_id, 940 __entry->version, 941 __entry->tsc_timestamp, 942 __entry->system_time, 943 __entry->tsc_to_system_mul, 944 __entry->tsc_shift, 945 __entry->flags) 946 ); 947 948 TRACE_EVENT(kvm_wait_lapic_expire, 949 TP_PROTO(unsigned int vcpu_id, s64 delta), 950 TP_ARGS(vcpu_id, delta), 951 952 TP_STRUCT__entry( 953 __field( unsigned int, vcpu_id ) 954 __field( s64, delta ) 955 ), 956 957 TP_fast_assign( 958 __entry->vcpu_id = vcpu_id; 959 __entry->delta = delta; 960 ), 961 962 TP_printk("vcpu %u: delta %lld (%s)", 963 __entry->vcpu_id, 964 __entry->delta, 965 __entry->delta < 0 ? "early" : "late") 966 ); 967 968 TRACE_EVENT(kvm_enter_smm, 969 TP_PROTO(unsigned int vcpu_id, u64 smbase, bool entering), 970 TP_ARGS(vcpu_id, smbase, entering), 971 972 TP_STRUCT__entry( 973 __field( unsigned int, vcpu_id ) 974 __field( u64, smbase ) 975 __field( bool, entering ) 976 ), 977 978 TP_fast_assign( 979 __entry->vcpu_id = vcpu_id; 980 __entry->smbase = smbase; 981 __entry->entering = entering; 982 ), 983 984 TP_printk("vcpu %u: %s SMM, smbase 0x%llx", 985 __entry->vcpu_id, 986 __entry->entering ? "entering" : "leaving", 987 __entry->smbase) 988 ); 989 990 /* 991 * Tracepoint for VT-d posted-interrupts. 992 */ 993 TRACE_EVENT(kvm_pi_irte_update, 994 TP_PROTO(unsigned int host_irq, unsigned int vcpu_id, 995 unsigned int gsi, unsigned int gvec, 996 u64 pi_desc_addr, bool set), 997 TP_ARGS(host_irq, vcpu_id, gsi, gvec, pi_desc_addr, set), 998 999 TP_STRUCT__entry( 1000 __field( unsigned int, host_irq ) 1001 __field( unsigned int, vcpu_id ) 1002 __field( unsigned int, gsi ) 1003 __field( unsigned int, gvec ) 1004 __field( u64, pi_desc_addr ) 1005 __field( bool, set ) 1006 ), 1007 1008 TP_fast_assign( 1009 __entry->host_irq = host_irq; 1010 __entry->vcpu_id = vcpu_id; 1011 __entry->gsi = gsi; 1012 __entry->gvec = gvec; 1013 __entry->pi_desc_addr = pi_desc_addr; 1014 __entry->set = set; 1015 ), 1016 1017 TP_printk("VT-d PI is %s for irq %u, vcpu %u, gsi: 0x%x, " 1018 "gvec: 0x%x, pi_desc_addr: 0x%llx", 1019 __entry->set ? "enabled and being updated" : "disabled", 1020 __entry->host_irq, 1021 __entry->vcpu_id, 1022 __entry->gsi, 1023 __entry->gvec, 1024 __entry->pi_desc_addr) 1025 ); 1026 1027 /* 1028 * Tracepoint for kvm_hv_notify_acked_sint. 1029 */ 1030 TRACE_EVENT(kvm_hv_notify_acked_sint, 1031 TP_PROTO(int vcpu_id, u32 sint), 1032 TP_ARGS(vcpu_id, sint), 1033 1034 TP_STRUCT__entry( 1035 __field(int, vcpu_id) 1036 __field(u32, sint) 1037 ), 1038 1039 TP_fast_assign( 1040 __entry->vcpu_id = vcpu_id; 1041 __entry->sint = sint; 1042 ), 1043 1044 TP_printk("vcpu_id %d sint %u", __entry->vcpu_id, __entry->sint) 1045 ); 1046 1047 /* 1048 * Tracepoint for synic_set_irq. 1049 */ 1050 TRACE_EVENT(kvm_hv_synic_set_irq, 1051 TP_PROTO(int vcpu_id, u32 sint, int vector, int ret), 1052 TP_ARGS(vcpu_id, sint, vector, ret), 1053 1054 TP_STRUCT__entry( 1055 __field(int, vcpu_id) 1056 __field(u32, sint) 1057 __field(int, vector) 1058 __field(int, ret) 1059 ), 1060 1061 TP_fast_assign( 1062 __entry->vcpu_id = vcpu_id; 1063 __entry->sint = sint; 1064 __entry->vector = vector; 1065 __entry->ret = ret; 1066 ), 1067 1068 TP_printk("vcpu_id %d sint %u vector %d ret %d", 1069 __entry->vcpu_id, __entry->sint, __entry->vector, 1070 __entry->ret) 1071 ); 1072 1073 /* 1074 * Tracepoint for kvm_hv_synic_send_eoi. 1075 */ 1076 TRACE_EVENT(kvm_hv_synic_send_eoi, 1077 TP_PROTO(int vcpu_id, int vector), 1078 TP_ARGS(vcpu_id, vector), 1079 1080 TP_STRUCT__entry( 1081 __field(int, vcpu_id) 1082 __field(u32, sint) 1083 __field(int, vector) 1084 __field(int, ret) 1085 ), 1086 1087 TP_fast_assign( 1088 __entry->vcpu_id = vcpu_id; 1089 __entry->vector = vector; 1090 ), 1091 1092 TP_printk("vcpu_id %d vector %d", __entry->vcpu_id, __entry->vector) 1093 ); 1094 1095 /* 1096 * Tracepoint for synic_set_msr. 1097 */ 1098 TRACE_EVENT(kvm_hv_synic_set_msr, 1099 TP_PROTO(int vcpu_id, u32 msr, u64 data, bool host), 1100 TP_ARGS(vcpu_id, msr, data, host), 1101 1102 TP_STRUCT__entry( 1103 __field(int, vcpu_id) 1104 __field(u32, msr) 1105 __field(u64, data) 1106 __field(bool, host) 1107 ), 1108 1109 TP_fast_assign( 1110 __entry->vcpu_id = vcpu_id; 1111 __entry->msr = msr; 1112 __entry->data = data; 1113 __entry->host = host 1114 ), 1115 1116 TP_printk("vcpu_id %d msr 0x%x data 0x%llx host %d", 1117 __entry->vcpu_id, __entry->msr, __entry->data, __entry->host) 1118 ); 1119 1120 /* 1121 * Tracepoint for stimer_set_config. 1122 */ 1123 TRACE_EVENT(kvm_hv_stimer_set_config, 1124 TP_PROTO(int vcpu_id, int timer_index, u64 config, bool host), 1125 TP_ARGS(vcpu_id, timer_index, config, host), 1126 1127 TP_STRUCT__entry( 1128 __field(int, vcpu_id) 1129 __field(int, timer_index) 1130 __field(u64, config) 1131 __field(bool, host) 1132 ), 1133 1134 TP_fast_assign( 1135 __entry->vcpu_id = vcpu_id; 1136 __entry->timer_index = timer_index; 1137 __entry->config = config; 1138 __entry->host = host; 1139 ), 1140 1141 TP_printk("vcpu_id %d timer %d config 0x%llx host %d", 1142 __entry->vcpu_id, __entry->timer_index, __entry->config, 1143 __entry->host) 1144 ); 1145 1146 /* 1147 * Tracepoint for stimer_set_count. 1148 */ 1149 TRACE_EVENT(kvm_hv_stimer_set_count, 1150 TP_PROTO(int vcpu_id, int timer_index, u64 count, bool host), 1151 TP_ARGS(vcpu_id, timer_index, count, host), 1152 1153 TP_STRUCT__entry( 1154 __field(int, vcpu_id) 1155 __field(int, timer_index) 1156 __field(u64, count) 1157 __field(bool, host) 1158 ), 1159 1160 TP_fast_assign( 1161 __entry->vcpu_id = vcpu_id; 1162 __entry->timer_index = timer_index; 1163 __entry->count = count; 1164 __entry->host = host; 1165 ), 1166 1167 TP_printk("vcpu_id %d timer %d count %llu host %d", 1168 __entry->vcpu_id, __entry->timer_index, __entry->count, 1169 __entry->host) 1170 ); 1171 1172 /* 1173 * Tracepoint for stimer_start(periodic timer case). 1174 */ 1175 TRACE_EVENT(kvm_hv_stimer_start_periodic, 1176 TP_PROTO(int vcpu_id, int timer_index, u64 time_now, u64 exp_time), 1177 TP_ARGS(vcpu_id, timer_index, time_now, exp_time), 1178 1179 TP_STRUCT__entry( 1180 __field(int, vcpu_id) 1181 __field(int, timer_index) 1182 __field(u64, time_now) 1183 __field(u64, exp_time) 1184 ), 1185 1186 TP_fast_assign( 1187 __entry->vcpu_id = vcpu_id; 1188 __entry->timer_index = timer_index; 1189 __entry->time_now = time_now; 1190 __entry->exp_time = exp_time; 1191 ), 1192 1193 TP_printk("vcpu_id %d timer %d time_now %llu exp_time %llu", 1194 __entry->vcpu_id, __entry->timer_index, __entry->time_now, 1195 __entry->exp_time) 1196 ); 1197 1198 /* 1199 * Tracepoint for stimer_start(one-shot timer case). 1200 */ 1201 TRACE_EVENT(kvm_hv_stimer_start_one_shot, 1202 TP_PROTO(int vcpu_id, int timer_index, u64 time_now, u64 count), 1203 TP_ARGS(vcpu_id, timer_index, time_now, count), 1204 1205 TP_STRUCT__entry( 1206 __field(int, vcpu_id) 1207 __field(int, timer_index) 1208 __field(u64, time_now) 1209 __field(u64, count) 1210 ), 1211 1212 TP_fast_assign( 1213 __entry->vcpu_id = vcpu_id; 1214 __entry->timer_index = timer_index; 1215 __entry->time_now = time_now; 1216 __entry->count = count; 1217 ), 1218 1219 TP_printk("vcpu_id %d timer %d time_now %llu count %llu", 1220 __entry->vcpu_id, __entry->timer_index, __entry->time_now, 1221 __entry->count) 1222 ); 1223 1224 /* 1225 * Tracepoint for stimer_timer_callback. 1226 */ 1227 TRACE_EVENT(kvm_hv_stimer_callback, 1228 TP_PROTO(int vcpu_id, int timer_index), 1229 TP_ARGS(vcpu_id, timer_index), 1230 1231 TP_STRUCT__entry( 1232 __field(int, vcpu_id) 1233 __field(int, timer_index) 1234 ), 1235 1236 TP_fast_assign( 1237 __entry->vcpu_id = vcpu_id; 1238 __entry->timer_index = timer_index; 1239 ), 1240 1241 TP_printk("vcpu_id %d timer %d", 1242 __entry->vcpu_id, __entry->timer_index) 1243 ); 1244 1245 /* 1246 * Tracepoint for stimer_expiration. 1247 */ 1248 TRACE_EVENT(kvm_hv_stimer_expiration, 1249 TP_PROTO(int vcpu_id, int timer_index, int direct, int msg_send_result), 1250 TP_ARGS(vcpu_id, timer_index, direct, msg_send_result), 1251 1252 TP_STRUCT__entry( 1253 __field(int, vcpu_id) 1254 __field(int, timer_index) 1255 __field(int, direct) 1256 __field(int, msg_send_result) 1257 ), 1258 1259 TP_fast_assign( 1260 __entry->vcpu_id = vcpu_id; 1261 __entry->timer_index = timer_index; 1262 __entry->direct = direct; 1263 __entry->msg_send_result = msg_send_result; 1264 ), 1265 1266 TP_printk("vcpu_id %d timer %d direct %d send result %d", 1267 __entry->vcpu_id, __entry->timer_index, 1268 __entry->direct, __entry->msg_send_result) 1269 ); 1270 1271 /* 1272 * Tracepoint for stimer_cleanup. 1273 */ 1274 TRACE_EVENT(kvm_hv_stimer_cleanup, 1275 TP_PROTO(int vcpu_id, int timer_index), 1276 TP_ARGS(vcpu_id, timer_index), 1277 1278 TP_STRUCT__entry( 1279 __field(int, vcpu_id) 1280 __field(int, timer_index) 1281 ), 1282 1283 TP_fast_assign( 1284 __entry->vcpu_id = vcpu_id; 1285 __entry->timer_index = timer_index; 1286 ), 1287 1288 TP_printk("vcpu_id %d timer %d", 1289 __entry->vcpu_id, __entry->timer_index) 1290 ); 1291 1292 TRACE_EVENT(kvm_apicv_update_request, 1293 TP_PROTO(bool activate, unsigned long bit), 1294 TP_ARGS(activate, bit), 1295 1296 TP_STRUCT__entry( 1297 __field(bool, activate) 1298 __field(unsigned long, bit) 1299 ), 1300 1301 TP_fast_assign( 1302 __entry->activate = activate; 1303 __entry->bit = bit; 1304 ), 1305 1306 TP_printk("%s bit=%lu", 1307 __entry->activate ? "activate" : "deactivate", 1308 __entry->bit) 1309 ); 1310 1311 /* 1312 * Tracepoint for AMD AVIC 1313 */ 1314 TRACE_EVENT(kvm_avic_incomplete_ipi, 1315 TP_PROTO(u32 vcpu, u32 icrh, u32 icrl, u32 id, u32 index), 1316 TP_ARGS(vcpu, icrh, icrl, id, index), 1317 1318 TP_STRUCT__entry( 1319 __field(u32, vcpu) 1320 __field(u32, icrh) 1321 __field(u32, icrl) 1322 __field(u32, id) 1323 __field(u32, index) 1324 ), 1325 1326 TP_fast_assign( 1327 __entry->vcpu = vcpu; 1328 __entry->icrh = icrh; 1329 __entry->icrl = icrl; 1330 __entry->id = id; 1331 __entry->index = index; 1332 ), 1333 1334 TP_printk("vcpu=%u, icrh:icrl=%#010x:%08x, id=%u, index=%u", 1335 __entry->vcpu, __entry->icrh, __entry->icrl, 1336 __entry->id, __entry->index) 1337 ); 1338 1339 TRACE_EVENT(kvm_avic_unaccelerated_access, 1340 TP_PROTO(u32 vcpu, u32 offset, bool ft, bool rw, u32 vec), 1341 TP_ARGS(vcpu, offset, ft, rw, vec), 1342 1343 TP_STRUCT__entry( 1344 __field(u32, vcpu) 1345 __field(u32, offset) 1346 __field(bool, ft) 1347 __field(bool, rw) 1348 __field(u32, vec) 1349 ), 1350 1351 TP_fast_assign( 1352 __entry->vcpu = vcpu; 1353 __entry->offset = offset; 1354 __entry->ft = ft; 1355 __entry->rw = rw; 1356 __entry->vec = vec; 1357 ), 1358 1359 TP_printk("vcpu=%u, offset=%#x(%s), %s, %s, vec=%#x", 1360 __entry->vcpu, 1361 __entry->offset, 1362 __print_symbolic(__entry->offset, kvm_trace_symbol_apic), 1363 __entry->ft ? "trap" : "fault", 1364 __entry->rw ? "write" : "read", 1365 __entry->vec) 1366 ); 1367 1368 TRACE_EVENT(kvm_avic_ga_log, 1369 TP_PROTO(u32 vmid, u32 vcpuid), 1370 TP_ARGS(vmid, vcpuid), 1371 1372 TP_STRUCT__entry( 1373 __field(u32, vmid) 1374 __field(u32, vcpuid) 1375 ), 1376 1377 TP_fast_assign( 1378 __entry->vmid = vmid; 1379 __entry->vcpuid = vcpuid; 1380 ), 1381 1382 TP_printk("vmid=%u, vcpuid=%u", 1383 __entry->vmid, __entry->vcpuid) 1384 ); 1385 1386 TRACE_EVENT(kvm_hv_timer_state, 1387 TP_PROTO(unsigned int vcpu_id, unsigned int hv_timer_in_use), 1388 TP_ARGS(vcpu_id, hv_timer_in_use), 1389 TP_STRUCT__entry( 1390 __field(unsigned int, vcpu_id) 1391 __field(unsigned int, hv_timer_in_use) 1392 ), 1393 TP_fast_assign( 1394 __entry->vcpu_id = vcpu_id; 1395 __entry->hv_timer_in_use = hv_timer_in_use; 1396 ), 1397 TP_printk("vcpu_id %x hv_timer %x", 1398 __entry->vcpu_id, 1399 __entry->hv_timer_in_use) 1400 ); 1401 1402 /* 1403 * Tracepoint for kvm_hv_flush_tlb. 1404 */ 1405 TRACE_EVENT(kvm_hv_flush_tlb, 1406 TP_PROTO(u64 processor_mask, u64 address_space, u64 flags), 1407 TP_ARGS(processor_mask, address_space, flags), 1408 1409 TP_STRUCT__entry( 1410 __field(u64, processor_mask) 1411 __field(u64, address_space) 1412 __field(u64, flags) 1413 ), 1414 1415 TP_fast_assign( 1416 __entry->processor_mask = processor_mask; 1417 __entry->address_space = address_space; 1418 __entry->flags = flags; 1419 ), 1420 1421 TP_printk("processor_mask 0x%llx address_space 0x%llx flags 0x%llx", 1422 __entry->processor_mask, __entry->address_space, 1423 __entry->flags) 1424 ); 1425 1426 /* 1427 * Tracepoint for kvm_hv_flush_tlb_ex. 1428 */ 1429 TRACE_EVENT(kvm_hv_flush_tlb_ex, 1430 TP_PROTO(u64 valid_bank_mask, u64 format, u64 address_space, u64 flags), 1431 TP_ARGS(valid_bank_mask, format, address_space, flags), 1432 1433 TP_STRUCT__entry( 1434 __field(u64, valid_bank_mask) 1435 __field(u64, format) 1436 __field(u64, address_space) 1437 __field(u64, flags) 1438 ), 1439 1440 TP_fast_assign( 1441 __entry->valid_bank_mask = valid_bank_mask; 1442 __entry->format = format; 1443 __entry->address_space = address_space; 1444 __entry->flags = flags; 1445 ), 1446 1447 TP_printk("valid_bank_mask 0x%llx format 0x%llx " 1448 "address_space 0x%llx flags 0x%llx", 1449 __entry->valid_bank_mask, __entry->format, 1450 __entry->address_space, __entry->flags) 1451 ); 1452 1453 /* 1454 * Tracepoints for kvm_hv_send_ipi. 1455 */ 1456 TRACE_EVENT(kvm_hv_send_ipi, 1457 TP_PROTO(u32 vector, u64 processor_mask), 1458 TP_ARGS(vector, processor_mask), 1459 1460 TP_STRUCT__entry( 1461 __field(u32, vector) 1462 __field(u64, processor_mask) 1463 ), 1464 1465 TP_fast_assign( 1466 __entry->vector = vector; 1467 __entry->processor_mask = processor_mask; 1468 ), 1469 1470 TP_printk("vector %x processor_mask 0x%llx", 1471 __entry->vector, __entry->processor_mask) 1472 ); 1473 1474 TRACE_EVENT(kvm_hv_send_ipi_ex, 1475 TP_PROTO(u32 vector, u64 format, u64 valid_bank_mask), 1476 TP_ARGS(vector, format, valid_bank_mask), 1477 1478 TP_STRUCT__entry( 1479 __field(u32, vector) 1480 __field(u64, format) 1481 __field(u64, valid_bank_mask) 1482 ), 1483 1484 TP_fast_assign( 1485 __entry->vector = vector; 1486 __entry->format = format; 1487 __entry->valid_bank_mask = valid_bank_mask; 1488 ), 1489 1490 TP_printk("vector %x format %llx valid_bank_mask 0x%llx", 1491 __entry->vector, __entry->format, 1492 __entry->valid_bank_mask) 1493 ); 1494 1495 TRACE_EVENT(kvm_pv_tlb_flush, 1496 TP_PROTO(unsigned int vcpu_id, bool need_flush_tlb), 1497 TP_ARGS(vcpu_id, need_flush_tlb), 1498 1499 TP_STRUCT__entry( 1500 __field( unsigned int, vcpu_id ) 1501 __field( bool, need_flush_tlb ) 1502 ), 1503 1504 TP_fast_assign( 1505 __entry->vcpu_id = vcpu_id; 1506 __entry->need_flush_tlb = need_flush_tlb; 1507 ), 1508 1509 TP_printk("vcpu %u need_flush_tlb %s", __entry->vcpu_id, 1510 __entry->need_flush_tlb ? "true" : "false") 1511 ); 1512 1513 /* 1514 * Tracepoint for failed nested VMX VM-Enter. 1515 */ 1516 TRACE_EVENT(kvm_nested_vmenter_failed, 1517 TP_PROTO(const char *msg, u32 err), 1518 TP_ARGS(msg, err), 1519 1520 TP_STRUCT__entry( 1521 __string(msg, msg) 1522 __field(u32, err) 1523 ), 1524 1525 TP_fast_assign( 1526 __assign_str(msg, msg); 1527 __entry->err = err; 1528 ), 1529 1530 TP_printk("%s%s", __get_str(msg), !__entry->err ? "" : 1531 __print_symbolic(__entry->err, VMX_VMENTER_INSTRUCTION_ERRORS)) 1532 ); 1533 1534 /* 1535 * Tracepoint for syndbg_set_msr. 1536 */ 1537 TRACE_EVENT(kvm_hv_syndbg_set_msr, 1538 TP_PROTO(int vcpu_id, u32 vp_index, u32 msr, u64 data), 1539 TP_ARGS(vcpu_id, vp_index, msr, data), 1540 1541 TP_STRUCT__entry( 1542 __field(int, vcpu_id) 1543 __field(u32, vp_index) 1544 __field(u32, msr) 1545 __field(u64, data) 1546 ), 1547 1548 TP_fast_assign( 1549 __entry->vcpu_id = vcpu_id; 1550 __entry->vp_index = vp_index; 1551 __entry->msr = msr; 1552 __entry->data = data; 1553 ), 1554 1555 TP_printk("vcpu_id %d vp_index %u msr 0x%x data 0x%llx", 1556 __entry->vcpu_id, __entry->vp_index, __entry->msr, 1557 __entry->data) 1558 ); 1559 1560 /* 1561 * Tracepoint for syndbg_get_msr. 1562 */ 1563 TRACE_EVENT(kvm_hv_syndbg_get_msr, 1564 TP_PROTO(int vcpu_id, u32 vp_index, u32 msr, u64 data), 1565 TP_ARGS(vcpu_id, vp_index, msr, data), 1566 1567 TP_STRUCT__entry( 1568 __field(int, vcpu_id) 1569 __field(u32, vp_index) 1570 __field(u32, msr) 1571 __field(u64, data) 1572 ), 1573 1574 TP_fast_assign( 1575 __entry->vcpu_id = vcpu_id; 1576 __entry->vp_index = vp_index; 1577 __entry->msr = msr; 1578 __entry->data = data; 1579 ), 1580 1581 TP_printk("vcpu_id %d vp_index %u msr 0x%x data 0x%llx", 1582 __entry->vcpu_id, __entry->vp_index, __entry->msr, 1583 __entry->data) 1584 ); 1585 #endif /* _TRACE_KVM_H */ 1586 1587 #undef TRACE_INCLUDE_PATH 1588 #define TRACE_INCLUDE_PATH ../../arch/x86/kvm 1589 #undef TRACE_INCLUDE_FILE 1590 #define TRACE_INCLUDE_FILE trace 1591 1592 /* This part must be outside protection */ 1593 #include <trace/define_trace.h> 1594