/tools/perf/arch/x86/util/ |
D | kvm-stat.c | 34 key->key = perf_evsel__intval(evsel, sample, "gpa"); in mmio_event_get_key() 35 key->info = perf_evsel__intval(evsel, sample, "type"); in mmio_event_get_key() 51 perf_evsel__intval(evsel, sample, "type") == KVM_TRACE_MMIO_WRITE) { in mmio_event_begin() 68 perf_evsel__intval(evsel, sample, "type") == KVM_TRACE_MMIO_READ) { in mmio_event_end() 97 key->key = perf_evsel__intval(evsel, sample, "port"); in ioport_event_get_key() 98 key->info = perf_evsel__intval(evsel, sample, "rw"); in ioport_event_get_key()
|
/tools/perf/arch/s390/util/ |
D | kvm-stat.c | 33 insn = perf_evsel__intval(evsel, sample, "instruction"); in event_icpt_insn_get_key() 42 key->key = perf_evsel__intval(evsel, sample, "order_code"); in event_sigp_get_key() 50 key->key = perf_evsel__intval(evsel, sample, "code"); in event_diag_get_key() 58 key->key = perf_evsel__intval(evsel, sample, "code"); in event_icpt_prog_get_key()
|
/tools/perf/ |
D | builtin-timechart.c | 582 u32 state = perf_evsel__intval(evsel, sample, "state"); in process_sample_cpu_idle() 583 u32 cpu_id = perf_evsel__intval(evsel, sample, "cpu_id"); in process_sample_cpu_idle() 598 u32 state = perf_evsel__intval(evsel, sample, "state"); in process_sample_cpu_frequency() 599 u32 cpu_id = perf_evsel__intval(evsel, sample, "cpu_id"); in process_sample_cpu_frequency() 611 u8 flags = perf_evsel__intval(evsel, sample, "common_flags"); in process_sample_sched_wakeup() 612 int waker = perf_evsel__intval(evsel, sample, "common_pid"); in process_sample_sched_wakeup() 613 int wakee = perf_evsel__intval(evsel, sample, "pid"); in process_sample_sched_wakeup() 625 int prev_pid = perf_evsel__intval(evsel, sample, "prev_pid"); in process_sample_sched_switch() 626 int next_pid = perf_evsel__intval(evsel, sample, "next_pid"); in process_sample_sched_switch() 627 u64 prev_state = perf_evsel__intval(evsel, sample, "prev_state"); in process_sample_sched_switch() [all …]
|
D | builtin-kmem.c | 175 unsigned long ptr = perf_evsel__intval(evsel, sample, "ptr"), in perf_evsel__process_alloc_event() 176 call_site = perf_evsel__intval(evsel, sample, "call_site"); in perf_evsel__process_alloc_event() 177 int bytes_req = perf_evsel__intval(evsel, sample, "bytes_req"), in perf_evsel__process_alloc_event() 178 bytes_alloc = perf_evsel__intval(evsel, sample, "bytes_alloc"); in perf_evsel__process_alloc_event() 198 node2 = perf_evsel__intval(evsel, sample, "node"); in perf_evsel__process_alloc_node_event() 238 unsigned long ptr = perf_evsel__intval(evsel, sample, "ptr"); in perf_evsel__process_free_event() 791 unsigned int order = perf_evsel__intval(evsel, sample, "order"); in perf_evsel__process_page_alloc_event() 792 unsigned int gfp_flags = perf_evsel__intval(evsel, sample, "gfp_flags"); in perf_evsel__process_page_alloc_event() 793 unsigned int migrate_type = perf_evsel__intval(evsel, sample, in perf_evsel__process_page_alloc_event() 805 page = perf_evsel__intval(evsel, sample, "pfn"); in perf_evsel__process_page_alloc_event() [all …]
|
D | builtin-sched.c | 812 const u32 pid = perf_evsel__intval(evsel, sample, "pid"); in replay_wakeup_event() 835 const u32 prev_pid = perf_evsel__intval(evsel, sample, "prev_pid"), in replay_switch_event() 836 next_pid = perf_evsel__intval(evsel, sample, "next_pid"); in replay_switch_event() 837 const u64 prev_state = perf_evsel__intval(evsel, sample, "prev_state"); in replay_switch_event() 1106 const u32 prev_pid = perf_evsel__intval(evsel, sample, "prev_pid"), in latency_switch_event() 1107 next_pid = perf_evsel__intval(evsel, sample, "next_pid"); in latency_switch_event() 1108 const u64 prev_state = perf_evsel__intval(evsel, sample, "prev_state"); in latency_switch_event() 1176 const u32 pid = perf_evsel__intval(evsel, sample, "pid"); in latency_runtime_event() 1177 const u64 runtime = perf_evsel__intval(evsel, sample, "runtime"); in latency_runtime_event() 1211 const u32 pid = perf_evsel__intval(evsel, sample, "pid"); in latency_wakeup_event() [all …]
|
D | builtin-lock.c | 408 u64 tmp = perf_evsel__intval(evsel, sample, "lockdep_addr"); in report_lock_acquire_event() 409 int flag = perf_evsel__intval(evsel, sample, "flag"); in report_lock_acquire_event() 481 u64 tmp = perf_evsel__intval(evsel, sample, "lockdep_addr"); in report_lock_acquired_event() 543 u64 tmp = perf_evsel__intval(evsel, sample, "lockdep_addr"); in report_lock_contended_event() 598 u64 tmp = perf_evsel__intval(evsel, sample, "lockdep_addr"); in report_lock_release_event()
|
D | builtin-kvm.c | 71 key->key = perf_evsel__intval(evsel, sample, kvm_exit_reason); in exit_event_get_key() 418 vcpu_record->vcpu_id = perf_evsel__intval(evsel, sample, in per_vcpu_record()
|
D | builtin-inject.c | 521 u32 pid = perf_evsel__intval(evsel, sample, "pid"); in perf_inject__sched_stat()
|
D | builtin-trace.c | 2340 u64 runtime = perf_evsel__intval(evsel, sample, "runtime"); in trace__sched_stat_runtime() 2360 (pid_t)perf_evsel__intval(evsel, sample, "pid"), in trace__sched_stat_runtime() 2362 perf_evsel__intval(evsel, sample, "vruntime")); in trace__sched_stat_runtime()
|
/tools/perf/tests/ |
D | openat-syscall-tp-fields.c | 116 tp_flags = perf_evsel__intval(evsel, &sample, "flags"); in test__syscall_openat_tp_fields()
|
D | switch-tracking.c | 137 next_tid = perf_evsel__intval(evsel, &sample, "next_pid"); in process_sample_event() 138 prev_tid = perf_evsel__intval(evsel, &sample, "prev_pid"); in process_sample_event()
|
/tools/perf/arch/powerpc/util/ |
D | kvm-stat.c | 42 key->key = perf_evsel__intval(evsel, sample, "req"); in hcall_event_get_key()
|
/tools/perf/util/ |
D | evsel.h | 236 u64 perf_evsel__intval(struct evsel *evsel, struct perf_sample *sample,
|
D | evsel.c | 2321 u64 perf_evsel__intval(struct evsel *evsel, struct perf_sample *sample, in perf_evsel__intval() function
|
D | intel-pt.c | 2415 tid = perf_evsel__intval(evsel, sample, "next_pid"); in intel_pt_process_switch()
|