/tools/perf/pmu-events/arch/x86/ |
D | mapfile.csv | 2 GenuineIntel-6-56,v5,broadwellde,core 3 GenuineIntel-6-3D,v17,broadwell,core 4 GenuineIntel-6-47,v17,broadwell,core 5 GenuineIntel-6-4F,v10,broadwellx,core 6 GenuineIntel-6-1C,v4,bonnell,core 7 GenuineIntel-6-26,v4,bonnell,core 8 GenuineIntel-6-27,v4,bonnell,core 9 GenuineIntel-6-36,v4,bonnell,core 10 GenuineIntel-6-35,v4,bonnell,core 11 GenuineIntel-6-5C,v8,goldmont,core [all …]
|
/tools/perf/tests/ |
D | parse-events.c | 51 TEST_ASSERT_VAL("wrong number of entries", 1 == evlist->core.nr_entries); in test__checkevent_tracepoint() 53 TEST_ASSERT_VAL("wrong type", PERF_TYPE_TRACEPOINT == evsel->core.attr.type); in test__checkevent_tracepoint() 55 PERF_TP_SAMPLE_TYPE == evsel->core.attr.sample_type); in test__checkevent_tracepoint() 56 TEST_ASSERT_VAL("wrong sample_period", 1 == evsel->core.attr.sample_period); in test__checkevent_tracepoint() 64 TEST_ASSERT_VAL("wrong number of entries", evlist->core.nr_entries > 1); in test__checkevent_tracepoint_multi() 69 PERF_TYPE_TRACEPOINT == evsel->core.attr.type); in test__checkevent_tracepoint_multi() 71 PERF_TP_SAMPLE_TYPE == evsel->core.attr.sample_type); in test__checkevent_tracepoint_multi() 73 1 == evsel->core.attr.sample_period); in test__checkevent_tracepoint_multi() 82 TEST_ASSERT_VAL("wrong number of entries", 1 == evlist->core.nr_entries); in test__checkevent_raw() 83 TEST_ASSERT_VAL("wrong type", PERF_TYPE_RAW == evsel->core.attr.type); in test__checkevent_raw() [all …]
|
D | task-exit.c | 81 perf_evlist__set_maps(&evlist->core, cpus, threads); in test__task_exit() 94 evsel->core.attr.task = 1; in test__task_exit() 96 evsel->core.attr.sample_freq = 1000000; in test__task_exit() 98 evsel->core.attr.sample_freq = 1; in test__task_exit() 100 evsel->core.attr.inherit = 0; in test__task_exit() 101 evsel->core.attr.watermark = 0; in test__task_exit() 102 evsel->core.attr.wakeup_events = 1; in test__task_exit() 103 evsel->core.attr.exclude_kernel = 1; in test__task_exit() 123 if (perf_mmap__read_init(&md->core) < 0) in test__task_exit() 126 while ((event = perf_mmap__read_event(&md->core)) != NULL) { in test__task_exit() [all …]
|
D | keep-tracking.c | 40 for (i = 0; i < evlist->core.nr_mmaps; i++) { in find_comm() 42 if (perf_mmap__read_init(&md->core) < 0) in find_comm() 44 while ((event = perf_mmap__read_event(&md->core)) != NULL) { in find_comm() 50 perf_mmap__consume(&md->core); in find_comm() 52 perf_mmap__read_done(&md->core); in find_comm() 90 perf_evlist__set_maps(&evlist->core, cpus, threads); in test__keep_tracking() 99 evsel->core.attr.comm = 1; in test__keep_tracking() 100 evsel->core.attr.disabled = 1; in test__keep_tracking() 101 evsel->core.attr.enable_on_exec = 0; in test__keep_tracking()
|
D | openat-syscall-tp-fields.c | 65 perf_thread_map__set_pid(evlist->core.threads, 0, getpid()); in test__syscall_openat_tp_fields() 91 for (i = 0; i < evlist->core.nr_mmaps; i++) { in test__syscall_openat_tp_fields() 96 if (perf_mmap__read_init(&md->core) < 0) in test__syscall_openat_tp_fields() 99 while ((event = perf_mmap__read_event(&md->core)) != NULL) { in test__syscall_openat_tp_fields() 107 perf_mmap__consume(&md->core); in test__syscall_openat_tp_fields() 127 perf_mmap__read_done(&md->core); in test__syscall_openat_tp_fields()
|
D | sw-clock.c | 72 perf_evlist__set_maps(&evlist->core, cpus, threads); in __test__sw_clock_freq() 103 if (perf_mmap__read_init(&md->core) < 0) in __test__sw_clock_freq() 106 while ((event = perf_mmap__read_event(&md->core)) != NULL) { in __test__sw_clock_freq() 121 perf_mmap__consume(&md->core); in __test__sw_clock_freq() 123 perf_mmap__read_done(&md->core); in __test__sw_clock_freq()
|
D | mmap-basic.c | 76 perf_evlist__set_maps(&evlist->core, cpus, threads); in test__basic_mmap() 88 evsels[i]->core.attr.wakeup_events = 1; in test__basic_mmap() 117 if (perf_mmap__read_init(&md->core) < 0) in test__basic_mmap() 120 while ((event = perf_mmap__read_event(&md->core)) != NULL) { in test__basic_mmap() 143 perf_mmap__consume(&md->core); in test__basic_mmap() 145 perf_mmap__read_done(&md->core); in test__basic_mmap()
|
D | switch-tracking.c | 271 for (i = 0; i < evlist->core.nr_mmaps; i++) { in process_events() 273 if (perf_mmap__read_init(&md->core) < 0) in process_events() 276 while ((event = perf_mmap__read_event(&md->core)) != NULL) { in process_events() 279 perf_mmap__consume(&md->core); in process_events() 283 perf_mmap__read_done(&md->core); in process_events() 362 perf_evlist__set_maps(&evlist->core, cpus, threads); in test__switch_tracking() 400 switch_evsel->core.system_wide = true; in test__switch_tracking() 429 tracking_evsel->core.attr.freq = 0; in test__switch_tracking() 430 tracking_evsel->core.attr.sample_period = 1; in test__switch_tracking() 444 if (!tracking_evsel->core.attr.mmap || !tracking_evsel->core.attr.comm) { in test__switch_tracking() [all …]
|
/tools/perf/pmu-events/arch/arm64/ |
D | mapfile.csv | 11 # Type is core, uncore etc 15 0x00000000410fd030,v1,arm/cortex-a53,core 16 0x00000000420f1000,v1,arm/cortex-a53,core 17 0x00000000410fd070,v1,arm/cortex-a57-a72,core 18 0x00000000410fd080,v1,arm/cortex-a57-a72,core 19 0x00000000410fd0b0,v1,arm/cortex-a76-n1,core 20 0x00000000410fd0c0,v1,arm/cortex-a76-n1,core 21 0x00000000420f5160,v1,cavium/thunderx2,core 22 0x00000000430f0af0,v1,cavium/thunderx2,core 23 0x00000000480fd010,v1,hisilicon/hip08,core [all …]
|
/tools/perf/pmu-events/arch/s390/ |
D | mapfile.csv | 2 ^IBM.209[78].*[13]\.[1-5].[[:xdigit:]]+$,1,cf_z10,core 3 ^IBM.281[78].*[13]\.[1-5].[[:xdigit:]]+$,1,cf_z196,core 4 ^IBM.282[78].*[13]\.[1-5].[[:xdigit:]]+$,1,cf_zec12,core 5 ^IBM.296[45].*[13]\.[1-5].[[:xdigit:]]+$,1,cf_z13,core 6 ^IBM.390[67].*[13]\.[1-5].[[:xdigit:]]+$,3,cf_z14,core 7 ^IBM.856[12].*3\.6.[[:xdigit:]]+$,3,cf_z15,core
|
/tools/perf/util/ |
D | mmap.c | 41 return perf_mmap__mmap_len(&map->core); in mmap__mmap_len() 175 ret = perf_mmap__aio_bind(map, i, map->core.cpu, mp->affinity); in perf_mmap__aio_mmap() 266 build_node_mask(cpu__get_node(map->core.cpu), &map->affinity_mask); in perf_mmap__setup_affinity_mask() 268 set_bit(map->core.cpu, map->affinity_mask.bits); in perf_mmap__setup_affinity_mask() 275 if (perf_mmap__mmap(&map->core, &mp->core, fd, cpu)) { in mmap__mmap() 291 map->core.flush = mp->flush; in mmap__mmap() 307 &mp->auxtrace_mp, map->core.base, fd)) in mmap__mmap() 316 u64 head = perf_mmap__read_head(&md->core); in perf_mmap__push() 317 unsigned char *data = md->core.base + page_size; in perf_mmap__push() 322 rc = perf_mmap__read_init(&md->core); in perf_mmap__push() [all …]
|
D | evlist.c | 56 #define FD(e, x, y) (*(int *)xyarray__entry(e->core.fd, x, y)) 57 #define SID(e, x, y) xyarray__entry(e->core.sample_id, x, y) 62 perf_evlist__init(&evlist->core); in evlist__init() 63 perf_evlist__set_maps(&evlist->core, cpus, threads); in evlist__init() 135 list_del_init(&pos->core.node); in evlist__purge() 140 evlist->core.nr_entries = 0; in evlist__purge() 147 perf_evlist__exit(&evlist->core); in evlist__exit() 165 entry->idx = evlist->core.nr_entries; in evlist__add() 168 perf_evlist__add(&evlist->core, &entry->core); in evlist__add() 170 if (evlist->core.nr_entries == 1) in evlist__add() [all …]
|
D | sideband_evlist.c | 25 evsel = evsel__new_idx(attr, evlist->core.nr_entries); in perf_evlist__add_sb_event() 58 for (i = 0; i < evlist->core.nr_mmaps; i++) { in perf_evlist__poll_thread() 62 if (perf_mmap__read_init(&map->core)) in perf_evlist__poll_thread() 64 while ((event = perf_mmap__read_event(&map->core)) != NULL) { in perf_evlist__poll_thread() 72 perf_mmap__consume(&map->core); in perf_evlist__poll_thread() 75 perf_mmap__read_done(&map->core); in perf_evlist__poll_thread() 89 evsel->core.attr.sample_id_all = 1; in evlist__set_cb() 90 evsel->core.attr.watermark = 1; in evlist__set_cb() 91 evsel->core.attr.wakeup_watermark = 1; in evlist__set_cb() 107 if (evlist->core.nr_entries > 1) { in perf_evlist__start_sb_thread() [all …]
|
D | evsel.h | 44 struct perf_evsel core; member 158 return perf_evsel__cpus(&evsel->core); in evsel__cpus() 272 (evsel->core.attr.type == PERF_TYPE_##t && \ 273 evsel->core.attr.config == PERF_COUNT_##c) 277 return (e1->core.attr.type == e2->core.attr.type) && in evsel__match2() 278 (e1->core.attr.config == e2->core.attr.config); in evsel__match2() 317 return list_entry(evsel->core.node.next, struct evsel, core.node); in evsel__next() 322 return list_entry(evsel->core.node.prev, struct evsel, core.node); in evsel__prev() 350 return evsel__is_group_leader(evsel) && evsel->core.nr_members > 1; in evsel__is_group_event() 377 for ((_evsel) = list_entry((_leader)->core.node.next, struct evsel, core.node); \ [all …]
|
D | evsel.c | 102 #define FD(e, x, y) (*(int *)xyarray__entry(e->core.fd, x, y)) 182 evsel->id_pos = __perf_evsel__calc_id_pos(evsel->core.attr.sample_type); in evsel__calc_id_pos() 183 evsel->is_pos = __perf_evsel__calc_is_pos(evsel->core.attr.sample_type); in evsel__calc_id_pos() 189 if (!(evsel->core.attr.sample_type & bit)) { in __evsel__set_sample_bit() 190 evsel->core.attr.sample_type |= bit; in __evsel__set_sample_bit() 199 if (evsel->core.attr.sample_type & bit) { in __evsel__reset_sample_bit() 200 evsel->core.attr.sample_type &= ~bit; in __evsel__reset_sample_bit() 215 evsel->core.attr.read_format |= PERF_FORMAT_ID; in evsel__set_sample_id() 239 perf_evsel__init(&evsel->core, attr); in evsel__init() 271 evsel->core.attr.sample_type |= (PERF_SAMPLE_RAW | PERF_SAMPLE_TIME | in evsel__new_idx() [all …]
|
D | record.c | 52 struct perf_event_attr *attr = &evsel->core.attr; in evsel__config_leader_sampling() 88 attr->sample_type = read_sampler->core.attr.sample_type | in evsel__config_leader_sampling() 89 leader->core.attr.sample_type; in evsel__config_leader_sampling() 107 if (evlist->core.cpus->map[0] < 0) in perf_evlist__config() 115 evsel->core.attr.comm_exec = 1; in perf_evlist__config() 130 } else if (evlist->core.nr_entries > 1) { in perf_evlist__config() 134 if (evsel->core.attr.sample_type == first->core.attr.sample_type) in perf_evlist__config() 238 if (!evlist || perf_cpu_map__empty(evlist->core.cpus)) { in perf_evlist__can_select_event() 244 cpu = evlist->core.cpus->map[0]; in perf_evlist__can_select_event() 248 fd = sys_perf_event_open(&evsel->core.attr, pid, cpu, -1, in perf_evlist__can_select_event()
|
D | evlist.h | 52 struct perf_evlist core; member 252 return list_empty(&evlist->core.entries); in perf_evlist__empty() 257 struct perf_evsel *evsel = perf_evlist__first(&evlist->core); in evlist__first() 259 return container_of(evsel, struct evsel, core); in evlist__first() 264 struct perf_evsel *evsel = perf_evlist__last(&evlist->core); in evlist__last() 266 return container_of(evsel, struct evsel, core); in evlist__last() 282 list_for_each_entry(evsel, list, core.node) 290 __evlist__for_each_entry(&(evlist)->core.entries, evsel) 298 list_for_each_entry_continue(evsel, list, core.node) 306 __evlist__for_each_entry_continue(&(evlist)->core.entries, evsel) [all …]
|
D | perf_api_probe.c | 32 fd = sys_perf_event_open(&evsel->core.attr, pid, cpu, -1, flags); in perf_do_probe_api() 46 fd = sys_perf_event_open(&evsel->core.attr, pid, cpu, -1, flags); in perf_do_probe_api() 83 evsel->core.attr.sample_type |= PERF_SAMPLE_IDENTIFIER; in perf_probe_sample_identifier() 88 evsel->core.attr.comm_exec = 1; in perf_probe_comm_exec() 93 evsel->core.attr.context_switch = 1; in perf_probe_context_switch() 98 evsel->core.attr.text_poke = 1; in perf_probe_text_poke()
|
/tools/testing/selftests/bpf/ |
D | test_kmod.sh | 23 sysctl -w net.core.bpf_jit_enable=$1 2>&1 > /dev/null 24 sysctl -w net.core.bpf_jit_harden=$2 2>&1 > /dev/null 50 JE=`sysctl -n net.core.bpf_jit_enable` 51 JH=`sysctl -n net.core.bpf_jit_harden` 56 sysctl -w net.core.bpf_jit_enable=$JE 2>&1 > /dev/null 57 sysctl -w net.core.bpf_jit_harden=$JH 2>&1 > /dev/null
|
/tools/perf/arch/x86/tests/ |
D | perf-time-to-tsc.c | 81 perf_evlist__set_maps(&evlist->core, cpus, threads); in test__perf_time_to_tsc() 89 evsel->core.attr.comm = 1; in test__perf_time_to_tsc() 90 evsel->core.attr.disabled = 1; in test__perf_time_to_tsc() 91 evsel->core.attr.enable_on_exec = 0; in test__perf_time_to_tsc() 97 pc = evlist->mmap[0].core.base; in test__perf_time_to_tsc() 119 for (i = 0; i < evlist->core.nr_mmaps; i++) { in test__perf_time_to_tsc() 121 if (perf_mmap__read_init(&md->core) < 0) in test__perf_time_to_tsc() 124 while ((event = perf_mmap__read_event(&md->core)) != NULL) { in test__perf_time_to_tsc() 141 perf_mmap__consume(&md->core); in test__perf_time_to_tsc() 143 perf_mmap__read_done(&md->core); in test__perf_time_to_tsc()
|
/tools/power/cpupower/lib/ |
D | cpupower.c | 113 else if (top1->core < top2->core) in __compare() 115 else if (top1->core > top2->core) in __compare() 147 cpu_top->core_info[cpu].core = -1; in get_cpu_topology() 153 &(cpu_top->core_info[cpu].core)) < 0) { in get_cpu_topology() 155 cpu_top->core_info[cpu].core = -1; in get_cpu_topology()
|
/tools/perf/pmu-events/arch/powerpc/ |
D | mapfile.csv | 10 # Type is core, uncore etc 16 004[bcd][[:xdigit:]]{4},1,power8,core 17 004e[[:xdigit:]]{4},1,power9,core
|
/tools/testing/selftests/powerpc/ptrace/ |
D | Makefile | 4 ptrace-tm-spd-vsx ptrace-tm-spr ptrace-hwbreak ptrace-pkey core-pkey \ 12 $(OUTPUT)/ptrace-pkey $(OUTPUT)/core-pkey: child.h 13 $(OUTPUT)/ptrace-pkey $(OUTPUT)/core-pkey: LDLIBS += -pthread
|
/tools/perf/arch/arm64/util/ |
D | arm-spe.c | 54 if (!session->evlist->core.nr_mmaps) in arm_spe_info_fill() 78 if (evsel->core.attr.type == arm_spe_pmu->type) { in arm_spe_recording_options() 83 evsel->core.attr.freq = 0; in arm_spe_recording_options() 84 evsel->core.attr.sample_period = 1; in arm_spe_recording_options() 141 tracking_evsel->core.attr.freq = 0; in arm_spe_recording_options() 142 tracking_evsel->core.attr.sample_period = 1; in arm_spe_recording_options()
|
/tools/perf/arch/x86/util/ |
D | intel-pt.c | 133 if (evsel->core.attr.type == intel_pt_pmu->type) { in intel_pt_read_config() 134 *res = intel_pt_masked_bits(mask, evsel->core.attr.config); in intel_pt_read_config() 286 if (evsel->core.attr.type == intel_pt_pmu->type) in intel_pt_find_filter() 364 if (!session->evlist->core.nr_mmaps) in intel_pt_info_fill() 367 pc = session->evlist->mmap[0].core.base; in intel_pt_info_fill() 380 per_cpu_mmaps = !perf_cpu_map__empty(session->evlist->core.cpus); in intel_pt_info_fill() 434 evsel->core.system_wide = true; in intel_pt_track_switches() 541 !(evsel->core.attr.config & 1)) { in intel_pt_validate_config() 543 evsel->core.attr.config |= 1; in intel_pt_validate_config() 548 evsel->core.attr.config); in intel_pt_validate_config() [all …]
|