Home
last modified time | relevance | path

Searched refs:user_requested_cpus (Results 1 – 21 of 21) sorted by relevance

/tools/perf/util/
Dtop.c98 perf_cpu_map__nr(top->evlist->core.user_requested_cpus) > 1 in perf_top__header_snprintf()
106 perf_cpu_map__nr(top->evlist->core.user_requested_cpus), in perf_top__header_snprintf()
107 perf_cpu_map__nr(top->evlist->core.user_requested_cpus) > 1 in perf_top__header_snprintf()
Dcputopo.c179 struct perf_cpu_map *user_requested_cpus; in cpu_topology__core_wide() local
188 user_requested_cpus = perf_cpu_map__new(user_requested_cpu_list); in cpu_topology__core_wide()
199 has_first = perf_cpu_map__has(user_requested_cpus, cpu); in cpu_topology__core_wide()
209 if (perf_cpu_map__has(user_requested_cpus, cpu) != has_first) { in cpu_topology__core_wide()
211 perf_cpu_map__put(user_requested_cpus); in cpu_topology__core_wide()
218 perf_cpu_map__put(user_requested_cpus); in cpu_topology__core_wide()
Drecord.c102 if (perf_cpu_map__cpu(evlist->core.user_requested_cpus, 0).cpu < 0) in evlist__config()
240 if (!evlist || perf_cpu_map__empty(evlist->core.user_requested_cpus)) { in evlist__can_select_event()
248 cpu = perf_cpu_map__cpu(evlist->core.user_requested_cpus, 0); in evlist__can_select_event()
Dbpf_ftrace.c41 ncpus = perf_cpu_map__nr(ftrace->evlist->core.user_requested_cpus); in perf_ftrace__latency_prepare_bpf()
66 cpu = perf_cpu_map__cpu(ftrace->evlist->core.user_requested_cpus, i).cpu; in perf_ftrace__latency_prepare_bpf()
Dsideband_evlist.c117 if (evsel__open(counter, evlist->core.user_requested_cpus, in evlist__start_sb_thread()
Devlist.c498 if (!cpu_map__is_dummy(evlist->core.user_requested_cpus)) { in __evlist__disable()
567 if (!cpu_map__is_dummy(evlist->core.user_requested_cpus)) { in __evlist__enable()
1325 if (!evlist->core.user_requested_cpus || in evlist__close()
1326 cpu_map__is_dummy(evlist->core.user_requested_cpus)) { in evlist__close()
1388 if (evlist->core.threads == NULL && evlist->core.user_requested_cpus == NULL) { in evlist__open()
2501 struct perf_cpu_map *user_requested_cpus; in evlist__warn_user_requested_cpus() local
2507 user_requested_cpus = perf_cpu_map__new(cpu_list); in evlist__warn_user_requested_cpus()
2508 if (!user_requested_cpus) in evlist__warn_user_requested_cpus()
2516 intersect = perf_cpu_map__intersect(to_test, user_requested_cpus); in evlist__warn_user_requested_cpus()
2517 if (!perf_cpu_map__equal(intersect, user_requested_cpus)) { in evlist__warn_user_requested_cpus()
[all …]
Dbpf_off_cpu.c147 ncpus = perf_cpu_map__nr(evlist->core.user_requested_cpus); in off_cpu_prepare()
213 cpu = perf_cpu_map__cpu(evlist->core.user_requested_cpus, i).cpu; in off_cpu_prepare()
Dbpf_lock_contention.c48 ncpus = perf_cpu_map__nr(evlist->core.user_requested_cpus); in lock_contention_prepare()
101 cpu = perf_cpu_map__cpu(evlist->core.user_requested_cpus, i).cpu; in lock_contention_prepare()
Dauxtrace.c177 bool per_cpu = !perf_cpu_map__empty(evlist->core.user_requested_cpus); in auxtrace_mmap_params__set_idx()
651 bool per_cpu_mmaps = !perf_cpu_map__empty(evlist->core.user_requested_cpus); in evlist__enable_event_idx()
Dstat-display.c1121 perf_cpu_map__for_each_cpu(cpu, all_idx, evlist->core.user_requested_cpus) { in print_no_aggr_metric()
Dsynthetic-events.c2256 err = perf_event__synthesize_cpu_map(tool, evlist->core.user_requested_cpus, process, NULL); in perf_event__synthesize_stat_events()
/tools/lib/perf/
Devlist.c49 evsel->cpus = perf_cpu_map__intersect(evlist->user_requested_cpus, evsel->own_cpus); in __perf_evlist__propagate_maps()
51 (!evsel->requires_cpu && perf_cpu_map__has_any_cpu(evlist->user_requested_cpus))) { in __perf_evlist__propagate_maps()
60 evsel->cpus = perf_cpu_map__get(evlist->user_requested_cpus); in __perf_evlist__propagate_maps()
153 perf_cpu_map__put(evlist->user_requested_cpus); in perf_evlist__exit()
156 evlist->user_requested_cpus = NULL; in perf_evlist__exit()
185 if (cpus != evlist->user_requested_cpus) { in perf_evlist__set_maps()
186 perf_cpu_map__put(evlist->user_requested_cpus); in perf_evlist__set_maps()
187 evlist->user_requested_cpus = perf_cpu_map__get(cpus); in perf_evlist__set_maps()
/tools/lib/perf/include/internal/
Devlist.h26 struct perf_cpu_map *user_requested_cpus; member
/tools/perf/arch/arm/util/
Dcs-etm.c201 struct perf_cpu_map *event_cpus = evsel->evlist->core.user_requested_cpus; in cs_etm_validate_config()
300 struct perf_cpu_map *cpus = evlist->core.user_requested_cpus; in cs_etm_recording_options()
520 struct perf_cpu_map *event_cpus = evlist->core.user_requested_cpus; in cs_etm_info_priv_size()
786 struct perf_cpu_map *event_cpus = session->evlist->core.user_requested_cpus; in cs_etm_info_fill()
/tools/perf/bench/
Devlist-open-close.c154 printf(" Number of cpus:\t%d\n", perf_cpu_map__nr(evlist->core.user_requested_cpus)); in bench_evlist_open_close__run()
/tools/perf/arch/arm64/util/
Darm-spe.c124 struct perf_cpu_map *cpus = evlist->core.user_requested_cpus; in arm_spe_recording_options()
/tools/perf/arch/x86/util/
Dintel-bts.c113 const struct perf_cpu_map *cpus = evlist->core.user_requested_cpus; in intel_bts_recording_options()
Dintel-pt.c381 per_cpu_mmaps = !perf_cpu_map__empty(session->evlist->core.user_requested_cpus); in intel_pt_info_fill()
626 const struct perf_cpu_map *cpus = evlist->core.user_requested_cpus; in intel_pt_recording_options()
/tools/perf/
Dbuiltin-stat.c714 if (!cpu_map__is_dummy(evsel_list->core.user_requested_cpus)) { in __run_perf_stat()
1596 stat_config.aggr_map = cpu_aggr_map__new(evsel_list->core.user_requested_cpus, in perf_stat_init_aggr_mode()
1625 if (!perf_cpu_map__empty(evsel_list->core.user_requested_cpus)) in perf_stat_init_aggr_mode()
1626 nr = perf_cpu_map__max(evsel_list->core.user_requested_cpus).cpu; in perf_stat_init_aggr_mode()
1912 stat_config.aggr_map = cpu_aggr_map__new(evsel_list->core.user_requested_cpus, in perf_stat_init_aggr_mode_file()
Dbuiltin-ftrace.c304 struct perf_cpu_map *cpumap = ftrace->evlist->core.user_requested_cpus; in set_tracing_cpu()
Dbuiltin-record.c1021 return cpu_map__is_dummy(evlist->core.user_requested_cpus); in evlist__per_thread()