/tools/perf/util/ |
D | map.c | 21 static void __maps__insert(struct maps *maps, struct map *map); 475 static void maps__init(struct maps *maps) in maps__init() argument 477 maps->entries = RB_ROOT; in maps__init() 478 pthread_rwlock_init(&maps->lock, NULL); in maps__init() 485 maps__init(&mg->maps[i]); in map_groups__init() 491 static void __maps__purge(struct maps *maps) in __maps__purge() argument 493 struct rb_root *root = &maps->entries; in __maps__purge() 505 static void maps__exit(struct maps *maps) in maps__exit() argument 507 pthread_rwlock_wrlock(&maps->lock); in maps__exit() 508 __maps__purge(maps); in maps__exit() [all …]
|
D | find-vdso-map.c | 3 FILE *maps; in find_vdso_map() local 7 maps = fopen("/proc/self/maps", "r"); in find_vdso_map() 8 if (!maps) { in find_vdso_map() 13 while (!found && fgets(line, sizeof(line), maps)) { in find_vdso_map() 28 fclose(maps); in find_vdso_map()
|
D | map.h | 62 struct maps { struct 68 struct maps maps[MAP__NR_TYPES]; argument 183 void maps__insert(struct maps *maps, struct map *map); 184 void maps__remove(struct maps *maps, struct map *map); 185 struct map *maps__find(struct maps *maps, u64 addr); 186 struct map *maps__first(struct maps *maps); 188 struct symbol *maps__find_symbol_by_name(struct maps *maps, const char *name, 196 int maps__set_kallsyms_ref_reloc_sym(struct map **maps, const char *symbol_name, 201 maps__insert(&mg->maps[map->type], map); in map_groups__insert() 207 maps__remove(&mg->maps[map->type], map); in map_groups__remove() [all …]
|
D | thread.c | 219 struct maps *maps = &thread->mg->maps[i]; in __thread__prepare_access() local 222 pthread_rwlock_rdlock(&maps->lock); in __thread__prepare_access() 224 for (map = maps__first(maps); map; map = map__next(map)) { in __thread__prepare_access() 230 pthread_rwlock_unlock(&maps->lock); in __thread__prepare_access()
|
D | symbol.c | 210 struct maps *maps = &mg->maps[type]; in __map_groups__fixup_end() local 213 pthread_rwlock_wrlock(&maps->lock); in __map_groups__fixup_end() 215 curr = maps__first(maps); in __map_groups__fixup_end() 231 pthread_rwlock_unlock(&maps->lock); in __map_groups__fixup_end() 1087 struct list_head maps; member 1102 list_add(&map->node, &md->maps); in kcore_mapfn() 1138 INIT_LIST_HEAD(&md.maps); in dso__load_kcore() 1154 if (list_empty(&md.maps)) { in dso__load_kcore() 1171 list_for_each_entry(new_map, &md.maps, node) { in dso__load_kcore() 1180 replacement_map = list_entry(md.maps.next, struct map, node); in dso__load_kcore() [all …]
|
D | evlist.c | 1042 struct perf_mmap *maps = evlist->mmap; in perf_evlist__mmap_per_evsel() local 1049 maps = evlist->backward_mmap; in perf_evlist__mmap_per_evsel() 1051 if (!maps) { in perf_evlist__mmap_per_evsel() 1052 maps = perf_evlist__alloc_mmap(evlist); in perf_evlist__mmap_per_evsel() 1053 if (!maps) in perf_evlist__mmap_per_evsel() 1055 evlist->backward_mmap = maps; in perf_evlist__mmap_per_evsel() 1073 if (perf_mmap__mmap(&maps[idx], mp, *output) < 0) in perf_evlist__mmap_per_evsel() 1079 perf_mmap__get(&maps[idx]); in perf_evlist__mmap_per_evsel() 1092 __perf_evlist__add_pollfd(evlist, fd, &maps[idx], revent) < 0) { in perf_evlist__mmap_per_evsel() 1093 perf_mmap__put(&maps[idx]); in perf_evlist__mmap_per_evsel()
|
D | event.c | 386 struct maps *maps = &kmaps->maps[MAP__FUNCTION]; in perf_event__synthesize_modules() local 406 for (pos = maps__first(maps); pos; pos = map__next(pos)) { in perf_event__synthesize_modules()
|
D | probe-event.c | 161 struct maps *maps = &grp->maps[MAP__FUNCTION]; in kernel_get_module_map() local 171 for (pos = maps__first(maps); pos; pos = map__next(pos)) { in kernel_get_module_map()
|
D | session.c | 1911 int maps__set_kallsyms_ref_reloc_sym(struct map **maps, in maps__set_kallsyms_ref_reloc_sym() argument 1935 struct kmap *kmap = map__kmap(maps[i]); in maps__set_kallsyms_ref_reloc_sym()
|
/tools/perf/tests/ |
D | vmlinux-kallsyms.c | 21 struct maps *maps = &vmlinux.kmaps.maps[type]; in test__vmlinux_matches_kallsyms() local 176 for (map = maps__first(maps); map; map = map__next(map)) { in test__vmlinux_matches_kallsyms() 201 for (map = maps__first(maps); map; map = map__next(map)) { in test__vmlinux_matches_kallsyms() 229 maps = &kallsyms.kmaps.maps[type]; in test__vmlinux_matches_kallsyms() 231 for (map = maps__first(maps); map; map = map__next(map)) { in test__vmlinux_matches_kallsyms()
|
/tools/lib/bpf/ |
D | libbpf.c | 201 struct bpf_map *maps; member 531 obj->maps = calloc(nr_maps, sizeof(obj->maps[0])); in bpf_object__init_maps() 532 if (!obj->maps) { in bpf_object__init_maps() 539 struct bpf_map_def *def = &obj->maps[i].def; in bpf_object__init_maps() 546 obj->maps[i].fd = -1; in bpf_object__init_maps() 582 obj->maps[map_idx].name = strdup(map_name); in bpf_object__init_maps_name() 583 if (!obj->maps[map_idx].name) { in bpf_object__init_maps_name() 588 obj->maps[map_idx].name); in bpf_object__init_maps_name() 793 struct bpf_map_def *def = &obj->maps[i].def; in bpf_object__create_maps() 794 int *pfd = &obj->maps[i].fd; in bpf_object__create_maps() [all …]
|
/tools/testing/selftests/x86/ |
D | test_vsyscall.c | 100 FILE *maps; in init_vsys() local 104 maps = fopen("/proc/self/maps", "r"); in init_vsys() 105 if (!maps) { in init_vsys() 111 while (fgets(line, sizeof(line), maps)) { in init_vsys() 142 fclose(maps); in init_vsys()
|
/tools/perf/pmu-events/ |
D | README | 15 to be a CSV file that maps a specific CPU to its set of PMU events. 52 - A 'mapping table' that maps each CPU of the architecture, to its
|
/tools/perf/ |
D | builtin-record.c | 517 struct perf_mmap *maps; in record__mmap_read_evlist() local 522 maps = backward ? evlist->backward_mmap : evlist->mmap; in record__mmap_read_evlist() 523 if (!maps) in record__mmap_read_evlist() 530 struct auxtrace_mmap *mm = &maps[i].auxtrace_mmap; in record__mmap_read_evlist() 532 if (maps[i].base) { in record__mmap_read_evlist() 533 if (record__mmap_read(rec, &maps[i], in record__mmap_read_evlist()
|
/tools/perf/Documentation/ |
D | perf-trace.txt | 172 - map type is either 'd' for non-executable maps or 'x' for executable maps;
|
D | perf-record.txt | 143 within a single mapping. MMAP events (or /proc/<pid>/maps) can be
|
D | intel-pt.txt | 225 that does not overlap executable memory maps.
|
/tools/perf/util/intel-pt-decoder/ |
D | x86-opcode-map.txt | 7 #<Opcode maps> 19 #<group maps>
|
/tools/objtool/arch/x86/insn/ |
D | x86-opcode-map.txt | 7 #<Opcode maps> 19 #<group maps>
|