• Home
  • Raw
  • Download

Lines Matching full:env

4 #include "env.h"
18 void perf_env__insert_bpf_prog_info(struct perf_env *env, in perf_env__insert_bpf_prog_info() argument
21 down_write(&env->bpf_progs.lock); in perf_env__insert_bpf_prog_info()
22 __perf_env__insert_bpf_prog_info(env, info_node); in perf_env__insert_bpf_prog_info()
23 up_write(&env->bpf_progs.lock); in perf_env__insert_bpf_prog_info()
26 void __perf_env__insert_bpf_prog_info(struct perf_env *env, struct bpf_prog_info_node *info_node) in __perf_env__insert_bpf_prog_info() argument
33 p = &env->bpf_progs.infos.rb_node; in __perf_env__insert_bpf_prog_info()
49 rb_insert_color(&info_node->rb_node, &env->bpf_progs.infos); in __perf_env__insert_bpf_prog_info()
50 env->bpf_progs.infos_cnt++; in __perf_env__insert_bpf_prog_info()
53 struct bpf_prog_info_node *perf_env__find_bpf_prog_info(struct perf_env *env, in perf_env__find_bpf_prog_info() argument
59 down_read(&env->bpf_progs.lock); in perf_env__find_bpf_prog_info()
60 n = env->bpf_progs.infos.rb_node; in perf_env__find_bpf_prog_info()
74 up_read(&env->bpf_progs.lock); in perf_env__find_bpf_prog_info()
78 bool perf_env__insert_btf(struct perf_env *env, struct btf_node *btf_node) in perf_env__insert_btf() argument
82 down_write(&env->bpf_progs.lock); in perf_env__insert_btf()
83 ret = __perf_env__insert_btf(env, btf_node); in perf_env__insert_btf()
84 up_write(&env->bpf_progs.lock); in perf_env__insert_btf()
88 bool __perf_env__insert_btf(struct perf_env *env, struct btf_node *btf_node) in __perf_env__insert_btf() argument
95 p = &env->bpf_progs.btfs.rb_node; in __perf_env__insert_btf()
111 rb_insert_color(&btf_node->rb_node, &env->bpf_progs.btfs); in __perf_env__insert_btf()
112 env->bpf_progs.btfs_cnt++; in __perf_env__insert_btf()
116 struct btf_node *perf_env__find_btf(struct perf_env *env, __u32 btf_id) in perf_env__find_btf() argument
120 down_read(&env->bpf_progs.lock); in perf_env__find_btf()
121 res = __perf_env__find_btf(env, btf_id); in perf_env__find_btf()
122 up_read(&env->bpf_progs.lock); in perf_env__find_btf()
126 struct btf_node *__perf_env__find_btf(struct perf_env *env, __u32 btf_id) in __perf_env__find_btf() argument
131 n = env->bpf_progs.btfs.rb_node; in __perf_env__find_btf()
146 static void perf_env__purge_bpf(struct perf_env *env) in perf_env__purge_bpf() argument
151 down_write(&env->bpf_progs.lock); in perf_env__purge_bpf()
153 root = &env->bpf_progs.infos; in perf_env__purge_bpf()
166 env->bpf_progs.infos_cnt = 0; in perf_env__purge_bpf()
168 root = &env->bpf_progs.btfs; in perf_env__purge_bpf()
180 env->bpf_progs.btfs_cnt = 0; in perf_env__purge_bpf()
182 up_write(&env->bpf_progs.lock); in perf_env__purge_bpf()
185 void perf_env__exit(struct perf_env *env) in perf_env__exit() argument
189 perf_env__purge_bpf(env); in perf_env__exit()
190 perf_env__purge_cgroups(env); in perf_env__exit()
191 zfree(&env->hostname); in perf_env__exit()
192 zfree(&env->os_release); in perf_env__exit()
193 zfree(&env->version); in perf_env__exit()
194 zfree(&env->arch); in perf_env__exit()
195 zfree(&env->cpu_desc); in perf_env__exit()
196 zfree(&env->cpuid); in perf_env__exit()
197 zfree(&env->cmdline); in perf_env__exit()
198 zfree(&env->cmdline_argv); in perf_env__exit()
199 zfree(&env->sibling_dies); in perf_env__exit()
200 zfree(&env->sibling_cores); in perf_env__exit()
201 zfree(&env->sibling_threads); in perf_env__exit()
202 zfree(&env->pmu_mappings); in perf_env__exit()
203 zfree(&env->cpu); in perf_env__exit()
204 zfree(&env->cpu_pmu_caps); in perf_env__exit()
205 zfree(&env->numa_map); in perf_env__exit()
207 for (i = 0; i < env->nr_numa_nodes; i++) in perf_env__exit()
208 perf_cpu_map__put(env->numa_nodes[i].map); in perf_env__exit()
209 zfree(&env->numa_nodes); in perf_env__exit()
211 for (i = 0; i < env->caches_cnt; i++) in perf_env__exit()
212 cpu_cache_level__free(&env->caches[i]); in perf_env__exit()
213 zfree(&env->caches); in perf_env__exit()
215 for (i = 0; i < env->nr_memory_nodes; i++) in perf_env__exit()
216 zfree(&env->memory_nodes[i].set); in perf_env__exit()
217 zfree(&env->memory_nodes); in perf_env__exit()
220 void perf_env__init(struct perf_env *env) in perf_env__init() argument
222 env->bpf_progs.infos = RB_ROOT; in perf_env__init()
223 env->bpf_progs.btfs = RB_ROOT; in perf_env__init()
224 init_rwsem(&env->bpf_progs.lock); in perf_env__init()
227 int perf_env__set_cmdline(struct perf_env *env, int argc, const char *argv[]) in perf_env__set_cmdline() argument
232 env->cmdline_argv = calloc(argc, sizeof(char *)); in perf_env__set_cmdline()
233 if (env->cmdline_argv == NULL) in perf_env__set_cmdline()
241 env->cmdline_argv[i] = argv[i]; in perf_env__set_cmdline()
242 if (env->cmdline_argv[i] == NULL) in perf_env__set_cmdline()
246 env->nr_cmdline = argc; in perf_env__set_cmdline()
250 zfree(&env->cmdline_argv); in perf_env__set_cmdline()
255 int perf_env__read_cpu_topology_map(struct perf_env *env) in perf_env__read_cpu_topology_map() argument
259 if (env->cpu != NULL) in perf_env__read_cpu_topology_map()
262 if (env->nr_cpus_avail == 0) in perf_env__read_cpu_topology_map()
263 env->nr_cpus_avail = cpu__max_present_cpu(); in perf_env__read_cpu_topology_map()
265 nr_cpus = env->nr_cpus_avail; in perf_env__read_cpu_topology_map()
269 env->cpu = calloc(nr_cpus, sizeof(env->cpu[0])); in perf_env__read_cpu_topology_map()
270 if (env->cpu == NULL) in perf_env__read_cpu_topology_map()
274 env->cpu[cpu].core_id = cpu_map__get_core_id(cpu); in perf_env__read_cpu_topology_map()
275 env->cpu[cpu].socket_id = cpu_map__get_socket_id(cpu); in perf_env__read_cpu_topology_map()
276 env->cpu[cpu].die_id = cpu_map__get_die_id(cpu); in perf_env__read_cpu_topology_map()
279 env->nr_cpus_avail = nr_cpus; in perf_env__read_cpu_topology_map()
283 int perf_env__read_cpuid(struct perf_env *env) in perf_env__read_cpuid() argument
291 free(env->cpuid); in perf_env__read_cpuid()
292 env->cpuid = strdup(cpuid); in perf_env__read_cpuid()
293 if (env->cpuid == NULL) in perf_env__read_cpuid()
298 static int perf_env__read_arch(struct perf_env *env) in perf_env__read_arch() argument
302 if (env->arch) in perf_env__read_arch()
306 env->arch = strdup(uts.machine); in perf_env__read_arch()
308 return env->arch ? 0 : -ENOMEM; in perf_env__read_arch()
311 static int perf_env__read_nr_cpus_avail(struct perf_env *env) in perf_env__read_nr_cpus_avail() argument
313 if (env->nr_cpus_avail == 0) in perf_env__read_nr_cpus_avail()
314 env->nr_cpus_avail = cpu__max_present_cpu(); in perf_env__read_nr_cpus_avail()
316 return env->nr_cpus_avail ? 0 : -ENOENT; in perf_env__read_nr_cpus_avail()
319 const char *perf_env__raw_arch(struct perf_env *env) in perf_env__raw_arch() argument
321 return env && !perf_env__read_arch(env) ? env->arch : "unknown"; in perf_env__raw_arch()
324 int perf_env__nr_cpus_avail(struct perf_env *env) in perf_env__nr_cpus_avail() argument
326 return env && !perf_env__read_nr_cpus_avail(env) ? env->nr_cpus_avail : 0; in perf_env__nr_cpus_avail()
368 const char *perf_env__arch(struct perf_env *env) in perf_env__arch() argument
372 if (!env || !env->arch) { /* Assume local operation */ in perf_env__arch()
378 arch_name = env->arch; in perf_env__arch()
384 int perf_env__numa_node(struct perf_env *env, int cpu) in perf_env__numa_node() argument
386 if (!env->nr_numa_map) { in perf_env__numa_node()
390 for (i = 0; i < env->nr_numa_nodes; i++) { in perf_env__numa_node()
391 nn = &env->numa_nodes[i]; in perf_env__numa_node()
401 env->numa_map = malloc(nr * sizeof(int)); in perf_env__numa_node()
402 if (!env->numa_map) in perf_env__numa_node()
406 env->numa_map[i] = -1; in perf_env__numa_node()
408 env->nr_numa_map = nr; in perf_env__numa_node()
410 for (i = 0; i < env->nr_numa_nodes; i++) { in perf_env__numa_node()
413 nn = &env->numa_nodes[i]; in perf_env__numa_node()
415 env->numa_map[j] = i; in perf_env__numa_node()
419 return cpu >= 0 && cpu < env->nr_numa_map ? env->numa_map[cpu] : -1; in perf_env__numa_node()