Home
last modified time | relevance | path

Searched refs:cgroups (Results 1 – 10 of 10) sorted by relevance

/tools/perf/util/
Dcgroup.c349 down_write(&env->cgroups.lock); in cgroup__findnew()
350 cgrp = __cgroup__findnew(&env->cgroups.tree, id, true, path); in cgroup__findnew()
351 up_write(&env->cgroups.lock); in cgroup__findnew()
359 down_read(&env->cgroups.lock); in cgroup__find()
360 cgrp = __cgroup__findnew(&env->cgroups.tree, id, false, NULL); in cgroup__find()
361 up_read(&env->cgroups.lock); in cgroup__find()
370 down_write(&env->cgroups.lock); in perf_env__purge_cgroups()
371 while (!RB_EMPTY_ROOT(&env->cgroups.tree)) { in perf_env__purge_cgroups()
372 node = rb_first(&env->cgroups.tree); in perf_env__purge_cgroups()
375 rb_erase(node, &env->cgroups.tree); in perf_env__purge_cgroups()
[all …]
Dcgroup.h28 int evlist__expand_cgroup(struct evlist *evlist, const char *cgroups,
Denv.h97 } cgroups; member
/tools/testing/selftests/vm/
Dcharge_reserved_hugetlb.sh501 echo Test normal case, multiple cgroups.
547 echo Test normal case with write, multiple cgroups.
/tools/testing/selftests/bpf/progs/
Dprofiler.inc.h257 struct kernfs_node* proc_kernfs = BPF_CORE_READ(task, cgroups, dfl_cgrp, kn); in populate_cgroup_info()
268 BPF_CORE_READ(task, cgroups, subsys[i]); in populate_cgroup_info()
632 struct kernfs_node* proc_kernfs = BPF_CORE_READ(task, cgroups, dfl_cgrp, kn); in raw_tracepoint__sched_process_exit()
/tools/bpf/bpftool/Documentation/
Dbpftool-cgroup.rst50 Iterate over all cgroups in *CGROUP_ROOT* and list all
/tools/perf/Documentation/
Dperf-trace.txt75 Look for cgroups to set at the /sys/fs/cgroup/perf_event directory, then
89 Multiple cgroups:
Dperf-stat.txt158 container "name" are monitored when they run on the monitored CPUs. Multiple cgroups
170 Expand event list for each cgroup in "name" (allow multiple cgroups separated
Dperf-top.txt280 --all-cgroups::
Dperf-record.txt345 container "name" are monitored when they run on the monitored CPUs. Multiple cgroups
400 --all-cgroups::