/tools/perf/tests/ |
D | mem2node.c | 49 struct memory_node nodes[3]; in test__mem2node() local 51 .memory_nodes = (struct memory_node *) &nodes[0], in test__mem2node() 52 .nr_memory_nodes = ARRAY_SIZE(nodes), in test__mem2node() 57 for (i = 0; i < ARRAY_SIZE(nodes); i++) { in test__mem2node() 58 nodes[i].node = test_nodes[i].node; in test__mem2node() 59 nodes[i].size = 10; in test__mem2node() 62 (nodes[i].set = get_bitmap(test_nodes[i].map, 10))); in test__mem2node() 74 for (i = 0; i < ARRAY_SIZE(nodes); i++) in test__mem2node() 75 zfree(&nodes[i].set); in test__mem2node()
|
/tools/testing/selftests/net/ |
D | reuseport_bpf_numa.c | 231 int *rcv_fd, nodes; in main() local 236 nodes = numa_max_node() + 1; in main() 238 rcv_fd = calloc(nodes, sizeof(int)); in main() 243 test(rcv_fd, nodes, AF_INET, SOCK_DGRAM); in main() 246 test(rcv_fd, nodes, AF_INET6, SOCK_DGRAM); in main() 249 test(rcv_fd, nodes, AF_INET, SOCK_STREAM); in main() 252 test(rcv_fd, nodes, AF_INET6, SOCK_STREAM); in main()
|
/tools/perf/util/ |
D | mem2node.c | 49 struct memory_node *n, *nodes = &env->memory_nodes[0]; in mem2node__init() local 58 n = &nodes[i]; in mem2node__init() 69 n = &nodes[i]; in mem2node__init()
|
D | cputopo.h | 25 struct numa_topology_node nodes[]; member
|
D | cputopo.c | 324 tp = zalloc(sizeof(*tp) + sizeof(tp->nodes[0])*nr); in numa_topology__new() 331 if (load_numa_node(&tp->nodes[i], node_map->map[i])) { in numa_topology__new() 350 zfree(&tp->nodes[i].cpus); in numa_topology__delete()
|
D | header.c | 688 struct numa_topology_node *n = &tp->nodes[i]; in write_numa_topology() 1293 static int build_mem_topology(struct memory_node *nodes, u64 size, u64 *cntp) in build_mem_topology() argument 1329 ret = memory_node__read(&nodes[cnt++], idx); in build_mem_topology() 1336 qsort(nodes, cnt, sizeof(nodes[0]), memory_node__sort); in build_mem_topology() 1361 static struct memory_node nodes[MAX_MEMORY_NODES]; in write_mem_topology() local 1370 ret = build_mem_topology(&nodes[0], MAX_MEMORY_NODES, &nr); in write_mem_topology() 1387 struct memory_node *n = &nodes[i]; in write_mem_topology() 2037 struct memory_node *nodes; in print_mem_topology() local 2040 nodes = ff->ph->env.memory_nodes; in print_mem_topology() 2047 memory_node__fprintf(&nodes[i], ff->ph->env.memory_bsize, fp); in print_mem_topology() [all …]
|
/tools/perf/bench/ |
D | numa.c | 892 int nodes; in count_process_nodes() local 910 nodes = 0; in count_process_nodes() 913 nodes += node_present[n]; in count_process_nodes() 915 return nodes; in count_process_nodes() 959 unsigned int nodes = count_process_nodes(p); in calc_convergence_compression() local 961 if (!nodes) { in calc_convergence_compression() 966 nodes_min = min(nodes, nodes_min); in calc_convergence_compression() 967 nodes_max = max(nodes, nodes_max); in calc_convergence_compression() 983 int nodes[MAX_NR_NODES]; in calc_convergence() local 998 nodes[node] = 0; in calc_convergence() [all …]
|
/tools/testing/selftests/zram/ |
D | README | 11 Statistics for individual zram devices are exported through sysfs nodes at
|
/tools/perf/ |
D | builtin-c2c.c | 87 unsigned long **nodes; member 1062 bitmap_and(set, c2c_he->cpuset, c2c.nodes[node], c2c.cpus_cnt); in node_entry() 1988 unsigned long **nodes; in setup_nodes() local 2002 nodes = zalloc(sizeof(unsigned long *) * c2c.nodes_cnt); in setup_nodes() 2003 if (!nodes) in setup_nodes() 2006 c2c.nodes = nodes; in setup_nodes() 2025 nodes[node] = set; in setup_nodes()
|
/tools/perf/Documentation/ |
D | perf-c2c.txt | 244 - nodes participating on the access (see NODE INFO section) 248 The 'Node' field displays nodes that accesses given cacheline
|
D | perf.data-file-format.txt | 205 } nodes[nr]; /* Variable length records */ 303 u64 count; // number of nodes 321 # memory nodes (nr 1, block size 0x8000000):
|
D | perf-stat.txt | 285 Aggregate counts per NUMA nodes for system-wide mode measurements. This 286 is a useful mode to detect imbalance between NUMA nodes. To enable this
|