/kernel/linux/linux-5.10/kernel/irq/ |
D | affinity.c | 48 masks = kcalloc(nr_node_ids, sizeof(cpumask_var_t), GFP_KERNEL); in alloc_node_to_cpumask() 52 for (node = 0; node < nr_node_ids; node++) { in alloc_node_to_cpumask() 70 for (node = 0; node < nr_node_ids; node++) in free_node_to_cpumask() 137 for (n = 0; n < nr_node_ids; n++) { in alloc_nodes_vectors() 156 sort(node_vectors, nr_node_ids, sizeof(node_vectors[0]), in alloc_nodes_vectors() 227 for (n = 0; n < nr_node_ids; n++) { in alloc_nodes_vectors() 280 node_vectors = kcalloc(nr_node_ids, in __irq_build_affinity_masks() 290 for (i = 0; i < nr_node_ids; i++) { in __irq_build_affinity_masks()
|
/kernel/linux/linux-5.10/arch/arm64/mm/ |
D | numa.c | 53 if (WARN_ON(node < 0 || node >= nr_node_ids)) in cpumask_of_node() 106 if (nr_node_ids == MAX_NUMNODES) in setup_node_to_cpumask_map() 110 for (node = 0; node < nr_node_ids; node++) { in setup_node_to_cpumask_map() 116 pr_debug("Node to cpumask map for %u nodes\n", nr_node_ids); in setup_node_to_cpumask_map() 282 size = nr_node_ids * nr_node_ids * sizeof(numa_distance[0]); in numa_alloc_distance() 291 numa_distance_cnt = nr_node_ids; in numa_alloc_distance()
|
/kernel/linux/linux-5.10/arch/x86/mm/ |
D | numa.c | 114 if (nr_node_ids == MAX_NUMNODES) in setup_node_to_cpumask_map() 118 for (node = 0; node < nr_node_ids; node++) in setup_node_to_cpumask_map() 122 pr_debug("Node to cpumask map for %u nodes\n", nr_node_ids); in setup_node_to_cpumask_map() 903 if ((unsigned)node >= nr_node_ids) { in cpumask_of_node() 906 node, nr_node_ids); in cpumask_of_node()
|
/kernel/linux/linux-5.10/arch/powerpc/mm/ |
D | numa.c | 74 if (nr_node_ids == MAX_NUMNODES) in setup_node_to_cpumask_map() 82 dbg("Node to cpumask map for %u nodes\n", nr_node_ids); in setup_node_to_cpumask_map() 239 if (nid == 0xffff || nid >= nr_node_ids) in associativity_to_nid() 452 if (nid == 0xffff || nid >= nr_node_ids) in of_drconf_to_nid_single()
|
/kernel/linux/linux-5.10/kernel/sched/ |
D | topology.c | 1553 for (i = 0; i < nr_node_ids; i++) { in sched_numa_warn() 1555 for (j = 0; j < nr_node_ids; j++) in sched_numa_warn() 1647 for (i = 0; i < nr_node_ids; i++) { in sched_init_numa() 1648 for (j = 0; j < nr_node_ids; j++) { in sched_init_numa() 1706 kzalloc(nr_node_ids * sizeof(void *), GFP_KERNEL); in sched_init_numa() 1710 for (j = 0; j < nr_node_ids; j++) { in sched_init_numa() 1781 for (j = 0; j < nr_node_ids; j++) { in sched_domains_numa_masks_set() 1793 for (j = 0; j < nr_node_ids; j++) in sched_domains_numa_masks_clear()
|
D | fair.c | 1280 return NR_NUMA_HINT_FAULT_TYPES * (s * nr_node_ids + nid) + priv; in task_faults_idx() 2524 4*nr_node_ids*sizeof(unsigned long); in task_numa_group() 2537 nr_node_ids; in task_numa_group() 2539 for (i = 0; i < NR_NUMA_HINT_FAULT_STATS * nr_node_ids; i++) in task_numa_group() 2597 for (i = 0; i < NR_NUMA_HINT_FAULT_STATS * nr_node_ids; i++) { in task_numa_group() 2640 for (i = 0; i < NR_NUMA_HINT_FAULT_STATS * nr_node_ids; i++) in task_numa_free() 2655 for (i = 0; i < NR_NUMA_HINT_FAULT_STATS * nr_node_ids; i++) in task_numa_free() 2682 NR_NUMA_HINT_FAULT_BUCKETS * nr_node_ids; in task_numa_fault()
|
/kernel/linux/linux-5.10/arch/x86/kernel/ |
D | setup_percpu.c | 175 NR_CPUS, nr_cpumask_bits, nr_cpu_ids, nr_node_ids); in setup_per_cpu_areas()
|
/kernel/linux/linux-5.10/include/linux/ |
D | nodemask.h | 448 extern unsigned int nr_node_ids; 489 #define nr_node_ids 1U macro
|
/kernel/linux/linux-5.10/drivers/hv/ |
D | hv.c | 89 hv_context.hv_numa_map = kcalloc(nr_node_ids, sizeof(struct cpumask), in hv_synic_alloc()
|
D | channel_mgmt.c | 699 if (numa_node == nr_node_ids) { in init_vp_index()
|
/kernel/linux/linux-5.10/mm/ |
D | ksm.c | 2963 buf = kcalloc(nr_node_ids + nr_node_ids, sizeof(*buf), in merge_across_nodes_store() 2970 root_unstable_tree = buf + nr_node_ids; in merge_across_nodes_store() 2977 ksm_nr_node_ids = knob ? 1 : nr_node_ids; in merge_across_nodes_store()
|
D | slab.h | 585 for (__node = 0; __node < nr_node_ids; __node++) \
|
D | list_lru.c | 604 lru->node = kcalloc(nr_node_ids, sizeof(*lru->node), GFP_KERNEL); in __list_lru_init()
|
D | mempolicy.c | 1442 unsigned int nbytes = BITS_TO_LONGS(nr_node_ids) * sizeof(long); in copy_nodes_to_user() 1616 if (nmask != NULL && maxnode < nr_node_ids) in kernel_get_mempolicy() 1654 nr_bits = min_t(unsigned long, maxnode-1, nr_node_ids); in COMPAT_SYSCALL_DEFINE5()
|
D | slub.c | 4384 nr_node_ids * sizeof(struct kmem_cache_node *), in kmem_cache_init() 4403 nr_cpu_ids, nr_node_ids); in kmem_cache_init() 4913 nodes = kcalloc(nr_node_ids, sizeof(unsigned long), GFP_KERNEL); in show_slab_objects() 5002 for (node = 0; node < nr_node_ids; node++) in show_slab_objects()
|
D | vmalloc.c | 3475 memset(counters, 0, nr_node_ids * sizeof(unsigned int)); in show_numa_info() 3580 nr_node_ids * sizeof(unsigned int), NULL); in proc_vmalloc_init()
|
D | swapfile.c | 2889 p = kvzalloc(struct_size(p, avail_lists, nr_node_ids), GFP_KERNEL); in alloc_swap_info() 3871 swap_avail_heads = kmalloc_array(nr_node_ids, sizeof(struct plist_head), in swapfile_init()
|
D | slab.c | 665 alc_ptr = kcalloc_node(nr_node_ids, sizeof(void *), gfp, node); in alloc_alien_cache() 1249 nr_node_ids * sizeof(struct kmem_cache_node *), in kmem_cache_init()
|
D | page_alloc.c | 390 unsigned int nr_node_ids __read_mostly = MAX_NUMNODES; 392 EXPORT_SYMBOL(nr_node_ids); 7206 nr_node_ids = highest + 1; in setup_nr_node_ids()
|
/kernel/linux/linux-5.10/drivers/infiniband/sw/siw/ |
D | siw_main.c | 138 int i, num_nodes = nr_node_ids; in siw_init_cpulist()
|
/kernel/linux/linux-5.10/kernel/ |
D | workqueue.c | 3937 ctx = kzalloc(struct_size(ctx, pwq_tbl, nr_node_ids), GFP_KERNEL); in apply_wqattrs_prepare() 4286 tbl_size = nr_node_ids * sizeof(wq->numa_pwq_tbl[0]); in alloc_workqueue() 5927 tbl = kcalloc(nr_node_ids, sizeof(tbl[0]), GFP_KERNEL); in wq_numa_init()
|
/kernel/linux/linux-5.10/fs/ |
D | io-wq.c | 1096 wq->wqes = kcalloc(nr_node_ids, sizeof(struct io_wqe *), GFP_KERNEL); in io_wq_create()
|
/kernel/linux/linux-5.10/drivers/net/ethernet/cavium/thunder/ |
D | nic_main.c | 1244 if (nr_node_ids > 1) in nic_num_sqs_en()
|
/kernel/linux/linux-5.10/net/sunrpc/ |
D | svc.c | 196 unsigned int maxpools = nr_node_ids; in svc_pool_map_init_pernode()
|
/kernel/linux/linux-5.10/drivers/char/ |
D | random.c | 879 pool = kcalloc(nr_node_ids, sizeof(*pool), GFP_KERNEL|__GFP_NOFAIL); in do_numa_crng_init()
|