Lines Matching refs:j
1689 int i,j; in sched_numa_warn() local
1700 for (j = 0; j < nr_node_ids; j++) { in sched_numa_warn()
1701 if (!node_state(i, N_CPU) || !node_state(j, N_CPU)) in sched_numa_warn()
1702 printk(KERN_CONT "(%02d) ", node_distance(i,j)); in sched_numa_warn()
1704 printk(KERN_CONT " %02d ", node_distance(i,j)); in sched_numa_warn()
1804 int i, j; in sched_init_numa() local
1818 for_each_cpu_node_but(j, offline_node) { in sched_init_numa()
1819 int distance = node_distance(i, j); in sched_init_numa()
1842 for (i = 0, j = 0; i < nr_levels; i++, j++) { in sched_init_numa()
1843 j = find_next_bit(distance_map, NR_DISTANCE_VALUES, j); in sched_init_numa()
1844 distances[i] = j; in sched_init_numa()
1881 for_each_cpu_node_but(j, offline_node) { in sched_init_numa()
1888 masks[i][j] = mask; in sched_init_numa()
1891 if (sched_debug() && (node_distance(j, k) != node_distance(k, j))) in sched_init_numa()
1894 if (node_distance(j, k) > sched_domains_numa_distance[i]) in sched_init_numa()
1929 for (j = 1; j < nr_levels; i++, j++) { in sched_init_numa()
1934 .numa_level = j, in sched_init_numa()
1963 int i, j; in sched_reset_numa() local
1970 for_each_node(j) in sched_reset_numa()
1971 kfree(masks[i][j]); in sched_reset_numa()
2005 int i, j; in sched_domains_numa_masks_set() local
2008 for (j = 0; j < nr_node_ids; j++) { in sched_domains_numa_masks_set()
2009 if (!node_state(j, N_CPU)) in sched_domains_numa_masks_set()
2013 if (node_distance(j, node) <= sched_domains_numa_distance[i]) in sched_domains_numa_masks_set()
2014 cpumask_set_cpu(cpu, sched_domains_numa_masks[i][j]); in sched_domains_numa_masks_set()
2021 int i, j; in sched_domains_numa_masks_clear() local
2024 for (j = 0; j < nr_node_ids; j++) { in sched_domains_numa_masks_clear()
2025 if (sched_domains_numa_masks[i][j]) in sched_domains_numa_masks_clear()
2026 cpumask_clear_cpu(cpu, sched_domains_numa_masks[i][j]); in sched_domains_numa_masks_clear()
2041 int i, j = cpu_to_node(cpu), found = nr_cpu_ids; in sched_numa_find_closest() local
2049 if (!masks[i][j]) in sched_numa_find_closest()
2051 cpu = cpumask_any_and(cpus, masks[i][j]); in sched_numa_find_closest()
2068 int j; in __sdt_alloc() local
2089 for_each_cpu(j, cpu_map) { in __sdt_alloc()
2096 GFP_KERNEL, cpu_to_node(j)); in __sdt_alloc()
2100 *per_cpu_ptr(sdd->sd, j) = sd; in __sdt_alloc()
2103 GFP_KERNEL, cpu_to_node(j)); in __sdt_alloc()
2107 *per_cpu_ptr(sdd->sds, j) = sds; in __sdt_alloc()
2110 GFP_KERNEL, cpu_to_node(j)); in __sdt_alloc()
2116 *per_cpu_ptr(sdd->sg, j) = sg; in __sdt_alloc()
2119 GFP_KERNEL, cpu_to_node(j)); in __sdt_alloc()
2124 sgc->id = j; in __sdt_alloc()
2127 *per_cpu_ptr(sdd->sgc, j) = sgc; in __sdt_alloc()
2137 int j; in __sdt_free() local
2142 for_each_cpu(j, cpu_map) { in __sdt_free()
2146 sd = *per_cpu_ptr(sdd->sd, j); in __sdt_free()
2149 kfree(*per_cpu_ptr(sdd->sd, j)); in __sdt_free()
2153 kfree(*per_cpu_ptr(sdd->sds, j)); in __sdt_free()
2155 kfree(*per_cpu_ptr(sdd->sg, j)); in __sdt_free()
2157 kfree(*per_cpu_ptr(sdd->sgc, j)); in __sdt_free()
2533 int i, j, n; in partition_sched_domains_locked() local
2559 for (j = 0; j < n && !new_topology; j++) { in partition_sched_domains_locked()
2560 if (cpumask_equal(doms_cur[i], doms_new[j]) && in partition_sched_domains_locked()
2561 dattrs_equal(dattr_cur, i, dattr_new, j)) { in partition_sched_domains_locked()
2591 for (j = 0; j < n && !new_topology; j++) { in partition_sched_domains_locked()
2592 if (cpumask_equal(doms_new[i], doms_cur[j]) && in partition_sched_domains_locked()
2593 dattrs_equal(dattr_new, i, dattr_cur, j)) in partition_sched_domains_locked()
2605 for (j = 0; j < n && !sched_energy_update; j++) { in partition_sched_domains_locked()
2606 if (cpumask_equal(doms_new[i], doms_cur[j]) && in partition_sched_domains_locked()
2607 cpu_rq(cpumask_first(doms_cur[j]))->rd->pd) { in partition_sched_domains_locked()