Lines Matching refs:mask
843 build_balance_mask(struct sched_domain *sd, struct sched_group *sg, struct cpumask *mask) in build_balance_mask() argument
850 cpumask_clear(mask); in build_balance_mask()
867 cpumask_set_cpu(i, mask); in build_balance_mask()
871 WARN_ON_ONCE(cpumask_empty(mask)); in build_balance_mask()
904 struct cpumask *mask = sched_domains_tmpmask2; in init_overlap_sched_group() local
909 build_balance_mask(sd, sg, mask); in init_overlap_sched_group()
910 cpu = cpumask_first_and(sched_group_span(sg), mask); in init_overlap_sched_group()
914 cpumask_copy(group_balance_mask(sg), mask); in init_overlap_sched_group()
916 WARN_ON_ONCE(!cpumask_equal(group_balance_mask(sg), mask)); in init_overlap_sched_group()
1333 sd_weight = cpumask_weight(tl->mask(cpu)); in sd_init()
1376 cpumask_and(sched_domain_span(sd), cpu_map, tl->mask(cpu)); in sd_init()
1453 for (tl = sched_domain_topology; tl->mask; tl++)
1649 struct cpumask *mask = kzalloc(cpumask_size(), GFP_KERNEL); in sched_init_numa() local
1650 if (!mask) in sched_init_numa()
1653 sched_domains_numa_masks[i][j] = mask; in sched_init_numa()
1659 cpumask_or(mask, mask, cpumask_of_node(k)); in sched_init_numa()
1665 for (i = 0; sched_domain_topology[i].mask; i++); in sched_init_numa()
1675 for (i = 0; sched_domain_topology[i].mask; i++) in sched_init_numa()
1682 .mask = sd_numa_mask, in sched_init_numa()
1692 .mask = sd_numa_mask, in sched_init_numa()
1926 for_each_cpu_and(j, tl->mask(i), cpu_map) { in asym_cpu_capacity_level()