• Home
  • Raw
  • Download

Lines Matching refs:mask

845 build_balance_mask(struct sched_domain *sd, struct sched_group *sg, struct cpumask *mask)  in build_balance_mask()  argument
852 cpumask_clear(mask); in build_balance_mask()
869 cpumask_set_cpu(i, mask); in build_balance_mask()
873 WARN_ON_ONCE(cpumask_empty(mask)); in build_balance_mask()
906 struct cpumask *mask = sched_domains_tmpmask2; in init_overlap_sched_group() local
911 build_balance_mask(sd, sg, mask); in init_overlap_sched_group()
912 cpu = cpumask_first_and(sched_group_span(sg), mask); in init_overlap_sched_group()
916 cpumask_copy(group_balance_mask(sg), mask); in init_overlap_sched_group()
918 WARN_ON_ONCE(!cpumask_equal(group_balance_mask(sg), mask)); in init_overlap_sched_group()
1330 sd_weight = cpumask_weight(tl->mask(cpu)); in sd_init()
1372 cpumask_and(sched_domain_span(sd), cpu_map, tl->mask(cpu)); in sd_init()
1440 for (tl = sched_domain_topology; tl->mask; tl++)
1628 struct cpumask *mask = kzalloc(cpumask_size(), GFP_KERNEL); in sched_init_numa() local
1631 if (!mask) in sched_init_numa()
1634 sched_domains_numa_masks[i][j] = mask; in sched_init_numa()
1643 cpumask_or(mask, mask, cpumask_of_node(k)); in sched_init_numa()
1649 for (i = 0; sched_domain_topology[i].mask; i++); in sched_init_numa()
1659 for (i = 0; sched_domain_topology[i].mask; i++) in sched_init_numa()
1666 .mask = sd_numa_mask, in sched_init_numa()
1676 .mask = sd_numa_mask, in sched_init_numa()
1900 if (!cpumask_equal(tl->mask(cpu), tl->mask(i)) && in topology_span_sane()
1901 cpumask_intersects(tl->mask(cpu), tl->mask(i))) in topology_span_sane()
1946 for_each_cpu_and(j, tl->mask(i), cpu_map) { in asym_cpu_capacity_level()