• Home
  • Raw
  • Download

Lines Matching refs:local

2178 	unsigned long local = p->numa_faults_locality[1];  in update_task_scan_period()  local
2187 if (local + shared == 0 || p->numa_faults_locality[2]) { in update_task_scan_period()
2204 lr_ratio = (local * NUMA_PERIOD_SLOTS) / (local + remote); in update_task_scan_period()
2637 int local = !!(flags & TNF_FAULT_LOCAL); in task_numa_fault() local
2680 if (!priv && !local && ng && ng->active_nodes > 1 && in task_numa_fault()
2683 local = 1; in task_numa_fault()
2701 p->numa_faults_locality[local] += pages; in task_numa_fault()
8667 struct sched_group *local; /* Local group in this sd */ member
8688 .local = NULL, in init_sd_lb_stats()
9236 unsigned int local; in update_sg_wakeup_stats() local
9241 local = task_running_on_cpu(i, p); in update_sg_wakeup_stats()
9242 sgs->sum_h_nr_running += rq->cfs.h_nr_running - local; in update_sg_wakeup_stats()
9244 nr_running = rq->nr_running - local; in update_sg_wakeup_stats()
9348 struct sched_group *idlest = NULL, *local = NULL, *group = sd->groups; in find_idlest_group() local
9374 local = group; in find_idlest_group()
9394 if (!local) in find_idlest_group()
9450 if (local->sgc->max_capacity >= idlest->sgc->max_capacity) in find_idlest_group()
9504 struct sg_lb_stats *local = &sds->local_stat; in update_sd_lb_stats() local
9514 sds->local = sg; in update_sd_lb_stats()
9515 sgs = local; in update_sd_lb_stats()
9591 struct sg_lb_stats *local, *busiest; in calculate_imbalance() local
9593 local = &sds->local_stat; in calculate_imbalance()
9629 if (local->group_type == group_has_spare) { in calculate_imbalance()
9641 env->imbalance = max(local->group_capacity, local->group_util) - in calculate_imbalance()
9642 local->group_util; in calculate_imbalance()
9666 lsub_positive(&nr_diff, local->sum_nr_running); in calculate_imbalance()
9675 env->imbalance = max_t(long, 0, (local->idle_cpus - in calculate_imbalance()
9682 local->sum_nr_running + 1, local->group_weight); in calculate_imbalance()
9692 if (local->group_type < group_overloaded) { in calculate_imbalance()
9698 local->avg_load = (local->group_load * SCHED_CAPACITY_SCALE) / in calculate_imbalance()
9699 local->group_capacity; in calculate_imbalance()
9705 if (local->avg_load >= busiest->avg_load) { in calculate_imbalance()
9717 if (local->avg_load >= sds->avg_load) { in calculate_imbalance()
9735 (sds->avg_load - local->avg_load) * local->group_capacity in calculate_imbalance()
9774 struct sg_lb_stats *local, *busiest; in find_busiest_group() local
9796 local = &sds.local_stat; in find_busiest_group()
9823 if (local->group_type > busiest->group_type) in find_busiest_group()
9830 if (local->group_type == group_overloaded) { in find_busiest_group()
9835 if (local->avg_load >= busiest->avg_load) in find_busiest_group()
9846 if (local->avg_load >= sds.avg_load) in find_busiest_group()
9854 env->sd->imbalance_pct * local->avg_load) in find_busiest_group()
9859 if (sds.prefer_sibling && local->group_type == group_has_spare && in find_busiest_group()
9860 busiest->sum_nr_running > local->sum_nr_running + 1) in find_busiest_group()
9873 local->idle_cpus <= (busiest->idle_cpus + 1)) in find_busiest_group()