• Home
  • Raw
  • Download

Lines Matching refs:busiest

7979 	struct sched_group *busiest;	/* Busiest group in this sd */  member
7999 .busiest = NULL, in init_sd_lb_stats()
8401 struct sg_lb_stats *busiest = &sds->busiest_stat; in update_sd_pick_busiest() local
8414 if (sgs->group_type > busiest->group_type) in update_sd_pick_busiest()
8417 if (sgs->group_type < busiest->group_type) in update_sd_pick_busiest()
8420 if (sgs->avg_load <= busiest->avg_load) in update_sd_pick_busiest()
8440 sgs->group_misfit_task_load < busiest->group_misfit_task_load) in update_sd_pick_busiest()
8458 if (!sds->busiest) in update_sd_pick_busiest()
8462 if (sched_asym_prefer(sds->busiest->asym_prefer_cpu, in update_sd_pick_busiest()
8556 sds->busiest = sg; in update_sd_lb_stats()
8633 if (!sds->busiest) in check_asym_packing()
8636 busiest_cpu = sds->busiest->asym_prefer_cpu; in check_asym_packing()
8658 struct sg_lb_stats *local, *busiest; in fix_small_imbalance() local
8661 busiest = &sds->busiest_stat; in fix_small_imbalance()
8665 else if (busiest->load_per_task > local->load_per_task) in fix_small_imbalance()
8669 (busiest->load_per_task * SCHED_CAPACITY_SCALE) / in fix_small_imbalance()
8670 busiest->group_capacity; in fix_small_imbalance()
8672 if (busiest->avg_load + scaled_busy_load_per_task >= in fix_small_imbalance()
8674 env->imbalance = busiest->load_per_task; in fix_small_imbalance()
8684 capa_now += busiest->group_capacity * in fix_small_imbalance()
8685 min(busiest->load_per_task, busiest->avg_load); in fix_small_imbalance()
8691 if (busiest->avg_load > scaled_busy_load_per_task) { in fix_small_imbalance()
8692 capa_move += busiest->group_capacity * in fix_small_imbalance()
8693 min(busiest->load_per_task, in fix_small_imbalance()
8694 busiest->avg_load - scaled_busy_load_per_task); in fix_small_imbalance()
8698 if (busiest->avg_load * busiest->group_capacity < in fix_small_imbalance()
8699 busiest->load_per_task * SCHED_CAPACITY_SCALE) { in fix_small_imbalance()
8700 tmp = (busiest->avg_load * busiest->group_capacity) / in fix_small_imbalance()
8703 tmp = (busiest->load_per_task * SCHED_CAPACITY_SCALE) / in fix_small_imbalance()
8712 env->imbalance = busiest->load_per_task; in fix_small_imbalance()
8723 busiest->group_type == group_overloaded && in fix_small_imbalance()
8724 busiest->sum_nr_running > busiest->group_weight && in fix_small_imbalance()
8726 local->group_capacity < busiest->group_capacity) in fix_small_imbalance()
8727 env->imbalance = busiest->load_per_task; in fix_small_imbalance()
8739 struct sg_lb_stats *local, *busiest; in calculate_imbalance() local
8742 busiest = &sds->busiest_stat; in calculate_imbalance()
8744 if (busiest->group_type == group_imbalanced) { in calculate_imbalance()
8749 busiest->load_per_task = in calculate_imbalance()
8750 min(busiest->load_per_task, sds->avg_load); in calculate_imbalance()
8759 if (busiest->group_type != group_misfit_task && in calculate_imbalance()
8760 (busiest->avg_load <= sds->avg_load || in calculate_imbalance()
8769 if (busiest->group_type == group_overloaded && in calculate_imbalance()
8771 load_above_capacity = busiest->sum_nr_running * SCHED_CAPACITY_SCALE; in calculate_imbalance()
8772 if (load_above_capacity > busiest->group_capacity) { in calculate_imbalance()
8773 load_above_capacity -= busiest->group_capacity; in calculate_imbalance()
8775 load_above_capacity /= busiest->group_capacity; in calculate_imbalance()
8787 max_pull = min(busiest->avg_load - sds->avg_load, load_above_capacity); in calculate_imbalance()
8791 max_pull * busiest->group_capacity, in calculate_imbalance()
8804 if (busiest->group_type == group_misfit_task && in calculate_imbalance()
8808 busiest->group_misfit_task_load); in calculate_imbalance()
8817 if (env->imbalance < busiest->load_per_task) in calculate_imbalance()
8836 struct sg_lb_stats *local, *busiest; in find_busiest_group() local
8851 trace_android_rvh_find_busiest_group(sds.busiest, env->dst_rq, in find_busiest_group()
8859 busiest = &sds.busiest_stat; in find_busiest_group()
8863 return sds.busiest; in find_busiest_group()
8866 if (!sds.busiest || busiest->sum_nr_running == 0) in find_busiest_group()
8878 if (busiest->group_type == group_imbalanced) in find_busiest_group()
8886 busiest->group_no_capacity) in find_busiest_group()
8890 if (busiest->group_type == group_misfit_task) in find_busiest_group()
8897 if (local->avg_load >= busiest->avg_load) in find_busiest_group()
8915 if ((busiest->group_type != group_overloaded) && in find_busiest_group()
8916 (local->idle_cpus <= (busiest->idle_cpus + 1))) in find_busiest_group()
8923 if (100 * busiest->avg_load <= in find_busiest_group()
8930 env->src_grp_type = busiest->group_type; in find_busiest_group()
8932 return env->imbalance ? sds.busiest : NULL; in find_busiest_group()
8945 struct rq *busiest = NULL, *rq; in find_busiest_queue() local
8985 busiest = rq; in find_busiest_queue()
9029 busiest = rq; in find_busiest_queue()
9033 return busiest; in find_busiest_queue()
9145 struct rq *busiest; in load_balance() local
9177 busiest = find_busiest_queue(&env, group); in load_balance()
9178 if (!busiest) { in load_balance()
9183 BUG_ON(busiest == env.dst_rq); in load_balance()
9187 env.src_cpu = busiest->cpu; in load_balance()
9188 env.src_rq = busiest; in load_balance()
9191 if (busiest->nr_running > 1) { in load_balance()
9199 env.loop_max = min(sysctl_sched_nr_migrate, busiest->nr_running); in load_balance()
9202 rq_lock_irqsave(busiest, &rf); in load_balance()
9203 update_rq_clock(busiest); in load_balance()
9219 rq_unlock(busiest, &rf); in load_balance()
9282 __cpumask_clear_cpu(cpu_of(busiest), cpus); in load_balance()
9315 raw_spin_lock_irqsave(&busiest->lock, flags); in load_balance()
9322 if (!cpumask_test_cpu(this_cpu, busiest->curr->cpus_ptr)) { in load_balance()
9323 raw_spin_unlock_irqrestore(&busiest->lock, in load_balance()
9334 if (!busiest->active_balance) { in load_balance()
9335 busiest->active_balance = 1; in load_balance()
9336 busiest->push_cpu = this_cpu; in load_balance()
9339 raw_spin_unlock_irqrestore(&busiest->lock, flags); in load_balance()
9342 stop_one_cpu_nowait(cpu_of(busiest), in load_balance()
9343 active_load_balance_cpu_stop, busiest, in load_balance()
9344 &busiest->active_balance_work); in load_balance()