Lines Matching refs:load
664 if (unlikely(se->load.weight != NICE_0_LOAD)) in calc_delta_fair()
665 delta = __calc_delta(delta, NICE_0_LOAD, &se->load); in calc_delta_fair()
697 struct load_weight *load; in sched_slice() local
701 load = &cfs_rq->load; in sched_slice()
704 lw = cfs_rq->load; in sched_slice()
706 update_load_add(&lw, se->load.weight); in sched_slice()
707 load = &lw; in sched_slice()
709 slice = __calc_delta(slice, se->load.weight, load); in sched_slice()
745 sa->runnable_load_avg = sa->load_avg = scale_load_down(se->load.weight); in init_entity_runnable_average()
747 se->runnable_weight = se->load.weight; in init_entity_runnable_average()
790 sa->util_avg = cfs_rq->avg.util_avg * se->load.weight; in post_init_entity_util_avg()
1481 unsigned long load; member
1498 ns->load += cpu_runnable_load(rq); in update_numa_stats()
1567 orig_src_load = env->src_stats.load; in load_too_imbalanced()
1568 orig_dst_load = env->dst_stats.load; in load_too_imbalanced()
1599 long load; in task_numa_compare() local
1679 load = task_h_load(env->p) - task_h_load(cur); in task_numa_compare()
1680 if (!load) in task_numa_compare()
1683 dst_load = env->dst_stats.load + load; in task_numa_compare()
1684 src_load = env->src_stats.load - load; in task_numa_compare()
1713 long src_load, dst_load, load; in task_numa_find_cpu() local
1717 load = task_h_load(env->p); in task_numa_find_cpu()
1718 dst_load = env->dst_stats.load + load; in task_numa_find_cpu()
1719 src_load = env->src_stats.load - load; in task_numa_find_cpu()
2759 update_load_add(&cfs_rq->load, se->load.weight); in account_entity_enqueue()
2774 update_load_sub(&cfs_rq->load, se->load.weight); in account_entity_dequeue()
2889 update_load_set(&se->load, weight); in reweight_entity()
2912 struct load_weight *load = &se->load; in reweight_task() local
2916 load->inv_weight = sched_prio_to_wmult[prio]; in reweight_task()
2996 long tg_weight, tg_shares, load, shares; in calc_group_shares() local
3001 load = max(scale_load_down(cfs_rq->load.weight), cfs_rq->avg.load_avg); in calc_group_shares()
3007 tg_weight += load; in calc_group_shares()
3009 shares = (tg_shares * load); in calc_group_shares()
3060 scale_load_down(cfs_rq->load.weight)); in calc_group_runnable()
3093 if (likely(se->load.weight == shares)) in update_cfs_group()
3334 if (scale_load_down(gcfs_rq->load.weight)) { in update_tg_cfs_runnable()
3336 scale_load_down(gcfs_rq->load.weight)); in update_tg_cfs_runnable()
4191 rq_of(cfs_rq)->cfs.load.weight >= 2*se->load.weight) { in set_next_entity()
4533 if (qcfs_rq->load.weight) in throttle_cfs_rq()
4587 if (!cfs_rq->load.weight) in unthrottle_cfs_rq()
5328 if (cfs_rq->load.weight) { in dequeue_task_fair()
5584 unsigned long load, avg_load, runnable_load; in find_idlest_group() local
5606 load = cpu_runnable_load(cpu_rq(i)); in find_idlest_group()
5607 runnable_load += load; in find_idlest_group()
5706 unsigned long load, min_load = ULONG_MAX; in find_idlest_group_cpu() local
5747 load = cpu_runnable_load(cpu_rq(i)); in find_idlest_group_cpu()
5748 if (load < min_load) { in find_idlest_group_cpu()
5749 min_load = load; in find_idlest_group_cpu()
7418 unsigned long load; in detach_tasks() local
7451 load = task_h_load(p); in detach_tasks()
7453 if (sched_feat(LB_MIN) && load < 16 && !env->sd->nr_balance_failed) in detach_tasks()
7456 if ((load / 2) > env->imbalance) in detach_tasks()
7463 env->imbalance -= load; in detach_tasks()
7613 if (cfs_rq->load.weight) in cfs_rq_is_decayed()
7678 unsigned long load; in update_cfs_rq_h_load() local
7697 load = cfs_rq->h_load; in update_cfs_rq_h_load()
7698 load = div64_ul(load * se->avg.load_avg, in update_cfs_rq_h_load()
7701 cfs_rq->h_load = load; in update_cfs_rq_h_load()
8746 unsigned long capacity, load; in find_busiest_queue() local
8800 load = cpu_runnable_load(rq); in find_busiest_queue()
8807 if (rq->nr_running == 1 && load > env->imbalance && in find_busiest_queue()
8822 if (load * busiest_capacity > busiest_load * capacity) { in find_busiest_queue()
8823 busiest_load = load; in find_busiest_queue()
10495 update_load_set(&se->load, NICE_0_LOAD); in init_tg_cfs_entry()
10562 if (rq->cfs.load.weight) in get_rr_interval_fair()