Lines Matching refs:load
710 if (unlikely(se->load.weight != NICE_0_LOAD)) { in calc_delta_fair()
711 delta = fair_calc_delta(delta, NICE_0_LOAD, &se->load); in calc_delta_fair()
752 struct load_weight *load; in sched_slice() local
756 load = &cfs_rq->load; in sched_slice()
759 lw = cfs_rq->load; in sched_slice()
761 update_load_add(&lw, se->load.weight); in sched_slice()
762 load = &lw; in sched_slice()
764 slice = fair_calc_delta(slice, se->load.weight, load); in sched_slice()
804 sa->load_avg = scale_load_down(se->load.weight); in init_entity_runnable_average()
848 sa->util_avg = cfs_rq->avg.util_avg * se->load.weight; in post_init_entity_util_avg()
1568 unsigned long load; member
1680 ns->load += cpu_load(rq); in update_numa_stats()
1776 orig_src_load = env->src_stats.load; in load_too_imbalanced()
1777 orig_dst_load = env->dst_stats.load; in load_too_imbalanced()
1807 long load; in task_numa_compare() local
1926 load = task_h_load(env->p) - task_h_load(cur); in task_numa_compare()
1927 if (!load) { in task_numa_compare()
1931 dst_load = env->dst_stats.load + load; in task_numa_compare()
1932 src_load = env->src_stats.load - load; in task_numa_compare()
2016 long src_load, dst_load, load; in task_numa_find_cpu() local
2021 load = task_h_load(env->p); in task_numa_find_cpu()
2022 dst_load = env->dst_stats.load + load; in task_numa_find_cpu()
2023 src_load = env->src_stats.load - load; in task_numa_find_cpu()
3109 update_load_add(&cfs_rq->load, se->load.weight); in account_entity_enqueue()
3123 update_load_sub(&cfs_rq->load, se->load.weight); in account_entity_dequeue()
3212 update_load_sub(&cfs_rq->load, se->load.weight); in reweight_entity()
3216 update_load_set(&se->load, weight); in reweight_entity()
3228 update_load_add(&cfs_rq->load, se->load.weight); in reweight_entity()
3236 struct load_weight *load = &se->load; in reweight_task() local
3240 load->inv_weight = sched_prio_to_wmult[prio]; in reweight_task()
3320 long tg_weight, tg_shares, load, shares; in calc_group_shares() local
3325 load = max(scale_load_down(cfs_rq->load.weight), cfs_rq->avg.load_avg); in calc_group_shares()
3331 tg_weight += load; in calc_group_shares()
3333 shares = (tg_shares * load); in calc_group_shares()
3375 if (likely(se->load.weight == shares)) { in update_cfs_group()
3644 if (scale_load_down(gcfs_rq->load.weight)) { in update_tg_cfs_load()
3645 load_sum = div_s64(gcfs_rq->avg.load_sum, scale_load_down(gcfs_rq->load.weight)); in update_tg_cfs_load()
4651 if (schedstat_enabled() && rq_of(cfs_rq)->cfs.load.weight >= 0x2 * se->load.weight) { in set_next_entity()
5037 if (qcfs_rq->load.weight) { in throttle_cfs_rq()
5078 if (!cfs_rq->load.weight) { in unthrottle_cfs_rq()
5936 if (cfs_rq->load.weight) { in dequeue_task_fair()
6021 unsigned int load; in cpu_load_without() local
6029 load = READ_ONCE(cfs_rq->avg.load_avg); in cpu_load_without()
6032 lsub_positive(&load, task_h_load(p)); in cpu_load_without()
6034 return load; in cpu_load_without()
6220 unsigned long load, min_load = ULONG_MAX; in find_idlest_group_cpu() local
6265 load = cpu_load(cpu_rq(i)); in find_idlest_group_cpu()
6266 if (load < min_load) { in find_idlest_group_cpu()
6267 min_load = load; in find_idlest_group_cpu()
8146 unsigned long util, load; in detach_tasks() local
8203 load = max_t(unsigned long, task_h_load(p), 1); in detach_tasks()
8205 if (sched_feat(LB_MIN) && load < 0x10 && !env->sd->nr_balance_failed) { in detach_tasks()
8215 if (shr_bound(load, env->sd->nr_balance_failed) > env->imbalance) { in detach_tasks()
8219 env->imbalance -= load; in detach_tasks()
8431 if (cfs_rq->load.weight) { in cfs_rq_is_decayed()
8504 unsigned long load; in update_cfs_rq_h_load() local
8525 load = cfs_rq->h_load; in update_cfs_rq_h_load()
8526 load = div64_ul(load * se->avg.load_avg, cfs_rq_load_avg(cfs_rq) + 1); in update_cfs_rq_h_load()
8528 cfs_rq->h_load = load; in update_cfs_rq_h_load()
9923 unsigned long capacity, load, util; in find_busiest_queue() local
9975 load = cpu_load(rq); in find_busiest_queue()
9976 if (nr_running == 1 && load > env->imbalance && !check_cpu_capacity(rq, env->sd)) { in find_busiest_queue()
9993 if (load * busiest_capacity > busiest_load * capacity) { in find_busiest_queue()
9994 busiest_load = load; in find_busiest_queue()
11982 update_load_set(&se->load, NICE_0_LOAD); in init_tg_cfs_entry()
12056 if (rq->cfs.load.weight) { in get_rr_interval_fair()