Lines Matching refs:load
705 if (unlikely(se->load.weight != NICE_0_LOAD)) in calc_delta_fair()
706 delta = __calc_delta(delta, NICE_0_LOAD, &se->load); in calc_delta_fair()
748 struct load_weight *load; in sched_slice() local
753 load = &qcfs_rq->load; in sched_slice()
756 lw = qcfs_rq->load; in sched_slice()
758 update_load_add(&lw, se->load.weight); in sched_slice()
759 load = &lw; in sched_slice()
761 slice = __calc_delta(slice, se->load.weight, load); in sched_slice()
807 sa->load_avg = scale_load_down(se->load.weight); in init_entity_runnable_average()
863 sa->util_avg = cfs_rq->avg.util_avg * se->load.weight; in post_init_entity_util_avg()
1692 unsigned long load; member
1801 ns->load += cpu_load(rq); in update_numa_stats()
1894 orig_src_load = env->src_stats.load; in load_too_imbalanced()
1895 orig_dst_load = env->dst_stats.load; in load_too_imbalanced()
1926 long load; in task_numa_compare() local
2049 load = task_h_load(env->p) - task_h_load(cur); in task_numa_compare()
2050 if (!load) in task_numa_compare()
2053 dst_load = env->dst_stats.load + load; in task_numa_compare()
2054 src_load = env->src_stats.load - load; in task_numa_compare()
2140 long src_load, dst_load, load; in task_numa_find_cpu() local
2145 load = task_h_load(env->p); in task_numa_find_cpu()
2146 dst_load = env->dst_stats.load + load; in task_numa_find_cpu()
2147 src_load = env->src_stats.load - load; in task_numa_find_cpu()
3217 update_load_add(&cfs_rq->load, se->load.weight); in account_entity_enqueue()
3234 update_load_sub(&cfs_rq->load, se->load.weight); in account_entity_dequeue()
3325 update_load_sub(&cfs_rq->load, se->load.weight); in reweight_entity()
3329 update_load_set(&se->load, weight); in reweight_entity()
3341 update_load_add(&cfs_rq->load, se->load.weight); in reweight_entity()
3349 struct load_weight *load = &se->load; in reweight_task() local
3353 load->inv_weight = sched_prio_to_wmult[prio]; in reweight_task()
3436 long tg_weight, tg_shares, load, shares; in calc_group_shares() local
3441 load = max(scale_load_down(cfs_rq->load.weight), cfs_rq->avg.load_avg); in calc_group_shares()
3447 tg_weight += load; in calc_group_shares()
3449 shares = (tg_shares * load); in calc_group_shares()
3487 if (likely(se->load.weight == shares)) in update_cfs_group()
3583 if (cfs_rq->load.weight) in cfs_rq_is_decayed()
3817 if (scale_load_down(gcfs_rq->load.weight)) { in update_tg_cfs_load()
3819 scale_load_down(gcfs_rq->load.weight)); in update_tg_cfs_load()
5016 rq_of(cfs_rq)->cfs.load.weight >= 2*se->load.weight) { in set_next_entity()
5413 if (qcfs_rq->load.weight) { in throttle_cfs_rq()
5470 if (!cfs_rq->load.weight) { in unthrottle_cfs_rq()
6252 if (cfs_rq->load.weight) { in dequeue_task_fair()
6338 unsigned int load; in cpu_load_without() local
6345 load = READ_ONCE(cfs_rq->avg.load_avg); in cpu_load_without()
6348 lsub_positive(&load, task_h_load(p)); in cpu_load_without()
6350 return load; in cpu_load_without()
6539 unsigned long load, min_load = ULONG_MAX; in find_idlest_group_cpu() local
6582 load = cpu_load(cpu_rq(i)); in find_idlest_group_cpu()
6583 if (load < min_load) { in find_idlest_group_cpu()
6584 min_load = load; in find_idlest_group_cpu()
8551 unsigned long util, load; in detach_tasks() local
8607 load = max_t(unsigned long, task_h_load(p), 1); in detach_tasks()
8610 load < 16 && !env->sd->nr_balance_failed) in detach_tasks()
8619 if (shr_bound(load, env->sd->nr_balance_failed) > env->imbalance) in detach_tasks()
8622 env->imbalance -= load; in detach_tasks()
8864 unsigned long load; in update_cfs_rq_h_load() local
8883 load = cfs_rq->h_load; in update_cfs_rq_h_load()
8884 load = div64_ul(load * se->avg.load_avg, in update_cfs_rq_h_load()
8887 cfs_rq->h_load = load; in update_cfs_rq_h_load()
9368 unsigned long load = cpu_load(rq); in update_sg_lb_stats() local
9370 sgs->group_load += load; in update_sg_lb_stats()
9409 if (sgs->group_misfit_task_load < load) in update_sg_lb_stats()
9410 sgs->group_misfit_task_load = load; in update_sg_lb_stats()
10413 unsigned long capacity, load, util; in find_busiest_queue() local
10471 load = cpu_load(rq); in find_busiest_queue()
10473 if (nr_running == 1 && load > env->imbalance && in find_busiest_queue()
10490 if (load * busiest_capacity > busiest_load * capacity) { in find_busiest_queue()
10491 busiest_load = load; in find_busiest_queue()
12396 update_load_set(&se->load, NICE_0_LOAD); in init_tg_cfs_entry()
12550 if (rq->cfs.load.weight) in get_rr_interval_fair()