Lines Matching refs:curr
247 rq->curr->sched_class->task_tick(rq, rq->curr, 1); in hrtick()
510 struct task_struct *curr = rq->curr; in resched_curr() local
515 if (test_tsk_need_resched(curr)) in resched_curr()
521 set_tsk_need_resched(curr); in resched_curr()
526 if (set_nr_and_not_polling(curr)) in resched_curr()
1421 if (p->sched_class == rq->curr->sched_class) { in check_preempt_curr()
1422 rq->curr->sched_class->check_preempt_curr(rq, p, flags); in check_preempt_curr()
1425 if (class == rq->curr->sched_class) in check_preempt_curr()
1438 if (task_on_rq_queued(rq->curr) && test_tsk_need_resched(rq->curr)) in check_preempt_curr()
2372 if (!is_idle_task(rcu_dereference(rq->curr))) in wake_up_if_idle()
2379 if (is_idle_task(rq->curr)) in wake_up_if_idle()
3025 static void __fire_sched_in_preempt_notifiers(struct task_struct *curr) in __fire_sched_in_preempt_notifiers() argument
3029 hlist_for_each_entry(notifier, &curr->preempt_notifiers, link) in __fire_sched_in_preempt_notifiers()
3033 static __always_inline void fire_sched_in_preempt_notifiers(struct task_struct *curr) in fire_sched_in_preempt_notifiers() argument
3036 __fire_sched_in_preempt_notifiers(curr); in fire_sched_in_preempt_notifiers()
3040 __fire_sched_out_preempt_notifiers(struct task_struct *curr, in __fire_sched_out_preempt_notifiers() argument
3045 hlist_for_each_entry(notifier, &curr->preempt_notifiers, link) in __fire_sched_out_preempt_notifiers()
3050 fire_sched_out_preempt_notifiers(struct task_struct *curr, in fire_sched_out_preempt_notifiers() argument
3054 __fire_sched_out_preempt_notifiers(curr, next); in fire_sched_out_preempt_notifiers()
3059 static inline void fire_sched_in_preempt_notifiers(struct task_struct *curr) in fire_sched_in_preempt_notifiers() argument
3064 fire_sched_out_preempt_notifiers(struct task_struct *curr, in fire_sched_out_preempt_notifiers() argument
3534 struct sched_entity *curr = (&p->se)->cfs_rq->curr; in prefetch_curr_exec_start() local
3536 struct sched_entity *curr = (&task_rq(p)->cfs)->curr; in prefetch_curr_exec_start()
3538 prefetch(curr); in prefetch_curr_exec_start()
3539 prefetch(&curr->exec_start); in prefetch_curr_exec_start()
3594 struct task_struct *curr = rq->curr; in scheduler_tick() local
3602 curr->sched_class->task_tick(rq, curr, 0); in scheduler_tick()
3659 struct task_struct *curr; in sched_tick_remote() local
3675 curr = rq->curr; in sched_tick_remote()
3676 if (is_idle_task(curr) || cpu_is_offline(cpu)) in sched_tick_remote()
3680 delta = rq_clock_task(rq) - curr->se.exec_start; in sched_tick_remote()
3687 curr->sched_class->task_tick(rq, curr, 0); in sched_tick_remote()
4009 prev = rq->curr; in __schedule()
4059 RCU_INIT_POINTER(rq->curr, next); in __schedule()
4337 int default_wake_function(wait_queue_entry_t *curr, unsigned mode, int wake_flags, in default_wake_function() argument
4340 return try_to_wake_up(curr->private, mode, wake_flags); in default_wake_function()
4422 WARN_ON(p != rq->curr); in rt_mutex_setprio()
4622 if (rq->curr != rq->idle) in idle_cpu()
5696 struct task_struct *curr = current; in yield_to() local
5721 if (!curr->sched_class->yield_to_task) in yield_to()
5724 if (curr->sched_class != p->sched_class) in yield_to()
5730 yielded = curr->sched_class->yield_to_task(rq, p, preempt); in yield_to()
6059 rcu_assign_pointer(rq->curr, idle); in init_idle()