Searched refs:last (Results 1 – 6 of 6) sorted by relevance
/kernel/ |
D | pid.c | 127 int i, offset, max_scan, pid, last = pid_ns->last_pid; in alloc_pidmap() local 130 pid = last + 1; in alloc_pidmap() 168 (i != max_scan || pid < last || in alloc_pidmap() 169 !((last+1) & BITS_PER_PAGE_MASK))); in alloc_pidmap() 177 if (unlikely(last == offset)) in alloc_pidmap() 185 int next_pidmap(struct pid_namespace *pid_ns, int last) in next_pidmap() argument 190 offset = (last + 1) & BITS_PER_PAGE_MASK; in next_pidmap() 191 map = &pid_ns->pidmap[(last + 1)/BITS_PER_PAGE]; in next_pidmap()
|
D | sched_fair.c | 359 struct rb_node *last = rb_last(&cfs_rq->tasks_timeline); in __pick_last_entity() local 361 if (!last) in __pick_last_entity() 364 return rb_entry(last, struct sched_entity, run_node); in __pick_last_entity() 724 if (cfs_rq->last == se) in __clear_buddies() 725 cfs_rq->last = NULL; in __clear_buddies() 827 if (cfs_rq->last && wakeup_preempt_entity(cfs_rq->last, se) < 1) in pick_next_entity() 828 return cfs_rq->last; in pick_next_entity() 1364 cfs_rq_of(se)->last = se; in set_last_buddy()
|
D | sched_debug.c | 166 struct sched_entity *last; in print_cfs_rq() local 190 last = __pick_last_entity(cfs_rq); in print_cfs_rq() 191 if (last) in print_cfs_rq() 192 max_vruntime = last->vruntime; in print_cfs_rq()
|
D | sched.c | 420 struct sched_entity *curr, *next, *last; member 1852 &p->se == cfs_rq_of(&p->se)->last)) in task_hot() 7162 struct sched_group *first = NULL, *last = NULL; in init_sched_build_groups() local 7187 if (last) in init_sched_build_groups() 7188 last->next = sg; in init_sched_build_groups() 7189 last = sg; in init_sched_build_groups() 7191 last->next = first; in init_sched_build_groups()
|
/kernel/power/ |
D | Kconfig | 44 This enables code to save the last PM event point across 63 This enables some cheesy code to save the last PM event point in the
|
/kernel/trace/ |
D | Kconfig | 267 buffer giving access to the last N branches for each cpu.
|