Home
last modified time | relevance | path

Searched refs:cur (Results 1 – 21 of 21) sorted by relevance

/kernel/
Dcrash_core.c41 char *cur = cmdline, *tmp; in parse_crashkernel_mem() local
48 start = memparse(cur, &tmp); in parse_crashkernel_mem()
49 if (cur == tmp) { in parse_crashkernel_mem()
53 cur = tmp; in parse_crashkernel_mem()
54 if (*cur != '-') { in parse_crashkernel_mem()
58 cur++; in parse_crashkernel_mem()
61 if (*cur != ':') { in parse_crashkernel_mem()
62 end = memparse(cur, &tmp); in parse_crashkernel_mem()
63 if (cur == tmp) { in parse_crashkernel_mem()
67 cur = tmp; in parse_crashkernel_mem()
[all …]
Dseccomp.c1014 struct seccomp_knotif *knotif = NULL, *cur; in seccomp_notify_recv() local
1032 list_for_each_entry(cur, &filter->notif->notifications, list) { in seccomp_notify_recv()
1033 if (cur->state == SECCOMP_NOTIFY_INIT) { in seccomp_notify_recv()
1034 knotif = cur; in seccomp_notify_recv()
1070 list_for_each_entry(cur, &filter->notif->notifications, list) { in seccomp_notify_recv()
1071 if (cur->id == unotif.id) { in seccomp_notify_recv()
1072 knotif = cur; in seccomp_notify_recv()
1091 struct seccomp_knotif *knotif = NULL, *cur; in seccomp_notify_send() local
1104 list_for_each_entry(cur, &filter->notif->notifications, list) { in seccomp_notify_send()
1105 if (cur->id == resp.id) { in seccomp_notify_send()
[all …]
Dsmpboot.c215 struct smp_hotplug_thread *cur; in smpboot_create_threads() local
219 list_for_each_entry(cur, &hotplug_threads, list) { in smpboot_create_threads()
220 ret = __smpboot_create_thread(cur, cpu); in smpboot_create_threads()
238 struct smp_hotplug_thread *cur; in smpboot_unpark_threads() local
241 list_for_each_entry(cur, &hotplug_threads, list) in smpboot_unpark_threads()
242 smpboot_unpark_thread(cur, cpu); in smpboot_unpark_threads()
257 struct smp_hotplug_thread *cur; in smpboot_park_threads() local
260 list_for_each_entry_reverse(cur, &hotplug_threads, list) in smpboot_park_threads()
261 smpboot_park_thread(cur, cpu); in smpboot_park_threads()
Dpadata.c322 struct padata_priv *cur; in padata_do_serial() local
326 list_for_each_entry_reverse(cur, &pqueue->reorder.list, list) in padata_do_serial()
327 if (cur->seq_nr < padata->seq_nr) in padata_do_serial()
329 list_add(&padata->list, &cur->list); in padata_do_serial()
Dkprobes.c1129 struct kprobe *cur = __this_cpu_read(kprobe_instance); in aggr_fault_handler() local
1135 if (cur && cur->fault_handler) { in aggr_fault_handler()
1136 if (cur->fault_handler(cur, regs, trapnr)) in aggr_fault_handler()
Dcpu.c2114 ssize_t cur, res = 0; in show_cpuhp_states() local
2122 cur = sprintf(buf, "%3d: %s\n", i, sp->name); in show_cpuhp_states()
2123 buf += cur; in show_cpuhp_states()
2124 res += cur; in show_cpuhp_states()
/kernel/power/
Dsnapshot.c382 struct bm_position cur; /* most recently used bit position */ member
550 bm->cur.zone = list_entry(bm->zones.next, struct mem_zone_bm_rtree, in memory_bm_position_reset()
552 bm->cur.node = list_entry(bm->cur.zone->leaves.next, in memory_bm_position_reset()
554 bm->cur.node_pfn = 0; in memory_bm_position_reset()
555 bm->cur.node_bit = 0; in memory_bm_position_reset()
595 struct mem_extent *ext, *cur, *aux; in create_mem_extents() local
626 cur = ext; in create_mem_extents()
627 list_for_each_entry_safe_continue(cur, aux, list, hook) { in create_mem_extents()
628 if (zone_end < cur->start) in create_mem_extents()
630 if (zone_end < cur->end) in create_mem_extents()
[all …]
Dswap.c96 struct swap_map_page *cur; member
404 if (handle->cur) in release_swap_writer()
405 free_page((unsigned long)handle->cur); in release_swap_writer()
406 handle->cur = NULL; in release_swap_writer()
419 handle->cur = (struct swap_map_page *)get_zeroed_page(GFP_KERNEL); in get_swap_writer()
420 if (!handle->cur) { in get_swap_writer()
446 if (!handle->cur) in swap_write_page()
452 handle->cur->entries[handle->k++] = offset; in swap_write_page()
457 handle->cur->next_swap = offset; in swap_write_page()
458 error = write_page(handle->cur, handle->cur_swap, hb); in swap_write_page()
[all …]
Dpower.h132 unsigned int cur; /* number of the block of PAGE_SIZE bytes the member
/kernel/locking/
Dmutex.c446 struct mutex_waiter *cur; in __ww_mutex_check_waiters() local
450 list_for_each_entry(cur, &lock->wait_list, list) { in __ww_mutex_check_waiters()
451 if (!cur->ww_ctx) in __ww_mutex_check_waiters()
454 if (__ww_mutex_die(lock, cur, ww_ctx) || in __ww_mutex_check_waiters()
455 __ww_mutex_wound(lock, cur->ww_ctx, ww_ctx)) in __ww_mutex_check_waiters()
810 struct mutex_waiter *cur; in __ww_mutex_check_kill() local
829 cur = waiter; in __ww_mutex_check_kill()
830 list_for_each_entry_continue_reverse(cur, &lock->wait_list, list) { in __ww_mutex_check_kill()
831 if (!cur->ww_ctx) in __ww_mutex_check_kill()
856 struct mutex_waiter *cur; in __ww_mutex_add_waiter() local
[all …]
/kernel/bpf/
Dbpf_lru_list.c170 struct list_head *cur, *last, *next = inactive; in __bpf_lru_list_rotate_inactive() local
181 cur = l->next_inactive_rotation; in __bpf_lru_list_rotate_inactive()
183 if (cur == inactive) { in __bpf_lru_list_rotate_inactive()
184 cur = cur->prev; in __bpf_lru_list_rotate_inactive()
188 node = list_entry(cur, struct bpf_lru_node, list); in __bpf_lru_list_rotate_inactive()
189 next = cur->prev; in __bpf_lru_list_rotate_inactive()
192 if (cur == last) in __bpf_lru_list_rotate_inactive()
194 cur = next; in __bpf_lru_list_rotate_inactive()
Dverifier.c428 struct bpf_verifier_state *cur = env->cur_state; in func() local
430 return cur->frame[reg->frameno]; in func()
779 struct bpf_verifier_state *cur = env->cur_state; in pop_stack() local
786 if (cur) { in pop_stack()
787 err = copy_verifier_state(cur, &head->st); in pop_stack()
807 struct bpf_verifier_state *cur = env->cur_state; in push_stack() local
820 err = copy_verifier_state(&elem->st, cur); in push_stack()
1417 struct bpf_verifier_state *cur) in push_jmp_history() argument
1419 u32 cnt = cur->jmp_history_cnt; in push_jmp_history()
1423 p = krealloc(cur->jmp_history, cnt * sizeof(*p), GFP_USER); in push_jmp_history()
[all …]
Dbtf.c2970 void *cur, *end; in btf_check_all_metas() local
2973 cur = btf->nohdr_data + hdr->type_off; in btf_check_all_metas()
2974 end = cur + hdr->type_len; in btf_check_all_metas()
2977 while (cur < end) { in btf_check_all_metas()
2978 struct btf_type *t = cur; in btf_check_all_metas()
2981 meta_size = btf_check_meta(env, t, end - cur); in btf_check_all_metas()
2986 cur += meta_size; in btf_check_all_metas()
/kernel/time/
Dtick-broadcast.c96 struct clock_event_device *cur = tick_broadcast_device.evtdev; in tick_install_broadcast_device() local
98 if (!tick_check_broadcast_device(cur, dev)) in tick_install_broadcast_device()
104 clockevents_exchange_device(cur, dev); in tick_install_broadcast_device()
105 if (cur) in tick_install_broadcast_device()
106 cur->event_handler = clockevents_handle_noop; in tick_install_broadcast_device()
/kernel/cgroup/
Dcpuset.c563 static int validate_change(struct cpuset *cur, struct cpuset *trial) in validate_change() argument
573 cpuset_for_each_child(c, css, cur) in validate_change()
579 if (cur == &top_cpuset) in validate_change()
582 par = parent_cs(cur); in validate_change()
596 c != cur && in validate_change()
600 c != cur && in validate_change()
610 if ((cgroup_is_populated(cur->css.cgroup) || cur->attach_in_progress)) { in validate_change()
611 if (!cpumask_empty(cur->cpus_allowed) && in validate_change()
614 if (!nodes_empty(cur->mems_allowed) && in validate_change()
624 if (is_cpu_exclusive(cur) && in validate_change()
[all …]
/kernel/sched/
Dfair.c1595 struct task_struct *cur; in task_numa_compare() local
1605 cur = rcu_dereference(dst_rq->curr); in task_numa_compare()
1606 if (cur && ((cur->flags & PF_EXITING) || is_idle_task(cur))) in task_numa_compare()
1607 cur = NULL; in task_numa_compare()
1613 if (cur == env->p) in task_numa_compare()
1616 if (!cur) { in task_numa_compare()
1631 if (!cpumask_test_cpu(env->src_cpu, cur->cpus_ptr)) in task_numa_compare()
1638 cur_ng = rcu_dereference(cur->numa_group); in task_numa_compare()
1640 imp = taskimp + task_weight(cur, env->src_nid, dist) - in task_numa_compare()
1641 task_weight(cur, env->dst_nid, dist); in task_numa_compare()
[all …]
Dcpufreq_schedutil.c127 policy->cur = next_freq; in sugov_fast_switch()
174 policy->cpuinfo.max_freq : policy->cur; in get_next_freq()
Dtopology.c2131 static int dattrs_equal(struct sched_domain_attr *cur, int idx_cur, in dattrs_equal() argument
2137 if (!new && !cur) in dattrs_equal()
2142 return !memcmp(cur ? (cur + idx_cur) : &tmp, in dattrs_equal()
Dcore.c1839 int migrate_swap(struct task_struct *cur, struct task_struct *p, in migrate_swap() argument
1846 .src_task = cur, in migrate_swap()
1868 trace_sched_swap_numa(cur, arg.src_cpu, p, arg.dst_cpu); in migrate_swap()
6083 int cpuset_cpumask_can_shrink(const struct cpumask *cur, in cpuset_cpumask_can_shrink() argument
6088 if (!cpumask_weight(cur)) in cpuset_cpumask_can_shrink()
6091 ret = dl_cpuset_cpumask_can_shrink(cur, trial); in cpuset_cpumask_can_shrink()
Ddeadline.c2749 int dl_cpuset_cpumask_can_shrink(const struct cpumask *cur, in dl_cpuset_cpumask_can_shrink() argument
2757 cur_dl_b = dl_bw_of(cpumask_any(cur)); in dl_cpuset_cpumask_can_shrink()
Dsched.h318 extern int dl_cpuset_cpumask_can_shrink(const struct cpumask *cur, const struct cpumask *trial);