Home
last modified time | relevance | path

Searched refs:cur (Results 1 – 25 of 26) sorted by relevance

12

/kernel/
Dcrash_core.c48 char *cur = cmdline, *tmp; in parse_crashkernel_mem() local
64 start = memparse(cur, &tmp); in parse_crashkernel_mem()
65 if (cur == tmp) { in parse_crashkernel_mem()
69 cur = tmp; in parse_crashkernel_mem()
70 if (*cur != '-') { in parse_crashkernel_mem()
74 cur++; in parse_crashkernel_mem()
77 if (*cur != ':') { in parse_crashkernel_mem()
78 end = memparse(cur, &tmp); in parse_crashkernel_mem()
79 if (cur == tmp) { in parse_crashkernel_mem()
83 cur = tmp; in parse_crashkernel_mem()
[all …]
Dsmpboot.c215 struct smp_hotplug_thread *cur; in smpboot_create_threads() local
219 list_for_each_entry(cur, &hotplug_threads, list) { in smpboot_create_threads()
220 ret = __smpboot_create_thread(cur, cpu); in smpboot_create_threads()
238 struct smp_hotplug_thread *cur; in smpboot_unpark_threads() local
241 list_for_each_entry(cur, &hotplug_threads, list) in smpboot_unpark_threads()
242 smpboot_unpark_thread(cur, cpu); in smpboot_unpark_threads()
257 struct smp_hotplug_thread *cur; in smpboot_park_threads() local
260 list_for_each_entry_reverse(cur, &hotplug_threads, list) in smpboot_park_threads()
261 smpboot_park_thread(cur, cpu); in smpboot_park_threads()
Dseccomp.c1441 struct seccomp_knotif *cur; in find_notification() local
1445 list_for_each_entry(cur, &filter->notif->notifications, list) { in find_notification()
1446 if (cur->id == id) in find_notification()
1447 return cur; in find_notification()
1457 struct seccomp_knotif *knotif = NULL, *cur; in seccomp_notify_recv() local
1475 list_for_each_entry(cur, &filter->notif->notifications, list) { in seccomp_notify_recv()
1476 if (cur->state == SECCOMP_NOTIFY_INIT) { in seccomp_notify_recv()
1477 knotif = cur; in seccomp_notify_recv()
1740 struct seccomp_knotif *cur; in seccomp_notify_poll() local
1747 list_for_each_entry(cur, &filter->notif->notifications, list) { in seccomp_notify_poll()
[all …]
Dresource.c1434 struct resource *cur; in merge_system_ram_resource() local
1443 cur = res->sibling; in merge_system_ram_resource()
1444 if (cur && system_ram_resources_mergeable(res, cur)) { in merge_system_ram_resource()
1445 res->end = cur->end; in merge_system_ram_resource()
1446 res->sibling = cur->sibling; in merge_system_ram_resource()
1447 free_resource(cur); in merge_system_ram_resource()
1451 cur = res->parent->child; in merge_system_ram_resource()
1452 while (cur && cur->sibling != res) in merge_system_ram_resource()
1453 cur = cur->sibling; in merge_system_ram_resource()
1454 if (cur && system_ram_resources_mergeable(cur, res)) { in merge_system_ram_resource()
[all …]
Dpadata.c124 struct padata_work *cur, *next; in padata_works_free() local
130 list_for_each_entry_safe(cur, next, works, pw_list) { in padata_works_free()
131 list_del(&cur->pw_list); in padata_works_free()
132 padata_work_free(cur); in padata_works_free()
392 struct padata_priv *cur; in padata_do_serial() local
398 cur = list_entry(pos, struct padata_priv, list); in padata_do_serial()
399 if (cur->seq_nr < padata->seq_nr) in padata_do_serial()
Dkprobes.c1953 struct llist_node **cur) in __kretprobe_find_ret_addr() argument
1956 struct llist_node *node = *cur; in __kretprobe_find_ret_addr()
1966 *cur = node; in __kretprobe_find_ret_addr()
1991 struct llist_node **cur) in kretprobe_find_ret_addr() argument
1996 if (WARN_ON_ONCE(!cur)) in kretprobe_find_ret_addr()
2000 ret = __kretprobe_find_ret_addr(tsk, cur); in kretprobe_find_ret_addr()
2003 ri = container_of(*cur, struct kretprobe_instance, llist); in kretprobe_find_ret_addr()
Dcpu.c2502 ssize_t cur, res = 0; in states_show() local
2510 cur = sprintf(buf, "%3d: %s\n", i, sp->name); in states_show()
2511 buf += cur; in states_show()
2512 res += cur; in states_show()
/kernel/locking/
Dww_mutex.h357 struct MUTEX_WAITER *cur; in __ww_mutex_check_waiters() local
361 for (cur = __ww_waiter_first(lock); cur; in __ww_mutex_check_waiters()
362 cur = __ww_waiter_next(lock, cur)) { in __ww_mutex_check_waiters()
364 if (!cur->ww_ctx) in __ww_mutex_check_waiters()
367 if (__ww_mutex_die(lock, cur, ww_ctx) || in __ww_mutex_check_waiters()
368 __ww_mutex_wound(lock, cur->ww_ctx, ww_ctx)) in __ww_mutex_check_waiters()
446 struct MUTEX_WAITER *cur; in __ww_mutex_check_kill() local
465 for (cur = __ww_waiter_prev(lock, waiter); cur; in __ww_mutex_check_kill()
466 cur = __ww_waiter_prev(lock, cur)) { in __ww_mutex_check_kill()
468 if (!cur->ww_ctx) in __ww_mutex_check_kill()
[all …]
Dlocktorture.c741 long cur; in __torture_print_stats() local
751 cur = data_race(statp[i].n_lock_acquired); in __torture_print_stats()
752 sum += cur; in __torture_print_stats()
753 if (max < cur) in __torture_print_stats()
754 max = cur; in __torture_print_stats()
755 if (min > cur) in __torture_print_stats()
756 min = cur; in __torture_print_stats()
/kernel/power/
Dsnapshot.c416 struct bm_position cur; /* most recently used bit position */ member
588 bm->cur.zone = list_entry(bm->zones.next, struct mem_zone_bm_rtree, in memory_bm_position_reset()
590 bm->cur.node = list_entry(bm->cur.zone->leaves.next, in memory_bm_position_reset()
592 bm->cur.node_pfn = 0; in memory_bm_position_reset()
593 bm->cur.cur_pfn = BM_END_OF_MAP; in memory_bm_position_reset()
594 bm->cur.node_bit = 0; in memory_bm_position_reset()
634 struct mem_extent *ext, *cur, *aux; in create_mem_extents() local
665 cur = ext; in create_mem_extents()
666 list_for_each_entry_safe_continue(cur, aux, list, hook) { in create_mem_extents()
667 if (zone_end < cur->start) in create_mem_extents()
[all …]
Dswap.c98 struct swap_map_page *cur; member
416 if (handle->cur) in release_swap_writer()
417 free_page((unsigned long)handle->cur); in release_swap_writer()
418 handle->cur = NULL; in release_swap_writer()
431 handle->cur = (struct swap_map_page *)get_zeroed_page(GFP_KERNEL); in get_swap_writer()
432 if (!handle->cur) { in get_swap_writer()
459 if (!handle->cur) in swap_write_page()
465 handle->cur->entries[handle->k++] = offset; in swap_write_page()
470 handle->cur->next_swap = offset; in swap_write_page()
473 error = write_page(handle->cur, handle->cur_swap, hb); in swap_write_page()
[all …]
Dpower.h135 unsigned int cur; /* number of the block of PAGE_SIZE bytes the member
/kernel/trace/
Drethook.c216 struct llist_node **cur) in __rethook_find_ret_addr() argument
219 struct llist_node *node = *cur; in __rethook_find_ret_addr()
229 *cur = node; in __rethook_find_ret_addr()
255 struct llist_node **cur) in rethook_find_ret_addr() argument
260 if (WARN_ON_ONCE(!cur)) in rethook_find_ret_addr()
267 ret = __rethook_find_ret_addr(tsk, cur); in rethook_find_ret_addr()
270 rhn = container_of(*cur, struct rethook_node, llist); in rethook_find_ret_addr()
/kernel/kcsan/
Dkcsan_test.c164 char *cur; in __report_matches() local
178 cur = expect[0]; in __report_matches()
180 cur += scnprintf(cur, end - cur, "BUG: KCSAN: %s in ", in __report_matches()
190 cur += scnprintf(cur, end - cur, "%ps / %ps", in __report_matches()
194 scnprintf(cur, end - cur, "%pS", r->access[0].fn); in __report_matches()
196 cur = strchr(expect[0], '+'); in __report_matches()
197 if (cur) in __report_matches()
198 *cur = '\0'; in __report_matches()
202 cur = expect[1]; in __report_matches()
205 cur += scnprintf(cur, end - cur, "race at unknown origin, with "); in __report_matches()
[all …]
Dreport.c280 char *cur; in get_stack_skipnr() local
291 cur = strnstr(buf, "kcsan_", len); in get_stack_skipnr()
292 if (cur) { in get_stack_skipnr()
293 cur += strlen("kcsan_"); in get_stack_skipnr()
294 if (!str_has_prefix(cur, "test")) in get_stack_skipnr()
/kernel/bpf/
Dbpf_lru_list.c175 struct list_head *cur, *last, *next = inactive; in __bpf_lru_list_rotate_inactive() local
186 cur = l->next_inactive_rotation; in __bpf_lru_list_rotate_inactive()
188 if (cur == inactive) { in __bpf_lru_list_rotate_inactive()
189 cur = cur->prev; in __bpf_lru_list_rotate_inactive()
193 node = list_entry(cur, struct bpf_lru_node, list); in __bpf_lru_list_rotate_inactive()
194 next = cur->prev; in __bpf_lru_list_rotate_inactive()
197 if (cur == last) in __bpf_lru_list_rotate_inactive()
199 cur = next; in __bpf_lru_list_rotate_inactive()
Dverifier.c644 struct bpf_verifier_state *cur = env->cur_state; in func() local
646 return cur->frame[reg->frameno]; in func()
1262 struct bpf_verifier_state *cur = env->cur_state; in pop_stack() local
1269 if (cur) { in pop_stack()
1270 err = copy_verifier_state(cur, &head->st); in pop_stack()
1292 struct bpf_verifier_state *cur = env->cur_state; in push_stack() local
1306 err = copy_verifier_state(&elem->st, cur); in push_stack()
2517 struct bpf_verifier_state *cur) in push_jmp_history() argument
2519 u32 cnt = cur->jmp_history_cnt; in push_jmp_history()
2525 p = krealloc(cur->jmp_history, alloc_size, GFP_USER); in push_jmp_history()
[all …]
Dbtf.c4622 void *cur, *end; in btf_check_all_metas() local
4625 cur = btf->nohdr_data + hdr->type_off; in btf_check_all_metas()
4626 end = cur + hdr->type_len; in btf_check_all_metas()
4629 while (cur < end) { in btf_check_all_metas()
4630 struct btf_type *t = cur; in btf_check_all_metas()
4633 meta_size = btf_check_meta(env, t, end - cur); in btf_check_all_metas()
4638 cur += meta_size; in btf_check_all_metas()
/kernel/time/
Dtick-broadcast.c165 struct clock_event_device *cur = tick_broadcast_device.evtdev; in tick_install_broadcast_device() local
170 if (!tick_check_broadcast_device(cur, dev)) in tick_install_broadcast_device()
176 clockevents_exchange_device(cur, dev); in tick_install_broadcast_device()
177 if (cur) in tick_install_broadcast_device()
178 cur->event_handler = clockevents_handle_noop; in tick_install_broadcast_device()
/kernel/cgroup/
Dcpuset.c720 static int validate_change_legacy(struct cpuset *cur, struct cpuset *trial) in validate_change_legacy() argument
730 cpuset_for_each_child(c, css, cur) in validate_change_legacy()
736 par = parent_cs(cur); in validate_change_legacy()
765 static int validate_change(struct cpuset *cur, struct cpuset *trial) in validate_change() argument
774 ret = validate_change_legacy(cur, trial); in validate_change()
779 if (cur == &top_cpuset) in validate_change()
782 par = parent_cs(cur); in validate_change()
789 if ((cgroup_is_populated(cur->css.cgroup) || cur->attach_in_progress)) { in validate_change()
790 if (!cpumask_empty(cur->cpus_allowed) && in validate_change()
793 if (!nodes_empty(cur->mems_allowed) && in validate_change()
[all …]
/kernel/sched/
Dfair.c1922 struct task_struct *cur; in task_numa_compare() local
1933 cur = rcu_dereference(dst_rq->curr); in task_numa_compare()
1934 if (cur && ((cur->flags & PF_EXITING) || is_idle_task(cur))) in task_numa_compare()
1935 cur = NULL; in task_numa_compare()
1941 if (cur == env->p) { in task_numa_compare()
1946 if (!cur) { in task_numa_compare()
1954 if (!cpumask_test_cpu(env->src_cpu, cur->cpus_ptr)) in task_numa_compare()
1963 cur->numa_preferred_nid != env->src_nid) { in task_numa_compare()
1977 cur_ng = rcu_dereference(cur->numa_group); in task_numa_compare()
1988 imp = taskimp + task_weight(cur, env->src_nid, dist) - in task_numa_compare()
[all …]
Dtopology.c2487 static int dattrs_equal(struct sched_domain_attr *cur, int idx_cur, in dattrs_equal() argument
2493 if (!new && !cur) in dattrs_equal()
2498 return !memcmp(cur ? (cur + idx_cur) : &tmp, in dattrs_equal()
Dcpufreq_schedutil.c147 policy->cpuinfo.max_freq : policy->cur; in get_next_freq()
Ddeadline.c3023 int dl_cpuset_cpumask_can_shrink(const struct cpumask *cur, in dl_cpuset_cpumask_can_shrink() argument
3031 cur_dl_b = dl_bw_of(cpumask_any(cur)); in dl_cpuset_cpumask_can_shrink()
Dcore.c3477 int migrate_swap(struct task_struct *cur, struct task_struct *p, in migrate_swap() argument
3484 .src_task = cur, in migrate_swap()
3506 trace_sched_swap_numa(cur, arg.src_cpu, p, arg.dst_cpu); in migrate_swap()
9276 int cpuset_cpumask_can_shrink(const struct cpumask *cur, in cpuset_cpumask_can_shrink() argument
9281 if (cpumask_empty(cur)) in cpuset_cpumask_can_shrink()
9284 ret = dl_cpuset_cpumask_can_shrink(cur, trial); in cpuset_cpumask_can_shrink()

12