/kernel/ |
D | kthread.c | 60 static inline struct kthread *to_kthread(struct task_struct *k) in to_kthread() argument 62 return __to_kthread(k->vfork_done); in to_kthread() 65 static struct kthread *to_live_kthread(struct task_struct *k) in to_live_kthread() argument 67 struct completion *vfork = ACCESS_ONCE(k->vfork_done); in to_live_kthread() 68 if (likely(vfork) && try_get_task_stack(k)) in to_live_kthread() 408 static void __kthread_unpark(struct task_struct *k, struct kthread *kthread) in __kthread_unpark() argument 419 __kthread_bind(k, kthread->cpu, TASK_PARKED); in __kthread_unpark() 420 wake_up_state(k, TASK_PARKED); in __kthread_unpark() 432 void kthread_unpark(struct task_struct *k) in kthread_unpark() argument 434 struct kthread *kthread = to_live_kthread(k); in kthread_unpark() [all …]
|
D | range.c | 126 int i, j, k = az - 1, nr_range = az; in clean_sort_range() local 128 for (i = 0; i < k; i++) { in clean_sort_range() 131 for (j = k; j > i; j--) { in clean_sort_range() 133 k = j; in clean_sort_range() 139 range[i].start = range[k].start; in clean_sort_range() 140 range[i].end = range[k].end; in clean_sort_range() 141 range[k].start = 0; in clean_sort_range() 142 range[k].end = 0; in clean_sort_range() 143 k--; in clean_sort_range()
|
D | audit.h | 286 #define audit_to_watch(k, p, l, o) (-EINVAL) argument 287 #define audit_add_watch(k, l) (-EINVAL) argument 288 #define audit_remove_watch_rule(k) BUG() argument 292 #define audit_alloc_mark(k, p, l) (ERR_PTR(-EINVAL)) argument 295 #define audit_remove_mark_rule(k) argument
|
D | kexec_file.c | 952 int i, k; in kexec_purgatory_find_symbol() local 972 for (k = 0; k < sechdrs[i].sh_size/sizeof(Elf_Sym); k++) { in kexec_purgatory_find_symbol() 973 if (ELF_ST_BIND(syms[k].st_info) != STB_GLOBAL) in kexec_purgatory_find_symbol() 976 if (strcmp(strtab + syms[k].st_name, name) != 0) in kexec_purgatory_find_symbol() 979 if (syms[k].st_shndx == SHN_UNDEF || in kexec_purgatory_find_symbol() 980 syms[k].st_shndx >= ehdr->e_shnum) { in kexec_purgatory_find_symbol() 982 name, syms[k].st_shndx); in kexec_purgatory_find_symbol() 987 return &syms[k]; in kexec_purgatory_find_symbol()
|
D | seccomp.c | 108 u32 k = ftest->k; in seccomp_check_filter() local 114 if (k >= sizeof(struct seccomp_data) || k & 3) in seccomp_check_filter() 119 ftest->k = sizeof(struct seccomp_data); in seccomp_check_filter() 123 ftest->k = sizeof(struct seccomp_data); in seccomp_check_filter()
|
D | cpuset.c | 638 int i, j, k; /* indices for partition finding loops */ in generate_sched_domains() local 717 for (k = 0; k < csn; k++) { in generate_sched_domains() 718 struct cpuset *c = csa[k]; in generate_sched_domains()
|
D | compat.c | 605 unsigned long *k; in compat_get_user_cpu_mask() local 612 k = cpumask_bits(new_mask); in compat_get_user_cpu_mask() 613 return compat_get_bitmap(k, user_mask_ptr, len * 8); in compat_get_user_cpu_mask()
|
D | signal.c | 3133 struct k_sigaction *k; in do_sigaction() local 3139 k = &p->sighand->action[sig-1]; in do_sigaction() 3143 *oact = *k; in do_sigaction() 3148 *k = *act; in do_sigaction()
|
/kernel/power/ |
D | swap.c | 100 unsigned int k; member 423 handle->k = 0; in get_swap_writer() 446 handle->cur->entries[handle->k++] = offset; in swap_write_page() 447 if (handle->k >= MAP_PAGE_ENTRIES) { in swap_write_page() 457 handle->k = 0; in swap_write_page() 1002 handle->k = 0; in get_swap_reader() 1016 offset = handle->cur->entries[handle->k]; in swap_read_page() 1022 if (++handle->k >= MAP_PAGE_ENTRIES) { in swap_read_page() 1023 handle->k = 0; in swap_read_page() 1304 handle->cur->entries[handle->k]) { in load_image_lzo()
|
D | hibernate.c | 242 unsigned int k; in swsusp_show_speed() local 250 k = nr_pages * (PAGE_SIZE / 1024); in swsusp_show_speed() 251 kps = (k * 100) / centisecs; in swsusp_show_speed() 253 msg, k, in swsusp_show_speed()
|
/kernel/bpf/ |
D | core.c | 58 void *bpf_internal_load_pointer_neg_helper(const struct sk_buff *skb, int k, unsigned int size) in bpf_internal_load_pointer_neg_helper() argument 62 if (k >= SKF_NET_OFF) in bpf_internal_load_pointer_neg_helper() 63 ptr = skb_network_header(skb) + k - SKF_NET_OFF; in bpf_internal_load_pointer_neg_helper() 64 else if (k >= SKF_LL_OFF) in bpf_internal_load_pointer_neg_helper() 65 ptr = skb_mac_header(skb) + k - SKF_LL_OFF; in bpf_internal_load_pointer_neg_helper()
|
/kernel/printk/ |
D | printk.c | 1022 unsigned long long k; in boot_delay_msec() local 1030 k = (unsigned long long)loops_per_msec * boot_delay; in boot_delay_msec() 1033 while (k) { in boot_delay_msec() 1034 k--; in boot_delay_msec()
|
/kernel/sched/ |
D | core.c | 7042 int i, j, k; in sched_init_numa() local 7058 for (k = 0; k < nr_node_ids; k++) { in sched_init_numa() 7059 int distance = node_distance(i, k); in sched_init_numa() 7071 if (sched_debug() && node_distance(k, i) != distance) in sched_init_numa() 7134 for_each_node(k) { in sched_init_numa() 7135 if (node_distance(j, k) > sched_domains_numa_distance[i]) in sched_init_numa() 7138 cpumask_or(mask, mask, cpumask_of_node(k)); in sched_init_numa()
|