/kernel/ |
D | notifier.c | 65 unsigned long val, void *v, in notifier_call_chain() argument 83 ret = nb->notifier_call(nb, val, v); in notifier_call_chain() 114 void *v) in notifier_call_chain_robust() argument 118 ret = notifier_call_chain(nl, val_up, v, -1, &nr); in notifier_call_chain_robust() 120 notifier_call_chain(nl, val_down, v, nr-1, NULL); in notifier_call_chain_robust() 176 unsigned long val_up, unsigned long val_down, void *v) in atomic_notifier_call_chain_robust() argument 186 ret = notifier_call_chain_robust(&nh->head, val_up, val_down, v); in atomic_notifier_call_chain_robust() 212 unsigned long val, void *v) in atomic_notifier_call_chain() argument 217 ret = notifier_call_chain(&nh->head, val, v, -1, NULL); in atomic_notifier_call_chain() 291 unsigned long val_up, unsigned long val_down, void *v) in blocking_notifier_call_chain_robust() argument [all …]
|
D | compat.c | 253 compat_sigset_t v; in get_compat_sigset() local 254 if (copy_from_user(&v, compat, sizeof(compat_sigset_t))) in get_compat_sigset() 257 case 4: set->sig[3] = v.sig[6] | (((long)v.sig[7]) << 32 ); in get_compat_sigset() 259 case 3: set->sig[2] = v.sig[4] | (((long)v.sig[5]) << 32 ); in get_compat_sigset() 261 case 2: set->sig[1] = v.sig[2] | (((long)v.sig[3]) << 32 ); in get_compat_sigset() 263 case 1: set->sig[0] = v.sig[0] | (((long)v.sig[1]) << 32 ); in get_compat_sigset()
|
D | fail_function.c | 189 static void fei_seq_stop(struct seq_file *m, void *v) in fei_seq_stop() argument 194 static void *fei_seq_next(struct seq_file *m, void *v, loff_t *pos) in fei_seq_next() argument 196 return seq_list_next(v, &fei_attr_list, pos); in fei_seq_next() 199 static int fei_seq_show(struct seq_file *m, void *v) in fei_seq_show() argument 201 struct fei_attr *attr = list_entry(v, struct fei_attr, list); in fei_seq_show()
|
D | jump_label.c | 118 int v, v1; in static_key_slow_inc_cpuslocked() local 135 for (v = atomic_read(&key->enabled); v > 0; v = v1) { in static_key_slow_inc_cpuslocked() 136 v1 = atomic_cmpxchg(&key->enabled, v, v + 1); in static_key_slow_inc_cpuslocked() 137 if (likely(v1 == v)) in static_key_slow_inc_cpuslocked()
|
D | dma.c | 118 static int proc_dma_show(struct seq_file *m, void *v) in proc_dma_show() argument 131 static int proc_dma_show(struct seq_file *m, void *v) in proc_dma_show() argument
|
/kernel/bpf/ |
D | prog_iter.c | 27 static void *bpf_prog_seq_next(struct seq_file *seq, void *v, loff_t *pos) in bpf_prog_seq_next() argument 33 bpf_prog_put((struct bpf_prog *)v); in bpf_prog_seq_next() 44 static int __bpf_prog_seq_show(struct seq_file *seq, void *v, bool in_stop) in DEFINE_BPF_ITER_FUNC() 52 ctx.prog = v; in DEFINE_BPF_ITER_FUNC() 61 static int bpf_prog_seq_show(struct seq_file *seq, void *v) in bpf_prog_seq_show() argument 63 return __bpf_prog_seq_show(seq, v, false); in bpf_prog_seq_show() 66 static void bpf_prog_seq_stop(struct seq_file *seq, void *v) in bpf_prog_seq_stop() argument 68 if (!v) in bpf_prog_seq_stop() 69 (void)__bpf_prog_seq_show(seq, v, true); in bpf_prog_seq_stop() 71 bpf_prog_put((struct bpf_prog *)v); in bpf_prog_seq_stop()
|
D | tnum.c | 88 u64 alpha, beta, v; in tnum_and() local 92 v = a.value & b.value; in tnum_and() 93 return TNUM(v, alpha & beta & ~v); in tnum_and() 98 u64 v, mu; in tnum_or() local 100 v = a.value | b.value; in tnum_or() 102 return TNUM(v, mu & ~v); in tnum_or() 107 u64 v, mu; in tnum_xor() local 109 v = a.value ^ b.value; in tnum_xor() 111 return TNUM(v & ~mu, mu); in tnum_xor() 143 u64 v, mu; in tnum_intersect() local [all …]
|
D | task_iter.c | 67 static void *task_seq_next(struct seq_file *seq, void *v, loff_t *pos) in task_seq_next() argument 74 put_task_struct((struct task_struct *)v); in task_seq_next() 107 static int task_seq_show(struct seq_file *seq, void *v) in task_seq_show() argument 109 return __task_seq_show(seq, v, false); in task_seq_show() 112 static void task_seq_stop(struct seq_file *seq, void *v) in task_seq_stop() argument 114 if (!v) in task_seq_stop() 115 (void)__task_seq_show(seq, v, true); in task_seq_stop() 117 put_task_struct((struct task_struct *)v); in task_seq_stop() 225 static void *task_file_seq_next(struct seq_file *seq, void *v, loff_t *pos) in task_file_seq_next() argument 231 fput((struct file *)v); in task_file_seq_next() [all …]
|
D | map_iter.c | 27 static void *bpf_map_seq_next(struct seq_file *seq, void *v, loff_t *pos) in bpf_map_seq_next() argument 33 bpf_map_put((struct bpf_map *)v); in bpf_map_seq_next() 44 static int __bpf_map_seq_show(struct seq_file *seq, void *v, bool in_stop) in DEFINE_BPF_ITER_FUNC() 52 ctx.map = v; in DEFINE_BPF_ITER_FUNC() 61 static int bpf_map_seq_show(struct seq_file *seq, void *v) in bpf_map_seq_show() argument 63 return __bpf_map_seq_show(seq, v, false); in bpf_map_seq_show() 66 static void bpf_map_seq_stop(struct seq_file *seq, void *v) in bpf_map_seq_stop() argument 68 if (!v) in bpf_map_seq_stop() 69 (void)__bpf_map_seq_show(seq, v, true); in bpf_map_seq_stop() 71 bpf_map_put((struct bpf_map *)v); in bpf_map_seq_stop()
|
D | btf.c | 372 const struct resolve_vertex *v); 1567 struct resolve_vertex *v; in env_stack_push() local 1577 v = &env->stack[env->top_stack++]; in env_stack_push() 1578 v->t = t; in env_stack_push() 1579 v->type_id = type_id; in env_stack_push() 1580 v->next_member = 0; in env_stack_push() 1803 const struct resolve_vertex *v) in btf_df_resolve() argument 1805 btf_verifier_log_basic(env, v->t, "Unsupported resolve"); in btf_df_resolve() 2292 const struct resolve_vertex *v) in btf_modifier_resolve() argument 2294 const struct btf_type *t = v->t; in btf_modifier_resolve() [all …]
|
/kernel/time/ |
D | itimer.c | 103 struct __kernel_old_itimerval v; in put_itimerval() local 105 v.it_interval.tv_sec = i->it_interval.tv_sec; in put_itimerval() 106 v.it_interval.tv_usec = i->it_interval.tv_nsec / NSEC_PER_USEC; in put_itimerval() 107 v.it_value.tv_sec = i->it_value.tv_sec; in put_itimerval() 108 v.it_value.tv_usec = i->it_value.tv_nsec / NSEC_PER_USEC; in put_itimerval() 109 return copy_to_user(o, &v, sizeof(struct __kernel_old_itimerval)) ? -EFAULT : 0; in put_itimerval() 255 struct itimerspec64 v = {}; in clear_itimer() local 259 do_setitimer(i, &v, NULL); in clear_itimer() 315 struct __kernel_old_itimerval v; in get_itimerval() local 317 if (copy_from_user(&v, i, sizeof(struct __kernel_old_itimerval))) in get_itimerval() [all …]
|
D | timer_list.c | 312 static int timer_list_show(struct seq_file *m, void *v) in timer_list_show() argument 314 struct timer_list_iter *iter = v; in timer_list_show() 359 static void *timer_list_next(struct seq_file *file, void *v, loff_t *offset) in timer_list_next() argument 366 static void timer_list_stop(struct seq_file *seq, void *v) in timer_list_stop() argument
|
/kernel/trace/ |
D | trace_printk.c | 123 find_next_mod_format(int start_index, void *v, const char **fmt, loff_t *pos) in find_next_mod_format() argument 136 if (!v || start_index == *pos) { in find_next_mod_format() 153 mod_fmt = container_of(v, typeof(*mod_fmt), fmt); in find_next_mod_format() 180 find_next_mod_format(int start_index, void *v, const char **fmt, loff_t *pos) in find_next_mod_format() argument 254 static const char **find_next(void *v, loff_t *pos) in find_next() argument 256 const char **fmt = v; in find_next() 284 return find_next_mod_format(start_index, v, fmt, pos); in find_next() 294 static void *t_next(struct seq_file *m, void * v, loff_t *pos) in t_next() argument 297 return find_next(v, pos); in t_next() 300 static int t_show(struct seq_file *m, void *v) in t_show() argument [all …]
|
D | trace_branch.c | 297 static int annotate_branch_stat_show(struct seq_file *m, void *v) in annotate_branch_stat_show() argument 299 struct ftrace_likely_data *p = v; in annotate_branch_stat_show() 323 annotated_branch_stat_next(void *v, int idx) in annotated_branch_stat_next() argument 325 struct ftrace_likely_data *p = v; in annotated_branch_stat_next() 413 all_branch_stat_next(void *v, int idx) in all_branch_stat_next() argument 415 struct ftrace_branch_data *p = v; in all_branch_stat_next() 425 static int all_branch_stat_show(struct seq_file *m, void *v) in all_branch_stat_show() argument 427 struct ftrace_branch_data *p = v; in all_branch_stat_show()
|
D | trace_boot.c | 30 unsigned long v = 0; in trace_boot_set_instance_options() local 45 if (kstrtoul(p, 10, &v)) in trace_boot_set_instance_options() 47 if (v) in trace_boot_set_instance_options() 61 v = memparse(p, NULL); in trace_boot_set_instance_options() 62 if (v < PAGE_SIZE) in trace_boot_set_instance_options() 64 if (tracing_resize_ring_buffer(tr, v, RING_BUFFER_ALL_CPUS) < 0) in trace_boot_set_instance_options()
|
D | trace_dynevent.c | 109 void *dyn_event_seq_next(struct seq_file *m, void *v, loff_t *pos) in dyn_event_seq_next() argument 111 return seq_list_next(v, &dyn_event_list, pos); in dyn_event_seq_next() 114 void dyn_event_seq_stop(struct seq_file *m, void *v) in dyn_event_seq_stop() argument 119 static int dyn_event_seq_show(struct seq_file *m, void *v) in dyn_event_seq_show() argument 121 struct dyn_event *ev = v; in dyn_event_seq_show()
|
D | trace_stack.c | 389 t_next(struct seq_file *m, void *v, loff_t *pos) in t_next() argument 436 static int t_show(struct seq_file *m, void *v) in t_show() argument 441 if (v == SEQ_START_TOKEN) { in t_show() 453 i = *(long *)v; in t_show()
|
/kernel/locking/ |
D | lockdep_proc.c | 38 static void *l_next(struct seq_file *m, void *v, loff_t *pos) in l_next() argument 40 struct lock_class *class = v; in l_next() 56 static void l_stop(struct seq_file *m, void *v) in l_stop() argument 77 static int l_show(struct seq_file *m, void *v) in l_show() argument 79 struct lock_class *class = v; in l_show() 84 if (v == lock_classes) in l_show() 137 static void *lc_next(struct seq_file *m, void *v, loff_t *pos) in lc_next() argument 143 static void lc_stop(struct seq_file *m, void *v) in lc_stop() argument 147 static int lc_show(struct seq_file *m, void *v) in lc_show() argument 149 struct lock_chain *chain = v; in lc_show() [all …]
|
/kernel/rcu/ |
D | rcu_segcblist.c | 82 static void rcu_segcblist_set_len(struct rcu_segcblist *rsclp, long v) in rcu_segcblist_set_len() argument 85 atomic_long_set(&rsclp->len, v); in rcu_segcblist_set_len() 87 WRITE_ONCE(rsclp->len, v); in rcu_segcblist_set_len() 98 static void rcu_segcblist_add_len(struct rcu_segcblist *rsclp, long v) in rcu_segcblist_add_len() argument 102 atomic_long_add(v, &rsclp->len); in rcu_segcblist_add_len() 106 WRITE_ONCE(rsclp->len, rsclp->len + v); in rcu_segcblist_add_len() 128 static long rcu_segcblist_xchg_len(struct rcu_segcblist *rsclp, long v) in rcu_segcblist_xchg_len() argument 131 return atomic_long_xchg(&rsclp->len, v); in rcu_segcblist_xchg_len() 136 WRITE_ONCE(rsclp->len, v); in rcu_segcblist_xchg_len()
|
/kernel/irq/ |
D | proc.c | 83 static int irq_affinity_hint_proc_show(struct seq_file *m, void *v) in irq_affinity_hint_proc_show() argument 104 static int irq_affinity_proc_show(struct seq_file *m, void *v) in irq_affinity_proc_show() argument 109 static int irq_affinity_list_proc_show(struct seq_file *m, void *v) in irq_affinity_list_proc_show() argument 218 static int irq_effective_aff_proc_show(struct seq_file *m, void *v) in irq_effective_aff_proc_show() argument 223 static int irq_effective_aff_list_proc_show(struct seq_file *m, void *v) in irq_effective_aff_list_proc_show() argument 229 static int default_affinity_show(struct seq_file *m, void *v) in default_affinity_show() argument 279 static int irq_node_proc_show(struct seq_file *m, void *v) in irq_node_proc_show() argument 288 static int irq_spurious_proc_show(struct seq_file *m, void *v) in irq_spurious_proc_show() argument 460 int show_interrupts(struct seq_file *p, void *v) in show_interrupts() argument 465 int i = *(loff_t *) v, j; in show_interrupts()
|
D | affinity.c | 292 unsigned int ncpus, v; in __irq_build_affinity_masks() local 310 for (v = 0; v < nv->nvectors; v++, curvec++) { in __irq_build_affinity_masks()
|
/kernel/cgroup/ |
D | debug.c | 40 static int current_css_set_read(struct seq_file *seq, void *v) in current_css_set_read() argument 87 static int current_css_set_cg_links_read(struct seq_file *seq, void *v) in current_css_set_cg_links_read() argument 114 static int cgroup_css_links_read(struct seq_file *seq, void *v) in cgroup_css_links_read() argument 203 static int cgroup_subsys_states_read(struct seq_file *seq, void *v) in cgroup_subsys_states_read() argument 255 static int cgroup_masks_read(struct seq_file *seq, void *v) in cgroup_masks_read() argument
|
/kernel/sched/ |
D | stats.c | 15 static int show_schedstat(struct seq_file *seq, void *v) in show_schedstat() argument 19 if (v == (void *)1) { in show_schedstat() 28 cpu = (unsigned long)(v - 2); in show_schedstat()
|
/kernel/gcov/ |
D | gcc_4_7.c | 407 static size_t store_gcov_u32(void *buffer, size_t off, u32 v) in store_gcov_u32() argument 413 *data = v; in store_gcov_u32() 431 static size_t store_gcov_u64(void *buffer, size_t off, u64 v) in store_gcov_u64() argument 438 data[0] = (v & 0xffffffffUL); in store_gcov_u64() 439 data[1] = (v >> 32); in store_gcov_u64()
|
D | clang.c | 481 static size_t store_gcov_u32(void *buffer, size_t off, u32 v) in store_gcov_u32() argument 487 *data = v; in store_gcov_u32() 505 static size_t store_gcov_u64(void *buffer, size_t off, u64 v) in store_gcov_u64() argument 512 data[0] = (v & 0xffffffffUL); in store_gcov_u64() 513 data[1] = (v >> 32); in store_gcov_u64()
|