/kernel/trace/ |
D | trace_export.c | 31 #define __field_struct(type, item) argument 34 #define __field(type, item) type item; argument 37 #define __field_desc(type, container, item) type item; argument 40 #define __array(type, item, size) type item[size]; argument 43 #define __array_desc(type, container, item, size) type item[size]; argument 46 #define __dynamic_array(type, item) type item[]; argument 75 #define __field(type, item) \ argument 76 ret = trace_define_field(event_call, #type, #item, \ 79 is_signed_type(type), filter_type); \ 84 #define __field_desc(type, container, item) \ argument [all …]
|
D | trace_probe.c | 39 #define DEFINE_BASIC_PRINT_TYPE_FUNC(type, fmt) \ argument 40 int PRINT_TYPE_FUNC_NAME(type)(struct trace_seq *s, const char *name, \ 43 trace_seq_printf(s, " %s=" fmt, name, *(type *)data); \ 46 const char PRINT_TYPE_FMT_NAME(type)[] = fmt; \ 47 NOKPROBE_SYMBOL(PRINT_TYPE_FUNC_NAME(type)); 85 #define DEFINE_FETCH_reg(type) \ argument 86 void FETCH_FUNC_NAME(reg, type)(struct pt_regs *regs, void *offset, void *dest) \ 88 *(type *)dest = (type)regs_get_register(regs, \ 91 NOKPROBE_SYMBOL(FETCH_FUNC_NAME(reg, type)); 97 #define DEFINE_FETCH_retval(type) \ argument [all …]
|
D | trace_probe.h | 54 #define DEFINE_FIELD(type, item, name, is_signed) \ argument 56 ret = trace_define_field(event_call, #type, name, \ 134 #define PRINT_TYPE_FUNC_NAME(type) print_type_##type argument 135 #define PRINT_TYPE_FMT_NAME(type) print_type_format_##type argument 138 #define DECLARE_BASIC_PRINT_TYPE_FUNC(type) \ argument 139 int PRINT_TYPE_FUNC_NAME(type)(struct trace_seq *s, const char *name, \ 141 extern const char PRINT_TYPE_FMT_NAME(type)[] 153 #define FETCH_FUNC_NAME(method, type) fetch_##method##_##type argument 156 #define DECLARE_FETCH_FUNC(method, type) \ argument 157 extern void FETCH_FUNC_NAME(method, type)(struct pt_regs *regs, \ [all …]
|
D | trace_syscalls.c | 17 enum trace_reg type, void *data); 19 enum trace_reg type, void *data); 127 if (entry->enter_event->event.type != ent->type) { in print_syscall_enter() 175 if (entry->exit_event->event.type != ent->type) { in print_syscall_exit() 189 #define SYSCALL_FIELD(type, name) \ argument 190 sizeof(type) != sizeof(trace.name) ? \ 192 #type, #name, offsetof(typeof(trace), name), \ 193 sizeof(trace.name), is_signed_type(type) 330 sys_data->enter_event->event.type, size, irq_flags, pc); in ftrace_syscall_enter() 375 sys_data->exit_event->event.type, sizeof(*entry), in ftrace_syscall_exit() [all …]
|
D | trace_output.c | 236 if (entry->type != event->event.type) { in trace_raw_output_prep() 611 struct trace_event *ftrace_find_event(int type) in ftrace_find_event() argument 616 key = type & (EVENT_HASHSIZE - 1); in ftrace_find_event() 619 if (event->type == type) in ftrace_find_event() 643 if (e->type != last + 1) in trace_search_list() 696 if (!event->type) { in register_trace_event() 701 event->type = trace_search_list(&list); in register_trace_event() 702 if (!event->type) in register_trace_event() 707 event->type = next_event_type++; in register_trace_event() 711 if (WARN_ON(ftrace_find_event(event->type))) in register_trace_event() [all …]
|
D | trace_kprobe.c | 134 #define DEFINE_FETCH_stack(type) \ argument 135 static void FETCH_FUNC_NAME(stack, type)(struct pt_regs *regs, \ 138 *(type *)dest = (type)regs_get_kernel_stack_nth(regs, \ 141 NOKPROBE_SYMBOL(FETCH_FUNC_NAME(stack, type)); 148 #define DEFINE_FETCH_memory(type) \ argument 149 static void FETCH_FUNC_NAME(memory, type)(struct pt_regs *regs, \ 152 type retval; \ 154 *(type *)dest = 0; \ 156 *(type *)dest = retval; \ 158 NOKPROBE_SYMBOL(FETCH_FUNC_NAME(memory, type)); [all …]
|
D | trace.h | 47 #define __field(type, item) type item; argument 50 #define __field_struct(type, item) __field(type, item) argument 53 #define __field_desc(type, container, item) argument 56 #define __array(type, item, size) type item[size]; argument 59 #define __array_desc(type, container, item, size) argument 62 #define __dynamic_array(type, item) type item[]; argument 283 #define FTRACE_CMP_TYPE(var, type) \ argument 284 __builtin_types_compatible_p(typeof(var), type *) 290 WARN_ON(id && (entry)->type != id); \ 573 int type, [all …]
|
D | trace_uprobe.c | 110 #define DEFINE_FETCH_stack(type) \ argument 111 static void FETCH_FUNC_NAME(stack, type)(struct pt_regs *regs, \ 114 *(type *)dest = (type)get_user_stack_nth(regs, \ 122 #define DEFINE_FETCH_memory(type) \ argument 123 static void FETCH_FUNC_NAME(memory, type)(struct pt_regs *regs, \ 126 type retval; \ 129 if (copy_from_user(&retval, vaddr, sizeof(type))) \ 130 *(type *)dest = 0; \ 132 *(type *) dest = retval; \ 195 #define DEFINE_FETCH_file_offset(type) \ argument [all …]
|
D | bpf_trace.c | 209 if (unlikely(event->attr.type != PERF_TYPE_HARDWARE && in bpf_perf_event_read() 210 event->attr.type != PERF_TYPE_RAW)) in bpf_perf_event_read() 249 if (unlikely(event->attr.type != PERF_TYPE_SOFTWARE || in bpf_perf_event_output() 308 static bool kprobe_prog_is_valid_access(int off, int size, enum bpf_access_type type) in kprobe_prog_is_valid_access() argument 315 if (type != BPF_READ) in kprobe_prog_is_valid_access() 332 .type = BPF_PROG_TYPE_KPROBE,
|
D | trace_event_perf.c | 215 if (tp_event->event.type == event_id && in perf_trace_init() 263 void *perf_trace_buf_prepare(int size, unsigned short type, in perf_trace_buf_prepare() argument 293 entry->type = type; in perf_trace_buf_prepare() 361 enum trace_reg type, void *data) in perf_ftrace_event_register() argument 363 switch (type) { in perf_ftrace_event_register()
|
/kernel/bpf/ |
D | inode.c | 30 static void *bpf_any_get(void *raw, enum bpf_type type) in bpf_any_get() argument 32 switch (type) { in bpf_any_get() 47 static void bpf_any_put(void *raw, enum bpf_type type) in bpf_any_put() argument 49 switch (type) { in bpf_any_put() 62 static void *bpf_fd_probe_obj(u32 ufd, enum bpf_type *type) in bpf_fd_probe_obj() argument 66 *type = BPF_TYPE_MAP; in bpf_fd_probe_obj() 69 *type = BPF_TYPE_PROG; in bpf_fd_probe_obj() 109 static int bpf_inode_type(const struct inode *inode, enum bpf_type *type) in bpf_inode_type() argument 111 *type = BPF_TYPE_UNSPEC; in bpf_inode_type() 113 *type = BPF_TYPE_PROG; in bpf_inode_type() [all …]
|
D | verifier.c | 142 enum bpf_reg_type type; member 258 t = env->cur_state.regs[i].type; in print_verifier_state() 273 reg_type_str[env->cur_state.spilled_regs[i / BPF_REG_SIZE].type]); in print_verifier_state() 489 regs[i].type = NOT_INIT; in init_reg_state() 495 regs[BPF_REG_FP].type = FRAME_PTR; in init_reg_state() 498 regs[BPF_REG_1].type = PTR_TO_CTX; in init_reg_state() 504 regs[regno].type = UNKNOWN_VALUE; in mark_reg_unknown_value() 525 if (regs[regno].type == NOT_INIT) { in check_reg_arg() 555 static bool is_spillable_regtype(enum bpf_reg_type type) in is_spillable_regtype() argument 557 switch (type) { in is_spillable_regtype() [all …]
|
D | syscall.c | 31 if (tl->type == attr->map_type) { in find_and_alloc_map() 436 static int find_prog_type(enum bpf_prog_type type, struct bpf_prog *prog) in find_prog_type() argument 441 if (tl->type == type) { in find_prog_type() 443 prog->type = type; in find_prog_type() 571 enum bpf_prog_type type = attr->prog_type; in bpf_prog_load() local 592 if (type == BPF_PROG_TYPE_KPROBE && in bpf_prog_load() 596 if (type != BPF_PROG_TYPE_SOCKET_FILTER && !capable(CAP_SYS_ADMIN)) in bpf_prog_load() 622 err = find_prog_type(type, prog); in bpf_prog_load()
|
/kernel/events/ |
D | hw_breakpoint.c | 64 static struct bp_cpuinfo *get_bp_info(int cpu, enum bp_type_idx type) in get_bp_info() argument 66 return per_cpu_ptr(bp_cpuinfo + type, cpu); in get_bp_info() 100 static unsigned int max_task_bp_pinned(int cpu, enum bp_type_idx type) in max_task_bp_pinned() argument 102 unsigned int *tsk_pinned = get_bp_info(cpu, type)->tsk_pinned; in max_task_bp_pinned() 105 for (i = nr_slots[type] - 1; i >= 0; i--) { in max_task_bp_pinned() 117 static int task_bp_pinned(int cpu, struct perf_event *bp, enum bp_type_idx type) in task_bp_pinned() argument 125 find_slot_idx(iter) == type && in task_bp_pinned() 146 enum bp_type_idx type) in fetch_bp_busy_slots() argument 152 struct bp_cpuinfo *info = get_bp_info(cpu, type); in fetch_bp_busy_slots() 157 nr += max_task_bp_pinned(cpu, type); in fetch_bp_busy_slots() [all …]
|
/kernel/ |
D | pid.c | 298 enum pid_type type; in alloc_pid() local 331 for (type = 0; type < PIDTYPE_MAX; ++type) in alloc_pid() 332 INIT_HLIST_HEAD(&pid->tasks[type]); in alloc_pid() 389 void attach_pid(struct task_struct *task, enum pid_type type) in attach_pid() argument 391 struct pid_link *link = &task->pids[type]; in attach_pid() 392 hlist_add_head_rcu(&link->node, &link->pid->tasks[type]); in attach_pid() 395 static void __change_pid(struct task_struct *task, enum pid_type type, in __change_pid() argument 402 link = &task->pids[type]; in __change_pid() 415 void detach_pid(struct task_struct *task, enum pid_type type) in detach_pid() argument 417 __change_pid(task, type, NULL); in detach_pid() [all …]
|
D | kcmp.c | 34 static long kptr_obfuscate(long v, int type) in kptr_obfuscate() argument 36 return (v ^ cookies[type][0]) * cookies[type][1]; in kptr_obfuscate() 45 static int kcmp_ptr(void *v1, void *v2, enum kcmp_type type) in kcmp_ptr() argument 49 t1 = kptr_obfuscate((long)v1, type); in kcmp_ptr() 50 t2 = kptr_obfuscate((long)v2, type); in kcmp_ptr() 97 SYSCALL_DEFINE5(kcmp, pid_t, pid1, pid_t, pid2, int, type, in SYSCALL_DEFINE5() argument 131 switch (type) { in SYSCALL_DEFINE5()
|
D | auditsc.c | 101 int type; member 459 switch (f->type) { in audit_filter_rules() 629 result = security_audit_rule_match(sid, f->type, in audit_filter_rules() 646 name->osid, f->type, f->op, in audit_filter_rules() 650 if (security_audit_rule_match(n->osid, f->type, in audit_filter_rules() 659 if (!ctx || ctx->type != AUDIT_IPC) in audit_filter_rules() 662 f->type, f->op, in audit_filter_rules() 672 result = audit_comparator(ctx->argv[f->type-AUDIT_ARG0], f->op, f->val); in audit_filter_rules() 1174 ab = audit_log_start(context, GFP_KERNEL, context->type); in show_special() 1178 switch (context->type) { in show_special() [all …]
|
D | auditfilter.c | 76 switch (f->type) { in audit_free_lsm_field() 335 switch(f->type) { in audit_field_valid() 343 switch(f->type) { in audit_field_valid() 445 f->type = data->fields[i]; in audit_data_to_entry() 449 if ((f->type == AUDIT_LOGINUID) && (f_val == AUDIT_UID_UNSET)) { in audit_data_to_entry() 450 f->type = AUDIT_LOGINUID_SET; in audit_data_to_entry() 460 switch (f->type) { in audit_data_to_entry() 501 err = security_audit_rule_init(f->type, f->op, str, in audit_data_to_entry() 622 data->fields[i] = f->type; in audit_krule_to_data() 624 switch(f->type) { in audit_krule_to_data() [all …]
|
D | profile.c | 167 int profile_event_register(enum profile_type type, struct notifier_block *n) in profile_event_register() argument 171 switch (type) { in profile_event_register() 186 int profile_event_unregister(enum profile_type type, struct notifier_block *n) in profile_event_unregister() argument 190 switch (type) { in profile_event_unregister() 282 static void do_profile_hits(int type, void *__pc, unsigned int nr_hits) in do_profile_hits() argument 394 static void do_profile_hits(int type, void *__pc, unsigned int nr_hits) in do_profile_hits() argument 402 void profile_hits(int type, void *__pc, unsigned int nr_hits) in profile_hits() argument 404 if (prof_on != type || !prof_buffer) in profile_hits() 406 do_profile_hits(type, __pc, nr_hits); in profile_hits() 410 void profile_tick(int type) in profile_tick() argument [all …]
|
D | kallsyms.c | 452 char type; member 461 &iter->type, iter->name, iter->module_name, in get_ksymbol_mod() 475 iter->type = kallsyms_get_symbol_type(off); in get_ksymbol_core() 537 char type; in s_show() local 543 type = iter->exported ? toupper(iter->type) : in s_show() 544 tolower(iter->type); in s_show() 546 type, iter->name, iter->module_name); in s_show() 549 iter->type, iter->name); in s_show()
|
D | kcov.c | 114 static void write_comp_data(u64 type, u64 arg1, u64 arg2, u64 ip) in write_comp_data() argument 139 area[start_index] = type; in write_comp_data() 204 u64 type = KCOV_CMP_CONST; in __sanitizer_cov_trace_switch() local 208 type |= KCOV_CMP_SIZE(0); in __sanitizer_cov_trace_switch() 211 type |= KCOV_CMP_SIZE(1); in __sanitizer_cov_trace_switch() 214 type |= KCOV_CMP_SIZE(2); in __sanitizer_cov_trace_switch() 217 type |= KCOV_CMP_SIZE(3); in __sanitizer_cov_trace_switch() 223 write_comp_data(type, cases[i + 2], val, _RET_IP_); in __sanitizer_cov_trace_switch()
|
/kernel/time/ |
D | alarmtimer.c | 178 struct alarm_base *base = &alarm_bases[alarm->type]; in alarmtimer_fired() 204 struct alarm_base *base = &alarm_bases[alarm->type]; in alarm_expires_remaining() 281 static void alarmtimer_freezerset(ktime_t absexp, enum alarmtimer_type type) in alarmtimer_freezerset() argument 285 struct alarm_base *base = &alarm_bases[type]; in alarmtimer_freezerset() 302 void alarm_init(struct alarm *alarm, enum alarmtimer_type type, in alarm_init() argument 306 hrtimer_init(&alarm->timer, alarm_bases[type].base_clockid, in alarm_init() 310 alarm->type = type; in alarm_init() 322 struct alarm_base *base = &alarm_bases[alarm->type]; in alarm_start() 340 struct alarm_base *base = &alarm_bases[alarm->type]; in alarm_start_relative() 349 struct alarm_base *base = &alarm_bases[alarm->type]; in alarm_restart() [all …]
|
/kernel/gcov/ |
D | gcc_3_4.c | 149 static int counter_active(struct gcov_info *info, unsigned int type) in counter_active() argument 151 return (1 << type) & info->ctr_mask; in counter_active() 339 unsigned int type; member 353 return &iter->type_info[iter->type]; in get_type() 403 iter->type = 0; in gcov_iter_start() 449 if (iter->count < get_func(iter)->n_ctrs[iter->type]) { in gcov_iter_next() 456 iter->type++; in gcov_iter_next() 459 if (iter->type < iter->num_types) { in gcov_iter_next() 464 iter->type = 0; in gcov_iter_next() 553 get_func(iter)->n_ctrs[iter->type] * 2); in gcov_iter_write() [all …]
|
/kernel/power/ |
D | qos.c | 72 .type = PM_QOS_MIN, 86 .type = PM_QOS_MIN, 101 .type = PM_QOS_MAX, 116 .type = PM_QOS_SUM, 157 switch (c->type) { in pm_qos_get_value() 193 char *type; in pm_qos_dbg_show_requests() local 215 switch (c->type) { in pm_qos_dbg_show_requests() 217 type = "Minimum"; in pm_qos_dbg_show_requests() 220 type = "Maximum"; in pm_qos_dbg_show_requests() 223 type = "Sum"; in pm_qos_dbg_show_requests() [all …]
|
/kernel/locking/ |
D | rwsem-spinlock.c | 20 enum rwsem_waiter_type type; member 73 if (waiter->type == RWSEM_WAITING_FOR_WRITE) { in __rwsem_do_wake() 103 } while (waiter->type != RWSEM_WAITING_FOR_WRITE); in __rwsem_do_wake() 148 waiter.type = RWSEM_WAITING_FOR_READ; in __down_read() 205 waiter.type = RWSEM_WAITING_FOR_WRITE; in __down_write_nested()
|