/kernel/trace/ |
D | trace_functions.c | 25 function_trace_call(unsigned long ip, unsigned long parent_ip, 28 function_stack_trace_call(unsigned long ip, unsigned long parent_ip, 31 function_no_repeats_trace_call(unsigned long ip, unsigned long parent_ip, 34 function_stack_no_repeats_trace_call(unsigned long ip, unsigned long parent_ip, 172 function_trace_call(unsigned long ip, unsigned long parent_ip, in function_trace_call() argument 184 bit = ftrace_test_recursion_trylock(ip, parent_ip); in function_trace_call() 193 trace_function(tr, ip, parent_ip, trace_ctx); in function_trace_call() 217 function_stack_trace_call(unsigned long ip, unsigned long parent_ip, in function_stack_trace_call() argument 241 trace_function(tr, ip, parent_ip, trace_ctx); in function_stack_trace_call() 251 unsigned long ip, unsigned long parent_ip) in is_repeat_check() argument [all …]
|
D | trace_recursion_record.c | 12 unsigned long ip; member 25 void ftrace_record_recursion(unsigned long ip, unsigned long parent_ip) in ftrace_record_recursion() argument 33 if (ip == cached_function) in ftrace_record_recursion() 64 if (recursed_functions[i].ip == ip) { in ftrace_record_recursion() 65 cached_function = ip; in ftrace_record_recursion() 70 cached_function = ip; in ftrace_record_recursion() 78 old = cmpxchg(&recursed_functions[index].ip, 0, ip); in ftrace_record_recursion() 81 if (old == ip) in ftrace_record_recursion() 112 cmpxchg(&recursed_functions[index].ip, ip, 0); in ftrace_record_recursion() 166 trace_seq_print_sym(tseq, record->ip, true); in recursed_function_seq_show()
|
D | ftrace.c | 125 void ftrace_ops_list_func(unsigned long ip, unsigned long parent_ip, 139 static void ftrace_pid_func(unsigned long ip, unsigned long parent_ip, in ftrace_pid_func() argument 154 op->saved_func(ip, parent_ip, op, fregs); in ftrace_pid_func() 384 unsigned long ip; member 525 kallsyms_lookup(rec->ip, NULL, NULL, NULL, str); in function_stat_show() 687 ftrace_find_profiled_func(struct ftrace_profile_stat *stat, unsigned long ip) in ftrace_find_profiled_func() argument 693 key = hash_long(ip, FTRACE_PROFILE_HASH_BITS); in ftrace_find_profiled_func() 700 if (rec->ip == ip) in ftrace_find_profiled_func() 712 key = hash_long(rec->ip, FTRACE_PROFILE_HASH_BITS); in ftrace_add_profile() 720 ftrace_profile_alloc(struct ftrace_profile_stat *stat, unsigned long ip) in ftrace_profile_alloc() argument [all …]
|
D | trace_entries.h | 64 __field_fn( unsigned long, ip ) 69 (void *)__entry->ip, (void *)__entry->parent_ip), 205 __field( unsigned long, ip ) 211 (void *)__entry->ip, __entry->fmt) 219 __field( unsigned long, ip ) 224 (void *)__entry->ip, __entry->buf), 247 __field( unsigned long, ip ) 252 (void *)__entry->ip, __entry->str) 350 __field( unsigned long, ip ) 358 (void *)__entry->ip,
|
D | trace_printk.c | 200 int __trace_bprintk(unsigned long ip, const char *fmt, ...) in __trace_bprintk() argument 212 ret = trace_vbprintk(ip, fmt, ap); in __trace_bprintk() 218 int __ftrace_vbprintk(unsigned long ip, const char *fmt, va_list ap) in __ftrace_vbprintk() argument 226 return trace_vbprintk(ip, fmt, ap); in __ftrace_vbprintk() 230 int __trace_printk(unsigned long ip, const char *fmt, ...) in __trace_printk() argument 239 ret = trace_vprintk(ip, fmt, ap); in __trace_printk() 245 int __ftrace_vprintk(unsigned long ip, const char *fmt, va_list ap) in __ftrace_vprintk() argument 250 return trace_vprintk(ip, fmt, ap); in __ftrace_vprintk()
|
D | trace_output.c | 385 unsigned long ip, unsigned long sym_flags) in seq_print_user_ip() argument 398 vma = find_vma(mm, ip); in seq_print_user_ip() 407 ip - vmstart); in seq_print_user_ip() 412 trace_seq_printf(s, " <" IP_FMT ">", ip); in seq_print_user_ip() 417 seq_print_ip_sym(struct trace_seq *s, unsigned long ip, unsigned long sym_flags) in seq_print_ip_sym() argument 419 if (!ip) { in seq_print_ip_sym() 424 trace_seq_print_sym(s, ip, sym_flags & TRACE_ITER_SYM_OFFSET); in seq_print_ip_sym() 427 trace_seq_printf(s, " <" IP_FMT ">", ip); in seq_print_ip_sym() 852 static void print_fn_trace(struct trace_seq *s, unsigned long ip, in print_fn_trace() argument 855 seq_print_ip_sym(s, ip, flags); in print_fn_trace() [all …]
|
D | trace_irqsoff.c | 140 irqsoff_tracer_call(unsigned long ip, unsigned long parent_ip, in irqsoff_tracer_call() argument 153 trace_function(tr, ip, parent_ip, trace_ctx); in irqsoff_tracer_call() 273 unsigned long ip, unsigned long parent_ip, in __trace_function() argument 277 trace_graph_function(tr, ip, parent_ip, trace_ctx); in __trace_function() 279 trace_function(tr, ip, parent_ip, trace_ctx); in __trace_function() 372 start_critical_timing(unsigned long ip, unsigned long parent_ip) in start_critical_timing() argument 395 data->critical_start = parent_ip ? : ip; in start_critical_timing() 397 __trace_function(tr, ip, parent_ip, tracing_gen_ctx()); in start_critical_timing() 405 stop_critical_timing(unsigned long ip, unsigned long parent_ip) in stop_critical_timing() argument 431 __trace_function(tr, ip, parent_ip, trace_ctx); in stop_critical_timing() [all …]
|
D | fprobe.c | 23 static void fprobe_handler(unsigned long ip, unsigned long parent_ip, in fprobe_handler() argument 36 bit = ftrace_test_recursion_trylock(ip, parent_ip); in fprobe_handler() 49 fpr->entry_ip = ip; in fprobe_handler() 55 fp->entry_handler(fp, ip, ftrace_get_regs(fregs), entry_data); in fprobe_handler() 65 static void fprobe_kprobe_handler(unsigned long ip, unsigned long parent_ip, in fprobe_kprobe_handler() argument 75 fprobe_handler(ip, parent_ip, ops, fregs); in fprobe_kprobe_handler()
|
D | trace.h | 140 unsigned long ip; member 287 unsigned long ip; member 640 unsigned long ip, 644 unsigned long ip, 782 trace_vbprintk(unsigned long ip, const char *fmt, va_list args); 784 trace_vprintk(unsigned long ip, const char *fmt, va_list args); 787 unsigned long ip, const char *fmt, va_list args); 789 unsigned long ip, const char *fmt, ...); 817 ftrace_lookup_ip(struct ftrace_hash *hash, unsigned long ip); 1046 void (*func)(unsigned long ip, [all …]
|
D | trace_stack.c | 155 static void check_stack(unsigned long ip, unsigned long *stack) in check_stack() argument 197 if (stack_dump_trace[i] == ip) in check_stack() 292 stack_trace_call(unsigned long ip, unsigned long parent_ip, in stack_trace_call() argument 308 ip += MCOUNT_INSN_SIZE; in stack_trace_call() 310 check_stack(ip, &stack); in stack_trace_call()
|
D | trace_selftest.c | 107 static void trace_selftest_test_probe1_func(unsigned long ip, in trace_selftest_test_probe1_func() argument 116 static void trace_selftest_test_probe2_func(unsigned long ip, in trace_selftest_test_probe2_func() argument 125 static void trace_selftest_test_probe3_func(unsigned long ip, in trace_selftest_test_probe3_func() argument 134 static void trace_selftest_test_global_func(unsigned long ip, in trace_selftest_test_global_func() argument 143 static void trace_selftest_test_dyn_func(unsigned long ip, in trace_selftest_test_dyn_func() argument 448 static void trace_selftest_test_recursion_func(unsigned long ip, in trace_selftest_test_recursion_func() argument 463 static void trace_selftest_test_recursion_safe_func(unsigned long ip, in trace_selftest_test_recursion_safe_func() argument 582 static void trace_selftest_test_regs_func(unsigned long ip, in trace_selftest_test_regs_func() argument
|
D | ftrace_internal.h | 14 int ftrace_ops_test(struct ftrace_ops *ops, unsigned long ip, void *regs); 36 ftrace_ops_test(struct ftrace_ops *ops, unsigned long ip, void *regs) in ftrace_ops_test() argument
|
/kernel/kcsan/ |
D | report.c | 35 unsigned long ip; member 315 replace_stack_entry(unsigned long stack_entries[], int num_entries, unsigned long ip, in replace_stack_entry() argument 322 if (kallsyms_lookup_size_offset(ip, &symbolsize, &offset)) in replace_stack_entry() 323 target_func = ip - offset; in replace_stack_entry() 336 stack_entries[skip] = ip; in replace_stack_entry() 343 WARN_ONCE(1, "Cannot find frame for %pS in stack trace", (void *)ip); in replace_stack_entry() 348 sanitize_stack_entries(unsigned long stack_entries[], int num_entries, unsigned long ip, in sanitize_stack_entries() argument 351 return ip ? replace_stack_entry(stack_entries, num_entries, ip, replaced) : in sanitize_stack_entries() 396 int skipnr = sanitize_stack_entries(stack_entries, num_stack_entries, ai->ip, &reordered_to); in print_report() 411 other_info->ai.ip, &other_reordered_to); in print_report() [all …]
|
D | core.c | 210 check_access(const volatile void *ptr, size_t size, int type, unsigned long ip); 224 scoped_access->type, scoped_access->ip); in kcsan_check_scoped_accesses() 402 int type, unsigned long ip) in find_reorder_access() argument 414 reorder_access->type == type && reorder_access->ip == ip; in find_reorder_access() 419 int type, unsigned long ip) in set_reorder_access() argument 436 reorder_access->ip = ip; in set_reorder_access() 458 unsigned long ip, in kcsan_found_watchpoint() argument 483 if (ctx->access_mask && !find_reorder_access(ctx, ptr, size, type, ip)) in kcsan_found_watchpoint() 507 kcsan_report_set_info(ptr, size, type, ip, watchpoint - watchpoints); in kcsan_found_watchpoint() 527 kcsan_setup_watchpoint(const volatile void *ptr, size_t size, int type, unsigned long ip) in kcsan_setup_watchpoint() argument [all …]
|
D | kcsan.h | 124 unsigned long ip, int watchpoint_idx); 132 unsigned long ip, enum kcsan_value_change value_change, 140 unsigned long ip, u64 old, u64 new, u64 mask);
|
/kernel/ |
D | rseq.c | 240 static bool in_rseq_cs(unsigned long ip, struct rseq_cs *rseq_cs) in in_rseq_cs() argument 242 return ip - rseq_cs->start_ip < rseq_cs->post_commit_offset; in in_rseq_cs() 247 unsigned long ip = instruction_pointer(regs); in rseq_ip_fixup() local 261 if (!in_rseq_cs(ip, &rseq_cs)) in rseq_ip_fixup() 269 trace_rseq_ip_fixup(ip, rseq_cs.start_ip, rseq_cs.post_commit_offset, in rseq_ip_fixup() 321 unsigned long ip = instruction_pointer(regs); in rseq_syscall() local 327 if (rseq_get_rseq_cs(t, &rseq_cs) || in_rseq_cs(ip, &rseq_cs)) in rseq_syscall()
|
D | softirq.c | 185 void __local_bh_disable_ip(unsigned long ip, unsigned int cnt) in __local_bh_disable_ip() argument 216 lockdep_softirqs_off(ip); in __local_bh_disable_ip() 245 void __local_bh_enable_ip(unsigned long ip, unsigned int cnt) in __local_bh_enable_ip() argument 344 void __local_bh_disable_ip(unsigned long ip, unsigned int cnt) in __local_bh_disable_ip() argument 363 lockdep_softirqs_off(ip); in __local_bh_disable_ip() 400 void __local_bh_enable_ip(unsigned long ip, unsigned int cnt) in __local_bh_enable_ip() argument 411 lockdep_softirqs_on(ip); in __local_bh_enable_ip()
|
D | kcov.c | 187 static notrace unsigned long canonicalize_ip(unsigned long ip) in canonicalize_ip() argument 190 ip -= kaslr_offset(); in canonicalize_ip() 192 return ip; in canonicalize_ip() 203 unsigned long ip = canonicalize_ip(_RET_IP_); in __sanitizer_cov_trace_pc() local 223 area[pos] = ip; in __sanitizer_cov_trace_pc() 229 static void notrace write_comp_data(u64 type, u64 arg1, u64 arg2, u64 ip) in write_comp_data() argument 239 ip = canonicalize_ip(ip); in write_comp_data() 260 area[start_index + 3] = ip; in write_comp_data()
|
/kernel/locking/ |
D | mutex.c | 543 static noinline void __sched __mutex_unlock_slowpath(struct mutex *lock, unsigned long ip); 592 struct lockdep_map *nest_lock, unsigned long ip, in __mutex_lock_common() argument 625 mutex_acquire_nest(&lock->dep_map, subclass, 0, nest_lock, ip); in __mutex_lock_common() 631 lock_acquired(&lock->dep_map, ip); in __mutex_lock_common() 656 lock_contended(&lock->dep_map, ip); in __mutex_lock_common() 746 lock_acquired(&lock->dep_map, ip); in __mutex_lock_common() 765 mutex_release(&lock->dep_map, ip); in __mutex_lock_common() 772 struct lockdep_map *nest_lock, unsigned long ip) in __mutex_lock() argument 774 return __mutex_lock_common(lock, state, subclass, nest_lock, ip, NULL, false); in __mutex_lock() 779 unsigned long ip, struct ww_acquire_ctx *ww_ctx) in __ww_mutex_lock() argument [all …]
|
D | lockdep.c | 251 static int lock_point(unsigned long points[], unsigned long ip) in lock_point() argument 257 points[i] = ip; in lock_point() 260 if (points[i] == ip) in lock_point() 4320 void noinstr lockdep_hardirqs_on(unsigned long ip) in lockdep_hardirqs_on() argument 4378 trace->hardirq_enable_ip = ip; in lockdep_hardirqs_on() 4387 void noinstr lockdep_hardirqs_off(unsigned long ip) in lockdep_hardirqs_off() argument 4417 trace->hardirq_disable_ip = ip; in lockdep_hardirqs_off() 4429 void lockdep_softirqs_on(unsigned long ip) in lockdep_softirqs_on() argument 4453 trace->softirq_enable_ip = ip; in lockdep_softirqs_on() 4469 void lockdep_softirqs_off(unsigned long ip) in lockdep_softirqs_off() argument [all …]
|
D | ww_rt_mutex.c | 39 unsigned int state, unsigned long ip) in __ww_rt_mutex_lock() argument 63 mutex_acquire_nest(&rtm->dep_map, 0, 0, nest_lock, ip); in __ww_rt_mutex_lock() 74 mutex_release(&rtm->dep_map, ip); in __ww_rt_mutex_lock()
|
D | lockdep_proc.c | 534 char ip[32]; in seq_stats() local 542 snprintf(ip, sizeof(ip), "[<%p>]", in seq_stats() 546 ip, (void *)class->contention_point[i]); in seq_stats() 549 char ip[32]; in seq_stats() local 557 snprintf(ip, sizeof(ip), "[<%p>]", in seq_stats() 561 ip, (void *)class->contending_point[i]); in seq_stats()
|
D | rtmutex_api.c | 511 unsigned long ip) in __mutex_lock_common() argument 516 mutex_acquire_nest(&lock->dep_map, subclass, 0, nest_lock, ip); in __mutex_lock_common() 519 mutex_release(&lock->dep_map, ip); in __mutex_lock_common() 521 lock_acquired(&lock->dep_map, ip); in __mutex_lock_common()
|
/kernel/bpf/ |
D | stackmap.c | 145 id_offs[i].ip = ips[i]; in stack_map_get_build_id_offset() 162 id_offs[i].ip = ips[i]; in stack_map_get_build_id_offset() 188 entry->nr = stack_trace_save_tsk(task, (unsigned long *)entry->ip, in get_callchain_entry_for_task() 196 unsigned long *from = (unsigned long *) entry->ip; in get_callchain_entry_for_task() 197 u64 *to = entry->ip; in get_callchain_entry_for_task() 230 ips = trace->ip + skip; in __bpf_get_stackid() 324 if (trace->ip[nr_kernel] == PERF_CONTEXT_USER) in count_kernel_ip() 447 ips = trace->ip + skip; in __bpf_get_stack()
|
/kernel/trace/rv/monitors/wip/ |
D | wip.c | 22 static void handle_preempt_disable(void *data, unsigned long ip, unsigned long parent_ip) in handle_preempt_disable() argument 27 static void handle_preempt_enable(void *data, unsigned long ip, unsigned long parent_ip) in handle_preempt_enable() argument
|