/kernel/ |
D | stacktrace.c | 78 unsigned int skip; member 89 if (c->skip > 0) { in stack_trace_consume_entry() 90 c->skip--; in stack_trace_consume_entry() 119 .skip = skipnr + 1, in stack_trace_save() 144 .skip = skipnr + (current == tsk), in stack_trace_save_tsk() 172 .skip = skipnr, in stack_trace_save_regs() 275 .skip = skipnr + 1, in stack_trace_save() 300 .skip = skipnr + (current == task), in stack_trace_save_tsk() 322 .skip = skipnr, in stack_trace_save_regs()
|
D | seccomp.c | 1221 goto skip; in __seccomp_filter() 1228 goto skip; in __seccomp_filter() 1240 goto skip; in __seccomp_filter() 1256 goto skip; in __seccomp_filter() 1260 goto skip; in __seccomp_filter() 1275 goto skip; in __seccomp_filter() 1311 skip: in __seccomp_filter()
|
D | resource.c | 937 goto skip; in __adjust_resource() 953 skip: in __adjust_resource()
|
/kernel/kcsan/ |
D | report.c | 281 int len, skip; in get_stack_skipnr() local 283 for (skip = 0; skip < num_entries; ++skip) { in get_stack_skipnr() 284 len = scnprintf(buf, sizeof(buf), "%ps", (void *)stack_entries[skip]); in get_stack_skipnr() 306 return skip; in get_stack_skipnr() 320 int skip; in replace_stack_entry() local 327 for (skip = 0; skip < num_entries; ++skip) { in replace_stack_entry() 328 unsigned long func = stack_entries[skip]; in replace_stack_entry() 335 *replaced = stack_entries[skip]; in replace_stack_entry() 336 stack_entries[skip] = ip; in replace_stack_entry() 337 return skip; in replace_stack_entry()
|
/kernel/bpf/ |
D | stackmap.c | 218 u32 skip = flags & BPF_F_SKIP_FIELD_MASK; in __bpf_get_stackid() local 224 if (trace->nr <= skip) in __bpf_get_stackid() 228 trace_nr = trace->nr - skip; in __bpf_get_stackid() 230 ips = trace->ip + skip; in __bpf_get_stackid() 287 u32 skip = flags & BPF_F_SKIP_FIELD_MASK; in BPF_CALL_3() local 296 max_depth += skip; in BPF_CALL_3() 367 u64 skip = flags & BPF_F_SKIP_FIELD_MASK; in BPF_CALL_3() local 369 skip += nr_kernel; in BPF_CALL_3() 370 if (skip > BPF_F_SKIP_FIELD_MASK) in BPF_CALL_3() 373 flags = (flags & ~BPF_F_SKIP_FIELD_MASK) | skip; in BPF_CALL_3() [all …]
|
/kernel/sched/ |
D | cpupri.c | 71 int skip = 0; in __cpupri_find() local 74 skip = 1; in __cpupri_find() 96 if (skip) in __cpupri_find()
|
D | rt.c | 931 int skip; in do_sched_rt_period_timer() local 940 skip = !rt_rq->rt_time && !rt_rq->rt_nr_running; in do_sched_rt_period_timer() 942 if (skip) in do_sched_rt_period_timer() 2538 goto skip; in pull_rt_task() 2555 skip: in pull_rt_task()
|
D | deadline.c | 2425 goto skip; in pull_dl_task() 2445 goto skip; in pull_dl_task() 2459 skip: in pull_dl_task()
|
D | core.c | 8383 bool skip = false; in sched_setaffinity() local 8412 trace_android_vh_sched_setaffinity_early(p, in_mask, &skip); in sched_setaffinity() 8413 if (skip) in sched_setaffinity() 8531 long skip = 0; in do_sched_yield() local 8533 trace_android_rvh_before_do_sched_yield(&skip); in do_sched_yield() 8534 if (skip) in do_sched_yield()
|
D | fair.c | 4858 if (cfs_rq->skip != se) in __clear_buddies_skip() 4861 cfs_rq->skip = NULL; in __clear_buddies_skip() 4873 if (cfs_rq->skip == se) in clear_buddies() 5062 if (cfs_rq->skip && cfs_rq->skip == se) { in pick_next_entity() 7715 cfs_rq_of(se)->skip = se; in set_skip_buddy()
|
D | sched.h | 588 struct sched_entity *skip; member
|
/kernel/debug/kdb/ |
D | kdb_main.c | 2089 int skip = 0; in kdb_dmesg() local 2126 skip = adjust; in kdb_dmesg() 2129 skip = n - lines - adjust; in kdb_dmesg() 2134 skip = n; in kdb_dmesg() 2135 } else if (skip < 0) { in kdb_dmesg() 2136 lines += skip; in kdb_dmesg() 2137 skip = 0; in kdb_dmesg() 2145 if (skip >= n || skip < 0) in kdb_dmesg() 2150 if (skip) { in kdb_dmesg() 2151 skip--; in kdb_dmesg()
|
/kernel/events/ |
D | hw_breakpoint_test.c | 80 static bool fill_bp_slots(struct kunit *test, int *id, int cpu, struct task_struct *tsk, int skip) in fill_bp_slots() argument 82 for (int i = 0; i < get_test_bp_slots() - skip; ++i) in fill_bp_slots()
|
/kernel/power/ |
D | hibernate.c | 872 goto skip; in hibernate_quiet_exec() 876 skip: in hibernate_quiet_exec()
|
D | swap.c | 457 bool skip = false; in swap_write_page() local 471 trace_android_vh_skip_swap_map_write(&skip); in swap_write_page() 472 if (!skip) { in swap_write_page()
|
/kernel/locking/ |
D | lockdep.c | 1718 bool (*skip)(struct lock_list *entry, void *data), in __bfs() 1782 if (skip && skip(lock, data)) in __bfs() 1827 bool (*skip)(struct lock_list *entry, void *data), in __bfs_forwards() 1830 return __bfs(src_entry, data, match, skip, target_entry, in __bfs_forwards() 1839 bool (*skip)(struct lock_list *entry, void *data), in __bfs_backwards() 1842 return __bfs(src_entry, data, match, skip, target_entry, in __bfs_backwards() 2126 bool (*skip)(struct lock_list *entry, void *data), in check_path() 2131 ret = __bfs_forwards(src_entry, target, match, skip, target_entry); in check_path()
|
/kernel/trace/ |
D | trace.c | 915 int skip, struct pt_regs *regs); 919 int skip, struct pt_regs *regs); 924 int skip, struct pt_regs *regs) in __ftrace_trace_stack() argument 930 int skip, struct pt_regs *regs) in ftrace_trace_stack() argument 3038 int skip, struct pt_regs *regs) in __ftrace_trace_stack() argument 3053 skip++; in __ftrace_trace_stack() 3078 size, skip); in __ftrace_trace_stack() 3080 nr_entries = stack_trace_save(fstack->calls, size, skip); in __ftrace_trace_stack() 3108 int skip, struct pt_regs *regs) in ftrace_trace_stack() argument 3113 __ftrace_trace_stack(buffer, trace_ctx, skip, regs); in ftrace_trace_stack() [all …]
|
D | trace_osnoise.c | 606 static void timerlat_save_stack(int skip) in timerlat_save_stack() argument 615 nr_entries = stack_trace_save(fstack->calls, size, skip); in timerlat_save_stack()
|
D | trace.h | 712 void __trace_stack(struct trace_array *tr, unsigned int trace_ctx, int skip); 715 int skip) in __trace_stack() argument
|
/kernel/printk/ |
D | printk.c | 1542 size_t skip; in syslog_print() local 1561 skip = syslog_partial; in syslog_print() 1579 err = copy_to_user(buf, text + skip, n); in syslog_print() 2725 goto skip; in console_emit_next_record() 2759 skip: in console_emit_next_record()
|