/kernel/bpf/ |
D | disasm.c | 112 insn->code, insn->dst_reg, in print_bpf_end_insn() 113 BPF_SRC(insn->code) == BPF_TO_BE ? "be" : "le", in print_bpf_end_insn() 122 u8 class = BPF_CLASS(insn->code); in print_bpf_insn() 125 if (BPF_OP(insn->code) == BPF_END) { in print_bpf_insn() 127 verbose(cbs->private_data, "BUG_alu64_%02x\n", insn->code); in print_bpf_insn() 130 } else if (BPF_OP(insn->code) == BPF_NEG) { in print_bpf_insn() 132 insn->code, class == BPF_ALU ? 'w' : 'r', in print_bpf_insn() 135 } else if (BPF_SRC(insn->code) == BPF_X) { in print_bpf_insn() 137 insn->code, class == BPF_ALU ? 'w' : 'r', in print_bpf_insn() 139 bpf_alu_string[BPF_OP(insn->code) >> 4], in print_bpf_insn() [all …]
|
D | verifier.c | 1503 if (insn[i].code != (BPF_JMP | BPF_CALL)) in check_subprogs() 1530 u8 code = insn[i].code; in check_subprogs() local 1532 if (code == (BPF_JMP | BPF_CALL) && in check_subprogs() 1536 if (BPF_CLASS(code) == BPF_LD && in check_subprogs() 1537 (BPF_MODE(code) == BPF_ABS || BPF_MODE(code) == BPF_IND)) in check_subprogs() 1539 if (BPF_CLASS(code) != BPF_JMP && BPF_CLASS(code) != BPF_JMP32) in check_subprogs() 1541 if (BPF_OP(code) == BPF_EXIT || BPF_OP(code) == BPF_CALL) in check_subprogs() 1554 if (code != (BPF_JMP | BPF_EXIT) && in check_subprogs() 1555 code != (BPF_JMP | BPF_JA)) { in check_subprogs() 1626 u8 code, class, op; in is_reg64() local [all …]
|
D | core.c | 301 dst[i].code == (BPF_LD | BPF_IMM | BPF_DW) && in bpf_prog_calc_tag() 307 dst[i].code == 0 && in bpf_prog_calc_tag() 391 u8 code; in bpf_adj_branches() local 401 code = insn->code; in bpf_adj_branches() 402 if ((BPF_CLASS(code) != BPF_JMP && in bpf_adj_branches() 403 BPF_CLASS(code) != BPF_JMP32) || in bpf_adj_branches() 404 BPF_OP(code) == BPF_EXIT) in bpf_adj_branches() 407 if (BPF_OP(code) == BPF_CALL) { in bpf_adj_branches() 1003 (from->code == (BPF_ALU | BPF_MOV | BPF_K) || in bpf_jit_blind_insn() 1004 from->code == (BPF_ALU64 | BPF_MOV | BPF_K))) { in bpf_jit_blind_insn() [all …]
|
D | syscall.c | 3343 u8 code; in bpf_insn_prepare_dump() local 3352 code = insns[i].code; in bpf_insn_prepare_dump() 3354 if (code == (BPF_JMP | BPF_TAIL_CALL)) { in bpf_insn_prepare_dump() 3355 insns[i].code = BPF_JMP | BPF_CALL; in bpf_insn_prepare_dump() 3359 if (code == (BPF_JMP | BPF_CALL) || in bpf_insn_prepare_dump() 3360 code == (BPF_JMP | BPF_CALL_ARGS)) { in bpf_insn_prepare_dump() 3361 if (code == (BPF_JMP | BPF_CALL_ARGS)) in bpf_insn_prepare_dump() 3362 insns[i].code = BPF_JMP | BPF_CALL; in bpf_insn_prepare_dump() 3367 if (BPF_CLASS(code) == BPF_LDX && BPF_MODE(code) == BPF_PROBE_MEM) { in bpf_insn_prepare_dump() 3368 insns[i].code = BPF_LDX | BPF_SIZE(code) | BPF_MEM; in bpf_insn_prepare_dump() [all …]
|
D | cgroup.c | 1758 BPF_SIZE(si->code), si->dst_reg, si->src_reg, in sysctl_convert_ctx_access() 1796 read_size = bpf_size_to_bytes(BPF_SIZE(si->code)); in sysctl_convert_ctx_access() 1798 BPF_SIZE(si->code), si->dst_reg, si->dst_reg, in sysctl_convert_ctx_access()
|
/kernel/trace/ |
D | trace_probe_tmpl.h | 7 fetch_store_raw(unsigned long val, struct fetch_insn *code, void *buf) in fetch_store_raw() argument 9 switch (code->size) { in fetch_store_raw() 29 fetch_apply_bitfield(struct fetch_insn *code, void *buf) in fetch_apply_bitfield() argument 31 switch (code->basesize) { in fetch_apply_bitfield() 33 *(u8 *)buf <<= code->lshift; in fetch_apply_bitfield() 34 *(u8 *)buf >>= code->rshift; in fetch_apply_bitfield() 37 *(u16 *)buf <<= code->lshift; in fetch_apply_bitfield() 38 *(u16 *)buf >>= code->rshift; in fetch_apply_bitfield() 41 *(u32 *)buf <<= code->lshift; in fetch_apply_bitfield() 42 *(u32 *)buf >>= code->rshift; in fetch_apply_bitfield() [all …]
|
D | trace_probe.c | 273 struct fetch_insn *code, unsigned int flags, int offs) in parse_probe_vars() argument 281 code->op = FETCH_OP_RETVAL; in parse_probe_vars() 288 code->op = FETCH_OP_STACKP; in parse_probe_vars() 298 code->op = FETCH_OP_STACK; in parse_probe_vars() 299 code->param = (unsigned int)param; in parse_probe_vars() 304 code->op = FETCH_OP_COMM; in parse_probe_vars() 316 code->op = FETCH_OP_ARG; in parse_probe_vars() 317 code->param = (unsigned int)param - 1; in parse_probe_vars() 358 struct fetch_insn *code = *pcode; in parse_probe_arg() local 367 ret = parse_probe_vars(arg + 1, type, code, flags, offs); in parse_probe_arg() [all …]
|
D | trace_uprobe.c | 220 process_fetch_insn(struct fetch_insn *code, void *rec, void *dest, in process_fetch_insn() argument 227 switch (code->op) { in process_fetch_insn() 229 val = regs_get_register(regs, code->param); in process_fetch_insn() 232 val = get_user_stack_nth(regs, code->param); in process_fetch_insn() 241 val = code->immediate; in process_fetch_insn() 247 val = (unsigned long)code->data; in process_fetch_insn() 250 val = translate_user_vaddr(code->immediate); in process_fetch_insn() 255 code++; in process_fetch_insn() 257 return process_fetch_insn_bottom(code, val, dest, base); in process_fetch_insn()
|
D | trace_kprobe.c | 1335 process_fetch_insn(struct fetch_insn *code, void *rec, void *dest, in process_fetch_insn() argument 1343 switch (code->op) { in process_fetch_insn() 1345 val = regs_get_register(regs, code->param); in process_fetch_insn() 1348 val = regs_get_kernel_stack_nth(regs, code->param); in process_fetch_insn() 1357 val = code->immediate; in process_fetch_insn() 1363 val = (unsigned long)code->data; in process_fetch_insn() 1367 val = regs_get_kernel_argument(regs, code->param); in process_fetch_insn() 1371 code++; in process_fetch_insn() 1376 code++; in process_fetch_insn() 1378 return process_fetch_insn_bottom(code, val, dest, base); in process_fetch_insn()
|
D | trace_probe.h | 215 struct fetch_insn *code; member
|
/kernel/ |
D | Kconfig.preempt | 25 "explicit preemption points" to the kernel code. These new 45 all kernel code (that is not executing in a critical section) 52 and a slight runtime overhead to kernel code. 68 low level and critical code paths (entry code, scheduler, low
|
D | seccomp.c | 273 u16 code = ftest->code; in seccomp_check_filter() local 276 switch (code) { in seccomp_check_filter() 278 ftest->code = BPF_LDX | BPF_W | BPF_ABS; in seccomp_check_filter() 284 ftest->code = BPF_LD | BPF_IMM; in seccomp_check_filter() 288 ftest->code = BPF_LDX | BPF_IMM; in seccomp_check_filter() 722 u16 code = insn->code; in seccomp_is_const_allow() local 725 switch (code) { in seccomp_is_const_allow() 749 switch (BPF_OP(code)) { in seccomp_is_const_allow()
|
D | exit.c | 763 void __noreturn do_exit(long code) in do_exit() argument 801 ptrace_event(PTRACE_EVENT_EXIT, code); in do_exit() 824 trace_android_vh_exit_check(current, code, group_dead); in do_exit() 832 tsk->signal->group_exit_code ?: (int)code); in do_exit() 841 acct_collect(code, group_dead); in do_exit() 846 tsk->exit_code = code; in do_exit() 942 void complete_and_exit(struct completion *comp, long code) in complete_and_exit() argument 947 do_exit(code); in complete_and_exit()
|
D | jump_label.c | 70 jea->code = jeb->code - delta; in jump_label_swap() 74 jeb->code = tmp.code + delta; in jump_label_swap()
|
D | signal.c | 1665 int force_sig_fault_to_task(int sig, int code, void __user *addr in force_sig_fault_to_task() argument 1675 info.si_code = code; in force_sig_fault_to_task() 1688 int force_sig_fault(int sig, int code, void __user *addr in force_sig_fault() argument 1692 return force_sig_fault_to_task(sig, code, addr in force_sig_fault() 1697 int send_sig_fault(int sig, int code, void __user *addr in send_sig_fault() argument 1707 info.si_code = code; in send_sig_fault() 1720 int force_sig_mceerr(int code, void __user *addr, short lsb) in force_sig_mceerr() argument 1724 WARN_ON((code != BUS_MCEERR_AO) && (code != BUS_MCEERR_AR)); in force_sig_mceerr() 1728 info.si_code = code; in force_sig_mceerr() 1734 int send_sig_mceerr(int code, void __user *addr, short lsb, struct task_struct *t) in send_sig_mceerr() argument [all …]
|
D | auditsc.c | 2718 void audit_seccomp(unsigned long syscall, long signr, int code) in audit_seccomp() argument 2728 in_compat_syscall(), KSTK_EIP(current), code); in audit_seccomp()
|
D | module.c | 348 void __noreturn __module_put_and_exit(struct module *mod, long code) in __module_put_and_exit() argument 351 do_exit(code); in __module_put_and_exit()
|
/kernel/gcov/ |
D | Kconfig | 10 This option enables gcov-based code profiling (e.g. for code coverage
|
/kernel/irq/ |
D | Kconfig | 3 # Options selectable by the architecture code 30 # and should not be used in new code. Use irq domains instead.
|
/kernel/rcu/ |
D | Kconfig | 77 This option enables generic infrastructure code supporting 103 hotplug code paths. It can force IPIs on online CPUs, including 109 This option enables RCU CPU stall code that is common between 132 code paths on small(er) systems.
|
D | Kconfig.debug | 112 bugs in arch-specific NO_HZ code.
|
/kernel/livepatch/ |
D | Kconfig | 20 to new function code contained in the patch module.
|
/kernel/power/ |
D | Kconfig | 184 code. This is helpful when debugging and reporting PM bugs, like 229 This enables code to save the last PM event point across 233 The architecture specific code must provide the extern 247 This enables some cheesy code to save the last PM event point in the
|
/kernel/debug/ |
D | debug_core.c | 1058 dbg_notify_reboot(struct notifier_block *this, unsigned long code, void *x) in dbg_notify_reboot() argument 1073 gdbstub_exit(code); in dbg_notify_reboot()
|
/kernel/dma/ |
D | Kconfig | 16 # IOMMU drivers that can bypass the IOMMU code and optionally use the direct
|