Home
last modified time | relevance | path

Searched refs:code (Results 1 – 25 of 26) sorted by relevance

12

/kernel/bpf/
Ddisasm.c112 insn->code, insn->dst_reg, in print_bpf_end_insn()
113 BPF_SRC(insn->code) == BPF_TO_BE ? "be" : "le", in print_bpf_end_insn()
122 u8 class = BPF_CLASS(insn->code); in print_bpf_insn()
125 if (BPF_OP(insn->code) == BPF_END) { in print_bpf_insn()
127 verbose(cbs->private_data, "BUG_alu64_%02x\n", insn->code); in print_bpf_insn()
130 } else if (BPF_OP(insn->code) == BPF_NEG) { in print_bpf_insn()
132 insn->code, class == BPF_ALU ? 'w' : 'r', in print_bpf_insn()
135 } else if (BPF_SRC(insn->code) == BPF_X) { in print_bpf_insn()
137 insn->code, class == BPF_ALU ? 'w' : 'r', in print_bpf_insn()
139 bpf_alu_string[BPF_OP(insn->code) >> 4], in print_bpf_insn()
[all …]
Dverifier.c1503 if (insn[i].code != (BPF_JMP | BPF_CALL)) in check_subprogs()
1530 u8 code = insn[i].code; in check_subprogs() local
1532 if (code == (BPF_JMP | BPF_CALL) && in check_subprogs()
1536 if (BPF_CLASS(code) == BPF_LD && in check_subprogs()
1537 (BPF_MODE(code) == BPF_ABS || BPF_MODE(code) == BPF_IND)) in check_subprogs()
1539 if (BPF_CLASS(code) != BPF_JMP && BPF_CLASS(code) != BPF_JMP32) in check_subprogs()
1541 if (BPF_OP(code) == BPF_EXIT || BPF_OP(code) == BPF_CALL) in check_subprogs()
1554 if (code != (BPF_JMP | BPF_EXIT) && in check_subprogs()
1555 code != (BPF_JMP | BPF_JA)) { in check_subprogs()
1626 u8 code, class, op; in is_reg64() local
[all …]
Dcore.c301 dst[i].code == (BPF_LD | BPF_IMM | BPF_DW) && in bpf_prog_calc_tag()
307 dst[i].code == 0 && in bpf_prog_calc_tag()
391 u8 code; in bpf_adj_branches() local
401 code = insn->code; in bpf_adj_branches()
402 if ((BPF_CLASS(code) != BPF_JMP && in bpf_adj_branches()
403 BPF_CLASS(code) != BPF_JMP32) || in bpf_adj_branches()
404 BPF_OP(code) == BPF_EXIT) in bpf_adj_branches()
407 if (BPF_OP(code) == BPF_CALL) { in bpf_adj_branches()
1003 (from->code == (BPF_ALU | BPF_MOV | BPF_K) || in bpf_jit_blind_insn()
1004 from->code == (BPF_ALU64 | BPF_MOV | BPF_K))) { in bpf_jit_blind_insn()
[all …]
Dsyscall.c3343 u8 code; in bpf_insn_prepare_dump() local
3352 code = insns[i].code; in bpf_insn_prepare_dump()
3354 if (code == (BPF_JMP | BPF_TAIL_CALL)) { in bpf_insn_prepare_dump()
3355 insns[i].code = BPF_JMP | BPF_CALL; in bpf_insn_prepare_dump()
3359 if (code == (BPF_JMP | BPF_CALL) || in bpf_insn_prepare_dump()
3360 code == (BPF_JMP | BPF_CALL_ARGS)) { in bpf_insn_prepare_dump()
3361 if (code == (BPF_JMP | BPF_CALL_ARGS)) in bpf_insn_prepare_dump()
3362 insns[i].code = BPF_JMP | BPF_CALL; in bpf_insn_prepare_dump()
3367 if (BPF_CLASS(code) == BPF_LDX && BPF_MODE(code) == BPF_PROBE_MEM) { in bpf_insn_prepare_dump()
3368 insns[i].code = BPF_LDX | BPF_SIZE(code) | BPF_MEM; in bpf_insn_prepare_dump()
[all …]
Dcgroup.c1758 BPF_SIZE(si->code), si->dst_reg, si->src_reg, in sysctl_convert_ctx_access()
1796 read_size = bpf_size_to_bytes(BPF_SIZE(si->code)); in sysctl_convert_ctx_access()
1798 BPF_SIZE(si->code), si->dst_reg, si->dst_reg, in sysctl_convert_ctx_access()
/kernel/trace/
Dtrace_probe_tmpl.h7 fetch_store_raw(unsigned long val, struct fetch_insn *code, void *buf) in fetch_store_raw() argument
9 switch (code->size) { in fetch_store_raw()
29 fetch_apply_bitfield(struct fetch_insn *code, void *buf) in fetch_apply_bitfield() argument
31 switch (code->basesize) { in fetch_apply_bitfield()
33 *(u8 *)buf <<= code->lshift; in fetch_apply_bitfield()
34 *(u8 *)buf >>= code->rshift; in fetch_apply_bitfield()
37 *(u16 *)buf <<= code->lshift; in fetch_apply_bitfield()
38 *(u16 *)buf >>= code->rshift; in fetch_apply_bitfield()
41 *(u32 *)buf <<= code->lshift; in fetch_apply_bitfield()
42 *(u32 *)buf >>= code->rshift; in fetch_apply_bitfield()
[all …]
Dtrace_probe.c273 struct fetch_insn *code, unsigned int flags, int offs) in parse_probe_vars() argument
281 code->op = FETCH_OP_RETVAL; in parse_probe_vars()
288 code->op = FETCH_OP_STACKP; in parse_probe_vars()
298 code->op = FETCH_OP_STACK; in parse_probe_vars()
299 code->param = (unsigned int)param; in parse_probe_vars()
304 code->op = FETCH_OP_COMM; in parse_probe_vars()
316 code->op = FETCH_OP_ARG; in parse_probe_vars()
317 code->param = (unsigned int)param - 1; in parse_probe_vars()
358 struct fetch_insn *code = *pcode; in parse_probe_arg() local
367 ret = parse_probe_vars(arg + 1, type, code, flags, offs); in parse_probe_arg()
[all …]
Dtrace_uprobe.c220 process_fetch_insn(struct fetch_insn *code, void *rec, void *dest, in process_fetch_insn() argument
227 switch (code->op) { in process_fetch_insn()
229 val = regs_get_register(regs, code->param); in process_fetch_insn()
232 val = get_user_stack_nth(regs, code->param); in process_fetch_insn()
241 val = code->immediate; in process_fetch_insn()
247 val = (unsigned long)code->data; in process_fetch_insn()
250 val = translate_user_vaddr(code->immediate); in process_fetch_insn()
255 code++; in process_fetch_insn()
257 return process_fetch_insn_bottom(code, val, dest, base); in process_fetch_insn()
Dtrace_kprobe.c1335 process_fetch_insn(struct fetch_insn *code, void *rec, void *dest, in process_fetch_insn() argument
1343 switch (code->op) { in process_fetch_insn()
1345 val = regs_get_register(regs, code->param); in process_fetch_insn()
1348 val = regs_get_kernel_stack_nth(regs, code->param); in process_fetch_insn()
1357 val = code->immediate; in process_fetch_insn()
1363 val = (unsigned long)code->data; in process_fetch_insn()
1367 val = regs_get_kernel_argument(regs, code->param); in process_fetch_insn()
1371 code++; in process_fetch_insn()
1376 code++; in process_fetch_insn()
1378 return process_fetch_insn_bottom(code, val, dest, base); in process_fetch_insn()
Dtrace_probe.h215 struct fetch_insn *code; member
/kernel/
DKconfig.preempt25 "explicit preemption points" to the kernel code. These new
45 all kernel code (that is not executing in a critical section)
52 and a slight runtime overhead to kernel code.
68 low level and critical code paths (entry code, scheduler, low
Dseccomp.c273 u16 code = ftest->code; in seccomp_check_filter() local
276 switch (code) { in seccomp_check_filter()
278 ftest->code = BPF_LDX | BPF_W | BPF_ABS; in seccomp_check_filter()
284 ftest->code = BPF_LD | BPF_IMM; in seccomp_check_filter()
288 ftest->code = BPF_LDX | BPF_IMM; in seccomp_check_filter()
722 u16 code = insn->code; in seccomp_is_const_allow() local
725 switch (code) { in seccomp_is_const_allow()
749 switch (BPF_OP(code)) { in seccomp_is_const_allow()
Dexit.c763 void __noreturn do_exit(long code) in do_exit() argument
801 ptrace_event(PTRACE_EVENT_EXIT, code); in do_exit()
824 trace_android_vh_exit_check(current, code, group_dead); in do_exit()
832 tsk->signal->group_exit_code ?: (int)code); in do_exit()
841 acct_collect(code, group_dead); in do_exit()
846 tsk->exit_code = code; in do_exit()
942 void complete_and_exit(struct completion *comp, long code) in complete_and_exit() argument
947 do_exit(code); in complete_and_exit()
Djump_label.c70 jea->code = jeb->code - delta; in jump_label_swap()
74 jeb->code = tmp.code + delta; in jump_label_swap()
Dsignal.c1665 int force_sig_fault_to_task(int sig, int code, void __user *addr in force_sig_fault_to_task() argument
1675 info.si_code = code; in force_sig_fault_to_task()
1688 int force_sig_fault(int sig, int code, void __user *addr in force_sig_fault() argument
1692 return force_sig_fault_to_task(sig, code, addr in force_sig_fault()
1697 int send_sig_fault(int sig, int code, void __user *addr in send_sig_fault() argument
1707 info.si_code = code; in send_sig_fault()
1720 int force_sig_mceerr(int code, void __user *addr, short lsb) in force_sig_mceerr() argument
1724 WARN_ON((code != BUS_MCEERR_AO) && (code != BUS_MCEERR_AR)); in force_sig_mceerr()
1728 info.si_code = code; in force_sig_mceerr()
1734 int send_sig_mceerr(int code, void __user *addr, short lsb, struct task_struct *t) in send_sig_mceerr() argument
[all …]
Dauditsc.c2718 void audit_seccomp(unsigned long syscall, long signr, int code) in audit_seccomp() argument
2728 in_compat_syscall(), KSTK_EIP(current), code); in audit_seccomp()
Dmodule.c348 void __noreturn __module_put_and_exit(struct module *mod, long code) in __module_put_and_exit() argument
351 do_exit(code); in __module_put_and_exit()
/kernel/gcov/
DKconfig10 This option enables gcov-based code profiling (e.g. for code coverage
/kernel/irq/
DKconfig3 # Options selectable by the architecture code
30 # and should not be used in new code. Use irq domains instead.
/kernel/rcu/
DKconfig77 This option enables generic infrastructure code supporting
103 hotplug code paths. It can force IPIs on online CPUs, including
109 This option enables RCU CPU stall code that is common between
132 code paths on small(er) systems.
DKconfig.debug112 bugs in arch-specific NO_HZ code.
/kernel/livepatch/
DKconfig20 to new function code contained in the patch module.
/kernel/power/
DKconfig184 code. This is helpful when debugging and reporting PM bugs, like
229 This enables code to save the last PM event point across
233 The architecture specific code must provide the extern
247 This enables some cheesy code to save the last PM event point in the
/kernel/debug/
Ddebug_core.c1058 dbg_notify_reboot(struct notifier_block *this, unsigned long code, void *x) in dbg_notify_reboot() argument
1073 gdbstub_exit(code); in dbg_notify_reboot()
/kernel/dma/
DKconfig16 # IOMMU drivers that can bypass the IOMMU code and optionally use the direct

12