• Home
  • Raw
  • Download

Lines Matching +full:0 +full:x48

34 	do { prog = emit_code(prog, bytes, len); } while (0)
42 do { EMIT1(b1); EMIT(off, 4); } while (0)
44 do { EMIT2(b1, b2); EMIT(off, 4); } while (0)
46 do { EMIT3(b1, b2, b3); EMIT(off, 4); } while (0)
48 do { EMIT4(b1, b2, b3, b4); EMIT(off, 4); } while (0)
75 EMIT3(add_2mod(0x48, DST, SRC), 0x89, add_2reg(0xC0, DST, SRC)); \
76 } while (0)
89 return 0; in bpf_size_to_x86_bytes()
94 * Add 0x10 (and an extra 0x0f) to generate far jumps (. + s32)
96 #define X86_JB 0x72
97 #define X86_JAE 0x73
98 #define X86_JE 0x74
99 #define X86_JNE 0x75
100 #define X86_JBE 0x76
101 #define X86_JA 0x77
102 #define X86_JL 0x7C
103 #define X86_JGE 0x7D
104 #define X86_JLE 0x7E
105 #define X86_JG 0x7F
122 [BPF_REG_0] = 0, /* RAX */
127 [BPF_REG_5] = 0, /* R8 */
216 [BPF_ADD] = 0x01,
217 [BPF_SUB] = 0x29,
218 [BPF_AND] = 0x21,
219 [BPF_OR] = 0x09,
220 [BPF_XOR] = 0x31,
221 [BPF_LSH] = 0xE0,
222 [BPF_RSH] = 0xE8,
223 [BPF_ARSH] = 0xF8,
229 memset(area, 0xcc, size); in jit_fill_hole()
234 return IS_ERR_OR_NULL(text_poke_set(dst, 0xcc, len)); in bpf_arch_text_invalidate()
262 if (callee_regs_used[0]) in push_callee_regs()
263 EMIT1(0x53); /* push rbx */ in push_callee_regs()
265 EMIT2(0x41, 0x55); /* push r13 */ in push_callee_regs()
267 EMIT2(0x41, 0x56); /* push r14 */ in push_callee_regs()
269 EMIT2(0x41, 0x57); /* push r15 */ in push_callee_regs()
278 EMIT2(0x41, 0x5F); /* pop r15 */ in pop_callee_regs()
280 EMIT2(0x41, 0x5E); /* pop r14 */ in pop_callee_regs()
282 EMIT2(0x41, 0x5D); /* pop r13 */ in pop_callee_regs()
283 if (callee_regs_used[0]) in pop_callee_regs()
284 EMIT1(0x5B); /* pop rbx */ in pop_callee_regs()
306 EMIT2(0x31, 0xC0); /* xor eax, eax */ in emit_prologue()
308 EMIT2(0x66, 0x90); /* nop2 */ in emit_prologue()
310 EMIT1(0x55); /* push rbp */ in emit_prologue()
311 EMIT3(0x48, 0x89, 0xE5); /* mov rbp, rsp */ in emit_prologue()
318 EMIT3_off32(0x48, 0x81, 0xEC, round_up(stack_depth, 8)); in emit_prologue()
320 EMIT1(0x50); /* push rax */ in emit_prologue()
336 return 0; in emit_patch()
341 return emit_patch(pprog, func, ip, 0xE8); in emit_call()
348 return emit_patch(pprog, func, ip, 0xE8); in emit_rsb_call()
353 return emit_patch(pprog, func, ip, 0xE9); in emit_jump()
392 ret = 0; in __bpf_arch_text_poke()
417 #define EMIT_LFENCE() EMIT3(0x0F, 0xAE, 0xE8)
425 EMIT2(0xFF, 0xE0 + reg); in emit_indirect_jump()
433 EMIT2(0xFF, 0xE0 + reg); /* jmp *%\reg */ in emit_indirect_jump()
435 EMIT1(0xCC); /* int3 */ in emit_indirect_jump()
448 EMIT1(0xC3); /* ret */ in emit_return()
450 EMIT1(0xCC); /* int3 */ in emit_return()
488 EMIT2(0x89, 0xD2); /* mov edx, edx */ in emit_bpf_tail_call_indirect()
489 EMIT3(0x39, 0x56, /* cmp dword ptr [rsi + 16], edx */ in emit_bpf_tail_call_indirect()
499 EMIT2_off32(0x8B, 0x85, tcc_off); /* mov eax, dword ptr [rbp - tcc_off] */ in emit_bpf_tail_call_indirect()
500 EMIT3(0x83, 0xF8, MAX_TAIL_CALL_CNT); /* cmp eax, MAX_TAIL_CALL_CNT */ in emit_bpf_tail_call_indirect()
504 EMIT3(0x83, 0xC0, 0x01); /* add eax, 1 */ in emit_bpf_tail_call_indirect()
505 EMIT2_off32(0x89, 0x85, tcc_off); /* mov dword ptr [rbp - tcc_off], eax */ in emit_bpf_tail_call_indirect()
508 EMIT4_off32(0x48, 0x8B, 0x8C, 0xD6, /* mov rcx, [rsi + rdx * 8 + offsetof(...)] */ in emit_bpf_tail_call_indirect()
515 EMIT3(0x48, 0x85, 0xC9); /* test rcx,rcx */ in emit_bpf_tail_call_indirect()
522 EMIT1(0x58); /* pop rax */ in emit_bpf_tail_call_indirect()
524 EMIT3_off32(0x48, 0x81, 0xC4, /* add rsp, sd */ in emit_bpf_tail_call_indirect()
528 EMIT4(0x48, 0x8B, 0x49, /* mov rcx, qword ptr [rcx + 32] */ in emit_bpf_tail_call_indirect()
530 EMIT4(0x48, 0x83, 0xC1, /* add rcx, X86_TAIL_CALL_OFFSET */ in emit_bpf_tail_call_indirect()
557 EMIT2_off32(0x8B, 0x85, tcc_off); /* mov eax, dword ptr [rbp - tcc_off] */ in emit_bpf_tail_call_direct()
558 EMIT3(0x83, 0xF8, MAX_TAIL_CALL_CNT); /* cmp eax, MAX_TAIL_CALL_CNT */ in emit_bpf_tail_call_direct()
562 EMIT3(0x83, 0xC0, 0x01); /* add eax, 1 */ in emit_bpf_tail_call_direct()
563 EMIT2_off32(0x89, 0x85, tcc_off); /* mov dword ptr [rbp - tcc_off], eax */ in emit_bpf_tail_call_direct()
574 EMIT1(0x58); /* pop rax */ in emit_bpf_tail_call_direct()
576 EMIT3_off32(0x48, 0x81, 0xC4, round_up(stack_depth, 8)); in emit_bpf_tail_call_direct()
594 for (i = 0; i < prog->aux->size_poke_tab; i++) { in bpf_tail_call_direct_fixup()
612 BUG_ON(ret < 0); in bpf_tail_call_direct_fixup()
617 BUG_ON(ret < 0); in bpf_tail_call_direct_fixup()
634 if (sign_propagate && (s32)imm32 < 0) { in emit_mov_imm32()
636 b1 = add_1mod(0x48, dst_reg); in emit_mov_imm32()
637 b2 = 0xC7; in emit_mov_imm32()
638 b3 = 0xC0; in emit_mov_imm32()
647 if (imm32 == 0) { in emit_mov_imm32()
649 EMIT1(add_2mod(0x40, dst_reg, dst_reg)); in emit_mov_imm32()
650 b2 = 0x31; /* xor */ in emit_mov_imm32()
651 b3 = 0xC0; in emit_mov_imm32()
658 EMIT1(add_1mod(0x40, dst_reg)); in emit_mov_imm32()
659 EMIT1_off32(add_1reg(0xB8, dst_reg), imm32); in emit_mov_imm32()
679 EMIT2(add_1mod(0x48, dst_reg), add_1reg(0xB8, dst_reg)); in emit_mov_imm64()
697 EMIT1(add_2mod(0x40, dst_reg, src_reg)); in emit_mov_reg()
698 EMIT2(0x89, add_2reg(0xC0, dst_reg, src_reg)); in emit_mov_reg()
712 EMIT4(add_2mod(0x48, src_reg, dst_reg), 0x0f, 0xbe, in emit_movsx_reg()
713 add_2reg(0xC0, src_reg, dst_reg)); in emit_movsx_reg()
715 EMIT4(add_2mod(0x48, src_reg, dst_reg), 0x0f, 0xbf, in emit_movsx_reg()
716 add_2reg(0xC0, src_reg, dst_reg)); in emit_movsx_reg()
718 EMIT3(add_2mod(0x48, src_reg, dst_reg), 0x63, in emit_movsx_reg()
719 add_2reg(0xC0, src_reg, dst_reg)); in emit_movsx_reg()
723 EMIT4(add_2mod(0x40, src_reg, dst_reg), 0x0f, 0xbe, in emit_movsx_reg()
724 add_2reg(0xC0, src_reg, dst_reg)); in emit_movsx_reg()
727 EMIT1(add_2mod(0x40, src_reg, dst_reg)); in emit_movsx_reg()
728 EMIT3(add_2mod(0x0f, src_reg, dst_reg), 0xbf, in emit_movsx_reg()
729 add_2reg(0xC0, src_reg, dst_reg)); in emit_movsx_reg()
744 * If off == 0 we could skip this and save one extra byte, but in emit_insn_suffix()
748 EMIT2(add_2reg(0x40, ptr_reg, val_reg), off); in emit_insn_suffix()
751 EMIT1_off32(add_2reg(0x80, ptr_reg, val_reg), off); in emit_insn_suffix()
764 EMIT1(add_2mod(0x48, dst_reg, src_reg)); in maybe_emit_mod()
766 EMIT1(add_2mod(0x40, dst_reg, src_reg)); in maybe_emit_mod()
778 EMIT1(add_1mod(0x48, reg)); in maybe_emit_1mod()
780 EMIT1(add_1mod(0x40, reg)); in maybe_emit_1mod()
792 EMIT3(add_2mod(0x48, src_reg, dst_reg), 0x0F, 0xB6); in emit_ldx()
796 EMIT3(add_2mod(0x48, src_reg, dst_reg), 0x0F, 0xB7); in emit_ldx()
799 /* Emit 'mov eax, dword ptr [rax+0x14]' */ in emit_ldx()
801 EMIT2(add_2mod(0x40, src_reg, dst_reg), 0x8B); in emit_ldx()
803 EMIT1(0x8B); in emit_ldx()
806 /* Emit 'mov rax, qword ptr [rax+0x14]' */ in emit_ldx()
807 EMIT2(add_2mod(0x48, src_reg, dst_reg), 0x8B); in emit_ldx()
822 EMIT3(add_2mod(0x48, src_reg, dst_reg), 0x0F, 0xBE); in emit_ldsx()
826 EMIT3(add_2mod(0x48, src_reg, dst_reg), 0x0F, 0xBF); in emit_ldsx()
829 /* Emit 'movsx rax, dword ptr [rax+0x14]' */ in emit_ldsx()
830 EMIT2(add_2mod(0x48, src_reg, dst_reg), 0x63); in emit_ldsx()
847 EMIT2(add_2mod(0x40, dst_reg, src_reg), 0x88); in emit_stx()
849 EMIT1(0x88); in emit_stx()
853 EMIT3(0x66, add_2mod(0x40, dst_reg, src_reg), 0x89); in emit_stx()
855 EMIT2(0x66, 0x89); in emit_stx()
859 EMIT2(add_2mod(0x40, dst_reg, src_reg), 0x89); in emit_stx()
861 EMIT1(0x89); in emit_stx()
864 EMIT2(add_2mod(0x48, dst_reg, src_reg), 0x89); in emit_stx()
876 EMIT1(0xF0); /* lock prefix */ in emit_atomic()
891 EMIT2(0x0F, 0xC1); in emit_atomic()
895 EMIT1(0x87); in emit_atomic()
899 EMIT2(0x0F, 0xB1); in emit_atomic()
909 return 0; in emit_atomic()
917 *(unsigned long *)((void *)regs + reg) = 0; in ex_handler_bpf()
918 regs->ip += x->fixup & 0xff; in ex_handler_bpf()
931 regs_used[0] = true; in detect_reg_usage()
946 while (len > 0) { in emit_nops()
952 for (i = 0; i < noplen; i++) in emit_nops()
965 * m: opcode map select, encoding escape bytes e.g. 0x0f38
969 * pp: opcode prefix (none, 0x66, 0xf2 or 0xf3)
975 const u8 b0 = 0xc4; /* first byte of 3-byte VEX prefix */ in emit_3vex()
987 * 7 0 in emit_3vex()
992 b1 = (!r << 7) | (!x << 6) | (!b << 5) | (m & 0x1f); in emit_3vex()
996 * 7 0 in emit_3vex()
1001 b2 = (w << 7) | ((~vvvv & 0xf) << 3) | (l << 2) | (pp & 3); in emit_3vex()
1012 u8 m = 2; /* escape code 0f38 */ in emit_shiftx()
1015 EMIT2(0xf7, add_2reg(0xC0, dst_reg, dst_reg)); in emit_shiftx()
1023 EMIT3_off32(0x48, 0x8B, 0x85, -round_up(stack, 8) - 8)
1035 int i, excnt = 0; in do_jit()
1036 int ilen, proglen = 0; in do_jit()
1048 bpf_prog->aux->func_idx != 0); in do_jit()
1055 addrs[0] = proglen; in do_jit()
1062 u8 b2 = 0, b3 = 0; in do_jit()
1085 EMIT2(b2, add_2reg(0xC0, dst_reg, src_reg)); in do_jit()
1090 if (insn->off == 0) in do_jit()
1105 EMIT2(0xF7, add_1reg(0xD8, dst_reg)); in do_jit()
1127 b3 = 0xC0; in do_jit()
1128 b2 = 0x05; in do_jit()
1131 b3 = 0xE8; in do_jit()
1132 b2 = 0x2D; in do_jit()
1135 b3 = 0xE0; in do_jit()
1136 b2 = 0x25; in do_jit()
1139 b3 = 0xC8; in do_jit()
1140 b2 = 0x0D; in do_jit()
1143 b3 = 0xF0; in do_jit()
1144 b2 = 0x35; in do_jit()
1149 EMIT3(0x83, add_1reg(b3, dst_reg), imm32); in do_jit()
1153 EMIT2_off32(0x81, add_1reg(b3, dst_reg), imm32); in do_jit()
1163 emit_mov_imm64(&prog, dst_reg, insn[1].imm, insn[0].imm); in do_jit()
1180 EMIT1(0x50); /* push rax */ in do_jit()
1182 EMIT1(0x52); /* push rdx */ in do_jit()
1193 EMIT3_off32(0x49, 0xC7, 0xC3, imm32); in do_jit()
1201 if (insn->off == 0) { in do_jit()
1206 EMIT2(0x31, 0xd2); in do_jit()
1210 EMIT2(0xF7, add_1reg(0xF0, src_reg)); in do_jit()
1213 EMIT1(0x99); /* cdq */ in do_jit()
1215 EMIT2(0x48, 0x99); /* cqo */ in do_jit()
1219 EMIT2(0xF7, add_1reg(0xF8, src_reg)); in do_jit()
1232 EMIT1(0x5A); /* pop rdx */ in do_jit()
1234 EMIT1(0x58); /* pop rax */ in do_jit()
1245 EMIT3(0x6B, add_2reg(0xC0, dst_reg, dst_reg), in do_jit()
1249 EMIT2_off32(0x69, in do_jit()
1250 add_2reg(0xC0, dst_reg, dst_reg), in do_jit()
1260 EMIT3(0x0F, 0xAF, add_2reg(0xC0, src_reg, dst_reg)); in do_jit()
1275 EMIT2(0xD1, add_1reg(b3, dst_reg)); in do_jit()
1277 EMIT3(0xC1, add_1reg(b3, dst_reg), imm32); in do_jit()
1294 op = 1; /* prefix 0x66 */ in do_jit()
1297 op = 3; /* prefix 0xf2 */ in do_jit()
1300 op = 2; /* prefix 0xf3 */ in do_jit()
1316 EMIT1(0x51); /* push rcx */ in do_jit()
1327 EMIT2(0xD3, add_1reg(b3, dst_reg)); in do_jit()
1334 EMIT1(0x59); /* pop rcx */ in do_jit()
1344 EMIT1(0x66); in do_jit()
1346 EMIT1(0x41); in do_jit()
1347 EMIT3(0xC1, add_1reg(0xC8, dst_reg), 8); in do_jit()
1351 EMIT3(0x45, 0x0F, 0xB7); in do_jit()
1353 EMIT2(0x0F, 0xB7); in do_jit()
1354 EMIT1(add_2reg(0xC0, dst_reg, dst_reg)); in do_jit()
1359 EMIT2(0x41, 0x0F); in do_jit()
1361 EMIT1(0x0F); in do_jit()
1362 EMIT1(add_1reg(0xC8, dst_reg)); in do_jit()
1366 EMIT3(add_1mod(0x48, dst_reg), 0x0F, in do_jit()
1367 add_1reg(0xC8, dst_reg)); in do_jit()
1380 EMIT3(0x45, 0x0F, 0xB7); in do_jit()
1382 EMIT2(0x0F, 0xB7); in do_jit()
1383 EMIT1(add_2reg(0xC0, dst_reg, dst_reg)); in do_jit()
1388 EMIT1(0x45); in do_jit()
1389 EMIT2(0x89, add_2reg(0xC0, dst_reg, dst_reg)); in do_jit()
1405 EMIT2(0x41, 0xC6); in do_jit()
1407 EMIT1(0xC6); in do_jit()
1411 EMIT3(0x66, 0x41, 0xC7); in do_jit()
1413 EMIT2(0x66, 0xC7); in do_jit()
1417 EMIT2(0x41, 0xC7); in do_jit()
1419 EMIT1(0xC7); in do_jit()
1422 EMIT2(add_1mod(0x48, dst_reg), 0xC7); in do_jit()
1425 EMIT2(add_1reg(0x40, dst_reg), insn->off); in do_jit()
1427 EMIT1_off32(add_1reg(0x80, dst_reg), insn->off); in do_jit()
1472 insn_off = 0; in do_jit()
1475 EMIT2(add_1mod(0x48, AUX_REG), add_1reg(0xB8, AUX_REG)); in do_jit()
1482 EMIT2_off32(0x81, add_1reg(0xC0, src_reg), insn->off); in do_jit()
1487 EMIT2(0x39, add_2reg(0xC0, src_reg, AUX_REG)); in do_jit()
1490 EMIT2(X86_JAE, 0); in do_jit()
1494 emit_mov_imm32(&prog, false, dst_reg, 0); in do_jit()
1496 EMIT2(0xEB, 0); in do_jit()
1525 EMIT2_off32(0x81, add_1reg(0xE8, src_reg), insn->off); in do_jit()
1558 * End result: x86 insn "mov rbx, qword ptr [rax+0x14]" in do_jit()
1598 add_2reg(0xC0, AUX_REG, real_src_reg)); in do_jit()
1683 EMIT2(0x39, add_2reg(0xC0, dst_reg, src_reg)); in do_jit()
1691 EMIT2(0x85, add_2reg(0xC0, dst_reg, src_reg)); in do_jit()
1699 EMIT2_off32(0xF7, add_1reg(0xC0, dst_reg), imm32); in do_jit()
1723 if (imm32 == 0) { in do_jit()
1726 EMIT2(0x85, add_2reg(0xC0, dst_reg, dst_reg)); in do_jit()
1735 EMIT3(0x83, add_1reg(0xF8, dst_reg), imm32); in do_jit()
1737 EMIT2_off32(0x81, add_1reg(0xF8, dst_reg), imm32); in do_jit()
1792 * "nops" is 0. in do_jit()
1802 if (nops != 0 && nops != 4) { in do_jit()
1811 EMIT2_off32(0x0F, jmp_cond + 0x10, jmp_offset); in do_jit()
1848 * to pad any extra byte (0 byte). in do_jit()
1856 if (nops != 0 && nops != 2 && nops != 5) { in do_jit()
1873 * jmp, there is nothing to pad (0 byte). in do_jit()
1881 if (nops != 0 && nops != 3) { in do_jit()
1888 EMIT2(0xEB, jmp_offset); in do_jit()
1890 EMIT1_off32(0xE9, jmp_offset); in do_jit()
1906 EMIT1(0xC9); /* leave */ in do_jit()
1993 /* mov DWORD PTR [rbp + off], 0 */ in clean_stack_garbage()
1995 EMIT2_off32(0xC7, 0x85, off); in clean_stack_garbage()
1997 EMIT3(0xC7, 0x45, off); in clean_stack_garbage()
1998 EMIT(0, 4); in clean_stack_garbage()
2006 int i, arg_regs, nr_used_regs = 0; in get_nr_used_regs()
2008 for (i = 0; i < min_t(int, m->nr_args, MAX_BPF_FUNC_ARGS); i++) { in get_nr_used_regs()
2023 int arg_regs, first_off = 0, nr_regs = 0, nr_stack_slots = 0; in save_args()
2028 * mov QWORD PTR [rbp-0x10],rdi in save_args()
2029 * mov QWORD PTR [rbp-0x8],rsi in save_args()
2031 for (i = 0; i < min_t(int, m->nr_args, MAX_BPF_FUNC_ARGS); i++) { in save_args()
2062 for (j = 0; j < arg_regs; j++) { in save_args()
2064 nr_stack_slots * 8 + 0x18); in save_args()
2084 for (j = 0; j < arg_regs; j++) { in save_args()
2100 int i, j, arg_regs, nr_regs = 0; in restore_regs()
2104 * EMIT4(0x48, 0x8B, 0x7D, 0xF0); mov rdi,QWORD PTR [rbp-0x10] in restore_regs()
2105 * EMIT4(0x48, 0x8B, 0x75, 0xF8); mov rsi,QWORD PTR [rbp-0x8] in restore_regs()
2109 for (i = 0; i < min_t(int, m->nr_args, MAX_BPF_FUNC_ARGS); i++) { in restore_regs()
2112 for (j = 0; j < arg_regs; j++) { in restore_regs()
2155 EMIT3_off32(0x48, 0x8D, 0xB5, -run_ctx_off); in invoke_bpf_prog()
2157 EMIT4(0x48, 0x8D, 0x75, -run_ctx_off); in invoke_bpf_prog()
2164 /* if (__bpf_prog_enter*(prog) == 0) in invoke_bpf_prog()
2167 EMIT3(0x48, 0x85, 0xC0); /* test rax,rax */ in invoke_bpf_prog()
2174 EMIT3_off32(0x48, 0x8D, 0xBD, -stack_size); in invoke_bpf_prog()
2176 EMIT4(0x48, 0x8D, 0x7D, -stack_size); in invoke_bpf_prog()
2198 jmp_insn[0] = X86_JE; in invoke_bpf_prog()
2207 EMIT3_off32(0x48, 0x8D, 0x95, -run_ctx_off); in invoke_bpf_prog()
2209 EMIT4(0x48, 0x8D, 0x55, -run_ctx_off); in invoke_bpf_prog()
2214 return 0; in invoke_bpf_prog()
2238 EMIT2_off32(0x0F, jmp_cond + 0x10, offset); in emit_cond_near_jump()
2240 return 0; in emit_cond_near_jump()
2250 for (i = 0; i < tl->nr_links; i++) { in invoke_bpf()
2256 return 0; in invoke_bpf()
2267 * Set this to 0 to avoid confusing the program. in invoke_bpf_mod_ret()
2269 emit_mov_imm32(&prog, false, BPF_REG_0, 0); in invoke_bpf_mod_ret()
2271 for (i = 0; i < tl->nr_links; i++) { in invoke_bpf_mod_ret()
2276 * if (*(u64 *)(rbp - 8) != 0) in invoke_bpf_mod_ret()
2279 /* cmp QWORD PTR [rbp - 0x8], 0x0 */ in invoke_bpf_mod_ret()
2280 EMIT4(0x48, 0x83, 0x7d, 0xf8); EMIT1(0x00); in invoke_bpf_mod_ret()
2292 return 0; in invoke_bpf_mod_ret()
2360 int i, ret, nr_regs = m->nr_args, stack_size = 0; in arch_prepare_bpf_trampoline()
2371 for (i = 0; i < m->nr_args; i++) in arch_prepare_bpf_trampoline()
2384 * RBP + 0 [ RBP ] in arch_prepare_bpf_trampoline()
2428 stack_size += (sizeof(struct bpf_tramp_run_ctx) + 7) & ~0x7; in arch_prepare_bpf_trampoline()
2440 stack_size += (stack_size % 16) ? 0 : 8; in arch_prepare_bpf_trampoline()
2462 EMIT1(0x55); /* push rbp */ in arch_prepare_bpf_trampoline()
2463 EMIT3(0x48, 0x89, 0xE5); /* mov rbp, rsp */ in arch_prepare_bpf_trampoline()
2466 EMIT3_off32(0x48, 0x81, 0xEC, stack_size); in arch_prepare_bpf_trampoline()
2469 EMIT4(0x48, 0x83, 0xEC, stack_size); in arch_prepare_bpf_trampoline()
2471 EMIT1(0x50); /* push rax */ in arch_prepare_bpf_trampoline()
2479 emit_mov_imm64(&prog, BPF_REG_0, 0, (u32) nr_regs); in arch_prepare_bpf_trampoline()
2532 EMIT2(0xff, 0xd3); /* call *rbx */ in arch_prepare_bpf_trampoline()
2557 for (i = 0; i < fmod_ret->nr_links; i++) in arch_prepare_bpf_trampoline()
2594 EMIT1(0xC9); /* leave */ in arch_prepare_bpf_trampoline()
2597 EMIT4(0x48, 0x83, 0xC4, 8); /* add rsp, 8 */ in arch_prepare_bpf_trampoline()
2621 EMIT1(add_1mod(0x48, BPF_REG_3)); /* cmp rdx,func */ in emit_bpf_dispatcher()
2624 EMIT2_off32(0x81, add_1reg(0xF8, BPF_REG_3), in emit_bpf_dispatcher()
2635 return 0; in emit_bpf_dispatcher()
2642 EMIT1(add_1mod(0x48, BPF_REG_3)); /* cmp rdx,func */ in emit_bpf_dispatcher()
2645 EMIT2_off32(0x81, add_1reg(0xF8, BPF_REG_3), progs[a + pivot]); in emit_bpf_dispatcher()
2650 EMIT2_off32(0x0F, X86_JG + 0x10, 0); in emit_bpf_dispatcher()
2652 EMIT2(X86_JG, 0); in emit_bpf_dispatcher()
2676 return 0; in emit_bpf_dispatcher()
2688 return 0; in cmp_ips()
2695 sort(funcs, num_funcs, sizeof(funcs[0]), cmp_ips, NULL); in arch_prepare_bpf_dispatcher()
2696 return emit_bpf_dispatcher(&prog, 0, num_funcs - 1, funcs, image, buf); in arch_prepare_bpf_dispatcher()
2717 int proglen, oldproglen = 0; in bpf_int_jit_compile()
2774 for (proglen = 0, i = 0; i <= prog->len; i++) { in bpf_int_jit_compile()
2787 for (pass = 0; pass < MAX_PASSES || image; pass++) { in bpf_int_jit_compile()
2791 if (proglen <= 0) { in bpf_int_jit_compile()
2803 prog->jited = 0; in bpf_int_jit_compile()
2804 prog->jited_len = 0; in bpf_int_jit_compile()
2952 BUG_ON(ret < 0); in bpf_arch_poke_desc_update()
2958 BUG_ON(ret < 0); in bpf_arch_poke_desc_update()
2965 BUG_ON(ret < 0); in bpf_arch_poke_desc_update()
2975 BUG_ON(ret < 0); in bpf_arch_poke_desc_update()