Lines Matching +full:0 +full:x48
34 do { prog = emit_code(prog, bytes, len); } while (0)
41 do { EMIT1(b1); EMIT4(b2, b3, b4, b5); } while (0)
44 do { EMIT1(b1); EMIT(off, 4); } while (0)
46 do { EMIT2(b1, b2); EMIT(off, 4); } while (0)
48 do { EMIT3(b1, b2, b3); EMIT(off, 4); } while (0)
50 do { EMIT4(b1, b2, b3, b4); EMIT(off, 4); } while (0)
70 * 211: 74 7d je 0x290
71 * 213: 48 8b 77 00 mov rsi,QWORD PTR [rdi+0x0]
74 * 28c: 74 17 je 0x2a5
75 * 28e: e9 7f ff ff ff jmp 0x212
76 * 293: bf 03 00 00 00 mov edi,0x3
77 * Note that insn at 0x211 is 2-byte cond jump insn for offset 0x7d (-125)
78 * and insn at 0x28e is 5-byte jmp insn with offset -129.
83 * 211: 0f 84 80 00 00 00 je 0x297
84 * 217: 48 8b 77 00 mov rsi,QWORD PTR [rdi+0x0]
87 * 290: 74 1a je 0x2ac
88 * 292: eb 84 jmp 0x218
89 * 294: bf 03 00 00 00 mov edi,0x3
90 * Note that insn at 0x211 is 6-byte cond jump insn now since its offset
91 * becomes 0x80 based on previous round (0x293 - 0x213 = 0x80).
92 * At the same time, insn at 0x292 is a 2-byte insn since its offset is
99 * cycle in the above. In the above example je offset <= 0x7c should work.
101 * For other cases, je <-> je needs offset <= 0x7b to avoid no convergence
102 * issue. For jmp <-> je and jmp <-> jmp cases, jmp offset <= 0x7c should
106 * to maximum 123 (0x7b). This way, the jit pass can eventually converge.
127 EMIT3(add_2mod(0x48, DST, SRC), 0x89, add_2reg(0xC0, DST, SRC)); \
128 } while (0)
141 return 0; in bpf_size_to_x86_bytes()
146 * Add 0x10 (and an extra 0x0f) to generate far jumps (. + s32)
148 #define X86_JB 0x72
149 #define X86_JAE 0x73
150 #define X86_JE 0x74
151 #define X86_JNE 0x75
152 #define X86_JBE 0x76
153 #define X86_JA 0x77
154 #define X86_JL 0x7C
155 #define X86_JGE 0x7D
156 #define X86_JLE 0x7E
157 #define X86_JG 0x7F
174 [BPF_REG_0] = 0, /* RAX */
179 [BPF_REG_5] = 0, /* R8 */
268 [BPF_ADD] = 0x01,
269 [BPF_SUB] = 0x29,
270 [BPF_AND] = 0x21,
271 [BPF_OR] = 0x09,
272 [BPF_XOR] = 0x31,
273 [BPF_LSH] = 0xE0,
274 [BPF_RSH] = 0xE8,
275 [BPF_ARSH] = 0xF8,
281 memset(area, 0xcc, size); in jit_fill_hole()
286 return IS_ERR_OR_NULL(text_poke_set(dst, 0xcc, len)); in bpf_arch_text_invalidate()
314 if (callee_regs_used[0]) in push_callee_regs()
315 EMIT1(0x53); /* push rbx */ in push_callee_regs()
317 EMIT2(0x41, 0x55); /* push r13 */ in push_callee_regs()
319 EMIT2(0x41, 0x56); /* push r14 */ in push_callee_regs()
321 EMIT2(0x41, 0x57); /* push r15 */ in push_callee_regs()
330 EMIT2(0x41, 0x5F); /* pop r15 */ in pop_callee_regs()
332 EMIT2(0x41, 0x5E); /* pop r14 */ in pop_callee_regs()
334 EMIT2(0x41, 0x5D); /* pop r13 */ in pop_callee_regs()
335 if (callee_regs_used[0]) in pop_callee_regs()
336 EMIT1(0x5B); /* pop rbx */ in pop_callee_regs()
358 EMIT2(0x31, 0xC0); /* xor eax, eax */ in emit_prologue()
360 EMIT2(0x66, 0x90); /* nop2 */ in emit_prologue()
362 EMIT1(0x55); /* push rbp */ in emit_prologue()
363 EMIT3(0x48, 0x89, 0xE5); /* mov rbp, rsp */ in emit_prologue()
370 EMIT3_off32(0x48, 0x81, 0xEC, round_up(stack_depth, 8)); in emit_prologue()
372 EMIT1(0x50); /* push rax */ in emit_prologue()
388 return 0; in emit_patch()
393 return emit_patch(pprog, func, ip, 0xE8); in emit_call()
400 return emit_patch(pprog, func, ip, 0xE8); in emit_rsb_call()
405 return emit_patch(pprog, func, ip, 0xE9); in emit_jump()
444 ret = 0; in __bpf_arch_text_poke()
469 #define EMIT_LFENCE() EMIT3(0x0F, 0xAE, 0xE8)
481 EMIT2(0xFF, 0xE0 + reg); in emit_indirect_jump()
489 EMIT2(0xFF, 0xE0 + reg); /* jmp *%\reg */ in emit_indirect_jump()
491 EMIT1(0xCC); /* int3 */ in emit_indirect_jump()
504 EMIT1(0xC3); /* ret */ in emit_return()
506 EMIT1(0xCC); /* int3 */ in emit_return()
544 EMIT2(0x89, 0xD2); /* mov edx, edx */ in emit_bpf_tail_call_indirect()
545 EMIT3(0x39, 0x56, /* cmp dword ptr [rsi + 16], edx */ in emit_bpf_tail_call_indirect()
555 EMIT2_off32(0x8B, 0x85, tcc_off); /* mov eax, dword ptr [rbp - tcc_off] */ in emit_bpf_tail_call_indirect()
556 EMIT3(0x83, 0xF8, MAX_TAIL_CALL_CNT); /* cmp eax, MAX_TAIL_CALL_CNT */ in emit_bpf_tail_call_indirect()
560 EMIT3(0x83, 0xC0, 0x01); /* add eax, 1 */ in emit_bpf_tail_call_indirect()
561 EMIT2_off32(0x89, 0x85, tcc_off); /* mov dword ptr [rbp - tcc_off], eax */ in emit_bpf_tail_call_indirect()
564 EMIT4_off32(0x48, 0x8B, 0x8C, 0xD6, /* mov rcx, [rsi + rdx * 8 + offsetof(...)] */ in emit_bpf_tail_call_indirect()
571 EMIT3(0x48, 0x85, 0xC9); /* test rcx,rcx */ in emit_bpf_tail_call_indirect()
578 EMIT1(0x58); /* pop rax */ in emit_bpf_tail_call_indirect()
580 EMIT3_off32(0x48, 0x81, 0xC4, /* add rsp, sd */ in emit_bpf_tail_call_indirect()
584 EMIT4(0x48, 0x8B, 0x49, /* mov rcx, qword ptr [rcx + 32] */ in emit_bpf_tail_call_indirect()
586 EMIT4(0x48, 0x83, 0xC1, /* add rcx, X86_TAIL_CALL_OFFSET */ in emit_bpf_tail_call_indirect()
613 EMIT2_off32(0x8B, 0x85, tcc_off); /* mov eax, dword ptr [rbp - tcc_off] */ in emit_bpf_tail_call_direct()
614 EMIT3(0x83, 0xF8, MAX_TAIL_CALL_CNT); /* cmp eax, MAX_TAIL_CALL_CNT */ in emit_bpf_tail_call_direct()
618 EMIT3(0x83, 0xC0, 0x01); /* add eax, 1 */ in emit_bpf_tail_call_direct()
619 EMIT2_off32(0x89, 0x85, tcc_off); /* mov dword ptr [rbp - tcc_off], eax */ in emit_bpf_tail_call_direct()
630 EMIT1(0x58); /* pop rax */ in emit_bpf_tail_call_direct()
632 EMIT3_off32(0x48, 0x81, 0xC4, round_up(stack_depth, 8)); in emit_bpf_tail_call_direct()
650 for (i = 0; i < prog->aux->size_poke_tab; i++) { in bpf_tail_call_direct_fixup()
668 BUG_ON(ret < 0); in bpf_tail_call_direct_fixup()
673 BUG_ON(ret < 0); in bpf_tail_call_direct_fixup()
690 if (sign_propagate && (s32)imm32 < 0) { in emit_mov_imm32()
692 b1 = add_1mod(0x48, dst_reg); in emit_mov_imm32()
693 b2 = 0xC7; in emit_mov_imm32()
694 b3 = 0xC0; in emit_mov_imm32()
703 if (imm32 == 0) { in emit_mov_imm32()
705 EMIT1(add_2mod(0x40, dst_reg, dst_reg)); in emit_mov_imm32()
706 b2 = 0x31; /* xor */ in emit_mov_imm32()
707 b3 = 0xC0; in emit_mov_imm32()
714 EMIT1(add_1mod(0x40, dst_reg)); in emit_mov_imm32()
715 EMIT1_off32(add_1reg(0xB8, dst_reg), imm32); in emit_mov_imm32()
735 EMIT2(add_1mod(0x48, dst_reg), add_1reg(0xB8, dst_reg)); in emit_mov_imm64()
753 EMIT1(add_2mod(0x40, dst_reg, src_reg)); in emit_mov_reg()
754 EMIT2(0x89, add_2reg(0xC0, dst_reg, src_reg)); in emit_mov_reg()
768 EMIT4(add_2mod(0x48, src_reg, dst_reg), 0x0f, 0xbe, in emit_movsx_reg()
769 add_2reg(0xC0, src_reg, dst_reg)); in emit_movsx_reg()
771 EMIT4(add_2mod(0x48, src_reg, dst_reg), 0x0f, 0xbf, in emit_movsx_reg()
772 add_2reg(0xC0, src_reg, dst_reg)); in emit_movsx_reg()
774 EMIT3(add_2mod(0x48, src_reg, dst_reg), 0x63, in emit_movsx_reg()
775 add_2reg(0xC0, src_reg, dst_reg)); in emit_movsx_reg()
779 EMIT4(add_2mod(0x40, src_reg, dst_reg), 0x0f, 0xbe, in emit_movsx_reg()
780 add_2reg(0xC0, src_reg, dst_reg)); in emit_movsx_reg()
783 EMIT1(add_2mod(0x40, src_reg, dst_reg)); in emit_movsx_reg()
784 EMIT3(add_2mod(0x0f, src_reg, dst_reg), 0xbf, in emit_movsx_reg()
785 add_2reg(0xC0, src_reg, dst_reg)); in emit_movsx_reg()
800 * If off == 0 we could skip this and save one extra byte, but in emit_insn_suffix()
804 EMIT2(add_2reg(0x40, ptr_reg, val_reg), off); in emit_insn_suffix()
807 EMIT1_off32(add_2reg(0x80, ptr_reg, val_reg), off); in emit_insn_suffix()
820 EMIT1(add_2mod(0x48, dst_reg, src_reg)); in maybe_emit_mod()
822 EMIT1(add_2mod(0x40, dst_reg, src_reg)); in maybe_emit_mod()
834 EMIT1(add_1mod(0x48, reg)); in maybe_emit_1mod()
836 EMIT1(add_1mod(0x40, reg)); in maybe_emit_1mod()
848 EMIT3(add_2mod(0x48, src_reg, dst_reg), 0x0F, 0xB6); in emit_ldx()
852 EMIT3(add_2mod(0x48, src_reg, dst_reg), 0x0F, 0xB7); in emit_ldx()
855 /* Emit 'mov eax, dword ptr [rax+0x14]' */ in emit_ldx()
857 EMIT2(add_2mod(0x40, src_reg, dst_reg), 0x8B); in emit_ldx()
859 EMIT1(0x8B); in emit_ldx()
862 /* Emit 'mov rax, qword ptr [rax+0x14]' */ in emit_ldx()
863 EMIT2(add_2mod(0x48, src_reg, dst_reg), 0x8B); in emit_ldx()
878 EMIT3(add_2mod(0x48, src_reg, dst_reg), 0x0F, 0xBE); in emit_ldsx()
882 EMIT3(add_2mod(0x48, src_reg, dst_reg), 0x0F, 0xBF); in emit_ldsx()
885 /* Emit 'movsx rax, dword ptr [rax+0x14]' */ in emit_ldsx()
886 EMIT2(add_2mod(0x48, src_reg, dst_reg), 0x63); in emit_ldsx()
903 EMIT2(add_2mod(0x40, dst_reg, src_reg), 0x88); in emit_stx()
905 EMIT1(0x88); in emit_stx()
909 EMIT3(0x66, add_2mod(0x40, dst_reg, src_reg), 0x89); in emit_stx()
911 EMIT2(0x66, 0x89); in emit_stx()
915 EMIT2(add_2mod(0x40, dst_reg, src_reg), 0x89); in emit_stx()
917 EMIT1(0x89); in emit_stx()
920 EMIT2(add_2mod(0x48, dst_reg, src_reg), 0x89); in emit_stx()
932 EMIT1(0xF0); /* lock prefix */ in emit_atomic()
947 EMIT2(0x0F, 0xC1); in emit_atomic()
951 EMIT1(0x87); in emit_atomic()
955 EMIT2(0x0F, 0xB1); in emit_atomic()
965 return 0; in emit_atomic()
973 *(unsigned long *)((void *)regs + reg) = 0; in ex_handler_bpf()
974 regs->ip += x->fixup & 0xff; in ex_handler_bpf()
987 regs_used[0] = true; in detect_reg_usage()
1002 while (len > 0) { in emit_nops()
1008 for (i = 0; i < noplen; i++) in emit_nops()
1021 * m: opcode map select, encoding escape bytes e.g. 0x0f38
1025 * pp: opcode prefix (none, 0x66, 0xf2 or 0xf3)
1031 const u8 b0 = 0xc4; /* first byte of 3-byte VEX prefix */ in emit_3vex()
1043 * 7 0 in emit_3vex()
1048 b1 = (!r << 7) | (!x << 6) | (!b << 5) | (m & 0x1f); in emit_3vex()
1052 * 7 0 in emit_3vex()
1057 b2 = (w << 7) | ((~vvvv & 0xf) << 3) | (l << 2) | (pp & 3); in emit_3vex()
1068 u8 m = 2; /* escape code 0f38 */ in emit_shiftx()
1071 EMIT2(0xf7, add_2reg(0xC0, dst_reg, dst_reg)); in emit_shiftx()
1079 EMIT3_off32(0x48, 0x8B, 0x85, -round_up(stack, 8) - 8)
1089 EMIT1(0x50); /* push rax */ in emit_spectre_bhb_barrier()
1090 EMIT1(0x51); /* push rcx */ in emit_spectre_bhb_barrier()
1098 EMIT1(0x59); /* pop rcx */ in emit_spectre_bhb_barrier()
1099 EMIT1(0x58); /* pop rax */ in emit_spectre_bhb_barrier()
1117 EMIT5(0xF3, 0x48, 0x0F, 0x1E, 0xF8); /* ibhf */ in emit_spectre_bhb_barrier()
1120 return 0; in emit_spectre_bhb_barrier()
1133 int i, excnt = 0; in do_jit()
1134 int ilen, proglen = 0; in do_jit()
1146 bpf_prog->aux->func_idx != 0); in do_jit()
1153 addrs[0] = proglen; in do_jit()
1160 u8 b2 = 0, b3 = 0; in do_jit()
1183 EMIT2(b2, add_2reg(0xC0, dst_reg, src_reg)); in do_jit()
1188 if (insn->off == 0) in do_jit()
1203 EMIT2(0xF7, add_1reg(0xD8, dst_reg)); in do_jit()
1225 b3 = 0xC0; in do_jit()
1226 b2 = 0x05; in do_jit()
1229 b3 = 0xE8; in do_jit()
1230 b2 = 0x2D; in do_jit()
1233 b3 = 0xE0; in do_jit()
1234 b2 = 0x25; in do_jit()
1237 b3 = 0xC8; in do_jit()
1238 b2 = 0x0D; in do_jit()
1241 b3 = 0xF0; in do_jit()
1242 b2 = 0x35; in do_jit()
1247 EMIT3(0x83, add_1reg(b3, dst_reg), imm32); in do_jit()
1251 EMIT2_off32(0x81, add_1reg(b3, dst_reg), imm32); in do_jit()
1261 emit_mov_imm64(&prog, dst_reg, insn[1].imm, insn[0].imm); in do_jit()
1278 EMIT1(0x50); /* push rax */ in do_jit()
1280 EMIT1(0x52); /* push rdx */ in do_jit()
1291 EMIT3_off32(0x49, 0xC7, 0xC3, imm32); in do_jit()
1299 if (insn->off == 0) { in do_jit()
1304 EMIT2(0x31, 0xd2); in do_jit()
1308 EMIT2(0xF7, add_1reg(0xF0, src_reg)); in do_jit()
1311 EMIT1(0x99); /* cdq */ in do_jit()
1313 EMIT2(0x48, 0x99); /* cqo */ in do_jit()
1317 EMIT2(0xF7, add_1reg(0xF8, src_reg)); in do_jit()
1330 EMIT1(0x5A); /* pop rdx */ in do_jit()
1332 EMIT1(0x58); /* pop rax */ in do_jit()
1343 EMIT3(0x6B, add_2reg(0xC0, dst_reg, dst_reg), in do_jit()
1347 EMIT2_off32(0x69, in do_jit()
1348 add_2reg(0xC0, dst_reg, dst_reg), in do_jit()
1358 EMIT3(0x0F, 0xAF, add_2reg(0xC0, src_reg, dst_reg)); in do_jit()
1373 EMIT2(0xD1, add_1reg(b3, dst_reg)); in do_jit()
1375 EMIT3(0xC1, add_1reg(b3, dst_reg), imm32); in do_jit()
1392 op = 1; /* prefix 0x66 */ in do_jit()
1395 op = 3; /* prefix 0xf2 */ in do_jit()
1398 op = 2; /* prefix 0xf3 */ in do_jit()
1414 EMIT1(0x51); /* push rcx */ in do_jit()
1425 EMIT2(0xD3, add_1reg(b3, dst_reg)); in do_jit()
1432 EMIT1(0x59); /* pop rcx */ in do_jit()
1442 EMIT1(0x66); in do_jit()
1444 EMIT1(0x41); in do_jit()
1445 EMIT3(0xC1, add_1reg(0xC8, dst_reg), 8); in do_jit()
1449 EMIT3(0x45, 0x0F, 0xB7); in do_jit()
1451 EMIT2(0x0F, 0xB7); in do_jit()
1452 EMIT1(add_2reg(0xC0, dst_reg, dst_reg)); in do_jit()
1457 EMIT2(0x41, 0x0F); in do_jit()
1459 EMIT1(0x0F); in do_jit()
1460 EMIT1(add_1reg(0xC8, dst_reg)); in do_jit()
1464 EMIT3(add_1mod(0x48, dst_reg), 0x0F, in do_jit()
1465 add_1reg(0xC8, dst_reg)); in do_jit()
1478 EMIT3(0x45, 0x0F, 0xB7); in do_jit()
1480 EMIT2(0x0F, 0xB7); in do_jit()
1481 EMIT1(add_2reg(0xC0, dst_reg, dst_reg)); in do_jit()
1486 EMIT1(0x45); in do_jit()
1487 EMIT2(0x89, add_2reg(0xC0, dst_reg, dst_reg)); in do_jit()
1503 EMIT2(0x41, 0xC6); in do_jit()
1505 EMIT1(0xC6); in do_jit()
1509 EMIT3(0x66, 0x41, 0xC7); in do_jit()
1511 EMIT2(0x66, 0xC7); in do_jit()
1515 EMIT2(0x41, 0xC7); in do_jit()
1517 EMIT1(0xC7); in do_jit()
1520 EMIT2(add_1mod(0x48, dst_reg), 0xC7); in do_jit()
1523 EMIT2(add_1reg(0x40, dst_reg), insn->off); in do_jit()
1525 EMIT1_off32(add_1reg(0x80, dst_reg), insn->off); in do_jit()
1577 EMIT2_off32(0x81, add_1reg(0xC0, AUX_REG), insn->off); in do_jit()
1582 EMIT2(0x29, add_2reg(0xC0, AUX_REG, BPF_REG_AX)); in do_jit()
1590 EMIT2(0x39, add_2reg(0xC0, AUX_REG, BPF_REG_AX)); in do_jit()
1593 EMIT2(X86_JA, 0); in do_jit()
1597 emit_mov_imm32(&prog, false, dst_reg, 0); in do_jit()
1599 EMIT2(0xEB, 0); in do_jit()
1649 * End result: x86 insn "mov rbx, qword ptr [rax+0x14]" in do_jit()
1689 add_2reg(0xC0, AUX_REG, real_src_reg)); in do_jit()
1774 EMIT2(0x39, add_2reg(0xC0, dst_reg, src_reg)); in do_jit()
1782 EMIT2(0x85, add_2reg(0xC0, dst_reg, src_reg)); in do_jit()
1790 EMIT2_off32(0xF7, add_1reg(0xC0, dst_reg), imm32); in do_jit()
1814 if (imm32 == 0) { in do_jit()
1817 EMIT2(0x85, add_2reg(0xC0, dst_reg, dst_reg)); in do_jit()
1826 EMIT3(0x83, add_1reg(0xF8, dst_reg), imm32); in do_jit()
1828 EMIT2_off32(0x81, add_1reg(0xF8, dst_reg), imm32); in do_jit()
1883 * "nops" is 0. in do_jit()
1893 if (nops != 0 && nops != 4) { in do_jit()
1902 EMIT2_off32(0x0F, jmp_cond + 0x10, jmp_offset); in do_jit()
1939 * to pad any extra byte (0 byte). in do_jit()
1947 if (nops != 0 && nops != 2 && nops != 5) { in do_jit()
1964 * jmp, there is nothing to pad (0 byte). in do_jit()
1972 if (nops != 0 && nops != 3) { in do_jit()
1979 EMIT2(0xEB, jmp_offset); in do_jit()
1981 EMIT1_off32(0xE9, jmp_offset); in do_jit()
2006 EMIT1(0xC9); /* leave */ in do_jit()
2093 /* mov DWORD PTR [rbp + off], 0 */ in clean_stack_garbage()
2095 EMIT2_off32(0xC7, 0x85, off); in clean_stack_garbage()
2097 EMIT3(0xC7, 0x45, off); in clean_stack_garbage()
2098 EMIT(0, 4); in clean_stack_garbage()
2106 int i, arg_regs, nr_used_regs = 0; in get_nr_used_regs()
2108 for (i = 0; i < min_t(int, m->nr_args, MAX_BPF_FUNC_ARGS); i++) { in get_nr_used_regs()
2123 int arg_regs, first_off = 0, nr_regs = 0, nr_stack_slots = 0; in save_args()
2128 * mov QWORD PTR [rbp-0x10],rdi in save_args()
2129 * mov QWORD PTR [rbp-0x8],rsi in save_args()
2131 for (i = 0; i < min_t(int, m->nr_args, MAX_BPF_FUNC_ARGS); i++) { in save_args()
2162 for (j = 0; j < arg_regs; j++) { in save_args()
2164 nr_stack_slots * 8 + 0x18); in save_args()
2184 for (j = 0; j < arg_regs; j++) { in save_args()
2200 int i, j, arg_regs, nr_regs = 0; in restore_regs()
2204 * EMIT4(0x48, 0x8B, 0x7D, 0xF0); mov rdi,QWORD PTR [rbp-0x10] in restore_regs()
2205 * EMIT4(0x48, 0x8B, 0x75, 0xF8); mov rsi,QWORD PTR [rbp-0x8] in restore_regs()
2209 for (i = 0; i < min_t(int, m->nr_args, MAX_BPF_FUNC_ARGS); i++) { in restore_regs()
2212 for (j = 0; j < arg_regs; j++) { in restore_regs()
2255 EMIT3_off32(0x48, 0x8D, 0xB5, -run_ctx_off); in invoke_bpf_prog()
2257 EMIT4(0x48, 0x8D, 0x75, -run_ctx_off); in invoke_bpf_prog()
2264 /* if (__bpf_prog_enter*(prog) == 0) in invoke_bpf_prog()
2267 EMIT3(0x48, 0x85, 0xC0); /* test rax,rax */ in invoke_bpf_prog()
2274 EMIT3_off32(0x48, 0x8D, 0xBD, -stack_size); in invoke_bpf_prog()
2276 EMIT4(0x48, 0x8D, 0x7D, -stack_size); in invoke_bpf_prog()
2298 jmp_insn[0] = X86_JE; in invoke_bpf_prog()
2307 EMIT3_off32(0x48, 0x8D, 0x95, -run_ctx_off); in invoke_bpf_prog()
2309 EMIT4(0x48, 0x8D, 0x55, -run_ctx_off); in invoke_bpf_prog()
2314 return 0; in invoke_bpf_prog()
2338 EMIT2_off32(0x0F, jmp_cond + 0x10, offset); in emit_cond_near_jump()
2340 return 0; in emit_cond_near_jump()
2350 for (i = 0; i < tl->nr_links; i++) { in invoke_bpf()
2356 return 0; in invoke_bpf()
2367 * Set this to 0 to avoid confusing the program. in invoke_bpf_mod_ret()
2369 emit_mov_imm32(&prog, false, BPF_REG_0, 0); in invoke_bpf_mod_ret()
2371 for (i = 0; i < tl->nr_links; i++) { in invoke_bpf_mod_ret()
2376 * if (*(u64 *)(rbp - 8) != 0) in invoke_bpf_mod_ret()
2379 /* cmp QWORD PTR [rbp - 0x8], 0x0 */ in invoke_bpf_mod_ret()
2380 EMIT4(0x48, 0x83, 0x7d, 0xf8); EMIT1(0x00); in invoke_bpf_mod_ret()
2392 return 0; in invoke_bpf_mod_ret()
2460 int i, ret, nr_regs = m->nr_args, stack_size = 0; in arch_prepare_bpf_trampoline()
2471 for (i = 0; i < m->nr_args; i++) in arch_prepare_bpf_trampoline()
2484 * RBP + 0 [ RBP ] in arch_prepare_bpf_trampoline()
2528 stack_size += (sizeof(struct bpf_tramp_run_ctx) + 7) & ~0x7; in arch_prepare_bpf_trampoline()
2540 stack_size += (stack_size % 16) ? 0 : 8; in arch_prepare_bpf_trampoline()
2562 EMIT1(0x55); /* push rbp */ in arch_prepare_bpf_trampoline()
2563 EMIT3(0x48, 0x89, 0xE5); /* mov rbp, rsp */ in arch_prepare_bpf_trampoline()
2566 EMIT3_off32(0x48, 0x81, 0xEC, stack_size); in arch_prepare_bpf_trampoline()
2569 EMIT4(0x48, 0x83, 0xEC, stack_size); in arch_prepare_bpf_trampoline()
2571 EMIT1(0x50); /* push rax */ in arch_prepare_bpf_trampoline()
2579 emit_mov_imm64(&prog, BPF_REG_0, 0, (u32) nr_regs); in arch_prepare_bpf_trampoline()
2632 EMIT2(0xff, 0xd3); /* call *rbx */ in arch_prepare_bpf_trampoline()
2657 for (i = 0; i < fmod_ret->nr_links; i++) in arch_prepare_bpf_trampoline()
2694 EMIT1(0xC9); /* leave */ in arch_prepare_bpf_trampoline()
2697 EMIT4(0x48, 0x83, 0xC4, 8); /* add rsp, 8 */ in arch_prepare_bpf_trampoline()
2721 EMIT1(add_1mod(0x48, BPF_REG_3)); /* cmp rdx,func */ in emit_bpf_dispatcher()
2724 EMIT2_off32(0x81, add_1reg(0xF8, BPF_REG_3), in emit_bpf_dispatcher()
2735 return 0; in emit_bpf_dispatcher()
2742 EMIT1(add_1mod(0x48, BPF_REG_3)); /* cmp rdx,func */ in emit_bpf_dispatcher()
2745 EMIT2_off32(0x81, add_1reg(0xF8, BPF_REG_3), progs[a + pivot]); in emit_bpf_dispatcher()
2750 EMIT2_off32(0x0F, X86_JG + 0x10, 0); in emit_bpf_dispatcher()
2752 EMIT2(X86_JG, 0); in emit_bpf_dispatcher()
2776 return 0; in emit_bpf_dispatcher()
2788 return 0; in cmp_ips()
2795 sort(funcs, num_funcs, sizeof(funcs[0]), cmp_ips, NULL); in arch_prepare_bpf_dispatcher()
2796 return emit_bpf_dispatcher(&prog, 0, num_funcs - 1, funcs, image, buf); in arch_prepare_bpf_dispatcher()
2817 int proglen, oldproglen = 0; in bpf_int_jit_compile()
2874 for (proglen = 0, i = 0; i <= prog->len; i++) { in bpf_int_jit_compile()
2887 for (pass = 0; pass < MAX_PASSES || image; pass++) { in bpf_int_jit_compile()
2891 if (proglen <= 0) { in bpf_int_jit_compile()
2903 prog->jited = 0; in bpf_int_jit_compile()
2904 prog->jited_len = 0; in bpf_int_jit_compile()
3052 BUG_ON(ret < 0); in bpf_arch_poke_desc_update()
3058 BUG_ON(ret < 0); in bpf_arch_poke_desc_update()
3065 BUG_ON(ret < 0); in bpf_arch_poke_desc_update()
3075 BUG_ON(ret < 0); in bpf_arch_poke_desc_update()