/tools/testing/selftests/bpf/prog_tests/ |
D | align.c | 55 BPF_ALU64_IMM(BPF_LSH, BPF_REG_3, 1), 56 BPF_ALU64_IMM(BPF_LSH, BPF_REG_3, 1), 57 BPF_ALU64_IMM(BPF_LSH, BPF_REG_3, 1), 58 BPF_ALU64_IMM(BPF_LSH, BPF_REG_3, 1), 59 BPF_ALU64_IMM(BPF_RSH, BPF_REG_3, 4), 61 BPF_ALU64_IMM(BPF_RSH, BPF_REG_4, 1), 62 BPF_ALU64_IMM(BPF_RSH, BPF_REG_4, 1), 63 BPF_ALU64_IMM(BPF_RSH, BPF_REG_4, 1), 64 BPF_ALU64_IMM(BPF_RSH, BPF_REG_4, 1), 89 BPF_ALU64_IMM(BPF_ADD, BPF_REG_3, 4), [all …]
|
D | flow_dissector_load_bytes.c | 14 BPF_ALU64_IMM(BPF_ADD, BPF_REG_3, -1), in serial_test_flow_dissector_load_bytes()
|
D | signal_pending.c | 26 prog[i] = BPF_ALU64_IMM(BPF_MOV, BPF_REG_0, 0); in test_signal_pending_by_type()
|
/tools/testing/selftests/bpf/verifier/ |
D | bpf_loop_inline.c | 42 BPF_ALU64_IMM(BPF_MOV, BPF_REG_1, 1), 44 BPF_ALU64_IMM(BPF_MOV, BPF_REG_1, 2), 48 BPF_ALU64_IMM(BPF_MOV, BPF_REG_3, 0), 49 BPF_ALU64_IMM(BPF_MOV, BPF_REG_4, 0), 51 BPF_ALU64_IMM(BPF_MOV, BPF_REG_0, 0), 54 BPF_ALU64_IMM(BPF_MOV, BPF_REG_0, 1), 75 BPF_ALU64_IMM(BPF_MOV, BPF_REG_4, 0), 77 BPF_ALU64_IMM(BPF_MOV, BPF_REG_1, 1), 80 BPF_ALU64_IMM(BPF_MOV, BPF_REG_3, 0), 82 BPF_ALU64_IMM(BPF_MOV, BPF_REG_0, 0), [all …]
|
D | precise.c | 8 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -8), 18 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -8), 30 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, 1), /* R2=scalar(umin=1, umax=8) */ 32 BPF_ALU64_IMM(BPF_ADD, BPF_REG_1, -8), 67 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -8), 77 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -8), 89 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, 1), /* R2=scalar(umin=1, umax=8) */ 91 BPF_ALU64_IMM(BPF_ADD, BPF_REG_1, -8), 226 BPF_ALU64_IMM(BPF_MOV, BPF_REG_6, 0x400), 227 BPF_ALU64_IMM(BPF_MOV, BPF_REG_7, 0), [all …]
|
D | map_kptr.c | 8 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -4), 28 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -4), 48 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -4), 79 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -4), 111 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -4), 117 BPF_ALU64_IMM(BPF_ADD, BPF_REG_0, 7), 133 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -4), 163 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -4), 172 BPF_ALU64_IMM(BPF_ADD, BPF_REG_1, 4), 187 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -4), [all …]
|
D | basic_instr.c | 5 BPF_ALU64_IMM(BPF_ADD, BPF_REG_1, 2), 8 BPF_ALU64_IMM(BPF_ADD, BPF_REG_1, -1), 9 BPF_ALU64_IMM(BPF_MUL, BPF_REG_1, 3), 20 BPF_ALU64_IMM(BPF_LSH, BPF_REG_2, 32), 21 BPF_ALU64_IMM(BPF_OR, BPF_REG_2, 0xffff), 78 BPF_ALU64_IMM(BPF_ARSH, BPF_REG_0, 5), 98 BPF_ALU64_IMM(BPF_LSH, BPF_REG_1, 0), 112 BPF_ALU64_IMM(BPF_RSH, BPF_REG_1, 0), 126 BPF_ALU64_IMM(BPF_ARSH, BPF_REG_1, 0),
|
D | calls.c | 28 BPF_ALU64_IMM(BPF_ADD, BPF_REG_1, -8), 43 BPF_ALU64_IMM(BPF_ADD, BPF_REG_1, -8), 58 BPF_ALU64_IMM(BPF_ADD, BPF_REG_1, -8), 73 BPF_ALU64_IMM(BPF_ADD, BPF_REG_1, -8), 88 BPF_ALU64_IMM(BPF_ADD, BPF_REG_1, -8), 103 BPF_ALU64_IMM(BPF_ADD, BPF_REG_1, -8), 122 BPF_ALU64_IMM(BPF_ADD, BPF_REG_1, -8), 127 BPF_ALU64_IMM(BPF_ADD, BPF_REG_0, 8), 165 BPF_ALU64_IMM(BPF_ADD, BPF_REG_1, -8), 172 BPF_ALU64_IMM(BPF_ADD, BPF_REG_1, -4), [all …]
|
D | atomic_or.c | 55 BPF_ALU64_IMM(BPF_SUB, BPF_REG_0, 1), 73 BPF_ALU64_IMM(BPF_ADD, BPF_REG_0, 1), 83 BPF_ALU64_IMM(BPF_SUB, BPF_REG_1, 1), 91 BPF_ALU64_IMM(BPF_LSH, BPF_REG_2, 32), 92 BPF_ALU64_IMM(BPF_SUB, BPF_REG_2, 1),
|
D | jit.c | 6 BPF_ALU64_IMM(BPF_LSH, BPF_REG_1, 1), 10 BPF_ALU64_IMM(BPF_RSH, BPF_REG_1, 1), 14 BPF_ALU64_IMM(BPF_ARSH, BPF_REG_1, 1), 52 BPF_ALU64_IMM(BPF_RSH, BPF_REG_1, 32), 90 BPF_ALU64_IMM(BPF_MUL, BPF_REG_3, 0xefefef), 140 BPF_ALU64_IMM(BPF_DIV, BPF_REG_3, 0xfefeeeULL), 145 BPF_ALU64_IMM(BPF_MOD, BPF_REG_1, 0xbeefULL),
|
D | jump.c | 59 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -8), 63 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -16), 67 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -32), 71 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -40), 75 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -48), 79 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -56), 381 BPF_ALU64_IMM(BPF_NEG, BPF_REG_3, 0), 382 BPF_ALU64_IMM(BPF_NEG, BPF_REG_3, 0), 383 BPF_ALU64_IMM(BPF_OR, BPF_REG_3, 32767), 388 BPF_ALU64_IMM(BPF_ADD, BPF_REG_3, -32767),
|
D | bpf_st_mem.c | 6 BPF_ALU64_IMM(BPF_ADD, BPF_REG_0, -42), 49 BPF_ALU64_IMM(BPF_SUB, BPF_REG_0, 32),
|
D | atomic_xor.c | 55 BPF_ALU64_IMM(BPF_SUB, BPF_REG_0, 1), 73 BPF_ALU64_IMM(BPF_ADD, BPF_REG_0, 1),
|
D | jmp32.c | 798 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -8), 807 BPF_ALU64_IMM(BPF_LSH, BPF_REG_6, 32), 808 BPF_ALU64_IMM(BPF_RSH, BPF_REG_6, 32), 825 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -8), 835 BPF_ALU64_IMM(BPF_LSH, BPF_REG_6, 32), 836 BPF_ALU64_IMM(BPF_RSH, BPF_REG_6, 32), 853 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -8), 863 BPF_ALU64_IMM(BPF_LSH, BPF_REG_6, 32), 864 BPF_ALU64_IMM(BPF_RSH, BPF_REG_6, 32), 880 BPF_ALU64_IMM(BPF_NEG, BPF_REG_2, 0), [all …]
|
D | atomic_fetch.c | 8 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -8), 32 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -8), 56 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -8), 79 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -8),
|
D | atomic_and.c | 55 BPF_ALU64_IMM(BPF_SUB, BPF_REG_0, 1), 73 BPF_ALU64_IMM(BPF_ADD, BPF_REG_0, 1),
|
D | ld_abs.c | 170 BPF_ALU64_IMM(BPF_MOV, BPF_REG_2, 2), 190 BPF_ALU64_IMM(BPF_MOV, BPF_REG_2, 2), 209 BPF_ALU64_IMM(BPF_MOV, BPF_REG_7, 0), 225 BPF_ALU64_IMM(BPF_MOV, BPF_REG_7, 0),
|
D | atomic_cmpxchg.c | 91 BPF_ALU64_IMM(BPF_SUB, BPF_REG_0, 1), 99 BPF_ALU64_IMM(BPF_LSH, BPF_REG_1, 32), 100 BPF_ALU64_IMM(BPF_SUB, BPF_REG_1, 1),
|
D | jset.c | 137 BPF_ALU64_IMM(BPF_OR, BPF_REG_0, 2), 154 BPF_ALU64_IMM(BPF_AND, BPF_REG_1, 0xff),
|
D | ctx_skb.c | 47 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -8), 71 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -8), 94 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -8), 323 BPF_ALU64_IMM(BPF_ADD, BPF_REG_0, 8), 340 BPF_ALU64_IMM(BPF_ADD, BPF_REG_0, 8), 357 BPF_ALU64_IMM(BPF_ADD, BPF_REG_0, 8), 360 BPF_ALU64_IMM(BPF_ADD, BPF_REG_1, 6), 1160 BPF_ALU64_IMM(BPF_ADD, BPF_REG_3, 42), // 3. r3 += 42 1163 BPF_ALU64_IMM(BPF_ADD, BPF_REG_4, 14), // 6. r4 += 14 1182 BPF_ALU64_IMM(BPF_ADD, BPF_REG_3, 42), // 3. r3 += 42 [all …]
|
D | event_output.c | 31 BPF_ALU64_IMM(BPF_ADD, BPF_REG_4, -8), \
|
/tools/testing/selftests/bpf/ |
D | test_sysctl.c | 134 BPF_ALU64_IMM(BPF_AND, BPF_REG_7, 1), 231 BPF_ALU64_IMM(BPF_ADD, BPF_REG_7, -8), 272 BPF_ALU64_IMM(BPF_ADD, BPF_REG_7, -8), 314 BPF_ALU64_IMM(BPF_ADD, BPF_REG_7, -24), 369 BPF_ALU64_IMM(BPF_ADD, BPF_REG_7, -16), 418 BPF_ALU64_IMM(BPF_ADD, BPF_REG_7, -8), 460 BPF_ALU64_IMM(BPF_ADD, BPF_REG_7, -8), 496 BPF_ALU64_IMM(BPF_ADD, BPF_REG_7, -8), 535 BPF_ALU64_IMM(BPF_ADD, BPF_REG_7, -8), 574 BPF_ALU64_IMM(BPF_ADD, BPF_REG_7, -8), [all …]
|
D | test_cgroup_storage.c | 25 BPF_ALU64_IMM(BPF_ADD, BPF_REG_3, 0x1), in main() 35 BPF_ALU64_IMM(BPF_AND, BPF_REG_1, 0x1), in main()
|
D | test_verifier.c | 227 insn[i] = BPF_ALU64_IMM(BPF_MOV, BPF_REG_0, 0xbef); in bpf_fill_ld_abs_vlan_push_pop() 274 insn[i++] = BPF_ALU64_IMM(BPF_RSH, BPF_REG_1, 32); in bpf_fill_rand_ld_dw() 304 insn[i++] = BPF_ALU64_IMM(BPF_MOV, BPF_REG_0, 42); in bpf_fill_scale1() 333 insn[i++] = BPF_ALU64_IMM(BPF_MOV, BPF_REG_0, 42); in bpf_fill_scale2() 439 insn[i++] = BPF_ALU64_IMM(BPF_MOV, BPF_REG_1, 1); in bpf_fill_big_prog_with_loop_1() 445 insn[i++] = BPF_ALU64_IMM(BPF_MOV, BPF_REG_3, 0); in bpf_fill_big_prog_with_loop_1() 446 insn[i++] = BPF_ALU64_IMM(BPF_MOV, BPF_REG_4, 0); in bpf_fill_big_prog_with_loop_1() 450 insn[i++] = BPF_ALU64_IMM(BPF_MOV, BPF_REG_0, 0); in bpf_fill_big_prog_with_loop_1() 454 insn[i++] = BPF_ALU64_IMM(BPF_MOV, BPF_REG_0, 0); in bpf_fill_big_prog_with_loop_1() 475 BPF_ALU64_IMM(BPF_ADD, BPF_REG_2, -48), \ [all …]
|
D | test_tag.c | 33 prog[i] = BPF_ALU64_IMM(BPF_MOV, i % BPF_REG_10, rand()); in bpf_gen_imm_prog() 49 prog[insns - 2] = BPF_ALU64_IMM(BPF_MOV, i % BPF_REG_10, 42); in bpf_gen_map_prog()
|