• Home
  • Raw
  • Download

Lines Matching refs:insn

90 	struct sock_filter *insn;  in bpf_fill_maxinsns1()  local
94 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); in bpf_fill_maxinsns1()
95 if (!insn) in bpf_fill_maxinsns1()
99 insn[i] = __BPF_STMT(BPF_RET | BPF_K, k); in bpf_fill_maxinsns1()
101 self->u.ptr.insns = insn; in bpf_fill_maxinsns1()
110 struct sock_filter *insn; in bpf_fill_maxinsns2() local
113 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); in bpf_fill_maxinsns2()
114 if (!insn) in bpf_fill_maxinsns2()
118 insn[i] = __BPF_STMT(BPF_RET | BPF_K, 0xfefefefe); in bpf_fill_maxinsns2()
120 self->u.ptr.insns = insn; in bpf_fill_maxinsns2()
129 struct sock_filter *insn; in bpf_fill_maxinsns3() local
133 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); in bpf_fill_maxinsns3()
134 if (!insn) in bpf_fill_maxinsns3()
142 insn[i] = __BPF_STMT(BPF_ALU | BPF_ADD | BPF_K, k); in bpf_fill_maxinsns3()
145 insn[len - 1] = __BPF_STMT(BPF_RET | BPF_A, 0); in bpf_fill_maxinsns3()
147 self->u.ptr.insns = insn; in bpf_fill_maxinsns3()
156 struct sock_filter *insn; in bpf_fill_maxinsns4() local
159 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); in bpf_fill_maxinsns4()
160 if (!insn) in bpf_fill_maxinsns4()
164 insn[i] = __BPF_STMT(BPF_RET | BPF_K, 0xfefefefe); in bpf_fill_maxinsns4()
166 self->u.ptr.insns = insn; in bpf_fill_maxinsns4()
175 struct sock_filter *insn; in bpf_fill_maxinsns5() local
178 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); in bpf_fill_maxinsns5()
179 if (!insn) in bpf_fill_maxinsns5()
182 insn[0] = __BPF_JUMP(BPF_JMP | BPF_JA, len - 2, 0, 0); in bpf_fill_maxinsns5()
185 insn[i] = __BPF_STMT(BPF_RET | BPF_K, 0xfefefefe); in bpf_fill_maxinsns5()
187 insn[len - 1] = __BPF_STMT(BPF_RET | BPF_K, 0xabababab); in bpf_fill_maxinsns5()
189 self->u.ptr.insns = insn; in bpf_fill_maxinsns5()
198 struct sock_filter *insn; in bpf_fill_maxinsns6() local
201 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); in bpf_fill_maxinsns6()
202 if (!insn) in bpf_fill_maxinsns6()
206 insn[i] = __BPF_STMT(BPF_LD | BPF_W | BPF_ABS, SKF_AD_OFF + in bpf_fill_maxinsns6()
209 insn[len - 1] = __BPF_STMT(BPF_RET | BPF_A, 0); in bpf_fill_maxinsns6()
211 self->u.ptr.insns = insn; in bpf_fill_maxinsns6()
220 struct sock_filter *insn; in bpf_fill_maxinsns7() local
223 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); in bpf_fill_maxinsns7()
224 if (!insn) in bpf_fill_maxinsns7()
228 insn[i] = __BPF_STMT(BPF_LD | BPF_W | BPF_ABS, SKF_AD_OFF + in bpf_fill_maxinsns7()
231 insn[len - 4] = __BPF_STMT(BPF_MISC | BPF_TAX, 0); in bpf_fill_maxinsns7()
232 insn[len - 3] = __BPF_STMT(BPF_LD | BPF_W | BPF_ABS, SKF_AD_OFF + in bpf_fill_maxinsns7()
234 insn[len - 2] = __BPF_STMT(BPF_ALU | BPF_SUB | BPF_X, 0); in bpf_fill_maxinsns7()
235 insn[len - 1] = __BPF_STMT(BPF_RET | BPF_A, 0); in bpf_fill_maxinsns7()
237 self->u.ptr.insns = insn; in bpf_fill_maxinsns7()
246 struct sock_filter *insn; in bpf_fill_maxinsns8() local
249 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); in bpf_fill_maxinsns8()
250 if (!insn) in bpf_fill_maxinsns8()
253 insn[0] = __BPF_STMT(BPF_LD | BPF_IMM, 0xffffffff); in bpf_fill_maxinsns8()
256 insn[i] = __BPF_JUMP(BPF_JMP | BPF_JGT, 0xffffffff, jmp_off--, 0); in bpf_fill_maxinsns8()
258 insn[len - 1] = __BPF_STMT(BPF_RET | BPF_A, 0); in bpf_fill_maxinsns8()
260 self->u.ptr.insns = insn; in bpf_fill_maxinsns8()
269 struct bpf_insn *insn; in bpf_fill_maxinsns9() local
272 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); in bpf_fill_maxinsns9()
273 if (!insn) in bpf_fill_maxinsns9()
276 insn[0] = BPF_JMP_IMM(BPF_JA, 0, 0, len - 2); in bpf_fill_maxinsns9()
277 insn[1] = BPF_ALU32_IMM(BPF_MOV, R0, 0xcbababab); in bpf_fill_maxinsns9()
278 insn[2] = BPF_EXIT_INSN(); in bpf_fill_maxinsns9()
281 insn[i] = BPF_ALU32_IMM(BPF_MOV, R0, 0xfefefefe); in bpf_fill_maxinsns9()
283 insn[len - 2] = BPF_EXIT_INSN(); in bpf_fill_maxinsns9()
284 insn[len - 1] = BPF_JMP_IMM(BPF_JA, 0, 0, -(len - 1)); in bpf_fill_maxinsns9()
286 self->u.ptr.insns = insn; in bpf_fill_maxinsns9()
295 struct bpf_insn *insn; in bpf_fill_maxinsns10() local
298 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); in bpf_fill_maxinsns10()
299 if (!insn) in bpf_fill_maxinsns10()
303 insn[i] = BPF_JMP_IMM(BPF_JA, 0, 0, hlen - 2 - 2 * i); in bpf_fill_maxinsns10()
305 insn[i] = BPF_JMP_IMM(BPF_JA, 0, 0, hlen - 1 - 2 * i); in bpf_fill_maxinsns10()
307 insn[hlen / 2] = BPF_JMP_IMM(BPF_JA, 0, 0, hlen / 2 - 1); in bpf_fill_maxinsns10()
308 insn[hlen] = BPF_ALU32_IMM(BPF_MOV, R0, 0xabababac); in bpf_fill_maxinsns10()
309 insn[hlen + 1] = BPF_EXIT_INSN(); in bpf_fill_maxinsns10()
311 self->u.ptr.insns = insn; in bpf_fill_maxinsns10()
320 struct sock_filter *insn; in __bpf_fill_ja() local
324 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); in __bpf_fill_ja()
325 if (!insn) in __bpf_fill_ja()
332 insn[i + j] = __BPF_JUMP(BPF_JMP | BPF_JA, in __bpf_fill_ja()
335 insn[i + j] = __BPF_JUMP(BPF_JMP | BPF_JA, rlen - 1 - j, in __bpf_fill_ja()
338 insn[len - 1] = __BPF_STMT(BPF_RET | BPF_K, 0xababcbac); in __bpf_fill_ja()
340 self->u.ptr.insns = insn; in __bpf_fill_ja()
355 struct sock_filter *insn; in bpf_fill_maxinsns12() local
358 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); in bpf_fill_maxinsns12()
359 if (!insn) in bpf_fill_maxinsns12()
362 insn[0] = __BPF_JUMP(BPF_JMP | BPF_JA, len - 2, 0, 0); in bpf_fill_maxinsns12()
365 insn[i] = __BPF_STMT(BPF_LDX | BPF_B | BPF_MSH, 0); in bpf_fill_maxinsns12()
367 insn[len - 1] = __BPF_STMT(BPF_RET | BPF_K, 0xabababab); in bpf_fill_maxinsns12()
369 self->u.ptr.insns = insn; in bpf_fill_maxinsns12()
378 struct sock_filter *insn; in bpf_fill_maxinsns13() local
381 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); in bpf_fill_maxinsns13()
382 if (!insn) in bpf_fill_maxinsns13()
386 insn[i] = __BPF_STMT(BPF_LDX | BPF_B | BPF_MSH, 0); in bpf_fill_maxinsns13()
388 insn[len - 3] = __BPF_STMT(BPF_LD | BPF_IMM, 0xabababab); in bpf_fill_maxinsns13()
389 insn[len - 2] = __BPF_STMT(BPF_ALU | BPF_XOR | BPF_X, 0); in bpf_fill_maxinsns13()
390 insn[len - 1] = __BPF_STMT(BPF_RET | BPF_A, 0); in bpf_fill_maxinsns13()
392 self->u.ptr.insns = insn; in bpf_fill_maxinsns13()
407 struct sock_filter *insn; in bpf_fill_ld_abs_get_processor_id() local
410 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); in bpf_fill_ld_abs_get_processor_id()
411 if (!insn) in bpf_fill_ld_abs_get_processor_id()
415 insn[i] = __BPF_STMT(BPF_LD | BPF_B | BPF_ABS, 0); in bpf_fill_ld_abs_get_processor_id()
416 insn[i + 1] = __BPF_STMT(BPF_LD | BPF_W | BPF_ABS, in bpf_fill_ld_abs_get_processor_id()
420 insn[len - 1] = __BPF_STMT(BPF_RET | BPF_K, 0xbee); in bpf_fill_ld_abs_get_processor_id()
422 self->u.ptr.insns = insn; in bpf_fill_ld_abs_get_processor_id()
431 struct bpf_insn *insn; in __bpf_fill_stxdw() local
434 insn = kmalloc_array(len, sizeof(*insn), GFP_KERNEL); in __bpf_fill_stxdw()
435 if (!insn) in __bpf_fill_stxdw()
438 insn[0] = BPF_ALU32_IMM(BPF_MOV, R0, 1); in __bpf_fill_stxdw()
439 insn[1] = BPF_ST_MEM(size, R10, -40, 42); in __bpf_fill_stxdw()
442 insn[i] = BPF_STX_XADD(size, R10, R0, -40); in __bpf_fill_stxdw()
444 insn[len - 2] = BPF_LDX_MEM(size, R0, R10, -40); in __bpf_fill_stxdw()
445 insn[len - 1] = BPF_EXIT_INSN(); in __bpf_fill_stxdw()
447 self->u.ptr.insns = insn; in __bpf_fill_stxdw()