• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 // SPDX-License-Identifier: (GPL-2.0-only OR BSD-2-Clause)
2 /* Copyright (C) 2016-2018 Netronome Systems, Inc. */
3 
4 #include <linux/bpf.h>
5 #include <linux/bpf_verifier.h>
6 #include <linux/kernel.h>
7 #include <linux/netdevice.h>
8 #include <linux/pkt_cls.h>
9 
10 #include "../nfp_app.h"
11 #include "../nfp_main.h"
12 #include "../nfp_net.h"
13 #include "fw.h"
14 #include "main.h"
15 
16 #define pr_vlog(env, fmt, ...)	\
17 	bpf_verifier_log_write(env, "[nfp] " fmt, ##__VA_ARGS__)
18 
19 struct nfp_insn_meta *
nfp_bpf_goto_meta(struct nfp_prog * nfp_prog,struct nfp_insn_meta * meta,unsigned int insn_idx)20 nfp_bpf_goto_meta(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta,
21 		  unsigned int insn_idx)
22 {
23 	unsigned int forward, backward, i;
24 
25 	backward = meta->n - insn_idx;
26 	forward = insn_idx - meta->n;
27 
28 	if (min(forward, backward) > nfp_prog->n_insns - insn_idx - 1) {
29 		backward = nfp_prog->n_insns - insn_idx - 1;
30 		meta = nfp_prog_last_meta(nfp_prog);
31 	}
32 	if (min(forward, backward) > insn_idx && backward > insn_idx) {
33 		forward = insn_idx;
34 		meta = nfp_prog_first_meta(nfp_prog);
35 	}
36 
37 	if (forward < backward)
38 		for (i = 0; i < forward; i++)
39 			meta = nfp_meta_next(meta);
40 	else
41 		for (i = 0; i < backward; i++)
42 			meta = nfp_meta_prev(meta);
43 
44 	return meta;
45 }
46 
47 static void
nfp_record_adjust_head(struct nfp_app_bpf * bpf,struct nfp_prog * nfp_prog,struct nfp_insn_meta * meta,const struct bpf_reg_state * reg2)48 nfp_record_adjust_head(struct nfp_app_bpf *bpf, struct nfp_prog *nfp_prog,
49 		       struct nfp_insn_meta *meta,
50 		       const struct bpf_reg_state *reg2)
51 {
52 	unsigned int location =	UINT_MAX;
53 	int imm;
54 
55 	/* Datapath usually can give us guarantees on how much adjust head
56 	 * can be done without the need for any checks.  Optimize the simple
57 	 * case where there is only one adjust head by a constant.
58 	 */
59 	if (reg2->type != SCALAR_VALUE || !tnum_is_const(reg2->var_off))
60 		goto exit_set_location;
61 	imm = reg2->var_off.value;
62 	/* Translator will skip all checks, we need to guarantee min pkt len */
63 	if (imm > ETH_ZLEN - ETH_HLEN)
64 		goto exit_set_location;
65 	if (imm > (int)bpf->adjust_head.guaranteed_add ||
66 	    imm < -bpf->adjust_head.guaranteed_sub)
67 		goto exit_set_location;
68 
69 	if (nfp_prog->adjust_head_location) {
70 		/* Only one call per program allowed */
71 		if (nfp_prog->adjust_head_location != meta->n)
72 			goto exit_set_location;
73 
74 		if (meta->arg2.reg.var_off.value != imm)
75 			goto exit_set_location;
76 	}
77 
78 	location = meta->n;
79 exit_set_location:
80 	nfp_prog->adjust_head_location = location;
81 }
82 
nfp_bpf_map_update_value_ok(struct bpf_verifier_env * env)83 static bool nfp_bpf_map_update_value_ok(struct bpf_verifier_env *env)
84 {
85 	const struct bpf_reg_state *reg1 = cur_regs(env) + BPF_REG_1;
86 	const struct bpf_reg_state *reg3 = cur_regs(env) + BPF_REG_3;
87 	struct bpf_offloaded_map *offmap;
88 	struct bpf_func_state *state;
89 	struct nfp_bpf_map *nfp_map;
90 	int off, i;
91 
92 	state = env->cur_state->frame[reg3->frameno];
93 
94 	/* We need to record each time update happens with non-zero words,
95 	 * in case such word is used in atomic operations.
96 	 * Implicitly depend on nfp_bpf_stack_arg_ok(reg3) being run before.
97 	 */
98 
99 	offmap = map_to_offmap(reg1->map_ptr);
100 	nfp_map = offmap->dev_priv;
101 	off = reg3->off + reg3->var_off.value;
102 
103 	for (i = 0; i < offmap->map.value_size; i++) {
104 		struct bpf_stack_state *stack_entry;
105 		unsigned int soff;
106 
107 		soff = -(off + i) - 1;
108 		stack_entry = &state->stack[soff / BPF_REG_SIZE];
109 		if (stack_entry->slot_type[soff % BPF_REG_SIZE] == STACK_ZERO)
110 			continue;
111 
112 		if (nfp_map->use_map[i / 4].type == NFP_MAP_USE_ATOMIC_CNT) {
113 			pr_vlog(env, "value at offset %d/%d may be non-zero, bpf_map_update_elem() is required to initialize atomic counters to zero to avoid offload endian issues\n",
114 				i, soff);
115 			return false;
116 		}
117 		nfp_map->use_map[i / 4].non_zero_update = 1;
118 	}
119 
120 	return true;
121 }
122 
123 static int
nfp_bpf_stack_arg_ok(const char * fname,struct bpf_verifier_env * env,const struct bpf_reg_state * reg,struct nfp_bpf_reg_state * old_arg)124 nfp_bpf_stack_arg_ok(const char *fname, struct bpf_verifier_env *env,
125 		     const struct bpf_reg_state *reg,
126 		     struct nfp_bpf_reg_state *old_arg)
127 {
128 	s64 off, old_off;
129 
130 	if (reg->type != PTR_TO_STACK) {
131 		pr_vlog(env, "%s: unsupported ptr type %d\n",
132 			fname, reg->type);
133 		return false;
134 	}
135 	if (!tnum_is_const(reg->var_off)) {
136 		pr_vlog(env, "%s: variable pointer\n", fname);
137 		return false;
138 	}
139 
140 	off = reg->var_off.value + reg->off;
141 	if (-off % 4) {
142 		pr_vlog(env, "%s: unaligned stack pointer %lld\n", fname, -off);
143 		return false;
144 	}
145 
146 	/* Rest of the checks is only if we re-parse the same insn */
147 	if (!old_arg)
148 		return true;
149 
150 	old_off = old_arg->reg.var_off.value + old_arg->reg.off;
151 	old_arg->var_off |= off != old_off;
152 
153 	return true;
154 }
155 
156 static bool
nfp_bpf_map_call_ok(const char * fname,struct bpf_verifier_env * env,struct nfp_insn_meta * meta,u32 helper_tgt,const struct bpf_reg_state * reg1)157 nfp_bpf_map_call_ok(const char *fname, struct bpf_verifier_env *env,
158 		    struct nfp_insn_meta *meta,
159 		    u32 helper_tgt, const struct bpf_reg_state *reg1)
160 {
161 	if (!helper_tgt) {
162 		pr_vlog(env, "%s: not supported by FW\n", fname);
163 		return false;
164 	}
165 
166 	return true;
167 }
168 
169 static int
nfp_bpf_check_helper_call(struct nfp_prog * nfp_prog,struct bpf_verifier_env * env,struct nfp_insn_meta * meta)170 nfp_bpf_check_helper_call(struct nfp_prog *nfp_prog,
171 			  struct bpf_verifier_env *env,
172 			  struct nfp_insn_meta *meta)
173 {
174 	const struct bpf_reg_state *reg1 = cur_regs(env) + BPF_REG_1;
175 	const struct bpf_reg_state *reg2 = cur_regs(env) + BPF_REG_2;
176 	const struct bpf_reg_state *reg3 = cur_regs(env) + BPF_REG_3;
177 	struct nfp_app_bpf *bpf = nfp_prog->bpf;
178 	u32 func_id = meta->insn.imm;
179 
180 	switch (func_id) {
181 	case BPF_FUNC_xdp_adjust_head:
182 		if (!bpf->adjust_head.off_max) {
183 			pr_vlog(env, "adjust_head not supported by FW\n");
184 			return -EOPNOTSUPP;
185 		}
186 		if (!(bpf->adjust_head.flags & NFP_BPF_ADJUST_HEAD_NO_META)) {
187 			pr_vlog(env, "adjust_head: FW requires shifting metadata, not supported by the driver\n");
188 			return -EOPNOTSUPP;
189 		}
190 
191 		nfp_record_adjust_head(bpf, nfp_prog, meta, reg2);
192 		break;
193 
194 	case BPF_FUNC_xdp_adjust_tail:
195 		if (!bpf->adjust_tail) {
196 			pr_vlog(env, "adjust_tail not supported by FW\n");
197 			return -EOPNOTSUPP;
198 		}
199 		break;
200 
201 	case BPF_FUNC_map_lookup_elem:
202 		if (!nfp_bpf_map_call_ok("map_lookup", env, meta,
203 					 bpf->helpers.map_lookup, reg1) ||
204 		    !nfp_bpf_stack_arg_ok("map_lookup", env, reg2,
205 					  meta->func_id ? &meta->arg2 : NULL))
206 			return -EOPNOTSUPP;
207 		break;
208 
209 	case BPF_FUNC_map_update_elem:
210 		if (!nfp_bpf_map_call_ok("map_update", env, meta,
211 					 bpf->helpers.map_update, reg1) ||
212 		    !nfp_bpf_stack_arg_ok("map_update", env, reg2,
213 					  meta->func_id ? &meta->arg2 : NULL) ||
214 		    !nfp_bpf_stack_arg_ok("map_update", env, reg3, NULL) ||
215 		    !nfp_bpf_map_update_value_ok(env))
216 			return -EOPNOTSUPP;
217 		break;
218 
219 	case BPF_FUNC_map_delete_elem:
220 		if (!nfp_bpf_map_call_ok("map_delete", env, meta,
221 					 bpf->helpers.map_delete, reg1) ||
222 		    !nfp_bpf_stack_arg_ok("map_delete", env, reg2,
223 					  meta->func_id ? &meta->arg2 : NULL))
224 			return -EOPNOTSUPP;
225 		break;
226 
227 	case BPF_FUNC_get_prandom_u32:
228 		if (bpf->pseudo_random)
229 			break;
230 		pr_vlog(env, "bpf_get_prandom_u32(): FW doesn't support random number generation\n");
231 		return -EOPNOTSUPP;
232 
233 	case BPF_FUNC_perf_event_output:
234 		BUILD_BUG_ON(NFP_BPF_SCALAR_VALUE != SCALAR_VALUE ||
235 			     NFP_BPF_MAP_VALUE != PTR_TO_MAP_VALUE ||
236 			     NFP_BPF_STACK != PTR_TO_STACK ||
237 			     NFP_BPF_PACKET_DATA != PTR_TO_PACKET);
238 
239 		if (!bpf->helpers.perf_event_output) {
240 			pr_vlog(env, "event_output: not supported by FW\n");
241 			return -EOPNOTSUPP;
242 		}
243 
244 		/* Force current CPU to make sure we can report the event
245 		 * wherever we get the control message from FW.
246 		 */
247 		if (reg3->var_off.mask & BPF_F_INDEX_MASK ||
248 		    (reg3->var_off.value & BPF_F_INDEX_MASK) !=
249 		    BPF_F_CURRENT_CPU) {
250 			char tn_buf[48];
251 
252 			tnum_strn(tn_buf, sizeof(tn_buf), reg3->var_off);
253 			pr_vlog(env, "event_output: must use BPF_F_CURRENT_CPU, var_off: %s\n",
254 				tn_buf);
255 			return -EOPNOTSUPP;
256 		}
257 
258 		/* Save space in meta, we don't care about arguments other
259 		 * than 4th meta, shove it into arg1.
260 		 */
261 		reg1 = cur_regs(env) + BPF_REG_4;
262 
263 		if (reg1->type != SCALAR_VALUE /* NULL ptr */ &&
264 		    reg1->type != PTR_TO_STACK &&
265 		    reg1->type != PTR_TO_MAP_VALUE &&
266 		    reg1->type != PTR_TO_PACKET) {
267 			pr_vlog(env, "event_output: unsupported ptr type: %d\n",
268 				reg1->type);
269 			return -EOPNOTSUPP;
270 		}
271 
272 		if (reg1->type == PTR_TO_STACK &&
273 		    !nfp_bpf_stack_arg_ok("event_output", env, reg1, NULL))
274 			return -EOPNOTSUPP;
275 
276 		/* Warn user that on offload NFP may return success even if map
277 		 * is not going to accept the event, since the event output is
278 		 * fully async and device won't know the state of the map.
279 		 * There is also FW limitation on the event length.
280 		 *
281 		 * Lost events will not show up on the perf ring, driver
282 		 * won't see them at all.  Events may also get reordered.
283 		 */
284 		dev_warn_once(&nfp_prog->bpf->app->pf->pdev->dev,
285 			      "bpf: note: return codes and behavior of bpf_event_output() helper differs for offloaded programs!\n");
286 		pr_vlog(env, "warning: return codes and behavior of event_output helper differ for offload!\n");
287 
288 		if (!meta->func_id)
289 			break;
290 
291 		if (reg1->type != meta->arg1.type) {
292 			pr_vlog(env, "event_output: ptr type changed: %d %d\n",
293 				meta->arg1.type, reg1->type);
294 			return -EINVAL;
295 		}
296 		break;
297 
298 	default:
299 		pr_vlog(env, "unsupported function id: %d\n", func_id);
300 		return -EOPNOTSUPP;
301 	}
302 
303 	meta->func_id = func_id;
304 	meta->arg1 = *reg1;
305 	meta->arg2.reg = *reg2;
306 
307 	return 0;
308 }
309 
310 static int
nfp_bpf_check_exit(struct nfp_prog * nfp_prog,struct bpf_verifier_env * env)311 nfp_bpf_check_exit(struct nfp_prog *nfp_prog,
312 		   struct bpf_verifier_env *env)
313 {
314 	const struct bpf_reg_state *reg0 = cur_regs(env) + BPF_REG_0;
315 	u64 imm;
316 
317 	if (nfp_prog->type == BPF_PROG_TYPE_XDP)
318 		return 0;
319 
320 	if (!(reg0->type == SCALAR_VALUE && tnum_is_const(reg0->var_off))) {
321 		char tn_buf[48];
322 
323 		tnum_strn(tn_buf, sizeof(tn_buf), reg0->var_off);
324 		pr_vlog(env, "unsupported exit state: %d, var_off: %s\n",
325 			reg0->type, tn_buf);
326 		return -EINVAL;
327 	}
328 
329 	imm = reg0->var_off.value;
330 	if (nfp_prog->type == BPF_PROG_TYPE_SCHED_CLS &&
331 	    imm <= TC_ACT_REDIRECT &&
332 	    imm != TC_ACT_SHOT && imm != TC_ACT_STOLEN &&
333 	    imm != TC_ACT_QUEUED) {
334 		pr_vlog(env, "unsupported exit state: %d, imm: %llx\n",
335 			reg0->type, imm);
336 		return -EINVAL;
337 	}
338 
339 	return 0;
340 }
341 
342 static int
nfp_bpf_check_stack_access(struct nfp_prog * nfp_prog,struct nfp_insn_meta * meta,const struct bpf_reg_state * reg,struct bpf_verifier_env * env)343 nfp_bpf_check_stack_access(struct nfp_prog *nfp_prog,
344 			   struct nfp_insn_meta *meta,
345 			   const struct bpf_reg_state *reg,
346 			   struct bpf_verifier_env *env)
347 {
348 	s32 old_off, new_off;
349 
350 	if (reg->frameno != env->cur_state->curframe)
351 		meta->flags |= FLAG_INSN_PTR_CALLER_STACK_FRAME;
352 
353 	if (!tnum_is_const(reg->var_off)) {
354 		pr_vlog(env, "variable ptr stack access\n");
355 		return -EINVAL;
356 	}
357 
358 	if (meta->ptr.type == NOT_INIT)
359 		return 0;
360 
361 	old_off = meta->ptr.off + meta->ptr.var_off.value;
362 	new_off = reg->off + reg->var_off.value;
363 
364 	meta->ptr_not_const |= old_off != new_off;
365 
366 	if (!meta->ptr_not_const)
367 		return 0;
368 
369 	if (old_off % 4 == new_off % 4)
370 		return 0;
371 
372 	pr_vlog(env, "stack access changed location was:%d is:%d\n",
373 		old_off, new_off);
374 	return -EINVAL;
375 }
376 
nfp_bpf_map_use_name(enum nfp_bpf_map_use use)377 static const char *nfp_bpf_map_use_name(enum nfp_bpf_map_use use)
378 {
379 	static const char * const names[] = {
380 		[NFP_MAP_UNUSED]	= "unused",
381 		[NFP_MAP_USE_READ]	= "read",
382 		[NFP_MAP_USE_WRITE]	= "write",
383 		[NFP_MAP_USE_ATOMIC_CNT] = "atomic",
384 	};
385 
386 	if (use >= ARRAY_SIZE(names) || !names[use])
387 		return "unknown";
388 	return names[use];
389 }
390 
391 static int
nfp_bpf_map_mark_used_one(struct bpf_verifier_env * env,struct nfp_bpf_map * nfp_map,unsigned int off,enum nfp_bpf_map_use use)392 nfp_bpf_map_mark_used_one(struct bpf_verifier_env *env,
393 			  struct nfp_bpf_map *nfp_map,
394 			  unsigned int off, enum nfp_bpf_map_use use)
395 {
396 	if (nfp_map->use_map[off / 4].type != NFP_MAP_UNUSED &&
397 	    nfp_map->use_map[off / 4].type != use) {
398 		pr_vlog(env, "map value use type conflict %s vs %s off: %u\n",
399 			nfp_bpf_map_use_name(nfp_map->use_map[off / 4].type),
400 			nfp_bpf_map_use_name(use), off);
401 		return -EOPNOTSUPP;
402 	}
403 
404 	if (nfp_map->use_map[off / 4].non_zero_update &&
405 	    use == NFP_MAP_USE_ATOMIC_CNT) {
406 		pr_vlog(env, "atomic counter in map value may already be initialized to non-zero value off: %u\n",
407 			off);
408 		return -EOPNOTSUPP;
409 	}
410 
411 	nfp_map->use_map[off / 4].type = use;
412 
413 	return 0;
414 }
415 
416 static int
nfp_bpf_map_mark_used(struct bpf_verifier_env * env,struct nfp_insn_meta * meta,const struct bpf_reg_state * reg,enum nfp_bpf_map_use use)417 nfp_bpf_map_mark_used(struct bpf_verifier_env *env, struct nfp_insn_meta *meta,
418 		      const struct bpf_reg_state *reg,
419 		      enum nfp_bpf_map_use use)
420 {
421 	struct bpf_offloaded_map *offmap;
422 	struct nfp_bpf_map *nfp_map;
423 	unsigned int size, off;
424 	int i, err;
425 
426 	if (!tnum_is_const(reg->var_off)) {
427 		pr_vlog(env, "map value offset is variable\n");
428 		return -EOPNOTSUPP;
429 	}
430 
431 	off = reg->var_off.value + meta->insn.off + reg->off;
432 	size = BPF_LDST_BYTES(&meta->insn);
433 	offmap = map_to_offmap(reg->map_ptr);
434 	nfp_map = offmap->dev_priv;
435 
436 	if (off + size > offmap->map.value_size) {
437 		pr_vlog(env, "map value access out-of-bounds\n");
438 		return -EINVAL;
439 	}
440 
441 	for (i = 0; i < size; i += 4 - (off + i) % 4) {
442 		err = nfp_bpf_map_mark_used_one(env, nfp_map, off + i, use);
443 		if (err)
444 			return err;
445 	}
446 
447 	return 0;
448 }
449 
450 static int
nfp_bpf_check_ptr(struct nfp_prog * nfp_prog,struct nfp_insn_meta * meta,struct bpf_verifier_env * env,u8 reg_no)451 nfp_bpf_check_ptr(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta,
452 		  struct bpf_verifier_env *env, u8 reg_no)
453 {
454 	const struct bpf_reg_state *reg = cur_regs(env) + reg_no;
455 	int err;
456 
457 	if (reg->type != PTR_TO_CTX &&
458 	    reg->type != PTR_TO_STACK &&
459 	    reg->type != PTR_TO_MAP_VALUE &&
460 	    reg->type != PTR_TO_PACKET) {
461 		pr_vlog(env, "unsupported ptr type: %d\n", reg->type);
462 		return -EINVAL;
463 	}
464 
465 	if (reg->type == PTR_TO_STACK) {
466 		err = nfp_bpf_check_stack_access(nfp_prog, meta, reg, env);
467 		if (err)
468 			return err;
469 	}
470 
471 	if (reg->type == PTR_TO_MAP_VALUE) {
472 		if (is_mbpf_load(meta)) {
473 			err = nfp_bpf_map_mark_used(env, meta, reg,
474 						    NFP_MAP_USE_READ);
475 			if (err)
476 				return err;
477 		}
478 		if (is_mbpf_store(meta)) {
479 			pr_vlog(env, "map writes not supported\n");
480 			return -EOPNOTSUPP;
481 		}
482 		if (is_mbpf_xadd(meta)) {
483 			err = nfp_bpf_map_mark_used(env, meta, reg,
484 						    NFP_MAP_USE_ATOMIC_CNT);
485 			if (err)
486 				return err;
487 		}
488 	}
489 
490 	if (meta->ptr.type != NOT_INIT && meta->ptr.type != reg->type) {
491 		pr_vlog(env, "ptr type changed for instruction %d -> %d\n",
492 			meta->ptr.type, reg->type);
493 		return -EINVAL;
494 	}
495 
496 	meta->ptr = *reg;
497 
498 	return 0;
499 }
500 
501 static int
nfp_bpf_check_store(struct nfp_prog * nfp_prog,struct nfp_insn_meta * meta,struct bpf_verifier_env * env)502 nfp_bpf_check_store(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta,
503 		    struct bpf_verifier_env *env)
504 {
505 	const struct bpf_reg_state *reg = cur_regs(env) + meta->insn.dst_reg;
506 
507 	if (reg->type == PTR_TO_CTX) {
508 		if (nfp_prog->type == BPF_PROG_TYPE_XDP) {
509 			/* XDP ctx accesses must be 4B in size */
510 			switch (meta->insn.off) {
511 			case offsetof(struct xdp_md, rx_queue_index):
512 				if (nfp_prog->bpf->queue_select)
513 					goto exit_check_ptr;
514 				pr_vlog(env, "queue selection not supported by FW\n");
515 				return -EOPNOTSUPP;
516 			}
517 		}
518 		pr_vlog(env, "unsupported store to context field\n");
519 		return -EOPNOTSUPP;
520 	}
521 exit_check_ptr:
522 	return nfp_bpf_check_ptr(nfp_prog, meta, env, meta->insn.dst_reg);
523 }
524 
525 static int
nfp_bpf_check_xadd(struct nfp_prog * nfp_prog,struct nfp_insn_meta * meta,struct bpf_verifier_env * env)526 nfp_bpf_check_xadd(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta,
527 		   struct bpf_verifier_env *env)
528 {
529 	const struct bpf_reg_state *sreg = cur_regs(env) + meta->insn.src_reg;
530 	const struct bpf_reg_state *dreg = cur_regs(env) + meta->insn.dst_reg;
531 
532 	if (dreg->type != PTR_TO_MAP_VALUE) {
533 		pr_vlog(env, "atomic add not to a map value pointer: %d\n",
534 			dreg->type);
535 		return -EOPNOTSUPP;
536 	}
537 	if (sreg->type != SCALAR_VALUE) {
538 		pr_vlog(env, "atomic add not of a scalar: %d\n", sreg->type);
539 		return -EOPNOTSUPP;
540 	}
541 
542 	meta->xadd_over_16bit |=
543 		sreg->var_off.value > 0xffff || sreg->var_off.mask > 0xffff;
544 	meta->xadd_maybe_16bit |=
545 		(sreg->var_off.value & ~sreg->var_off.mask) <= 0xffff;
546 
547 	return nfp_bpf_check_ptr(nfp_prog, meta, env, meta->insn.dst_reg);
548 }
549 
550 static int
nfp_bpf_check_alu(struct nfp_prog * nfp_prog,struct nfp_insn_meta * meta,struct bpf_verifier_env * env)551 nfp_bpf_check_alu(struct nfp_prog *nfp_prog, struct nfp_insn_meta *meta,
552 		  struct bpf_verifier_env *env)
553 {
554 	const struct bpf_reg_state *sreg =
555 		cur_regs(env) + meta->insn.src_reg;
556 	const struct bpf_reg_state *dreg =
557 		cur_regs(env) + meta->insn.dst_reg;
558 
559 	meta->umin_src = min(meta->umin_src, sreg->umin_value);
560 	meta->umax_src = max(meta->umax_src, sreg->umax_value);
561 	meta->umin_dst = min(meta->umin_dst, dreg->umin_value);
562 	meta->umax_dst = max(meta->umax_dst, dreg->umax_value);
563 
564 	/* NFP supports u16 and u32 multiplication.
565 	 *
566 	 * For ALU64, if either operand is beyond u32's value range, we reject
567 	 * it. One thing to note, if the source operand is BPF_K, then we need
568 	 * to check "imm" field directly, and we'd reject it if it is negative.
569 	 * Because for ALU64, "imm" (with s32 type) is expected to be sign
570 	 * extended to s64 which NFP mul doesn't support.
571 	 *
572 	 * For ALU32, it is fine for "imm" be negative though, because the
573 	 * result is 32-bits and there is no difference on the low halve of
574 	 * the result for signed/unsigned mul, so we will get correct result.
575 	 */
576 	if (is_mbpf_mul(meta)) {
577 		if (meta->umax_dst > U32_MAX) {
578 			pr_vlog(env, "multiplier is not within u32 value range\n");
579 			return -EINVAL;
580 		}
581 		if (mbpf_src(meta) == BPF_X && meta->umax_src > U32_MAX) {
582 			pr_vlog(env, "multiplicand is not within u32 value range\n");
583 			return -EINVAL;
584 		}
585 		if (mbpf_class(meta) == BPF_ALU64 &&
586 		    mbpf_src(meta) == BPF_K && meta->insn.imm < 0) {
587 			pr_vlog(env, "sign extended multiplicand won't be within u32 value range\n");
588 			return -EINVAL;
589 		}
590 	}
591 
592 	/* NFP doesn't have divide instructions, we support divide by constant
593 	 * through reciprocal multiplication. Given NFP support multiplication
594 	 * no bigger than u32, we'd require divisor and dividend no bigger than
595 	 * that as well.
596 	 *
597 	 * Also eBPF doesn't support signed divide and has enforced this on C
598 	 * language level by failing compilation. However LLVM assembler hasn't
599 	 * enforced this, so it is possible for negative constant to leak in as
600 	 * a BPF_K operand through assembly code, we reject such cases as well.
601 	 */
602 	if (is_mbpf_div(meta)) {
603 		if (meta->umax_dst > U32_MAX) {
604 			pr_vlog(env, "dividend is not within u32 value range\n");
605 			return -EINVAL;
606 		}
607 		if (mbpf_src(meta) == BPF_X) {
608 			if (meta->umin_src != meta->umax_src) {
609 				pr_vlog(env, "divisor is not constant\n");
610 				return -EINVAL;
611 			}
612 			if (meta->umax_src > U32_MAX) {
613 				pr_vlog(env, "divisor is not within u32 value range\n");
614 				return -EINVAL;
615 			}
616 		}
617 		if (mbpf_src(meta) == BPF_K && meta->insn.imm < 0) {
618 			pr_vlog(env, "divide by negative constant is not supported\n");
619 			return -EINVAL;
620 		}
621 	}
622 
623 	return 0;
624 }
625 
nfp_verify_insn(struct bpf_verifier_env * env,int insn_idx,int prev_insn_idx)626 int nfp_verify_insn(struct bpf_verifier_env *env, int insn_idx,
627 		    int prev_insn_idx)
628 {
629 	struct nfp_prog *nfp_prog = env->prog->aux->offload->dev_priv;
630 	struct nfp_insn_meta *meta = nfp_prog->verifier_meta;
631 
632 	meta = nfp_bpf_goto_meta(nfp_prog, meta, insn_idx);
633 	nfp_prog->verifier_meta = meta;
634 
635 	if (!nfp_bpf_supported_opcode(meta->insn.code)) {
636 		pr_vlog(env, "instruction %#02x not supported\n",
637 			meta->insn.code);
638 		return -EINVAL;
639 	}
640 
641 	if (meta->insn.src_reg >= MAX_BPF_REG ||
642 	    meta->insn.dst_reg >= MAX_BPF_REG) {
643 		pr_vlog(env, "program uses extended registers - jit hardening?\n");
644 		return -EINVAL;
645 	}
646 
647 	if (is_mbpf_helper_call(meta))
648 		return nfp_bpf_check_helper_call(nfp_prog, env, meta);
649 	if (meta->insn.code == (BPF_JMP | BPF_EXIT))
650 		return nfp_bpf_check_exit(nfp_prog, env);
651 
652 	if (is_mbpf_load(meta))
653 		return nfp_bpf_check_ptr(nfp_prog, meta, env,
654 					 meta->insn.src_reg);
655 	if (is_mbpf_store(meta))
656 		return nfp_bpf_check_store(nfp_prog, meta, env);
657 
658 	if (is_mbpf_xadd(meta))
659 		return nfp_bpf_check_xadd(nfp_prog, meta, env);
660 
661 	if (is_mbpf_alu(meta))
662 		return nfp_bpf_check_alu(nfp_prog, meta, env);
663 
664 	return 0;
665 }
666 
667 static int
nfp_assign_subprog_idx_and_regs(struct bpf_verifier_env * env,struct nfp_prog * nfp_prog)668 nfp_assign_subprog_idx_and_regs(struct bpf_verifier_env *env,
669 				struct nfp_prog *nfp_prog)
670 {
671 	struct nfp_insn_meta *meta;
672 	int index = 0;
673 
674 	list_for_each_entry(meta, &nfp_prog->insns, l) {
675 		if (nfp_is_subprog_start(meta))
676 			index++;
677 		meta->subprog_idx = index;
678 
679 		if (meta->insn.dst_reg >= BPF_REG_6 &&
680 		    meta->insn.dst_reg <= BPF_REG_9)
681 			nfp_prog->subprog[index].needs_reg_push = 1;
682 	}
683 
684 	if (index + 1 != nfp_prog->subprog_cnt) {
685 		pr_vlog(env, "BUG: number of processed BPF functions is not consistent (processed %d, expected %d)\n",
686 			index + 1, nfp_prog->subprog_cnt);
687 		return -EFAULT;
688 	}
689 
690 	return 0;
691 }
692 
nfp_bpf_get_stack_usage(struct nfp_prog * nfp_prog)693 static unsigned int nfp_bpf_get_stack_usage(struct nfp_prog *nfp_prog)
694 {
695 	struct nfp_insn_meta *meta = nfp_prog_first_meta(nfp_prog);
696 	unsigned int max_depth = 0, depth = 0, frame = 0;
697 	struct nfp_insn_meta *ret_insn[MAX_CALL_FRAMES];
698 	unsigned short frame_depths[MAX_CALL_FRAMES];
699 	unsigned short ret_prog[MAX_CALL_FRAMES];
700 	unsigned short idx = meta->subprog_idx;
701 
702 	/* Inspired from check_max_stack_depth() from kernel verifier.
703 	 * Starting from main subprogram, walk all instructions and recursively
704 	 * walk all callees that given subprogram can call. Since recursion is
705 	 * prevented by the kernel verifier, this algorithm only needs a local
706 	 * stack of MAX_CALL_FRAMES to remember callsites.
707 	 */
708 process_subprog:
709 	frame_depths[frame] = nfp_prog->subprog[idx].stack_depth;
710 	frame_depths[frame] = round_up(frame_depths[frame], STACK_FRAME_ALIGN);
711 	depth += frame_depths[frame];
712 	max_depth = max(max_depth, depth);
713 
714 continue_subprog:
715 	for (; meta != nfp_prog_last_meta(nfp_prog) && meta->subprog_idx == idx;
716 	     meta = nfp_meta_next(meta)) {
717 		if (!is_mbpf_pseudo_call(meta))
718 			continue;
719 
720 		/* We found a call to a subprogram. Remember instruction to
721 		 * return to and subprog id.
722 		 */
723 		ret_insn[frame] = nfp_meta_next(meta);
724 		ret_prog[frame] = idx;
725 
726 		/* Find the callee and start processing it. */
727 		meta = nfp_bpf_goto_meta(nfp_prog, meta,
728 					 meta->n + 1 + meta->insn.imm);
729 		idx = meta->subprog_idx;
730 		frame++;
731 		goto process_subprog;
732 	}
733 	/* End of for() loop means the last instruction of the subprog was
734 	 * reached. If we popped all stack frames, return; otherwise, go on
735 	 * processing remaining instructions from the caller.
736 	 */
737 	if (frame == 0)
738 		return max_depth;
739 
740 	depth -= frame_depths[frame];
741 	frame--;
742 	meta = ret_insn[frame];
743 	idx = ret_prog[frame];
744 	goto continue_subprog;
745 }
746 
nfp_bpf_insn_flag_zext(struct nfp_prog * nfp_prog,struct bpf_insn_aux_data * aux)747 static void nfp_bpf_insn_flag_zext(struct nfp_prog *nfp_prog,
748 				   struct bpf_insn_aux_data *aux)
749 {
750 	struct nfp_insn_meta *meta;
751 
752 	list_for_each_entry(meta, &nfp_prog->insns, l) {
753 		if (aux[meta->n].zext_dst)
754 			meta->flags |= FLAG_INSN_DO_ZEXT;
755 	}
756 }
757 
nfp_bpf_finalize(struct bpf_verifier_env * env)758 int nfp_bpf_finalize(struct bpf_verifier_env *env)
759 {
760 	struct bpf_subprog_info *info;
761 	struct nfp_prog *nfp_prog;
762 	unsigned int max_stack;
763 	struct nfp_net *nn;
764 	int i;
765 
766 	nfp_prog = env->prog->aux->offload->dev_priv;
767 	nfp_prog->subprog_cnt = env->subprog_cnt;
768 	nfp_prog->subprog = kcalloc(nfp_prog->subprog_cnt,
769 				    sizeof(nfp_prog->subprog[0]), GFP_KERNEL);
770 	if (!nfp_prog->subprog)
771 		return -ENOMEM;
772 
773 	nfp_assign_subprog_idx_and_regs(env, nfp_prog);
774 
775 	info = env->subprog_info;
776 	for (i = 0; i < nfp_prog->subprog_cnt; i++) {
777 		nfp_prog->subprog[i].stack_depth = info[i].stack_depth;
778 
779 		if (i == 0)
780 			continue;
781 
782 		/* Account for size of return address. */
783 		nfp_prog->subprog[i].stack_depth += REG_WIDTH;
784 		/* Account for size of saved registers, if necessary. */
785 		if (nfp_prog->subprog[i].needs_reg_push)
786 			nfp_prog->subprog[i].stack_depth += BPF_REG_SIZE * 4;
787 	}
788 
789 	nn = netdev_priv(env->prog->aux->offload->netdev);
790 	max_stack = nn_readb(nn, NFP_NET_CFG_BPF_STACK_SZ) * 64;
791 	nfp_prog->stack_size = nfp_bpf_get_stack_usage(nfp_prog);
792 	if (nfp_prog->stack_size > max_stack) {
793 		pr_vlog(env, "stack too large: program %dB > FW stack %dB\n",
794 			nfp_prog->stack_size, max_stack);
795 		return -EOPNOTSUPP;
796 	}
797 
798 	nfp_bpf_insn_flag_zext(nfp_prog, env->insn_aux_data);
799 	return 0;
800 }
801 
nfp_bpf_opt_replace_insn(struct bpf_verifier_env * env,u32 off,struct bpf_insn * insn)802 int nfp_bpf_opt_replace_insn(struct bpf_verifier_env *env, u32 off,
803 			     struct bpf_insn *insn)
804 {
805 	struct nfp_prog *nfp_prog = env->prog->aux->offload->dev_priv;
806 	struct bpf_insn_aux_data *aux_data = env->insn_aux_data;
807 	struct nfp_insn_meta *meta = nfp_prog->verifier_meta;
808 
809 	meta = nfp_bpf_goto_meta(nfp_prog, meta, aux_data[off].orig_idx);
810 	nfp_prog->verifier_meta = meta;
811 
812 	/* conditional jump to jump conversion */
813 	if (is_mbpf_cond_jump(meta) &&
814 	    insn->code == (BPF_JMP | BPF_JA | BPF_K)) {
815 		unsigned int tgt_off;
816 
817 		tgt_off = off + insn->off + 1;
818 
819 		if (!insn->off) {
820 			meta->jmp_dst = list_next_entry(meta, l);
821 			meta->jump_neg_op = false;
822 		} else if (meta->jmp_dst->n != aux_data[tgt_off].orig_idx) {
823 			pr_vlog(env, "branch hard wire at %d changes target %d -> %d\n",
824 				off, meta->jmp_dst->n,
825 				aux_data[tgt_off].orig_idx);
826 			return -EINVAL;
827 		}
828 		return 0;
829 	}
830 
831 	pr_vlog(env, "unsupported instruction replacement %hhx -> %hhx\n",
832 		meta->insn.code, insn->code);
833 	return -EINVAL;
834 }
835 
nfp_bpf_opt_remove_insns(struct bpf_verifier_env * env,u32 off,u32 cnt)836 int nfp_bpf_opt_remove_insns(struct bpf_verifier_env *env, u32 off, u32 cnt)
837 {
838 	struct nfp_prog *nfp_prog = env->prog->aux->offload->dev_priv;
839 	struct bpf_insn_aux_data *aux_data = env->insn_aux_data;
840 	struct nfp_insn_meta *meta = nfp_prog->verifier_meta;
841 	unsigned int i;
842 
843 	meta = nfp_bpf_goto_meta(nfp_prog, meta, aux_data[off].orig_idx);
844 
845 	for (i = 0; i < cnt; i++) {
846 		if (WARN_ON_ONCE(&meta->l == &nfp_prog->insns))
847 			return -EINVAL;
848 
849 		/* doesn't count if it already has the flag */
850 		if (meta->flags & FLAG_INSN_SKIP_VERIFIER_OPT)
851 			i--;
852 
853 		meta->flags |= FLAG_INSN_SKIP_VERIFIER_OPT;
854 		meta = list_next_entry(meta, l);
855 	}
856 
857 	return 0;
858 }
859