Home
last modified time | relevance | path

Searched refs:xdp_buff (Results 1 – 25 of 74) sorted by relevance

123

/drivers/net/ethernet/freescale/dpaa2/
Ddpaa2-xsk.c39 struct xdp_buff *xdp_buff; in dpaa2_xsk_run_xdp() local
50 xdp_buff = swa->xsk.xdp_buff; in dpaa2_xsk_run_xdp()
52 xdp_buff->data_hard_start = vaddr; in dpaa2_xsk_run_xdp()
53 xdp_buff->data = vaddr + dpaa2_fd_get_offset(fd); in dpaa2_xsk_run_xdp()
54 xdp_buff->data_end = xdp_buff->data + dpaa2_fd_get_len(fd); in dpaa2_xsk_run_xdp()
55 xdp_set_data_meta_invalid(xdp_buff); in dpaa2_xsk_run_xdp()
56 xdp_buff->rxq = &ch->xdp_rxq; in dpaa2_xsk_run_xdp()
58 xsk_buff_dma_sync_for_cpu(xdp_buff, ch->xsk_pool); in dpaa2_xsk_run_xdp()
59 xdp_act = bpf_prog_run_xdp(xdp_prog, xdp_buff); in dpaa2_xsk_run_xdp()
62 dpaa2_fd_set_offset(fd, xdp_buff->data - vaddr); in dpaa2_xsk_run_xdp()
[all …]
/drivers/net/ethernet/broadcom/bnxt/
Dbnxt_xdp.h18 struct xdp_buff *xdp);
21 struct xdp_buff xdp, struct page *page, u8 **data_ptr,
31 struct xdp_buff *xdp);
33 struct xdp_buff *xdp);
36 struct xdp_buff *xdp,
Dbnxt_xdp.c28 struct xdp_buff *xdp) in bnxt_xmit_bd()
97 struct xdp_buff *xdp) in __bnxt_xmit_xdp()
184 struct xdp_buff *xdp) in bnxt_xdp_buff_init()
204 struct xdp_buff *xdp) in bnxt_xdp_buff_frags_free()
225 struct xdp_buff xdp, struct page *page, u8 **data_ptr, in bnxt_rx_xdp()
465 struct page_pool *pool, struct xdp_buff *xdp, in bnxt_xdp_build_skb()
/drivers/net/ethernet/intel/i40e/
Di40e_xsk.c15 static struct xdp_buff **i40e_rx_bi(struct i40e_ring *rx_ring, u32 idx) in i40e_rx_bi()
197 static int i40e_run_xdp_zc(struct i40e_ring *rx_ring, struct xdp_buff *xdp, in i40e_run_xdp_zc()
244 struct xdp_buff **xdp; in i40e_alloc_rx_buffers_zc()
289 struct xdp_buff *xdp) in i40e_construct_skb_zc()
345 struct xdp_buff *xdp_buff, in i40e_handle_xdp_result_zc() argument
355 *rx_bytes = xdp_get_buff_len(xdp_buff); in i40e_handle_xdp_result_zc()
366 xsk_buff_free(xdp_buff); in i40e_handle_xdp_result_zc()
375 skb = i40e_construct_skb_zc(rx_ring, xdp_buff); in i40e_handle_xdp_result_zc()
400 i40e_add_xsk_frag(struct i40e_ring *rx_ring, struct xdp_buff *first, in i40e_add_xsk_frag()
401 struct xdp_buff *xdp, const unsigned int size) in i40e_add_xsk_frag()
[all …]
Di40e_trace.h165 struct xdp_buff *xdp),
193 struct xdp_buff *xdp),
201 struct xdp_buff *xdp),
Di40e_txrx.h335 struct xdp_buff **rx_bi_zc;
347 struct xdp_buff xdp;
Di40e_txrx_common.h9 int i40e_xmit_xdp_tx_ring(struct xdp_buff *xdp, struct i40e_ring *xdp_ring);
/drivers/net/ethernet/intel/ice/
Dice_xsk.c15 static struct xdp_buff **ice_xdp_buf(struct ice_rx_ring *rx_ring, u32 idx) in ice_xdp_buf()
451 static u16 ice_fill_rx_descs(struct xsk_buff_pool *pool, struct xdp_buff **xdp, in ice_fill_rx_descs()
488 struct xdp_buff **xdp; in __ice_alloc_rx_bufs_zc()
555 ice_construct_skb_zc(struct ice_rx_ring *rx_ring, struct xdp_buff *xdp) in ice_construct_skb_zc()
681 static int ice_xmit_xdp_tx_zc(struct xdp_buff *xdp, in ice_xmit_xdp_tx_zc()
689 struct xdp_buff *head; in ice_xmit_xdp_tx_zc()
762 ice_run_xdp_zc(struct ice_rx_ring *rx_ring, struct xdp_buff *xdp, in ice_run_xdp_zc()
806 ice_add_xsk_frag(struct ice_rx_ring *rx_ring, struct xdp_buff *first, in ice_add_xsk_frag()
807 struct xdp_buff *xdp, const unsigned int size) in ice_add_xsk_frag()
847 struct xdp_buff *first = NULL; in ice_clean_rx_irq_zc()
[all …]
Dice_txrx_lib.h18 ice_set_rx_bufs_act(struct xdp_buff *xdp, const struct ice_rx_ring *rx_ring, in ice_set_rx_bufs_act()
157 int ice_xmit_xdp_buff(struct xdp_buff *xdp, struct ice_tx_ring *xdp_ring);
158 int __ice_xmit_xdp_ring(struct xdp_buff *xdp, struct ice_tx_ring *xdp_ring,
Dice_txrx.h174 struct xdp_buff *xdp; /* used for XDP_TX ZC */
316 struct xdp_buff **xdp_buf;
318 struct xdp_buff xdp;
Dice_txrx.c385 struct xdp_buff *xdp = &rx_ring->xdp; in ice_clean_rx_ring()
535 ice_run_xdp(struct ice_rx_ring *rx_ring, struct xdp_buff *xdp, in ice_run_xdp()
585 struct xdp_buff xdp; in ice_xmit_xdp_ring()
846 ice_add_xdp_frag(struct ice_rx_ring *rx_ring, struct xdp_buff *xdp, in ice_add_xdp_frag()
950 ice_build_skb(struct ice_rx_ring *rx_ring, struct xdp_buff *xdp) in ice_build_skb()
1003 ice_construct_skb(struct ice_rx_ring *rx_ring, struct xdp_buff *xdp) in ice_construct_skb()
1120 struct xdp_buff *xdp = &rx_ring->xdp; in ice_clean_rx_irq()
/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/
Drx.c11 static struct mlx5e_xdp_buff *xsk_buff_to_mxbuf(struct xdp_buff *xdp) in xsk_buff_to_mxbuf()
25 struct xdp_buff **xsk_buffs; in mlx5e_xsk_alloc_rx_mpwqe()
34 xsk_buffs = (struct xdp_buff **)wi->alloc_units.xsk_buffs; in mlx5e_xsk_alloc_rx_mpwqe()
163 struct xdp_buff **buffs; in mlx5e_xsk_alloc_rx_wqes_batched()
225 static struct sk_buff *mlx5e_xsk_construct_skb(struct mlx5e_rq *rq, struct xdp_buff *xdp) in mlx5e_xsk_construct_skb()
/drivers/net/vmxnet3/
Dvmxnet3_xdp.c251 vmxnet3_run_xdp(struct vmxnet3_rx_queue *rq, struct xdp_buff *xdp, in vmxnet3_run_xdp()
304 const struct xdp_buff *xdp) in vmxnet3_build_skb()
331 struct xdp_buff xdp; in vmxnet3_process_xdp_small()
377 struct xdp_buff xdp; in vmxnet3_process_xdp()
/drivers/net/ethernet/netronome/nfp/
Dnfp_net_xsk.c18 struct xdp_buff *xdp) in nfp_net_xsk_rx_bufs_stash()
62 struct xdp_buff *xdp; in nfp_net_xsk_rx_ring_fill_freelist()
/drivers/net/ethernet/freescale/enetc/
Denetc.c1420 struct xdp_buff *xdp_buff, u16 size) in enetc_map_rx_buff_to_xdp() argument
1428 xdp_prepare_buff(xdp_buff, hard_start - rx_ring->buffer_offset, in enetc_map_rx_buff_to_xdp()
1433 u16 size, struct xdp_buff *xdp_buff) in enetc_add_rx_buff_to_xdp() argument
1435 struct skb_shared_info *shinfo = xdp_get_shared_info_from_buff(xdp_buff); in enetc_add_rx_buff_to_xdp()
1442 if (!xdp_buff_has_frags(xdp_buff)) { in enetc_add_rx_buff_to_xdp()
1443 xdp_buff_set_frags_flag(xdp_buff); in enetc_add_rx_buff_to_xdp()
1451 xdp_buff_set_frag_pfmemalloc(xdp_buff); in enetc_add_rx_buff_to_xdp()
1462 int *cleaned_cnt, struct xdp_buff *xdp_buff) in enetc_build_xdp_buff() argument
1466 xdp_init_buff(xdp_buff, ENETC_RXB_TRUESIZE, &rx_ring->xdp.rxq); in enetc_build_xdp_buff()
1468 enetc_map_rx_buff_to_xdp(rx_ring, *i, xdp_buff, size); in enetc_build_xdp_buff()
[all …]
/drivers/net/ethernet/engleder/
Dtsnep.h109 struct xdp_buff *xdp;
132 struct xdp_buff **xdp_batch;
/drivers/net/ethernet/stmicro/stmmac/
Dstmmac.h90 struct xdp_buff *xdp;
97 struct xdp_buff xdp;
/drivers/net/ethernet/netronome/nfp/nfd3/
Dnfd3.h72 struct xdp_buff *xdp;
/drivers/net/ethernet/google/gve/
Dgve_rx.c616 struct xdp_buff *xdp; in gve_xsk_pool_redirect()
637 struct xdp_buff *orig, struct bpf_prog *xdp_prog) in gve_xdp_redirect()
641 struct xdp_buff new; in gve_xdp_redirect()
670 struct xdp_buff *xdp, struct bpf_prog *xprog, in gve_xdp_done()
725 struct xdp_buff xdp; in gve_rx()
/drivers/net/ethernet/aquantia/atlantic/
Daq_ring.c24 struct xdp_buff *xdp) in aq_get_rxpages_xdp()
386 static struct sk_buff *aq_xdp_build_skb(struct xdp_buff *xdp, in aq_xdp_build_skb()
406 struct xdp_buff *xdp, in aq_xdp_run_prog()
487 struct xdp_buff *xdp) in aq_add_rx_fragment()
720 struct xdp_buff xdp; in __aq_ring_xdp_clean()
/drivers/net/ethernet/microchip/lan966x/
Dlan966x_xdp.c79 struct xdp_buff xdp; in lan966x_xdp_run()
/drivers/net/ethernet/amazon/ena/
Dena_xdp.h80 static inline int ena_xdp_execute(struct ena_ring *rx_ring, struct xdp_buff *xdp) in ena_xdp_execute()
/drivers/net/ethernet/intel/igc/
Digc.h526 struct xdp_buff *xdp;
532 struct xdp_buff xdp;
/drivers/net/ethernet/mellanox/mlx5/core/en/
Dxdp.h48 struct xdp_buff xdp;
/drivers/net/
Dveth.c122 struct xdp_buff xdp;
592 static int veth_xdp_tx(struct veth_rq *rq, struct xdp_buff *xdp, in veth_xdp_tx()
620 struct xdp_buff *xdp = &vxbuf.xdp; in veth_xdp_rcv_one()
709 static void veth_xdp_get(struct xdp_buff *xdp) in veth_xdp_get()
723 struct xdp_buff *xdp, in veth_convert_skb_to_xdp_buff()
831 struct xdp_buff *xdp = &vxbuf.xdp; in veth_xdp_rcv_skb()

123