/drivers/net/ethernet/freescale/dpaa2/ |
D | dpaa2-xsk.c | 39 struct xdp_buff *xdp_buff; in dpaa2_xsk_run_xdp() local 50 xdp_buff = swa->xsk.xdp_buff; in dpaa2_xsk_run_xdp() 52 xdp_buff->data_hard_start = vaddr; in dpaa2_xsk_run_xdp() 53 xdp_buff->data = vaddr + dpaa2_fd_get_offset(fd); in dpaa2_xsk_run_xdp() 54 xdp_buff->data_end = xdp_buff->data + dpaa2_fd_get_len(fd); in dpaa2_xsk_run_xdp() 55 xdp_set_data_meta_invalid(xdp_buff); in dpaa2_xsk_run_xdp() 56 xdp_buff->rxq = &ch->xdp_rxq; in dpaa2_xsk_run_xdp() 58 xsk_buff_dma_sync_for_cpu(xdp_buff, ch->xsk_pool); in dpaa2_xsk_run_xdp() 59 xdp_act = bpf_prog_run_xdp(xdp_prog, xdp_buff); in dpaa2_xsk_run_xdp() 62 dpaa2_fd_set_offset(fd, xdp_buff->data - vaddr); in dpaa2_xsk_run_xdp() [all …]
|
/drivers/net/ethernet/broadcom/bnxt/ |
D | bnxt_xdp.h | 18 struct xdp_buff *xdp); 21 struct xdp_buff xdp, struct page *page, u8 **data_ptr, 31 struct xdp_buff *xdp); 33 struct xdp_buff *xdp); 36 struct xdp_buff *xdp,
|
D | bnxt_xdp.c | 28 struct xdp_buff *xdp) in bnxt_xmit_bd() 97 struct xdp_buff *xdp) in __bnxt_xmit_xdp() 184 struct xdp_buff *xdp) in bnxt_xdp_buff_init() 204 struct xdp_buff *xdp) in bnxt_xdp_buff_frags_free() 225 struct xdp_buff xdp, struct page *page, u8 **data_ptr, in bnxt_rx_xdp() 465 struct page_pool *pool, struct xdp_buff *xdp, in bnxt_xdp_build_skb()
|
/drivers/net/ethernet/intel/i40e/ |
D | i40e_xsk.c | 15 static struct xdp_buff **i40e_rx_bi(struct i40e_ring *rx_ring, u32 idx) in i40e_rx_bi() 197 static int i40e_run_xdp_zc(struct i40e_ring *rx_ring, struct xdp_buff *xdp, in i40e_run_xdp_zc() 244 struct xdp_buff **xdp; in i40e_alloc_rx_buffers_zc() 289 struct xdp_buff *xdp) in i40e_construct_skb_zc() 345 struct xdp_buff *xdp_buff, in i40e_handle_xdp_result_zc() argument 355 *rx_bytes = xdp_get_buff_len(xdp_buff); in i40e_handle_xdp_result_zc() 366 xsk_buff_free(xdp_buff); in i40e_handle_xdp_result_zc() 375 skb = i40e_construct_skb_zc(rx_ring, xdp_buff); in i40e_handle_xdp_result_zc() 400 i40e_add_xsk_frag(struct i40e_ring *rx_ring, struct xdp_buff *first, in i40e_add_xsk_frag() 401 struct xdp_buff *xdp, const unsigned int size) in i40e_add_xsk_frag() [all …]
|
D | i40e_trace.h | 165 struct xdp_buff *xdp), 193 struct xdp_buff *xdp), 201 struct xdp_buff *xdp),
|
D | i40e_txrx.h | 335 struct xdp_buff **rx_bi_zc; 347 struct xdp_buff xdp;
|
D | i40e_txrx_common.h | 9 int i40e_xmit_xdp_tx_ring(struct xdp_buff *xdp, struct i40e_ring *xdp_ring);
|
/drivers/net/ethernet/intel/ice/ |
D | ice_xsk.c | 15 static struct xdp_buff **ice_xdp_buf(struct ice_rx_ring *rx_ring, u32 idx) in ice_xdp_buf() 451 static u16 ice_fill_rx_descs(struct xsk_buff_pool *pool, struct xdp_buff **xdp, in ice_fill_rx_descs() 488 struct xdp_buff **xdp; in __ice_alloc_rx_bufs_zc() 555 ice_construct_skb_zc(struct ice_rx_ring *rx_ring, struct xdp_buff *xdp) in ice_construct_skb_zc() 681 static int ice_xmit_xdp_tx_zc(struct xdp_buff *xdp, in ice_xmit_xdp_tx_zc() 689 struct xdp_buff *head; in ice_xmit_xdp_tx_zc() 762 ice_run_xdp_zc(struct ice_rx_ring *rx_ring, struct xdp_buff *xdp, in ice_run_xdp_zc() 806 ice_add_xsk_frag(struct ice_rx_ring *rx_ring, struct xdp_buff *first, in ice_add_xsk_frag() 807 struct xdp_buff *xdp, const unsigned int size) in ice_add_xsk_frag() 847 struct xdp_buff *first = NULL; in ice_clean_rx_irq_zc() [all …]
|
D | ice_txrx_lib.h | 18 ice_set_rx_bufs_act(struct xdp_buff *xdp, const struct ice_rx_ring *rx_ring, in ice_set_rx_bufs_act() 157 int ice_xmit_xdp_buff(struct xdp_buff *xdp, struct ice_tx_ring *xdp_ring); 158 int __ice_xmit_xdp_ring(struct xdp_buff *xdp, struct ice_tx_ring *xdp_ring,
|
D | ice_txrx.h | 174 struct xdp_buff *xdp; /* used for XDP_TX ZC */ 316 struct xdp_buff **xdp_buf; 318 struct xdp_buff xdp;
|
D | ice_txrx.c | 385 struct xdp_buff *xdp = &rx_ring->xdp; in ice_clean_rx_ring() 535 ice_run_xdp(struct ice_rx_ring *rx_ring, struct xdp_buff *xdp, in ice_run_xdp() 585 struct xdp_buff xdp; in ice_xmit_xdp_ring() 846 ice_add_xdp_frag(struct ice_rx_ring *rx_ring, struct xdp_buff *xdp, in ice_add_xdp_frag() 950 ice_build_skb(struct ice_rx_ring *rx_ring, struct xdp_buff *xdp) in ice_build_skb() 1003 ice_construct_skb(struct ice_rx_ring *rx_ring, struct xdp_buff *xdp) in ice_construct_skb() 1120 struct xdp_buff *xdp = &rx_ring->xdp; in ice_clean_rx_irq()
|
/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/ |
D | rx.c | 11 static struct mlx5e_xdp_buff *xsk_buff_to_mxbuf(struct xdp_buff *xdp) in xsk_buff_to_mxbuf() 25 struct xdp_buff **xsk_buffs; in mlx5e_xsk_alloc_rx_mpwqe() 34 xsk_buffs = (struct xdp_buff **)wi->alloc_units.xsk_buffs; in mlx5e_xsk_alloc_rx_mpwqe() 163 struct xdp_buff **buffs; in mlx5e_xsk_alloc_rx_wqes_batched() 225 static struct sk_buff *mlx5e_xsk_construct_skb(struct mlx5e_rq *rq, struct xdp_buff *xdp) in mlx5e_xsk_construct_skb()
|
/drivers/net/vmxnet3/ |
D | vmxnet3_xdp.c | 251 vmxnet3_run_xdp(struct vmxnet3_rx_queue *rq, struct xdp_buff *xdp, in vmxnet3_run_xdp() 304 const struct xdp_buff *xdp) in vmxnet3_build_skb() 331 struct xdp_buff xdp; in vmxnet3_process_xdp_small() 377 struct xdp_buff xdp; in vmxnet3_process_xdp()
|
/drivers/net/ethernet/netronome/nfp/ |
D | nfp_net_xsk.c | 18 struct xdp_buff *xdp) in nfp_net_xsk_rx_bufs_stash() 62 struct xdp_buff *xdp; in nfp_net_xsk_rx_ring_fill_freelist()
|
/drivers/net/ethernet/freescale/enetc/ |
D | enetc.c | 1420 struct xdp_buff *xdp_buff, u16 size) in enetc_map_rx_buff_to_xdp() argument 1428 xdp_prepare_buff(xdp_buff, hard_start - rx_ring->buffer_offset, in enetc_map_rx_buff_to_xdp() 1433 u16 size, struct xdp_buff *xdp_buff) in enetc_add_rx_buff_to_xdp() argument 1435 struct skb_shared_info *shinfo = xdp_get_shared_info_from_buff(xdp_buff); in enetc_add_rx_buff_to_xdp() 1442 if (!xdp_buff_has_frags(xdp_buff)) { in enetc_add_rx_buff_to_xdp() 1443 xdp_buff_set_frags_flag(xdp_buff); in enetc_add_rx_buff_to_xdp() 1451 xdp_buff_set_frag_pfmemalloc(xdp_buff); in enetc_add_rx_buff_to_xdp() 1462 int *cleaned_cnt, struct xdp_buff *xdp_buff) in enetc_build_xdp_buff() argument 1466 xdp_init_buff(xdp_buff, ENETC_RXB_TRUESIZE, &rx_ring->xdp.rxq); in enetc_build_xdp_buff() 1468 enetc_map_rx_buff_to_xdp(rx_ring, *i, xdp_buff, size); in enetc_build_xdp_buff() [all …]
|
/drivers/net/ethernet/engleder/ |
D | tsnep.h | 109 struct xdp_buff *xdp; 132 struct xdp_buff **xdp_batch;
|
/drivers/net/ethernet/stmicro/stmmac/ |
D | stmmac.h | 90 struct xdp_buff *xdp; 97 struct xdp_buff xdp;
|
/drivers/net/ethernet/netronome/nfp/nfd3/ |
D | nfd3.h | 72 struct xdp_buff *xdp;
|
/drivers/net/ethernet/google/gve/ |
D | gve_rx.c | 616 struct xdp_buff *xdp; in gve_xsk_pool_redirect() 637 struct xdp_buff *orig, struct bpf_prog *xdp_prog) in gve_xdp_redirect() 641 struct xdp_buff new; in gve_xdp_redirect() 670 struct xdp_buff *xdp, struct bpf_prog *xprog, in gve_xdp_done() 725 struct xdp_buff xdp; in gve_rx()
|
/drivers/net/ethernet/aquantia/atlantic/ |
D | aq_ring.c | 24 struct xdp_buff *xdp) in aq_get_rxpages_xdp() 386 static struct sk_buff *aq_xdp_build_skb(struct xdp_buff *xdp, in aq_xdp_build_skb() 406 struct xdp_buff *xdp, in aq_xdp_run_prog() 487 struct xdp_buff *xdp) in aq_add_rx_fragment() 720 struct xdp_buff xdp; in __aq_ring_xdp_clean()
|
/drivers/net/ethernet/microchip/lan966x/ |
D | lan966x_xdp.c | 79 struct xdp_buff xdp; in lan966x_xdp_run()
|
/drivers/net/ethernet/amazon/ena/ |
D | ena_xdp.h | 80 static inline int ena_xdp_execute(struct ena_ring *rx_ring, struct xdp_buff *xdp) in ena_xdp_execute()
|
/drivers/net/ethernet/intel/igc/ |
D | igc.h | 526 struct xdp_buff *xdp; 532 struct xdp_buff xdp;
|
/drivers/net/ethernet/mellanox/mlx5/core/en/ |
D | xdp.h | 48 struct xdp_buff xdp;
|
/drivers/net/ |
D | veth.c | 122 struct xdp_buff xdp; 592 static int veth_xdp_tx(struct veth_rq *rq, struct xdp_buff *xdp, in veth_xdp_tx() 620 struct xdp_buff *xdp = &vxbuf.xdp; in veth_xdp_rcv_one() 709 static void veth_xdp_get(struct xdp_buff *xdp) in veth_xdp_get() 723 struct xdp_buff *xdp, in veth_convert_skb_to_xdp_buff() 831 struct xdp_buff *xdp = &vxbuf.xdp; in veth_xdp_rcv_skb()
|