Searched refs:mlx5_wqe_data_seg (Results 1 – 16 of 16) sorted by relevance
145 struct mlx5_wqe_data_seg *dseg = in mlx5e_xdp_mpwqe_add_dseg()146 (struct mlx5_wqe_data_seg *)session->wqe + session->ds_count; in mlx5e_xdp_mpwqe_add_dseg()
321 struct mlx5_wqe_data_seg *dseg = wqe->data; in mlx5e_xmit_xdp_frame()
243 srq->msrq.max_gs * sizeof(struct mlx5_wqe_data_seg); in mlx5_ib_create_srq()253 sizeof(struct mlx5_wqe_data_seg); in mlx5_ib_create_srq()428 struct mlx5_wqe_data_seg *scat; in mlx5_ib_post_srq_recv()461 scat = (struct mlx5_wqe_data_seg *)(next + 1); in mlx5_ib_post_srq_recv()
161 static void set_data_ptr_seg(struct mlx5_wqe_data_seg *dseg, struct ib_sge *sg) in set_data_ptr_seg()438 static void set_reg_data_seg(struct mlx5_wqe_data_seg *dseg, in set_reg_data_seg()904 *seg += sizeof(struct mlx5_wqe_data_seg); in set_reg_wr()905 *size += (sizeof(struct mlx5_wqe_data_seg) / 16); in set_reg_wr()1411 (struct mlx5_wqe_data_seg *)seg, in mlx5_ib_post_send()1413 size += sizeof(struct mlx5_wqe_data_seg) / 16; in mlx5_ib_post_send()1414 seg += sizeof(struct mlx5_wqe_data_seg); in mlx5_ib_post_send()1463 struct mlx5_wqe_data_seg *scat; in mlx5_ib_post_recv()
376 sizeof(struct mlx5_wqe_data_seg) < in set_rq_size()381 sizeof(struct mlx5_wqe_data_seg) - in set_rq_size()388 wqe_size += cap->max_recv_sge * sizeof(struct mlx5_wqe_data_seg); in set_rq_size()403 sizeof(struct mlx5_wqe_data_seg) - in set_rq_size()478 size += attr->cap.max_send_sge * sizeof(struct mlx5_wqe_data_seg); in calc_send_wqe()494 sizeof(struct mlx5_wqe_data_seg); in get_send_sge()500 sizeof(struct mlx5_wqe_data_seg); in get_send_sge()503 sizeof(struct mlx5_wqe_data_seg); in get_send_sge()506 sizeof(struct mlx5_wqe_data_seg)); in get_send_sge()
1074 struct mlx5_wqe_data_seg *dseg = wqe; in pagefault_data_segments()
946 sizeof(struct mlx5_wqe_data_seg); in mlx5_ib_query_device()950 sizeof(struct mlx5_wqe_data_seg); in mlx5_ib_query_device()
113 struct mlx5_wqe_data_seg data[];
215 struct mlx5_wqe_data_seg *dseg) in mlx5e_txwqe_build_dsegs()417 struct mlx5_wqe_data_seg *dseg; in mlx5e_sq_xmit_wqe()512 struct mlx5_wqe_data_seg *dseg; in mlx5e_tx_mpwqe_add_dseg()514 dseg = (struct mlx5_wqe_data_seg *)session->wqe + session->ds_count; in mlx5e_tx_mpwqe_add_dseg()932 struct mlx5_wqe_data_seg *dseg; in mlx5i_sq_xmit()
202 struct mlx5_wqe_data_seg data[0];207 struct mlx5_wqe_data_seg data[];211 struct mlx5_wqe_data_seg data[0];
1521 struct mlx5_wqe_data_seg *dseg; in mlx5e_open_xdpsq()1531 dseg = (struct mlx5_wqe_data_seg *)cseg + (ds_cnt - 1); in mlx5e_open_xdpsq()2133 int sz = sizeof(struct mlx5_wqe_data_seg) * ndsegs; in mlx5e_get_rqwq_log_stride()
9 struct mlx5_wqe_data_seg data;303 struct mlx5_wqe_data_seg *dseg; in tx_post_resync_dump()
352 struct mlx5_wqe_data_seg { struct
102 struct mlx5_wqe_data_seg *data; in mlx5_fpga_conn_post_recv()145 struct mlx5_wqe_data_seg *data; in mlx5_fpga_conn_post_send()
229 struct mlx5_wqe_data_seg *wq_dseg; in dr_rdma_segments()
366 num_ent * sizeof(struct mlx5_wqe_data_seg), &vqp->frag_buf, in rq_buf_alloc()