Home
last modified time | relevance | path

Searched refs:wr_id (Results 1 – 25 of 50) sorted by relevance

12

/drivers/infiniband/ulp/ipoib/
Dipoib_ib.c138 priv->rx_wr.wr_id = id | IPOIB_OP_RECV; in ipoib_ib_post_receive()
231 unsigned int wr_id = wc->wr_id & ~IPOIB_OP_RECV; in ipoib_ib_handle_rx_wc() local
237 wr_id, wc->status); in ipoib_ib_handle_rx_wc()
239 if (unlikely(wr_id >= ipoib_recvq_size)) { in ipoib_ib_handle_rx_wc()
241 wr_id, ipoib_recvq_size); in ipoib_ib_handle_rx_wc()
245 skb = priv->rx_ring[wr_id].skb; in ipoib_ib_handle_rx_wc()
251 wc->status, wr_id, wc->vendor_err); in ipoib_ib_handle_rx_wc()
252 ipoib_ud_dma_unmap_rx(priv, priv->rx_ring[wr_id].mapping); in ipoib_ib_handle_rx_wc()
254 priv->rx_ring[wr_id].skb = NULL; in ipoib_ib_handle_rx_wc()
265 memcpy(mapping, priv->rx_ring[wr_id].mapping, in ipoib_ib_handle_rx_wc()
[all …]
Dipoib_cm.c73 .wr_id = IPOIB_CM_RX_DRAIN_WRID,
97 priv->cm.rx_wr.wr_id = id | IPOIB_OP_CM | IPOIB_OP_RECV; in ipoib_cm_post_receive_srq()
123 wr->wr_id = id | IPOIB_OP_CM | IPOIB_OP_RECV; in ipoib_cm_post_receive_nonsrq()
560 unsigned int wr_id = wc->wr_id & ~(IPOIB_OP_CM | IPOIB_OP_RECV); in ipoib_cm_handle_rx_wc() local
570 wr_id, wc->status); in ipoib_cm_handle_rx_wc()
572 if (unlikely(wr_id >= ipoib_recvq_size)) { in ipoib_cm_handle_rx_wc()
573 if (wr_id == (IPOIB_CM_RX_DRAIN_WRID & ~(IPOIB_OP_CM | IPOIB_OP_RECV))) { in ipoib_cm_handle_rx_wc()
581 wr_id, ipoib_recvq_size); in ipoib_cm_handle_rx_wc()
590 skb = rx_ring[wr_id].skb; in ipoib_cm_handle_rx_wc()
595 wc->status, wr_id, wc->vendor_err); in ipoib_cm_handle_rx_wc()
[all …]
/drivers/infiniband/ulp/srpt/
Dib_srpt.h144 static inline enum srpt_opcode opcode_from_wr_id(u64 wr_id) in opcode_from_wr_id() argument
146 return wr_id >> 32; in opcode_from_wr_id()
148 static inline u32 idx_from_wr_id(u64 wr_id) in idx_from_wr_id() argument
150 return (u32)wr_id; in idx_from_wr_id()
Dib_srpt.c782 wr.wr_id = encode_wr_id(SRPT_RECV, ioctx->ioctx.index); in srpt_post_recv()
824 wr.wr_id = encode_wr_id(SRPT_SEND, ioctx->ioctx.index); in srpt_post_send()
1388 static void srpt_handle_send_err_comp(struct srpt_rdma_ch *ch, u64 wr_id) in srpt_handle_send_err_comp() argument
1397 index = idx_from_wr_id(wr_id); in srpt_handle_send_err_comp()
1952 index = idx_from_wr_id(wc->wr_id); in srpt_process_rcv_completion()
1990 index = idx_from_wr_id(wc->wr_id); in srpt_process_send_completion()
1991 opcode = opcode_from_wr_id(wc->wr_id); in srpt_process_send_completion()
2005 srpt_handle_send_err_comp(ch, wc->wr_id); in srpt_process_send_completion()
2037 if (opcode_from_wr_id(wc[i].wr_id) == SRPT_RECV) in srpt_process_completion()
2864 wr.wr_id = encode_wr_id(i == n_rdma - 1 ? in srpt_perform_rdmas()
[all …]
/drivers/infiniband/hw/ehca/
Dehca_reqs.c55 static u64 replace_wr_id(u64 wr_id, u16 idx) in replace_wr_id() argument
59 ret = wr_id & ~QMAP_IDX_MASK; in replace_wr_id()
65 static u16 get_app_wr_id(u64 wr_id) in get_app_wr_id() argument
67 return wr_id & QMAP_IDX_MASK; in get_app_wr_id()
87 wqe_p->work_request_id = replace_wr_id(recv_wr->wr_id, rq_map_idx); in ehca_write_rwqe()
121 "send_flags=%x opcode=%x", idx, send_wr->wr_id, in trace_send_wr_ud()
177 wqe_p->work_request_id = replace_wr_id(send_wr->wr_id, sq_map_idx); in ehca_write_swqe()
179 qmap_entry->app_wr_id = get_app_wr_id(send_wr->wr_id); in ehca_write_swqe()
560 qmap_entry->app_wr_id = get_app_wr_id(recv_wr->wr_id); in internal_post_recv()
743 wc->wr_id = replace_wr_id(cqe->work_request_id, qmap_entry->app_wr_id); in ehca_poll_cq_one()
[all …]
/drivers/infiniband/hw/mlx4/
Dmad.c562 wr.wr_id = ((u64) tun_tx_ix) | MLX4_TUN_SET_WRID_QPN(dest_qpt); in mlx4_ib_send_to_slave()
1047 recv_wr.wr_id = (u64) index | MLX4_TUN_WRID_RECV | in mlx4_ib_post_pv_qp_buf()
1164 wr.wr_id = ((u64) wire_tx_ix) | MLX4_TUN_SET_WRID_QPN(src_qpnum); in mlx4_ib_send_to_wire()
1180 struct mlx4_ib_demux_pv_qp *tun_qp = &ctx->qp[MLX4_TUN_WRID_QPN(wc->wr_id)]; in mlx4_ib_multiplex_mad()
1181 int wr_ix = wc->wr_id & (MLX4_NUM_TUNNEL_BUFS - 1); in mlx4_ib_multiplex_mad()
1411 tun_qp = &ctx->qp[MLX4_TUN_WRID_QPN(wc.wr_id)]; in mlx4_ib_tunnel_comp_worker()
1417 wc.wr_id & in mlx4_ib_tunnel_comp_worker()
1421 "buf:%lld\n", wc.wr_id); in mlx4_ib_tunnel_comp_worker()
1426 wc.wr_id, wc.status); in mlx4_ib_tunnel_comp_worker()
1427 ib_destroy_ah(tun_qp->tx_ring[wc.wr_id & in mlx4_ib_tunnel_comp_worker()
[all …]
Dcq.c678 wc->wr_id = wq->wrid[wq->tail & (wq->wqe_cnt - 1)]; in mlx4_ib_poll_one()
683 wc->wr_id = srq->wrid[wqe_ctr]; in mlx4_ib_poll_one()
688 wc->wr_id = srq->wrid[wqe_ctr]; in mlx4_ib_poll_one()
693 wc->wr_id = wq->wrid[tail]; in mlx4_ib_poll_one()
Dmr.c243 wr.wr_id = mw_bind->wr_id; in mlx4_ib_bind_mw()
/drivers/infiniband/hw/qib/
Dqib_srq.c80 wqe->wr_id = wr->wr_id; in qib_post_srq_receive()
286 p->wr_id = wqe->wr_id; in qib_modify_srq()
Dqib_ruc.c117 wc.wr_id = wqe->wr_id; in qib_init_sge()
189 qp->r_wr_id = wqe->wr_id; in qib_get_rwqe()
554 wc.wr_id = qp->r_wr_id; in qib_ruc_loopback()
795 wc.wr_id = wqe->wr.wr_id; in qib_send_complete()
Dqib_cq.c81 wc->uqueue[head].wr_id = entry->wr_id; in qib_cq_enter()
/drivers/infiniband/hw/ipath/
Dipath_srq.c80 wqe->wr_id = wr->wr_id; in ipath_post_srq_receive()
290 p->wr_id = wqe->wr_id; in ipath_modify_srq()
Dipath_ruc.c146 wc.wr_id = wqe->wr_id; in ipath_init_sge()
212 qp->r_wr_id = wqe->wr_id; in ipath_get_rwqe()
436 wc.wr_id = qp->r_wr_id; in ipath_ruc_loopback()
714 wc.wr_id = wqe->wr.wr_id; in ipath_send_complete()
Dipath_ud.c149 wc.wr_id = wqe->wr_id; in ipath_ud_loopback()
557 wc.wr_id = qp->r_wr_id; in ipath_ud_rcv()
Dipath_cq.c81 wc->uqueue[head].wr_id = entry->wr_id; in ipath_cq_enter()
/drivers/infiniband/ulp/iser/
Diser_verbs.c755 rx_wr.wr_id = (unsigned long)ib_conn->login_resp_buf; in iser_post_recvl()
778 rx_wr->wr_id = (unsigned long)rx_desc; in iser_post_recvm()
813 send_wr.wr_id = (unsigned long)tx_desc; in iser_post_send()
861 tx_desc = (struct iser_tx_desc *) (unsigned long) wc.wr_id; in iser_drain_tx_cq()
871 wc.wr_id, wc.status, wc.vendor_err); in iser_drain_tx_cq()
895 desc = (struct iser_rx_desc *) (unsigned long) wc.wr_id; in iser_cq_tasklet_fn()
908 wc.wr_id, wc.status, wc.vendor_err); in iser_cq_tasklet_fn()
/drivers/net/ethernet/ibm/ehea/
Dehea_qmr.h101 u64 wr_id; member
140 u64 wr_id; /* work request ID */ member
163 u64 wr_id; /* work request ID from WQE */ member
Dehea_main.c489 rwqe->wr_id = EHEA_BMASK_SET(EHEA_WR_ID_TYPE, wqe_type) in ehea_refill_rq_def()
566 int skb_index = EHEA_BMASK_GET(EHEA_WR_ID_INDEX, cqe->wr_id); in get_skb_by_index()
781 swqe->wr_id = SWQE_RESTART_CHECK; in check_sqs()
820 if (cqe->wr_id == SWQE_RESTART_CHECK) { in ehea_proc_cqes()
843 if (likely(EHEA_BMASK_GET(EHEA_WR_ID_TYPE, cqe->wr_id) in ehea_proc_cqes()
846 index = EHEA_BMASK_GET(EHEA_WR_ID_INDEX, cqe->wr_id); in ehea_proc_cqes()
852 swqe_av += EHEA_BMASK_GET(EHEA_WR_ID_REFILL, cqe->wr_id); in ehea_proc_cqes()
2068 swqe->wr_id = EHEA_BMASK_SET(EHEA_WR_ID_TYPE, EHEA_SWQE3_TYPE) in ehea_start_xmit()
2071 swqe->wr_id |= EHEA_BMASK_SET(EHEA_WR_ID_REFILL, in ehea_start_xmit()
2078 swqe->wr_id = in ehea_start_xmit()
[all …]
/drivers/infiniband/hw/cxgb3/
Diwch_qp.c281 qhp->wq.rq_size_log2)].wr_id = wr->wr_id; in build_rdma_recv()
344 qhp->wq.rq_size_log2)].wr_id = wr->wr_id; in build_zero_stag_recv()
436 sqp->wr_id = wr->wr_id; in iwch_post_send()
447 __func__, (unsigned long long) wr->wr_id, idx, in iwch_post_send()
511 "wqe %p \n", __func__, (unsigned long long) wr->wr_id, in iwch_post_receive()
590 sqp->wr_id = mw_bind->wr_id; in iwch_bind_mw()
Diwch_cq.c81 wc->wr_id = cookie; in iwch_poll_cq_one()
/drivers/infiniband/core/
Dmad.c655 u64 wr_id, u16 slid, u16 pkey_index, u8 port_num, in build_smp_wc() argument
659 wc->wr_id = wr_id; in build_smp_wc()
735 send_wr->wr_id, be16_to_cpu(smp->dr_slid), in handle_outgoing_dr_smp()
904 mad_send_wr->send_wr.wr_id = (unsigned long) mad_send_wr; in ib_create_send_mad()
1016 mad_send_wr->send_wr.wr_id = (unsigned long)&mad_send_wr->mad_list; in ib_send_mad()
1829 mad_recv_wc->wc->wr_id = (unsigned long) &mad_send_wr->send_buf; in ib_mad_complete_recv()
1876 mad_list = (struct ib_mad_list_head *)(unsigned long)wc->wr_id; in ib_mad_recv_done_handler()
2130 mad_list = (struct ib_mad_list_head *)(unsigned long)wc->wr_id; in ib_mad_send_done_handler()
2203 mad_list = (struct ib_mad_list_head *)(unsigned long)wc->wr_id; in mad_error_handler()
2594 recv_wr.wr_id = (unsigned long)&mad_priv->header.mad_list; in ib_mad_post_receive_mads()
/drivers/infiniband/ulp/isert/
Dib_isert.c613 rx_wr->wr_id = (unsigned long)rx_desc; in isert_post_recv()
647 send_wr.wr_id = (unsigned long)tx_desc; in isert_post_send()
716 send_wr->wr_id = (unsigned long)&isert_cmd->tx_desc; in isert_init_send_wr()
739 rx_wr.wr_id = (unsigned long)isert_conn->login_req_buf; in isert_rdma_post_recvl()
1440 tx_desc = (struct iser_tx_desc *)(unsigned long)wc.wr_id; in isert_cq_tx_work()
1479 rx_desc = (struct iser_rx_desc *)(unsigned long)wc.wr_id; in isert_cq_rx_work()
1670 send_wr->wr_id = (unsigned long)&isert_cmd->tx_desc; in isert_build_rdma_wr()
/drivers/infiniband/hw/cxgb4/
Dcq.c528 *cookie = wq->sq.sw_sq[wq->sq.cidx].wr_id; in poll_cq()
532 *cookie = wq->rq.sw_rq[wq->rq.cidx].wr_id; in poll_cq()
591 wc->wr_id = cookie; in c4iw_poll_cq_one()
Dt4.h265 u64 wr_id; member
306 u64 wr_id; member
/drivers/infiniband/hw/mthca/
Dmthca_cq.c542 entry->wr_id = (*cur_qp)->wrid[wqe_index + in mthca_poll_one()
549 entry->wr_id = srq->wrid[wqe_index]; in mthca_poll_one()
563 entry->wr_id = (*cur_qp)->wrid[wqe_index]; in mthca_poll_one()

12