Home
last modified time | relevance | path

Searched refs:recv_cq (Results 1 – 25 of 59) sorted by relevance

123

/kernel/linux/linux-5.10/drivers/infiniband/hw/hns/
Dhns_roce_qp.c209 struct ib_cq *send_cq, struct ib_cq *recv_cq) in add_qp_to_list() argument
215 hr_recv_cq = recv_cq ? to_hr_cq(recv_cq) : NULL; in add_qp_to_list()
246 init_attr->recv_cq); in hns_roce_qp_store()
1206 void hns_roce_lock_cqs(struct hns_roce_cq *send_cq, struct hns_roce_cq *recv_cq) in hns_roce_lock_cqs() argument
1207 __acquires(&send_cq->lock) __acquires(&recv_cq->lock) in hns_roce_lock_cqs()
1209 if (unlikely(send_cq == NULL && recv_cq == NULL)) { in hns_roce_lock_cqs()
1211 __acquire(&recv_cq->lock); in hns_roce_lock_cqs()
1212 } else if (unlikely(send_cq != NULL && recv_cq == NULL)) { in hns_roce_lock_cqs()
1214 __acquire(&recv_cq->lock); in hns_roce_lock_cqs()
1215 } else if (unlikely(send_cq == NULL && recv_cq != NULL)) { in hns_roce_lock_cqs()
[all …]
Dhns_roce_hw_v1.c364 hr_qp->ibqp.recv_cq)) { in hns_roce_v1_post_recv()
792 hr_qp->ibqp.recv_cq = cq; in hns_roce_v1_rsv_lp_qp()
2574 to_hr_cq(ibqp->recv_cq)->cqn); in hns_roce_v1_m_sqp()
2619 hns_roce_v1_cq_clean(to_hr_cq(ibqp->recv_cq), hr_qp->qpn, in hns_roce_v1_m_sqp()
2621 if (ibqp->send_cq != ibqp->recv_cq) in hns_roce_v1_m_sqp()
2752 to_hr_cq(ibqp->recv_cq)->cqn); in hns_roce_v1_m_qp()
2818 to_hr_cq(ibqp->recv_cq)->cqn); in hns_roce_v1_m_qp()
3243 hns_roce_v1_cq_clean(to_hr_cq(ibqp->recv_cq), hr_qp->qpn, in hns_roce_v1_m_qp()
3245 if (ibqp->send_cq != ibqp->recv_cq) in hns_roce_v1_m_qp()
3544 struct hns_roce_cq *send_cq, *recv_cq; in hns_roce_v1_destroy_qp() local
[all …]
/kernel/linux/linux-5.10/drivers/infiniband/core/
Duverbs_std_types_qp.c94 struct ib_cq *recv_cq = NULL; in UVERBS_HANDLER() local
181 recv_cq = uverbs_attr_get_obj(attrs, in UVERBS_HANDLER()
183 if (IS_ERR(recv_cq)) in UVERBS_HANDLER()
184 return PTR_ERR(recv_cq); in UVERBS_HANDLER()
235 attr.recv_cq = recv_cq; in UVERBS_HANDLER()
266 if (attr.recv_cq) in UVERBS_HANDLER()
267 atomic_inc(&attr.recv_cq->usecnt); in UVERBS_HANDLER()
Dcore_priv.h346 qp->recv_cq = attr->recv_cq; in _ib_create_qp()
Dverbs.c1170 qp->send_cq = qp->recv_cq = NULL; in create_xrc_qp_user()
1208 (qp_init_attr->recv_cq || in ib_create_qp()
1248 qp->recv_cq = NULL; in ib_create_qp()
1251 qp->recv_cq = qp_init_attr->recv_cq; in ib_create_qp()
1252 if (qp_init_attr->recv_cq) in ib_create_qp()
1253 atomic_inc(&qp_init_attr->recv_cq->usecnt); in ib_create_qp()
1941 rcq = qp->recv_cq; in ib_destroy_qp_user()
2740 struct ib_cq *cq = qp->recv_cq; in __ib_drain_rq()
2823 trace_cq_drain_complete(qp->recv_cq); in ib_drain_rq()
/kernel/linux/linux-5.10/drivers/infiniband/ulp/ipoib/
Dipoib_verbs.c178 priv->recv_cq = ib_create_cq(priv->ca, ipoib_ib_rx_completion, NULL, in ipoib_transport_dev_init()
180 if (IS_ERR(priv->recv_cq)) { in ipoib_transport_dev_init()
194 if (ib_req_notify_cq(priv->recv_cq, IB_CQ_NEXT_COMP)) in ipoib_transport_dev_init()
198 init_attr.recv_cq = priv->recv_cq; in ipoib_transport_dev_init()
247 ib_destroy_cq(priv->recv_cq); in ipoib_transport_dev_init()
267 ib_destroy_cq(priv->recv_cq); in ipoib_transport_dev_cleanup()
Dipoib_ethtool.c99 ret = rdma_set_cq_moderation(priv->recv_cq, in ipoib_set_coalesce()
Dipoib_ib.c466 n = ib_poll_cq(priv->recv_cq, t, priv->ibwc); in ipoib_rx_poll()
488 if (unlikely(ib_req_notify_cq(priv->recv_cq, in ipoib_rx_poll()
843 ib_req_notify_cq(priv->recv_cq, IB_CQ_NEXT_COMP); in ipoib_ib_dev_stop_default()
976 n = ib_poll_cq(priv->recv_cq, IPOIB_NUM_WC, priv->ibwc); in ipoib_drain_cq()
Dipoib_cm.c255 .send_cq = priv->recv_cq, /* For drain WR */ in ipoib_cm_create_rx_qp()
256 .recv_cq = priv->recv_cq, in ipoib_cm_create_rx_qp()
1062 .recv_cq = priv->recv_cq, in ipoib_cm_create_tx_qp()
/kernel/linux/linux-5.10/drivers/infiniband/hw/mthca/
Dmthca_qp.c774 qp_context->cqn_rcv = cpu_to_be32(to_mcq(ibqp->recv_cq)->cqn); in __mthca_modify_qp()
835 mthca_cq_clean(dev, to_mcq(qp->ibqp.recv_cq), qp->qpn, in __mthca_modify_qp()
837 if (qp->ibqp.send_cq != qp->ibqp.recv_cq) in __mthca_modify_qp()
1162 struct mthca_cq *recv_cq, in mthca_alloc_qp_common() argument
1291 struct mthca_cq *recv_cq, in mthca_alloc_qp() argument
1318 err = mthca_alloc_qp_common(dev, pd, send_cq, recv_cq, in mthca_alloc_qp()
1333 static void mthca_lock_cqs(struct mthca_cq *send_cq, struct mthca_cq *recv_cq) in mthca_lock_cqs() argument
1334 __acquires(&send_cq->lock) __acquires(&recv_cq->lock) in mthca_lock_cqs()
1336 if (send_cq == recv_cq) { in mthca_lock_cqs()
1338 __acquire(&recv_cq->lock); in mthca_lock_cqs()
[all …]
Dmthca_dev.h536 struct mthca_cq *recv_cq,
545 struct mthca_cq *recv_cq,
/kernel/linux/linux-5.10/drivers/infiniband/hw/mlx5/
Dqp.c75 struct mlx5_ib_cq **send_cq, struct mlx5_ib_cq **recv_cq);
745 struct mlx5_ib_cq *recv_cq);
747 struct mlx5_ib_cq *recv_cq);
1930 struct mlx5_ib_cq *recv_cq; in create_user_qp() local
2003 int rcqe_sz = mlx5_ib_get_cqe_size(init_attr->recv_cq); in create_user_qp()
2050 if (init_attr->recv_cq) in create_user_qp()
2051 MLX5_SET(qpc, qpc, cqn_rcv, to_mcq(init_attr->recv_cq)->mcq.cqn); in create_user_qp()
2085 get_cqs(qp->type, init_attr->send_cq, init_attr->recv_cq, in create_user_qp()
2086 &send_cq, &recv_cq); in create_user_qp()
2088 mlx5_ib_lock_cqs(send_cq, recv_cq); in create_user_qp()
[all …]
Dgsi.c163 gsi->rx_qp->recv_cq = hw_init_attr.recv_cq; in mlx5_ib_create_gsi()
228 .recv_cq = gsi->rx_qp->recv_cq, in create_gsi_ud_qp()
Dmem.c342 qp_init_attr.recv_cq = cq; in mlx5_ib_test_wc()
/kernel/linux/linux-5.10/drivers/infiniband/hw/mlx4/
Dqp.c53 struct mlx4_ib_cq *recv_cq);
55 struct mlx4_ib_cq *recv_cq);
922 to_mcq(init_attr->recv_cq)); in create_rq()
932 mcq = to_mcq(init_attr->recv_cq); in create_rq()
935 to_mcq(init_attr->recv_cq)); in create_rq()
1196 to_mcq(init_attr->recv_cq)); in create_qp_common()
1206 mcq = to_mcq(init_attr->recv_cq); in create_qp_common()
1209 to_mcq(init_attr->recv_cq)); in create_qp_common()
1263 static void mlx4_ib_lock_cqs(struct mlx4_ib_cq *send_cq, struct mlx4_ib_cq *recv_cq) in mlx4_ib_lock_cqs() argument
1264 __acquires(&send_cq->lock) __acquires(&recv_cq->lock) in mlx4_ib_lock_cqs()
[all …]
/kernel/linux/linux-5.10/net/sunrpc/xprtrdma/
Dverbs.c366 if (ep->re_attr.recv_cq) in rpcrdma_ep_destroy()
367 ib_free_cq(ep->re_attr.recv_cq); in rpcrdma_ep_destroy()
368 ep->re_attr.recv_cq = NULL; in rpcrdma_ep_destroy()
456 ep->re_attr.recv_cq = ib_alloc_cq_any(device, r_xprt, in rpcrdma_ep_create()
459 if (IS_ERR(ep->re_attr.recv_cq)) { in rpcrdma_ep_create()
460 rc = PTR_ERR(ep->re_attr.recv_cq); in rpcrdma_ep_create()
461 ep->re_attr.recv_cq = NULL; in rpcrdma_ep_create()
/kernel/linux/linux-5.10/drivers/infiniband/hw/vmw_pvrdma/
Dpvrdma_qp.c59 struct pvrdma_cq **recv_cq) in get_cqs() argument
62 *recv_cq = to_vcq(qp->ibqp.recv_cq); in get_cqs()
367 cmd->recv_cq_handle = to_vcq(init_attr->recv_cq)->cq_handle; in pvrdma_create_qp()
1051 init_attr->recv_cq = qp->ibqp.recv_cq; in pvrdma_query_qp()
/kernel/linux/linux-5.10/drivers/infiniband/sw/rxe/
Drxe_qp.c65 if (!init->recv_cq || !init->send_cq) { in rxe_qp_chk_init()
319 struct rxe_cq *rcq = to_rcq(init->recv_cq); in rxe_qp_from_init()
372 init->recv_cq = qp->ibqp.recv_cq; in rxe_qp_to_init()
/kernel/linux/linux-5.10/fs/cifs/
Dsmbdirect.c1381 ib_free_cq(info->recv_cq); in smbd_destroy()
1568 info->recv_cq = NULL; in _smbd_get_connection()
1577 info->recv_cq = in _smbd_get_connection()
1580 if (IS_ERR(info->recv_cq)) { in _smbd_get_connection()
1581 info->recv_cq = NULL; in _smbd_get_connection()
1596 qp_attr.recv_cq = info->recv_cq; in _smbd_get_connection()
1713 if (info->recv_cq) in _smbd_get_connection()
1714 ib_free_cq(info->recv_cq); in _smbd_get_connection()
Dsmbdirect.h59 struct ib_cq *send_cq, *recv_cq; member
/kernel/linux/linux-5.10/drivers/infiniband/ulp/srp/
Dib_srp.h150 struct ib_cq *recv_cq; member
Dib_srp.c534 struct ib_cq *recv_cq, *send_cq; in srp_create_ch_ib() local
545 recv_cq = ib_alloc_cq(dev->dev, ch, target->queue_size + 1, in srp_create_ch_ib()
547 if (IS_ERR(recv_cq)) { in srp_create_ch_ib()
548 ret = PTR_ERR(recv_cq); in srp_create_ch_ib()
567 init_attr->recv_cq = recv_cq; in srp_create_ch_ib()
602 if (ch->recv_cq) in srp_create_ch_ib()
603 ib_free_cq(ch->recv_cq); in srp_create_ch_ib()
608 ch->recv_cq = recv_cq; in srp_create_ch_ib()
630 ib_free_cq(recv_cq); in srp_create_ch_ib()
673 ib_free_cq(ch->recv_cq); in srp_free_ch_ib()
[all …]
/kernel/linux/linux-5.10/drivers/net/ethernet/ibm/ehea/
Dehea_main.c197 arr[i++].fwh = pr->recv_cq->fw_handle; in ehea_update_firmware_handles()
888 ehea_reset_cq_ep(pr->recv_cq); in ehea_poll()
890 ehea_reset_cq_n1(pr->recv_cq); in ehea_poll()
1465 pr->recv_cq = ehea_create_cq(adapter, pr_cfg->max_entries_rcq, in ehea_init_port_res()
1468 if (!pr->recv_cq) { in ehea_init_port_res()
1484 pr->recv_cq->attr.act_nr_of_cqes); in ehea_init_port_res()
1509 init_attr->recv_cq_handle = pr->recv_cq->fw_handle; in ehea_init_port_res()
1559 ehea_destroy_cq(pr->recv_cq); in ehea_init_port_res()
1576 ehea_destroy_cq(pr->recv_cq); in ehea_clean_portres()
Dehea.h351 struct ehea_cq *recv_cq; member
/kernel/linux/linux-5.10/drivers/infiniband/sw/siw/
Dsiw_verbs.c343 if (!attrs->send_cq || (!attrs->recv_cq && !attrs->srq)) { in siw_create_qp()
399 qp->rcq = to_siw_cq(attrs->recv_cq); in siw_create_qp()
533 qp_init_attr->recv_cq = base_qp->recv_cq; in siw_query_qp()

123