Home
last modified time | relevance | path

Searched refs:send_cq (Results 1 – 25 of 49) sorted by relevance

12

/drivers/infiniband/hw/hns/
Dhns_roce_qp.c184 struct ib_cq *scq = init_attr->send_cq; in get_least_load_bankid_for_qp()
292 struct ib_cq *send_cq, struct ib_cq *recv_cq) in add_qp_to_list() argument
297 hr_send_cq = send_cq ? to_hr_cq(send_cq) : NULL; in add_qp_to_list()
328 add_qp_to_list(hr_dev, hr_qp, init_attr->send_cq, in hns_roce_qp_store()
1358 void hns_roce_lock_cqs(struct hns_roce_cq *send_cq, struct hns_roce_cq *recv_cq) in hns_roce_lock_cqs() argument
1359 __acquires(&send_cq->lock) __acquires(&recv_cq->lock) in hns_roce_lock_cqs()
1361 if (unlikely(send_cq == NULL && recv_cq == NULL)) { in hns_roce_lock_cqs()
1362 __acquire(&send_cq->lock); in hns_roce_lock_cqs()
1364 } else if (unlikely(send_cq != NULL && recv_cq == NULL)) { in hns_roce_lock_cqs()
1365 spin_lock_irq(&send_cq->lock); in hns_roce_lock_cqs()
[all …]
Dhns_roce_hw_v1.c109 if (hns_roce_wq_overflow(&qp->sq, nreq, qp->ibqp.send_cq)) { in hns_roce_v1_post_send()
880 hr_qp->ibqp.send_cq = cq; in hns_roce_v1_rsv_lp_qp()
2649 to_hr_cq(ibqp->send_cq)->cqn); in hns_roce_v1_m_sqp()
2692 if (ibqp->send_cq != ibqp->recv_cq) in hns_roce_v1_m_sqp()
2693 hns_roce_v1_cq_clean(to_hr_cq(ibqp->send_cq), in hns_roce_v1_m_sqp()
2819 to_hr_cq(ibqp->send_cq)->cqn); in hns_roce_v1_m_qp()
2884 to_hr_cq(ibqp->send_cq)->cqn); in hns_roce_v1_m_qp()
3312 if (ibqp->send_cq != ibqp->recv_cq) in hns_roce_v1_m_qp()
3313 hns_roce_v1_cq_clean(to_hr_cq(ibqp->send_cq), in hns_roce_v1_m_qp()
3612 struct hns_roce_cq *send_cq, *recv_cq; in hns_roce_v1_destroy_qp() local
[all …]
Dhns_roce_hw_v2.c696 if (hns_roce_wq_overflow(&qp->sq, nreq, qp->ibqp.send_cq)) { in hns_roce_v2_post_send()
4127 hr_reg_write(context, QPC_TX_CQN, get_cqn(ibqp->send_cq)); in modify_qp_reset_to_init()
4156 hr_reg_write(context, QPC_TX_CQN, get_cqn(ibqp->send_cq)); in modify_qp_init_to_init()
4954 if (ibqp->send_cq) in clear_qp()
4955 hns_roce_v2_cq_clean(to_hr_cq(ibqp->send_cq), in clear_qp()
4958 if (ibqp->recv_cq && ibqp->recv_cq != ibqp->send_cq) in clear_qp()
5199 qp_init_attr->send_cq = ibqp->send_cq; in hns_roce_v2_query_qp()
5223 struct hns_roce_cq *send_cq, *recv_cq; in hns_roce_v2_destroy_qp_common() local
5237 send_cq = hr_qp->ibqp.send_cq ? to_hr_cq(hr_qp->ibqp.send_cq) : NULL; in hns_roce_v2_destroy_qp_common()
5241 hns_roce_lock_cqs(send_cq, recv_cq); in hns_roce_v2_destroy_qp_common()
[all …]
Dhns_roce_device.h1235 void hns_roce_lock_cqs(struct hns_roce_cq *send_cq,
1237 void hns_roce_unlock_cqs(struct hns_roce_cq *send_cq,
/drivers/infiniband/core/
Duverbs_std_types_qp.c95 struct ib_cq *send_cq = NULL; in UVERBS_HANDLER() local
168 send_cq = uverbs_attr_get_obj(attrs, in UVERBS_HANDLER()
170 if (IS_ERR(send_cq)) in UVERBS_HANDLER()
171 return PTR_ERR(send_cq); in UVERBS_HANDLER()
175 send_cq = uverbs_attr_get_obj(attrs, in UVERBS_HANDLER()
177 if (IS_ERR(send_cq)) in UVERBS_HANDLER()
178 return PTR_ERR(send_cq); in UVERBS_HANDLER()
234 attr.send_cq = send_cq; in UVERBS_HANDLER()
Dverbs.c1185 qp->send_cq = qp->recv_cq = NULL; in create_xrc_qp_user()
1237 qp->send_cq = attr->send_cq; in create_qp()
1251 qp->send_cq = attr->send_cq; in create_qp()
1311 if (qp->send_cq) in ib_qp_usecnt_inc()
1312 atomic_inc(&qp->send_cq->usecnt); in ib_qp_usecnt_inc()
1330 if (qp->send_cq) in ib_qp_usecnt_dec()
1331 atomic_dec(&qp->send_cq->usecnt); in ib_qp_usecnt_dec()
2757 struct ib_cq *cq = qp->send_cq; in __ib_drain_sq()
2850 trace_cq_drain_complete(qp->send_cq); in ib_drain_sq()
/drivers/infiniband/ulp/ipoib/
Dipoib_verbs.c187 priv->send_cq = ib_create_cq(priv->ca, ipoib_ib_tx_completion, NULL, in ipoib_transport_dev_init()
189 if (IS_ERR(priv->send_cq)) { in ipoib_transport_dev_init()
197 init_attr.send_cq = priv->send_cq; in ipoib_transport_dev_init()
218 if (ib_req_notify_cq(priv->send_cq, IB_CQ_NEXT_COMP)) in ipoib_transport_dev_init()
244 ib_destroy_cq(priv->send_cq); in ipoib_transport_dev_init()
266 ib_destroy_cq(priv->send_cq); in ipoib_transport_dev_cleanup()
Dipoib_ib.c439 n = ib_poll_cq(priv->send_cq, MAX_SEND_CQE, priv->send_wc); in poll_tx()
507 n = ib_poll_cq(priv->send_cq, MAX_SEND_CQE, priv->send_wc); in ipoib_tx_poll()
519 if (unlikely(ib_req_notify_cq(priv->send_cq, IB_CQ_NEXT_COMP | in ipoib_tx_poll()
649 if (ib_req_notify_cq(priv->send_cq, IB_CQ_NEXT_COMP | in ipoib_send()
Dipoib_cm.c255 .send_cq = priv->recv_cq, /* For drain WR */ in ipoib_cm_create_rx_qp()
770 rc = ib_req_notify_cq(priv->send_cq, IB_CQ_NEXT_COMP | in ipoib_cm_send()
1061 .send_cq = priv->send_cq, in ipoib_cm_create_tx_qp()
/drivers/infiniband/hw/mlx5/
Dqp.c75 struct mlx5_ib_cq **send_cq, struct mlx5_ib_cq **recv_cq);
744 static void mlx5_ib_lock_cqs(struct mlx5_ib_cq *send_cq,
746 static void mlx5_ib_unlock_cqs(struct mlx5_ib_cq *send_cq,
1213 static int get_sq_ts_format(struct mlx5_ib_dev *dev, struct mlx5_ib_cq *send_cq) in get_sq_ts_format() argument
1217 return get_ts_format(dev, send_cq, fr_supported(ts_cap), in get_sq_ts_format()
1221 static int get_qp_ts_format(struct mlx5_ib_dev *dev, struct mlx5_ib_cq *send_cq, in get_qp_ts_format() argument
1230 send_cq ? get_ts_format(dev, send_cq, fr_sup, rt_sup) : in get_qp_ts_format()
1507 to_mcq(init_attr->send_cq)); in create_raw_packet_qp()
1848 scqe_sz = mlx5_ib_get_cqe_size(init_attr->send_cq); in configure_requester_scat_cqe()
1994 struct mlx5_ib_cq *send_cq; in create_dci() local
[all …]
Dgsi.c50 struct ib_cq *gsi_cq = mqp->ibqp.send_cq; in generate_completions()
141 hw_init_attr.send_cq = gsi->cq; in mlx5_ib_create_gsi()
205 .send_cq = gsi->cq, in create_gsi_ud_qp()
Dmem.c247 qp_init_attr.send_cq = cq; in mlx5_ib_test_wc()
/drivers/infiniband/hw/mthca/
Dmthca_qp.c735 qp_context->cqn_snd = cpu_to_be32(to_mcq(ibqp->send_cq)->cqn); in __mthca_modify_qp()
837 if (qp->ibqp.send_cq != qp->ibqp.recv_cq) in __mthca_modify_qp()
838 mthca_cq_clean(dev, to_mcq(qp->ibqp.send_cq), qp->qpn, NULL); in __mthca_modify_qp()
1164 struct mthca_cq *send_cq, in mthca_alloc_qp_common() argument
1293 struct mthca_cq *send_cq, in mthca_alloc_qp() argument
1321 err = mthca_alloc_qp_common(dev, pd, send_cq, recv_cq, in mthca_alloc_qp()
1336 static void mthca_lock_cqs(struct mthca_cq *send_cq, struct mthca_cq *recv_cq) in mthca_lock_cqs() argument
1337 __acquires(&send_cq->lock) __acquires(&recv_cq->lock) in mthca_lock_cqs()
1339 if (send_cq == recv_cq) { in mthca_lock_cqs()
1340 spin_lock_irq(&send_cq->lock); in mthca_lock_cqs()
[all …]
Dmthca_dev.h535 struct mthca_cq *send_cq,
544 struct mthca_cq *send_cq,
Dmthca_provider.c509 to_mcq(init_attr->send_cq), in mthca_create_qp()
534 to_mcq(init_attr->send_cq), in mthca_create_qp()
/drivers/infiniband/hw/mlx4/
Dqp.c52 static void mlx4_ib_lock_cqs(struct mlx4_ib_cq *send_cq,
54 static void mlx4_ib_unlock_cqs(struct mlx4_ib_cq *send_cq,
727 if (init_attr->send_cq || init_attr->cap.max_send_wr) { in _mlx4_ib_create_qp_rss()
925 mlx4_ib_lock_cqs(to_mcq(init_attr->send_cq), in create_rq()
934 mcq = to_mcq(init_attr->send_cq); in create_rq()
938 mlx4_ib_unlock_cqs(to_mcq(init_attr->send_cq), in create_rq()
1199 mlx4_ib_lock_cqs(to_mcq(init_attr->send_cq), in create_qp_common()
1208 mcq = to_mcq(init_attr->send_cq); in create_qp_common()
1212 mlx4_ib_unlock_cqs(to_mcq(init_attr->send_cq), in create_qp_common()
1267 static void mlx4_ib_lock_cqs(struct mlx4_ib_cq *send_cq, struct mlx4_ib_cq *recv_cq) in mlx4_ib_lock_cqs() argument
[all …]
/drivers/net/ethernet/ibm/ehea/
Dehea_main.c196 arr[i++].fwh = pr->send_cq->fw_handle; in ehea_update_firmware_handles()
801 struct ehea_cq *send_cq = pr->send_cq; in ehea_proc_cqes() local
810 cqe = ehea_poll_cq(send_cq); in ehea_proc_cqes()
812 ehea_inc_cq(send_cq); in ehea_proc_cqes()
852 cqe = ehea_poll_cq(send_cq); in ehea_proc_cqes()
855 ehea_update_feca(send_cq, cqe_counter); in ehea_proc_cqes()
890 ehea_reset_cq_ep(pr->send_cq); in ehea_poll()
892 ehea_reset_cq_n1(pr->send_cq); in ehea_poll()
895 cqe_skb = ehea_poll_cq(pr->send_cq); in ehea_poll()
1474 pr->send_cq = ehea_create_cq(adapter, pr_cfg->max_entries_scq, in ehea_init_port_res()
[all …]
Dehea.h350 struct ehea_cq *send_cq; member
/drivers/infiniband/hw/vmw_pvrdma/
Dpvrdma_qp.c58 static inline void get_cqs(struct pvrdma_qp *qp, struct pvrdma_cq **send_cq, in get_cqs() argument
61 *send_cq = to_vcq(qp->ibqp.send_cq); in get_cqs()
359 cmd->send_cq_handle = to_vcq(init_attr->send_cq)->cq_handle; in pvrdma_create_qp()
1042 init_attr->send_cq = qp->ibqp.send_cq; in pvrdma_query_qp()
/drivers/infiniband/sw/rxe/
Drxe_qp.c76 if (!init->recv_cq || !init->send_cq) { in rxe_qp_chk_init()
333 struct rxe_cq *scq = to_rcq(init->send_cq); in rxe_qp_from_init()
385 init->send_cq = qp->ibqp.send_cq; in rxe_qp_to_init()
/drivers/infiniband/ulp/srp/
Dib_srp.c521 ib_process_cq_direct(ch->send_cq, -1); in srp_destroy_qp()
534 struct ib_cq *recv_cq, *send_cq; in srp_create_ch_ib() local
552 send_cq = ib_alloc_cq(dev->dev, ch, m * target->queue_size, in srp_create_ch_ib()
554 if (IS_ERR(send_cq)) { in srp_create_ch_ib()
555 ret = PTR_ERR(send_cq); in srp_create_ch_ib()
566 init_attr->send_cq = send_cq; in srp_create_ch_ib()
604 if (ch->send_cq) in srp_create_ch_ib()
605 ib_free_cq(ch->send_cq); in srp_create_ch_ib()
609 ch->send_cq = send_cq; in srp_create_ch_ib()
627 ib_free_cq(send_cq); in srp_create_ch_ib()
[all …]
Dib_srp.h149 struct ib_cq *send_cq; member
/drivers/infiniband/sw/siw/
Dsiw_verbs.c346 if (!attrs->send_cq || (!attrs->recv_cq && !attrs->srq)) { in siw_create_qp()
397 qp->scq = to_siw_cq(attrs->send_cq); in siw_create_qp()
530 qp_init_attr->send_cq = base_qp->send_cq; in siw_query_qp()
/drivers/infiniband/hw/qedr/
Dqedr_roce_cm.c62 dev->gsi_sqcq = get_qedr_cq(attrs->send_cq); in qedr_store_gsi_qp_cq()
139 cq = get_qedr_cq(attrs->send_cq); in qedr_destroy_gsi_cq()
/drivers/infiniband/hw/hfi1/
Dqp.c653 ib_cq_head(qp->ibqp.send_cq), in qp_iter_print()
654 ib_cq_tail(qp->ibqp.send_cq), in qp_iter_print()

12