Home
last modified time | relevance | path

Searched refs:cqn (Results 1 – 25 of 53) sorted by relevance

123

/drivers/net/ethernet/mellanox/mlx4/
Dcq.c104 void mlx4_cq_completion(struct mlx4_dev *dev, u32 cqn) in mlx4_cq_completion() argument
110 cqn & (dev->caps.num_cqs - 1)); in mlx4_cq_completion()
114 mlx4_dbg(dev, "Completion event for bogus CQ %08x\n", cqn); in mlx4_cq_completion()
126 void mlx4_cq_event(struct mlx4_dev *dev, u32 cqn, int event_type) in mlx4_cq_event() argument
132 cq = radix_tree_lookup(&cq_table->tree, cqn & (dev->caps.num_cqs - 1)); in mlx4_cq_event()
136 mlx4_dbg(dev, "Async event for bogus CQ %08x\n", cqn); in mlx4_cq_event()
184 err = mlx4_MODIFY_CQ(dev, mailbox, cq->cqn, 1); in mlx4_cq_modify()
210 err = mlx4_MODIFY_CQ(dev, mailbox, cq->cqn, 0); in mlx4_cq_resize()
217 int __mlx4_cq_alloc_icm(struct mlx4_dev *dev, int *cqn) in __mlx4_cq_alloc_icm() argument
223 *cqn = mlx4_bitmap_alloc(&cq_table->bitmap); in __mlx4_cq_alloc_icm()
[all …]
Den_resources.c41 int is_tx, int rss, int qpn, int cqn, in mlx4_en_fill_qp_context() argument
72 context->cqn_send = cpu_to_be32(cqn); in mlx4_en_fill_qp_context()
73 context->cqn_recv = cpu_to_be32(cqn); in mlx4_en_fill_qp_context()
Dresource_tracker.c1671 static int cq_res_start_move_to(struct mlx4_dev *dev, int slave, int cqn, in cq_res_start_move_to() argument
1680 r = res_tracker_lookup(&tracker->res_tree[RES_CQ], cqn); in cq_res_start_move_to()
1938 int cqn; in cq_alloc_res() local
1947 err = __mlx4_cq_alloc_icm(dev, &cqn); in cq_alloc_res()
1953 err = add_res_range(dev, slave, cqn, 1, RES_CQ, 0); in cq_alloc_res()
1956 __mlx4_cq_free_icm(dev, cqn); in cq_alloc_res()
1960 set_param_l(out_param, cqn); in cq_alloc_res()
2464 int cqn; in cq_free_res() local
2469 cqn = get_param_l(&in_param); in cq_free_res()
2470 err = rem_res_range(dev, slave, cqn, 1, RES_CQ, 0); in cq_free_res()
[all …]
Dsrq.c162 int mlx4_srq_alloc(struct mlx4_dev *dev, u32 pdn, u32 cqn, u16 xrcd, in mlx4_srq_alloc() argument
192 srq_context->pg_offset_cqn = cpu_to_be32(cqn & 0xffffff); in mlx4_srq_alloc()
Deq.c498 int cqn; in mlx4_eq_int() local
521 cqn = be32_to_cpu(eqe->event.comp.cqn) & 0xffffff; in mlx4_eq_int()
522 mlx4_cq_completion(dev, cqn); in mlx4_eq_int()
688 be32_to_cpu(eqe->event.cq_err.cqn) & 0xffffff); in mlx4_eq_int()
692 be32_to_cpu(eqe->event.cq_err.cqn) in mlx4_eq_int()
707 be32_to_cpu(eqe->event.cq_err.cqn) in mlx4_eq_int()
/drivers/infiniband/hw/hns/
Dhns_roce_cq.c59 ret = hns_roce_bitmap_alloc(&cq_table->bitmap, &hr_cq->cqn); in alloc_cqc()
66 ret = hns_roce_table_get(hr_dev, &cq_table->table, hr_cq->cqn); in alloc_cqc()
69 hr_cq->cqn, ret); in alloc_cqc()
73 ret = xa_err(xa_store(&cq_table->array, hr_cq->cqn, hr_cq, GFP_KERNEL)); in alloc_cqc()
89 ret = hns_roce_cmd_mbox(hr_dev, mailbox->dma, 0, hr_cq->cqn, 0, in alloc_cqc()
95 hr_cq->cqn, ret); in alloc_cqc()
108 xa_erase(&cq_table->array, hr_cq->cqn); in alloc_cqc()
111 hns_roce_table_put(hr_dev, &cq_table->table, hr_cq->cqn); in alloc_cqc()
114 hns_roce_bitmap_free(&cq_table->bitmap, hr_cq->cqn, BITMAP_NO_RR); in alloc_cqc()
124 ret = hns_roce_cmd_mbox(hr_dev, 0, 0, hr_cq->cqn, 1, in free_cqc()
[all …]
Dhns_roce_hw_v2_dfx.c9 int hns_roce_v2_query_cqc_info(struct hns_roce_dev *hr_dev, u32 cqn, in hns_roce_v2_query_cqc_info() argument
21 ret = hns_roce_cmd_mbox(hr_dev, 0, mailbox->dma, cqn, 0, in hns_roce_v2_query_cqc_info()
Dhns_roce_srq.c81 u32 pdn, u32 cqn, u16 xrcd, u64 db_rec_addr) in alloc_srqc() argument
136 hr_dev->hw->write_srqc(hr_dev, srq, pdn, xrcd, cqn, mailbox->buf, in alloc_srqc()
293 u32 cqn; in hns_roce_create_srq() local
338 cqn = ib_srq_has_cq(init_attr->srq_type) ? in hns_roce_create_srq()
339 to_hr_cq(init_attr->ext.cq)->cqn : 0; in hns_roce_create_srq()
342 ret = alloc_srqc(hr_dev, srq, to_hr_pd(ib_srq->pd)->pdn, cqn, 0, 0); in hns_roce_create_srq()
Dhns_roce_device.h469 unsigned long cqn; member
621 u32 cqn; member
879 int (*query_cqc_info)(struct hns_roce_dev *hr_dev, u32 cqn,
944 struct hns_roce_srq *srq, u32 pdn, u16 xrcd, u32 cqn,
1276 void hns_roce_cq_completion(struct hns_roce_dev *hr_dev, u32 cqn);
1277 void hns_roce_cq_event(struct hns_roce_dev *hr_dev, u32 cqn, int event_type);
Dhns_roce_hw_v1.c2004 ROCEE_DB_OTHERS_H_ROCEE_DB_OTH_INP_H_S, hr_cq->cqn); in hns_roce_v1_cq_set_ci()
2078 offset = hr_cq->cqn * HNS_ROCE_V1_TPTR_ENTRY_SIZE; in hns_roce_v1_write_cqc()
2087 CQ_CONTEXT_CQC_BYTE_4_CQN_S, hr_cq->cqn); in hns_roce_v1_write_cqc()
2163 hr_cq->cqn | notification_flag); in hns_roce_v1_req_notify_cq()
2213 hr_cq->cqn, (qpn & HNS_ROCE_CQE_QPN_MASK)); in hns_roce_v1_poll_one()
2645 to_hr_cq(ibqp->recv_cq)->cqn); in hns_roce_v1_m_sqp()
2649 to_hr_cq(ibqp->send_cq)->cqn); in hns_roce_v1_m_sqp()
2819 to_hr_cq(ibqp->send_cq)->cqn); in hns_roce_v1_m_qp()
2823 to_hr_cq(ibqp->recv_cq)->cqn); in hns_roce_v1_m_qp()
2884 to_hr_cq(ibqp->send_cq)->cqn); in hns_roce_v1_m_qp()
[all …]
Dhns_roce_restrack.c95 ret = hr_dev->dfx->query_cqc_info(hr_dev, hr_cq->cqn, (int *)context); in hns_roce_fill_res_cq_entry()
Dhns_roce_hw_v2.c3068 V2_CQC_BYTE_8_CQN_S, hr_cq->cqn); in hns_roce_v2_write_cqc()
3142 hr_cq->cqn); in hns_roce_v2_req_notify_cq()
3340 hr_cq->cqn, qpn & HNS_ROCE_V2_CQE_QPN_MASK); in hns_roce_v2_poll_one()
3854 V2_QPC_BYTE_80_RX_CQN_S, to_hr_cq(ibqp->recv_cq)->cqn); in modify_qp_reset_to_init()
3867 V2_QPC_BYTE_252_TX_CQN_S, to_hr_cq(ibqp->send_cq)->cqn); in modify_qp_reset_to_init()
3939 V2_QPC_BYTE_80_RX_CQN_S, to_hr_cq(ibqp->recv_cq)->cqn); in modify_qp_init_to_init()
3944 V2_QPC_BYTE_252_TX_CQN_S, to_hr_cq(ibqp->send_cq)->cqn); in modify_qp_init_to_init()
5146 u32 cqn, void *mb_buf, u64 *mtts_wqe, in hns_roce_v2_write_srqc() argument
5224 cqn); in hns_roce_v2_write_srqc()
5357 ret = hns_roce_cmd_mbox(hr_dev, mailbox->dma, 0, hr_cq->cqn, 1, in hns_roce_v2_modify_cq()
[all …]
/drivers/infiniband/hw/mthca/
Dmthca_cq.c76 __be32 cqn; member
211 mthca_write64(MTHCA_TAVOR_CQ_DB_INC_CI | cq->cqn, incr - 1, in update_cons_index()
217 void mthca_cq_completion(struct mthca_dev *dev, u32 cqn) in mthca_cq_completion() argument
221 cq = mthca_array_get(&dev->cq_table.cq, cqn & (dev->limits.num_cqs - 1)); in mthca_cq_completion()
224 mthca_warn(dev, "Completion event for bogus CQ %08x\n", cqn); in mthca_cq_completion()
233 void mthca_cq_event(struct mthca_dev *dev, u32 cqn, in mthca_cq_event() argument
241 cq = mthca_array_get(&dev->cq_table.cq, cqn & (dev->limits.num_cqs - 1)); in mthca_cq_event()
248 mthca_warn(dev, "Async event for bogus CQ %08x\n", cqn); in mthca_cq_event()
297 qpn, cq->cqn, cq->cons_index, prod_index); in mthca_cq_clean()
384 cq->cqn, cq->cons_index); in handle_error_cqe()
[all …]
Dmthca_eq.c132 __be32 cqn; member
149 __be32 cqn; member
219 static inline void disarm_cq(struct mthca_dev *dev, int eqn, int cqn) in disarm_cq() argument
222 mthca_write64(MTHCA_EQ_DB_DISARM_CQ | eqn, cqn, in disarm_cq()
276 disarm_cqn = be32_to_cpu(eqe->event.comp.cqn) & 0xffffff; in mthca_eq_int()
343 be32_to_cpu(eqe->event.cq_err.cqn) & 0xffffff); in mthca_eq_int()
344 mthca_cq_event(dev, be32_to_cpu(eqe->event.cq_err.cqn), in mthca_eq_int()
/drivers/net/ethernet/mellanox/mlx5/core/
Dcq.c107 cq->cqn = MLX5_GET(create_cq_out, out, cqn); in mlx5_core_create_cq()
134 cq->cqn); in mlx5_core_create_cq()
145 MLX5_SET(destroy_cq_in, din, cqn, cq->cqn); in mlx5_core_create_cq()
163 MLX5_SET(destroy_cq_in, in, cqn, cq->cqn); in mlx5_core_destroy_cq()
183 MLX5_SET(query_cq_in, in, cqn, cq->cqn); in mlx5_core_query_cq()
207 MLX5_SET(modify_cq_in, in, cqn, cq->cqn); in mlx5_core_modify_cq_moderation()
Deq.c112 static struct mlx5_core_cq *mlx5_eq_cq_get(struct mlx5_eq *eq, u32 cqn) in mlx5_eq_cq_get() argument
118 cq = radix_tree_lookup(&table->tree, cqn); in mlx5_eq_cq_get()
135 u32 cqn = -1; in mlx5_eq_comp_int() local
149 cqn = be32_to_cpu(eqe->data.comp.cqn) & 0xffffff; in mlx5_eq_comp_int()
151 cq = mlx5_eq_cq_get(eq, cqn); in mlx5_eq_comp_int()
158 "Completion event for bogus CQ 0x%x\n", cqn); in mlx5_eq_comp_int()
168 if (cqn != -1) in mlx5_eq_comp_int()
425 err = radix_tree_insert(&table->tree, cq->cqn, cq); in mlx5_eq_add_cq()
437 tmp = radix_tree_delete(&table->tree, cq->cqn); in mlx5_eq_del_cq()
442 eq->eqn, cq->cqn); in mlx5_eq_del_cq()
[all …]
/drivers/infiniband/hw/mlx4/
Dsrq.c82 u32 cqn; in mlx4_ib_create_srq() local
176 cqn = ib_srq_has_cq(init_attr->srq_type) ? in mlx4_ib_create_srq()
177 to_mcq(init_attr->ext.cq)->mcq.cqn : 0; in mlx4_ib_create_srq()
181 err = mlx4_srq_alloc(dev->dev, to_mpd(ib_srq->pd)->pdn, cqn, xrcdn, in mlx4_ib_create_srq()
/drivers/infiniband/hw/mlx5/
Dcq.c57 type, mcq->cqn); in mlx5_ib_cq_event()
520 "Requestor" : "Responder", cq->mcq.cqn); in mlx5_poll_one()
557 cq->mcq.cqn, sig->err_item.key, in mlx5_poll_one()
583 cq->mcq.cqn); in poll_soft_wc()
993 mlx5_ib_dbg(dev, "cqn 0x%x\n", cq->mcq.cqn); in mlx5_ib_create_cq()
1003 if (ib_copy_to_udata(udata, &cq->mcq.cqn, sizeof(__u32))) { in mlx5_ib_create_cq()
1122 mlx5_ib_warn(dev, "modify cq 0x%x failed\n", mcq->mcq.cqn); in mlx5_ib_modify_cq()
1233 cq->mcq.cqn); in copy_resize_cqes()
1326 MLX5_SET(modify_cq_in, in, cqn, cq->mcq.cqn); in mlx5_ib_resize_cq()
Dsrq.c295 in.cqn = to_mcq(init_attr->ext.cq)->mcq.cqn; in mlx5_ib_create_srq()
297 in.cqn = to_mcq(dev->devr.c0)->mcq.cqn; in mlx5_ib_create_srq()
Dsrq.h25 u32 cqn; member
Dqp.c1237 MLX5_SET(sqc, sqc, cqn, MLX5_GET(qpc, qpc, cqn_snd)); in create_raw_packet_qp_sq()
1327 MLX5_SET(rqc, rqc, cqn, MLX5_GET(qpc, qpc, cqn_rcv)); in create_raw_packet_qp_rq()
1881 MLX5_SET(qpc, qpc, cqn_rcv, to_mcq(devr->c0)->mcq.cqn); in create_xrc_tgt_qp()
1882 MLX5_SET(qpc, qpc, cqn_snd, to_mcq(devr->c0)->mcq.cqn); in create_xrc_tgt_qp()
2033 MLX5_SET(qpc, qpc, cqn_rcv, to_mcq(devr->c0)->mcq.cqn); in create_user_qp()
2048 MLX5_SET(qpc, qpc, cqn_snd, to_mcq(init_attr->send_cq)->mcq.cqn); in create_user_qp()
2051 MLX5_SET(qpc, qpc, cqn_rcv, to_mcq(init_attr->recv_cq)->mcq.cqn); in create_user_qp()
2191 MLX5_SET(qpc, qpc, cqn_snd, to_mcq(attr->send_cq)->mcq.cqn); in create_kernel_qp()
2194 MLX5_SET(qpc, qpc, cqn_rcv, to_mcq(attr->recv_cq)->mcq.cqn); in create_kernel_qp()
2243 if (send_cq->mcq.cqn < recv_cq->mcq.cqn) { in mlx5_ib_lock_cqs()
[all …]
Dsrq_cmd.c52 MLX5_SET(srqc, srqc, cqn, in->cqn); in set_srqc()
449 MLX5_SET(xrqc, xrqc, cqn, in->cqn); in create_xrq_cmd()
Drestrack.c156 return fill_res_raw(msg, dev, MLX5_SGMT_TYPE_PRM_QUERY_CQ, cq->mcq.cqn); in fill_res_cq_entry_raw()
/drivers/net/ethernet/mellanox/mlx5/core/steering/
Ddr_send.c44 u32 cqn; member
166 MLX5_SET(qpc, qpc, cqn_snd, attr->cqn); in dr_create_rc_qp()
167 MLX5_SET(qpc, qpc, cqn_rcv, attr->cqn); in dr_create_rc_qp()
702 pr_err("CQ completion CQ: #%u\n", mcq->cqn); in dr_cq_complete()
886 init_attr.cqn = dmn->send_ring->cq->mcq.cqn; in mlx5dr_send_ring_alloc()
/drivers/net/ethernet/mellanox/mlx5/core/fpga/
Dconn.c481 mlx5_fpga_dbg(fdev, "Created CQ #0x%x\n", conn->cq.mcq.cqn); in mlx5_fpga_conn_create_cq()
574 MLX5_SET(qpc, qpc, cqn_snd, conn->cq.mcq.cqn); in mlx5_fpga_conn_create_qp()
575 MLX5_SET(qpc, qpc, cqn_rcv, conn->cq.mcq.cqn); in mlx5_fpga_conn_create_qp()
685 MLX5_SET(qpc, qpc, cqn_snd, conn->cq.mcq.cqn); in mlx5_fpga_conn_init_qp()
686 MLX5_SET(qpc, qpc, cqn_rcv, conn->cq.mcq.cqn); in mlx5_fpga_conn_init_qp()

123