/kernel/linux/linux-5.10/drivers/infiniband/sw/rdmavt/ |
D | cq.c | 96 if (head >= (unsigned)cq->ibcq.cqe) { in rvt_cq_enter() 97 head = cq->ibcq.cqe; in rvt_cq_enter() 110 if (cq->ibcq.event_handler) { in rvt_cq_enter() 113 ev.device = cq->ibcq.device; in rvt_cq_enter() 114 ev.element.cq = &cq->ibcq; in rvt_cq_enter() 116 cq->ibcq.event_handler(&ev, cq->ibcq.cq_context); in rvt_cq_enter() 182 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in send_complete() 200 int rvt_create_cq(struct ib_cq *ibcq, const struct ib_cq_init_attr *attr, in rvt_create_cq() argument 203 struct ib_device *ibdev = ibcq->device; in rvt_create_cq() 205 struct rvt_cq *cq = ibcq_to_rvtcq(ibcq); in rvt_create_cq() [all …]
|
D | cq.h | 54 int rvt_create_cq(struct ib_cq *ibcq, const struct ib_cq_init_attr *attr, 56 int rvt_destroy_cq(struct ib_cq *ibcq, struct ib_udata *udata); 57 int rvt_req_notify_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags notify_flags); 58 int rvt_resize_cq(struct ib_cq *ibcq, int cqe, struct ib_udata *udata); 59 int rvt_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *entry);
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/mlx4/ |
D | cq.c | 45 struct ib_cq *ibcq = &to_mibcq(cq)->ibcq; in mlx4_ib_cq_comp() local 46 ibcq->comp_handler(ibcq, ibcq->cq_context); in mlx4_ib_cq_comp() 52 struct ib_cq *ibcq; in mlx4_ib_cq_event() local 60 ibcq = &to_mibcq(cq)->ibcq; in mlx4_ib_cq_event() 61 if (ibcq->event_handler) { in mlx4_ib_cq_event() 62 event.device = ibcq->device; in mlx4_ib_cq_event() 64 event.element.cq = ibcq; in mlx4_ib_cq_event() 65 ibcq->event_handler(&event, ibcq->cq_context); in mlx4_ib_cq_event() 81 struct mlx4_cqe *cqe = get_cqe(cq, n & cq->ibcq.cqe); in get_sw_cqe() 85 !!(n & (cq->ibcq.cqe + 1))) ? NULL : cqe; in get_sw_cqe() [all …]
|
D | mlx4_ib.h | 116 struct ib_cq ibcq; member 688 static inline struct mlx4_ib_cq *to_mcq(struct ib_cq *ibcq) in to_mcq() argument 690 return container_of(ibcq, struct mlx4_ib_cq, ibcq); in to_mcq() 766 int mlx4_ib_resize_cq(struct ib_cq *ibcq, int entries, struct ib_udata *udata); 767 int mlx4_ib_create_cq(struct ib_cq *ibcq, const struct ib_cq_init_attr *attr, 770 int mlx4_ib_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc);
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/vmw_pvrdma/ |
D | pvrdma_cq.c | 63 int pvrdma_req_notify_cq(struct ib_cq *ibcq, in pvrdma_req_notify_cq() argument 66 struct pvrdma_dev *dev = to_vdev(ibcq->device); in pvrdma_req_notify_cq() 67 struct pvrdma_cq *cq = to_vcq(ibcq); in pvrdma_req_notify_cq() 83 cq->ibcq.cqe, &head); in pvrdma_req_notify_cq() 101 int pvrdma_create_cq(struct ib_cq *ibcq, const struct ib_cq_init_attr *attr, in pvrdma_create_cq() argument 104 struct ib_device *ibdev = ibcq->device; in pvrdma_create_cq() 107 struct pvrdma_cq *cq = to_vcq(ibcq); in pvrdma_create_cq() 129 cq->ibcq.cqe = entries; in pvrdma_create_cq() 192 cq->ibcq.cqe = resp->cqe; in pvrdma_create_cq() 206 pvrdma_destroy_cq(&cq->ibcq, udata); in pvrdma_create_cq() [all …]
|
D | pvrdma_main.c | 185 INIT_RDMA_OBJ_SIZE(ib_cq, pvrdma_cq, ibcq), 348 if (cq && cq->ibcq.event_handler) { in pvrdma_cq_event() 349 struct ib_cq *ibcq = &cq->ibcq; in pvrdma_cq_event() local 352 e.device = ibcq->device; in pvrdma_cq_event() 353 e.element.cq = ibcq; in pvrdma_cq_event() 355 ibcq->event_handler(&e, ibcq->cq_context); in pvrdma_cq_event() 520 if (cq && cq->ibcq.comp_handler) in pvrdma_intrx_handler() 521 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in pvrdma_intrx_handler()
|
D | pvrdma.h | 87 struct ib_cq ibcq; member 274 static inline struct pvrdma_cq *to_vcq(struct ib_cq *ibcq) in to_vcq() argument 276 return container_of(ibcq, struct pvrdma_cq, ibcq); in to_vcq()
|
/kernel/linux/linux-5.10/drivers/infiniband/sw/rxe/ |
D | rxe_cq.c | 54 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in rxe_send_complete() 86 cq->ibcq.cqe = cqe; in rxe_cq_from_init() 100 cq->ibcq.cqe = cqe; in rxe_cq_resize_queue() 114 if (cq->ibcq.event_handler) { in rxe_cq_post() 115 ev.device = cq->ibcq.device; in rxe_cq_post() 116 ev.element.cq = &cq->ibcq; in rxe_cq_post() 118 cq->ibcq.event_handler(&ev, cq->ibcq.cq_context); in rxe_cq_post()
|
D | rxe_verbs.c | 752 static int rxe_create_cq(struct ib_cq *ibcq, const struct ib_cq_init_attr *attr, in rxe_create_cq() argument 756 struct ib_device *dev = ibcq->device; in rxe_create_cq() 758 struct rxe_cq *cq = to_rcq(ibcq); in rxe_create_cq() 782 static int rxe_destroy_cq(struct ib_cq *ibcq, struct ib_udata *udata) in rxe_destroy_cq() argument 784 struct rxe_cq *cq = to_rcq(ibcq); in rxe_destroy_cq() 792 static int rxe_resize_cq(struct ib_cq *ibcq, int cqe, struct ib_udata *udata) in rxe_resize_cq() argument 795 struct rxe_cq *cq = to_rcq(ibcq); in rxe_resize_cq() 796 struct rxe_dev *rxe = to_rdev(ibcq->device); in rxe_resize_cq() 819 static int rxe_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc) in rxe_poll_cq() argument 822 struct rxe_cq *cq = to_rcq(ibcq); in rxe_poll_cq() [all …]
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/mthca/ |
D | mthca_cq.c | 181 return cqe_sw(get_cqe(cq, cq->cons_index & cq->ibcq.cqe)); in next_cqe_sw() 230 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in mthca_cq_completion() 254 event.element.cq = &cq->ibcq; in mthca_cq_event() 255 if (cq->ibcq.event_handler) in mthca_cq_event() 256 cq->ibcq.event_handler(&event, cq->ibcq.cq_context); in mthca_cq_event() 290 cqe_sw(get_cqe(cq, prod_index & cq->ibcq.cqe)); in mthca_cq_clean() 292 if (prod_index == cq->cons_index + cq->ibcq.cqe) in mthca_cq_clean() 304 cqe = get_cqe(cq, prod_index & cq->ibcq.cqe); in mthca_cq_clean() 310 memcpy(get_cqe(cq, (prod_index + nfreed) & cq->ibcq.cqe), in mthca_cq_clean() 316 set_cqe_hw(get_cqe(cq, (cq->cons_index + i) & cq->ibcq.cqe)); in mthca_cq_clean() [all …]
|
D | mthca_provider.h | 184 struct ib_cq ibcq; member 301 static inline struct mthca_cq *to_mcq(struct ib_cq *ibcq) in to_mcq() argument 303 return container_of(ibcq, struct mthca_cq, ibcq); in to_mcq()
|
D | mthca_provider.c | 601 static int mthca_create_cq(struct ib_cq *ibcq, in mthca_create_cq() argument 605 struct ib_device *ibdev = ibcq->device; in mthca_create_cq() 637 cq = to_mcq(ibcq); in mthca_create_cq() 722 static int mthca_resize_cq(struct ib_cq *ibcq, int entries, struct ib_udata *udata) in mthca_resize_cq() argument 724 struct mthca_dev *dev = to_mdev(ibcq->device); in mthca_resize_cq() 725 struct mthca_cq *cq = to_mcq(ibcq); in mthca_resize_cq() 736 if (entries == ibcq->cqe + 1) { in mthca_resize_cq() 776 tcqe = cq->ibcq.cqe; in mthca_resize_cq() 778 cq->ibcq.cqe = cq->resize_buf->cqe; in mthca_resize_cq() 790 ibcq->cqe = entries - 1; in mthca_resize_cq() [all …]
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/mlx5/ |
D | cq.c | 43 struct ib_cq *ibcq = &to_mibcq(cq)->ibcq; in mlx5_ib_cq_comp() local 45 ibcq->comp_handler(ibcq, ibcq->cq_context); in mlx5_ib_cq_comp() 51 struct mlx5_ib_dev *dev = to_mdev(cq->ibcq.device); in mlx5_ib_cq_event() 52 struct ib_cq *ibcq = &cq->ibcq; in mlx5_ib_cq_event() local 61 if (ibcq->event_handler) { in mlx5_ib_cq_event() 64 event.element.cq = ibcq; in mlx5_ib_cq_event() 65 ibcq->event_handler(&event, ibcq->cq_context); in mlx5_ib_cq_event() 81 void *cqe = get_cqe(cq, n & cq->ibcq.cqe); in get_sw_cqe() 87 !((cqe64->op_own & MLX5_CQE_OWNER_MASK) ^ !!(n & (cq->ibcq.cqe + 1)))) { in get_sw_cqe() 445 struct mlx5_ib_dev *dev = to_mdev(cq->ibcq.device); in mlx5_poll_one() [all …]
|
D | restrack.c | 151 static int fill_res_cq_entry_raw(struct sk_buff *msg, struct ib_cq *ibcq) in fill_res_cq_entry_raw() argument 153 struct mlx5_ib_dev *dev = to_mdev(ibcq->device); in fill_res_cq_entry_raw() 154 struct mlx5_ib_cq *cq = to_mcq(ibcq); in fill_res_cq_entry_raw()
|
D | mlx5_ib.h | 493 struct ib_cq ibcq; member 1053 static inline struct mlx5_ib_cq *to_mcq(struct ib_cq *ibcq) in to_mcq() argument 1055 return container_of(ibcq, struct mlx5_ib_cq, ibcq); in to_mcq() 1166 int mlx5_ib_create_cq(struct ib_cq *ibcq, const struct ib_cq_init_attr *attr, 1169 int mlx5_ib_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc); 1170 int mlx5_ib_arm_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags flags); 1172 int mlx5_ib_resize_cq(struct ib_cq *ibcq, int entries, struct ib_udata *udata); 1243 int mlx5_ib_get_cqe_size(struct ib_cq *ibcq); 1356 int mlx5_ib_generate_wc(struct ib_cq *ibcq, struct ib_wc *wc);
|
/kernel/linux/linux-5.10/include/rdma/ |
D | rdmavt_cq.h | 47 struct ib_cq ibcq; member 60 static inline struct rvt_cq *ibcq_to_rvtcq(struct ib_cq *ibcq) in ibcq_to_rvtcq() argument 62 return container_of(ibcq, struct rvt_cq, ibcq); in ibcq_to_rvtcq()
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/hns/ |
D | hns_roce_cq.c | 351 struct ib_cq *ibcq; in hns_roce_cq_completion() local 362 ibcq = &hr_cq->ib_cq; in hns_roce_cq_completion() 363 if (ibcq->comp_handler) in hns_roce_cq_completion() 364 ibcq->comp_handler(ibcq, ibcq->cq_context); in hns_roce_cq_completion() 372 struct ib_cq *ibcq; in hns_roce_cq_event() local 391 ibcq = &hr_cq->ib_cq; in hns_roce_cq_event() 392 if (ibcq->event_handler) { in hns_roce_cq_event() 393 event.device = ibcq->device; in hns_roce_cq_event() 394 event.element.cq = ibcq; in hns_roce_cq_event() 396 ibcq->event_handler(&event, ibcq->cq_context); in hns_roce_cq_event()
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/cxgb4/ |
D | ev.c | 105 event.device = chp->ibcq.device; in post_qp_event() 107 event.element.cq = &chp->ibcq; in post_qp_event() 115 (*chp->ibcq.comp_handler)(&chp->ibcq, chp->ibcq.cq_context); in post_qp_event() 235 (*chp->ibcq.comp_handler)(&chp->ibcq, chp->ibcq.cq_context); in c4iw_ev_handler()
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/qedr/ |
D | qedr_roce_cm.c | 82 cq->ibcq.comp_handler ? "Yes" : "No"); in qedr_ll2_complete_tx_packet() 92 if (cq->ibcq.comp_handler) in qedr_ll2_complete_tx_packet() 93 (*cq->ibcq.comp_handler) (&cq->ibcq, cq->ibcq.cq_context); in qedr_ll2_complete_tx_packet() 121 if (cq->ibcq.comp_handler) in qedr_ll2_complete_rx_packet() 122 (*cq->ibcq.comp_handler) (&cq->ibcq, cq->ibcq.cq_context); in qedr_ll2_complete_rx_packet() 672 int qedr_gsi_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc) in qedr_gsi_poll_cq() argument 674 struct qedr_dev *dev = get_qedr_dev(ibcq->device); in qedr_gsi_poll_cq() 675 struct qedr_cq *cq = get_qedr_cq(ibcq); in qedr_gsi_poll_cq()
|
D | verbs.h | 53 int qedr_create_cq(struct ib_cq *ibcq, const struct ib_cq_init_attr *attr, 56 int qedr_destroy_cq(struct ib_cq *ibcq, struct ib_udata *udata); 57 int qedr_arm_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags flags);
|
D | main.c | 238 INIT_RDMA_OBJ_SIZE(ib_cq, qedr_cq, ibcq), 505 if (!cq->destroyed && cq->ibcq.comp_handler) in qedr_irq_handler() 506 (*cq->ibcq.comp_handler) in qedr_irq_handler() 507 (&cq->ibcq, cq->ibcq.cq_context); in qedr_irq_handler() 687 struct ib_cq *ibcq; in qedr_affiliated_event() local 758 ibcq = &cq->ibcq; in qedr_affiliated_event() 759 if (ibcq->event_handler) { in qedr_affiliated_event() 760 event.device = ibcq->device; in qedr_affiliated_event() 761 event.element.cq = ibcq; in qedr_affiliated_event() 762 ibcq->event_handler(&event, ibcq->cq_context); in qedr_affiliated_event()
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/efa/ |
D | efa.h | 83 struct ib_cq ibcq; member 138 int efa_destroy_cq(struct ib_cq *ibcq, struct ib_udata *udata); 139 int efa_create_cq(struct ib_cq *ibcq, const struct ib_cq_init_attr *attr,
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/i40iw/ |
D | i40iw_hw.c | 175 if (i40iwcq->ibcq.comp_handler) in i40iw_iwarp_ce_handler() 176 i40iwcq->ibcq.comp_handler(&i40iwcq->ibcq, in i40iw_iwarp_ce_handler() 177 i40iwcq->ibcq.cq_context); in i40iw_iwarp_ce_handler() 380 if (iwcq->ibcq.event_handler) { in i40iw_process_aeq() 383 ibevent.device = iwcq->ibcq.device; in i40iw_process_aeq() 385 ibevent.element.cq = &iwcq->ibcq; in i40iw_process_aeq() 386 iwcq->ibcq.event_handler(&ibevent, iwcq->ibcq.cq_context); in i40iw_process_aeq()
|
/kernel/linux/linux-5.10/drivers/infiniband/hw/ocrdma/ |
D | ocrdma_verbs.h | 72 int ocrdma_create_cq(struct ib_cq *ibcq, const struct ib_cq_init_attr *attr, 75 int ocrdma_destroy_cq(struct ib_cq *ibcq, struct ib_udata *udata);
|
D | ocrdma.h | 316 struct ib_cq ibcq; member 471 static inline struct ocrdma_cq *get_ocrdma_cq(struct ib_cq *ibcq) in get_ocrdma_cq() argument 473 return container_of(ibcq, struct ocrdma_cq, ibcq); in get_ocrdma_cq()
|