/drivers/infiniband/hw/ipath/ |
D | ipath_cq.c | 62 if (head >= (unsigned) cq->ibcq.cqe) { in ipath_cq_enter() 63 head = cq->ibcq.cqe; in ipath_cq_enter() 69 if (cq->ibcq.event_handler) { in ipath_cq_enter() 72 ev.device = cq->ibcq.device; in ipath_cq_enter() 73 ev.element.cq = &cq->ibcq; in ipath_cq_enter() 75 cq->ibcq.event_handler(&ev, cq->ibcq.cq_context); in ipath_cq_enter() 114 to_idev(cq->ibcq.device)->n_wqe_errs++; in ipath_cq_enter() 128 int ipath_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *entry) in ipath_poll_cq() argument 130 struct ipath_cq *cq = to_icq(ibcq); in ipath_poll_cq() 146 if (tail > (u32) cq->ibcq.cqe) in ipath_poll_cq() [all …]
|
D | ipath_verbs.h | 215 struct ib_cq ibcq; member 670 static inline struct ipath_cq *to_icq(struct ib_cq *ibcq) in to_icq() argument 672 return container_of(ibcq, struct ipath_cq, ibcq); in to_icq() 808 int ipath_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *entry); 814 int ipath_destroy_cq(struct ib_cq *ibcq); 816 int ipath_req_notify_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags notify_flags); 818 int ipath_resize_cq(struct ib_cq *ibcq, int cqe, struct ib_udata *udata);
|
/drivers/infiniband/hw/mlx4/ |
D | cq.c | 42 struct ib_cq *ibcq = &to_mibcq(cq)->ibcq; in mlx4_ib_cq_comp() local 43 ibcq->comp_handler(ibcq, ibcq->cq_context); in mlx4_ib_cq_comp() 49 struct ib_cq *ibcq; in mlx4_ib_cq_event() local 57 ibcq = &to_mibcq(cq)->ibcq; in mlx4_ib_cq_event() 58 if (ibcq->event_handler) { in mlx4_ib_cq_event() 59 event.device = ibcq->device; in mlx4_ib_cq_event() 61 event.element.cq = ibcq; in mlx4_ib_cq_event() 62 ibcq->event_handler(&event, ibcq->cq_context); in mlx4_ib_cq_event() 78 struct mlx4_cqe *cqe = get_cqe(cq, n & cq->ibcq.cqe); in get_sw_cqe() 81 !!(n & (cq->ibcq.cqe + 1))) ? NULL : cqe; in get_sw_cqe() [all …]
|
D | mlx4_ib.h | 70 struct ib_cq ibcq; member 194 static inline struct mlx4_ib_cq *to_mcq(struct ib_cq *ibcq) in to_mcq() argument 196 return container_of(ibcq, struct mlx4_ib_cq, ibcq); in to_mcq() 261 int mlx4_ib_resize_cq(struct ib_cq *ibcq, int entries, struct ib_udata *udata); 266 int mlx4_ib_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc);
|
/drivers/infiniband/hw/mthca/ |
D | mthca_cq.c | 180 return cqe_sw(get_cqe(cq, cq->cons_index & cq->ibcq.cqe)); in next_cqe_sw() 234 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in mthca_cq_completion() 258 event.element.cq = &cq->ibcq; in mthca_cq_event() 259 if (cq->ibcq.event_handler) in mthca_cq_event() 260 cq->ibcq.event_handler(&event, cq->ibcq.cq_context); in mthca_cq_event() 294 cqe_sw(get_cqe(cq, prod_index & cq->ibcq.cqe)); in mthca_cq_clean() 296 if (prod_index == cq->cons_index + cq->ibcq.cqe) in mthca_cq_clean() 308 cqe = get_cqe(cq, prod_index & cq->ibcq.cqe); in mthca_cq_clean() 314 memcpy(get_cqe(cq, (prod_index + nfreed) & cq->ibcq.cqe), in mthca_cq_clean() 320 set_cqe_hw(get_cqe(cq, (cq->cons_index + i) & cq->ibcq.cqe)); in mthca_cq_clean() [all …]
|
D | mthca_provider.h | 201 struct ib_cq ibcq; member 323 static inline struct mthca_cq *to_mcq(struct ib_cq *ibcq) in to_mcq() argument 325 return container_of(ibcq, struct mthca_cq, ibcq); in to_mcq()
|
D | mthca_provider.c | 730 return &cq->ibcq; in mthca_create_cq() 793 static int mthca_resize_cq(struct ib_cq *ibcq, int entries, struct ib_udata *udata) in mthca_resize_cq() argument 795 struct mthca_dev *dev = to_mdev(ibcq->device); in mthca_resize_cq() 796 struct mthca_cq *cq = to_mcq(ibcq); in mthca_resize_cq() 808 if (entries == ibcq->cqe + 1) { in mthca_resize_cq() 850 tcqe = cq->ibcq.cqe; in mthca_resize_cq() 852 cq->ibcq.cqe = cq->resize_buf->cqe; in mthca_resize_cq() 864 ibcq->cqe = entries - 1; in mthca_resize_cq()
|
D | mthca_dev.h | 492 int mthca_poll_cq(struct ib_cq *ibcq, int num_entries,
|
/drivers/infiniband/hw/cxgb3/ |
D | iwch_ev.c | 80 event.device = chp->ibcq.device; in post_qp_event() 82 event.element.cq = &chp->ibcq; in post_qp_event() 171 (*chp->ibcq.comp_handler)(&chp->ibcq, chp->ibcq.cq_context); in iwch_ev_dispatch() 187 (*chp->ibcq.comp_handler)(&chp->ibcq, chp->ibcq.cq_context); in iwch_ev_dispatch()
|
D | iwch_provider.h | 102 struct ib_cq ibcq; member 111 static inline struct iwch_cq *to_iwch_cq(struct ib_cq *ibcq) in to_iwch_cq() argument 113 return container_of(ibcq, struct iwch_cq, ibcq); in to_iwch_cq() 333 int iwch_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc);
|
D | iwch_cq.c | 195 int iwch_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc) in iwch_poll_cq() argument 203 chp = to_iwch_cq(ibcq); in iwch_poll_cq()
|
D | iwch_provider.c | 193 chp->ibcq.cqe = 1 << chp->cq.size_log2; in iwch_create_cq() 204 iwch_destroy_cq(&chp->ibcq); in iwch_create_cq() 215 iwch_destroy_cq(&chp->ibcq); in iwch_create_cq() 227 return &chp->ibcq; in iwch_create_cq() 278 chp->ibcq.cqe = (1<<chp->cq.size_log2) - 1; in iwch_resize_cq() 298 static int iwch_arm_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags flags) in iwch_arm_cq() argument 307 chp = to_iwch_cq(ibcq); in iwch_arm_cq()
|
D | iwch_qp.c | 818 (*rchp->ibcq.comp_handler)(&rchp->ibcq, rchp->ibcq.cq_context); in __flush_qp() 829 (*schp->ibcq.comp_handler)(&schp->ibcq, schp->ibcq.cq_context); in __flush_qp()
|
/drivers/infiniband/hw/amso1100/ |
D | c2_cq.c | 76 (*cq->ibcq.comp_handler) (&cq->ibcq, cq->ibcq.cq_context); in c2_cq_event() 199 int c2_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *entry) in c2_poll_cq() argument 201 struct c2_dev *c2dev = to_c2dev(ibcq->device); in c2_poll_cq() 202 struct c2_cq *cq = to_c2cq(ibcq); in c2_poll_cq() 220 int c2_arm_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags notify_flags) in c2_arm_cq() argument 227 cq = to_c2cq(ibcq); in c2_arm_cq() 295 cq->ibcq.cqe = entries - 1; in c2_init_cq()
|
D | c2_provider.h | 92 struct ib_cq ibcq; member 153 static inline struct c2_cq *to_c2cq(struct ib_cq *ibcq) in to_c2cq() argument 155 return container_of(ibcq, struct c2_cq, ibcq); in to_c2cq()
|
D | c2_ae.c | 303 ib_event.element.cq = &cq->ibcq; in c2_ae_event() 306 if (cq->ibcq.event_handler) in c2_ae_event() 307 cq->ibcq.event_handler(&ib_event, in c2_ae_event() 308 cq->ibcq.cq_context); in c2_ae_event()
|
D | c2.h | 520 extern int c2_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *entry); 521 extern int c2_arm_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags flags);
|
D | c2_provider.c | 316 return &cq->ibcq; in c2_create_cq()
|
/drivers/infiniband/hw/ehca/ |
D | ehca_main.c | 530 struct ib_cq *ibcq; in ehca_create_aqp1() local 540 ibcq = ib_create_cq(&shca->ib_device, NULL, NULL, (void *)(-1), 10, 0); in ehca_create_aqp1() 541 if (IS_ERR(ibcq)) { in ehca_create_aqp1() 543 return PTR_ERR(ibcq); in ehca_create_aqp1() 545 sport->ibcq_aqp1 = ibcq; in ehca_create_aqp1() 554 qp_init_attr.send_cq = ibcq; in ehca_create_aqp1() 555 qp_init_attr.recv_cq = ibcq; in ehca_create_aqp1()
|
/drivers/infiniband/hw/nes/ |
D | nes.h | 478 static inline struct nes_cq *to_nescq(struct ib_cq *ibcq) in to_nescq() argument 480 return container_of(ibcq, struct nes_cq, ibcq); in to_nescq()
|
D | nes_verbs.h | 108 struct ib_cq ibcq; member
|
D | nes_verbs.c | 1608 nescq->ibcq.cqe = nescq->hw_cq.cq_size - 1; in nes_create_cq() 1795 return &nescq->ibcq; in nes_create_cq() 3434 static int nes_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *entry) in nes_poll_cq() argument 3440 struct nes_vnic *nesvnic = to_nesvnic(ibcq->device); in nes_poll_cq() 3442 struct nes_cq *nescq = to_nescq(ibcq); in nes_poll_cq() 3573 static int nes_req_notify_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags notify_flags) in nes_req_notify_cq() argument 3575 struct nes_vnic *nesvnic = to_nesvnic(ibcq->device); in nes_req_notify_cq() 3577 struct nes_cq *nescq = to_nescq(ibcq); in nes_req_notify_cq()
|
D | nes_hw.c | 3239 if (nescq->ibcq.comp_handler) in nes_iwarp_ce_handler() 3240 nescq->ibcq.comp_handler(&nescq->ibcq, nescq->ibcq.cq_context); in nes_iwarp_ce_handler()
|