Lines Matching refs:ib_cq
81 struct ib_cq *ib_cq; member
140 static void nvme_rdma_recv_done(struct ib_cq *cq, struct ib_wc *wc);
270 init_attr.send_cq = queue->ib_cq; in nvme_rdma_create_qp()
271 init_attr.recv_cq = queue->ib_cq; in nvme_rdma_create_qp()
419 ib_free_cq(queue->ib_cq); in nvme_rdma_destroy_queue_ib()
463 queue->ib_cq = ib_alloc_cq(ibdev, queue, in nvme_rdma_create_queue_ib()
466 if (IS_ERR(queue->ib_cq)) { in nvme_rdma_create_queue_ib()
467 ret = PTR_ERR(queue->ib_cq); in nvme_rdma_create_queue_ib()
509 ib_free_cq(queue->ib_cq); in nvme_rdma_create_queue_ib()
1094 static void nvme_rdma_wr_error(struct ib_cq *cq, struct ib_wc *wc, in nvme_rdma_wr_error()
1108 static void nvme_rdma_memreg_done(struct ib_cq *cq, struct ib_wc *wc) in nvme_rdma_memreg_done()
1114 static void nvme_rdma_inv_rkey_done(struct ib_cq *cq, struct ib_wc *wc) in nvme_rdma_inv_rkey_done()
1323 static void nvme_rdma_send_done(struct ib_cq *cq, struct ib_wc *wc) in nvme_rdma_send_done()
1406 static void nvme_rdma_async_done(struct ib_cq *cq, struct ib_wc *wc) in nvme_rdma_async_done()
1483 static void nvme_rdma_recv_done(struct ib_cq *cq, struct ib_wc *wc) in nvme_rdma_recv_done()
1804 return ib_process_cq_direct(queue->ib_cq, -1); in nvme_rdma_poll()