/drivers/infiniband/hw/cxgb3/ |
D | iwch.h | 127 static inline int t3b_device(const struct iwch_dev *rhp) in t3b_device() argument 129 return rhp->rdev.t3cdev_p->type == T3B; in t3b_device() 132 static inline int t3a_device(const struct iwch_dev *rhp) in t3a_device() argument 134 return rhp->rdev.t3cdev_p->type == T3A; in t3a_device() 137 static inline struct iwch_cq *get_chp(struct iwch_dev *rhp, u32 cqid) in get_chp() argument 139 return idr_find(&rhp->cqidr, cqid); in get_chp() 142 static inline struct iwch_qp *get_qhp(struct iwch_dev *rhp, u32 qpid) in get_qhp() argument 144 return idr_find(&rhp->qpidr, qpid); in get_qhp() 147 static inline struct iwch_mr *get_mhp(struct iwch_dev *rhp, u32 mmid) in get_mhp() argument 149 return idr_find(&rhp->mmidr, mmid); in get_mhp() [all …]
|
D | iwch_provider.c | 97 struct iwch_dev *rhp = to_iwch_dev(context->device); in iwch_dealloc_ucontext() local 104 cxio_release_ucontext(&rhp->rdev, &ucontext->uctx); in iwch_dealloc_ucontext() 113 struct iwch_dev *rhp = to_iwch_dev(ibdev); in iwch_alloc_ucontext() local 119 cxio_init_ucontext(&rhp->rdev, &context->uctx); in iwch_alloc_ucontext() 132 remove_handle(chp->rhp, &chp->rhp->cqidr, chp->cq.cqid); in iwch_destroy_cq() 136 cxio_destroy_cq(&chp->rhp->rdev, &chp->cq); in iwch_destroy_cq() 145 struct iwch_dev *rhp; in iwch_create_cq() local 154 rhp = to_iwch_dev(ibdev); in iwch_create_cq() 161 if (!t3a_device(rhp)) { in iwch_create_cq() 170 if (t3a_device(rhp)) { in iwch_create_cq() [all …]
|
D | iwch_mem.c | 52 return insert_handle(mhp->rhp, &mhp->rhp->mmidr, mhp, mmid); in iwch_finish_mem_reg() 55 int iwch_register_mem(struct iwch_dev *rhp, struct iwch_pd *php, in iwch_register_mem() argument 61 if (cxio_register_phys_mem(&rhp->rdev, in iwch_register_mem() 73 cxio_dereg_mem(&rhp->rdev, mhp->attr.stag, mhp->attr.pbl_size, in iwch_register_mem() 78 int iwch_reregister_mem(struct iwch_dev *rhp, struct iwch_pd *php, in iwch_reregister_mem() argument 91 if (cxio_reregister_phys_mem(&rhp->rdev, in iwch_reregister_mem() 103 cxio_dereg_mem(&rhp->rdev, mhp->attr.stag, mhp->attr.pbl_size, in iwch_reregister_mem() 111 mhp->attr.pbl_addr = cxio_hal_pblpool_alloc(&mhp->rhp->rdev, in iwch_alloc_pbl() 124 cxio_hal_pblpool_free(&mhp->rhp->rdev, mhp->attr.pbl_addr, in iwch_free_pbl() 130 return cxio_write_pbl(&mhp->rhp->rdev, pages, in iwch_write_pbl()
|
D | iwch_cq.c | 44 static int iwch_poll_cq_one(struct iwch_dev *rhp, struct iwch_cq *chp, in iwch_poll_cq_one() argument 60 qhp = get_qhp(rhp, CQE_QPID(*rd_cqe)); in iwch_poll_cq_one() 69 if (t3a_device(chp->rhp) && credit) { in iwch_poll_cq_one() 72 cxio_hal_cq_op(&rhp->rdev, &chp->cq, CQ_CREDIT_UPDATE, credit); in iwch_poll_cq_one() 197 struct iwch_dev *rhp; in iwch_poll_cq() local 204 rhp = chp->rhp; in iwch_poll_cq() 218 err = iwch_poll_cq_one(rhp, chp, wc + npolled); in iwch_poll_cq()
|
D | iwch_provider.h | 47 struct iwch_dev *rhp; member 77 struct iwch_dev *rhp; member 91 struct iwch_dev *rhp; member 103 struct iwch_dev *rhp; member 161 struct iwch_dev *rhp; member 258 int iwch_modify_qp(struct iwch_dev *rhp, 340 int iwch_register_mem(struct iwch_dev *rhp, struct iwch_pd *php, 342 int iwch_reregister_mem(struct iwch_dev *rhp, struct iwch_pd *php,
|
D | iwch_qp.c | 200 static int iwch_sgl2pbl_map(struct iwch_dev *rhp, struct ib_sge *sg_list, in iwch_sgl2pbl_map() argument 208 mhp = get_mhp(rhp, (sg_list[i].lkey) >> 8); in iwch_sgl2pbl_map() 240 rhp->rdev.rnic_info.pbl_base) >> 3) + in iwch_sgl2pbl_map() 254 err = iwch_sgl2pbl_map(qhp->rhp, wr->sg_list, wr->num_sge, pbl_addr, in build_rdma_recv() 301 pbl_addr = cxio_hal_pblpool_alloc(&qhp->rhp->rdev, T3_STAG0_PBL_SIZE); in build_zero_stag_recv() 308 pbl_offset = (pbl_addr - qhp->rhp->rdev.rnic_info.pbl_base) >> 3; in build_zero_stag_recv() 532 struct iwch_dev *rhp; in iwch_bind_mw() local 548 rhp = qhp->rhp; in iwch_bind_mw() 583 err = iwch_sgl2pbl_map(rhp, &sgl, 1, &pbl_addr, &page_size); in iwch_bind_mw() 768 return iwch_cxgb3_ofld_send(ep->com.qp->rhp->rdev.t3cdev_p, skb); in iwch_post_zb_read() [all …]
|
D | iwch_cm.c | 925 err = iwch_modify_qp(ep->com.qp->rhp, in process_mpa_reply() 1484 iwch_modify_qp(ep->com.qp->rhp, ep->com.qp, in peer_close() 1499 iwch_modify_qp(ep->com.qp->rhp, ep->com.qp, in peer_close() 1594 ret = iwch_modify_qp(ep->com.qp->rhp, in peer_abort() 1661 iwch_modify_qp(ep->com.qp->rhp, in close_con_rpl() 1726 iwch_modify_qp(ep->com.qp->rhp, in ec_status() 1756 iwch_modify_qp(ep->com.qp->rhp, in ep_timeout() 1812 if ((conn_param->ord > qp->rhp->attr.max_rdma_read_qp_depth) || in iwch_accept_cr() 1813 (conn_param->ird > qp->rhp->attr.max_rdma_reads_per_qp)) { in iwch_accept_cr() 1845 err = iwch_modify_qp(ep->com.qp->rhp, in iwch_accept_cr()
|
D | iwch_ev.c | 82 iwch_modify_qp(qhp->rhp, qhp, IWCH_QP_ATTR_NEXT_STATE, in post_qp_event()
|
D | iwch.c | 79 ring_doorbell(qhp->rhp->rdev.ctrl_qp.doorbell, qhp->wq.qpid); in enable_qp_db()
|
/drivers/infiniband/hw/cxgb4/ |
D | mem.c | 362 return insert_handle(mhp->rhp, &mhp->rhp->mmidr, mhp, mmid); in finish_mem_reg() 365 static int register_mem(struct c4iw_dev *rhp, struct c4iw_pd *php, in register_mem() argument 371 ret = write_tpt_entry(&rhp->rdev, 0, &stag, 1, mhp->attr.pdid, in register_mem() 381 dereg_mem(&rhp->rdev, mhp->attr.stag, mhp->attr.pbl_size, in register_mem() 386 static int reregister_mem(struct c4iw_dev *rhp, struct c4iw_pd *php, in reregister_mem() argument 396 ret = write_tpt_entry(&rhp->rdev, 0, &stag, 1, mhp->attr.pdid, in reregister_mem() 406 dereg_mem(&rhp->rdev, mhp->attr.stag, mhp->attr.pbl_size, in reregister_mem() 414 mhp->attr.pbl_addr = c4iw_pblpool_alloc(&mhp->rhp->rdev, in alloc_pbl() 500 struct c4iw_dev *rhp; in c4iw_reregister_phys_mem() local 514 rhp = mhp->rhp; in c4iw_reregister_phys_mem() [all …]
|
D | provider.c | 92 struct c4iw_dev *rhp = to_c4iw_dev(context->device); in c4iw_dealloc_ucontext() local 99 c4iw_release_dev_ucontext(&rhp->rdev, &ucontext->uctx); in c4iw_dealloc_ucontext() 108 struct c4iw_dev *rhp = to_c4iw_dev(ibdev); in c4iw_alloc_ucontext() local 121 c4iw_init_dev_ucontext(&rhp->rdev, &context->uctx); in c4iw_alloc_ucontext() 128 rhp->rdev.flags |= T4_STATUS_PAGE_DISABLED; in c4iw_alloc_ucontext() 149 mm->addr = virt_to_phys(rhp->rdev.status_page); in c4iw_alloc_ucontext() 233 struct c4iw_dev *rhp; in c4iw_deallocate_pd() local 237 rhp = php->rhp; in c4iw_deallocate_pd() 239 c4iw_put_resource(&rhp->rdev.resource.pdid_table, php->pdid); in c4iw_deallocate_pd() 240 mutex_lock(&rhp->rdev.stats.lock); in c4iw_deallocate_pd() [all …]
|
D | qp.c | 692 spin_lock_irqsave(&qhp->rhp->lock, flags); in ring_kernel_sq_db() 694 if (qhp->rhp->db_state == NORMAL) in ring_kernel_sq_db() 696 is_t5(qhp->rhp->rdev.lldi.adapter_type), NULL); in ring_kernel_sq_db() 698 add_to_fc_list(&qhp->rhp->db_fc_list, &qhp->db_fc_entry); in ring_kernel_sq_db() 702 spin_unlock_irqrestore(&qhp->rhp->lock, flags); in ring_kernel_sq_db() 710 spin_lock_irqsave(&qhp->rhp->lock, flags); in ring_kernel_rq_db() 712 if (qhp->rhp->db_state == NORMAL) in ring_kernel_rq_db() 714 is_t5(qhp->rhp->rdev.lldi.adapter_type), NULL); in ring_kernel_rq_db() 716 add_to_fc_list(&qhp->rhp->db_fc_list, &qhp->db_fc_entry); in ring_kernel_rq_db() 720 spin_unlock_irqrestore(&qhp->rhp->lock, flags); in ring_kernel_rq_db() [all …]
|
D | iw_cxgb4.h | 280 static inline struct c4iw_cq *get_chp(struct c4iw_dev *rhp, u32 cqid) in get_chp() argument 282 return idr_find(&rhp->cqidr, cqid); in get_chp() 285 static inline struct c4iw_qp *get_qhp(struct c4iw_dev *rhp, u32 qpid) in get_qhp() argument 287 return idr_find(&rhp->qpidr, qpid); in get_qhp() 290 static inline struct c4iw_mr *get_mhp(struct c4iw_dev *rhp, u32 mmid) in get_mhp() argument 292 return idr_find(&rhp->mmidr, mmid); in get_mhp() 295 static inline int _insert_handle(struct c4iw_dev *rhp, struct idr *idr, in _insert_handle() argument 302 spin_lock_irq(&rhp->lock); in _insert_handle() 308 spin_unlock_irq(&rhp->lock); in _insert_handle() 316 static inline int insert_handle(struct c4iw_dev *rhp, struct idr *idr, in insert_handle() argument [all …]
|
D | cq.c | 350 qhp = get_qhp(chp->rhp, CQE_QPID(hw_cqe)); in c4iw_flush_hw_cq() 692 qhp = get_qhp(chp->rhp, CQE_QPID(rd_cqe)); in c4iw_poll_cq_one() 850 remove_handle(chp->rhp, &chp->rhp->cqidr, chp->cq.cqid); in c4iw_destroy_cq() 856 destroy_cq(&chp->rhp->rdev, &chp->cq, in c4iw_destroy_cq() 866 struct c4iw_dev *rhp; in c4iw_create_cq() local 876 rhp = to_c4iw_dev(ibdev); in c4iw_create_cq() 878 if (vector >= rhp->rdev.lldi.nciq) in c4iw_create_cq() 902 hwentries = min(entries * 2, rhp->rdev.hw_queue.t4_max_iq_size); in c4iw_create_cq() 922 ret = create_cq(&rhp->rdev, &chp->cq, in c4iw_create_cq() 923 ucontext ? &ucontext->uctx : &rhp->rdev.uctx); in c4iw_create_cq() [all …]
|
D | device.c | 1261 is_t5(qp->rhp->rdev.lldi.adapter_type), NULL); in resume_rc_qp() 1264 is_t5(qp->rhp->rdev.lldi.adapter_type), NULL); in resume_rc_qp() 1364 spin_lock_irq(&qp->rhp->lock); in recover_lost_dbs() 1366 ret = cxgb4_sync_txq_pidx(qp->rhp->rdev.lldi.ports[0], in recover_lost_dbs() 1376 spin_unlock_irq(&qp->rhp->lock); in recover_lost_dbs() 1381 ret = cxgb4_sync_txq_pidx(qp->rhp->rdev.lldi.ports[0], in recover_lost_dbs() 1392 spin_unlock_irq(&qp->rhp->lock); in recover_lost_dbs() 1397 spin_unlock_irq(&qp->rhp->lock); in recover_lost_dbs() 1400 while (cxgb4_dbfifo_count(qp->rhp->rdev.lldi.ports[0], 1) > 0) { in recover_lost_dbs()
|
D | ev.c | 100 c4iw_modify_qp(qhp->rhp, qhp, C4IW_QP_ATTR_NEXT_STATE, in post_qp_event()
|
D | cm.c | 1456 err = c4iw_modify_qp(ep->com.qp->rhp, in process_mpa_reply() 1471 err = c4iw_modify_qp(ep->com.qp->rhp, ep->com.qp, in process_mpa_reply() 1491 err = c4iw_modify_qp(ep->com.qp->rhp, ep->com.qp, in process_mpa_reply() 1694 c4iw_modify_qp(ep->com.qp->rhp, ep->com.qp, in rx_data() 2468 ret = c4iw_modify_qp(ep->com.qp->rhp, ep->com.qp, in peer_close() 2486 c4iw_modify_qp(ep->com.qp->rhp, ep->com.qp, in peer_close() 2575 ret = c4iw_modify_qp(ep->com.qp->rhp, in peer_abort() 2655 c4iw_modify_qp(ep->com.qp->rhp, in close_con_rpl() 2692 c4iw_modify_qp(ep->com.qp->rhp, ep->com.qp, in terminate() 2848 err = c4iw_modify_qp(ep->com.qp->rhp, in c4iw_accept_cr() [all …]
|
/drivers/media/usb/pvrusb2/ |
D | pvrusb2-v4l2.c | 56 struct pvr2_ioread *rhp; member 976 if (fhp->rhp) { in pvr2_v4l2_release() 979 sp = pvr2_ioread_get_stream(fhp->rhp); in pvr2_v4l2_release() 981 pvr2_ioread_destroy(fhp->rhp); in pvr2_v4l2_release() 982 fhp->rhp = NULL; in pvr2_v4l2_release() 1124 if (fh->rhp) return 0; in pvr2_v4l2_iosetup() 1140 fh->rhp = pvr2_channel_create_mpeg_stream(fh->pdi->stream); in pvr2_v4l2_iosetup() 1141 if (!fh->rhp) { in pvr2_v4l2_iosetup() 1151 return pvr2_ioread_set_enabled(fh->rhp,!0); in pvr2_v4l2_iosetup() 1194 if (!fh->rhp) { in pvr2_v4l2_read() [all …]
|