/drivers/infiniband/sw/rxe/ |
D | rxe_cq.c | 59 struct rxe_create_cq_resp __user *uresp) in rxe_cq_from_init() argument 70 err = do_mmap_info(rxe, uresp ? &uresp->mi : NULL, udata, in rxe_cq_from_init() 78 if (uresp) in rxe_cq_from_init() 91 struct rxe_resize_cq_resp __user *uresp, in rxe_cq_resize_queue() argument 98 uresp ? &uresp->mi : NULL, NULL, &cq->cq_lock); in rxe_cq_resize_queue()
|
D | rxe_srq.c | 76 struct rxe_create_srq_resp __user *uresp) in rxe_srq_from_init() argument 103 err = do_mmap_info(rxe, uresp ? &uresp->mi : NULL, udata, q->buf, in rxe_srq_from_init() 111 if (uresp) { in rxe_srq_from_init() 112 if (copy_to_user(&uresp->srq_num, &srq->srq_num, in rxe_srq_from_init() 113 sizeof(uresp->srq_num))) { in rxe_srq_from_init()
|
D | rxe_verbs.c | 266 struct rxe_create_srq_resp __user *uresp = NULL; in rxe_create_srq() local 269 if (udata->outlen < sizeof(*uresp)) in rxe_create_srq() 271 uresp = udata->outbuf; in rxe_create_srq() 285 err = rxe_srq_from_init(rxe, srq, init, udata, uresp); in rxe_create_srq() 387 struct rxe_create_qp_resp __user *uresp = NULL; in rxe_create_qp() local 390 if (udata->outlen < sizeof(*uresp)) in rxe_create_qp() 392 uresp = udata->outbuf; in rxe_create_qp() 415 err = rxe_qp_from_init(rxe, qp, pd, init, uresp, ibpd, udata); in rxe_create_qp() 759 struct rxe_create_cq_resp __user *uresp = NULL; in rxe_create_cq() local 762 if (udata->outlen < sizeof(*uresp)) in rxe_create_cq() [all …]
|
D | rxe_loc.h | 30 struct rxe_create_cq_resp __user *uresp); 33 struct rxe_resize_cq_resp __user *uresp, 128 struct rxe_create_qp_resp __user *uresp, 196 struct rxe_create_srq_resp __user *uresp);
|
D | rxe_qp.c | 204 struct rxe_create_qp_resp __user *uresp) in rxe_qp_init_req() argument 237 err = do_mmap_info(rxe, uresp ? &uresp->sq_mi : NULL, udata, in rxe_qp_init_req() 267 struct rxe_create_qp_resp __user *uresp) in rxe_qp_init_resp() argument 287 err = do_mmap_info(rxe, uresp ? &uresp->rq_mi : NULL, udata, in rxe_qp_init_resp() 310 struct rxe_create_qp_resp __user *uresp, in rxe_qp_from_init() argument 332 err = rxe_qp_init_req(rxe, qp, init, udata, uresp); in rxe_qp_from_init() 336 err = rxe_qp_init_resp(rxe, qp, init, udata, uresp); in rxe_qp_from_init()
|
/drivers/infiniband/hw/ocrdma/ |
D | ocrdma_verbs.c | 929 struct ocrdma_create_cq_uresp uresp; in ocrdma_copy_cq_uresp() local 935 memset(&uresp, 0, sizeof(uresp)); in ocrdma_copy_cq_uresp() 936 uresp.cq_id = cq->id; in ocrdma_copy_cq_uresp() 937 uresp.page_size = PAGE_ALIGN(cq->len); in ocrdma_copy_cq_uresp() 938 uresp.num_pages = 1; in ocrdma_copy_cq_uresp() 939 uresp.max_hw_cqe = cq->max_hw_cqe; in ocrdma_copy_cq_uresp() 940 uresp.page_addr[0] = virt_to_phys(cq->va); in ocrdma_copy_cq_uresp() 941 uresp.db_page_addr = ocrdma_get_db_addr(dev, uctx->cntxt_pd->id); in ocrdma_copy_cq_uresp() 942 uresp.db_page_size = dev->nic_info.db_page_size; in ocrdma_copy_cq_uresp() 943 uresp.phase_change = cq->phase_change ? 1 : 0; in ocrdma_copy_cq_uresp() [all …]
|
/drivers/infiniband/sw/siw/ |
D | siw_verbs.c | 83 struct siw_uresp_alloc_ctx uresp = {}; in siw_alloc_ucontext() local 92 uresp.dev_id = sdev->vendor_part_id; in siw_alloc_ucontext() 94 if (udata->outlen < sizeof(uresp)) { in siw_alloc_ucontext() 98 rv = ib_copy_to_udata(udata, &uresp, sizeof(uresp)); in siw_alloc_ucontext() 435 struct siw_uresp_create_qp uresp = {}; in siw_create_qp() local 437 uresp.num_sqe = num_sqe; in siw_create_qp() 438 uresp.num_rqe = num_rqe; in siw_create_qp() 439 uresp.qp_id = qp_id(qp); in siw_create_qp() 445 length, &uresp.sq_key); in siw_create_qp() 456 length, &uresp.rq_key); in siw_create_qp() [all …]
|
/drivers/infiniband/hw/cxgb4/ |
D | provider.c | 81 struct c4iw_alloc_ucontext_resp uresp; in c4iw_alloc_ucontext() local 90 if (udata->outlen < sizeof(uresp) - sizeof(uresp.reserved)) { in c4iw_alloc_ucontext() 100 uresp.status_page_size = PAGE_SIZE; in c4iw_alloc_ucontext() 103 uresp.status_page_key = context->key; in c4iw_alloc_ucontext() 107 ret = ib_copy_to_udata(udata, &uresp, in c4iw_alloc_ucontext() 108 sizeof(uresp) - sizeof(uresp.reserved)); in c4iw_alloc_ucontext() 112 mm->key = uresp.status_page_key; in c4iw_alloc_ucontext() 224 struct c4iw_alloc_pd_resp uresp = {.pdid = php->pdid}; in c4iw_allocate_pd() local 226 if (ib_copy_to_udata(udata, &uresp, sizeof(uresp))) { in c4iw_allocate_pd()
|
D | cq.c | 1000 struct c4iw_create_cq_resp uresp; in c4iw_create_cq() local 1098 memset(&uresp, 0, sizeof(uresp)); in c4iw_create_cq() 1099 uresp.qid_mask = rhp->rdev.cqmask; in c4iw_create_cq() 1100 uresp.cqid = chp->cq.cqid; in c4iw_create_cq() 1101 uresp.size = chp->cq.size; in c4iw_create_cq() 1102 uresp.memsize = chp->cq.memsize; in c4iw_create_cq() 1104 uresp.key = ucontext->key; in c4iw_create_cq() 1106 uresp.gts_key = ucontext->key; in c4iw_create_cq() 1111 uresp.flags |= C4IW_64B_CQE; in c4iw_create_cq() 1114 ret = ib_copy_to_udata(udata, &uresp, in c4iw_create_cq() [all …]
|
D | qp.c | 2120 struct c4iw_create_qp_resp uresp; in c4iw_create_qp() local 2251 memset(&uresp, 0, sizeof(uresp)); in c4iw_create_qp() 2259 uresp.flags = C4IW_QPF_ONCHIP; in c4iw_create_qp() 2262 uresp.flags |= C4IW_QPF_WRITE_W_IMM; in c4iw_create_qp() 2263 uresp.qid_mask = rhp->rdev.qpmask; in c4iw_create_qp() 2264 uresp.sqid = qhp->wq.sq.qid; in c4iw_create_qp() 2265 uresp.sq_size = qhp->wq.sq.size; in c4iw_create_qp() 2266 uresp.sq_memsize = qhp->wq.sq.memsize; in c4iw_create_qp() 2268 uresp.rqid = qhp->wq.rq.qid; in c4iw_create_qp() 2269 uresp.rq_size = qhp->wq.rq.size; in c4iw_create_qp() [all …]
|
/drivers/infiniband/hw/qedr/ |
D | verbs.c | 264 struct qedr_alloc_ucontext_resp uresp = {}; in qedr_alloc_ucontext() local 317 uresp.dpm_flags = 0; in qedr_alloc_ucontext() 319 uresp.dpm_flags = QEDR_DPM_TYPE_IWARP_LEGACY; in qedr_alloc_ucontext() 321 uresp.dpm_flags = QEDR_DPM_TYPE_ROCE_ENHANCED | in qedr_alloc_ucontext() 326 uresp.dpm_flags |= QEDR_DPM_SIZES_SET; in qedr_alloc_ucontext() 327 uresp.ldpm_limit_size = QEDR_LDPM_MAX_SIZE; in qedr_alloc_ucontext() 328 uresp.edpm_trans_size = QEDR_EDPM_TRANS_SIZE; in qedr_alloc_ucontext() 329 uresp.edpm_limit_size = QEDR_EDPM_MAX_SIZE; in qedr_alloc_ucontext() 332 uresp.wids_enabled = 1; in qedr_alloc_ucontext() 333 uresp.wid_count = oparams.wid_count; in qedr_alloc_ucontext() [all …]
|
/drivers/infiniband/hw/i40iw/ |
D | i40iw_verbs.c | 124 struct i40iw_alloc_ucontext_resp uresp = {}; in i40iw_alloc_ucontext() local 135 uresp.max_qps = iwdev->max_qp; in i40iw_alloc_ucontext() 136 uresp.max_pds = iwdev->max_pd; in i40iw_alloc_ucontext() 137 uresp.wq_size = iwdev->max_qp_wr * 2; in i40iw_alloc_ucontext() 138 uresp.kernel_ver = req.userspace_ver; in i40iw_alloc_ucontext() 143 if (ib_copy_to_udata(udata, &uresp, sizeof(uresp))) in i40iw_alloc_ucontext() 264 struct i40iw_alloc_pd_resp uresp; in i40iw_alloc_pd() local 285 memset(&uresp, 0, sizeof(uresp)); in i40iw_alloc_pd() 286 uresp.pd_id = pd_id; in i40iw_alloc_pd() 287 if (ib_copy_to_udata(udata, &uresp, sizeof(uresp))) { in i40iw_alloc_pd() [all …]
|
/drivers/infiniband/hw/vmw_pvrdma/ |
D | pvrdma_verbs.c | 322 struct pvrdma_alloc_ucontext_resp uresp = {}; in pvrdma_alloc_ucontext() local 350 uresp.qp_tab_size = vdev->dsr->caps.max_qp; in pvrdma_alloc_ucontext() 351 ret = ib_copy_to_udata(udata, &uresp, sizeof(uresp)); in pvrdma_alloc_ucontext()
|
/drivers/infiniband/hw/mthca/ |
D | mthca_provider.c | 298 struct mthca_alloc_ucontext_resp uresp = {}; in mthca_alloc_ucontext() local 305 uresp.qp_tab_size = to_mdev(ibdev)->limits.num_qps; in mthca_alloc_ucontext() 307 uresp.uarc_size = to_mdev(ibdev)->uar_table.uarc_size; in mthca_alloc_ucontext() 309 uresp.uarc_size = 0; in mthca_alloc_ucontext() 322 if (ib_copy_to_udata(udata, &uresp, sizeof(uresp))) { in mthca_alloc_ucontext()
|