Home
last modified time | relevance | path

Searched refs:udata (Results 1 – 25 of 108) sorted by relevance

12345

/drivers/infiniband/sw/siw/
Dsiw_verbs.h37 int siw_alloc_ucontext(struct ib_ucontext *base_ctx, struct ib_udata *udata);
44 struct ib_udata *udata);
46 struct ib_udata *udata);
52 int siw_alloc_pd(struct ib_pd *base_pd, struct ib_udata *udata);
53 void siw_dealloc_pd(struct ib_pd *base_pd, struct ib_udata *udata);
56 struct ib_udata *udata);
60 int attr_mask, struct ib_udata *udata);
61 int siw_destroy_qp(struct ib_qp *base_qp, struct ib_udata *udata);
66 void siw_destroy_cq(struct ib_cq *base_cq, struct ib_udata *udata);
70 u64 rnic_va, int rights, struct ib_udata *udata);
[all …]
Dsiw_verbs.c97 int siw_alloc_ucontext(struct ib_ucontext *base_ctx, struct ib_udata *udata) in siw_alloc_ucontext() argument
114 if (udata->outlen < sizeof(uresp)) { in siw_alloc_ucontext()
118 rv = ib_copy_to_udata(udata, &uresp, sizeof(uresp)); in siw_alloc_ucontext()
155 struct ib_udata *udata) in siw_query_device() argument
159 if (udata->inlen || udata->outlen) in siw_query_device()
265 int siw_alloc_pd(struct ib_pd *pd, struct ib_udata *udata) in siw_alloc_pd() argument
278 void siw_dealloc_pd(struct ib_pd *pd, struct ib_udata *udata) in siw_dealloc_pd() argument
308 struct ib_udata *udata) in siw_create_qp() argument
315 rdma_udata_to_drv_context(udata, struct siw_ucontext, in siw_create_qp()
382 qp->kernel_verbs = !udata; in siw_create_qp()
[all …]
/drivers/infiniband/hw/qedr/
Dverbs.h36 struct ib_device_attr *attr, struct ib_udata *udata);
46 int qedr_alloc_ucontext(struct ib_ucontext *uctx, struct ib_udata *udata);
50 int qedr_alloc_pd(struct ib_pd *pd, struct ib_udata *udata);
51 void qedr_dealloc_pd(struct ib_pd *pd, struct ib_udata *udata);
54 struct ib_udata *udata);
56 void qedr_destroy_cq(struct ib_cq *ibcq, struct ib_udata *udata);
61 int attr_mask, struct ib_udata *udata);
64 int qedr_destroy_qp(struct ib_qp *ibqp, struct ib_udata *udata);
67 struct ib_udata *udata);
69 enum ib_srq_attr_mask attr_mask, struct ib_udata *udata);
[all …]
/drivers/infiniband/hw/bnxt_re/
Dib_verbs.h147 struct ib_udata *udata);
164 int bnxt_re_alloc_pd(struct ib_pd *pd, struct ib_udata *udata);
165 void bnxt_re_dealloc_pd(struct ib_pd *pd, struct ib_udata *udata);
167 struct ib_udata *udata);
173 struct ib_udata *udata);
176 struct ib_udata *udata);
178 void bnxt_re_destroy_srq(struct ib_srq *srq, struct ib_udata *udata);
183 struct ib_udata *udata);
185 int qp_attr_mask, struct ib_udata *udata);
188 int bnxt_re_destroy_qp(struct ib_qp *qp, struct ib_udata *udata);
[all …]
/drivers/infiniband/hw/usnic/
Dusnic_ib_verbs.h53 int usnic_ib_alloc_pd(struct ib_pd *ibpd, struct ib_udata *udata);
54 void usnic_ib_dealloc_pd(struct ib_pd *pd, struct ib_udata *udata);
57 struct ib_udata *udata);
58 int usnic_ib_destroy_qp(struct ib_qp *qp, struct ib_udata *udata);
60 int attr_mask, struct ib_udata *udata);
62 struct ib_udata *udata);
63 void usnic_ib_destroy_cq(struct ib_cq *cq, struct ib_udata *udata);
66 struct ib_udata *udata);
67 int usnic_ib_dereg_mr(struct ib_mr *ibmr, struct ib_udata *udata);
68 int usnic_ib_alloc_ucontext(struct ib_ucontext *uctx, struct ib_udata *udata);
/drivers/infiniband/hw/ocrdma/
Docrdma_verbs.h66 int ocrdma_alloc_ucontext(struct ib_ucontext *uctx, struct ib_udata *udata);
71 int ocrdma_alloc_pd(struct ib_pd *pd, struct ib_udata *udata);
72 void ocrdma_dealloc_pd(struct ib_pd *pd, struct ib_udata *udata);
75 struct ib_udata *udata);
77 void ocrdma_destroy_cq(struct ib_cq *ibcq, struct ib_udata *udata);
85 int attr_mask, struct ib_udata *udata);
89 int ocrdma_destroy_qp(struct ib_qp *ibqp, struct ib_udata *udata);
93 struct ib_udata *udata);
97 void ocrdma_destroy_srq(struct ib_srq *ibsrq, struct ib_udata *udata);
101 int ocrdma_dereg_mr(struct ib_mr *ib_mr, struct ib_udata *udata);
[all …]
/drivers/infiniband/sw/rdmavt/
Dsrq.c77 struct ib_udata *udata) in rvt_create_srq() argument
101 dev->dparms.node, udata)) { in rvt_create_srq()
110 if (udata && udata->outlen >= sizeof(__u64)) { in rvt_create_srq()
113 srq->ip = rvt_create_mmap_info(dev, s, udata, srq->rq.wq); in rvt_create_srq()
119 ret = ib_copy_to_udata(udata, &srq->ip->offset, in rvt_create_srq()
168 struct ib_udata *udata) in rvt_modify_srq() argument
190 udata)) in rvt_modify_srq()
193 if (udata && udata->inlen >= sizeof(__u64)) { in rvt_modify_srq()
197 ret = ib_copy_from_udata(&offset_addr, udata, in rvt_modify_srq()
201 udata->outbuf = (void __user *) in rvt_modify_srq()
[all …]
Dcq.c201 struct ib_udata *udata) in rvt_create_cq() argument
231 if (udata && udata->outlen >= sizeof(__u64)) { in rvt_create_cq()
249 if (udata && udata->outlen >= sizeof(__u64)) { in rvt_create_cq()
250 cq->ip = rvt_create_mmap_info(rdi, sz, udata, u_wc); in rvt_create_cq()
256 err = ib_copy_to_udata(udata, &cq->ip->offset, in rvt_create_cq()
318 void rvt_destroy_cq(struct ib_cq *ibcq, struct ib_udata *udata) in rvt_destroy_cq() argument
379 int rvt_resize_cq(struct ib_cq *ibcq, int cqe, struct ib_udata *udata) in rvt_resize_cq() argument
397 if (udata && udata->outlen >= sizeof(__u64)) { in rvt_resize_cq()
411 if (udata && udata->outlen >= sizeof(__u64)) { in rvt_resize_cq()
414 ret = ib_copy_to_udata(udata, &offset, sizeof(offset)); in rvt_resize_cq()
[all …]
Dqp.h57 struct ib_udata *udata);
59 int attr_mask, struct ib_udata *udata);
60 int rvt_destroy_qp(struct ib_qp *ibqp, struct ib_udata *udata);
72 struct ib_udata *udata);
/drivers/infiniband/hw/efa/
Defa.h121 struct ib_udata *udata);
131 int efa_alloc_pd(struct ib_pd *ibpd, struct ib_udata *udata);
132 void efa_dealloc_pd(struct ib_pd *ibpd, struct ib_udata *udata);
133 int efa_destroy_qp(struct ib_qp *ibqp, struct ib_udata *udata);
136 struct ib_udata *udata);
137 void efa_destroy_cq(struct ib_cq *ibcq, struct ib_udata *udata);
139 struct ib_udata *udata);
142 struct ib_udata *udata);
143 int efa_dereg_mr(struct ib_mr *ibmr, struct ib_udata *udata);
146 int efa_alloc_ucontext(struct ib_ucontext *ibucontext, struct ib_udata *udata);
[all …]
Defa_verbs.c277 struct ib_udata *udata) in efa_query_device() argument
284 if (udata && udata->inlen && in efa_query_device()
285 !ib_is_udata_cleared(udata, 0, udata->inlen)) { in efa_query_device()
310 if (udata && udata->outlen) { in efa_query_device()
316 err = ib_copy_to_udata(udata, &resp, in efa_query_device()
317 min(sizeof(resp), udata->outlen)); in efa_query_device()
428 int efa_alloc_pd(struct ib_pd *ibpd, struct ib_udata *udata) in efa_alloc_pd() argument
436 if (udata->inlen && in efa_alloc_pd()
437 !ib_is_udata_cleared(udata, 0, udata->inlen)) { in efa_alloc_pd()
451 if (udata->outlen) { in efa_alloc_pd()
[all …]
/drivers/infiniband/hw/vmw_pvrdma/
Dpvrdma_verbs.h385 struct ib_udata *udata);
399 int pvrdma_alloc_ucontext(struct ib_ucontext *uctx, struct ib_udata *udata);
401 int pvrdma_alloc_pd(struct ib_pd *pd, struct ib_udata *udata);
402 void pvrdma_dealloc_pd(struct ib_pd *ibpd, struct ib_udata *udata);
406 struct ib_udata *udata);
407 int pvrdma_dereg_mr(struct ib_mr *mr, struct ib_udata *udata);
409 u32 max_num_sg, struct ib_udata *udata);
413 struct ib_udata *udata);
414 void pvrdma_destroy_cq(struct ib_cq *cq, struct ib_udata *udata);
418 struct ib_udata *udata);
[all …]
Dpvrdma_srq.c100 struct ib_udata *udata) in pvrdma_create_srq() argument
113 if (!udata) { in pvrdma_create_srq()
144 if (ib_copy_from_udata(&ucmd, udata, sizeof(ucmd))) { in pvrdma_create_srq()
149 srq->umem = ib_umem_get(udata, ucmd.buf_addr, ucmd.buf_size, 0, 0); in pvrdma_create_srq()
198 if (ib_copy_to_udata(udata, &srq_resp, sizeof(srq_resp))) { in pvrdma_create_srq()
200 pvrdma_destroy_srq(&srq->ibsrq, udata); in pvrdma_create_srq()
243 void pvrdma_destroy_srq(struct ib_srq *srq, struct ib_udata *udata) in pvrdma_destroy_srq() argument
274 enum ib_srq_attr_mask attr_mask, struct ib_udata *udata) in pvrdma_modify_srq() argument
/drivers/infiniband/hw/mthca/
Dmthca_provider.c305 struct ib_udata *udata) in mthca_alloc_ucontext() argument
332 if (ib_copy_to_udata(udata, &uresp, sizeof(uresp))) { in mthca_alloc_ucontext()
366 static int mthca_alloc_pd(struct ib_pd *ibpd, struct ib_udata *udata) in mthca_alloc_pd() argument
372 err = mthca_pd_alloc(to_mdev(ibdev), !udata, pd); in mthca_alloc_pd()
376 if (udata) { in mthca_alloc_pd()
377 if (ib_copy_to_udata(udata, &pd->pd_num, sizeof (__u32))) { in mthca_alloc_pd()
386 static void mthca_dealloc_pd(struct ib_pd *pd, struct ib_udata *udata) in mthca_dealloc_pd() argument
392 u32 flags, struct ib_udata *udata) in mthca_ah_create() argument
408 struct ib_udata *udata) in mthca_create_srq() argument
412 udata, struct mthca_ucontext, ibucontext); in mthca_create_srq()
[all …]
Dmthca_srq.c101 struct ib_udata *udata) in mthca_tavor_init_srq_context() argument
104 udata, struct mthca_ucontext, ibucontext); in mthca_tavor_init_srq_context()
112 if (udata) in mthca_tavor_init_srq_context()
122 struct ib_udata *udata) in mthca_arbel_init_srq_context() argument
125 udata, struct mthca_ucontext, ibucontext); in mthca_arbel_init_srq_context()
140 if (udata) in mthca_arbel_init_srq_context()
155 struct mthca_srq *srq, struct ib_udata *udata) in mthca_alloc_srq_buf() argument
162 if (udata) in mthca_alloc_srq_buf()
208 struct ib_udata *udata) in mthca_alloc_srq() argument
246 if (!udata) { in mthca_alloc_srq()
[all …]
/drivers/infiniband/hw/hns/
Dhns_roce_cq.c208 struct ib_udata *udata, in hns_roce_ib_get_cq_umem() argument
216 *umem = ib_umem_get(udata, buf_addr, cqe * hr_dev->caps.cq_entry_sz, in hns_roce_ib_get_cq_umem()
301 struct ib_udata *udata, in create_user_cq() argument
309 udata, struct hns_roce_ucontext, ibucontext); in create_user_cq()
311 if (ib_copy_from_udata(&ucmd, udata, sizeof(ucmd))) { in create_user_cq()
317 ret = hns_roce_ib_get_cq_umem(hr_dev, udata, &hr_cq->hr_buf, in create_user_cq()
326 (udata->outlen >= sizeof(*resp))) { in create_user_cq()
327 ret = hns_roce_db_map_user(context, udata, ucmd.db_addr, in create_user_cq()
385 struct ib_udata *udata, in destroy_user_cq() argument
389 udata, struct hns_roce_ucontext, ibucontext); in destroy_user_cq()
[all …]
Dhns_roce_srq.c178 static int create_user_srq(struct hns_roce_srq *srq, struct ib_udata *udata, in create_user_srq() argument
186 if (ib_copy_from_udata(&ucmd, udata, sizeof(ucmd))) in create_user_srq()
189 srq->umem = ib_umem_get(udata, ucmd.buf_addr, srq_buf_size, 0, 0); in create_user_srq()
208 srq->idx_que.umem = ib_umem_get(udata, ucmd.que_addr, in create_user_srq()
362 struct ib_udata *udata) in hns_roce_create_srq() argument
394 if (udata) { in hns_roce_create_srq()
395 ret = create_user_srq(srq, udata, srq_buf_size); in hns_roce_create_srq()
421 if (udata) { in hns_roce_create_srq()
422 if (ib_copy_to_udata(udata, &resp, in hns_roce_create_srq()
423 min(udata->outlen, sizeof(resp)))) { in hns_roce_create_srq()
[all …]
/drivers/infiniband/hw/mlx5/
Dsrq.c46 struct ib_udata *udata, int buf_size) in create_srq_user() argument
51 udata, struct mlx5_ib_ucontext, ibucontext); in create_srq_user()
60 ucmdlen = min(udata->inlen, sizeof(ucmd)); in create_srq_user()
62 if (ib_copy_from_udata(&ucmd, udata, ucmdlen)) { in create_srq_user()
70 if (udata->inlen > sizeof(ucmd) && in create_srq_user()
71 !ib_is_udata_cleared(udata, sizeof(ucmd), in create_srq_user()
72 udata->inlen - sizeof(ucmd))) in create_srq_user()
76 err = get_srq_user_index(ucontext, &ucmd, udata->inlen, &uidx); in create_srq_user()
83 srq->umem = ib_umem_get(udata, ucmd.buf_addr, buf_size, 0, 0); in create_srq_user()
107 err = mlx5_ib_db_map_user(ucontext, udata, ucmd.db_addr, &srq->db); in create_srq_user()
[all …]
Dmlx5_ib.h1019 static inline struct mlx5_ib_dev *mlx5_udata_to_mdev(struct ib_udata *udata) in mlx5_udata_to_mdev() argument
1022 udata, struct mlx5_ib_ucontext, ibucontext); in mlx5_udata_to_mdev()
1099 struct ib_udata *udata, unsigned long virt,
1106 struct ib_udata *udata);
1110 struct ib_udata *udata);
1112 enum ib_srq_attr_mask attr_mask, struct ib_udata *udata);
1114 void mlx5_ib_destroy_srq(struct ib_srq *srq, struct ib_udata *udata);
1121 struct ib_udata *udata);
1123 int attr_mask, struct ib_udata *udata);
1126 int mlx5_ib_destroy_qp(struct ib_qp *qp, struct ib_udata *udata);
[all …]
/drivers/infiniband/sw/rxe/
Drxe_verbs.c145 static int rxe_alloc_ucontext(struct ib_ucontext *uctx, struct ib_udata *udata) in rxe_alloc_ucontext() argument
179 static int rxe_alloc_pd(struct ib_pd *ibpd, struct ib_udata *udata) in rxe_alloc_pd() argument
187 static void rxe_dealloc_pd(struct ib_pd *ibpd, struct ib_udata *udata) in rxe_dealloc_pd() argument
195 u32 flags, struct ib_udata *udata) in rxe_create_ah() argument
293 struct ib_udata *udata) in rxe_create_srq() argument
301 if (udata) { in rxe_create_srq()
302 if (udata->outlen < sizeof(*uresp)) in rxe_create_srq()
304 uresp = udata->outbuf; in rxe_create_srq()
318 err = rxe_srq_from_init(rxe, srq, init, udata, uresp); in rxe_create_srq()
333 struct ib_udata *udata) in rxe_modify_srq() argument
[all …]
/drivers/infiniband/hw/mlx4/
Dsrq.c74 struct ib_udata *udata) in mlx4_ib_create_srq() argument
78 udata, struct mlx4_ib_ucontext, ibucontext); in mlx4_ib_create_srq()
107 if (udata) { in mlx4_ib_create_srq()
110 if (ib_copy_from_udata(&ucmd, udata, sizeof(ucmd))) in mlx4_ib_create_srq()
113 srq->umem = ib_umem_get(udata, ucmd.buf_addr, buf_size, 0, 0); in mlx4_ib_create_srq()
126 err = mlx4_ib_db_map_user(udata, ucmd.db_addr, &srq->db); in mlx4_ib_create_srq()
187 if (udata) in mlx4_ib_create_srq()
188 if (ib_copy_to_udata(udata, &srq->msrq.srqn, sizeof (__u32))) { in mlx4_ib_create_srq()
198 if (udata) in mlx4_ib_create_srq()
212 if (!udata) in mlx4_ib_create_srq()
[all …]
Dmlx4_ib.h726 int mlx4_ib_db_map_user(struct ib_udata *udata, unsigned long virt,
735 struct ib_udata *udata);
736 int mlx4_ib_dereg_mr(struct ib_mr *mr, struct ib_udata *udata);
738 struct ib_udata *udata);
741 u32 max_num_sg, struct ib_udata *udata);
745 int mlx4_ib_resize_cq(struct ib_cq *ibcq, int entries, struct ib_udata *udata);
747 struct ib_udata *udata);
748 void mlx4_ib_destroy_cq(struct ib_cq *cq, struct ib_udata *udata);
755 struct ib_udata *udata);
762 struct ib_udata *udata);
[all …]
Dcq.c138 static int mlx4_ib_get_cq_umem(struct mlx4_ib_dev *dev, struct ib_udata *udata, in mlx4_ib_get_cq_umem() argument
147 *umem = ib_umem_get(udata, buf_addr, cqe * cqe_size, in mlx4_ib_get_cq_umem()
176 struct ib_udata *udata) in mlx4_ib_create_cq() argument
187 udata, struct mlx4_ib_ucontext, ibucontext); in mlx4_ib_create_cq()
205 if (udata) { in mlx4_ib_create_cq()
208 if (ib_copy_from_udata(&ucmd, udata, sizeof ucmd)) { in mlx4_ib_create_cq()
214 err = mlx4_ib_get_cq_umem(dev, udata, &cq->buf, &cq->umem, in mlx4_ib_create_cq()
219 err = mlx4_ib_db_map_user(udata, ucmd.db_addr, &cq->db); in mlx4_ib_create_cq()
252 buf_addr, !!udata); in mlx4_ib_create_cq()
256 if (udata) in mlx4_ib_create_cq()
[all …]
/drivers/infiniband/hw/cxgb3/
Diwch_provider.c78 struct ib_udata *udata) in iwch_alloc_ucontext() argument
91 static void iwch_destroy_cq(struct ib_cq *ib_cq, struct ib_udata *udata) in iwch_destroy_cq() argument
107 struct ib_udata *udata) in iwch_create_cq() argument
122 if (udata) { in iwch_create_cq()
124 if (ib_copy_from_udata(&ureq, udata, sizeof(ureq))) in iwch_create_cq()
147 if (cxio_create_cq(&rhp->rdev, &chp->cq, !udata)) in iwch_create_cq()
161 if (udata) { in iwch_create_cq()
164 udata, struct iwch_ucontext, ibucontext); in iwch_create_cq()
168 iwch_destroy_cq(&chp->ibcq, udata); in iwch_create_cq()
179 if (udata->outlen < sizeof(uresp)) { in iwch_create_cq()
[all …]
/drivers/xen/
Dprivcmd.c70 static long privcmd_ioctl_hypercall(struct file *file, void __user *udata) in privcmd_ioctl_hypercall() argument
80 if (copy_from_user(&hypercall, udata, sizeof(hypercall))) in privcmd_ioctl_hypercall()
253 static long privcmd_ioctl_mmap(struct file *file, void __user *udata) in privcmd_ioctl_mmap() argument
267 if (copy_from_user(&mmapcmd, udata, sizeof(mmapcmd))) in privcmd_ioctl_mmap()
446 struct file *file, void __user *udata, int version) in privcmd_ioctl_mmap_batch() argument
459 if (copy_from_user(&m, udata, sizeof(struct privcmd_mmapbatch))) in privcmd_ioctl_mmap_batch()
467 if (copy_from_user(&m, udata, sizeof(struct privcmd_mmapbatch_v2))) in privcmd_ioctl_mmap_batch()
625 static long privcmd_ioctl_dm_op(struct file *file, void __user *udata) in privcmd_ioctl_dm_op() argument
636 if (copy_from_user(&kdata, udata, sizeof(kdata))) in privcmd_ioctl_dm_op()
710 static long privcmd_ioctl_restrict(struct file *file, void __user *udata) in privcmd_ioctl_restrict() argument
[all …]

12345