Home
last modified time | relevance | path

Searched refs:rdma_entry (Results 1 – 14 of 14) sorted by relevance

/drivers/infiniband/hw/mlx5/
Ddm.c158 context, &mentry->rdma_entry, size, in add_dm_mmap_entry()
180 page_idx = op_entry->mentry.rdma_entry.start_pgoff & 0xFFFF; in copy_op_to_user()
267 rdma_user_mmap_entry_remove(&op_entry->mentry.rdma_entry); in UVERBS_HANDLER()
317 page_idx = dm->mentry.rdma_entry.start_pgoff & 0xFFFF; in handle_alloc_dm_memic()
333 rdma_user_mmap_entry_remove(&dm->mentry.rdma_entry); in handle_alloc_dm_memic()
439 rdma_user_mmap_entry_remove(&entry->mentry.rdma_entry); in dm_memic_remove_ops()
447 rdma_user_mmap_entry_remove(&dm->mentry.rdma_entry); in mlx5_dm_memic_dealloc()
498 page_idx = memic->mentry.rdma_entry.start_pgoff & 0xFFFF; in UVERBS_HANDLER()
Dmlx5_ib.h616 struct rdma_user_mmap_entry rdma_entry; member
1193 to_mmmap(struct rdma_user_mmap_entry *rdma_entry) in to_mmmap() argument
1195 return container_of(rdma_entry, in to_mmmap()
1196 struct mlx5_user_mmap_entry, rdma_entry); in to_mmmap()
Dmain.c2296 rdma_user_mmap_entry_put(&mentry->rdma_entry); in mlx5_ib_mmap_offset()
2302 u64 cmd = (entry->rdma_entry.start_pgoff >> 16) & 0xFFFF; in mlx5_entry_to_mmap_offset()
2303 u64 index = entry->rdma_entry.start_pgoff & 0xFFFF; in mlx5_entry_to_mmap_offset()
3350 rdma_user_mmap_entry_remove(&obj->rdma_entry); in mmap_obj_cleanup()
3359 &c->ibucontext, &entry->rdma_entry, length, in mlx5_rdma_user_mmap_entry_insert()
3430 length = entry->rdma_entry.npages * PAGE_SIZE; in UVERBS_HANDLER()
3557 length = entry->rdma_entry.npages * PAGE_SIZE; in UVERBS_HANDLER()
/drivers/infiniband/hw/efa/
Defa_verbs.c28 struct rdma_user_mmap_entry rdma_entry; member
159 to_emmap(struct rdma_user_mmap_entry *rdma_entry) in to_emmap() argument
161 return container_of(rdma_entry, struct efa_user_mmap_entry, rdma_entry); in to_emmap()
470 err = rdma_user_mmap_entry_insert(ucontext, &entry->rdma_entry, in efa_user_mmap_entry_insert()
476 *offset = rdma_user_mmap_get_offset(&entry->rdma_entry); in efa_user_mmap_entry_insert()
478 return &entry->rdma_entry; in efa_user_mmap_entry_insert()
1737 void efa_mmap_free(struct rdma_user_mmap_entry *rdma_entry) in efa_mmap_free() argument
1739 struct efa_user_mmap_entry *entry = to_emmap(rdma_entry); in efa_mmap_free()
1747 struct rdma_user_mmap_entry *rdma_entry; in __efa_mmap() local
1753 rdma_entry = rdma_user_mmap_entry_get(&ucontext->ibucontext, vma); in __efa_mmap()
[all …]
Defa.h149 void efa_mmap_free(struct rdma_user_mmap_entry *rdma_entry);
/drivers/infiniband/hw/qedr/
Dqedr.h499 struct rdma_user_mmap_entry rdma_entry; member
637 get_qedr_mmap_entry(struct rdma_user_mmap_entry *rdma_entry) in get_qedr_mmap_entry() argument
639 return container_of(rdma_entry, struct qedr_user_mmap_entry, in get_qedr_mmap_entry()
640 rdma_entry); in get_qedr_mmap_entry()
Dverbs.h49 void qedr_mmap_free(struct rdma_user_mmap_entry *rdma_entry);
Dverbs.c309 rc = rdma_user_mmap_entry_insert(uctx, &entry->rdma_entry, in qedr_alloc_ucontext()
315 ctx->db_mmap_entry = &entry->rdma_entry; in qedr_alloc_ucontext()
374 void qedr_mmap_free(struct rdma_user_mmap_entry *rdma_entry) in qedr_mmap_free() argument
376 struct qedr_user_mmap_entry *entry = get_qedr_mmap_entry(rdma_entry); in qedr_mmap_free()
391 struct rdma_user_mmap_entry *rdma_entry; in qedr_mmap() local
400 rdma_entry = rdma_user_mmap_entry_get(ucontext, vma); in qedr_mmap()
401 if (!rdma_entry) { in qedr_mmap()
406 entry = get_qedr_mmap_entry(rdma_entry); in qedr_mmap()
416 rdma_entry); in qedr_mmap()
431 rdma_user_mmap_entry_put(rdma_entry); in qedr_mmap()
[all …]
/drivers/infiniband/sw/siw/
Dsiw_verbs.c37 void siw_mmap_free(struct rdma_user_mmap_entry *rdma_entry) in siw_mmap_free() argument
39 struct siw_user_mmap_entry *entry = to_siw_mmap_entry(rdma_entry); in siw_mmap_free()
48 struct rdma_user_mmap_entry *rdma_entry; in siw_mmap() local
59 rdma_entry = rdma_user_mmap_entry_get(&uctx->base_ucontext, vma); in siw_mmap()
60 if (!rdma_entry) { in siw_mmap()
65 entry = to_siw_mmap_entry(rdma_entry); in siw_mmap()
74 rdma_user_mmap_entry_put(rdma_entry); in siw_mmap()
271 &entry->rdma_entry, in siw_mmap_entry_insert()
278 *offset = rdma_user_mmap_get_offset(&entry->rdma_entry); in siw_mmap_entry_insert()
280 return &entry->rdma_entry; in siw_mmap_entry_insert()
Dsiw_verbs.h84 void siw_mmap_free(struct rdma_user_mmap_entry *rdma_entry);
Dsiw.h488 struct rdma_user_mmap_entry rdma_entry; member
594 return container_of(rdma_mmap, struct siw_user_mmap_entry, rdma_entry); in to_siw_mmap_entry()
/drivers/infiniband/hw/irdma/
Dmain.h367 to_irdma_mmap_entry(struct rdma_user_mmap_entry *rdma_entry) in to_irdma_mmap_entry() argument
369 return container_of(rdma_entry, struct irdma_user_mmap_entry, in to_irdma_mmap_entry()
370 rdma_entry); in to_irdma_mmap_entry()
Dverbs.h216 struct rdma_user_mmap_entry rdma_entry; member
Dverbs.c132 static void irdma_mmap_free(struct rdma_user_mmap_entry *rdma_entry) in irdma_mmap_free() argument
134 struct irdma_user_mmap_entry *entry = to_irdma_mmap_entry(rdma_entry); in irdma_mmap_free()
153 &entry->rdma_entry, PAGE_SIZE); in irdma_user_mmap_entry_insert()
158 *mmap_offset = rdma_user_mmap_get_offset(&entry->rdma_entry); in irdma_user_mmap_entry_insert()
160 return &entry->rdma_entry; in irdma_user_mmap_entry_insert()
170 struct rdma_user_mmap_entry *rdma_entry; in irdma_mmap() local
182 rdma_entry = rdma_user_mmap_entry_get(&ucontext->ibucontext, vma); in irdma_mmap()
183 if (!rdma_entry) { in irdma_mmap()
190 entry = to_irdma_mmap_entry(rdma_entry); in irdma_mmap()
202 rdma_entry); in irdma_mmap()
[all …]