Home
last modified time | relevance | path

Searched full:unmap (Results 1 – 25 of 1537) sorted by relevance

12345678910>>...62

/kernel/linux/linux-4.19/crypto/async_tx/
Dasync_xor.c36 do_async_xor(struct dma_chan *chan, struct dmaengine_unmap_data *unmap, in do_async_xor() argument
45 int src_cnt = unmap->to_cnt; in do_async_xor()
47 dma_addr_t dma_dest = unmap->addr[unmap->to_cnt]; in do_async_xor()
48 dma_addr_t *src_list = unmap->addr; in do_async_xor()
76 if (src_list > unmap->addr) in do_async_xor()
79 xor_src_cnt, unmap->len, in do_async_xor()
90 xor_src_cnt, unmap->len, in do_async_xor()
95 dma_set_unmap(tx, unmap); in do_async_xor()
180 struct dmaengine_unmap_data *unmap = NULL; in async_xor() local
185 unmap = dmaengine_get_unmap_data(device->dev, src_cnt+1, GFP_NOWAIT); in async_xor()
[all …]
Dasync_pq.c53 struct dmaengine_unmap_data *unmap, in do_async_gen_syndrome() argument
92 dma_dest[0] = unmap->addr[disks - 2]; in do_async_gen_syndrome()
93 dma_dest[1] = unmap->addr[disks - 1]; in do_async_gen_syndrome()
95 &unmap->addr[src_off], in do_async_gen_syndrome()
97 &scfs[src_off], unmap->len, in do_async_gen_syndrome()
105 dma_set_unmap(tx, unmap); in do_async_gen_syndrome()
187 struct dmaengine_unmap_data *unmap = NULL; in async_gen_syndrome() local
192 unmap = dmaengine_get_unmap_data(device->dev, disks, GFP_NOWAIT); in async_gen_syndrome()
195 if (unmap && !(submit->flags & ASYNC_TX_PQ_XOR_DST) && in async_gen_syndrome()
211 unmap->len = len; in async_gen_syndrome()
[all …]
Dasync_memcpy.c53 struct dmaengine_unmap_data *unmap = NULL; in async_memcpy() local
56 unmap = dmaengine_get_unmap_data(device->dev, 2, GFP_NOWAIT); in async_memcpy()
58 if (unmap && is_dma_copy_aligned(device, src_offset, dest_offset, len)) { in async_memcpy()
66 unmap->to_cnt = 1; in async_memcpy()
67 unmap->addr[0] = dma_map_page(device->dev, src, src_offset, len, in async_memcpy()
69 unmap->from_cnt = 1; in async_memcpy()
70 unmap->addr[1] = dma_map_page(device->dev, dest, dest_offset, len, in async_memcpy()
72 unmap->len = len; in async_memcpy()
74 tx = device->device_prep_dma_memcpy(chan, unmap->addr[1], in async_memcpy()
75 unmap->addr[0], len, in async_memcpy()
[all …]
Dasync_raid6_recov.c38 struct dmaengine_unmap_data *unmap = NULL; in async_sum_product() local
44 unmap = dmaengine_get_unmap_data(dma->dev, 3, GFP_NOWAIT); in async_sum_product()
46 if (unmap) { in async_sum_product()
54 unmap->addr[0] = dma_map_page(dev, srcs[0], 0, len, DMA_TO_DEVICE); in async_sum_product()
55 unmap->addr[1] = dma_map_page(dev, srcs[1], 0, len, DMA_TO_DEVICE); in async_sum_product()
56 unmap->to_cnt = 2; in async_sum_product()
58 unmap->addr[2] = dma_map_page(dev, dest, 0, len, DMA_BIDIRECTIONAL); in async_sum_product()
59 unmap->bidi_cnt = 1; in async_sum_product()
61 pq[1] = unmap->addr[2]; in async_sum_product()
63 unmap->len = len; in async_sum_product()
[all …]
/kernel/linux/linux-5.10/crypto/async_tx/
Dasync_xor.c23 do_async_xor(struct dma_chan *chan, struct dmaengine_unmap_data *unmap, in do_async_xor() argument
32 int src_cnt = unmap->to_cnt; in do_async_xor()
34 dma_addr_t dma_dest = unmap->addr[unmap->to_cnt]; in do_async_xor()
35 dma_addr_t *src_list = unmap->addr; in do_async_xor()
63 if (src_list > unmap->addr) in do_async_xor()
66 xor_src_cnt, unmap->len, in do_async_xor()
77 xor_src_cnt, unmap->len, in do_async_xor()
82 dma_set_unmap(tx, unmap); in do_async_xor()
190 struct dmaengine_unmap_data *unmap = NULL; in async_xor_offs() local
195 unmap = dmaengine_get_unmap_data(device->dev, src_cnt+1, GFP_NOWAIT); in async_xor_offs()
[all …]
Dasync_pq.c37 struct dmaengine_unmap_data *unmap, in do_async_gen_syndrome() argument
76 dma_dest[0] = unmap->addr[disks - 2]; in do_async_gen_syndrome()
77 dma_dest[1] = unmap->addr[disks - 1]; in do_async_gen_syndrome()
79 &unmap->addr[src_off], in do_async_gen_syndrome()
81 &scfs[src_off], unmap->len, in do_async_gen_syndrome()
89 dma_set_unmap(tx, unmap); in do_async_gen_syndrome()
185 struct dmaengine_unmap_data *unmap = NULL; in async_gen_syndrome() local
190 unmap = dmaengine_get_unmap_data(device->dev, disks, GFP_NOWAIT); in async_gen_syndrome()
193 if (unmap && !(submit->flags & ASYNC_TX_PQ_XOR_DST) && in async_gen_syndrome()
209 unmap->len = len; in async_gen_syndrome()
[all …]
Dasync_memcpy.c40 struct dmaengine_unmap_data *unmap = NULL; in async_memcpy() local
43 unmap = dmaengine_get_unmap_data(device->dev, 2, GFP_NOWAIT); in async_memcpy()
45 if (unmap && is_dma_copy_aligned(device, src_offset, dest_offset, len)) { in async_memcpy()
53 unmap->to_cnt = 1; in async_memcpy()
54 unmap->addr[0] = dma_map_page(device->dev, src, src_offset, len, in async_memcpy()
56 unmap->from_cnt = 1; in async_memcpy()
57 unmap->addr[1] = dma_map_page(device->dev, dest, dest_offset, len, in async_memcpy()
59 unmap->len = len; in async_memcpy()
61 tx = device->device_prep_dma_memcpy(chan, unmap->addr[1], in async_memcpy()
62 unmap->addr[0], len, in async_memcpy()
[all …]
Dasync_raid6_recov.c25 struct dmaengine_unmap_data *unmap = NULL; in async_sum_product() local
31 unmap = dmaengine_get_unmap_data(dma->dev, 3, GFP_NOWAIT); in async_sum_product()
33 if (unmap) { in async_sum_product()
41 unmap->addr[0] = dma_map_page(dev, srcs[0], src_offs[0], in async_sum_product()
43 unmap->addr[1] = dma_map_page(dev, srcs[1], src_offs[1], in async_sum_product()
45 unmap->to_cnt = 2; in async_sum_product()
47 unmap->addr[2] = dma_map_page(dev, dest, d_off, in async_sum_product()
49 unmap->bidi_cnt = 1; in async_sum_product()
51 pq[1] = unmap->addr[2]; in async_sum_product()
53 unmap->len = len; in async_sum_product()
[all …]
/kernel/linux/linux-4.19/tools/testing/selftests/vm/
Dmlock2-tests.c210 goto unmap; in test_mlock_lock()
214 goto unmap; in test_mlock_lock()
219 goto unmap; in test_mlock_lock()
224 unmap: in test_mlock_lock()
273 goto unmap; in test_mlock_onfault()
277 goto unmap; in test_mlock_onfault()
286 goto unmap; in test_mlock_onfault()
290 unmap: in test_mlock_onfault()
317 goto unmap; in test_lock_onfault_of_present()
323 goto unmap; in test_lock_onfault_of_present()
[all …]
/kernel/linux/linux-5.10/tools/testing/selftests/vm/
Dmlock2-tests.c210 goto unmap; in test_mlock_lock()
214 goto unmap; in test_mlock_lock()
219 goto unmap; in test_mlock_lock()
224 unmap: in test_mlock_lock()
273 goto unmap; in test_mlock_onfault()
277 goto unmap; in test_mlock_onfault()
286 goto unmap; in test_mlock_onfault()
290 unmap: in test_mlock_onfault()
317 goto unmap; in test_lock_onfault_of_present()
323 goto unmap; in test_lock_onfault_of_present()
[all …]
Dmremap_dontunmap.c62 "unable to unmap destination mapping"); in kernel_support_for_mremap_dontunmap()
66 "unable to unmap source mapping"); in kernel_support_for_mremap_dontunmap()
125 "unable to unmap destination mapping"); in mremap_dontunmap_simple()
127 "unable to unmap source mapping"); in mremap_dontunmap_simple()
159 // The dest mapping will have been unmap by mremap so we expect the Xs in mremap_dontunmap_simple_fixed()
171 "unable to unmap destination mapping"); in mremap_dontunmap_simple_fixed()
173 "unable to unmap source mapping"); in mremap_dontunmap_simple_fixed()
221 "unable to unmap destination mapping"); in mremap_dontunmap_partial_mapping()
223 "unable to unmap source mapping"); in mremap_dontunmap_partial_mapping()
280 "unable to unmap destination mapping"); in mremap_dontunmap_partial_mapping_overwrite()
[all …]
/kernel/linux/linux-5.10/drivers/xen/xenbus/
Dxenbus_client.c80 struct gnttab_unmap_grant_ref unmap[XENBUS_MAX_RING_GRANTS]; member
92 int (*unmap)(struct xenbus_device *dev, void *vaddr); member
538 gnttab_set_unmap_op(&info->unmap[j], in __xenbus_map_ring()
545 if (HYPERVISOR_grant_table_op(GNTTABOP_unmap_grant_ref, info->unmap, j)) in __xenbus_map_ring()
550 if (info->unmap[i].status != GNTST_okay) { in __xenbus_map_ring()
564 * @vaddrs: addresses to unmap
566 * Unmap memory in this domain that was imported from another domain.
573 struct gnttab_unmap_grant_ref unmap[XENBUS_MAX_RING_GRANTS]; in xenbus_unmap_ring() local
581 gnttab_set_unmap_op(&unmap[i], vaddrs[i], in xenbus_unmap_ring()
584 if (HYPERVISOR_grant_table_op(GNTTABOP_unmap_grant_ref, unmap, i)) in xenbus_unmap_ring()
[all …]
/kernel/linux/linux-4.19/drivers/xen/xenbus/
Dxenbus_client.c79 int (*unmap)(struct xenbus_device *dev, void *vaddr); member
486 struct gnttab_unmap_grant_ref unmap[XENBUS_MAX_RING_GRANTS]; in __xenbus_map_ring() local
518 memset(&unmap[j], 0, sizeof(unmap[j])); in __xenbus_map_ring()
519 gnttab_set_unmap_op(&unmap[j], (phys_addr_t)addrs[i], in __xenbus_map_ring()
525 if (HYPERVISOR_grant_table_op(GNTTABOP_unmap_grant_ref, unmap, j)) in __xenbus_map_ring()
530 if (unmap[i].status != GNTST_okay) { in __xenbus_map_ring()
672 * @vaddr: addr to unmap
675 * Unmap a page of memory in this domain that was imported from another domain.
683 return ring_ops->unmap(dev, vaddr); in xenbus_unmap_ring_vfree()
749 struct gnttab_unmap_grant_ref unmap[XENBUS_MAX_RING_GRANTS]; in xenbus_unmap_ring_vfree_pv() local
[all …]
/kernel/linux/linux-4.19/arch/arm/xen/
Dp2m.c94 struct gnttab_unmap_grant_ref unmap; in set_foreign_p2m_mapping() local
108 unmap.host_addr = map_ops[i].host_addr, in set_foreign_p2m_mapping()
109 unmap.handle = map_ops[i].handle; in set_foreign_p2m_mapping()
112 unmap.dev_bus_addr = map_ops[i].dev_bus_addr; in set_foreign_p2m_mapping()
114 unmap.dev_bus_addr = 0; in set_foreign_p2m_mapping()
120 unmap.status = 1; in set_foreign_p2m_mapping()
123 &unmap, 1); in set_foreign_p2m_mapping()
124 if (rc || unmap.status != GNTST_okay) in set_foreign_p2m_mapping()
125 pr_err_once("gnttab unmap failed: rc=%d st=%d\n", in set_foreign_p2m_mapping()
126 rc, unmap.status); in set_foreign_p2m_mapping()
/kernel/linux/linux-5.10/arch/arm/xen/
Dp2m.c96 struct gnttab_unmap_grant_ref unmap; in set_foreign_p2m_mapping() local
110 unmap.host_addr = map_ops[i].host_addr, in set_foreign_p2m_mapping()
111 unmap.handle = map_ops[i].handle; in set_foreign_p2m_mapping()
114 unmap.dev_bus_addr = map_ops[i].dev_bus_addr; in set_foreign_p2m_mapping()
116 unmap.dev_bus_addr = 0; in set_foreign_p2m_mapping()
122 unmap.status = 1; in set_foreign_p2m_mapping()
125 &unmap, 1); in set_foreign_p2m_mapping()
126 if (rc || unmap.status != GNTST_okay) in set_foreign_p2m_mapping()
127 pr_err_once("gnttab unmap failed: rc=%d st=%d\n", in set_foreign_p2m_mapping()
128 rc, unmap.status); in set_foreign_p2m_mapping()
/kernel/linux/linux-4.19/drivers/vfio/
Dvfio_iommu_type1.c761 * Sync if the number of fast-unmap regions hits the limit in unmap_unpin_fast()
834 * largest contiguous physical memory chunk to unmap. in vfio_unmap_unpin()
844 * First, try to use fast unmap/unpin. In case of failure, in vfio_unmap_unpin()
845 * switch to slow unmap/unpin path. in vfio_unmap_unpin()
892 * That way the user will be able to map/unmap buffers whose size/ in vfio_pgsize_bitmap()
906 struct vfio_iommu_type1_dma_unmap *unmap) in vfio_dma_do_unmap() argument
915 if (unmap->iova & mask) in vfio_dma_do_unmap()
917 if (!unmap->size || unmap->size & mask) in vfio_dma_do_unmap()
919 if (unmap->iova + unmap->size - 1 < unmap->iova || in vfio_dma_do_unmap()
920 unmap->size > SIZE_MAX) in vfio_dma_do_unmap()
[all …]
/kernel/linux/linux-5.10/drivers/net/ethernet/brocade/bna/
Dbnad.c93 struct bnad_tx_unmap *unmap; in bnad_tx_buff_unmap() local
97 unmap = &unmap_q[index]; in bnad_tx_buff_unmap()
98 nvecs = unmap->nvecs; in bnad_tx_buff_unmap()
100 skb = unmap->skb; in bnad_tx_buff_unmap()
101 unmap->skb = NULL; in bnad_tx_buff_unmap()
102 unmap->nvecs = 0; in bnad_tx_buff_unmap()
104 dma_unmap_addr(&unmap->vectors[0], dma_addr), in bnad_tx_buff_unmap()
106 dma_unmap_addr_set(&unmap->vectors[0], dma_addr, 0); in bnad_tx_buff_unmap()
115 unmap = &unmap_q[index]; in bnad_tx_buff_unmap()
119 dma_unmap_addr(&unmap->vectors[vector], dma_addr), in bnad_tx_buff_unmap()
[all …]
/kernel/linux/linux-4.19/drivers/net/ethernet/brocade/bna/
Dbnad.c101 struct bnad_tx_unmap *unmap; in bnad_tx_buff_unmap() local
105 unmap = &unmap_q[index]; in bnad_tx_buff_unmap()
106 nvecs = unmap->nvecs; in bnad_tx_buff_unmap()
108 skb = unmap->skb; in bnad_tx_buff_unmap()
109 unmap->skb = NULL; in bnad_tx_buff_unmap()
110 unmap->nvecs = 0; in bnad_tx_buff_unmap()
112 dma_unmap_addr(&unmap->vectors[0], dma_addr), in bnad_tx_buff_unmap()
114 dma_unmap_addr_set(&unmap->vectors[0], dma_addr, 0); in bnad_tx_buff_unmap()
123 unmap = &unmap_q[index]; in bnad_tx_buff_unmap()
127 dma_unmap_addr(&unmap->vectors[vector], dma_addr), in bnad_tx_buff_unmap()
[all …]
/kernel/linux/linux-5.10/drivers/xen/
Dxen-front-pgdir-shbuf.c68 /* Unmap grant references of the buffer. */
69 int (*unmap)(struct xen_front_pgdir_shbuf *buf); member
114 * Unmap granted references of the shared buffer.
118 * shared by the frontend itself) or unmap the provided granted
126 if (buf->ops && buf->ops->unmap) in xen_front_pgdir_shbuf_unmap()
127 return buf->ops->unmap(buf); in xen_front_pgdir_shbuf_unmap()
129 /* No need to unmap own grant references. */ in xen_front_pgdir_shbuf_unmap()
203 * Unmap the buffer previously mapped with grant references
236 "Failed to unmap page %d: %d\n", in backend_unmap()
242 "Failed to unmap grant references, ret %d", ret); in backend_unmap()
[all …]
/kernel/linux/linux-4.19/include/linux/
Dzpool.h28 * changed memory back out on unmap. Write-only does not copy
34 ZPOOL_MM_RO, /* read-only (no copy-out at unmap time) */
75 * @unmap: unmap a handle.
102 void (*unmap)(void *pool, unsigned long handle); member
/kernel/linux/linux-4.19/drivers/dma/
Ddmaengine.c1179 #define __UNMAP_POOL(x) { .size = x, .name = "dmaengine-unmap-" __stringify(x) }
1212 struct dmaengine_unmap_data *unmap = container_of(kref, typeof(*unmap), kref); in dmaengine_unmap() local
1213 struct device *dev = unmap->dev; in dmaengine_unmap()
1216 cnt = unmap->to_cnt; in dmaengine_unmap()
1218 dma_unmap_page(dev, unmap->addr[i], unmap->len, in dmaengine_unmap()
1220 cnt += unmap->from_cnt; in dmaengine_unmap()
1222 dma_unmap_page(dev, unmap->addr[i], unmap->len, in dmaengine_unmap()
1224 cnt += unmap->bidi_cnt; in dmaengine_unmap()
1226 if (unmap->addr[i] == 0) in dmaengine_unmap()
1228 dma_unmap_page(dev, unmap->addr[i], unmap->len, in dmaengine_unmap()
[all …]
/kernel/linux/linux-5.10/include/linux/
Dzpool.h28 * changed memory back out on unmap. Write-only does not copy
34 ZPOOL_MM_RO, /* read-only (no copy-out at unmap time) */
78 * @unmap: unmap a handle.
107 void (*unmap)(void *pool, unsigned long handle); member
/kernel/linux/linux-5.10/drivers/misc/habanalabs/common/
Dmmu.c96 * @page_size: size of the page to unmap
101 * - Unmap the virt addr and frees pgts if possible
137 * size is bigger, we break it to sub-pages and unmap them separately. in hl_mmu_unmap()
143 "page size of %u is not %uKB aligned, can't unmap\n", in hl_mmu_unmap()
153 rc = hdev->mmu_func.unmap(ctx, real_virt_addr, is_dram_addr); in hl_mmu_unmap()
219 "page size of %u is not %uKB aligned, can't unmap\n", in hl_mmu_map()
252 if (hdev->mmu_func.unmap(ctx, real_virt_addr, is_dram_addr)) in hl_mmu_map()
254 "failed to unmap va: 0x%llx\n", real_virt_addr); in hl_mmu_map()
/kernel/linux/linux-4.19/drivers/gpu/drm/xen/
Dxen_drm_front_shbuf.c43 /* Unmap grant references of the buffer. */
44 int (*unmap)(struct xen_drm_front_shbuf *buf); member
66 if (buf->ops->unmap) in xen_drm_front_shbuf_unmap()
67 return buf->ops->unmap(buf); in xen_drm_front_shbuf_unmap()
69 /* no need to unmap own grant references */ in xen_drm_front_shbuf_unmap()
155 DRM_ERROR("Failed to unmap page %d: %d\n", in backend_unmap()
160 DRM_ERROR("Failed to unmap grant references, ret %d", ret); in backend_unmap()
218 /* save handles even if error, so we can unmap */ in backend_map()
368 .unmap = backend_unmap
371 /* For locally granted references we do not need to map/unmap the references */
/kernel/linux/linux-5.10/drivers/vfio/
Dvfio_iommu_type1.c855 * Sync if the number of fast-unmap regions hits the limit in unmap_unpin_fast()
930 * largest contiguous physical memory chunk to unmap. in vfio_unmap_unpin()
940 * First, try to use fast unmap/unpin. In case of failure, in vfio_unmap_unpin()
941 * switch to slow unmap/unpin path. in vfio_unmap_unpin()
992 * That way the user will be able to map/unmap buffers whose size/ in vfio_update_pgsize_bitmap()
1096 struct vfio_iommu_type1_dma_unmap *unmap, in vfio_dma_do_unmap() argument
1109 if (unmap->iova & (pgsize - 1)) { in vfio_dma_do_unmap()
1114 if (!unmap->size || unmap->size & (pgsize - 1)) { in vfio_dma_do_unmap()
1119 if (unmap->iova + unmap->size - 1 < unmap->iova || in vfio_dma_do_unmap()
1120 unmap->size > SIZE_MAX) { in vfio_dma_do_unmap()
[all …]

12345678910>>...62