| /kernel/linux/linux-5.10/crypto/async_tx/ |
| D | async_xor.c | 23 do_async_xor(struct dma_chan *chan, struct dmaengine_unmap_data *unmap, in do_async_xor() argument 32 int src_cnt = unmap->to_cnt; in do_async_xor() 34 dma_addr_t dma_dest = unmap->addr[unmap->to_cnt]; in do_async_xor() 35 dma_addr_t *src_list = unmap->addr; in do_async_xor() 63 if (src_list > unmap->addr) in do_async_xor() 66 xor_src_cnt, unmap->len, in do_async_xor() 77 xor_src_cnt, unmap->len, in do_async_xor() 82 dma_set_unmap(tx, unmap); in do_async_xor() 190 struct dmaengine_unmap_data *unmap = NULL; in async_xor_offs() local 195 unmap = dmaengine_get_unmap_data(device->dev, src_cnt+1, GFP_NOWAIT); in async_xor_offs() [all …]
|
| D | async_pq.c | 37 struct dmaengine_unmap_data *unmap, in do_async_gen_syndrome() argument 76 dma_dest[0] = unmap->addr[disks - 2]; in do_async_gen_syndrome() 77 dma_dest[1] = unmap->addr[disks - 1]; in do_async_gen_syndrome() 79 &unmap->addr[src_off], in do_async_gen_syndrome() 81 &scfs[src_off], unmap->len, in do_async_gen_syndrome() 89 dma_set_unmap(tx, unmap); in do_async_gen_syndrome() 185 struct dmaengine_unmap_data *unmap = NULL; in async_gen_syndrome() local 190 unmap = dmaengine_get_unmap_data(device->dev, disks, GFP_NOWAIT); in async_gen_syndrome() 193 if (unmap && !(submit->flags & ASYNC_TX_PQ_XOR_DST) && in async_gen_syndrome() 209 unmap->len = len; in async_gen_syndrome() [all …]
|
| D | async_memcpy.c | 40 struct dmaengine_unmap_data *unmap = NULL; in async_memcpy() local 43 unmap = dmaengine_get_unmap_data(device->dev, 2, GFP_NOWAIT); in async_memcpy() 45 if (unmap && is_dma_copy_aligned(device, src_offset, dest_offset, len)) { in async_memcpy() 53 unmap->to_cnt = 1; in async_memcpy() 54 unmap->addr[0] = dma_map_page(device->dev, src, src_offset, len, in async_memcpy() 56 unmap->from_cnt = 1; in async_memcpy() 57 unmap->addr[1] = dma_map_page(device->dev, dest, dest_offset, len, in async_memcpy() 59 unmap->len = len; in async_memcpy() 61 tx = device->device_prep_dma_memcpy(chan, unmap->addr[1], in async_memcpy() 62 unmap->addr[0], len, in async_memcpy() [all …]
|
| D | async_raid6_recov.c | 25 struct dmaengine_unmap_data *unmap = NULL; in async_sum_product() local 31 unmap = dmaengine_get_unmap_data(dma->dev, 3, GFP_NOWAIT); in async_sum_product() 33 if (unmap) { in async_sum_product() 41 unmap->addr[0] = dma_map_page(dev, srcs[0], src_offs[0], in async_sum_product() 43 unmap->addr[1] = dma_map_page(dev, srcs[1], src_offs[1], in async_sum_product() 45 unmap->to_cnt = 2; in async_sum_product() 47 unmap->addr[2] = dma_map_page(dev, dest, d_off, in async_sum_product() 49 unmap->bidi_cnt = 1; in async_sum_product() 51 pq[1] = unmap->addr[2]; in async_sum_product() 53 unmap->len = len; in async_sum_product() [all …]
|
| /kernel/linux/linux-5.10/tools/testing/selftests/vm/ |
| D | mlock2-tests.c | 210 goto unmap; in test_mlock_lock() 214 goto unmap; in test_mlock_lock() 219 goto unmap; in test_mlock_lock() 224 unmap: in test_mlock_lock() 273 goto unmap; in test_mlock_onfault() 277 goto unmap; in test_mlock_onfault() 286 goto unmap; in test_mlock_onfault() 290 unmap: in test_mlock_onfault() 317 goto unmap; in test_lock_onfault_of_present() 323 goto unmap; in test_lock_onfault_of_present() [all …]
|
| D | mremap_dontunmap.c | 62 "unable to unmap destination mapping"); in kernel_support_for_mremap_dontunmap() 66 "unable to unmap source mapping"); in kernel_support_for_mremap_dontunmap() 125 "unable to unmap destination mapping"); in mremap_dontunmap_simple() 127 "unable to unmap source mapping"); in mremap_dontunmap_simple() 159 // The dest mapping will have been unmap by mremap so we expect the Xs in mremap_dontunmap_simple_fixed() 171 "unable to unmap destination mapping"); in mremap_dontunmap_simple_fixed() 173 "unable to unmap source mapping"); in mremap_dontunmap_simple_fixed() 221 "unable to unmap destination mapping"); in mremap_dontunmap_partial_mapping() 223 "unable to unmap source mapping"); in mremap_dontunmap_partial_mapping() 280 "unable to unmap destination mapping"); in mremap_dontunmap_partial_mapping_overwrite() [all …]
|
| /kernel/linux/linux-5.10/drivers/xen/xenbus/ |
| D | xenbus_client.c | 80 struct gnttab_unmap_grant_ref unmap[XENBUS_MAX_RING_GRANTS]; member 92 int (*unmap)(struct xenbus_device *dev, void *vaddr); member 538 gnttab_set_unmap_op(&info->unmap[j], in __xenbus_map_ring() 545 if (HYPERVISOR_grant_table_op(GNTTABOP_unmap_grant_ref, info->unmap, j)) in __xenbus_map_ring() 550 if (info->unmap[i].status != GNTST_okay) { in __xenbus_map_ring() 564 * @vaddrs: addresses to unmap 566 * Unmap memory in this domain that was imported from another domain. 573 struct gnttab_unmap_grant_ref unmap[XENBUS_MAX_RING_GRANTS]; in xenbus_unmap_ring() local 581 gnttab_set_unmap_op(&unmap[i], vaddrs[i], in xenbus_unmap_ring() 584 if (HYPERVISOR_grant_table_op(GNTTABOP_unmap_grant_ref, unmap, i)) in xenbus_unmap_ring() [all …]
|
| /kernel/linux/linux-5.10/arch/arm/xen/ |
| D | p2m.c | 96 struct gnttab_unmap_grant_ref unmap; in set_foreign_p2m_mapping() local 110 unmap.host_addr = map_ops[i].host_addr, in set_foreign_p2m_mapping() 111 unmap.handle = map_ops[i].handle; in set_foreign_p2m_mapping() 114 unmap.dev_bus_addr = map_ops[i].dev_bus_addr; in set_foreign_p2m_mapping() 116 unmap.dev_bus_addr = 0; in set_foreign_p2m_mapping() 122 unmap.status = 1; in set_foreign_p2m_mapping() 125 &unmap, 1); in set_foreign_p2m_mapping() 126 if (rc || unmap.status != GNTST_okay) in set_foreign_p2m_mapping() 127 pr_err_once("gnttab unmap failed: rc=%d st=%d\n", in set_foreign_p2m_mapping() 128 rc, unmap.status); in set_foreign_p2m_mapping()
|
| /kernel/linux/linux-5.10/drivers/net/ethernet/brocade/bna/ |
| D | bnad.c | 93 struct bnad_tx_unmap *unmap; in bnad_tx_buff_unmap() local 97 unmap = &unmap_q[index]; in bnad_tx_buff_unmap() 98 nvecs = unmap->nvecs; in bnad_tx_buff_unmap() 100 skb = unmap->skb; in bnad_tx_buff_unmap() 101 unmap->skb = NULL; in bnad_tx_buff_unmap() 102 unmap->nvecs = 0; in bnad_tx_buff_unmap() 104 dma_unmap_addr(&unmap->vectors[0], dma_addr), in bnad_tx_buff_unmap() 106 dma_unmap_addr_set(&unmap->vectors[0], dma_addr, 0); in bnad_tx_buff_unmap() 115 unmap = &unmap_q[index]; in bnad_tx_buff_unmap() 119 dma_unmap_addr(&unmap->vectors[vector], dma_addr), in bnad_tx_buff_unmap() [all …]
|
| /kernel/linux/linux-5.10/drivers/xen/ |
| D | xen-front-pgdir-shbuf.c | 68 /* Unmap grant references of the buffer. */ 69 int (*unmap)(struct xen_front_pgdir_shbuf *buf); member 114 * Unmap granted references of the shared buffer. 118 * shared by the frontend itself) or unmap the provided granted 126 if (buf->ops && buf->ops->unmap) in xen_front_pgdir_shbuf_unmap() 127 return buf->ops->unmap(buf); in xen_front_pgdir_shbuf_unmap() 129 /* No need to unmap own grant references. */ in xen_front_pgdir_shbuf_unmap() 203 * Unmap the buffer previously mapped with grant references 236 "Failed to unmap page %d: %d\n", in backend_unmap() 242 "Failed to unmap grant references, ret %d", ret); in backend_unmap() [all …]
|
| D | arm-device.c | 70 goto unmap; in xen_map_device_mmio() 94 goto unmap; in xen_map_device_mmio() 99 unmap: in xen_map_device_mmio() 126 (action == BUS_NOTIFY_DEL_DEVICE ? "unmap" : "?"), in xen_platform_notifier() 168 (action == BUS_NOTIFY_DEL_DEVICE ? "unmap" : "?"), in xen_amba_notifier()
|
| /kernel/linux/linux-5.10/include/linux/ |
| D | zpool.h | 28 * changed memory back out on unmap. Write-only does not copy 34 ZPOOL_MM_RO, /* read-only (no copy-out at unmap time) */ 78 * @unmap: unmap a handle. 107 void (*unmap)(void *pool, unsigned long handle); member
|
| /kernel/linux/linux-5.10/drivers/misc/habanalabs/common/ |
| D | mmu.c | 96 * @page_size: size of the page to unmap 101 * - Unmap the virt addr and frees pgts if possible 137 * size is bigger, we break it to sub-pages and unmap them separately. in hl_mmu_unmap() 143 "page size of %u is not %uKB aligned, can't unmap\n", in hl_mmu_unmap() 153 rc = hdev->mmu_func.unmap(ctx, real_virt_addr, is_dram_addr); in hl_mmu_unmap() 219 "page size of %u is not %uKB aligned, can't unmap\n", in hl_mmu_map() 252 if (hdev->mmu_func.unmap(ctx, real_virt_addr, is_dram_addr)) in hl_mmu_map() 254 "failed to unmap va: 0x%llx\n", real_virt_addr); in hl_mmu_map()
|
| /kernel/linux/linux-5.10/drivers/vfio/ |
| D | vfio_iommu_type1.c | 855 * Sync if the number of fast-unmap regions hits the limit in unmap_unpin_fast() 930 * largest contiguous physical memory chunk to unmap. in vfio_unmap_unpin() 940 * First, try to use fast unmap/unpin. In case of failure, in vfio_unmap_unpin() 941 * switch to slow unmap/unpin path. in vfio_unmap_unpin() 992 * That way the user will be able to map/unmap buffers whose size/ in vfio_update_pgsize_bitmap() 1096 struct vfio_iommu_type1_dma_unmap *unmap, in vfio_dma_do_unmap() argument 1109 if (unmap->iova & (pgsize - 1)) { in vfio_dma_do_unmap() 1114 if (!unmap->size || unmap->size & (pgsize - 1)) { in vfio_dma_do_unmap() 1119 if (unmap->iova + unmap->size - 1 < unmap->iova || in vfio_dma_do_unmap() 1120 unmap->size > SIZE_MAX) { in vfio_dma_do_unmap() [all …]
|
| /kernel/linux/linux-5.10/Documentation/vm/ |
| D | transhuge.rst | 125 - map/unmap of the pages with PTE entry increment/decrement ->_mapcount 128 - map/unmap of the whole compound page is accounted for in compound_mapcount 131 last unmap of subpages. 137 get race-free detection of unmap of subpages when we have them mapped with 142 map/unmap of the whole compound page. 177 Partial unmap and deferred_split_huge_page() 186 the place where we can detect partial unmap. It also might be 187 counterproductive since in many cases partial unmap happens during exit(2) if
|
| /kernel/linux/linux-5.10/arch/x86/xen/ |
| D | p2m.c | 712 struct gnttab_unmap_grant_ref unmap[2]; in set_foreign_p2m_mapping() local 739 unmap[0].host_addr = map_ops[i].host_addr, in set_foreign_p2m_mapping() 740 unmap[0].handle = map_ops[i].handle; in set_foreign_p2m_mapping() 743 unmap[0].dev_bus_addr = map_ops[i].dev_bus_addr; in set_foreign_p2m_mapping() 745 unmap[0].dev_bus_addr = 0; in set_foreign_p2m_mapping() 749 unmap[1].host_addr = kmap_ops[i].host_addr, in set_foreign_p2m_mapping() 750 unmap[1].handle = kmap_ops[i].handle; in set_foreign_p2m_mapping() 753 unmap[1].dev_bus_addr = kmap_ops[i].dev_bus_addr; in set_foreign_p2m_mapping() 755 unmap[1].dev_bus_addr = 0; in set_foreign_p2m_mapping() 762 unmap[0].status = 1; in set_foreign_p2m_mapping() [all …]
|
| /kernel/linux/linux-5.10/drivers/dma/ |
| D | dmaengine.c | 1375 #define __UNMAP_POOL(x) { .size = x, .name = "dmaengine-unmap-" __stringify(x) } 1408 struct dmaengine_unmap_data *unmap = container_of(kref, typeof(*unmap), kref); in dmaengine_unmap() local 1409 struct device *dev = unmap->dev; in dmaengine_unmap() 1412 cnt = unmap->to_cnt; in dmaengine_unmap() 1414 dma_unmap_page(dev, unmap->addr[i], unmap->len, in dmaengine_unmap() 1416 cnt += unmap->from_cnt; in dmaengine_unmap() 1418 dma_unmap_page(dev, unmap->addr[i], unmap->len, in dmaengine_unmap() 1420 cnt += unmap->bidi_cnt; in dmaengine_unmap() 1422 if (unmap->addr[i] == 0) in dmaengine_unmap() 1424 dma_unmap_page(dev, unmap->addr[i], unmap->len, in dmaengine_unmap() [all …]
|
| D | mv_xor.c | 769 struct dmaengine_unmap_data *unmap; in mv_chan_memcpy_self_test() local 792 unmap = dmaengine_get_unmap_data(dma_chan->device->dev, 2, GFP_KERNEL); in mv_chan_memcpy_self_test() 793 if (!unmap) { in mv_chan_memcpy_self_test() 801 unmap->addr[0] = src_dma; in mv_chan_memcpy_self_test() 808 unmap->to_cnt = 1; in mv_chan_memcpy_self_test() 813 unmap->addr[1] = dest_dma; in mv_chan_memcpy_self_test() 820 unmap->from_cnt = 1; in mv_chan_memcpy_self_test() 821 unmap->len = PAGE_SIZE; in mv_chan_memcpy_self_test() 862 dmaengine_unmap_put(unmap); in mv_chan_memcpy_self_test() 880 struct dmaengine_unmap_data *unmap; in mv_chan_xor_self_test() local [all …]
|
| /kernel/linux/linux-5.10/drivers/scsi/ |
| D | sd.h | 56 SD_LBP_UNMAP, /* Use UNMAP command */ 57 SD_LBP_WS16, /* Use WRITE SAME(16) with UNMAP bit */ 58 SD_LBP_WS10, /* Use WRITE SAME(10) with UNMAP bit */ 66 SD_ZERO_WS16_UNMAP, /* Use WRITE SAME(16) with UNMAP */ 67 SD_ZERO_WS10_UNMAP, /* Use WRITE SAME(10) with UNMAP */ 165 case UNMAP: in scsi_medium_access_command()
|
| /kernel/linux/linux-5.10/mm/ |
| D | percpu-vm.c | 140 * pcpu_unmap_pages - unmap pages out of a pcpu_chunk 143 * @page_start: page index of the first page to unmap 144 * @page_end: page index of the last page to unmap + 1 146 * For each cpu, unmap pages [@page_start,@page_end) out of @chunk. 297 * pcpu_depopulate_chunk - depopulate and unmap an area of a pcpu_chunk 302 * For each cpu, depopulate and unmap pages [@page_start,@page_end) 321 /* unmap and free */ in pcpu_depopulate_chunk()
|
| /kernel/linux/linux-5.10/lib/ |
| D | devres.c | 103 * @dev: Generic device to unmap for 104 * @addr: Address to unmap 288 * @dev: Generic device to unmap for 289 * @addr: Address to unmap 380 * @addr: Address to unmap 482 * pcim_iounmap_regions - Unmap and release PCI BARs 484 * @mask: Mask of BARs to unmap and release 486 * Unmap and release regions specified by @mask.
|
| /kernel/linux/linux-5.10/drivers/acpi/ |
| D | nvs.c | 78 bool unmap; member 137 if (entry->unmap) { in suspend_nvs_free() 139 entry->unmap = false; in suspend_nvs_free() 183 entry->unmap = !!entry->kaddr; in suspend_nvs_save()
|
| /kernel/linux/linux-5.10/include/uapi/xen/ |
| D | gntalloc.h | 49 /* Number of references to unmap */ 54 * Sets up an unmap notification within the page, so that the other side can do 73 /* Action(s) to take on unmap */
|
| /kernel/linux/linux-5.10/include/xen/ |
| D | grant_table.h | 164 gnttab_set_unmap_op(struct gnttab_unmap_grant_ref *unmap, phys_addr_t addr, in gnttab_set_unmap_op() argument 168 unmap->host_addr = addr; in gnttab_set_unmap_op() 170 unmap->host_addr = __pa(addr); in gnttab_set_unmap_op() 172 unmap->host_addr = addr; in gnttab_set_unmap_op() 174 unmap->handle = handle; in gnttab_set_unmap_op() 175 unmap->dev_bus_addr = 0; in gnttab_set_unmap_op()
|
| /kernel/linux/linux-5.10/Documentation/core-api/ |
| D | dma-api-howto.rst | 386 To unmap and free such a DMA region, you call:: 489 packets, map/unmap them with the DMA_TO_DEVICE direction 490 specifier. For receive packets, just the opposite, map/unmap them 497 context. There are two versions of each map/unmap, one which will 498 map/unmap a single memory region, and one which will map/unmap a 518 and to unmap it:: 534 map/unmap interface pair akin to dma_{map,unmap}_single(). These 590 To unmap a scatterlist, just call:: 642 After the last DMA transfer call one of the DMA unmap routines 747 - unmap pages that are already mapped, when mapping error occurs in the middle [all …]
|