/drivers/acpi/ |
D | nvs.c | 78 void *kaddr; member 138 if (entry->kaddr) { in suspend_nvs_free() 140 iounmap(entry->kaddr); in suspend_nvs_free() 143 acpi_os_unmap_iomem(entry->kaddr, in suspend_nvs_free() 146 entry->kaddr = NULL; in suspend_nvs_free() 182 entry->kaddr = acpi_os_get_iomem(phys, size); in suspend_nvs_save() 183 if (!entry->kaddr) { in suspend_nvs_save() 184 entry->kaddr = acpi_os_ioremap(phys, size); in suspend_nvs_save() 185 entry->unmap = !!entry->kaddr; in suspend_nvs_save() 187 if (!entry->kaddr) { in suspend_nvs_save() [all …]
|
/drivers/misc/ |
D | open-dice.c | 45 void *kaddr; in open_dice_wipe() local 48 kaddr = devm_memremap(drvdata->misc.this_device, drvdata->rmem->base, in open_dice_wipe() 50 if (IS_ERR(kaddr)) { in open_dice_wipe() 52 return PTR_ERR(kaddr); in open_dice_wipe() 55 memset(kaddr, 0, drvdata->rmem->size); in open_dice_wipe() 56 devm_memunmap(drvdata->misc.this_device, kaddr); in open_dice_wipe()
|
/drivers/tee/amdtee/ |
D | shm_pool.c | 26 shm->kaddr = (void *)va; in pool_op_alloc() 34 shm->kaddr = NULL; in pool_op_alloc() 45 free_pages((unsigned long)shm->kaddr, get_order(shm->size)); in pool_op_free() 46 shm->kaddr = NULL; in pool_op_free()
|
D | core.c | 161 if (shmdata->kaddr == shm->kaddr) { in get_buffer_id() 357 shmem.kaddr = shm->kaddr; in amdtee_map_shmem() 371 shmnode->kaddr = shm->kaddr; in amdtee_map_shmem() 378 pr_debug("buf_id :[%x] kaddr[%p]\n", shmnode->buf_id, shmnode->kaddr); in amdtee_map_shmem()
|
D | amdtee_private.h | 82 void *kaddr; member 93 void *kaddr; member
|
D | call.c | 302 if (!start[i].kaddr || (start[i].size & (PAGE_SIZE - 1))) { in handle_map_shmem() 307 if ((u64)start[i].kaddr & (PAGE_SIZE - 1)) { in handle_map_shmem() 309 (u64)start[i].kaddr); in handle_map_shmem() 319 paddr = __psp_pa(start[i].kaddr); in handle_map_shmem()
|
/drivers/gpu/drm/amd/pm/powerplay/smumgr/ |
D | smu8_smumgr.c | 333 struct TOC *toc = (struct TOC *)smu8_smu->toc_buffer.kaddr; in smu8_smu_populate_single_scratch_task() 355 (struct smu8_ih_meta_data *)smu8_smu->scratch_buffer[i].kaddr; in smu8_smu_populate_single_scratch_task() 370 struct TOC *toc = (struct TOC *)smu8_smu->toc_buffer.kaddr; in smu8_smu_populate_single_ucode_load_task() 409 struct TOC *toc = (struct TOC *)smu8_smu->toc_buffer.kaddr; in smu8_smu_initialize_toc_empty_job_list() 420 struct TOC *toc = (struct TOC *)smu8_smu->toc_buffer.kaddr; in smu8_smu_construct_toc_for_vddgfx_enter() 438 struct TOC *toc = (struct TOC *)smu8_smu->toc_buffer.kaddr; in smu8_smu_construct_toc_for_vddgfx_exit() 590 entry->kaddr = (char *) smu8_smu->smu_buffer.kaddr + in smu8_smu_populate_single_scratch_entry() 611 *table = (struct SMU8_Fusion_ClkTable *)smu8_smu->scratch_buffer[i].kaddr; in smu8_download_pptable_settings() 781 &smu8_smu->toc_buffer.kaddr); in smu8_smu_init() 791 &smu8_smu->smu_buffer.kaddr); in smu8_smu_init() [all …]
|
D | smu7_smumgr.c | 411 memcpy_toio(smu_data->header_buffer.kaddr, smu_data->toc, in smu7_request_smu_load_fw() 553 &smu_data->header_buffer.kaddr); in smu7_init() 568 &smu_data->smu_buffer.kaddr); in smu7_init() 573 &smu_data->header_buffer.kaddr); in smu7_init() 591 &smu_data->header_buffer.kaddr); in smu7_smu_fini() 596 &smu_data->smu_buffer.kaddr); in smu7_smu_fini()
|
D | smu8_smumgr.h | 64 void *kaddr; member
|
D | smu7_smumgr.h | 35 void *kaddr; member
|
/drivers/misc/lkdtm/ |
D | usercopy.c | 351 static void do_usercopy_page_span(const char *name, void *kaddr) in do_usercopy_page_span() argument 363 memset(kaddr, 0xAA, PAGE_SIZE); in do_usercopy_page_span() 366 kaddr += PAGE_SIZE / 2; in do_usercopy_page_span() 369 name, kaddr); in do_usercopy_page_span() 370 if (copy_to_user((void __user *)uaddr, kaddr, in do_usercopy_page_span() 377 name, kaddr); in do_usercopy_page_span() 378 if (copy_to_user((void __user *)uaddr, kaddr, unconst + PAGE_SIZE)) { in do_usercopy_page_span()
|
/drivers/tee/optee/ |
D | core.c | 41 shm->kaddr = page_address(page); in optee_pool_op_alloc_helper() 59 (unsigned long)shm->kaddr); in optee_pool_op_alloc_helper() 68 free_pages((unsigned long)shm->kaddr, order); in optee_pool_op_alloc_helper() 78 free_pages((unsigned long)shm->kaddr, get_order(shm->size)); in optee_pool_op_free_helper() 79 shm->kaddr = NULL; in optee_pool_op_free_helper()
|
/drivers/tee/ |
D | tee_shm_pool.c | 26 shm->kaddr = (void *)va; in pool_op_gen_alloc() 39 gen_pool_free(pool->private_data, (unsigned long)shm->kaddr, in pool_op_gen_free() 41 shm->kaddr = NULL; in pool_op_gen_free()
|
/drivers/vfio/ |
D | iova_bitmap.c | 417 void *kaddr; in iova_bitmap_set() local 422 kaddr = kmap_local_page(mapped->pages[page_idx]); in iova_bitmap_set() 423 bitmap_set(kaddr, offset, nbits); in iova_bitmap_set() 424 kunmap_local(kaddr); in iova_bitmap_set()
|
/drivers/pci/ |
D | p2pdma.c | 100 void *kaddr; in p2pmem_alloc_mmap() local 125 kaddr = (void *)gen_pool_alloc_owner(p2pdma->pool, len, (void **)&ref); in p2pmem_alloc_mmap() 126 if (!kaddr) { in p2pmem_alloc_mmap() 143 ret = vm_insert_page(vma, vaddr, virt_to_page(kaddr)); in p2pmem_alloc_mmap() 145 gen_pool_free(p2pdma->pool, (uintptr_t)kaddr, len); in p2pmem_alloc_mmap() 149 put_page(virt_to_page(kaddr)); in p2pmem_alloc_mmap() 150 kaddr += PAGE_SIZE; in p2pmem_alloc_mmap() 158 gen_pool_free(p2pdma->pool, (uintptr_t)kaddr, len); in p2pmem_alloc_mmap()
|
/drivers/s390/block/ |
D | dcssblk.c | 35 long nr_pages, enum dax_access_mode mode, void **kaddr, 52 void *kaddr; in dcssblk_dax_zero_page_range() local 55 &kaddr, NULL); in dcssblk_dax_zero_page_range() 59 memset(kaddr, 0, nr_pages << PAGE_SHIFT); in dcssblk_dax_zero_page_range() 60 dax_flush(dax_dev, kaddr, nr_pages << PAGE_SHIFT); in dcssblk_dax_zero_page_range() 913 long nr_pages, void **kaddr, pfn_t *pfn) in __dcssblk_direct_access() argument 919 if (kaddr) in __dcssblk_direct_access() 920 *kaddr = (void *) dev_info->start + offset; in __dcssblk_direct_access() 930 long nr_pages, enum dax_access_mode mode, void **kaddr, in dcssblk_dax_direct_access() argument 935 return __dcssblk_direct_access(dev_info, pgoff, nr_pages, kaddr, pfn); in dcssblk_dax_direct_access()
|
/drivers/fpga/ |
D | stratix10-soc.c | 120 static void s10_unlock_bufs(struct s10_priv *priv, void *kaddr) in s10_unlock_bufs() argument 124 if (!kaddr) in s10_unlock_bufs() 128 if (priv->svc_bufs[i].buf == kaddr) { in s10_unlock_bufs() 134 WARN(1, "Unknown buffer returned from service layer %p\n", kaddr); in s10_unlock_bufs()
|
/drivers/scsi/isci/ |
D | request.c | 1412 void *kaddr; in sci_stp_request_pio_data_in_copy_data_buffer() local 1426 kaddr = kmap_atomic(page); in sci_stp_request_pio_data_in_copy_data_buffer() 1427 memcpy(kaddr + sg->offset, src_addr, copy_len); in sci_stp_request_pio_data_in_copy_data_buffer() 1428 kunmap_atomic(kaddr); in sci_stp_request_pio_data_in_copy_data_buffer() 1768 void *frame_header, *kaddr; in sci_io_request_frame_handler() local 1774 kaddr = kmap_atomic(sg_page(sg)); in sci_io_request_frame_handler() 1775 rsp = kaddr + sg->offset; in sci_io_request_frame_handler() 1811 kunmap_atomic(kaddr); in sci_io_request_frame_handler() 2915 void *kaddr; in isci_request_io_request_complete() local 2920 kaddr = kmap_atomic(sg_page(sg)); in isci_request_io_request_complete() [all …]
|
/drivers/nvdimm/ |
D | pmem.c | 244 long nr_pages, enum dax_access_mode mode, void **kaddr, in __pmem_direct_access() argument 254 if (kaddr) in __pmem_direct_access() 255 *kaddr = pmem->virt_addr + offset; in __pmem_direct_access() 306 void **kaddr, pfn_t *pfn) in pmem_dax_direct_access() argument 310 return __pmem_direct_access(pmem, pgoff, nr_pages, mode, kaddr, pfn); in pmem_dax_direct_access()
|
D | pmem.h | 33 long nr_pages, enum dax_access_mode mode, void **kaddr,
|
/drivers/md/ |
D | dm-linear.c | 170 long nr_pages, enum dax_access_mode mode, void **kaddr, in linear_dax_direct_access() argument 175 return dax_direct_access(dax_dev, pgoff, nr_pages, mode, kaddr, pfn); in linear_dax_direct_access()
|
D | dm-target.c | 160 long nr_pages, enum dax_access_mode mode, void **kaddr, in io_err_dax_direct_access() argument
|
/drivers/net/ethernet/chelsio/inline_crypto/chtls/ |
D | chtls_hw.c | 367 int kaddr; in chtls_setkey() local 395 kaddr = keyid_to_addr(cdev->kmap.start, keyid); in chtls_setkey() 414 kwr->req.lock_addr = cpu_to_be32(ULP_MEMIO_ADDR_V(kaddr)); in chtls_setkey()
|
/drivers/infiniband/sw/siw/ |
D | siw_qp_tx.c | 500 void *kaddr; in siw_tx_hdt() local 521 void *kaddr = kmap_local_page(p); in siw_tx_hdt() local 525 iov[seg].iov_base = kaddr + fp_off; in siw_tx_hdt() 534 kaddr = kmap_local_page(p); in siw_tx_hdt() 536 kaddr + fp_off, in siw_tx_hdt() 538 kunmap_local(kaddr); in siw_tx_hdt()
|
/drivers/vhost/ |
D | vringh.c | 1294 void *kaddr = kmap_local_page(ivec.iov.bvec[0].bv_page); in getu16_iotlb() local 1295 void *from = kaddr + ivec.iov.bvec[0].bv_offset; in getu16_iotlb() 1298 kunmap_local(kaddr); in getu16_iotlb() 1333 void *kaddr = kmap_local_page(ivec.iov.bvec[0].bv_page); in putu16_iotlb() local 1334 void *to = kaddr + ivec.iov.bvec[0].bv_offset; in putu16_iotlb() 1337 kunmap_local(kaddr); in putu16_iotlb()
|