Home
last modified time | relevance | path

Searched refs:kaddr (Results 1 – 25 of 38) sorted by relevance

12

/drivers/acpi/
Dnvs.c78 void *kaddr; member
138 if (entry->kaddr) { in suspend_nvs_free()
140 iounmap(entry->kaddr); in suspend_nvs_free()
143 acpi_os_unmap_iomem(entry->kaddr, in suspend_nvs_free()
146 entry->kaddr = NULL; in suspend_nvs_free()
182 entry->kaddr = acpi_os_get_iomem(phys, size); in suspend_nvs_save()
183 if (!entry->kaddr) { in suspend_nvs_save()
184 entry->kaddr = acpi_os_ioremap(phys, size); in suspend_nvs_save()
185 entry->unmap = !!entry->kaddr; in suspend_nvs_save()
187 if (!entry->kaddr) { in suspend_nvs_save()
[all …]
/drivers/misc/
Dopen-dice.c45 void *kaddr; in open_dice_wipe() local
48 kaddr = devm_memremap(drvdata->misc.this_device, drvdata->rmem->base, in open_dice_wipe()
50 if (IS_ERR(kaddr)) { in open_dice_wipe()
52 return PTR_ERR(kaddr); in open_dice_wipe()
55 memset(kaddr, 0, drvdata->rmem->size); in open_dice_wipe()
56 devm_memunmap(drvdata->misc.this_device, kaddr); in open_dice_wipe()
/drivers/tee/amdtee/
Dshm_pool.c26 shm->kaddr = (void *)va; in pool_op_alloc()
34 shm->kaddr = NULL; in pool_op_alloc()
45 free_pages((unsigned long)shm->kaddr, get_order(shm->size)); in pool_op_free()
46 shm->kaddr = NULL; in pool_op_free()
Dcore.c161 if (shmdata->kaddr == shm->kaddr) { in get_buffer_id()
357 shmem.kaddr = shm->kaddr; in amdtee_map_shmem()
371 shmnode->kaddr = shm->kaddr; in amdtee_map_shmem()
378 pr_debug("buf_id :[%x] kaddr[%p]\n", shmnode->buf_id, shmnode->kaddr); in amdtee_map_shmem()
Damdtee_private.h82 void *kaddr; member
93 void *kaddr; member
Dcall.c302 if (!start[i].kaddr || (start[i].size & (PAGE_SIZE - 1))) { in handle_map_shmem()
307 if ((u64)start[i].kaddr & (PAGE_SIZE - 1)) { in handle_map_shmem()
309 (u64)start[i].kaddr); in handle_map_shmem()
319 paddr = __psp_pa(start[i].kaddr); in handle_map_shmem()
/drivers/gpu/drm/amd/pm/powerplay/smumgr/
Dsmu8_smumgr.c333 struct TOC *toc = (struct TOC *)smu8_smu->toc_buffer.kaddr; in smu8_smu_populate_single_scratch_task()
355 (struct smu8_ih_meta_data *)smu8_smu->scratch_buffer[i].kaddr; in smu8_smu_populate_single_scratch_task()
370 struct TOC *toc = (struct TOC *)smu8_smu->toc_buffer.kaddr; in smu8_smu_populate_single_ucode_load_task()
409 struct TOC *toc = (struct TOC *)smu8_smu->toc_buffer.kaddr; in smu8_smu_initialize_toc_empty_job_list()
420 struct TOC *toc = (struct TOC *)smu8_smu->toc_buffer.kaddr; in smu8_smu_construct_toc_for_vddgfx_enter()
438 struct TOC *toc = (struct TOC *)smu8_smu->toc_buffer.kaddr; in smu8_smu_construct_toc_for_vddgfx_exit()
590 entry->kaddr = (char *) smu8_smu->smu_buffer.kaddr + in smu8_smu_populate_single_scratch_entry()
611 *table = (struct SMU8_Fusion_ClkTable *)smu8_smu->scratch_buffer[i].kaddr; in smu8_download_pptable_settings()
781 &smu8_smu->toc_buffer.kaddr); in smu8_smu_init()
791 &smu8_smu->smu_buffer.kaddr); in smu8_smu_init()
[all …]
Dsmu7_smumgr.c411 memcpy_toio(smu_data->header_buffer.kaddr, smu_data->toc, in smu7_request_smu_load_fw()
553 &smu_data->header_buffer.kaddr); in smu7_init()
568 &smu_data->smu_buffer.kaddr); in smu7_init()
573 &smu_data->header_buffer.kaddr); in smu7_init()
591 &smu_data->header_buffer.kaddr); in smu7_smu_fini()
596 &smu_data->smu_buffer.kaddr); in smu7_smu_fini()
Dsmu8_smumgr.h64 void *kaddr; member
Dsmu7_smumgr.h35 void *kaddr; member
/drivers/misc/lkdtm/
Dusercopy.c351 static void do_usercopy_page_span(const char *name, void *kaddr) in do_usercopy_page_span() argument
363 memset(kaddr, 0xAA, PAGE_SIZE); in do_usercopy_page_span()
366 kaddr += PAGE_SIZE / 2; in do_usercopy_page_span()
369 name, kaddr); in do_usercopy_page_span()
370 if (copy_to_user((void __user *)uaddr, kaddr, in do_usercopy_page_span()
377 name, kaddr); in do_usercopy_page_span()
378 if (copy_to_user((void __user *)uaddr, kaddr, unconst + PAGE_SIZE)) { in do_usercopy_page_span()
/drivers/tee/optee/
Dcore.c41 shm->kaddr = page_address(page); in optee_pool_op_alloc_helper()
59 (unsigned long)shm->kaddr); in optee_pool_op_alloc_helper()
68 free_pages((unsigned long)shm->kaddr, order); in optee_pool_op_alloc_helper()
78 free_pages((unsigned long)shm->kaddr, get_order(shm->size)); in optee_pool_op_free_helper()
79 shm->kaddr = NULL; in optee_pool_op_free_helper()
/drivers/tee/
Dtee_shm_pool.c26 shm->kaddr = (void *)va; in pool_op_gen_alloc()
39 gen_pool_free(pool->private_data, (unsigned long)shm->kaddr, in pool_op_gen_free()
41 shm->kaddr = NULL; in pool_op_gen_free()
/drivers/vfio/
Diova_bitmap.c417 void *kaddr; in iova_bitmap_set() local
422 kaddr = kmap_local_page(mapped->pages[page_idx]); in iova_bitmap_set()
423 bitmap_set(kaddr, offset, nbits); in iova_bitmap_set()
424 kunmap_local(kaddr); in iova_bitmap_set()
/drivers/pci/
Dp2pdma.c100 void *kaddr; in p2pmem_alloc_mmap() local
125 kaddr = (void *)gen_pool_alloc_owner(p2pdma->pool, len, (void **)&ref); in p2pmem_alloc_mmap()
126 if (!kaddr) { in p2pmem_alloc_mmap()
143 ret = vm_insert_page(vma, vaddr, virt_to_page(kaddr)); in p2pmem_alloc_mmap()
145 gen_pool_free(p2pdma->pool, (uintptr_t)kaddr, len); in p2pmem_alloc_mmap()
149 put_page(virt_to_page(kaddr)); in p2pmem_alloc_mmap()
150 kaddr += PAGE_SIZE; in p2pmem_alloc_mmap()
158 gen_pool_free(p2pdma->pool, (uintptr_t)kaddr, len); in p2pmem_alloc_mmap()
/drivers/s390/block/
Ddcssblk.c35 long nr_pages, enum dax_access_mode mode, void **kaddr,
52 void *kaddr; in dcssblk_dax_zero_page_range() local
55 &kaddr, NULL); in dcssblk_dax_zero_page_range()
59 memset(kaddr, 0, nr_pages << PAGE_SHIFT); in dcssblk_dax_zero_page_range()
60 dax_flush(dax_dev, kaddr, nr_pages << PAGE_SHIFT); in dcssblk_dax_zero_page_range()
913 long nr_pages, void **kaddr, pfn_t *pfn) in __dcssblk_direct_access() argument
919 if (kaddr) in __dcssblk_direct_access()
920 *kaddr = (void *) dev_info->start + offset; in __dcssblk_direct_access()
930 long nr_pages, enum dax_access_mode mode, void **kaddr, in dcssblk_dax_direct_access() argument
935 return __dcssblk_direct_access(dev_info, pgoff, nr_pages, kaddr, pfn); in dcssblk_dax_direct_access()
/drivers/fpga/
Dstratix10-soc.c120 static void s10_unlock_bufs(struct s10_priv *priv, void *kaddr) in s10_unlock_bufs() argument
124 if (!kaddr) in s10_unlock_bufs()
128 if (priv->svc_bufs[i].buf == kaddr) { in s10_unlock_bufs()
134 WARN(1, "Unknown buffer returned from service layer %p\n", kaddr); in s10_unlock_bufs()
/drivers/scsi/isci/
Drequest.c1412 void *kaddr; in sci_stp_request_pio_data_in_copy_data_buffer() local
1426 kaddr = kmap_atomic(page); in sci_stp_request_pio_data_in_copy_data_buffer()
1427 memcpy(kaddr + sg->offset, src_addr, copy_len); in sci_stp_request_pio_data_in_copy_data_buffer()
1428 kunmap_atomic(kaddr); in sci_stp_request_pio_data_in_copy_data_buffer()
1768 void *frame_header, *kaddr; in sci_io_request_frame_handler() local
1774 kaddr = kmap_atomic(sg_page(sg)); in sci_io_request_frame_handler()
1775 rsp = kaddr + sg->offset; in sci_io_request_frame_handler()
1811 kunmap_atomic(kaddr); in sci_io_request_frame_handler()
2915 void *kaddr; in isci_request_io_request_complete() local
2920 kaddr = kmap_atomic(sg_page(sg)); in isci_request_io_request_complete()
[all …]
/drivers/nvdimm/
Dpmem.c244 long nr_pages, enum dax_access_mode mode, void **kaddr, in __pmem_direct_access() argument
254 if (kaddr) in __pmem_direct_access()
255 *kaddr = pmem->virt_addr + offset; in __pmem_direct_access()
306 void **kaddr, pfn_t *pfn) in pmem_dax_direct_access() argument
310 return __pmem_direct_access(pmem, pgoff, nr_pages, mode, kaddr, pfn); in pmem_dax_direct_access()
Dpmem.h33 long nr_pages, enum dax_access_mode mode, void **kaddr,
/drivers/md/
Ddm-linear.c170 long nr_pages, enum dax_access_mode mode, void **kaddr, in linear_dax_direct_access() argument
175 return dax_direct_access(dax_dev, pgoff, nr_pages, mode, kaddr, pfn); in linear_dax_direct_access()
Ddm-target.c160 long nr_pages, enum dax_access_mode mode, void **kaddr, in io_err_dax_direct_access() argument
/drivers/net/ethernet/chelsio/inline_crypto/chtls/
Dchtls_hw.c367 int kaddr; in chtls_setkey() local
395 kaddr = keyid_to_addr(cdev->kmap.start, keyid); in chtls_setkey()
414 kwr->req.lock_addr = cpu_to_be32(ULP_MEMIO_ADDR_V(kaddr)); in chtls_setkey()
/drivers/infiniband/sw/siw/
Dsiw_qp_tx.c500 void *kaddr; in siw_tx_hdt() local
521 void *kaddr = kmap_local_page(p); in siw_tx_hdt() local
525 iov[seg].iov_base = kaddr + fp_off; in siw_tx_hdt()
534 kaddr = kmap_local_page(p); in siw_tx_hdt()
536 kaddr + fp_off, in siw_tx_hdt()
538 kunmap_local(kaddr); in siw_tx_hdt()
/drivers/vhost/
Dvringh.c1294 void *kaddr = kmap_local_page(ivec.iov.bvec[0].bv_page); in getu16_iotlb() local
1295 void *from = kaddr + ivec.iov.bvec[0].bv_offset; in getu16_iotlb()
1298 kunmap_local(kaddr); in getu16_iotlb()
1333 void *kaddr = kmap_local_page(ivec.iov.bvec[0].bv_page); in putu16_iotlb() local
1334 void *to = kaddr + ivec.iov.bvec[0].bv_offset; in putu16_iotlb()
1337 kunmap_local(kaddr); in putu16_iotlb()

12