| /drivers/gpu/drm/amd/amdgpu/ |
| D | amdgpu_virt.c | 85 struct amdgpu_virt *virt = &adev->virt; in amdgpu_virt_request_full_gpu() local 88 if (virt->ops && virt->ops->req_full_gpu) { in amdgpu_virt_request_full_gpu() 89 r = virt->ops->req_full_gpu(adev, init); in amdgpu_virt_request_full_gpu() 95 adev->virt.caps &= ~AMDGPU_SRIOV_CAPS_RUNTIME; in amdgpu_virt_request_full_gpu() 110 struct amdgpu_virt *virt = &adev->virt; in amdgpu_virt_release_full_gpu() local 113 if (virt->ops && virt->ops->rel_full_gpu) { in amdgpu_virt_release_full_gpu() 114 r = virt->ops->rel_full_gpu(adev, init); in amdgpu_virt_release_full_gpu() 118 adev->virt.caps |= AMDGPU_SRIOV_CAPS_RUNTIME; in amdgpu_virt_release_full_gpu() 131 struct amdgpu_virt *virt = &adev->virt; in amdgpu_virt_reset_gpu() local 134 if (virt->ops && virt->ops->reset_gpu) { in amdgpu_virt_reset_gpu() [all …]
|
| D | amdgpu_vf_error.c | 41 mutex_lock(&adev->virt.vf_errors.lock); in amdgpu_vf_error_put() 42 index = adev->virt.vf_errors.write_count % AMDGPU_VF_ERROR_ENTRY_SIZE; in amdgpu_vf_error_put() 43 adev->virt.vf_errors.code [index] = error_code; in amdgpu_vf_error_put() 44 adev->virt.vf_errors.flags [index] = error_flags; in amdgpu_vf_error_put() 45 adev->virt.vf_errors.data [index] = error_data; in amdgpu_vf_error_put() 46 adev->virt.vf_errors.write_count ++; in amdgpu_vf_error_put() 47 mutex_unlock(&adev->virt.vf_errors.lock); in amdgpu_vf_error_put() 58 (!adev->virt.ops) || (!adev->virt.ops->trans_msg)) { in amdgpu_vf_error_trans_all() 69 mutex_lock(&adev->virt.vf_errors.lock); in amdgpu_vf_error_trans_all() 71 …if (adev->virt.vf_errors.write_count - adev->virt.vf_errors.read_count > AMDGPU_VF_ERROR_ENTRY_SIZ… in amdgpu_vf_error_trans_all() [all …]
|
| D | mxgpu_nv.c | 195 adev->virt.req_init_data_ver = 0; in xgpu_nv_send_access_requests_with_param() 198 adev->virt.req_init_data_ver = in xgpu_nv_send_access_requests_with_param() 202 if (adev->virt.req_init_data_ver < 1) in xgpu_nv_send_access_requests_with_param() 203 adev->virt.req_init_data_ver = 1; in xgpu_nv_send_access_requests_with_param() 209 adev->virt.fw_reserve.checksum_key = in xgpu_nv_send_access_requests_with_param() 312 struct amdgpu_virt *virt = container_of(work, struct amdgpu_virt, flr_work); in xgpu_nv_mailbox_flr_work() local 313 struct amdgpu_device *adev = container_of(virt, struct amdgpu_device, virt); in xgpu_nv_mailbox_flr_work() 363 &adev->virt.flr_work), in xgpu_nv_mailbox_rcv_irq() 393 adev->virt.ack_irq.num_types = 1; in xgpu_nv_mailbox_set_irq_funcs() 394 adev->virt.ack_irq.funcs = &xgpu_nv_mailbox_ack_irq_funcs; in xgpu_nv_mailbox_set_irq_funcs() [all …]
|
| D | mxgpu_ai.c | 181 adev->virt.fw_reserve.checksum_key = in xgpu_ai_send_access_requests() 189 adev->virt.req_init_data_ver = 0; in xgpu_ai_send_access_requests() 275 struct amdgpu_virt *virt = container_of(work, struct amdgpu_virt, flr_work); in xgpu_ai_mailbox_flr_work() local 276 struct amdgpu_device *adev = container_of(virt, struct amdgpu_device, virt); in xgpu_ai_mailbox_flr_work() 320 &adev->virt.flr_work), in xgpu_ai_mailbox_rcv_irq() 353 adev->virt.ack_irq.num_types = 1; in xgpu_ai_mailbox_set_irq_funcs() 354 adev->virt.ack_irq.funcs = &xgpu_ai_mailbox_ack_irq_funcs; in xgpu_ai_mailbox_set_irq_funcs() 355 adev->virt.rcv_irq.num_types = 1; in xgpu_ai_mailbox_set_irq_funcs() 356 adev->virt.rcv_irq.funcs = &xgpu_ai_mailbox_rcv_irq_funcs; in xgpu_ai_mailbox_set_irq_funcs() 363 r = amdgpu_irq_add_id(adev, SOC15_IH_CLIENTID_BIF, 135, &adev->virt.rcv_irq); in xgpu_ai_mailbox_add_irq_id() [all …]
|
| D | amdgpu_virt.h | 285 ((adev)->virt.caps & AMDGPU_SRIOV_CAPS_ENABLE_IOV) 288 ((adev)->virt.caps & AMDGPU_SRIOV_CAPS_IS_VF) 291 ((adev)->virt.caps & AMDGPU_SRIOV_CAPS_SRIOV_VBIOS) 294 ((adev)->virt.caps & AMDGPU_SRIOV_CAPS_RUNTIME) 301 ((adev)->virt.gim_feature & (AMDGIM_FEATURE_INDIRECT_REG_ACCESS))) 305 ((adev)->virt.reg_access & (AMDGIM_FEATURE_IH_REG_PSP_EN))) 309 ((adev)->virt.reg_access & (AMDGIM_FEATURE_MMHUB_REG_RLC_EN))) 313 ((adev)->virt.reg_access & (AMDGIM_FEATURE_GC_REG_RLC_EN))) 319 ((adev)->virt.caps & AMDGPU_PASSTHROUGH_MODE) 322 ((adev)->virt.caps & AMDGPU_VF_MMIO_ACCESS_PROTECT) [all …]
|
| D | mxgpu_vi.c | 515 struct amdgpu_virt *virt = container_of(work, struct amdgpu_virt, flr_work); in xgpu_vi_mailbox_flr_work() local 516 struct amdgpu_device *adev = container_of(virt, struct amdgpu_device, virt); in xgpu_vi_mailbox_flr_work() 560 &adev->virt.flr_work), in xgpu_vi_mailbox_rcv_irq() 580 adev->virt.ack_irq.num_types = 1; in xgpu_vi_mailbox_set_irq_funcs() 581 adev->virt.ack_irq.funcs = &xgpu_vi_mailbox_ack_irq_funcs; in xgpu_vi_mailbox_set_irq_funcs() 582 adev->virt.rcv_irq.num_types = 1; in xgpu_vi_mailbox_set_irq_funcs() 583 adev->virt.rcv_irq.funcs = &xgpu_vi_mailbox_rcv_irq_funcs; in xgpu_vi_mailbox_set_irq_funcs() 590 r = amdgpu_irq_add_id(adev, AMDGPU_IRQ_CLIENTID_LEGACY, 135, &adev->virt.rcv_irq); in xgpu_vi_mailbox_add_irq_id() 594 r = amdgpu_irq_add_id(adev, AMDGPU_IRQ_CLIENTID_LEGACY, 138, &adev->virt.ack_irq); in xgpu_vi_mailbox_add_irq_id() 596 amdgpu_irq_put(adev, &adev->virt.rcv_irq, 0); in xgpu_vi_mailbox_add_irq_id() [all …]
|
| /drivers/staging/media/atomisp/pci/hmm/ |
| D | hmm.c | 227 void hmm_free(ia_css_ptr virt) in hmm_free() argument 231 if (WARN_ON(virt == mmgr_EXCEPTION)) in hmm_free() 234 bo = hmm_bo_device_search_start(&bo_device, (unsigned int)virt); in hmm_free() 239 (unsigned int)virt); in hmm_free() 273 static int load_and_flush_by_kmap(ia_css_ptr virt, void *data, in load_and_flush_by_kmap() argument 281 bo = hmm_bo_device_search_in_range(&bo_device, virt); in load_and_flush_by_kmap() 282 ret = hmm_check_bo(bo, virt); in load_and_flush_by_kmap() 288 idx = (virt - bo->start) >> PAGE_SHIFT; in load_and_flush_by_kmap() 289 offset = (virt - bo->start) - (idx << PAGE_SHIFT); in load_and_flush_by_kmap() 301 virt += len; /* update virt for next loop */ in load_and_flush_by_kmap() [all …]
|
| /drivers/gpio/ |
| D | gpio-stp-xway.c | 82 void __iomem *virt; member 105 return (xway_stp_r32(chip->virt, XWAY_STP_CPU0) & BIT(gpio)); in xway_stp_get() 124 xway_stp_w32(chip->virt, chip->shadow, XWAY_STP_CPU0); in xway_stp_set() 126 xway_stp_w32_mask(chip->virt, 0, XWAY_STP_CON_SWU, XWAY_STP_CON0); in xway_stp_set() 170 xway_stp_w32(chip->virt, 0, XWAY_STP_AR); in xway_stp_hw_init() 171 xway_stp_w32(chip->virt, 0, XWAY_STP_CPU0); in xway_stp_hw_init() 172 xway_stp_w32(chip->virt, 0, XWAY_STP_CPU1); in xway_stp_hw_init() 173 xway_stp_w32(chip->virt, XWAY_STP_CON_SWU, XWAY_STP_CON0); in xway_stp_hw_init() 174 xway_stp_w32(chip->virt, 0, XWAY_STP_CON1); in xway_stp_hw_init() 177 xway_stp_w32_mask(chip->virt, XWAY_STP_EDGE_MASK, in xway_stp_hw_init() [all …]
|
| /drivers/scsi/lpfc/ |
| D | lpfc_mem.c | 130 pool->elements[i].virt = dma_pool_alloc(phba->lpfc_mbuf_pool, in lpfc_mem_alloc() 132 if (!pool->elements[i].virt) in lpfc_mem_alloc() 203 dma_pool_free(phba->lpfc_mbuf_pool, pool->elements[i].virt, in lpfc_mem_alloc() 275 dma_pool_free(phba->lpfc_mbuf_pool, pool->elements[i].virt, in lpfc_mem_free() 349 phba->cgn_i->virt, phba->cgn_i->phys); in lpfc_mem_free_all() 398 ret = pool->elements[pool->current_count].virt; in lpfc_mbuf_alloc() 420 __lpfc_mbuf_free(struct lpfc_hba * phba, void *virt, dma_addr_t dma) in __lpfc_mbuf_free() argument 425 pool->elements[pool->current_count].virt = virt; in __lpfc_mbuf_free() 429 dma_pool_free(phba->lpfc_mbuf_pool, virt, dma); in __lpfc_mbuf_free() 448 lpfc_mbuf_free(struct lpfc_hba * phba, void *virt, dma_addr_t dma) in lpfc_mbuf_free() argument [all …]
|
| /drivers/net/ethernet/qlogic/qed/ |
| D | qed_chain.c | 73 void *virt, *virt_next; in qed_chain_free_next_ptr() local 77 virt = chain->p_virt_addr; in qed_chain_free_next_ptr() 81 if (!virt) in qed_chain_free_next_ptr() 84 next = virt + size; in qed_chain_free_next_ptr() 88 dma_free_coherent(dev, chain->page_size, virt, phys); in qed_chain_free_next_ptr() 90 virt = virt_next; in qed_chain_free_next_ptr() 202 void *virt, *virt_prev = NULL; in qed_chain_alloc_next_ptr() local 207 virt = dma_alloc_coherent(dev, chain->page_size, &phys, in qed_chain_alloc_next_ptr() 209 if (!virt) in qed_chain_alloc_next_ptr() 213 qed_chain_init_mem(chain, virt, phys); in qed_chain_alloc_next_ptr() [all …]
|
| /drivers/staging/media/atomisp/include/hmm/ |
| D | hmm.h | 43 int hmm_load(ia_css_ptr virt, void *data, unsigned int bytes); 44 int hmm_store(ia_css_ptr virt, const void *data, unsigned int bytes); 45 int hmm_set(ia_css_ptr virt, int c, unsigned int bytes); 46 int hmm_flush(ia_css_ptr virt, unsigned int bytes); 51 phys_addr_t hmm_virt_to_phys(ia_css_ptr virt); 60 void *hmm_vmap(ia_css_ptr virt, bool cached); 61 void hmm_vunmap(ia_css_ptr virt); 67 void hmm_flush_vmap(ia_css_ptr virt); 77 int hmm_mmap(struct vm_area_struct *vma, ia_css_ptr virt);
|
| /drivers/mtd/maps/ |
| D | nettel.c | 220 nettel_amd_map.virt = ioremap(amdaddr, maxsize); in nettel_init() 221 if (!nettel_amd_map.virt) { in nettel_init() 280 iounmap(nettel_amd_map.virt); in nettel_init() 281 nettel_amd_map.virt = NULL; in nettel_init() 306 nettel_intel_map.virt = ioremap(intel0addr, maxsize); in nettel_init() 307 if (!nettel_intel_map.virt) { in nettel_init() 337 iounmap(nettel_intel_map.virt); in nettel_init() 340 nettel_intel_map.virt = ioremap(intel0addr, maxsize); in nettel_init() 341 if (!nettel_intel_map.virt) { in nettel_init() 412 iounmap(nettel_intel_map.virt); in nettel_init() [all …]
|
| D | dc21285.c | 57 val.x[0] = *(uint8_t*)(map->virt + ofs); in dc21285_read8() 64 val.x[0] = *(uint16_t*)(map->virt + ofs); in dc21285_read16() 71 val.x[0] = *(uint32_t*)(map->virt + ofs); in dc21285_read32() 77 memcpy(to, (void*)(map->virt + from), len); in dc21285_copy_from() 86 *(uint8_t*)(map->virt + adr) = d.x[0]; in dc21285_write8() 95 *(uint16_t*)(map->virt + adr) = d.x[0]; in dc21285_write16() 102 *(uint32_t*)(map->virt + adr) = d.x[0]; in dc21285_write32() 179 dc21285_map.virt = ioremap(DC21285_FLASH, 16*1024*1024); in init_dc21285() 180 if (!dc21285_map.virt) { in init_dc21285() 192 iounmap(dc21285_map.virt); in init_dc21285() [all …]
|
| D | ts5500_flash.c | 59 ts5500_map.virt = ioremap(ts5500_map.phys, ts5500_map.size); in init_ts5500_map() 61 if (!ts5500_map.virt) { in init_ts5500_map() 84 iounmap(ts5500_map.virt); in init_ts5500_map() 96 if (ts5500_map.virt) { in cleanup_ts5500_map() 97 iounmap(ts5500_map.virt); in cleanup_ts5500_map() 98 ts5500_map.virt = NULL; in cleanup_ts5500_map()
|
| D | netsc520.c | 85 netsc520_map.virt = ioremap(netsc520_map.phys, netsc520_map.size); in init_netsc520() 87 if (!netsc520_map.virt) { in init_netsc520() 101 iounmap(netsc520_map.virt); in init_netsc520() 116 if (netsc520_map.virt) { in cleanup_netsc520() 117 iounmap(netsc520_map.virt); in cleanup_netsc520() 118 netsc520_map.virt = NULL; in cleanup_netsc520()
|
| D | cfi_flagadm.c | 97 flagadm_map.virt = ioremap(FLASH_PHYS_ADDR, in init_flagadm() 100 if (!flagadm_map.virt) { in init_flagadm() 115 iounmap((void __iomem *)flagadm_map.virt); in init_flagadm() 125 if (flagadm_map.virt) { in cleanup_flagadm() 126 iounmap((void __iomem *)flagadm_map.virt); in cleanup_flagadm() 127 flagadm_map.virt = NULL; in cleanup_flagadm()
|
| D | amd76xrom.c | 32 void __iomem *virt; member 94 if (window->virt) { in amd76xrom_cleanup() 95 iounmap(window->virt); in amd76xrom_cleanup() 96 window->virt = NULL; in amd76xrom_cleanup() 166 window->virt = ioremap(window->phys, window->size); in amd76xrom_init_one() 167 if (!window->virt) { in amd76xrom_init_one() 200 map->map.virt = (void __iomem *) in amd76xrom_init_one() 201 (((unsigned long)(window->virt)) + offset); in amd76xrom_init_one() 255 map->map.virt = window->virt; in amd76xrom_init_one()
|
| /drivers/media/platform/samsung/s5p-mfc/ |
| D | s5p_mfc_opr.c | 51 b->virt = dev->mem_virt + offset; in s5p_mfc_alloc_priv_buf() 58 b->virt = dma_alloc_coherent(mem_dev, b->size, &b->dma, GFP_KERNEL); in s5p_mfc_alloc_priv_buf() 59 if (!b->virt) in s5p_mfc_alloc_priv_buf() 64 dma_free_coherent(mem_dev, b->size, b->virt, b->dma); in s5p_mfc_alloc_priv_buf() 69 mfc_debug(3, "Allocated addr %p %pad\n", b->virt, &b->dma); in s5p_mfc_alloc_priv_buf() 84 b->virt = dma_alloc_coherent(mem_dev, b->size, &b->dma, GFP_KERNEL); in s5p_mfc_alloc_generic_buf() 85 if (!b->virt) in s5p_mfc_alloc_generic_buf() 88 mfc_debug(3, "Allocated addr %p %pad\n", b->virt, &b->dma); in s5p_mfc_alloc_generic_buf() 106 dma_free_coherent(mem_dev, b->size, b->virt, b->dma); in s5p_mfc_release_priv_buf() 108 b->virt = NULL; in s5p_mfc_release_priv_buf() [all …]
|
| /drivers/net/ipa/ |
| D | gsi.c | 200 iowrite32(val, gsi->virt + reg_offset(reg)); in gsi_irq_type_update() 228 iowrite32(~0, gsi->virt + reg_offset(reg)); in gsi_irq_ev_ctrl_enable() 231 iowrite32(val, gsi->virt + reg_offset(reg)); in gsi_irq_ev_ctrl_enable() 243 iowrite32(0, gsi->virt + reg_offset(reg)); in gsi_irq_ev_ctrl_disable() 261 iowrite32(~0, gsi->virt + reg_offset(reg)); in gsi_irq_ch_ctrl_enable() 264 iowrite32(val, gsi->virt + reg_offset(reg)); in gsi_irq_ch_ctrl_enable() 277 iowrite32(0, gsi->virt + reg_offset(reg)); in gsi_irq_ch_ctrl_disable() 290 iowrite32(val, gsi->virt + reg_offset(reg)); in gsi_irq_ieob_enable_one() 310 iowrite32(val, gsi->virt + reg_offset(reg)); in gsi_irq_ieob_disable() 328 iowrite32(ERROR_INT, gsi->virt + reg_offset(reg)); in gsi_irq_enable() [all …]
|
| /drivers/firmware/ |
| D | iscsi_ibft_find.c | 55 void *virt = NULL; in reserve_ibft_region() local 74 if (virt) in reserve_ibft_region() 75 early_memunmap(virt, PAGE_SIZE); in reserve_ibft_region() 76 virt = early_memremap_ro(pos, PAGE_SIZE); in reserve_ibft_region() 81 if (memcmp(virt + (pos - virt_pos), ibft_signs[i].sign, in reserve_ibft_region() 84 (unsigned long *)(virt + pos - virt_pos + 4); in reserve_ibft_region() 99 early_memunmap(virt, PAGE_SIZE); in reserve_ibft_region()
|
| /drivers/media/platform/amphion/ |
| D | vpu_helpers.c | 242 void *virt; in vpu_helper_copy_from_stream_buffer() local 253 virt = stream_buffer->virt; in vpu_helper_copy_from_stream_buffer() 259 memcpy(dst, virt + (offset - start), size); in vpu_helper_copy_from_stream_buffer() 261 memcpy(dst, virt + (offset - start), end - offset); in vpu_helper_copy_from_stream_buffer() 262 memcpy(dst + end - offset, virt, size + offset - end); in vpu_helper_copy_from_stream_buffer() 276 void *virt; in vpu_helper_copy_to_stream_buffer() local 287 virt = stream_buffer->virt; in vpu_helper_copy_to_stream_buffer() 292 memcpy(virt + (offset - start), src, size); in vpu_helper_copy_to_stream_buffer() 294 memcpy(virt + (offset - start), src, end - offset); in vpu_helper_copy_to_stream_buffer() 295 memcpy(virt, src + end - offset, size + offset - end); in vpu_helper_copy_to_stream_buffer() [all …]
|
| /drivers/scsi/elx/libefc/ |
| D | efc_device.c | 539 struct fc_frame_header *hdr = cbdata->header->dma.virt; in __efc_d_init() 542 efc_node_save_sparms(node, cbdata->payload->dma.virt); in __efc_d_init() 544 cbdata->header->dma.virt, in __efc_d_init() 568 struct fc_frame_header *hdr = cbdata->header->dma.virt; in __efc_d_init() 573 cbdata->payload->dma.virt, in __efc_d_init() 592 struct fc_frame_header *hdr = cbdata->header->dma.virt; in __efc_d_init() 617 struct fc_frame_header *hdr = cbdata->header->dma.virt; in __efc_d_init() 708 efc_node_save_sparms(node, cbdata->payload->dma.virt); in __efc_d_wait_plogi_rsp() 710 cbdata->header->dma.virt, in __efc_d_wait_plogi_rsp() 730 efc_process_prli_payload(node, cbdata->payload->dma.virt); in __efc_d_wait_plogi_rsp() [all …]
|
| /drivers/gpu/drm/tegra/ |
| D | vic.c | 92 hdr = vic->falcon.firmware.virt; in vic_boot() 97 hdr = vic->falcon.firmware.virt + in vic_boot() 194 vic->falcon.firmware.virt, in vic_exit() 198 vic->falcon.firmware.virt, in vic_exit() 218 void *virt; in vic_load_firmware() local 223 if (vic->falcon.firmware.virt) { in vic_load_firmware() 235 virt = dma_alloc_coherent(vic->dev, size, &iova, GFP_KERNEL); in vic_load_firmware() 236 if (!virt) { in vic_load_firmware() 241 virt = tegra_drm_alloc(tegra, size, &iova); in vic_load_firmware() 242 if (IS_ERR(virt)) { in vic_load_firmware() [all …]
|
| /drivers/platform/x86/dell/ |
| D | dcdbas.c | 56 smi_buffer->virt = dma_alloc_coherent(&dcdbas_pdev->dev, size, in dcdbas_smi_alloc() 58 if (!smi_buffer->virt) { in dcdbas_smi_alloc() 75 if (!smi_buffer->virt) in dcdbas_smi_free() 81 smi_buffer->virt, smi_buffer->dma); in dcdbas_smi_free() 82 smi_buffer->virt = NULL; in dcdbas_smi_free() 93 if (!smi_buf.virt || wsmt_enabled) in smi_data_buf_free() 119 if (smi_buf.virt) in smi_data_buf_realloc() 120 memcpy(tmp.virt, smi_buf.virt, smi_buf.size); in smi_data_buf_realloc() 171 ret = memory_read_from_buffer(buf, count, &pos, smi_buf.virt, in smi_data_read() 192 memcpy(smi_buf.virt + pos, buf, count); in smi_data_write() [all …]
|
| /drivers/virt/coco/ |
| D | Kconfig | 10 source "drivers/virt/coco/efi_secret/Kconfig" 12 source "drivers/virt/coco/pkvm-guest/Kconfig" 14 source "drivers/virt/coco/gunyah-guest/Kconfig" 16 source "drivers/virt/coco/sev-guest/Kconfig" 18 source "drivers/virt/coco/tdx-guest/Kconfig"
|