/drivers/scsi/qedi/ |
D | qedi_dbg.c | 14 va_list va; in qedi_dbg_err() local 17 va_start(va, fmt); in qedi_dbg_err() 20 vaf.va = &va; in qedi_dbg_err() 28 va_end(va); in qedi_dbg_err() 35 va_list va; in qedi_dbg_warn() local 38 va_start(va, fmt); in qedi_dbg_warn() 41 vaf.va = &va; in qedi_dbg_warn() 53 va_end(va); in qedi_dbg_warn() 60 va_list va; in qedi_dbg_notice() local 63 va_start(va, fmt); in qedi_dbg_notice() [all …]
|
/drivers/scsi/qedf/ |
D | qedf_dbg.c | 13 va_list va; in qedf_dbg_err() local 16 va_start(va, fmt); in qedf_dbg_err() 19 vaf.va = &va; in qedf_dbg_err() 27 va_end(va); in qedf_dbg_err() 34 va_list va; in qedf_dbg_warn() local 37 va_start(va, fmt); in qedf_dbg_warn() 40 vaf.va = &va; in qedf_dbg_warn() 52 va_end(va); in qedf_dbg_warn() 59 va_list va; in qedf_dbg_notice() local 62 va_start(va, fmt); in qedf_dbg_notice() [all …]
|
/drivers/dio/ |
D | dio.c | 128 void *va; in dio_find() local 140 va = (void *)(pa + DIO_VIRADDRBASE); in dio_find() 142 va = ioremap(pa, PAGE_SIZE); in dio_find() 145 (unsigned char *)va + DIO_IDOFF, 1)) { in dio_find() 147 iounmap(va); in dio_find() 151 prid = DIO_ID(va); in dio_find() 154 secid = DIO_SECID(va); in dio_find() 161 iounmap(va); in dio_find() 202 u_char *va; in dio_init() local 214 va = (void *)(pa + DIO_VIRADDRBASE); in dio_init() [all …]
|
/drivers/video/fbdev/vermilion/ |
D | vermilion.c | 73 static int vmlfb_alloc_vram_area(struct vram_area *va, unsigned max_order, in vmlfb_alloc_vram_area() argument 89 va->logical = in vmlfb_alloc_vram_area() 91 } while (va->logical == 0 && max_order > min_order); in vmlfb_alloc_vram_area() 93 if (!va->logical) in vmlfb_alloc_vram_area() 96 va->phys = virt_to_phys((void *)va->logical); in vmlfb_alloc_vram_area() 97 va->size = PAGE_SIZE << max_order; in vmlfb_alloc_vram_area() 98 va->order = max_order; in vmlfb_alloc_vram_area() 107 memset((void *)va->logical, 0x00, va->size); in vmlfb_alloc_vram_area() 108 for (i = va->logical; i < va->logical + va->size; i += PAGE_SIZE) { in vmlfb_alloc_vram_area() 116 set_pages_uc(virt_to_page(va->logical), va->size >> PAGE_SHIFT); in vmlfb_alloc_vram_area() [all …]
|
/drivers/media/platform/mtk-vcodec/vdec/ |
D | vdec_vp9_if.c | 36 unsigned long va; member 226 if (fb->base_y.va == addr) { in vp9_rm_from_fb_use_list() 277 vsi->frm_bufs[ref_idx].buf.fb->base_y.va); in vp9_ref_cnt_fb() 295 if (vsi->sf_ref_fb[i].fb.base_y.va) { in vp9_free_all_sf_ref_fb() 320 if (vsi->sf_ref_fb[idx].fb.base_y.va && in vp9_get_sf_ref_fb() 329 if (vsi->sf_ref_fb[idx].fb.base_y.va == NULL) in vp9_get_sf_ref_fb() 394 if (mem->va) in vp9_alloc_work_buf() 406 vsi->mv_buf.va = (unsigned long)mem->va; in vp9_alloc_work_buf() 412 if (mem->va) in vp9_alloc_work_buf() 423 vsi->seg_id_buf.va = (unsigned long)mem->va; in vp9_alloc_work_buf() [all …]
|
/drivers/firmware/efi/ |
D | mokvar-table.c | 103 void *va = NULL; in efi_mokvar_table_init() local 137 mokvar_entry = va + cur_offset; in efi_mokvar_table_init() 140 if (va) in efi_mokvar_table_init() 141 early_memunmap(va, map_size); in efi_mokvar_table_init() 150 va = early_memremap(efi.mokvar_table, map_size); in efi_mokvar_table_init() 151 if (!va) { in efi_mokvar_table_init() 156 mokvar_entry = va + cur_offset; in efi_mokvar_table_init() 177 if (va) in efi_mokvar_table_init() 178 early_memunmap(va, map_size); in efi_mokvar_table_init()
|
/drivers/video/fbdev/matrox/ |
D | matroxfb_base.h | 128 static inline unsigned int mga_readb(vaddr_t va, unsigned int offs) { in mga_readb() argument 129 return readb(va.vaddr + offs); in mga_readb() 132 static inline void mga_writeb(vaddr_t va, unsigned int offs, u_int8_t value) { in mga_writeb() argument 133 writeb(value, va.vaddr + offs); in mga_writeb() 136 static inline void mga_writew(vaddr_t va, unsigned int offs, u_int16_t value) { in mga_writew() argument 137 writew(value, va.vaddr + offs); in mga_writew() 140 static inline u_int32_t mga_readl(vaddr_t va, unsigned int offs) { in mga_readl() argument 141 return readl(va.vaddr + offs); in mga_readl() 144 static inline void mga_writel(vaddr_t va, unsigned int offs, u_int32_t value) { in mga_writel() argument 145 writel(value, va.vaddr + offs); in mga_writel() [all …]
|
/drivers/media/platform/mtk-vcodec/ |
D | mtk_vcodec_util.c | 43 mem->va = dma_alloc_coherent(dev, size, &mem->dma_addr, GFP_KERNEL); in mtk_vcodec_mem_alloc() 44 if (!mem->va) { in mtk_vcodec_mem_alloc() 50 mtk_v4l2_debug(3, "[%d] - va = %p", ctx->id, mem->va); in mtk_vcodec_mem_alloc() 66 if (!mem->va) { in mtk_vcodec_mem_free() 72 mtk_v4l2_debug(3, "[%d] - va = %p", ctx->id, mem->va); in mtk_vcodec_mem_free() 77 dma_free_coherent(dev, size, mem->va, mem->dma_addr); in mtk_vcodec_mem_free() 78 mem->va = NULL; in mtk_vcodec_mem_free()
|
/drivers/tee/amdtee/ |
D | shm_pool.c | 15 unsigned long va; in pool_op_alloc() local 18 va = __get_free_pages(GFP_KERNEL | __GFP_ZERO, order); in pool_op_alloc() 19 if (!va) in pool_op_alloc() 22 shm->kaddr = (void *)va; in pool_op_alloc() 23 shm->paddr = __psp_pa((void *)va); in pool_op_alloc() 29 free_pages(va, order); in pool_op_alloc()
|
/drivers/gpu/drm/lima/ |
D | lima_vm.c | 29 #define LIMA_PDE(va) (va >> LIMA_VM_PD_SHIFT) argument 30 #define LIMA_PTE(va) ((va & LIMA_VM_PT_MASK) >> LIMA_VM_PT_SHIFT) argument 31 #define LIMA_PBE(va) (va >> LIMA_VM_PB_SHIFT) argument 32 #define LIMA_BTE(va) ((va & LIMA_VM_BT_MASK) >> LIMA_VM_BT_SHIFT) argument 47 static int lima_vm_map_page(struct lima_vm *vm, dma_addr_t pa, u32 va) in lima_vm_map_page() argument 49 u32 pbe = LIMA_PBE(va); in lima_vm_map_page() 50 u32 bte = LIMA_BTE(va); in lima_vm_map_page() 81 list_for_each_entry(bo_va, &bo->va, list) { in lima_vm_bo_find() 138 list_add_tail(&bo_va->list, &bo->va); in lima_vm_bo_add()
|
/drivers/infiniband/hw/usnic/ |
D | usnic_uiom.c | 199 long unsigned va, size; in usnic_uiom_unmap_sorted_intervals() local 202 va = interval->start << PAGE_SHIFT; in usnic_uiom_unmap_sorted_intervals() 206 usnic_dbg("va 0x%lx size 0x%lx", va, PAGE_SIZE); in usnic_uiom_unmap_sorted_intervals() 207 iommu_unmap(pd->domain, va, PAGE_SIZE); in usnic_uiom_unmap_sorted_intervals() 208 va += PAGE_SIZE; in usnic_uiom_unmap_sorted_intervals() 225 vpn_start = (uiomr->va & PAGE_MASK) >> PAGE_SHIFT; in __usnic_uiom_reg_release() 256 long int va = uiomr->va & PAGE_MASK; in usnic_uiom_map_sorted_intervals() local 264 for (i = 0; i < chunk->nents; i++, va += PAGE_SIZE) { in usnic_uiom_map_sorted_intervals() 266 if ((va >> PAGE_SHIFT) < interval_node->start) in usnic_uiom_map_sorted_intervals() 269 if ((va >> PAGE_SHIFT) == interval_node->start) { in usnic_uiom_map_sorted_intervals() [all …]
|
/drivers/net/wireless/ath/wil6210/ |
D | txrx_edma.c | 52 if (!wil->srings[i].va) in wil_find_free_sring() 65 if (!sring || !sring->va) in wil_sring_free() 71 sz, sring->va, &sring->pa); in wil_sring_free() 73 dma_free_coherent(dev, sz, (void *)sring->va, sring->pa); in wil_sring_free() 75 sring->va = NULL; in wil_sring_free() 96 sring->va = dma_alloc_coherent(dev, sz, &sring->pa, GFP_KERNEL); in wil_sring_alloc() 97 if (!sring->va) in wil_sring_alloc() 100 wil_dbg_misc(wil, "status_ring[%d] 0x%p:%pad\n", sring->size, sring->va, in wil_sring_alloc() 165 &ring->va[i].rx.enhanced; in wil_ring_alloc_skb_edma() 217 (sring->va + (sring->elem_size * sring->swhead)); in wil_get_next_rx_status_msg() [all …]
|
D | debug.c | 17 vaf.va = &args; in __wil_err() 33 vaf.va = &args; in __wil_err_ratelimited() 49 vaf.va = &args; in wil_dbg_ratelimited() 62 vaf.va = &args; in __wil_info() 75 vaf.va = &args; in wil_dbg_trace()
|
D | pmc.c | 18 void *va; member 145 pmc->descriptors[i].va = dma_alloc_coherent(dev, in wil_pmc_alloc() 150 if (unlikely(!pmc->descriptors[i].va)) { in wil_pmc_alloc() 156 u32 *p = (u32 *)pmc->descriptors[i].va + j; in wil_pmc_alloc() 196 for (i = 0; i < num_descriptors && pmc->descriptors[i].va; i++) { in wil_pmc_alloc() 199 pmc->descriptors[i].va, in wil_pmc_alloc() 202 pmc->descriptors[i].va = NULL; in wil_pmc_alloc() 279 i < pmc->num_descriptors && pmc->descriptors[i].va; i++) { in wil_pmc_free() 282 pmc->descriptors[i].va, in wil_pmc_free() 284 pmc->descriptors[i].va = NULL; in wil_pmc_free() [all …]
|
D | txrx.c | 78 if (!vring->va || !txdata->enabled) { in wil_is_tx_idle() 99 if (!vring->va || !txdata->enabled) in wil_is_tx_idle() 113 size_t sz = vring->size * sizeof(vring->va[0]); in wil_vring_alloc() 118 BUILD_BUG_ON(sizeof(vring->va[0]) != 32); in wil_vring_alloc() 124 vring->va = NULL; in wil_vring_alloc() 143 vring->va = dma_alloc_coherent(dev, sz, &vring->pa, GFP_KERNEL); in wil_vring_alloc() 144 if (!vring->va) { in wil_vring_alloc() 160 &vring->va[i].tx.legacy; in wil_vring_alloc() 166 vring->va, &vring->pa, vring->ctx); in wil_vring_alloc() 193 size_t sz = vring->size * sizeof(vring->va[0]); in wil_vring_free() [all …]
|
/drivers/scsi/be2iscsi/ |
D | be_mgmt.c | 42 struct be_bsg_vendor_cmd *req = nonemb_cmd->va; in mgmt_vendor_specific_fw_cmd() 46 memset(nonemb_cmd->va, 0, nonemb_cmd->size); in mgmt_vendor_specific_fw_cmd() 64 nonemb_cmd->va + offset, job->request_len); in mgmt_vendor_specific_fw_cmd() 156 req = nonemb_cmd->va; in mgmt_open_connection() 257 tag_mem->va = nonemb_cmd->va; in beiscsi_exec_nemb_cmd() 271 memcpy(resp_buf, nonemb_cmd->va, resp_buf_len); in beiscsi_exec_nemb_cmd() 289 nonemb_cmd->va, nonemb_cmd->dma); in beiscsi_exec_nemb_cmd() 297 cmd->va = dma_alloc_coherent(&phba->ctrl.pdev->dev, size, &cmd->dma, in beiscsi_prep_nemb_cmd() 299 if (!cmd->va) { in beiscsi_prep_nemb_cmd() 305 be_cmd_hdr_prepare(cmd->va, subsystem, opcode, size); in beiscsi_prep_nemb_cmd() [all …]
|
/drivers/net/ethernet/amd/ |
D | hplance.c | 130 unsigned long va = (d->resource.start + DIO_VIRADDRBASE); in hplance_init() local 135 out_8(va + DIO_IDOFF, 0xff); in hplance_init() 139 dev->base_addr = va; in hplance_init() 147 dev->dev_addr[i] = ((in_8(va + HPLANCE_NVRAMOFF + i*4 + 1) & 0xF) << 4) in hplance_init() 148 | (in_8(va + HPLANCE_NVRAMOFF + i*4 + 3) & 0xF); in hplance_init() 153 lp->lance.base = va; in hplance_init() 154 lp->lance.init_block = (struct lance_init_block *)(va + HPLANCE_MEMOFF); /* CPU addr */ in hplance_init()
|
/drivers/gpu/drm/i915/gvt/ |
D | opregion.c | 228 vgpu_opregion(vgpu)->va = (void *)__get_free_pages(GFP_KERNEL | in intel_vgpu_init_opregion() 231 if (!vgpu_opregion(vgpu)->va) { in intel_vgpu_init_opregion() 237 buf = (u8 *)vgpu_opregion(vgpu)->va; in intel_vgpu_init_opregion() 264 mfn = intel_gvt_hypervisor_virt_to_mfn(vgpu_opregion(vgpu)->va in map_vgpu_opregion() 336 if (!vgpu_opregion(vgpu)->va) in intel_vgpu_clean_opregion() 345 free_pages((unsigned long)vgpu_opregion(vgpu)->va, in intel_vgpu_clean_opregion() 348 vgpu_opregion(vgpu)->va = NULL; in intel_vgpu_clean_opregion() 475 scic = *((u32 *)vgpu_opregion(vgpu)->va + in intel_vgpu_emulate_opregion_request() 477 parm = *((u32 *)vgpu_opregion(vgpu)->va + in intel_vgpu_emulate_opregion_request() 540 *((u32 *)vgpu_opregion(vgpu)->va + in intel_vgpu_emulate_opregion_request() [all …]
|
/drivers/tee/ |
D | tee_shm_pool.c | 15 unsigned long va; in pool_op_gen_alloc() local 19 va = gen_pool_alloc(genpool, s); in pool_op_gen_alloc() 20 if (!va) in pool_op_gen_alloc() 23 memset((void *)va, 0, s); in pool_op_gen_alloc() 24 shm->kaddr = (void *)va; in pool_op_gen_alloc() 25 shm->paddr = gen_pool_virt_to_phys(genpool, va); in pool_op_gen_alloc()
|
/drivers/net/ethernet/emulex/benet/ |
D | be_cmds.c | 670 struct be_mcc_mailbox *mbox = mbox_mem->va; in be_mbox_notify_wait() 822 int i, buf_pages = min(PAGES_4K_SPANNED(mem->va, mem->size), max_pages); in be_cmd_page_addrs_prepare() 836 = &((struct be_mcc_mailbox *)(mbox_mem->va))->wrb; in wrb_from_mbox() 1016 req->num_pages = cpu_to_le16(PAGES_4K_SPANNED(q_mem->va, q_mem->size)); in be_cmd_eq_create() 1178 req->num_pages = cpu_to_le16(PAGES_4K_SPANNED(q_mem->va, q_mem->size)); in be_cmd_cq_create() 1256 req->num_pages = cpu_to_le16(PAGES_4K_SPANNED(q_mem->va, q_mem->size)); in be_cmd_mccq_ext_create() 1321 req->num_pages = cpu_to_le16(PAGES_4K_SPANNED(q_mem->va, q_mem->size)); in be_cmd_mccq_org_create() 1383 req->num_pages = PAGES_4K_SPANNED(q_mem->va, q_mem->size); in be_cmd_txq_create() 1602 hdr = nonemb_cmd->va; in be_cmd_get_stats() 1646 req = nonemb_cmd->va; in lancer_cmd_get_pport_stats() [all …]
|
/drivers/net/ethernet/intel/ice/ |
D | ice_controlq.c | 82 cq->sq.desc_buf.va = dmam_alloc_coherent(ice_hw_to_dev(hw), size, in ice_alloc_ctrlq_sq_ring() 85 if (!cq->sq.desc_buf.va) in ice_alloc_ctrlq_sq_ring() 93 cq->sq.desc_buf.va, cq->sq.desc_buf.pa); in ice_alloc_ctrlq_sq_ring() 94 cq->sq.desc_buf.va = NULL; in ice_alloc_ctrlq_sq_ring() 113 cq->rq.desc_buf.va = dmam_alloc_coherent(ice_hw_to_dev(hw), size, in ice_alloc_ctrlq_rq_ring() 116 if (!cq->rq.desc_buf.va) in ice_alloc_ctrlq_rq_ring() 133 ring->desc_buf.va, ring->desc_buf.pa); in ice_free_cq_ring() 134 ring->desc_buf.va = NULL; in ice_free_cq_ring() 164 bi->va = dmam_alloc_coherent(ice_hw_to_dev(hw), in ice_alloc_rq_bufs() 167 if (!bi->va) in ice_alloc_rq_bufs() [all …]
|
/drivers/soc/samsung/ |
D | s3c-pm-debug.c | 26 va_list va; in s3c_pm_dbg() local 29 va_start(va, fmt); in s3c_pm_dbg() 30 vsnprintf(buff, sizeof(buff), fmt, va); in s3c_pm_dbg() 31 va_end(va); in s3c_pm_dbg()
|
/drivers/net/wireless/ath/ath11k/ |
D | debug.c | 18 vaf.va = &args; in ath11k_info() 33 vaf.va = &args; in ath11k_err() 48 vaf.va = &args; in ath11k_warn() 66 vaf.va = &args; in __ath11k_dbg()
|
/drivers/infiniband/hw/ocrdma/ |
D | ocrdma_stats.c | 76 mem->va = dma_alloc_coherent(&dev->nic_info.pdev->dev, mem->size, in ocrdma_alloc_stats_resources() 78 if (!mem->va) { in ocrdma_alloc_stats_resources() 95 if (mem->va) in ocrdma_release_stats_resources() 97 mem->va, mem->pa); in ocrdma_release_stats_resources() 98 mem->va = NULL; in ocrdma_release_stats_resources() 106 (struct ocrdma_rdma_stats_resp *)dev->stats_mem.va; in ocrdma_resource_stats() 223 (struct ocrdma_rdma_stats_resp *)dev->stats_mem.va; in ocrdma_rx_stats() 269 (struct ocrdma_rdma_stats_resp *)dev->stats_mem.va; in ocrdma_sysfs_rcv_pkts() 280 (struct ocrdma_rdma_stats_resp *)dev->stats_mem.va; in ocrdma_sysfs_rcv_data() 291 (struct ocrdma_rdma_stats_resp *)dev->stats_mem.va; in ocrdma_tx_stats() [all …]
|
D | ocrdma.h | 120 void *va; member 126 void *va; member 131 void *va; member 188 u64 va; member 221 void *va; member 264 struct ocrdma_av *va; member 317 struct ocrdma_cqe *va; member 362 u8 *va; /* virtual address */ member 441 u32 *va; member
|