/drivers/char/hw_random/ |
D | n2-drv.c | 148 unsigned long ra = __pa(&np->scratch_control[0]); in n2rng_control_settle_v2() 390 unsigned long ra = __pa(&np->test_data); in n2rng_data_read() 424 unsigned long ra = __pa(&np->test_data); in n2rng_guest_check() 434 unsigned long post_ctl_ra = __pa(post_control); in n2rng_entropy_diag_read() 435 unsigned long pre_ctl_ra = __pa(pre_control); in n2rng_entropy_diag_read() 436 unsigned long buffer_ra = __pa(buffer); in n2rng_entropy_diag_read() 558 unsigned long ctl_ra = __pa(&up->control[0]); in n2rng_control_configure_units()
|
/drivers/crypto/nx/ |
D | nx.c | 107 sg_addr = __pa(sg_addr); in nx_build_sg_list() 263 nx_ctx->op.csbcpb = __pa(nx_ctx->csbcpb); in nx_ctx_init() 264 nx_ctx->op.in = __pa(nx_ctx->in_sg); in nx_ctx_init() 265 nx_ctx->op.out = __pa(nx_ctx->out_sg); in nx_ctx_init() 271 nx_ctx->op_aead.csbcpb = __pa(nx_ctx->csbcpb_aead); in nx_ctx_init() 272 nx_ctx->op_aead.in = __pa(nx_ctx->in_sg); in nx_ctx_init() 273 nx_ctx->op_aead.out = __pa(nx_ctx->out_sg); in nx_ctx_init()
|
/drivers/s390/net/ |
D | claw.c | 1060 p_last->r_TIC_1.cda =(__u32)__pa(&p_end->read2_nop1); in add_claw_reads() 1061 p_last->r_TIC_2.cda =(__u32)__pa(&p_end->read2_nop1); in add_claw_reads() 1070 p_last->r_TIC_1.cda = (__u32)__pa(&p_end->read1_nop1); in add_claw_reads() 1071 p_last->r_TIC_2.cda = (__u32)__pa(&p_end->read1_nop1); in add_claw_reads() 1083 temp_ccw.cda= (__u32)__pa(&p_first->read); in add_claw_reads() 1100 (__u32)__pa(&p_first->read); in add_claw_reads() 1102 (__u32)__pa(&p_first->read); in add_claw_reads() 1110 (__u32)__pa(&p_first->read); in add_claw_reads() 1112 (__u32)__pa(&p_first->read); in add_claw_reads() 1358 (__u32)__pa(&p_this_ccw->write); in claw_hw_tx() [all …]
|
/drivers/net/ethernet/ibm/ehea/ |
D | ehea_phyp.c | 144 __pa(cb_addr), /* R8 */ in ehea_h_query_ehea_qp() 418 __pa(cb_addr), /* R8 */ in ehea_h_modify_ehea_qp() 531 cb_logaddr = __pa(cb_addr); in ehea_h_query_ehea() 548 u64 cb_logaddr = __pa(cb_addr); in ehea_h_query_ehea_port() 570 u64 cb_logaddr = __pa(cb_addr); in ehea_h_modify_ehea_port() 624 __pa(rblock), /* R6 */ in ehea_h_error_data()
|
D | ehea_qmr.c | 165 rpage = __pa(vpage); in ehea_create_cq() 290 rpage = __pa(vpage); in ehea_create_eq() 395 rpage = __pa(vpage); in ehea_qp_alloc_register() 788 index = __pa(caddr) >> SECTION_SIZE_BITS; in ehea_map_vaddr() 820 u64 pt_abs = __pa(pt); in ehea_reg_mr_section() 828 pt[m] = __pa(pg); in ehea_reg_mr_section()
|
/drivers/tty/serial/ |
D | sunhv.c | 62 unsigned long ra = __pa(xmit->buf + xmit->tail); in transmit_chars_write() 124 unsigned long ra = __pa(con_read_page); in receive_chars_read() 439 unsigned long ra = __pa(con_write_page); in sunhv_console_write_paged() 554 port->membase = (unsigned char __iomem *) __pa(port); in hv_probe()
|
/drivers/s390/char/ |
D | raw3270.c | 157 rq->ccw.cda = __pa(rq->buffer); in raw3270_request_alloc() 182 rq->ccw.cda = __pa(rq->buffer); in raw3270_request_reset() 216 rq->ccw.cda = __pa(data); in raw3270_request_set_data() 226 rq->ccw.cda = __pa(ib->data); in raw3270_request_set_idal() 593 rp->init_readmod.ccw.cda = (__u32) __pa(rp->init_data); in raw3270_read_modified() 612 rp->init_readpart.ccw.cda = (__u32) __pa(&rp->init_data); in raw3270_writesf_readpart() 651 rp->init_reset.ccw.cda = (__u32) __pa(rp->init_data); in __raw3270_reset_device()
|
/drivers/cpuidle/ |
D | cpuidle-calxeda.c | 37 return psci_ops.cpu_suspend(ps, __pa(cpu_resume)); in calxeda_idle_finish()
|
/drivers/infiniband/hw/ehca/ |
D | hcp_if.c | 379 u64 r_cb = __pa(query_port_response_block); in hipz_h_query_port() 421 u64 r_cb = __pa(query_hca_rblock); in hipz_h_query_hca() 560 __pa(mqpcb), /* r7 */ in hipz_h_modify_qp() 578 __pa(qqpcb), /* r6 */ in hipz_h_query_qp() 927 u64 r_cb = __pa(rblock); in hipz_h_error_data()
|
D | ehca_eq.c | 104 rpage = __pa(vpage); in ehca_create_eq()
|
/drivers/staging/goldfish/ |
D | README | 10 - Use dma coherent memory not kmalloc/__pa for the memory (this is just
|
/drivers/watchdog/ |
D | wdrtas.c | 124 WDRTAS_SP_SPI, __pa(rtas_data_buf), in wdrtas_get_interval() 176 (void *)__pa(wdrtas_logbuffer), in wdrtas_timer_keepalive()
|
/drivers/char/ |
D | mspec.c | 119 nid = nasid_to_cnodeid(get_node_number(__pa(addr))); in mspec_zero_block() 379 phys = __pa(scratch_page[nid]); in mspec_init()
|
D | mem.c | 52 return addr + count <= __pa(high_memory); in valid_phys_addr_range() 285 return addr >= __pa(high_memory); in uncached_access() 381 pfn = __pa((u64)vma->vm_pgoff << PAGE_SHIFT) >> PAGE_SHIFT; in mmap_kmem()
|
/drivers/lguest/x86/ |
D | core.c | 103 pages->state.host_cr3 = __pa(current->mm->pgd); in copy_in_guest_info() 174 "1"(__pa(cpu->lg->pgdirs[cpu->cpu_pgd].pgdir)), in run_guest_once()
|
/drivers/s390/block/ |
D | xpram.c | 100 : "+d" (cc) : "a" (__pa(page_addr)), "d" (xpage_index) : "cc"); in xpram_page_in() 130 : "+d" (cc) : "a" (__pa(page_addr)), "d" (xpage_index) : "cc"); in xpram_page_out()
|
/drivers/crypto/ |
D | n2_core.c | 555 ent->src_addr = __pa(walk.data); in n2_do_async_digest() 557 ent->auth_iv_addr = __pa(hash_loc); in n2_do_async_digest() 561 ent->dest_addr = __pa(hash_loc); in n2_do_async_digest() 568 ent->src_addr = __pa(walk.data); in n2_do_async_digest() 642 __pa(&ctx->hash_key), in n2_hmac_async_digest() 852 ent->enc_key_addr = __pa(&ctx->key); in __n2_crypt_chunk() 1046 iv_paddr = __pa(rctx->walk.iv); in n2_do_chaining() 1063 iv_paddr = __pa(rctx->walk.iv); in n2_do_chaining() 1673 hv_ret = sun4v_ncs_qconf(q_type, __pa(p->q), in spu_queue_register()
|
/drivers/misc/sgi-xp/ |
D | xp_sn2.c | 83 return __pa(addr); in xp_pa_sn2()
|
/drivers/iommu/ |
D | ipmmu-vmsa.c | 338 ttbr = __pa(domain->pgd); in ipmmu_domain_init_context() 533 *pmd = __pmd(__pa(pte) | PMD_NSTABLE | PMD_TYPE_TABLE); in ipmmu_alloc_pte() 553 *pud = __pud(__pa(pmd) | PMD_NSTABLE | PMD_TYPE_TABLE); in ipmmu_alloc_pmd() 760 *pmd = __pmd(__pa(start) | PMD_NSTABLE | PMD_TYPE_TABLE); in ipmmu_split_pmd()
|
D | amd_iommu_v2.c | 436 __pa(empty_page_table)); in mn_invalidate_range_start() 458 __pa(pasid_state->mm->pgd)); in mn_invalidate_range_end() 683 __pa(pasid_state->mm->pgd)); in amd_iommu_bind_pasid()
|
/drivers/net/ethernet/brocade/bna/ |
D | bfa_ioc.h | 66 #define bfa_alen_set(__alen, __len, __pa) \ argument 67 __bfa_alen_set(__alen, __len, (u64)__pa)
|
/drivers/lguest/ |
D | page_tables.c | 325 set_pgd(spgd, __pgd(__pa(ptepage) | pgd_flags)); in find_spte() 359 set_pmd(spmd, __pmd(__pa(ptepage) | pmd_flags)); in find_spte() 1166 regs_page = pfn_to_page(__pa(cpu->regs_page) >> PAGE_SHIFT); in map_switcher_in_guest()
|
/drivers/platform/goldfish/ |
D | goldfish_pipe.c | 172 if (paddr != (__pa(aps))) in valid_batchbuffer_addr() 189 paddr = __pa(aps); in setup_access_params_addr()
|
D | goldfish_pipe_v2.c | 732 (u64)(unsigned long)__pa(pipe->command_buffer); in goldfish_pipe_open() 1099 u64 paddr = __pa(&dev->buffers->signalled_pipe_buffers); in goldfish_pipe_device_init_v2() 1104 paddr = __pa(&dev->buffers->open_command_params); in goldfish_pipe_device_init_v2()
|
/drivers/clk/ |
D | clk-xgene.c | 238 pclk->name, __pa(pclk->param.csr_reg), in xgene_clk_enable() 249 pclk->name, __pa(pclk->param.csr_reg), in xgene_clk_enable()
|