Home
last modified time | relevance | path

Searched refs:__pa (Results 1 – 25 of 67) sorted by relevance

123

/drivers/char/hw_random/
Dn2-drv.c148 unsigned long ra = __pa(&np->scratch_control[0]); in n2rng_control_settle_v2()
390 unsigned long ra = __pa(&np->test_data); in n2rng_data_read()
424 unsigned long ra = __pa(&np->test_data); in n2rng_guest_check()
434 unsigned long post_ctl_ra = __pa(post_control); in n2rng_entropy_diag_read()
435 unsigned long pre_ctl_ra = __pa(pre_control); in n2rng_entropy_diag_read()
436 unsigned long buffer_ra = __pa(buffer); in n2rng_entropy_diag_read()
558 unsigned long ctl_ra = __pa(&up->control[0]); in n2rng_control_configure_units()
/drivers/crypto/nx/
Dnx.c107 sg_addr = __pa(sg_addr); in nx_build_sg_list()
263 nx_ctx->op.csbcpb = __pa(nx_ctx->csbcpb); in nx_ctx_init()
264 nx_ctx->op.in = __pa(nx_ctx->in_sg); in nx_ctx_init()
265 nx_ctx->op.out = __pa(nx_ctx->out_sg); in nx_ctx_init()
271 nx_ctx->op_aead.csbcpb = __pa(nx_ctx->csbcpb_aead); in nx_ctx_init()
272 nx_ctx->op_aead.in = __pa(nx_ctx->in_sg); in nx_ctx_init()
273 nx_ctx->op_aead.out = __pa(nx_ctx->out_sg); in nx_ctx_init()
/drivers/s390/net/
Dclaw.c1060 p_last->r_TIC_1.cda =(__u32)__pa(&p_end->read2_nop1); in add_claw_reads()
1061 p_last->r_TIC_2.cda =(__u32)__pa(&p_end->read2_nop1); in add_claw_reads()
1070 p_last->r_TIC_1.cda = (__u32)__pa(&p_end->read1_nop1); in add_claw_reads()
1071 p_last->r_TIC_2.cda = (__u32)__pa(&p_end->read1_nop1); in add_claw_reads()
1083 temp_ccw.cda= (__u32)__pa(&p_first->read); in add_claw_reads()
1100 (__u32)__pa(&p_first->read); in add_claw_reads()
1102 (__u32)__pa(&p_first->read); in add_claw_reads()
1110 (__u32)__pa(&p_first->read); in add_claw_reads()
1112 (__u32)__pa(&p_first->read); in add_claw_reads()
1358 (__u32)__pa(&p_this_ccw->write); in claw_hw_tx()
[all …]
/drivers/net/ethernet/ibm/ehea/
Dehea_phyp.c144 __pa(cb_addr), /* R8 */ in ehea_h_query_ehea_qp()
418 __pa(cb_addr), /* R8 */ in ehea_h_modify_ehea_qp()
531 cb_logaddr = __pa(cb_addr); in ehea_h_query_ehea()
548 u64 cb_logaddr = __pa(cb_addr); in ehea_h_query_ehea_port()
570 u64 cb_logaddr = __pa(cb_addr); in ehea_h_modify_ehea_port()
624 __pa(rblock), /* R6 */ in ehea_h_error_data()
Dehea_qmr.c165 rpage = __pa(vpage); in ehea_create_cq()
290 rpage = __pa(vpage); in ehea_create_eq()
395 rpage = __pa(vpage); in ehea_qp_alloc_register()
788 index = __pa(caddr) >> SECTION_SIZE_BITS; in ehea_map_vaddr()
820 u64 pt_abs = __pa(pt); in ehea_reg_mr_section()
828 pt[m] = __pa(pg); in ehea_reg_mr_section()
/drivers/tty/serial/
Dsunhv.c62 unsigned long ra = __pa(xmit->buf + xmit->tail); in transmit_chars_write()
124 unsigned long ra = __pa(con_read_page); in receive_chars_read()
439 unsigned long ra = __pa(con_write_page); in sunhv_console_write_paged()
554 port->membase = (unsigned char __iomem *) __pa(port); in hv_probe()
/drivers/s390/char/
Draw3270.c157 rq->ccw.cda = __pa(rq->buffer); in raw3270_request_alloc()
182 rq->ccw.cda = __pa(rq->buffer); in raw3270_request_reset()
216 rq->ccw.cda = __pa(data); in raw3270_request_set_data()
226 rq->ccw.cda = __pa(ib->data); in raw3270_request_set_idal()
593 rp->init_readmod.ccw.cda = (__u32) __pa(rp->init_data); in raw3270_read_modified()
612 rp->init_readpart.ccw.cda = (__u32) __pa(&rp->init_data); in raw3270_writesf_readpart()
651 rp->init_reset.ccw.cda = (__u32) __pa(rp->init_data); in __raw3270_reset_device()
/drivers/cpuidle/
Dcpuidle-calxeda.c37 return psci_ops.cpu_suspend(ps, __pa(cpu_resume)); in calxeda_idle_finish()
/drivers/infiniband/hw/ehca/
Dhcp_if.c379 u64 r_cb = __pa(query_port_response_block); in hipz_h_query_port()
421 u64 r_cb = __pa(query_hca_rblock); in hipz_h_query_hca()
560 __pa(mqpcb), /* r7 */ in hipz_h_modify_qp()
578 __pa(qqpcb), /* r6 */ in hipz_h_query_qp()
927 u64 r_cb = __pa(rblock); in hipz_h_error_data()
Dehca_eq.c104 rpage = __pa(vpage); in ehca_create_eq()
/drivers/staging/goldfish/
DREADME10 - Use dma coherent memory not kmalloc/__pa for the memory (this is just
/drivers/watchdog/
Dwdrtas.c124 WDRTAS_SP_SPI, __pa(rtas_data_buf), in wdrtas_get_interval()
176 (void *)__pa(wdrtas_logbuffer), in wdrtas_timer_keepalive()
/drivers/char/
Dmspec.c119 nid = nasid_to_cnodeid(get_node_number(__pa(addr))); in mspec_zero_block()
379 phys = __pa(scratch_page[nid]); in mspec_init()
Dmem.c52 return addr + count <= __pa(high_memory); in valid_phys_addr_range()
285 return addr >= __pa(high_memory); in uncached_access()
381 pfn = __pa((u64)vma->vm_pgoff << PAGE_SHIFT) >> PAGE_SHIFT; in mmap_kmem()
/drivers/lguest/x86/
Dcore.c103 pages->state.host_cr3 = __pa(current->mm->pgd); in copy_in_guest_info()
174 "1"(__pa(cpu->lg->pgdirs[cpu->cpu_pgd].pgdir)), in run_guest_once()
/drivers/s390/block/
Dxpram.c100 : "+d" (cc) : "a" (__pa(page_addr)), "d" (xpage_index) : "cc"); in xpram_page_in()
130 : "+d" (cc) : "a" (__pa(page_addr)), "d" (xpage_index) : "cc"); in xpram_page_out()
/drivers/crypto/
Dn2_core.c555 ent->src_addr = __pa(walk.data); in n2_do_async_digest()
557 ent->auth_iv_addr = __pa(hash_loc); in n2_do_async_digest()
561 ent->dest_addr = __pa(hash_loc); in n2_do_async_digest()
568 ent->src_addr = __pa(walk.data); in n2_do_async_digest()
642 __pa(&ctx->hash_key), in n2_hmac_async_digest()
852 ent->enc_key_addr = __pa(&ctx->key); in __n2_crypt_chunk()
1046 iv_paddr = __pa(rctx->walk.iv); in n2_do_chaining()
1063 iv_paddr = __pa(rctx->walk.iv); in n2_do_chaining()
1673 hv_ret = sun4v_ncs_qconf(q_type, __pa(p->q), in spu_queue_register()
/drivers/misc/sgi-xp/
Dxp_sn2.c83 return __pa(addr); in xp_pa_sn2()
/drivers/iommu/
Dipmmu-vmsa.c338 ttbr = __pa(domain->pgd); in ipmmu_domain_init_context()
533 *pmd = __pmd(__pa(pte) | PMD_NSTABLE | PMD_TYPE_TABLE); in ipmmu_alloc_pte()
553 *pud = __pud(__pa(pmd) | PMD_NSTABLE | PMD_TYPE_TABLE); in ipmmu_alloc_pmd()
760 *pmd = __pmd(__pa(start) | PMD_NSTABLE | PMD_TYPE_TABLE); in ipmmu_split_pmd()
Damd_iommu_v2.c436 __pa(empty_page_table)); in mn_invalidate_range_start()
458 __pa(pasid_state->mm->pgd)); in mn_invalidate_range_end()
683 __pa(pasid_state->mm->pgd)); in amd_iommu_bind_pasid()
/drivers/net/ethernet/brocade/bna/
Dbfa_ioc.h66 #define bfa_alen_set(__alen, __len, __pa) \ argument
67 __bfa_alen_set(__alen, __len, (u64)__pa)
/drivers/lguest/
Dpage_tables.c325 set_pgd(spgd, __pgd(__pa(ptepage) | pgd_flags)); in find_spte()
359 set_pmd(spmd, __pmd(__pa(ptepage) | pmd_flags)); in find_spte()
1166 regs_page = pfn_to_page(__pa(cpu->regs_page) >> PAGE_SHIFT); in map_switcher_in_guest()
/drivers/platform/goldfish/
Dgoldfish_pipe.c172 if (paddr != (__pa(aps))) in valid_batchbuffer_addr()
189 paddr = __pa(aps); in setup_access_params_addr()
Dgoldfish_pipe_v2.c732 (u64)(unsigned long)__pa(pipe->command_buffer); in goldfish_pipe_open()
1099 u64 paddr = __pa(&dev->buffers->signalled_pipe_buffers); in goldfish_pipe_device_init_v2()
1104 paddr = __pa(&dev->buffers->open_command_params); in goldfish_pipe_device_init_v2()
/drivers/clk/
Dclk-xgene.c238 pclk->name, __pa(pclk->param.csr_reg), in xgene_clk_enable()
249 pclk->name, __pa(pclk->param.csr_reg), in xgene_clk_enable()

123