Home
last modified time | relevance | path

Searched refs:raddr (Results 1 – 22 of 22) sorted by relevance

/arch/mips/alchemy/devboards/
Dbcsr.c20 void __iomem *raddr; member
38 bcsr_regs[i].raddr = (void __iomem *)bcsr2_phys + in bcsr_init()
41 bcsr_regs[i].raddr = (void __iomem *)bcsr1_phys + in bcsr_init()
54 r = __raw_readw(bcsr_regs[reg].raddr); in bcsr_read()
65 __raw_writew(val, bcsr_regs[reg].raddr); in bcsr_write()
77 r = __raw_readw(bcsr_regs[reg].raddr); in bcsr_mod()
80 __raw_writew(r, bcsr_regs[reg].raddr); in bcsr_mod()
/arch/parisc/kernel/
Dperf.c202 static int perf_stop_counters(uint32_t *raddr);
231 uint32_t raddr[4]; in perf_config() local
234 error = perf_stop_counters(raddr); in perf_config()
444 uint32_t raddr[4]; in perf_ioctl() local
455 error_start = perf_stop_counters(raddr); in perf_ioctl()
463 if (copy_to_user((void __user *)arg, raddr, in perf_ioctl()
464 sizeof (raddr)) != 0) { in perf_ioctl()
565 static int perf_stop_counters(uint32_t *raddr) in perf_stop_counters() argument
585 raddr[0] = (uint32_t)tmp64; in perf_stop_counters()
591 raddr[1] = (uint32_t)tmp64; in perf_stop_counters()
[all …]
/arch/powerpc/kvm/
Dtrace_pr.h71 __field( ulong, raddr )
80 __entry->raddr = pte->pte.raddr;
88 __entry->vpage, __entry->raddr, __entry->flags)
100 __field( ulong, raddr )
109 __entry->raddr = pte->pte.raddr;
117 __entry->vpage, __entry->raddr, __entry->flags)
Dbook3s_32_mmu.c179 pte->raddr = bat->brpn | (eaddr & ~bat->bepi_mask); in kvmppc_mmu_book3s_32_xlate_bat()
237 pte->raddr = (pte1 & ~(0xFFFULL)) | (eaddr & 0xFFF); in kvmppc_mmu_book3s_32_xlate_pte()
320 pte->raddr = vcpu->arch.magic_page_pa | (pte->raddr & 0xfff); in kvmppc_mmu_book3s_32_xlate()
321 pte->raddr &= KVM_PAM; in kvmppc_mmu_book3s_32_xlate()
Dbook3s_64_mmu_host.c100 unsigned long gfn = orig_pte->raddr >> PAGE_SHIFT; in kvmppc_mmu_map_page()
108 pfn = kvmppc_gpa_to_pfn(vcpu, orig_pte->raddr, iswrite, &writable); in kvmppc_mmu_map_page()
111 orig_pte->raddr); in kvmppc_mmu_map_page()
155 hpaddr |= orig_pte->raddr & (~0xfffULL & ~PAGE_MASK); in kvmppc_mmu_map_page()
Dbook3s_64_mmu.c232 gpte->raddr = vcpu->arch.magic_page_pa | (gpte->raddr & 0xfff); in kvmppc_mmu_book3s_64_xlate()
233 gpte->raddr &= KVM_PAM; in kvmppc_mmu_book3s_64_xlate()
313 gpte->raddr = (r & HPTE_R_RPN & ~eaddr_mask) | (eaddr & eaddr_mask); in kvmppc_mmu_book3s_64_xlate()
339 eaddr, avpn, gpte->vpage, gpte->raddr); in kvmppc_mmu_book3s_64_xlate()
Dbook3s_32_mmu_host.c160 hpaddr = kvmppc_gpa_to_pfn(vcpu, orig_pte->raddr, iswrite, &writable); in kvmppc_mmu_map_page()
163 orig_pte->raddr); in kvmppc_mmu_map_page()
212 mark_page_dirty(vcpu->kvm, orig_pte->raddr >> PAGE_SHIFT); in kvmppc_mmu_map_page()
Dbook3s_mmu_hpte.c323 if ((pte->pte.raddr >= pa_start) && in kvmppc_mmu_pte_pflush()
324 (pte->pte.raddr < pa_end)) in kvmppc_mmu_pte_pflush()
Dpowerpc.c326 *eaddr = pte.raddr; in kvmppc_st()
333 ((pte.raddr & KVM_PAM & PAGE_MASK) == mp_pa) && in kvmppc_st()
341 if (kvm_write_guest(vcpu->kvm, pte.raddr, ptr, size)) in kvmppc_st()
362 *eaddr = pte.raddr; in kvmppc_ld()
372 ((pte.raddr & KVM_PAM & PAGE_MASK) == mp_pa) && in kvmppc_ld()
380 if (kvm_read_guest(vcpu->kvm, pte.raddr, ptr, size)) in kvmppc_ld()
Dbook3s_pr.c495 hpage = gfn_to_page(vcpu->kvm, pte->raddr >> PAGE_SHIFT); in kvmppc_patch_dcbz()
499 hpage_offset = pte->raddr & ~PAGE_MASK; in kvmppc_patch_dcbz()
555 pte.raddr = eaddr & KVM_PAM; in kvmppc_handle_pagefault()
568 ((pte.raddr & SPLIT_HACK_MASK) == SPLIT_HACK_OFFS)) in kvmppc_handle_pagefault()
569 pte.raddr &= ~SPLIT_HACK_MASK; in kvmppc_handle_pagefault()
618 kvmppc_visible_gpa(vcpu, pte.raddr)) { in kvmppc_handle_pagefault()
641 vcpu->arch.paddr_accessed = pte.raddr; in kvmppc_handle_pagefault()
Dbook3s.c410 pte->raddr = eaddr & KVM_PAM; in kvmppc_xlate()
421 pte->raddr &= ~SPLIT_HACK_MASK; in kvmppc_xlate()
De500_mmu.c439 u64 raddr = get_tlb_raddr(gtlbe); in kvmppc_e500_emul_tlbwe() local
447 kvmppc_mmu_map(vcpu, eaddr, raddr, index_of(tlbsel, esel)); in kvmppc_e500_emul_tlbwe()
Dbooke.c1944 pte->raddr = (vcpu->arch.magic_page_pa & PAGE_MASK) | in kvmppc_xlate()
1974 pte->raddr = (gpaddr & PAGE_MASK) | (eaddr & ~PAGE_MASK); in kvmppc_xlate()
Dbook3s_64_mmu_hv.c369 gpte->raddr = kvmppc_mmu_get_real_addr(v, gr, eaddr); in kvmppc_mmu_book3s_64_hv_xlate()
/arch/mips/include/asm/octeon/
Dcvmx-ipd.h199 ipd_pwp_ptr_fifo_ctl.s.raddr = in cvmx_ipd_free_ptr()
246 ipd_prc_port_ptr_fifo_ctl.s.raddr = in cvmx_ipd_free_ptr()
275 ipd_prc_hold_ptr_fifo_ctl.s.raddr = in cvmx_ipd_free_ptr()
303 ipd_pwp_ptr_fifo_ctl.s.raddr = in cvmx_ipd_free_ptr()
Dcvmx-ipd-defs.h716 uint64_t raddr:8; member
718 uint64_t raddr:8;
754 uint64_t raddr:3; member
756 uint64_t raddr:3;
1378 uint64_t raddr:7; member
1380 uint64_t raddr:7;
1481 uint64_t raddr:3; member
1483 uint64_t raddr:3;
1516 uint64_t raddr:7; member
1518 uint64_t raddr:7;
[all …]
/arch/s390/kvm/
Dgaccess.c544 union raddress raddr = {.addr = gva}; in guest_translate() local
637 raddr.rfaa = rtte.fc1.rfaa; in guest_translate()
664 raddr.sfaa = ste.fc1.sfaa; in guest_translate()
682 raddr.pfra = pte.pfra; in guest_translate()
684 raddr.addr = kvm_s390_real_to_abs(vcpu, raddr.addr); in guest_translate()
688 if (kvm_is_error_gpa(vcpu->kvm, raddr.addr)) in guest_translate()
690 *gpa = raddr.addr; in guest_translate()
/arch/mips/mm/
Duasm-micromips.c212 long raddr = (long)rel->addr; in __resolve_relocs() local
217 *rel->addr |= (build_bimm(laddr - (raddr + 4)) << 16); in __resolve_relocs()
219 *rel->addr |= build_bimm(laddr - (raddr + 4)); in __resolve_relocs()
Duasm-mips.c235 long raddr = (long)rel->addr; in __resolve_relocs() local
239 *rel->addr |= build_bimm(laddr - (raddr + 4)); in __resolve_relocs()
/arch/sparc/kernel/
Dsys_sparc_64.c385 ulong raddr; in SYSCALL_DEFINE6() local
386 err = do_shmat(first, ptr, (int)second, &raddr, SHMLBA); in SYSCALL_DEFINE6()
388 if (put_user(raddr, in SYSCALL_DEFINE6()
/arch/x86/kernel/kprobes/
Dcore.c122 s32 raddr; in __synthesize_relative_insn() member
126 insn->raddr = (s32)((long)(to) - ((long)(from) + 5)); in __synthesize_relative_insn()
/arch/powerpc/include/asm/
Dkvm_host.h342 ulong raddr; member