/arch/mips/alchemy/devboards/ |
D | bcsr.c | 22 void __iomem *raddr; member 40 bcsr_regs[i].raddr = (void __iomem *)bcsr2_phys + in bcsr_init() 43 bcsr_regs[i].raddr = (void __iomem *)bcsr1_phys + in bcsr_init() 56 r = __raw_readw(bcsr_regs[reg].raddr); in bcsr_read() 67 __raw_writew(val, bcsr_regs[reg].raddr); in bcsr_write() 79 r = __raw_readw(bcsr_regs[reg].raddr); in bcsr_mod() 82 __raw_writew(r, bcsr_regs[reg].raddr); in bcsr_mod()
|
/arch/s390/mm/ |
D | gmap.c | 1211 static int gmap_protect_rmap(struct gmap *sg, unsigned long raddr, in gmap_protect_rmap() argument 1230 rmap->raddr = raddr; in gmap_protect_rmap() 1292 static void gmap_unshadow_page(struct gmap *sg, unsigned long raddr) in gmap_unshadow_page() argument 1297 table = gmap_table_walk(sg, raddr, 0); /* get page table pointer */ in gmap_unshadow_page() 1300 gmap_call_notifier(sg, raddr, raddr + _PAGE_SIZE - 1); in gmap_unshadow_page() 1301 ptep_unshadow_pte(sg->mm, raddr, (pte_t *) table); in gmap_unshadow_page() 1312 static void __gmap_unshadow_pgt(struct gmap *sg, unsigned long raddr, in __gmap_unshadow_pgt() argument 1318 for (i = 0; i < _PAGE_ENTRIES; i++, raddr += _PAGE_SIZE) in __gmap_unshadow_pgt() 1329 static void gmap_unshadow_pgt(struct gmap *sg, unsigned long raddr) in gmap_unshadow_pgt() argument 1335 ste = gmap_table_walk(sg, raddr, 1); /* get segment pointer */ in gmap_unshadow_pgt() [all …]
|
/arch/parisc/kernel/ |
D | perf.c | 189 static int perf_stop_counters(uint32_t *raddr); 218 uint32_t raddr[4]; in perf_config() local 221 error = perf_stop_counters(raddr); in perf_config() 430 uint32_t raddr[4]; in perf_ioctl() local 441 error_start = perf_stop_counters(raddr); in perf_ioctl() 449 if (copy_to_user((void __user *)arg, raddr, in perf_ioctl() 450 sizeof (raddr)) != 0) { in perf_ioctl() 549 static int perf_stop_counters(uint32_t *raddr) in perf_stop_counters() argument 569 raddr[0] = (uint32_t)tmp64; in perf_stop_counters() 575 raddr[1] = (uint32_t)tmp64; in perf_stop_counters() [all …]
|
/arch/powerpc/kvm/ |
D | trace_pr.h | 72 __field( ulong, raddr ) 81 __entry->raddr = pte->pte.raddr; 89 __entry->vpage, __entry->raddr, __entry->flags) 101 __field( ulong, raddr ) 110 __entry->raddr = pte->pte.raddr; 118 __entry->vpage, __entry->raddr, __entry->flags)
|
D | book3s_32_mmu.c | 162 pte->raddr = bat->brpn | (eaddr & ~bat->bepi_mask); in kvmppc_mmu_book3s_32_xlate_bat() 221 pte->raddr = (pte1 & ~(0xFFFULL)) | (eaddr & 0xFFF); in kvmppc_mmu_book3s_32_xlate_pte() 305 pte->raddr = vcpu->arch.magic_page_pa | (pte->raddr & 0xfff); in kvmppc_mmu_book3s_32_xlate() 306 pte->raddr &= KVM_PAM; in kvmppc_mmu_book3s_32_xlate()
|
D | book3s_hv_rm_xics.c | 802 unsigned int *raddr; in this_cpu_inc_rm() local 805 raddr = per_cpu_ptr(addr, cpu); in this_cpu_inc_rm() 806 l = (unsigned long)raddr; in this_cpu_inc_rm() 809 l = vmalloc_to_phys(raddr); in this_cpu_inc_rm() 810 raddr = (unsigned int *)l; in this_cpu_inc_rm() 812 ++*raddr; in this_cpu_inc_rm()
|
D | book3s_64_mmu_host.c | 89 unsigned long gfn = orig_pte->raddr >> PAGE_SHIFT; in kvmppc_mmu_map_page() 97 pfn = kvmppc_gpa_to_pfn(vcpu, orig_pte->raddr, iswrite, &writable); in kvmppc_mmu_map_page() 100 orig_pte->raddr); in kvmppc_mmu_map_page() 147 hpaddr |= orig_pte->raddr & (~0xfffULL & ~PAGE_MASK); in kvmppc_mmu_map_page()
|
D | book3s_64_mmu.c | 215 gpte->raddr = vcpu->arch.magic_page_pa | (gpte->raddr & 0xfff); in kvmppc_mmu_book3s_64_xlate() 216 gpte->raddr &= KVM_PAM; in kvmppc_mmu_book3s_64_xlate() 297 gpte->raddr = (r & HPTE_R_RPN & ~eaddr_mask) | (eaddr & eaddr_mask); in kvmppc_mmu_book3s_64_xlate() 324 eaddr, avpn, gpte->vpage, gpte->raddr); in kvmppc_mmu_book3s_64_xlate()
|
D | book3s_32_mmu_host.c | 148 hpaddr = kvmppc_gpa_to_pfn(vcpu, orig_pte->raddr, iswrite, &writable); in kvmppc_mmu_map_page() 151 orig_pte->raddr); in kvmppc_mmu_map_page() 200 mark_page_dirty(vcpu->kvm, orig_pte->raddr >> PAGE_SHIFT); in kvmppc_mmu_map_page()
|
D | book3s_mmu_hpte.c | 312 if ((pte->pte.raddr >= pa_start) && in kvmppc_mmu_pte_pflush() 313 (pte->pte.raddr < pa_end)) in kvmppc_mmu_pte_pflush()
|
D | book3s_pr.c | 647 hpage = gfn_to_page(vcpu->kvm, pte->raddr >> PAGE_SHIFT); in kvmppc_patch_dcbz() 651 hpage_offset = pte->raddr & ~PAGE_MASK; in kvmppc_patch_dcbz() 706 pte.raddr = eaddr & KVM_PAM; in kvmppc_handle_pagefault() 720 ((pte.raddr & SPLIT_HACK_MASK) == SPLIT_HACK_OFFS)) in kvmppc_handle_pagefault() 721 pte.raddr &= ~SPLIT_HACK_MASK; in kvmppc_handle_pagefault() 765 } else if (kvmppc_visible_gpa(vcpu, pte.raddr)) { in kvmppc_handle_pagefault() 788 vcpu->arch.paddr_accessed = pte.raddr; in kvmppc_handle_pagefault()
|
D | powerpc.c | 345 *eaddr = pte.raddr; in kvmppc_st() 352 ((pte.raddr & KVM_PAM & PAGE_MASK) == mp_pa) && in kvmppc_st() 360 if (kvm_write_guest(vcpu->kvm, pte.raddr, ptr, size)) in kvmppc_st() 388 *eaddr = pte.raddr; in kvmppc_ld() 398 ((pte.raddr & KVM_PAM & PAGE_MASK) == mp_pa) && in kvmppc_ld() 407 rc = kvm_read_guest(vcpu->kvm, pte.raddr, ptr, size); in kvmppc_ld()
|
D | book3s.c | 466 pte->raddr = eaddr & KVM_PAM; in kvmppc_xlate() 477 pte->raddr &= ~SPLIT_HACK_MASK; in kvmppc_xlate()
|
D | e500_mmu.c | 436 u64 raddr = get_tlb_raddr(gtlbe); in kvmppc_e500_emul_tlbwe() local 444 kvmppc_mmu_map(vcpu, eaddr, raddr, index_of(tlbsel, esel)); in kvmppc_e500_emul_tlbwe()
|
D | book3s_hv_nested.c | 1435 gpte.raddr, kvm->arch.lpid); in kvmhv_handle_nested_set_rc() 1548 gpa = gpte.raddr; in __kvmhv_nested_page_fault()
|
/arch/mips/include/asm/octeon/ |
D | cvmx-ipd.h | 200 ipd_pwp_ptr_fifo_ctl.s.raddr = in cvmx_ipd_free_ptr() 247 ipd_prc_port_ptr_fifo_ctl.s.raddr = in cvmx_ipd_free_ptr() 276 ipd_prc_hold_ptr_fifo_ctl.s.raddr = in cvmx_ipd_free_ptr() 304 ipd_pwp_ptr_fifo_ctl.s.raddr = in cvmx_ipd_free_ptr()
|
D | cvmx-ipd-defs.h | 594 uint64_t raddr:8; member 596 uint64_t raddr:8; 628 uint64_t raddr:3; member 630 uint64_t raddr:3; 1098 uint64_t raddr:7; member 1100 uint64_t raddr:7; 1164 uint64_t raddr:3; member 1166 uint64_t raddr:3; 1184 uint64_t raddr:7; member 1186 uint64_t raddr:7; [all …]
|
/arch/powerpc/kernel/ |
D | fadump.c | 1038 unsigned long raddr, rstart, rend, rlast, hole_size; in fadump_relocate() local 1043 raddr = paddr; in fadump_relocate() 1050 raddr += fw_dump.boot_mem_dest_addr - hole_size; in fadump_relocate() 1057 pr_debug("vmcoreinfo: paddr = 0x%lx, raddr = 0x%lx\n", paddr, raddr); in fadump_relocate() 1058 return raddr; in fadump_relocate() 1063 unsigned long long raddr, offset; in fadump_create_elfcore_headers() local 1111 raddr = fw_dump.boot_mem_addr[0]; in fadump_create_elfcore_headers() 1126 if (mbase == raddr) { in fadump_create_elfcore_headers() 1135 raddr = fw_dump.boot_mem_addr[++j]; in fadump_create_elfcore_headers()
|
/arch/mips/mm/ |
D | uasm-micromips.c | 217 long raddr = (long)rel->addr; in __resolve_relocs() local 222 *rel->addr |= (build_bimm(laddr - (raddr + 4)) << 16); in __resolve_relocs() 224 *rel->addr |= build_bimm(laddr - (raddr + 4)); in __resolve_relocs()
|
D | uasm-mips.c | 279 long raddr = (long)rel->addr; in __resolve_relocs() local 283 *rel->addr |= build_bimm(laddr - (raddr + 4)); in __resolve_relocs()
|
/arch/s390/kvm/ |
D | gaccess.c | 619 union raddress raddr = {.addr = gva}; in guest_translate() local 715 raddr.rfaa = rtte.fc1.rfaa; in guest_translate() 743 raddr.sfaa = ste.fc1.sfaa; in guest_translate() 760 raddr.pfra = pte.pfra; in guest_translate() 762 raddr.addr = kvm_s390_real_to_abs(vcpu, raddr.addr); in guest_translate() 772 if (kvm_is_error_gpa(vcpu->kvm, raddr.addr)) in guest_translate() 774 *gpa = raddr.addr; in guest_translate()
|
/arch/s390/include/asm/ |
D | gmap.h | 80 unsigned long raddr; member
|
/arch/sparc/kernel/ |
D | sys_sparc_64.c | 392 ulong raddr; in SYSCALL_DEFINE6() local 393 err = do_shmat(first, ptr, (int)second, &raddr, SHMLBA); in SYSCALL_DEFINE6() 395 if (put_user(raddr, in SYSCALL_DEFINE6()
|
/arch/x86/kernel/kprobes/ |
D | core.c | 114 s32 raddr; in __synthesize_relative_insn() member 118 insn->raddr = (s32)((long)(to) - ((long)(from) + 5)); in __synthesize_relative_insn()
|
/arch/powerpc/include/asm/ |
D | kvm_host.h | 376 ulong raddr; member
|