/arch/arm/mach-rpc/include/mach/ |
D | acornfb.h | 51 u_int rr, v, p; in acornfb_vidc20_find_pll() local 54 rr = 41667 * r; in acornfb_vidc20_find_pll() 56 v = (rr + pixclk / 2) / pixclk; in acornfb_vidc20_find_pll() 61 p = (rr + v / 2) / v; in acornfb_vidc20_find_pll()
|
/arch/powerpc/kvm/ |
D | book3s_32_mmu_host.c | 151 register int rr = 0; in kvmppc_mmu_map_page() local 181 if (rr == 16) { in kvmppc_mmu_map_page() 184 rr = 0; in kvmppc_mmu_map_page() 190 if (!evict && (pteg[rr] & PTE_V)) { in kvmppc_mmu_map_page() 191 rr += 2; in kvmppc_mmu_map_page() 195 dprintk_mmu("KVM: old PTEG: %p (%d)\n", pteg, rr); in kvmppc_mmu_map_page() 221 if (pteg[rr]) { in kvmppc_mmu_map_page() 222 pteg[rr] = 0; in kvmppc_mmu_map_page() 225 pteg[rr + 1] = pteg1; in kvmppc_mmu_map_page() 226 pteg[rr] = pteg0; in kvmppc_mmu_map_page() [all …]
|
/arch/ia64/kvm/ |
D | process.c | 161 union ia64_rr rr, rr1; in vcpu_get_itir_on_fault() local 163 rr.val = vcpu_get_rr(vcpu, ifa); in vcpu_get_itir_on_fault() 165 rr1.ps = rr.ps; in vcpu_get_itir_on_fault() 166 rr1.rid = rr.rid; in vcpu_get_itir_on_fault() 470 ia64_set_rr(7UL << 61, vcpu->arch.host.rr[7]); in vmm_fp_emulate() 781 union ia64_rr rr; in physical_tlb_miss() local 783 rr.val = ia64_get_rr(vadr); in physical_tlb_miss() 786 thash_vhpt_insert(vcpu, pte, (u64)(rr.ps << 2), vadr, type); in physical_tlb_miss() 795 u64 vhpt_adr, gppa, pteval, rr, itir; in kvm_page_fault() local 840 rr = vcpu_get_rr(v, vadr); in kvm_page_fault() [all …]
|
D | vtlb.c | 162 union ia64_rr rr; in vhpt_insert() local 167 rr.val = ia64_get_rr(ifa); in vhpt_insert() 176 head->itir = rr.ps << 2; in vhpt_insert() 303 union ia64_rr rr; in vhpt_purge() local 306 rr.val = ia64_get_rr(va); in vhpt_purge() 307 size = PSIZE(rr.ps); in vhpt_purge() 308 num = 1UL << ((ps < rr.ps) ? 0 : (ps - rr.ps)); in vhpt_purge()
|
D | trampoline.S | 311 mov r16=rr[r0]; \ 315 mov r17=rr[r18]; \ 319 mov r16=rr[r18]; \ 323 mov r17=rr[r18]; \ 327 mov r16=rr[r18]; \ 331 mov r17=rr[r18]; \ 335 mov r16=rr[r18]; \ 361 mov rr[r18]=r20; \ 364 mov rr[r18]=r21; \ 367 mov rr[r18]=r22; \ [all …]
|
D | vcpu.h | 626 union ia64_rr rr; in vrrtomrr() local 627 rr.val = val; in vrrtomrr() 628 rr.rid = (rr.rid << 4) | 0xe; in vrrtomrr() 629 if (rr.ps > PAGE_SHIFT) in vrrtomrr() 630 rr.ps = PAGE_SHIFT; in vrrtomrr() 631 rr.ve = 1; in vrrtomrr() 632 return rr.val; in vrrtomrr()
|
D | optvfault.S | 274 (p7) mov rr[r28]=r19 335 mov rr[r0]=r26 337 mov rr[r28]=r27 389 mov rr[r0]=r26 391 mov rr[r28]=r27 470 mov rr[r0]=r26 472 mov rr[r28]=r27
|
D | kvm-ia64.c | 1210 p_ctx->rr[0] = 0x30; in kvm_arch_vcpu_init() 1211 p_ctx->rr[1] = 0x30; in kvm_arch_vcpu_init() 1212 p_ctx->rr[2] = 0x30; in kvm_arch_vcpu_init() 1213 p_ctx->rr[3] = 0x30; in kvm_arch_vcpu_init() 1214 p_ctx->rr[4] = 0x30; in kvm_arch_vcpu_init() 1215 p_ctx->rr[5] = 0x30; in kvm_arch_vcpu_init() 1216 p_ctx->rr[7] = 0x30; in kvm_arch_vcpu_init()
|
D | asm-offsets.c | 215 DEFINE(VMM_CTX_RR0_OFFSET, offsetof(union context, rr[0])); in foo()
|
D | vcpu.c | 1262 p->u.ptc_g_data.rr = vcpu_get_rr(vcpu, va); in vcpu_ptc_ga()
|
/arch/x86/mm/ |
D | numa.c | 583 int rr, i; in numa_init_array() local 585 rr = first_node(node_online_map); in numa_init_array() 589 numa_set_node(i, rr); in numa_init_array() 590 rr = next_node(rr, node_online_map); in numa_init_array() 591 if (rr == MAX_NUMNODES) in numa_init_array() 592 rr = first_node(node_online_map); in numa_init_array()
|
/arch/mips/kernel/ |
D | branch.c | 368 if (inst.rr.func == MIPS16e_jr_func) { in __MIPS16e_compute_return_epc() 370 if (inst.rr.ra) in __MIPS16e_compute_return_epc() 374 regs->regs[reg16to32[inst.rr.rx]]; in __MIPS16e_compute_return_epc() 376 if (inst.rr.l) { in __MIPS16e_compute_return_epc() 377 if (inst.rr.nd) in __MIPS16e_compute_return_epc()
|
/arch/arm/mach-shmobile/ |
D | intc-sh73a0.c | 294 static void pint_demux(void __iomem *rr, void __iomem *er, int base_irq) in pint_demux() argument 296 unsigned long value = ioread32(rr) & ioread32(er); in pint_demux() 302 iowrite32(~(1 << (31 - k)), rr); in pint_demux()
|
/arch/ia64/mm/ |
D | tlb.c | 400 if (va_rid != RR_TO_RID(p->rr)) in is_tr_overlap() 501 p->rr = ia64_get_rr(va); in ia64_itr_entry() 510 p->rr = ia64_get_rr(va); in ia64_itr_entry()
|
/arch/ia64/include/uapi/asm/ |
D | kvm.h | 86 unsigned long rr[8]; member
|
/arch/x86/crypto/ |
D | cast5-avx-x86_64-asm_64.S | 38 #define rr ((16*4)+16) macro 264 movzbl rr(CTX), %eax; 323 movzbl rr(CTX), %eax;
|
/arch/ia64/include/asm/ |
D | sal.h | 382 rr : 1, member 390 u64 rr[8]; member 863 u64 rr[8]; /* Region Registers */ member
|
D | tlb.h | 72 u64 rr; member
|
D | kvm_host.h | 305 unsigned long rr; member
|
D | pal.h | 450 rr : 1, /* Region regs valid */ member 690 #define pmci_proc_region_regs_valid pme_processor.rr
|
/arch/ia64/kernel/ |
D | head.S | 66 mov _reg=rr[_tmp] 94 mov rr[_idx2]=_tmp;; \ 167 mov rr[_tmp1]=_tmp2 327 mov r17=rr[r2]
|
D | mca.c | 1238 if (old_rr != p->rr) { in mca_insert_tr() 1239 ia64_set_rr(p->ifa, p->rr); in mca_insert_tr() 1252 if (old_rr != p->rr) { in mca_insert_tr()
|
/arch/mips/include/uapi/asm/ |
D | inst.h | 973 struct m16e_rr rr; member
|