/arch/powerpc/kvm/ |
D | book3s_32_mmu.c | 89 u64 *vsid); 99 u64 vsid; in kvmppc_mmu_book3s_32_ea_to_vp() local 105 kvmppc_mmu_book3s_32_esid_to_vsid(vcpu, eaddr >> SID_SHIFT, &vsid); in kvmppc_mmu_book3s_32_ea_to_vp() 106 return (((u64)eaddr >> 12) & 0xffff) | (vsid << 16); in kvmppc_mmu_book3s_32_ea_to_vp() 175 u64 vsid; in kvmppc_mmu_book3s_32_xlate_bat() local 177 eaddr >> SID_SHIFT, &vsid); in kvmppc_mmu_book3s_32_xlate_bat() 178 vsid <<= 16; in kvmppc_mmu_book3s_32_xlate_bat() 179 pte->vpage = (((u64)eaddr >> 12) & 0xffff) | vsid; in kvmppc_mmu_book3s_32_xlate_bat() 354 u64 *vsid) in kvmppc_mmu_book3s_32_esid_to_vsid() argument 371 *vsid = VSID_REAL | esid; in kvmppc_mmu_book3s_32_esid_to_vsid() [all …]
|
D | book3s_64_mmu.c | 67 if (vcpu->arch.slb[i].vsid) in kvmppc_mmu_book3s_64_find_slbe() 73 vcpu->arch.slb[i].vsid); in kvmppc_mmu_book3s_64_find_slbe() 90 (((u64)slb->vsid) << 28); in kvmppc_mmu_book3s_64_ea_to_vp() 92 return (((u64)eaddr >> 12) & 0xffff) | (((u64)slb->vsid) << 16); in kvmppc_mmu_book3s_64_ea_to_vp() 118 hash = slbe->vsid ^ page; in kvmppc_mmu_book3s_64_get_pteg() 129 page, vcpu_book3s->sdr1, pteg, slbe->vsid); in kvmppc_mmu_book3s_64_get_pteg() 149 avpn |= slbe->vsid << (28 - p); in kvmppc_mmu_book3s_64_get_avpn() 337 slbe->vsid = rs >> 12; in kvmppc_mmu_book3s_64_slbmte() 461 u64 *vsid) in kvmppc_mmu_book3s_64_esid_to_vsid() argument 471 gvsid = slb->vsid; in kvmppc_mmu_book3s_64_esid_to_vsid() [all …]
|
D | book3s_32_mmu_host.c | 117 static u32 *kvmppc_mmu_get_pteg(struct kvm_vcpu *vcpu, u32 vsid, u32 eaddr, in kvmppc_mmu_get_pteg() argument 125 hash = ((vsid ^ page) << 6); in kvmppc_mmu_get_pteg() 145 u64 vsid; in kvmppc_mmu_map_page() local 167 vcpu->arch.mmu.esid_to_vsid(vcpu, orig_pte->eaddr >> SID_SHIFT, &vsid); in kvmppc_mmu_map_page() 168 map = find_sid_vsid(vcpu, vsid); in kvmppc_mmu_map_page() 171 map = find_sid_vsid(vcpu, vsid); in kvmppc_mmu_map_page() 175 vsid = map->host_vsid; in kvmppc_mmu_map_page() 176 va = (vsid << SID_SHIFT) | (eaddr & ~ESID_MASK); in kvmppc_mmu_map_page() 185 pteg = kvmppc_mmu_get_pteg(vcpu, vsid, eaddr, primary); in kvmppc_mmu_map_page() 203 pteg0 = ((eaddr & 0x0fffffff) >> 22) | (vsid << 7) | PTE_V | in kvmppc_mmu_map_page()
|
D | book3s_64_mmu_host.c | 85 u64 vsid; in kvmppc_mmu_map_page() local 104 vcpu->arch.mmu.esid_to_vsid(vcpu, orig_pte->eaddr >> SID_SHIFT, &vsid); in kvmppc_mmu_map_page() 105 map = find_sid_vsid(vcpu, vsid); in kvmppc_mmu_map_page() 109 map = find_sid_vsid(vcpu, vsid); in kvmppc_mmu_map_page() 113 vsid, orig_pte->eaddr); in kvmppc_mmu_map_page() 119 vsid = map->host_vsid; in kvmppc_mmu_map_page() 120 va = hpt_va(orig_pte->eaddr, vsid, MMU_SEGSIZE_256M); in kvmppc_mmu_map_page() 289 svcpu->slb[slb_index].vsid = slb_vsid; in kvmppc_mmu_map_segment()
|
D | book3s_hv_rm_mmu.c | 658 unsigned long vsid, hash; in kvmppc_hv_find_lock_hpte() local 675 vsid = (slb_v & ~SLB_VSID_B) >> SLB_VSID_SHIFT_1T; in kvmppc_hv_find_lock_hpte() 676 vsid ^= vsid << 25; in kvmppc_hv_find_lock_hpte() 679 vsid = (slb_v & ~SLB_VSID_B) >> SLB_VSID_SHIFT; in kvmppc_hv_find_lock_hpte() 681 hash = (vsid ^ ((eaddr & somask) >> pshift)) & HPT_HASH_MASK; in kvmppc_hv_find_lock_hpte()
|
D | book3s_pr.c | 273 u64 vsid; in kvmppc_handle_pagefault() local 295 vcpu->arch.mmu.esid_to_vsid(vcpu, eaddr >> SID_SHIFT, &vsid); in kvmppc_handle_pagefault() 301 pte.vpage |= vsid; in kvmppc_handle_pagefault() 303 if (vsid == -1) in kvmppc_handle_pagefault()
|
/arch/powerpc/mm/ |
D | stab.c | 37 static int make_ste(unsigned long stab, unsigned long esid, unsigned long vsid) in make_ste() argument 45 vsid_data = vsid << STE_VSID_SHIFT; in make_ste() 120 unsigned long vsid; in __ste_allocate() local 126 vsid = get_kernel_vsid(ea, MMU_SEGSIZE_256M); in __ste_allocate() 131 vsid = get_vsid(mm->context.id, ea, MMU_SEGSIZE_256M); in __ste_allocate() 134 stab_entry = make_ste(get_paca()->stab_addr, GET_ESID(ea), vsid); in __ste_allocate() 274 unsigned long vsid = get_kernel_vsid(PAGE_OFFSET, MMU_SEGSIZE_256M); in stab_initialize() local 278 make_ste(stab, GET_ESID(PAGE_OFFSET), vsid); in stab_initialize()
|
D | tlb_hash64.c | 46 unsigned long vsid, vaddr; in hpte_need_flush() local 83 vsid = get_vsid(mm->context.id, addr, ssize); in hpte_need_flush() 84 WARN_ON(vsid == 0); in hpte_need_flush() 86 vsid = get_kernel_vsid(addr, mmu_kernel_ssize); in hpte_need_flush() 89 vaddr = hpt_va(addr, vsid, ssize); in hpte_need_flush()
|
D | hash_utils_64.c | 194 unsigned long vsid = get_kernel_vsid(vaddr, ssize); in htab_bolt_mapping() local 195 unsigned long va = hpt_va(vaddr, vsid, ssize); in htab_bolt_mapping() 893 unsigned long vsid, unsigned long trap, in hash_failure_debug() argument 901 trap, vsid, ssize, psize, pte); in hash_failure_debug() 913 unsigned long vsid; in hash_page() local 940 vsid = get_vsid(mm->context.id, ea, ssize); in hash_page() 944 vsid = get_kernel_vsid(ea, mmu_kernel_ssize); in hash_page() 957 DBG_LOW(" mm=%p, mm->pgdir=%p, vsid=%016lx\n", mm, mm->pgd, vsid); in hash_page() 1001 return __hash_page_huge(ea, access, vsid, ptep, trap, local, in hash_page() 1057 rc = __hash_page_64K(ea, access, vsid, ptep, trap, local, ssize); in hash_page() [all …]
|
D | hugetlbpage-hash64.c | 17 int __hash_page_huge(unsigned long ea, unsigned long access, unsigned long vsid, in __hash_page_huge() argument 28 va = hpt_va(ea, vsid, ssize); in __hash_page_huge() 130 hash_failure_debug(ea, access, vsid, trap, ssize, in __hash_page_huge()
|
D | hash_native_64.c | 298 unsigned long vsid, va; in native_hpte_updateboltedpp() local 302 vsid = get_kernel_vsid(ea, ssize); in native_hpte_updateboltedpp() 303 va = hpt_va(ea, vsid, ssize); in native_hpte_updateboltedpp() 387 unsigned long vpi, vsid, pteg; in hpte_decode() local 397 vsid = avpn >> 40; in hpte_decode() 398 vpi = (vsid ^ (vsid << 25) ^ pteg) & htab_hash_mask; in hpte_decode()
|
D | slb.c | 69 get_slb_shadow()->save_area[entry].vsid = mk_vsid_data(ea, ssize, flags); in slb_shadow_update() 115 ksp_vsid_data = get_slb_shadow()->save_area[2].vsid; in __slb_flush_and_rebolt()
|
/arch/powerpc/include/asm/ |
D | mmu-hash64.h | 221 static inline unsigned long hpt_va(unsigned long ea, unsigned long vsid, in hpt_va() argument 225 return (vsid << 28) | (ea & 0xfffffffUL); in hpt_va() 226 return (vsid << 40) | (ea & 0xffffffffffUL); in hpt_va() 236 unsigned long hash, vsid; in hpt_hash() local 241 vsid = va >> 40; in hpt_hash() 242 hash = vsid ^ (vsid << 25) ^ ((va & 0xffffffffffUL) >> shift); in hpt_hash() 248 unsigned long vsid, pte_t *ptep, unsigned long trap, 251 unsigned long vsid, pte_t *ptep, unsigned long trap, 256 int __hash_page_huge(unsigned long ea, unsigned long access, unsigned long vsid, 260 unsigned long vsid, unsigned long trap,
|
D | mmu-hash32.h | 65 unsigned long vsid:24; /* Virtual segment identifier */ member
|
D | kvm_book3s_asm.h | 118 u64 vsid; member
|
D | lppaca.h | 180 u64 vsid; member
|
D | kvm_host.h | 299 int (*esid_to_vsid)(struct kvm_vcpu *vcpu, ulong esid, u64 *vsid); 306 u64 vsid; member
|
D | kvm_book3s.h | 74 u64 vsid; member
|
/arch/microblaze/include/asm/ |
D | mmu.h | 26 unsigned long vsid:24; /* Virtual segment identifier */ member 54 unsigned long vsid:24; /* Virtual Segment Identifier */ member
|
/arch/powerpc/platforms/pseries/ |
D | lpar.c | 332 unsigned long lpar_rc, slot, vsid, va, flags; in pSeries_lpar_hpte_updateboltedpp() local 334 vsid = get_kernel_vsid(ea, ssize); in pSeries_lpar_hpte_updateboltedpp() 335 va = hpt_va(ea, vsid, ssize); in pSeries_lpar_hpte_updateboltedpp() 367 unsigned long slot, vsid, va; in pSeries_lpar_hpte_removebolted() local 369 vsid = get_kernel_vsid(ea, ssize); in pSeries_lpar_hpte_removebolted() 370 va = hpt_va(ea, vsid, ssize); in pSeries_lpar_hpte_removebolted()
|
/arch/powerpc/platforms/cell/ |
D | spu_base.c | 80 u64 esid, vsid; member 157 __func__, slbe, slb->vsid, slb->esid); in spu_load_slb() 163 out_be64(&priv2->slb_vsid_RW, slb->vsid); in spu_load_slb() 185 slb.vsid = (get_vsid(mm->context.id, ea, MMU_SEGSIZE_256M) in __spu_trap_data_seg() 193 slb.vsid = (get_kernel_vsid(ea, MMU_SEGSIZE_256M) in __spu_trap_data_seg() 198 slb.vsid = (get_kernel_vsid(ea, MMU_SEGSIZE_256M) in __spu_trap_data_seg() 208 slb.vsid |= mmu_psize_defs[psize].sllp; in __spu_trap_data_seg() 266 slb->vsid = (get_kernel_vsid(ea, MMU_SEGSIZE_256M) << SLB_VSID_SHIFT) | in __spu_kernel_slb()
|
D | beat_htab.c | 258 unsigned long lpar_rc, slot, vsid, va; in beat_lpar_hpte_updateboltedpp() local 261 vsid = get_kernel_vsid(ea, MMU_SEGSIZE_256M); in beat_lpar_hpte_updateboltedpp() 262 va = (vsid << 28) | (ea & 0x0fffffff); in beat_lpar_hpte_updateboltedpp()
|
/arch/powerpc/xmon/ |
D | xmon.c | 2596 unsigned long esid,vsid,valid; in dump_slb() local 2603 asm volatile("slbmfev %0,%1" : "=r" (vsid) : "r" (i)); in dump_slb() 2605 if (valid | esid | vsid) { in dump_slb() 2606 printf("%02d %016lx %016lx", i, esid, vsid); in dump_slb() 2608 llp = vsid & SLB_VSID_LLP; in dump_slb() 2609 if (vsid & SLB_VSID_B_1T) { in dump_slb() 2612 (vsid & ~SLB_VSID_B) >> SLB_VSID_SHIFT_1T, in dump_slb() 2617 (vsid & ~SLB_VSID_B) >> SLB_VSID_SHIFT, in dump_slb()
|
/arch/powerpc/kernel/ |
D | asm-offsets.c | 188 offsetof(struct slb_shadow, save_area[SLB_NUM_BOLTED - 1].vsid)); in main()
|