/arch/powerpc/mm/ |
D | hash64_64k.c | 57 unsigned long old_pte, new_pte, subpg_pte; in __hash_page_4K() local 79 new_pte = old_pte | H_PAGE_BUSY | _PAGE_ACCESSED | H_PAGE_COMBO; in __hash_page_4K() 81 new_pte |= _PAGE_DIRTY; in __hash_page_4K() 82 } while (!pte_xchg(ptep, __pte(old_pte), __pte(new_pte))); in __hash_page_4K() 87 subpg_pte = new_pte & ~subpg_prot; in __hash_page_4K() 120 new_pte &= ~(H_PAGE_HASHPTE | H_PAGE_F_GIX | H_PAGE_F_SECOND); in __hash_page_4K() 146 *ptep = __pte(new_pte & ~H_PAGE_BUSY); in __hash_page_4K() 210 new_pte = mark_subptegroup_valid(new_pte, subpg_index); in __hash_page_4K() 211 new_pte |= H_PAGE_HASHPTE; in __hash_page_4K() 216 *ptep = __pte(new_pte & ~H_PAGE_BUSY); in __hash_page_4K() [all …]
|
D | hugetlbpage-hash64.c | 27 unsigned long old_pte, new_pte; in __hash_page_huge() local 59 new_pte = old_pte | H_PAGE_BUSY | _PAGE_ACCESSED; in __hash_page_huge() 61 new_pte |= _PAGE_DIRTY; in __hash_page_huge() 62 } while(!pte_xchg(ptep, __pte(old_pte), __pte(new_pte))); in __hash_page_huge() 64 rflags = htab_convert_pte_flags(new_pte); in __hash_page_huge() 94 new_pte = (new_pte & ~_PAGE_HPTEFLAGS) | H_PAGE_HASHPTE; in __hash_page_huge() 110 new_pte |= (slot << H_PAGE_F_GIX_SHIFT) & in __hash_page_huge() 117 *ptep = __pte(new_pte & ~H_PAGE_BUSY); in __hash_page_huge()
|
D | hash64_4k.c | 25 unsigned long old_pte, new_pte; in __hash_page_4K() local 47 new_pte = old_pte | H_PAGE_BUSY | _PAGE_ACCESSED; in __hash_page_4K() 49 new_pte |= _PAGE_DIRTY; in __hash_page_4K() 50 } while (!pte_xchg(ptep, __pte(old_pte), __pte(new_pte))); in __hash_page_4K() 56 rflags = htab_convert_pte_flags(new_pte); in __hash_page_4K() 120 new_pte = (new_pte & ~_PAGE_HPTEFLAGS) | H_PAGE_HASHPTE; in __hash_page_4K() 121 new_pte |= (slot << H_PAGE_F_GIX_SHIFT) & in __hash_page_4K() 124 *ptep = __pte(new_pte & ~H_PAGE_BUSY); in __hash_page_4K()
|
/arch/powerpc/include/asm/book3s/64/ |
D | radix.h | 131 unsigned long old_pte, new_pte; in __radix_pte_update() local 136 new_pte = (old_pte | set) & ~clr; in __radix_pte_update() 138 } while (!pte_xchg(ptep, __pte(old_pte), __pte(new_pte))); in __radix_pte_update() 154 unsigned long new_pte; in radix__pte_update() local 160 new_pte = (old_pte | set) & ~clr; in radix__pte_update() 162 if (new_pte) in radix__pte_update() 163 __radix_pte_update(ptep, 0, new_pte); in radix__pte_update() 207 unsigned long old_pte, new_pte; in radix__ptep_set_access_flags() local 213 new_pte = old_pte | set; in radix__ptep_set_access_flags() 215 __radix_pte_update(ptep, 0, new_pte); in radix__ptep_set_access_flags()
|
/arch/unicore32/mm/ |
D | pgd.c | 31 pte_t *new_pte, *init_pte; in get_pgd_slow() local 57 new_pte = pte_alloc_map(mm, new_pmd, 0); in get_pgd_slow() 58 if (!new_pte) in get_pgd_slow() 63 set_pte(new_pte, *init_pte); in get_pgd_slow() 65 pte_unmap(new_pte); in get_pgd_slow()
|
/arch/arm/mm/ |
D | pgd.c | 38 pte_t *new_pte, *init_pte; in pgd_alloc() local 83 new_pte = pte_alloc_map(mm, new_pmd, 0); in pgd_alloc() 84 if (!new_pte) in pgd_alloc() 100 set_pte_ext(new_pte + 0, init_pte[0], 0); in pgd_alloc() 101 set_pte_ext(new_pte + 1, init_pte[1], 0); in pgd_alloc() 103 pte_unmap(new_pte); in pgd_alloc()
|
/arch/powerpc/include/asm/ |
D | kvm_book3s_64.h | 238 pte_t old_pte, new_pte = __pte(0); in kvmppc_read_update_linux_pte() local 256 new_pte = pte_mkyoung(old_pte); in kvmppc_read_update_linux_pte() 258 new_pte = pte_mkdirty(new_pte); in kvmppc_read_update_linux_pte() 260 if (pte_xchg(ptep, old_pte, new_pte)) in kvmppc_read_update_linux_pte() 263 return new_pte; in kvmppc_read_update_linux_pte()
|
/arch/arm/include/asm/ |
D | kvm_mmu.h | 75 static inline void kvm_set_pte(pte_t *pte, pte_t new_pte) in kvm_set_pte() argument 77 *pte = new_pte; in kvm_set_pte()
|
/arch/x86/include/asm/ |
D | pgtable_64.h | 56 void set_pte_vaddr_p4d(p4d_t *p4d_page, unsigned long vaddr, pte_t new_pte); 57 void set_pte_vaddr_pud(pud_t *pud_page, unsigned long vaddr, pte_t new_pte);
|
/arch/x86/mm/ |
D | init_64.c | 248 static void __set_pte_vaddr(pud_t *pud, unsigned long vaddr, pte_t new_pte) in __set_pte_vaddr() argument 253 set_pte(pte, new_pte); in __set_pte_vaddr() 262 void set_pte_vaddr_p4d(p4d_t *p4d_page, unsigned long vaddr, pte_t new_pte) in set_pte_vaddr_p4d() argument 267 __set_pte_vaddr(pud, vaddr, new_pte); in set_pte_vaddr_p4d() 270 void set_pte_vaddr_pud(pud_t *pud_page, unsigned long vaddr, pte_t new_pte) in set_pte_vaddr_pud() argument 274 __set_pte_vaddr(pud, vaddr, new_pte); in set_pte_vaddr_pud()
|
D | pageattr.c | 522 pte_t new_pte, old_pte, *tmp; in try_preserve_large_page() local 641 new_pte = pfn_pte(old_pfn, new_prot); in try_preserve_large_page() 642 __set_pmd_pte(kpte, address, new_pte); in try_preserve_large_page() 1230 pte_t new_pte; in __change_page_attr() local 1256 new_pte = pfn_pte(pfn, canon_pgprot(new_prot)); in __change_page_attr() 1261 if (pte_val(old_pte) != pte_val(new_pte)) { in __change_page_attr() 1262 set_pte_atomic(kpte, new_pte); in __change_page_attr()
|
/arch/mips/kvm/ |
D | mmu.c | 161 pte_t *new_pte; in kvm_mips_walk_pgd() local 165 new_pte = mmu_memory_cache_alloc(cache); in kvm_mips_walk_pgd() 166 clear_page(new_pte); in kvm_mips_walk_pgd() 167 pmd_populate_kernel(NULL, pmd, new_pte); in kvm_mips_walk_pgd()
|
/arch/x86/kvm/ |
D | paging_tmpl.h | 136 pt_element_t orig_pte, pt_element_t new_pte) in FNAME() 149 ret = CMPXCHG(&table[index], orig_pte, new_pte); in FNAME()
|