Home
last modified time | relevance | path

Searched refs:new_pte (Results 1 – 13 of 13) sorted by relevance

/arch/powerpc/mm/
Dhash64_64k.c57 unsigned long old_pte, new_pte, subpg_pte; in __hash_page_4K() local
79 new_pte = old_pte | H_PAGE_BUSY | _PAGE_ACCESSED | H_PAGE_COMBO; in __hash_page_4K()
81 new_pte |= _PAGE_DIRTY; in __hash_page_4K()
82 } while (!pte_xchg(ptep, __pte(old_pte), __pte(new_pte))); in __hash_page_4K()
87 subpg_pte = new_pte & ~subpg_prot; in __hash_page_4K()
120 new_pte &= ~(H_PAGE_HASHPTE | H_PAGE_F_GIX | H_PAGE_F_SECOND); in __hash_page_4K()
146 *ptep = __pte(new_pte & ~H_PAGE_BUSY); in __hash_page_4K()
210 new_pte = mark_subptegroup_valid(new_pte, subpg_index); in __hash_page_4K()
211 new_pte |= H_PAGE_HASHPTE; in __hash_page_4K()
216 *ptep = __pte(new_pte & ~H_PAGE_BUSY); in __hash_page_4K()
[all …]
Dhugetlbpage-hash64.c27 unsigned long old_pte, new_pte; in __hash_page_huge() local
59 new_pte = old_pte | H_PAGE_BUSY | _PAGE_ACCESSED; in __hash_page_huge()
61 new_pte |= _PAGE_DIRTY; in __hash_page_huge()
62 } while(!pte_xchg(ptep, __pte(old_pte), __pte(new_pte))); in __hash_page_huge()
64 rflags = htab_convert_pte_flags(new_pte); in __hash_page_huge()
94 new_pte = (new_pte & ~_PAGE_HPTEFLAGS) | H_PAGE_HASHPTE; in __hash_page_huge()
110 new_pte |= (slot << H_PAGE_F_GIX_SHIFT) & in __hash_page_huge()
117 *ptep = __pte(new_pte & ~H_PAGE_BUSY); in __hash_page_huge()
Dhash64_4k.c25 unsigned long old_pte, new_pte; in __hash_page_4K() local
47 new_pte = old_pte | H_PAGE_BUSY | _PAGE_ACCESSED; in __hash_page_4K()
49 new_pte |= _PAGE_DIRTY; in __hash_page_4K()
50 } while (!pte_xchg(ptep, __pte(old_pte), __pte(new_pte))); in __hash_page_4K()
56 rflags = htab_convert_pte_flags(new_pte); in __hash_page_4K()
120 new_pte = (new_pte & ~_PAGE_HPTEFLAGS) | H_PAGE_HASHPTE; in __hash_page_4K()
121 new_pte |= (slot << H_PAGE_F_GIX_SHIFT) & in __hash_page_4K()
124 *ptep = __pte(new_pte & ~H_PAGE_BUSY); in __hash_page_4K()
/arch/powerpc/include/asm/book3s/64/
Dradix.h131 unsigned long old_pte, new_pte; in __radix_pte_update() local
136 new_pte = (old_pte | set) & ~clr; in __radix_pte_update()
138 } while (!pte_xchg(ptep, __pte(old_pte), __pte(new_pte))); in __radix_pte_update()
154 unsigned long new_pte; in radix__pte_update() local
160 new_pte = (old_pte | set) & ~clr; in radix__pte_update()
162 if (new_pte) in radix__pte_update()
163 __radix_pte_update(ptep, 0, new_pte); in radix__pte_update()
207 unsigned long old_pte, new_pte; in radix__ptep_set_access_flags() local
213 new_pte = old_pte | set; in radix__ptep_set_access_flags()
215 __radix_pte_update(ptep, 0, new_pte); in radix__ptep_set_access_flags()
/arch/unicore32/mm/
Dpgd.c31 pte_t *new_pte, *init_pte; in get_pgd_slow() local
57 new_pte = pte_alloc_map(mm, new_pmd, 0); in get_pgd_slow()
58 if (!new_pte) in get_pgd_slow()
63 set_pte(new_pte, *init_pte); in get_pgd_slow()
65 pte_unmap(new_pte); in get_pgd_slow()
/arch/arm/mm/
Dpgd.c38 pte_t *new_pte, *init_pte; in pgd_alloc() local
83 new_pte = pte_alloc_map(mm, new_pmd, 0); in pgd_alloc()
84 if (!new_pte) in pgd_alloc()
100 set_pte_ext(new_pte + 0, init_pte[0], 0); in pgd_alloc()
101 set_pte_ext(new_pte + 1, init_pte[1], 0); in pgd_alloc()
103 pte_unmap(new_pte); in pgd_alloc()
/arch/powerpc/include/asm/
Dkvm_book3s_64.h238 pte_t old_pte, new_pte = __pte(0); in kvmppc_read_update_linux_pte() local
256 new_pte = pte_mkyoung(old_pte); in kvmppc_read_update_linux_pte()
258 new_pte = pte_mkdirty(new_pte); in kvmppc_read_update_linux_pte()
260 if (pte_xchg(ptep, old_pte, new_pte)) in kvmppc_read_update_linux_pte()
263 return new_pte; in kvmppc_read_update_linux_pte()
/arch/arm/include/asm/
Dkvm_mmu.h75 static inline void kvm_set_pte(pte_t *pte, pte_t new_pte) in kvm_set_pte() argument
77 *pte = new_pte; in kvm_set_pte()
/arch/x86/include/asm/
Dpgtable_64.h56 void set_pte_vaddr_p4d(p4d_t *p4d_page, unsigned long vaddr, pte_t new_pte);
57 void set_pte_vaddr_pud(pud_t *pud_page, unsigned long vaddr, pte_t new_pte);
/arch/x86/mm/
Dinit_64.c248 static void __set_pte_vaddr(pud_t *pud, unsigned long vaddr, pte_t new_pte) in __set_pte_vaddr() argument
253 set_pte(pte, new_pte); in __set_pte_vaddr()
262 void set_pte_vaddr_p4d(p4d_t *p4d_page, unsigned long vaddr, pte_t new_pte) in set_pte_vaddr_p4d() argument
267 __set_pte_vaddr(pud, vaddr, new_pte); in set_pte_vaddr_p4d()
270 void set_pte_vaddr_pud(pud_t *pud_page, unsigned long vaddr, pte_t new_pte) in set_pte_vaddr_pud() argument
274 __set_pte_vaddr(pud, vaddr, new_pte); in set_pte_vaddr_pud()
Dpageattr.c522 pte_t new_pte, old_pte, *tmp; in try_preserve_large_page() local
641 new_pte = pfn_pte(old_pfn, new_prot); in try_preserve_large_page()
642 __set_pmd_pte(kpte, address, new_pte); in try_preserve_large_page()
1230 pte_t new_pte; in __change_page_attr() local
1256 new_pte = pfn_pte(pfn, canon_pgprot(new_prot)); in __change_page_attr()
1261 if (pte_val(old_pte) != pte_val(new_pte)) { in __change_page_attr()
1262 set_pte_atomic(kpte, new_pte); in __change_page_attr()
/arch/mips/kvm/
Dmmu.c161 pte_t *new_pte; in kvm_mips_walk_pgd() local
165 new_pte = mmu_memory_cache_alloc(cache); in kvm_mips_walk_pgd()
166 clear_page(new_pte); in kvm_mips_walk_pgd()
167 pmd_populate_kernel(NULL, pmd, new_pte); in kvm_mips_walk_pgd()
/arch/x86/kvm/
Dpaging_tmpl.h136 pt_element_t orig_pte, pt_element_t new_pte) in FNAME()
149 ret = CMPXCHG(&table[index], orig_pte, new_pte); in FNAME()