Lines Matching +full:tlb +full:- +full:split
1 // SPDX-License-Identifier: GPL-2.0-or-later
7 * Copyright (C) 1995-1996 Gary Thomas (gdt@linuxppc.org)
27 #include <asm/tlb.h>
32 return current->thread.regs && TRAP(current->thread.regs) == 0x400; in is_exec_fault()
67 /* Server-style MMU handles coherency when hashing if HW exec permission
68 * is supposed per page (currently 64-bit only). If not, then, we always
84 if (!test_bit(PG_arch_1, &pg->flags)) { in set_pte_filter_hash()
86 set_bit(PG_arch_1, &pg->flags); in set_pte_filter_hash()
119 if (test_bit(PG_arch_1, &pg->flags)) in set_pte_filter()
125 set_bit(PG_arch_1, &pg->flags); in set_pte_filter()
154 if (WARN_ON(!(vma->vm_flags & VM_EXEC))) in set_access_flags_filter()
164 if (test_bit(PG_arch_1, &pg->flags)) in set_access_flags_filter()
169 set_bit(PG_arch_1, &pg->flags); in set_access_flags_filter()
183 * tlb flush for this update. in set_pte_at()
187 /* Note: mm->context.id might not yet have been assigned as in set_pte_at()
211 assert_pte_locked(vma->vm_mm, address); in ptep_set_access_flags()
226 * TLB entry. Without this, platforms that don't do a write of the TLB in huge_ptep_set_access_flags()
227 * entry in the TLB miss handler asm will fault ad infinitum. in huge_ptep_set_access_flags()
243 assert_spin_locked(huge_pte_lockptr(h, vma->vm_mm, ptep)); in huge_ptep_set_access_flags()
265 pte_basic_t *entry = &ptep->pte; in set_huge_pte_at()
270 * tlb flush for this update. in set_huge_pte_at()
296 pgd = mm->pgd + pgd_index(addr); in assert_pte_locked()
335 * when we have MSR[EE] = 0 but the paca->irq_soft_mask = IRQS_ENABLED
356 * value don't get updated by a parallel THP split/collapse, in __find_linux_pte()
414 * A hugepage split is captured by this condition, see in __find_linux_pte()