Lines Matching refs:vma
87 update_mmu_cache(struct vm_area_struct *vma, unsigned long address, pte_t *ptep) in update_mmu_cache() argument
304 __flush_cache_page(struct vm_area_struct *vma, unsigned long vmaddr, in __flush_cache_page() argument
309 if (vma->vm_flags & VM_EXEC) in __flush_cache_page()
315 __purge_cache_page(struct vm_area_struct *vma, unsigned long vmaddr, in __purge_cache_page() argument
320 if (vma->vm_flags & VM_EXEC) in __purge_cache_page()
524 struct vm_area_struct *vma; in mm_total_size() local
527 for (vma = mm->mmap; vma; vma = vma->vm_next) in mm_total_size()
528 usize += vma->vm_end - vma->vm_start; in mm_total_size()
549 struct vm_area_struct *vma; in flush_cache_mm() local
563 for (vma = mm->mmap; vma; vma = vma->vm_next) { in flush_cache_mm()
564 flush_user_dcache_range_asm(vma->vm_start, vma->vm_end); in flush_cache_mm()
565 if (vma->vm_flags & VM_EXEC) in flush_cache_mm()
566 flush_user_icache_range_asm(vma->vm_start, vma->vm_end); in flush_cache_mm()
567 flush_tlb_range(vma, vma->vm_start, vma->vm_end); in flush_cache_mm()
573 for (vma = mm->mmap; vma; vma = vma->vm_next) { in flush_cache_mm()
576 for (addr = vma->vm_start; addr < vma->vm_end; in flush_cache_mm()
586 flush_tlb_page(vma, addr); in flush_cache_mm()
587 __flush_cache_page(vma, addr, PFN_PHYS(pfn)); in flush_cache_mm()
589 __purge_cache_page(vma, addr, PFN_PHYS(pfn)); in flush_cache_mm()
595 void flush_cache_range(struct vm_area_struct *vma, in flush_cache_range() argument
603 if (vma->vm_mm->context) in flush_cache_range()
604 flush_tlb_range(vma, start, end); in flush_cache_range()
609 if (vma->vm_mm->context == mfsp(3)) { in flush_cache_range()
611 if (vma->vm_flags & VM_EXEC) in flush_cache_range()
613 flush_tlb_range(vma, start, end); in flush_cache_range()
617 pgd = vma->vm_mm->pgd; in flush_cache_range()
618 for (addr = vma->vm_start; addr < vma->vm_end; addr += PAGE_SIZE) { in flush_cache_range()
625 if (unlikely(vma->vm_mm->context)) { in flush_cache_range()
626 flush_tlb_page(vma, addr); in flush_cache_range()
627 __flush_cache_page(vma, addr, PFN_PHYS(pfn)); in flush_cache_range()
629 __purge_cache_page(vma, addr, PFN_PHYS(pfn)); in flush_cache_range()
636 flush_cache_page(struct vm_area_struct *vma, unsigned long vmaddr, unsigned long pfn) in flush_cache_page() argument
639 if (likely(vma->vm_mm->context)) { in flush_cache_page()
640 flush_tlb_page(vma, vmaddr); in flush_cache_page()
641 __flush_cache_page(vma, vmaddr, PFN_PHYS(pfn)); in flush_cache_page()
643 __purge_cache_page(vma, vmaddr, PFN_PHYS(pfn)); in flush_cache_page()