Searched refs:need_flush (Results 1 – 6 of 6) sorted by relevance
214 u16 *new_asid, bool *need_flush) in choose_new_asid() argument220 *need_flush = true; in choose_new_asid()233 *need_flush = (this_cpu_read(cpu_tlbstate.ctxs[asid].tlb_gen) < in choose_new_asid()247 *need_flush = true; in choose_new_asid()276 static void load_new_mm_cr3(pgd_t *pgdir, u16 new_asid, bool need_flush) in load_new_mm_cr3() argument280 if (need_flush) { in load_new_mm_cr3()496 bool need_flush; in switch_mm_irqs_off() local591 need_flush = true; in switch_mm_irqs_off()617 choose_new_asid(next, next_tlb_gen, &new_asid, &need_flush); in switch_mm_irqs_off()624 if (need_flush) { in switch_mm_irqs_off()
368 unsigned int subregions, bool need_flush) in mpu_setup_region() argument387 if (need_flush) in mpu_setup_region()441 bool need_flush = region == PMSAv7_RAM_REGION; in pmsav7_setup() local448 xip[i].subreg, need_flush); in pmsav7_setup()
89 static bool need_flush; /* global flush state. set for each gart wrap */ variable106 need_flush = true; in alloc_iommu()115 need_flush = true; in alloc_iommu()119 need_flush = true; in alloc_iommu()144 if (need_flush) { in flush_gart()146 need_flush = false; in flush_gart()
19 static inline bool need_flush(struct iommu_map_table *iommu) in need_flush() function206 (n < pool->hint || need_flush(iommu))) { in iommu_tbl_range_alloc()
960 bool need_flush = false; in stage2_unmap_walker() local976 need_flush = !stage2_has_fwb(pgt); in stage2_unmap_walker()986 if (need_flush && mm_ops->dcache_clean_inval_poc) in stage2_unmap_walker()
1453 int need_flush = 0; in kvm_set_pte_rmapp() local1465 need_flush = 1; in kvm_set_pte_rmapp()1479 if (need_flush && kvm_available_flush_tlb_with_range()) { in kvm_set_pte_rmapp()1484 return need_flush; in kvm_set_pte_rmapp()