/arch/x86/include/asm/ |
D | tlbflush.h | 98 if (mm == current->active_mm) in flush_tlb_mm() 105 if (vma->vm_mm == current->active_mm) in flush_tlb_page() 112 if (vma->vm_mm == current->active_mm) in flush_tlb_range() 153 struct mm_struct *active_mm; member
|
D | mmu_context_64.h | 23 write_pda(active_mm, next); in switch_mm() 34 if (read_pda(active_mm) != next) in switch_mm()
|
D | mmu_context_32.h | 23 x86_write_percpu(cpu_tlbstate.active_mm, next); in switch_mm() 39 BUG_ON(x86_read_percpu(cpu_tlbstate.active_mm) != next); in switch_mm()
|
/arch/x86/kernel/ |
D | tlb_32.c | 38 cpu_clear(cpu, x86_read_percpu(cpu_tlbstate.active_mm)->cpu_vm_mask); in leave_mm() 106 if (flush_mm == x86_read_percpu(cpu_tlbstate.active_mm)) { in smp_invalidate_interrupt() 200 if (current->active_mm == mm) { in flush_tlb_mm() 221 if (current->active_mm == mm) { in flush_tlb_page() 254 per_cpu(cpu_tlbstate, cpu).active_mm = &init_mm; in reset_lazy_tlbstate()
|
D | tlb_64.c | 67 cpu_clear(cpu, read_pda(active_mm)->cpu_vm_mask); in leave_mm() 145 if (f->flush_mm == read_pda(active_mm)) { in smp_invalidate_interrupt() 238 if (current->active_mm == mm) { in flush_tlb_mm() 259 if (current->active_mm == mm) { in flush_tlb_page()
|
D | ldt.c | 26 if (current->active_mm == current_mm) in flush_ldt() 27 load_LDT(¤t->active_mm->context); in flush_ldt() 131 if (mm == current->active_mm) in destroy_context()
|
/arch/alpha/include/asm/ |
D | tlbflush.h | 83 flush_tlb_current(current->active_mm); in flush_tlb() 108 if (mm == current->active_mm) in flush_tlb_mm() 120 if (mm == current->active_mm) in flush_tlb_page()
|
D | cacheflush.h | 52 if (current->active_mm == mm) in flush_icache_user_range()
|
/arch/sparc/include/asm/ |
D | mmu_context_32.h | 38 #define activate_mm(active_mm, mm) switch_mm((active_mm), (mm), NULL) argument
|
/arch/ia64/mm/ |
D | tlb.c | 230 struct mm_struct *active_mm = current->active_mm; in ia64_global_tlb_purge() local 234 if (mm != active_mm) { in ia64_global_tlb_purge() 236 if (mm && active_mm) { in ia64_global_tlb_purge() 259 if (mm != active_mm) { in ia64_global_tlb_purge() 260 activate_context(active_mm); in ia64_global_tlb_purge() 296 if (mm != current->active_mm) { in flush_tlb_range() 312 if (mm != current->active_mm || cpus_weight(mm->cpu_vm_mask) != 1) { in flush_tlb_range()
|
/arch/ia64/include/asm/ |
D | tlbflush.h | 41 if (mm == current->active_mm) in local_finish_flush_tlb_mm() 80 if (vma->vm_mm == current->active_mm) in flush_tlb_page()
|
/arch/m68k/include/asm/ |
D | tlbflush_mm.h | 69 if (mm == current->active_mm) in flush_tlb_mm() 75 if (vma->vm_mm == current->active_mm) { in flush_tlb_page() 86 if (vma->vm_mm == current->active_mm) in flush_tlb_range()
|
/arch/um/sys-i386/ |
D | ldt.c | 36 if (!current->active_mm || current->active_mm == &init_mm || in write_ldt_entry() 37 mm_idp != ¤t->active_mm->context.id) in write_ldt_entry() 82 if (current->active_mm && current->active_mm != &init_mm && in write_ldt_entry() 83 mm_idp != ¤t->active_mm->context.id) in write_ldt_entry() 84 __switch_mm(¤t->active_mm->context.id); in write_ldt_entry()
|
/arch/alpha/kernel/ |
D | smp.c | 147 current->active_mm = &init_mm; in smp_callin() 167 cpuid, current, current->active_mm)); in smp_callin() 689 if (mm == current->active_mm && !asn_locked()) in ipi_flush_tlb_mm() 700 if (mm == current->active_mm) { in flush_tlb_mm() 735 if (mm == current->active_mm && !asn_locked()) in ipi_flush_tlb_page() 749 if (mm == current->active_mm) { in flush_tlb_page() 788 if (mm == current->active_mm && !asn_locked()) in ipi_flush_icache_page() 805 if (mm == current->active_mm) { in flush_icache_user_range()
|
/arch/powerpc/kernel/ |
D | swsusp.c | 37 switch_mmu_context(NULL, current->active_mm); in restore_processor_state()
|
/arch/avr32/kernel/ |
D | asm-offsets.c | 23 OFFSET(TSK_active_mm, task_struct, active_mm); in foo()
|
/arch/arm/mm/ |
D | tlb-v3.S | 33 act_mm r3 @ get current->active_mm
|
D | tlb-v4.S | 34 act_mm r3 @ get current->active_mm
|
D | tlb-v4wb.S | 34 act_mm r3 @ get current->active_mm
|
/arch/x86/xen/ |
D | mmu.c | 1064 struct mm_struct *active_mm; in drop_other_mm_ref() local 1067 active_mm = read_pda(active_mm); in drop_other_mm_ref() 1069 active_mm = __get_cpu_var(cpu_tlbstate).active_mm; in drop_other_mm_ref() 1072 if (active_mm == mm) in drop_other_mm_ref() 1088 if (current->active_mm == mm) { in xen_drop_mm_ref() 1125 if (current->active_mm == mm) in xen_drop_mm_ref()
|
/arch/parisc/include/asm/ |
D | tlbflush.h | 57 if (mm == current->active_mm) in flush_tlb_mm()
|
/arch/sh/mm/ |
D | tlb-sh3.c | 39 if (vma && current->active_mm != vma->vm_mm) in update_mmu_cache()
|
D | tlb-sh4.c | 26 if (vma && current->active_mm != vma->vm_mm) in update_mmu_cache()
|
/arch/s390/mm/ |
D | pgtable.c | 268 tsk->mm != tsk->active_mm || !hlist_empty(&tsk->mm->ioctx_list)) { in s390_enable_sie() 284 tsk->mm != tsk->active_mm || !hlist_empty(&tsk->mm->ioctx_list)) { in s390_enable_sie() 292 tsk->mm = tsk->active_mm = mm; in s390_enable_sie()
|
/arch/s390/include/asm/ |
D | tlbflush.h | 97 if (atomic_read(&mm->mm_users) <= 1 && mm == current->active_mm) in __tlb_flush_mm_cond()
|