/arch/mips/include/asm/ |
D | mmu_context.h | 76 #define cpu_context(cpu, mm) ((mm)->context[cpu]) macro 77 #define cpu_asid(cpu, mm) (cpu_context((cpu), (mm)) & ASID_MASK) 105 cpu_context(cpu, mm) = asid_cache(cpu) = asid; in get_new_mmu_context() 124 cpu_context(i, mm) = 0; in init_new_context() 145 if ((cpu_context(cpu, next) ^ asid_cache(cpu)) & ASID_VERSION_MASK) in switch_mm() 168 | (cpu_context(cpu, next) & ASID_MASK)); in switch_mm() 172 write_c0_entryhi(cpu_context(cpu, next)); in switch_mm() 228 (cpu_context(cpu, next) & ASID_MASK)); in activate_mm() 232 write_c0_entryhi(cpu_context(cpu, next)); in activate_mm() 282 cpu_context(cpu, mm) = 0; in drop_mmu_context() [all …]
|
/arch/avr32/include/asm/ |
D | processor.h | 102 struct cpu_context { struct 118 struct cpu_context cpu_context; argument 124 .cpu_context = { \ 153 #define thread_saved_pc(tsk) ((tsk)->thread.cpu_context.pc) 164 #define KSTK_EIP(tsk) ((tsk)->thread.cpu_context.pc) 165 #define KSTK_ESP(tsk) ((tsk)->thread.cpu_context.ksp)
|
D | system.h | 52 struct cpu_context; 55 struct cpu_context *, 56 struct cpu_context *); 59 last = __switch_to(prev, &prev->thread.cpu_context + 1, \ 60 &next->thread.cpu_context); \
|
/arch/mips/mm/ |
D | tlb-r3k.c | 70 if (cpu_context(cpu, mm) != 0) { in local_flush_tlb_mm() 72 printk("[tlbmm<%lu>]", (unsigned long)cpu_context(cpu, mm)); in local_flush_tlb_mm() 84 if (cpu_context(cpu, mm) != 0) { in local_flush_tlb_range() 90 cpu_context(cpu, mm) & ASID_MASK, start, end); in local_flush_tlb_range() 96 int newpid = cpu_context(cpu, mm) & ASID_MASK; in local_flush_tlb_range() 163 if (!vma || cpu_context(cpu, vma->vm_mm) != 0) { in local_flush_tlb_page() 168 printk("[tlbpage<%lu,0x%08lx>]", cpu_context(cpu, vma->vm_mm), page); in local_flush_tlb_page() 170 newpid = cpu_context(cpu, vma->vm_mm) & ASID_MASK; in local_flush_tlb_page() 204 if ((pid != (cpu_context(cpu, vma->vm_mm) & ASID_MASK)) || (cpu_context(cpu, vma->vm_mm) == 0)) { in __update_tlb() 206 (cpu_context(cpu, vma->vm_mm)), pid); in __update_tlb()
|
D | tlb-r8k.c | 58 if (cpu_context(cpu, mm) != 0) in local_flush_tlb_mm() 70 if (!cpu_context(cpu, mm)) in local_flush_tlb_range() 157 if (!cpu_context(cpu, vma->vm_mm)) in local_flush_tlb_page()
|
D | c-tx39.c | 154 if (cpu_context(smp_processor_id(), mm) != 0) in tx39_flush_cache_mm() 163 if (!(cpu_context(smp_processor_id(), vma->vm_mm))) in tx39_flush_cache_range() 182 if (cpu_context(smp_processor_id(), mm) == 0) in tx39_flush_cache_page()
|
D | c-r3k.c | 249 cpu_context(smp_processor_id(), mm), addr); in r3k_flush_cache_page() 252 if (cpu_context(smp_processor_id(), mm) == 0) in r3k_flush_cache_page()
|
D | tlb-r4k.c | 106 if (cpu_context(cpu, mm) != 0) { in local_flush_tlb_mm() 119 if (cpu_context(cpu, mm) != 0) { in local_flush_tlb_range() 207 if (cpu_context(cpu, vma->vm_mm) != 0) { in local_flush_tlb_page()
|
D | c-r4k.c | 362 if (cpu_context(i, mm)) in has_valid_asid() 367 return cpu_context(smp_processor_id(), mm); in has_valid_asid() 501 if (cpu_context(cpu, mm) != 0) in local_r4k_flush_cache_page()
|
/arch/sh/include/asm/ |
D | mmu_context.h | 33 #define cpu_context(cpu, mm) ((mm)->context.id[cpu]) macro 36 (cpu_context((cpu), (mm)) & MMU_CONTEXT_ASID_MASK) 57 if (((cpu_context(cpu, mm) ^ asid) & MMU_CONTEXT_VERSION_MASK) == 0) in get_mmu_context() 85 cpu_context(cpu, mm) = asid_cache(cpu) = asid; in get_mmu_context() 98 cpu_context(i, mm) = NO_CONTEXT; in init_new_context()
|
/arch/arm/kernel/ |
D | xscale-cp0.c | 40 thread->cpu_context.extra[0] = 0; in dsp_do() 41 thread->cpu_context.extra[1] = 0; in dsp_do() 45 dsp_save_state(current_thread_info()->cpu_context.extra); in dsp_do() 46 dsp_load_state(thread->cpu_context.extra); in dsp_do()
|
D | process.c | 386 memset(&thread->cpu_context, 0, sizeof(struct cpu_context_save)); in copy_thread() 387 thread->cpu_context.sp = (unsigned long)childregs; in copy_thread() 388 thread->cpu_context.pc = (unsigned long)ret_from_fork; in copy_thread()
|
D | asm-offsets.c | 50 DEFINE(TI_CPU_SAVE, offsetof(struct thread_info, cpu_context)); in main()
|
/arch/sh/mm/ |
D | tlbflush_32.c | 19 if (vma->vm_mm && cpu_context(cpu, vma->vm_mm) != NO_CONTEXT) { in local_flush_tlb_page() 45 if (cpu_context(cpu, mm) != NO_CONTEXT) { in local_flush_tlb_range() 52 cpu_context(cpu, mm) = NO_CONTEXT; in local_flush_tlb_range() 112 if (cpu_context(cpu, mm) != NO_CONTEXT) { in local_flush_tlb_mm() 116 cpu_context(cpu, mm) = NO_CONTEXT; in local_flush_tlb_mm()
|
D | tlbflush_64.c | 397 if (cpu_context(cpu, mm) == NO_CONTEXT) in local_flush_tlb_range() 441 if (cpu_context(cpu, mm) == NO_CONTEXT) in local_flush_tlb_mm() 446 cpu_context(cpu, mm) = NO_CONTEXT; in local_flush_tlb_mm()
|
/arch/mips/kernel/ |
D | smp.c | 301 if (cpu_context(cpu, mm)) in flush_tlb_mm() 302 cpu_context(cpu, mm) = 0; in flush_tlb_mm() 341 if (cpu_context(cpu, mm)) in flush_tlb_range() 342 cpu_context(cpu, mm) = 0; in flush_tlb_range() 388 if (cpu_context(cpu, vma->vm_mm)) in flush_tlb_page() 389 cpu_context(cpu, vma->vm_mm) = 0; in flush_tlb_page()
|
D | smtc.c | 1335 cpu_context(i, mm) = asid_cache(i) = asid; in smtc_get_new_mmu_context()
|
/arch/arm/include/asm/ |
D | thread_info.h | 58 struct cpu_context_save cpu_context; /* cpu context */ member 101 ((unsigned long)(task_thread_info(tsk)->cpu_context.pc)) 103 ((unsigned long)(task_thread_info(tsk)->cpu_context.fp))
|
/arch/avr32/kernel/ |
D | process.c | 175 fp = tsk->thread.cpu_context.r7; in show_trace_log_lvl() 232 sp = tsk->thread.cpu_context.ksp; in show_stack_log_lvl() 351 p->thread.cpu_context.sr = MODE_SUPERVISOR | SR_GM; in copy_thread() 352 p->thread.cpu_context.ksp = (unsigned long)childregs; in copy_thread() 353 p->thread.cpu_context.pc = (unsigned long)ret_from_fork; in copy_thread() 428 unsigned long fp = p->thread.cpu_context.r7; in get_wchan() 441 unsigned long sp = p->thread.cpu_context.ksp + 16; in get_wchan()
|
/arch/sh/kernel/ |
D | smp.c | 267 cpu_context(i, mm) = 0; in flush_tlb_mm() 304 cpu_context(i, mm) = 0; in flush_tlb_range() 347 cpu_context(i, vma->vm_mm) = 0; in flush_tlb_page()
|