/arch/mips/include/asm/ |
D | switch_to.h | 33 struct task_struct *next, struct thread_info *next_ti); 62 next->thread.emulated_fp = 0; \ 82 #define __sanitize_fcr31(next) \ argument 84 unsigned long fcr31 = mask_fcr31_x(next->thread.fpu.fcr31); \ 88 pc = (void __user *)task_pt_regs(next)->cp0_epc; \ 89 next->thread.fpu.fcr31 &= ~fcr31; \ 90 force_fcr31_sig(fcr31, pc, next); \ 100 #define switch_to(prev, next, last) \ argument 104 if (tsk_used_math(next)) \ 105 __sanitize_fcr31(next); \ [all …]
|
/arch/blackfin/mm/ |
D | sram-alloc.c | 31 struct sram_piece *next; member 78 per_cpu(free_l1_ssram_head, cpu).next = in l1sram_init() 80 if (!per_cpu(free_l1_ssram_head, cpu).next) { in l1sram_init() 85 per_cpu(free_l1_ssram_head, cpu).next->paddr = (void *)get_l1_scratch_start_cpu(cpu) + reserve; in l1sram_init() 86 per_cpu(free_l1_ssram_head, cpu).next->size = L1_SCRATCH_LENGTH - reserve; in l1sram_init() 87 per_cpu(free_l1_ssram_head, cpu).next->pid = 0; in l1sram_init() 88 per_cpu(free_l1_ssram_head, cpu).next->next = NULL; in l1sram_init() 90 per_cpu(used_l1_ssram_head, cpu).next = NULL; in l1sram_init() 106 per_cpu(free_l1_data_A_sram_head, cpu).next = in l1_data_sram_init() 108 if (!per_cpu(free_l1_data_A_sram_head, cpu).next) { in l1_data_sram_init() [all …]
|
/arch/powerpc/include/asm/ |
D | mmu_context.h | 44 struct mm_struct *next); 46 struct mm_struct *next, in switch_mmu_context() argument 50 return radix__switch_mmu_context(prev, next); in switch_mmu_context() 51 return switch_slb(tsk, next); in switch_mmu_context() 58 extern void switch_mmu_context(struct mm_struct *prev, struct mm_struct *next, 65 extern void switch_cop(struct mm_struct *next); 74 struct mm_struct *next, in switch_mm_irqs_off() argument 78 if (!cpumask_test_cpu(smp_processor_id(), mm_cpumask(next))) { in switch_mm_irqs_off() 79 cpumask_set_cpu(smp_processor_id(), mm_cpumask(next)); in switch_mm_irqs_off() 101 tsk->thread.pgdir = next->pgd; in switch_mm_irqs_off() [all …]
|
/arch/x86/mm/ |
D | ident_map.c | 23 unsigned long next; in ident_pud_init() local 25 for (; addr < end; addr = next) { in ident_pud_init() 29 next = (addr & PUD_MASK) + PUD_SIZE; in ident_pud_init() 30 if (next > end) in ident_pud_init() 31 next = end; in ident_pud_init() 35 ident_pmd_init(info, pmd, addr, next); in ident_pud_init() 41 ident_pmd_init(info, pmd, addr, next); in ident_pud_init() 53 unsigned long next; in kernel_ident_mapping_init() local 56 for (; addr < end; addr = next) { in kernel_ident_mapping_init() 60 next = (addr & PGDIR_MASK) + PGDIR_SIZE; in kernel_ident_mapping_init() [all …]
|
/arch/parisc/include/asm/ |
D | mmu_context.h | 53 struct mm_struct *next, struct task_struct *tsk) in switch_mm_irqs_off() argument 55 if (prev != next) { in switch_mm_irqs_off() 56 mtctl(__pa(next->pgd), 25); in switch_mm_irqs_off() 57 load_context(next->context); in switch_mm_irqs_off() 62 struct mm_struct *next, struct task_struct *tsk) in switch_mm() argument 67 switch_mm_irqs_off(prev, next, tsk); in switch_mm() 74 static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next) in activate_mm() argument 84 BUG_ON(next == &init_mm); /* Should never happen */ in activate_mm() 86 if (next->context == 0) in activate_mm() 87 next->context = alloc_sid(); in activate_mm() [all …]
|
/arch/ia64/include/asm/ |
D | switch_to.h | 44 #define __switch_to(prev,next,last) do { \ argument 47 if (IA64_HAS_EXTRA_STATE(next)) \ 48 ia64_load_extra(next); \ 49 ia64_psr(task_pt_regs(next))->dfh = !ia64_is_local_fpu_owner(next); \ 50 (last) = ia64_switch_to((next)); \ 60 # define switch_to(prev,next,last) do { \ argument 66 __switch_to(prev, next, last); \ 76 # define switch_to(prev,next,last) __switch_to(prev, next, last) argument
|
/arch/tile/include/asm/ |
D | cacheflush.h | 81 char *next = (char *)((long)buffer & -L2_CACHE_BYTES); in __flush_buffer() local 83 while (next < finish) { in __flush_buffer() 84 __insn_flush(next); in __flush_buffer() 85 next += CHIP_FLUSH_STRIDE(); in __flush_buffer() 92 char *next = (char *)((long)buffer & -L2_CACHE_BYTES); in __finv_buffer() local 94 while (next < finish) { in __finv_buffer() 95 __insn_finv(next); in __finv_buffer() 96 next += CHIP_FINV_STRIDE(); in __finv_buffer() 125 char *next = (char *)((long)buffer & -L2_CACHE_BYTES); in __inv_buffer() local 127 while (next < finish) { in __inv_buffer() [all …]
|
D | switch_to.h | 36 #define prepare_arch_switch(next) _prepare_arch_switch(next) argument 37 void _prepare_arch_switch(struct task_struct *next); 40 #define switch_to(prev, next, last) ((last) = _switch_to((prev), (next))) argument 42 struct task_struct *next); 46 struct task_struct *next,
|
/arch/score/include/asm/ |
D | mmu_context.h | 72 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, in switch_mm() argument 78 if ((next->context ^ asid_cache) & ASID_VERSION_MASK) in switch_mm() 79 get_new_mmu_context(next); in switch_mm() 81 pevn_set(next->context); in switch_mm() 82 TLBMISS_HANDLER_SETUP_PGD(next->pgd); in switch_mm() 102 activate_mm(struct mm_struct *prev, struct mm_struct *next) in activate_mm() argument 107 get_new_mmu_context(next); in activate_mm() 108 pevn_set(next->context); in activate_mm() 109 TLBMISS_HANDLER_SETUP_PGD(next->pgd); in activate_mm()
|
/arch/mn10300/kernel/ |
D | kprobes.c | 133 u8 opc, *pc, *sp, *next; in find_nextpc() local 135 next = 0; in find_nextpc() 144 next = pc + size; in find_nextpc() 152 next = pc + x8; in find_nextpc() 160 next = pc + x16; in find_nextpc() 168 next = pc + x32; in find_nextpc() 174 next = (u8 *)regs->mdr; in find_nextpc() 183 next = (u8 *)x32; in find_nextpc() 189 next = pc + 2; in find_nextpc() 195 next = (u8 *)regs->a0; in find_nextpc() [all …]
|
/arch/sh/mm/ |
D | tlb-sh5.c | 27 cpu_data->dtlb.next = cpu_data->dtlb.first; in sh64_tlb_init() 38 cpu_data->itlb.next = cpu_data->itlb.first; in sh64_tlb_init() 51 return cpu_data->dtlb.next; in sh64_next_free_dtlb_entry() 62 cpu_data->dtlb.next += cpu_data->dtlb.step; in sh64_get_wired_dtlb_entry() 104 cpu_data->dtlb.next = entry; in sh64_put_wired_dtlb_entry() 191 unsigned long long next; in __update_tlb() local 214 next = tlbp->next; in __update_tlb() 215 __flush_tlb_slot(next); in __update_tlb() 218 : : "r" (next), "r" (pteh), "r" (ptel) ); in __update_tlb() 220 next += TLB_STEP; in __update_tlb() [all …]
|
D | gup.c | 122 unsigned long next; in gup_pmd_range() local 129 next = pmd_addr_end(addr, end); in gup_pmd_range() 132 if (!gup_pte_range(pmd, addr, next, write, pages, nr)) in gup_pmd_range() 134 } while (pmdp++, addr = next, addr != end); in gup_pmd_range() 142 unsigned long next; in gup_pud_range() local 149 next = pud_addr_end(addr, end); in gup_pud_range() 152 if (!gup_pmd_range(pud, addr, next, write, pages, nr)) in gup_pud_range() 154 } while (pudp++, addr = next, addr != end); in gup_pud_range() 168 unsigned long next; in __get_user_pages_fast() local 190 next = pgd_addr_end(addr, end); in __get_user_pages_fast() [all …]
|
/arch/nios2/mm/ |
D | mmu_context.c | 80 void switch_mm(struct mm_struct *prev, struct mm_struct *next, in switch_mm() argument 89 if (unlikely(CTX_VERSION(next->context) != in switch_mm() 91 next->context = get_new_context(); in switch_mm() 94 pgd_current = next->pgd; in switch_mm() 97 set_context(next->context); in switch_mm() 106 void activate_mm(struct mm_struct *prev, struct mm_struct *next) in activate_mm() argument 108 next->context = get_new_context(); in activate_mm() 109 set_context(next->context); in activate_mm() 110 pgd_current = next->pgd; in activate_mm()
|
/arch/frv/include/asm/ |
D | mmu_context.h | 35 #define switch_mm(prev, next, tsk) \ argument 37 if (prev != next) \ 38 change_mm_context(&prev->context, &next->context, next->pgd); \ 41 #define activate_mm(prev, next) \ argument 43 change_mm_context(&prev->context, &next->context, next->pgd); \
|
/arch/x86/include/asm/ |
D | switch_to.h | 7 struct task_struct *next); 10 struct task_struct *next); 17 struct task_struct *next) in prepare_switch_to() argument 33 READ_ONCE(*(unsigned char *)next->thread.sp); in prepare_switch_to() 60 #define switch_to(prev, next, last) \ argument 62 prepare_switch_to(prev, next); \ 64 ((last) = __switch_to_asm((prev), (next))); \
|
/arch/arm64/include/asm/ |
D | mmu_context.h | 34 static inline void contextidr_thread_switch(struct task_struct *next) in contextidr_thread_switch() argument 39 write_sysreg(task_pid_nr(next), contextidr_el1); in contextidr_thread_switch() 192 static inline void __switch_mm(struct mm_struct *next) in __switch_mm() argument 200 if (next == &init_mm) { in __switch_mm() 205 check_and_switch_context(next, cpu); in __switch_mm() 209 switch_mm(struct mm_struct *prev, struct mm_struct *next, in switch_mm() argument 212 if (prev != next) in switch_mm() 213 __switch_mm(next); in switch_mm() 222 if (next != &init_mm) in switch_mm() 223 update_saved_ttbr0(tsk, next); in switch_mm() [all …]
|
/arch/sh/include/asm/ |
D | switch_to_64.h | 21 struct task_struct *next, 24 #define switch_to(prev,next,last) \ argument 26 if (last_task_used_math != next) { \ 27 struct pt_regs *regs = next->thread.uregs; \ 30 last = sh64_switch_to(prev, &prev->thread, next, \ 31 &next->thread); \
|
/arch/blackfin/include/asm/ |
D | switch_to.h | 21 asmlinkage struct task_struct *resume(struct task_struct *prev, struct task_struct *next); 24 #define switch_to(prev,next,last) \ argument 28 memcpy (L1_SCRATCH_TASK_INFO, &task_thread_info(next)->l1_task_info, \ 30 (last) = resume (prev, next); \ 33 #define switch_to(prev, next, last) \ argument 35 (last) = resume(prev, next); \
|
/arch/avr32/include/asm/ |
D | switch_to.h | 18 #define ocd_switch(prev, next) \ argument 21 ocd_write(PID, next->pid); \ 24 #define ocd_switch(prev, next) argument 41 #define switch_to(prev, next, last) \ argument 43 ocd_switch(prev, next); \ 45 &next->thread.cpu_context); \
|
/arch/arm/mm/ |
D | idmap.c | 25 unsigned long next; in idmap_add_pmd() local 46 next = pmd_addr_end(addr, end); in idmap_add_pmd() 49 } while (pmd++, addr = next, addr != end); in idmap_add_pmd() 69 unsigned long next; in idmap_add_pud() local 72 next = pud_addr_end(addr, end); in idmap_add_pud() 73 idmap_add_pmd(pud, addr, next, prot); in idmap_add_pud() 74 } while (pud++, addr = next, addr != end); in idmap_add_pud() 81 unsigned long next; in identity_mapping_add() local 94 next = pgd_addr_end(addr, end); in identity_mapping_add() 95 idmap_add_pud(pgd, addr, next, prot); in identity_mapping_add() [all …]
|
/arch/m32r/include/asm/ |
D | mmu_context.h | 122 struct mm_struct *next, struct task_struct *tsk) in switch_mm() argument 128 if (prev != next) { in switch_mm() 130 cpumask_set_cpu(cpu, mm_cpumask(next)); in switch_mm() 133 *(volatile unsigned long *)MPTB = (unsigned long)next->pgd; in switch_mm() 134 activate_context(next); in switch_mm() 138 if (!cpumask_test_and_set_cpu(cpu, mm_cpumask(next))) in switch_mm() 139 activate_context(next); in switch_mm() 145 #define activate_mm(prev, next) \ argument 146 switch_mm((prev), (next), NULL) 155 #define switch_mm(prev,next,tsk) do { } while (0) argument [all …]
|
/arch/hexagon/include/asm/ |
D | mmu_context.h | 66 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, in switch_mm() argument 75 if (next->context.generation < prev->context.generation) { in switch_mm() 77 next->pgd[l1] = init_mm.pgd[l1]; in switch_mm() 79 next->context.generation = prev->context.generation; in switch_mm() 82 __vmnewmap((void *)next->context.ptbase); in switch_mm() 88 static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next) in activate_mm() argument 93 switch_mm(prev, next, current_thread_info()->task); in activate_mm()
|
/arch/arm64/mm/ |
D | kasan_init.c | 41 unsigned long next; in kasan_early_pte_populate() local 48 next = addr + PAGE_SIZE; in kasan_early_pte_populate() 51 } while (pte++, addr = next, addr != end && pte_none(*pte)); in kasan_early_pte_populate() 59 unsigned long next; in kasan_early_pmd_populate() local 66 next = pmd_addr_end(addr, end); in kasan_early_pmd_populate() 67 kasan_early_pte_populate(pmd, addr, next); in kasan_early_pmd_populate() 68 } while (pmd++, addr = next, addr != end && pmd_none(*pmd)); in kasan_early_pmd_populate() 76 unsigned long next; in kasan_early_pud_populate() local 83 next = pud_addr_end(addr, end); in kasan_early_pud_populate() 84 kasan_early_pmd_populate(pud, addr, next); in kasan_early_pud_populate() [all …]
|
/arch/sparc/mm/ |
D | gup.c | 109 unsigned long next; in gup_pmd_range() local 116 next = pmd_addr_end(addr, end); in gup_pmd_range() 120 if (!gup_huge_pmd(pmdp, pmd, addr, next, in gup_pmd_range() 123 } else if (!gup_pte_range(pmd, addr, next, write, in gup_pmd_range() 126 } while (pmdp++, addr = next, addr != end); in gup_pmd_range() 134 unsigned long next; in gup_pud_range() local 141 next = pud_addr_end(addr, end); in gup_pud_range() 144 if (!gup_pmd_range(pud, addr, next, write, pages, nr)) in gup_pud_range() 146 } while (pudp++, addr = next, addr != end); in gup_pud_range() 156 unsigned long next, flags; in __get_user_pages_fast() local [all …]
|
/arch/mn10300/include/asm/ |
D | switch_to.h | 22 #define switch_fpu(prev, next) \ argument 31 #define switch_fpu(prev, next) do {} while (0) argument 37 struct thread_struct *next, 40 #define switch_to(prev, next, last) \ argument 42 switch_fpu(prev, next); \ 44 (last) = __switch_to(&(prev)->thread, &(next)->thread, (prev)); \
|