Home
last modified time | relevance | path

Searched refs:next (Results 1 – 25 of 620) sorted by relevance

12345678910>>...25

/arch/mips/include/asm/
Dswitch_to.h33 struct task_struct *next, struct thread_info *next_ti);
62 next->thread.emulated_fp = 0; \
82 #define __sanitize_fcr31(next) \ argument
84 unsigned long fcr31 = mask_fcr31_x(next->thread.fpu.fcr31); \
88 pc = (void __user *)task_pt_regs(next)->cp0_epc; \
89 next->thread.fpu.fcr31 &= ~fcr31; \
90 force_fcr31_sig(fcr31, pc, next); \
100 #define switch_to(prev, next, last) \ argument
104 if (tsk_used_math(next)) \
105 __sanitize_fcr31(next); \
[all …]
/arch/blackfin/mm/
Dsram-alloc.c31 struct sram_piece *next; member
78 per_cpu(free_l1_ssram_head, cpu).next = in l1sram_init()
80 if (!per_cpu(free_l1_ssram_head, cpu).next) { in l1sram_init()
85 per_cpu(free_l1_ssram_head, cpu).next->paddr = (void *)get_l1_scratch_start_cpu(cpu) + reserve; in l1sram_init()
86 per_cpu(free_l1_ssram_head, cpu).next->size = L1_SCRATCH_LENGTH - reserve; in l1sram_init()
87 per_cpu(free_l1_ssram_head, cpu).next->pid = 0; in l1sram_init()
88 per_cpu(free_l1_ssram_head, cpu).next->next = NULL; in l1sram_init()
90 per_cpu(used_l1_ssram_head, cpu).next = NULL; in l1sram_init()
106 per_cpu(free_l1_data_A_sram_head, cpu).next = in l1_data_sram_init()
108 if (!per_cpu(free_l1_data_A_sram_head, cpu).next) { in l1_data_sram_init()
[all …]
/arch/powerpc/include/asm/
Dmmu_context.h44 struct mm_struct *next);
46 struct mm_struct *next, in switch_mmu_context() argument
50 return radix__switch_mmu_context(prev, next); in switch_mmu_context()
51 return switch_slb(tsk, next); in switch_mmu_context()
58 extern void switch_mmu_context(struct mm_struct *prev, struct mm_struct *next,
65 extern void switch_cop(struct mm_struct *next);
74 struct mm_struct *next, in switch_mm_irqs_off() argument
78 if (!cpumask_test_cpu(smp_processor_id(), mm_cpumask(next))) { in switch_mm_irqs_off()
79 cpumask_set_cpu(smp_processor_id(), mm_cpumask(next)); in switch_mm_irqs_off()
101 tsk->thread.pgdir = next->pgd; in switch_mm_irqs_off()
[all …]
/arch/x86/mm/
Dident_map.c23 unsigned long next; in ident_pud_init() local
25 for (; addr < end; addr = next) { in ident_pud_init()
29 next = (addr & PUD_MASK) + PUD_SIZE; in ident_pud_init()
30 if (next > end) in ident_pud_init()
31 next = end; in ident_pud_init()
35 ident_pmd_init(info, pmd, addr, next); in ident_pud_init()
41 ident_pmd_init(info, pmd, addr, next); in ident_pud_init()
53 unsigned long next; in kernel_ident_mapping_init() local
56 for (; addr < end; addr = next) { in kernel_ident_mapping_init()
60 next = (addr & PGDIR_MASK) + PGDIR_SIZE; in kernel_ident_mapping_init()
[all …]
/arch/parisc/include/asm/
Dmmu_context.h53 struct mm_struct *next, struct task_struct *tsk) in switch_mm_irqs_off() argument
55 if (prev != next) { in switch_mm_irqs_off()
56 mtctl(__pa(next->pgd), 25); in switch_mm_irqs_off()
57 load_context(next->context); in switch_mm_irqs_off()
62 struct mm_struct *next, struct task_struct *tsk) in switch_mm() argument
67 switch_mm_irqs_off(prev, next, tsk); in switch_mm()
74 static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next) in activate_mm() argument
84 BUG_ON(next == &init_mm); /* Should never happen */ in activate_mm()
86 if (next->context == 0) in activate_mm()
87 next->context = alloc_sid(); in activate_mm()
[all …]
/arch/ia64/include/asm/
Dswitch_to.h44 #define __switch_to(prev,next,last) do { \ argument
47 if (IA64_HAS_EXTRA_STATE(next)) \
48 ia64_load_extra(next); \
49 ia64_psr(task_pt_regs(next))->dfh = !ia64_is_local_fpu_owner(next); \
50 (last) = ia64_switch_to((next)); \
60 # define switch_to(prev,next,last) do { \ argument
66 __switch_to(prev, next, last); \
76 # define switch_to(prev,next,last) __switch_to(prev, next, last) argument
/arch/tile/include/asm/
Dcacheflush.h81 char *next = (char *)((long)buffer & -L2_CACHE_BYTES); in __flush_buffer() local
83 while (next < finish) { in __flush_buffer()
84 __insn_flush(next); in __flush_buffer()
85 next += CHIP_FLUSH_STRIDE(); in __flush_buffer()
92 char *next = (char *)((long)buffer & -L2_CACHE_BYTES); in __finv_buffer() local
94 while (next < finish) { in __finv_buffer()
95 __insn_finv(next); in __finv_buffer()
96 next += CHIP_FINV_STRIDE(); in __finv_buffer()
125 char *next = (char *)((long)buffer & -L2_CACHE_BYTES); in __inv_buffer() local
127 while (next < finish) { in __inv_buffer()
[all …]
Dswitch_to.h36 #define prepare_arch_switch(next) _prepare_arch_switch(next) argument
37 void _prepare_arch_switch(struct task_struct *next);
40 #define switch_to(prev, next, last) ((last) = _switch_to((prev), (next))) argument
42 struct task_struct *next);
46 struct task_struct *next,
/arch/score/include/asm/
Dmmu_context.h72 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, in switch_mm() argument
78 if ((next->context ^ asid_cache) & ASID_VERSION_MASK) in switch_mm()
79 get_new_mmu_context(next); in switch_mm()
81 pevn_set(next->context); in switch_mm()
82 TLBMISS_HANDLER_SETUP_PGD(next->pgd); in switch_mm()
102 activate_mm(struct mm_struct *prev, struct mm_struct *next) in activate_mm() argument
107 get_new_mmu_context(next); in activate_mm()
108 pevn_set(next->context); in activate_mm()
109 TLBMISS_HANDLER_SETUP_PGD(next->pgd); in activate_mm()
/arch/mn10300/kernel/
Dkprobes.c133 u8 opc, *pc, *sp, *next; in find_nextpc() local
135 next = 0; in find_nextpc()
144 next = pc + size; in find_nextpc()
152 next = pc + x8; in find_nextpc()
160 next = pc + x16; in find_nextpc()
168 next = pc + x32; in find_nextpc()
174 next = (u8 *)regs->mdr; in find_nextpc()
183 next = (u8 *)x32; in find_nextpc()
189 next = pc + 2; in find_nextpc()
195 next = (u8 *)regs->a0; in find_nextpc()
[all …]
/arch/sh/mm/
Dtlb-sh5.c27 cpu_data->dtlb.next = cpu_data->dtlb.first; in sh64_tlb_init()
38 cpu_data->itlb.next = cpu_data->itlb.first; in sh64_tlb_init()
51 return cpu_data->dtlb.next; in sh64_next_free_dtlb_entry()
62 cpu_data->dtlb.next += cpu_data->dtlb.step; in sh64_get_wired_dtlb_entry()
104 cpu_data->dtlb.next = entry; in sh64_put_wired_dtlb_entry()
191 unsigned long long next; in __update_tlb() local
214 next = tlbp->next; in __update_tlb()
215 __flush_tlb_slot(next); in __update_tlb()
218 : : "r" (next), "r" (pteh), "r" (ptel) ); in __update_tlb()
220 next += TLB_STEP; in __update_tlb()
[all …]
Dgup.c122 unsigned long next; in gup_pmd_range() local
129 next = pmd_addr_end(addr, end); in gup_pmd_range()
132 if (!gup_pte_range(pmd, addr, next, write, pages, nr)) in gup_pmd_range()
134 } while (pmdp++, addr = next, addr != end); in gup_pmd_range()
142 unsigned long next; in gup_pud_range() local
149 next = pud_addr_end(addr, end); in gup_pud_range()
152 if (!gup_pmd_range(pud, addr, next, write, pages, nr)) in gup_pud_range()
154 } while (pudp++, addr = next, addr != end); in gup_pud_range()
168 unsigned long next; in __get_user_pages_fast() local
190 next = pgd_addr_end(addr, end); in __get_user_pages_fast()
[all …]
/arch/nios2/mm/
Dmmu_context.c80 void switch_mm(struct mm_struct *prev, struct mm_struct *next, in switch_mm() argument
89 if (unlikely(CTX_VERSION(next->context) != in switch_mm()
91 next->context = get_new_context(); in switch_mm()
94 pgd_current = next->pgd; in switch_mm()
97 set_context(next->context); in switch_mm()
106 void activate_mm(struct mm_struct *prev, struct mm_struct *next) in activate_mm() argument
108 next->context = get_new_context(); in activate_mm()
109 set_context(next->context); in activate_mm()
110 pgd_current = next->pgd; in activate_mm()
/arch/frv/include/asm/
Dmmu_context.h35 #define switch_mm(prev, next, tsk) \ argument
37 if (prev != next) \
38 change_mm_context(&prev->context, &next->context, next->pgd); \
41 #define activate_mm(prev, next) \ argument
43 change_mm_context(&prev->context, &next->context, next->pgd); \
/arch/x86/include/asm/
Dswitch_to.h7 struct task_struct *next);
10 struct task_struct *next);
17 struct task_struct *next) in prepare_switch_to() argument
33 READ_ONCE(*(unsigned char *)next->thread.sp); in prepare_switch_to()
60 #define switch_to(prev, next, last) \ argument
62 prepare_switch_to(prev, next); \
64 ((last) = __switch_to_asm((prev), (next))); \
/arch/arm64/include/asm/
Dmmu_context.h34 static inline void contextidr_thread_switch(struct task_struct *next) in contextidr_thread_switch() argument
39 write_sysreg(task_pid_nr(next), contextidr_el1); in contextidr_thread_switch()
192 static inline void __switch_mm(struct mm_struct *next) in __switch_mm() argument
200 if (next == &init_mm) { in __switch_mm()
205 check_and_switch_context(next, cpu); in __switch_mm()
209 switch_mm(struct mm_struct *prev, struct mm_struct *next, in switch_mm() argument
212 if (prev != next) in switch_mm()
213 __switch_mm(next); in switch_mm()
222 if (next != &init_mm) in switch_mm()
223 update_saved_ttbr0(tsk, next); in switch_mm()
[all …]
/arch/sh/include/asm/
Dswitch_to_64.h21 struct task_struct *next,
24 #define switch_to(prev,next,last) \ argument
26 if (last_task_used_math != next) { \
27 struct pt_regs *regs = next->thread.uregs; \
30 last = sh64_switch_to(prev, &prev->thread, next, \
31 &next->thread); \
/arch/blackfin/include/asm/
Dswitch_to.h21 asmlinkage struct task_struct *resume(struct task_struct *prev, struct task_struct *next);
24 #define switch_to(prev,next,last) \ argument
28 memcpy (L1_SCRATCH_TASK_INFO, &task_thread_info(next)->l1_task_info, \
30 (last) = resume (prev, next); \
33 #define switch_to(prev, next, last) \ argument
35 (last) = resume(prev, next); \
/arch/avr32/include/asm/
Dswitch_to.h18 #define ocd_switch(prev, next) \ argument
21 ocd_write(PID, next->pid); \
24 #define ocd_switch(prev, next) argument
41 #define switch_to(prev, next, last) \ argument
43 ocd_switch(prev, next); \
45 &next->thread.cpu_context); \
/arch/arm/mm/
Didmap.c25 unsigned long next; in idmap_add_pmd() local
46 next = pmd_addr_end(addr, end); in idmap_add_pmd()
49 } while (pmd++, addr = next, addr != end); in idmap_add_pmd()
69 unsigned long next; in idmap_add_pud() local
72 next = pud_addr_end(addr, end); in idmap_add_pud()
73 idmap_add_pmd(pud, addr, next, prot); in idmap_add_pud()
74 } while (pud++, addr = next, addr != end); in idmap_add_pud()
81 unsigned long next; in identity_mapping_add() local
94 next = pgd_addr_end(addr, end); in identity_mapping_add()
95 idmap_add_pud(pgd, addr, next, prot); in identity_mapping_add()
[all …]
/arch/m32r/include/asm/
Dmmu_context.h122 struct mm_struct *next, struct task_struct *tsk) in switch_mm() argument
128 if (prev != next) { in switch_mm()
130 cpumask_set_cpu(cpu, mm_cpumask(next)); in switch_mm()
133 *(volatile unsigned long *)MPTB = (unsigned long)next->pgd; in switch_mm()
134 activate_context(next); in switch_mm()
138 if (!cpumask_test_and_set_cpu(cpu, mm_cpumask(next))) in switch_mm()
139 activate_context(next); in switch_mm()
145 #define activate_mm(prev, next) \ argument
146 switch_mm((prev), (next), NULL)
155 #define switch_mm(prev,next,tsk) do { } while (0) argument
[all …]
/arch/hexagon/include/asm/
Dmmu_context.h66 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, in switch_mm() argument
75 if (next->context.generation < prev->context.generation) { in switch_mm()
77 next->pgd[l1] = init_mm.pgd[l1]; in switch_mm()
79 next->context.generation = prev->context.generation; in switch_mm()
82 __vmnewmap((void *)next->context.ptbase); in switch_mm()
88 static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next) in activate_mm() argument
93 switch_mm(prev, next, current_thread_info()->task); in activate_mm()
/arch/arm64/mm/
Dkasan_init.c41 unsigned long next; in kasan_early_pte_populate() local
48 next = addr + PAGE_SIZE; in kasan_early_pte_populate()
51 } while (pte++, addr = next, addr != end && pte_none(*pte)); in kasan_early_pte_populate()
59 unsigned long next; in kasan_early_pmd_populate() local
66 next = pmd_addr_end(addr, end); in kasan_early_pmd_populate()
67 kasan_early_pte_populate(pmd, addr, next); in kasan_early_pmd_populate()
68 } while (pmd++, addr = next, addr != end && pmd_none(*pmd)); in kasan_early_pmd_populate()
76 unsigned long next; in kasan_early_pud_populate() local
83 next = pud_addr_end(addr, end); in kasan_early_pud_populate()
84 kasan_early_pmd_populate(pud, addr, next); in kasan_early_pud_populate()
[all …]
/arch/sparc/mm/
Dgup.c109 unsigned long next; in gup_pmd_range() local
116 next = pmd_addr_end(addr, end); in gup_pmd_range()
120 if (!gup_huge_pmd(pmdp, pmd, addr, next, in gup_pmd_range()
123 } else if (!gup_pte_range(pmd, addr, next, write, in gup_pmd_range()
126 } while (pmdp++, addr = next, addr != end); in gup_pmd_range()
134 unsigned long next; in gup_pud_range() local
141 next = pud_addr_end(addr, end); in gup_pud_range()
144 if (!gup_pmd_range(pud, addr, next, write, pages, nr)) in gup_pud_range()
146 } while (pudp++, addr = next, addr != end); in gup_pud_range()
156 unsigned long next, flags; in __get_user_pages_fast() local
[all …]
/arch/mn10300/include/asm/
Dswitch_to.h22 #define switch_fpu(prev, next) \ argument
31 #define switch_fpu(prev, next) do {} while (0) argument
37 struct thread_struct *next,
40 #define switch_to(prev, next, last) \ argument
42 switch_fpu(prev, next); \
44 (last) = __switch_to(&(prev)->thread, &(next)->thread, (prev)); \

12345678910>>...25