Home
last modified time | relevance | path

Searched refs:next (Results 1 – 25 of 417) sorted by relevance

12345678910>>...17

/arch/blackfin/mm/
Dsram-alloc.c55 struct sram_piece *next; member
87 per_cpu(free_l1_ssram_head, cpu).next = in l1sram_init()
89 if (!per_cpu(free_l1_ssram_head, cpu).next) { in l1sram_init()
94 per_cpu(free_l1_ssram_head, cpu).next->paddr = (void *)get_l1_scratch_start_cpu(cpu); in l1sram_init()
95 per_cpu(free_l1_ssram_head, cpu).next->size = L1_SCRATCH_LENGTH; in l1sram_init()
96 per_cpu(free_l1_ssram_head, cpu).next->pid = 0; in l1sram_init()
97 per_cpu(free_l1_ssram_head, cpu).next->next = NULL; in l1sram_init()
99 per_cpu(used_l1_ssram_head, cpu).next = NULL; in l1sram_init()
115 per_cpu(free_l1_data_A_sram_head, cpu).next = in l1_data_sram_init()
117 if (!per_cpu(free_l1_data_A_sram_head, cpu).next) { in l1_data_sram_init()
[all …]
/arch/x86/include/asm/
Dmmu_context_64.h14 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, in switch_mm() argument
18 if (likely(prev != next)) { in switch_mm()
23 write_pda(active_mm, next); in switch_mm()
25 cpu_set(cpu, next->cpu_vm_mask); in switch_mm()
26 load_cr3(next->pgd); in switch_mm()
28 if (unlikely(next->context.ldt != prev->context.ldt)) in switch_mm()
29 load_LDT_nolock(&next->context); in switch_mm()
34 if (read_pda(active_mm) != next) in switch_mm()
36 if (!cpu_test_and_set(cpu, next->cpu_vm_mask)) { in switch_mm()
41 load_cr3(next->pgd); in switch_mm()
[all …]
Dmmu_context_32.h13 struct mm_struct *next, in switch_mm() argument
18 if (likely(prev != next)) { in switch_mm()
23 x86_write_percpu(cpu_tlbstate.active_mm, next); in switch_mm()
25 cpu_set(cpu, next->cpu_vm_mask); in switch_mm()
28 load_cr3(next->pgd); in switch_mm()
33 if (unlikely(prev->context.ldt != next->context.ldt)) in switch_mm()
34 load_LDT_nolock(&next->context); in switch_mm()
39 BUG_ON(x86_read_percpu(cpu_tlbstate.active_mm) != next); in switch_mm()
41 if (!cpu_test_and_set(cpu, next->cpu_vm_mask)) { in switch_mm()
45 load_cr3(next->pgd); in switch_mm()
[all …]
Dmmu_context.h13 struct mm_struct *next) in paravirt_activate_mm() argument
30 #define activate_mm(prev, next) \ argument
32 paravirt_activate_mm((prev), (next)); \
33 switch_mm((prev), (next), NULL); \
/arch/powerpc/include/asm/
Dmmu_context.h21 extern void switch_mmu_context(struct mm_struct *prev, struct mm_struct *next);
30 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, in switch_mm() argument
34 cpu_set(smp_processor_id(), next->cpu_vm_mask); in switch_mm()
38 tsk->thread.pgdir = next->pgd; in switch_mm()
42 if (prev == next) in switch_mm()
58 switch_slb(tsk, next); in switch_mm()
60 switch_stab(tsk, next); in switch_mm()
63 switch_mmu_context(prev, next); in switch_mm()
74 static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next) in activate_mm() argument
79 switch_mm(prev, next, current); in activate_mm()
/arch/parisc/include/asm/
Dmmu_context.h52 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, struct task_struct *ts… in switch_mm() argument
55 if (prev != next) { in switch_mm()
56 mtctl(__pa(next->pgd), 25); in switch_mm()
57 load_context(next->context); in switch_mm()
63 static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next) in activate_mm() argument
73 BUG_ON(next == &init_mm); /* Should never happen */ in activate_mm()
75 if (next->context == 0) in activate_mm()
76 next->context = alloc_sid(); in activate_mm()
78 switch_mm(prev,next,current); in activate_mm()
/arch/mn10300/kernel/
Dkprobes.c133 u8 opc, *pc, *sp, *next; in find_nextpc() local
135 next = 0; in find_nextpc()
144 next = pc + size; in find_nextpc()
152 next = pc + x8; in find_nextpc()
160 next = pc + x16; in find_nextpc()
168 next = pc + x32; in find_nextpc()
174 next = (u8 *)regs->mdr; in find_nextpc()
183 next = (u8 *)x32; in find_nextpc()
189 next = pc + 2; in find_nextpc()
195 next = (u8 *)regs->a0; in find_nextpc()
[all …]
/arch/arm/include/asm/
Dmmu_context.h96 switch_mm(struct mm_struct *prev, struct mm_struct *next, in switch_mm() argument
104 if (!cpus_empty(next->cpu_vm_mask) && !cpu_isset(cpu, next->cpu_vm_mask)) in switch_mm()
107 if (!cpu_test_and_set(cpu, next->cpu_vm_mask) || prev != next) { in switch_mm()
108 check_context(next); in switch_mm()
109 cpu_switch_mm(next->pgd, next); in switch_mm()
117 #define activate_mm(prev,next) switch_mm(prev, next, NULL) argument
/arch/s390/kernel/
Ds390_ext.c47 p->next = ext_int_hash[index]; in register_external_interrupt()
62 p->next = ext_int_hash[index]; in register_early_external_interrupt()
79 p = p->next; in unregister_external_interrupt()
84 q->next = p->next; in unregister_external_interrupt()
86 ext_int_hash[index] = p->next; in unregister_external_interrupt()
103 if (q->next == p) in unregister_early_external_interrupt()
105 q = q->next; in unregister_early_external_interrupt()
109 q->next = p->next; in unregister_early_external_interrupt()
111 ext_int_hash[index] = p->next; in unregister_early_external_interrupt()
129 for (p = ext_int_hash[index]; p; p = p->next) { in do_extint()
/arch/sh/include/asm/
Dsystem_64.h22 struct task_struct *next,
25 #define switch_to(prev,next,last) \ argument
27 if (last_task_used_math != next) { \
28 struct pt_regs *regs = next->thread.uregs; \
31 last = sh64_switch_to(prev, &prev->thread, next, \
32 &next->thread); \
Dmmu_context.h114 struct mm_struct *next, in switch_mm() argument
119 if (likely(prev != next)) { in switch_mm()
120 cpu_set(cpu, next->cpu_vm_mask); in switch_mm()
121 set_TTB(next->pgd); in switch_mm()
122 activate_context(next, cpu); in switch_mm()
124 if (!cpu_test_and_set(cpu, next->cpu_vm_mask)) in switch_mm()
125 activate_context(next, cpu); in switch_mm()
138 #define switch_mm(prev,next,tsk) do { } while (0) argument
141 #define activate_mm(prev, next) switch_mm((prev),(next),NULL) argument
/arch/xtensa/include/asm/
Dmmu_context.h96 activate_mm(struct mm_struct *prev, struct mm_struct *next) in activate_mm() argument
100 __get_new_mmu_context(next); in activate_mm()
101 __load_mmu_context(next); in activate_mm()
105 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, in switch_mm() argument
112 if (next->context == NO_CONTEXT || ((next->context^asid) & ~ASID_MASK)) in switch_mm()
113 __get_new_mmu_context(next); in switch_mm()
115 __load_mmu_context(next); in switch_mm()
/arch/alpha/kernel/
Derr_common.c158 struct el_subpacket *next = NULL; in el_process_subpacket_reg() local
161 for (; h && h->class != header->class; h = h->next); in el_process_subpacket_reg()
162 if (h) next = h->handler(header); in el_process_subpacket_reg()
164 return next; in el_process_subpacket_reg()
196 struct el_subpacket *next = NULL; in el_process_subpacket() local
203 next = el_process_header_subpacket(header); in el_process_subpacket()
206 if (NULL == (next = el_process_subpacket_reg(header))) { in el_process_subpacket()
215 return next; in el_process_subpacket()
224 for (a = subpacket_annotation_list; a; a = a->next) { in el_annotate_subpacket()
285 for (; a->next != NULL; a = a->next) { in cdl_register_subpacket_annotation()
[all …]
/arch/powerpc/boot/dtc-src/
Dlivetree.c34 new->next = NULL; in build_property()
43 assert(first->next == NULL); in chain_property()
45 first->next = list; in chain_property()
53 struct property *next; in reverse_properties() local
56 next = p->next; in reverse_properties()
57 p->next = head; in reverse_properties()
59 p = next; in reverse_properties()
104 prop->next = NULL; in add_property()
108 p = &((*p)->next); in add_property()
135 new->next = NULL; in build_reserve_entry()
[all …]
/arch/mips/include/asm/
Dmmu_context.h129 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, in switch_mm() argument
145 if ((cpu_context(cpu, next) ^ asid_cache(cpu)) & ASID_VERSION_MASK) in switch_mm()
146 get_new_mmu_context(next, cpu); in switch_mm()
168 | (cpu_context(cpu, next) & ASID_MASK)); in switch_mm()
172 write_c0_entryhi(cpu_context(cpu, next)); in switch_mm()
174 TLBMISS_HANDLER_SETUP_PGD(next->pgd); in switch_mm()
181 cpu_set(cpu, next->cpu_vm_mask); in switch_mm()
201 activate_mm(struct mm_struct *prev, struct mm_struct *next) in activate_mm() argument
215 get_new_mmu_context(next, cpu); in activate_mm()
228 (cpu_context(cpu, next) & ASID_MASK)); in activate_mm()
[all …]
/arch/ia64/include/asm/
Dmmu_context.h36 unsigned int next; /* next context number to use */ member
91 if (ia64_ctx.next >= ia64_ctx.limit) { in get_mmu_context()
92 ia64_ctx.next = find_next_zero_bit(ia64_ctx.bitmap, in get_mmu_context()
93 ia64_ctx.max_ctx, ia64_ctx.next); in get_mmu_context()
95 ia64_ctx.max_ctx, ia64_ctx.next); in get_mmu_context()
96 if (ia64_ctx.next >= ia64_ctx.max_ctx) in get_mmu_context()
99 mm->context = context = ia64_ctx.next++; in get_mmu_context()
185 activate_mm (struct mm_struct *prev, struct mm_struct *next) in activate_mm() argument
191 ia64_set_kr(IA64_KR_PT_BASE, __pa(next->pgd)); in activate_mm()
192 activate_context(next); in activate_mm()
Dsystem.h221 extern void ia64_account_on_switch (struct task_struct *prev, struct task_struct *next);
238 #define __switch_to(prev,next,last) do { \ argument
239 IA64_ACCOUNT_ON_SWITCH(prev, next); \
242 if (IA64_HAS_EXTRA_STATE(next)) \
243 ia64_load_extra(next); \
244 ia64_psr(task_pt_regs(next))->dfh = !ia64_is_local_fpu_owner(next); \
245 (last) = ia64_switch_to((next)); \
255 # define switch_to(prev,next,last) do { \ argument
261 __switch_to(prev, next, last); \
271 # define switch_to(prev,next,last) __switch_to(prev, next, last) argument
/arch/sh/mm/
Dfault_64.c51 unsigned long long next; in __do_tlb_refill() local
78 next = tlbp->next; in __do_tlb_refill()
79 __flush_tlb_slot(next); in __do_tlb_refill()
82 : : "r" (next), "r" (pteh), "r" (ptel) ); in __do_tlb_refill()
84 next += TLB_STEP; in __do_tlb_refill()
85 if (next > tlbp->last) next = tlbp->first; in __do_tlb_refill()
86 tlbp->next = next; in __do_tlb_refill()
/arch/x86/mm/
Dgup.c143 unsigned long next; in gup_pmd_range() local
150 next = pmd_addr_end(addr, end); in gup_pmd_range()
154 if (!gup_huge_pmd(pmd, addr, next, write, pages, nr)) in gup_pmd_range()
157 if (!gup_pte_range(pmd, addr, next, write, pages, nr)) in gup_pmd_range()
160 } while (pmdp++, addr = next, addr != end); in gup_pmd_range()
200 unsigned long next; in gup_pud_range() local
207 next = pud_addr_end(addr, end); in gup_pud_range()
211 if (!gup_huge_pud(pud, addr, next, write, pages, nr)) in gup_pud_range()
214 if (!gup_pmd_range(pud, addr, next, write, pages, nr)) in gup_pud_range()
217 } while (pudp++, addr = next, addr != end); in gup_pud_range()
[all …]
/arch/um/include/asm/
Dmmu_context.h32 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, in switch_mm() argument
37 if(prev != next){ in switch_mm()
39 cpu_set(cpu, next->cpu_vm_mask); in switch_mm()
40 if(next != &init_mm) in switch_mm()
41 __switch_mm(&next->context.id); in switch_mm()
Dsystem.h6 extern void *switch_to(void *prev, void *next, void *last);
32 extern void *_switch_to(void *prev, void *next, void *last);
33 #define switch_to(prev, next, last) prev = _switch_to(prev, next, last) argument
/arch/avr32/include/asm/
Dmmu_context.h109 struct mm_struct *next, in switch_mm() argument
112 if (likely(prev != next)) { in switch_mm()
113 unsigned long __pgdir = (unsigned long)next->pgd; in switch_mm()
116 activate_context(next); in switch_mm()
122 #define activate_mm(prev, next) switch_mm((prev), (next), NULL) argument
/arch/powerpc/mm/
Dgup.c110 unsigned long next; in gup_pmd_range() local
117 next = pmd_addr_end(addr, end); in gup_pmd_range()
120 if (!gup_pte_range(pmd, addr, next, write, pages, nr)) in gup_pmd_range()
122 } while (pmdp++, addr = next, addr != end); in gup_pmd_range()
130 unsigned long next; in gup_pud_range() local
137 next = pud_addr_end(addr, end); in gup_pud_range()
140 if (!gup_pmd_range(pud, addr, next, write, pages, nr)) in gup_pud_range()
142 } while (pudp++, addr = next, addr != end); in gup_pud_range()
152 unsigned long next; in get_user_pages_fast() local
242 next = pgd_addr_end(addr, end); in get_user_pages_fast()
[all …]
Dmmu_context_nohash.c173 void switch_mmu_context(struct mm_struct *prev, struct mm_struct *next) in switch_mmu_context() argument
183 cpu, next, next->context.active, next->context.id); in switch_mmu_context()
188 next->context.active++; in switch_mmu_context()
200 id = next->context.id; in switch_mmu_context()
231 context_mm[id] = next; in switch_mmu_context()
232 next->context.id = id; in switch_mmu_context()
247 cpu, id, next); in switch_mmu_context()
248 local_flush_tlb_mm(next); in switch_mmu_context()
255 set_context(id, next->pgd); in switch_mmu_context()
/arch/x86/kernel/
Dprocess_64.c479 struct thread_struct *prev, *next; in __switch_to_xtra() local
482 next = &next_p->thread; in __switch_to_xtra()
487 else if (next->debugctlmsr != prev->debugctlmsr) in __switch_to_xtra()
488 update_debugctlmsr(next->debugctlmsr); in __switch_to_xtra()
491 loaddebug(next, 0); in __switch_to_xtra()
492 loaddebug(next, 1); in __switch_to_xtra()
493 loaddebug(next, 2); in __switch_to_xtra()
494 loaddebug(next, 3); in __switch_to_xtra()
496 loaddebug(next, 6); in __switch_to_xtra()
497 loaddebug(next, 7); in __switch_to_xtra()
[all …]

12345678910>>...17