/arch/blackfin/mm/ |
D | sram-alloc.c | 55 struct sram_piece *next; member 87 per_cpu(free_l1_ssram_head, cpu).next = in l1sram_init() 89 if (!per_cpu(free_l1_ssram_head, cpu).next) { in l1sram_init() 94 per_cpu(free_l1_ssram_head, cpu).next->paddr = (void *)get_l1_scratch_start_cpu(cpu); in l1sram_init() 95 per_cpu(free_l1_ssram_head, cpu).next->size = L1_SCRATCH_LENGTH; in l1sram_init() 96 per_cpu(free_l1_ssram_head, cpu).next->pid = 0; in l1sram_init() 97 per_cpu(free_l1_ssram_head, cpu).next->next = NULL; in l1sram_init() 99 per_cpu(used_l1_ssram_head, cpu).next = NULL; in l1sram_init() 115 per_cpu(free_l1_data_A_sram_head, cpu).next = in l1_data_sram_init() 117 if (!per_cpu(free_l1_data_A_sram_head, cpu).next) { in l1_data_sram_init() [all …]
|
/arch/x86/include/asm/ |
D | mmu_context_64.h | 14 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, in switch_mm() argument 18 if (likely(prev != next)) { in switch_mm() 23 write_pda(active_mm, next); in switch_mm() 25 cpu_set(cpu, next->cpu_vm_mask); in switch_mm() 26 load_cr3(next->pgd); in switch_mm() 28 if (unlikely(next->context.ldt != prev->context.ldt)) in switch_mm() 29 load_LDT_nolock(&next->context); in switch_mm() 34 if (read_pda(active_mm) != next) in switch_mm() 36 if (!cpu_test_and_set(cpu, next->cpu_vm_mask)) { in switch_mm() 41 load_cr3(next->pgd); in switch_mm() [all …]
|
D | mmu_context_32.h | 13 struct mm_struct *next, in switch_mm() argument 18 if (likely(prev != next)) { in switch_mm() 23 x86_write_percpu(cpu_tlbstate.active_mm, next); in switch_mm() 25 cpu_set(cpu, next->cpu_vm_mask); in switch_mm() 28 load_cr3(next->pgd); in switch_mm() 33 if (unlikely(prev->context.ldt != next->context.ldt)) in switch_mm() 34 load_LDT_nolock(&next->context); in switch_mm() 39 BUG_ON(x86_read_percpu(cpu_tlbstate.active_mm) != next); in switch_mm() 41 if (!cpu_test_and_set(cpu, next->cpu_vm_mask)) { in switch_mm() 45 load_cr3(next->pgd); in switch_mm() [all …]
|
D | mmu_context.h | 13 struct mm_struct *next) in paravirt_activate_mm() argument 30 #define activate_mm(prev, next) \ argument 32 paravirt_activate_mm((prev), (next)); \ 33 switch_mm((prev), (next), NULL); \
|
/arch/powerpc/include/asm/ |
D | mmu_context.h | 21 extern void switch_mmu_context(struct mm_struct *prev, struct mm_struct *next); 30 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, in switch_mm() argument 34 cpu_set(smp_processor_id(), next->cpu_vm_mask); in switch_mm() 38 tsk->thread.pgdir = next->pgd; in switch_mm() 42 if (prev == next) in switch_mm() 58 switch_slb(tsk, next); in switch_mm() 60 switch_stab(tsk, next); in switch_mm() 63 switch_mmu_context(prev, next); in switch_mm() 74 static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next) in activate_mm() argument 79 switch_mm(prev, next, current); in activate_mm()
|
/arch/parisc/include/asm/ |
D | mmu_context.h | 52 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, struct task_struct *ts… in switch_mm() argument 55 if (prev != next) { in switch_mm() 56 mtctl(__pa(next->pgd), 25); in switch_mm() 57 load_context(next->context); in switch_mm() 63 static inline void activate_mm(struct mm_struct *prev, struct mm_struct *next) in activate_mm() argument 73 BUG_ON(next == &init_mm); /* Should never happen */ in activate_mm() 75 if (next->context == 0) in activate_mm() 76 next->context = alloc_sid(); in activate_mm() 78 switch_mm(prev,next,current); in activate_mm()
|
/arch/mn10300/kernel/ |
D | kprobes.c | 133 u8 opc, *pc, *sp, *next; in find_nextpc() local 135 next = 0; in find_nextpc() 144 next = pc + size; in find_nextpc() 152 next = pc + x8; in find_nextpc() 160 next = pc + x16; in find_nextpc() 168 next = pc + x32; in find_nextpc() 174 next = (u8 *)regs->mdr; in find_nextpc() 183 next = (u8 *)x32; in find_nextpc() 189 next = pc + 2; in find_nextpc() 195 next = (u8 *)regs->a0; in find_nextpc() [all …]
|
/arch/arm/include/asm/ |
D | mmu_context.h | 96 switch_mm(struct mm_struct *prev, struct mm_struct *next, in switch_mm() argument 104 if (!cpus_empty(next->cpu_vm_mask) && !cpu_isset(cpu, next->cpu_vm_mask)) in switch_mm() 107 if (!cpu_test_and_set(cpu, next->cpu_vm_mask) || prev != next) { in switch_mm() 108 check_context(next); in switch_mm() 109 cpu_switch_mm(next->pgd, next); in switch_mm() 117 #define activate_mm(prev,next) switch_mm(prev, next, NULL) argument
|
/arch/s390/kernel/ |
D | s390_ext.c | 47 p->next = ext_int_hash[index]; in register_external_interrupt() 62 p->next = ext_int_hash[index]; in register_early_external_interrupt() 79 p = p->next; in unregister_external_interrupt() 84 q->next = p->next; in unregister_external_interrupt() 86 ext_int_hash[index] = p->next; in unregister_external_interrupt() 103 if (q->next == p) in unregister_early_external_interrupt() 105 q = q->next; in unregister_early_external_interrupt() 109 q->next = p->next; in unregister_early_external_interrupt() 111 ext_int_hash[index] = p->next; in unregister_early_external_interrupt() 129 for (p = ext_int_hash[index]; p; p = p->next) { in do_extint()
|
/arch/sh/include/asm/ |
D | system_64.h | 22 struct task_struct *next, 25 #define switch_to(prev,next,last) \ argument 27 if (last_task_used_math != next) { \ 28 struct pt_regs *regs = next->thread.uregs; \ 31 last = sh64_switch_to(prev, &prev->thread, next, \ 32 &next->thread); \
|
D | mmu_context.h | 114 struct mm_struct *next, in switch_mm() argument 119 if (likely(prev != next)) { in switch_mm() 120 cpu_set(cpu, next->cpu_vm_mask); in switch_mm() 121 set_TTB(next->pgd); in switch_mm() 122 activate_context(next, cpu); in switch_mm() 124 if (!cpu_test_and_set(cpu, next->cpu_vm_mask)) in switch_mm() 125 activate_context(next, cpu); in switch_mm() 138 #define switch_mm(prev,next,tsk) do { } while (0) argument 141 #define activate_mm(prev, next) switch_mm((prev),(next),NULL) argument
|
/arch/xtensa/include/asm/ |
D | mmu_context.h | 96 activate_mm(struct mm_struct *prev, struct mm_struct *next) in activate_mm() argument 100 __get_new_mmu_context(next); in activate_mm() 101 __load_mmu_context(next); in activate_mm() 105 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, in switch_mm() argument 112 if (next->context == NO_CONTEXT || ((next->context^asid) & ~ASID_MASK)) in switch_mm() 113 __get_new_mmu_context(next); in switch_mm() 115 __load_mmu_context(next); in switch_mm()
|
/arch/alpha/kernel/ |
D | err_common.c | 158 struct el_subpacket *next = NULL; in el_process_subpacket_reg() local 161 for (; h && h->class != header->class; h = h->next); in el_process_subpacket_reg() 162 if (h) next = h->handler(header); in el_process_subpacket_reg() 164 return next; in el_process_subpacket_reg() 196 struct el_subpacket *next = NULL; in el_process_subpacket() local 203 next = el_process_header_subpacket(header); in el_process_subpacket() 206 if (NULL == (next = el_process_subpacket_reg(header))) { in el_process_subpacket() 215 return next; in el_process_subpacket() 224 for (a = subpacket_annotation_list; a; a = a->next) { in el_annotate_subpacket() 285 for (; a->next != NULL; a = a->next) { in cdl_register_subpacket_annotation() [all …]
|
/arch/powerpc/boot/dtc-src/ |
D | livetree.c | 34 new->next = NULL; in build_property() 43 assert(first->next == NULL); in chain_property() 45 first->next = list; in chain_property() 53 struct property *next; in reverse_properties() local 56 next = p->next; in reverse_properties() 57 p->next = head; in reverse_properties() 59 p = next; in reverse_properties() 104 prop->next = NULL; in add_property() 108 p = &((*p)->next); in add_property() 135 new->next = NULL; in build_reserve_entry() [all …]
|
/arch/mips/include/asm/ |
D | mmu_context.h | 129 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, in switch_mm() argument 145 if ((cpu_context(cpu, next) ^ asid_cache(cpu)) & ASID_VERSION_MASK) in switch_mm() 146 get_new_mmu_context(next, cpu); in switch_mm() 168 | (cpu_context(cpu, next) & ASID_MASK)); in switch_mm() 172 write_c0_entryhi(cpu_context(cpu, next)); in switch_mm() 174 TLBMISS_HANDLER_SETUP_PGD(next->pgd); in switch_mm() 181 cpu_set(cpu, next->cpu_vm_mask); in switch_mm() 201 activate_mm(struct mm_struct *prev, struct mm_struct *next) in activate_mm() argument 215 get_new_mmu_context(next, cpu); in activate_mm() 228 (cpu_context(cpu, next) & ASID_MASK)); in activate_mm() [all …]
|
/arch/ia64/include/asm/ |
D | mmu_context.h | 36 unsigned int next; /* next context number to use */ member 91 if (ia64_ctx.next >= ia64_ctx.limit) { in get_mmu_context() 92 ia64_ctx.next = find_next_zero_bit(ia64_ctx.bitmap, in get_mmu_context() 93 ia64_ctx.max_ctx, ia64_ctx.next); in get_mmu_context() 95 ia64_ctx.max_ctx, ia64_ctx.next); in get_mmu_context() 96 if (ia64_ctx.next >= ia64_ctx.max_ctx) in get_mmu_context() 99 mm->context = context = ia64_ctx.next++; in get_mmu_context() 185 activate_mm (struct mm_struct *prev, struct mm_struct *next) in activate_mm() argument 191 ia64_set_kr(IA64_KR_PT_BASE, __pa(next->pgd)); in activate_mm() 192 activate_context(next); in activate_mm()
|
D | system.h | 221 extern void ia64_account_on_switch (struct task_struct *prev, struct task_struct *next); 238 #define __switch_to(prev,next,last) do { \ argument 239 IA64_ACCOUNT_ON_SWITCH(prev, next); \ 242 if (IA64_HAS_EXTRA_STATE(next)) \ 243 ia64_load_extra(next); \ 244 ia64_psr(task_pt_regs(next))->dfh = !ia64_is_local_fpu_owner(next); \ 245 (last) = ia64_switch_to((next)); \ 255 # define switch_to(prev,next,last) do { \ argument 261 __switch_to(prev, next, last); \ 271 # define switch_to(prev,next,last) __switch_to(prev, next, last) argument
|
/arch/sh/mm/ |
D | fault_64.c | 51 unsigned long long next; in __do_tlb_refill() local 78 next = tlbp->next; in __do_tlb_refill() 79 __flush_tlb_slot(next); in __do_tlb_refill() 82 : : "r" (next), "r" (pteh), "r" (ptel) ); in __do_tlb_refill() 84 next += TLB_STEP; in __do_tlb_refill() 85 if (next > tlbp->last) next = tlbp->first; in __do_tlb_refill() 86 tlbp->next = next; in __do_tlb_refill()
|
/arch/x86/mm/ |
D | gup.c | 143 unsigned long next; in gup_pmd_range() local 150 next = pmd_addr_end(addr, end); in gup_pmd_range() 154 if (!gup_huge_pmd(pmd, addr, next, write, pages, nr)) in gup_pmd_range() 157 if (!gup_pte_range(pmd, addr, next, write, pages, nr)) in gup_pmd_range() 160 } while (pmdp++, addr = next, addr != end); in gup_pmd_range() 200 unsigned long next; in gup_pud_range() local 207 next = pud_addr_end(addr, end); in gup_pud_range() 211 if (!gup_huge_pud(pud, addr, next, write, pages, nr)) in gup_pud_range() 214 if (!gup_pmd_range(pud, addr, next, write, pages, nr)) in gup_pud_range() 217 } while (pudp++, addr = next, addr != end); in gup_pud_range() [all …]
|
/arch/um/include/asm/ |
D | mmu_context.h | 32 static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, in switch_mm() argument 37 if(prev != next){ in switch_mm() 39 cpu_set(cpu, next->cpu_vm_mask); in switch_mm() 40 if(next != &init_mm) in switch_mm() 41 __switch_mm(&next->context.id); in switch_mm()
|
D | system.h | 6 extern void *switch_to(void *prev, void *next, void *last); 32 extern void *_switch_to(void *prev, void *next, void *last); 33 #define switch_to(prev, next, last) prev = _switch_to(prev, next, last) argument
|
/arch/avr32/include/asm/ |
D | mmu_context.h | 109 struct mm_struct *next, in switch_mm() argument 112 if (likely(prev != next)) { in switch_mm() 113 unsigned long __pgdir = (unsigned long)next->pgd; in switch_mm() 116 activate_context(next); in switch_mm() 122 #define activate_mm(prev, next) switch_mm((prev), (next), NULL) argument
|
/arch/powerpc/mm/ |
D | gup.c | 110 unsigned long next; in gup_pmd_range() local 117 next = pmd_addr_end(addr, end); in gup_pmd_range() 120 if (!gup_pte_range(pmd, addr, next, write, pages, nr)) in gup_pmd_range() 122 } while (pmdp++, addr = next, addr != end); in gup_pmd_range() 130 unsigned long next; in gup_pud_range() local 137 next = pud_addr_end(addr, end); in gup_pud_range() 140 if (!gup_pmd_range(pud, addr, next, write, pages, nr)) in gup_pud_range() 142 } while (pudp++, addr = next, addr != end); in gup_pud_range() 152 unsigned long next; in get_user_pages_fast() local 242 next = pgd_addr_end(addr, end); in get_user_pages_fast() [all …]
|
D | mmu_context_nohash.c | 173 void switch_mmu_context(struct mm_struct *prev, struct mm_struct *next) in switch_mmu_context() argument 183 cpu, next, next->context.active, next->context.id); in switch_mmu_context() 188 next->context.active++; in switch_mmu_context() 200 id = next->context.id; in switch_mmu_context() 231 context_mm[id] = next; in switch_mmu_context() 232 next->context.id = id; in switch_mmu_context() 247 cpu, id, next); in switch_mmu_context() 248 local_flush_tlb_mm(next); in switch_mmu_context() 255 set_context(id, next->pgd); in switch_mmu_context()
|
/arch/x86/kernel/ |
D | process_64.c | 479 struct thread_struct *prev, *next; in __switch_to_xtra() local 482 next = &next_p->thread; in __switch_to_xtra() 487 else if (next->debugctlmsr != prev->debugctlmsr) in __switch_to_xtra() 488 update_debugctlmsr(next->debugctlmsr); in __switch_to_xtra() 491 loaddebug(next, 0); in __switch_to_xtra() 492 loaddebug(next, 1); in __switch_to_xtra() 493 loaddebug(next, 2); in __switch_to_xtra() 494 loaddebug(next, 3); in __switch_to_xtra() 496 loaddebug(next, 6); in __switch_to_xtra() 497 loaddebug(next, 7); in __switch_to_xtra() [all …]
|