/arch/x86/kernel/ |
D | espfix_64.c | 139 void *stack_page; in init_espfix_ap() local 150 stack_page = READ_ONCE(espfix_pages[page]); in init_espfix_ap() 151 if (likely(stack_page)) in init_espfix_ap() 157 stack_page = READ_ONCE(espfix_pages[page]); in init_espfix_ap() 158 if (stack_page) in init_espfix_ap() 189 stack_page = page_address(alloc_pages_node(node, GFP_KERNEL, 0)); in init_espfix_ap() 194 pte = __pte(__pa(stack_page) | ((__PAGE_KERNEL_RO | _PAGE_ENC) & ptemask)); in init_espfix_ap() 199 WRITE_ONCE(espfix_pages[page], stack_page); in init_espfix_ap() 205 per_cpu(espfix_waddr, cpu) = (unsigned long)stack_page in init_espfix_ap()
|
/arch/mips/kernel/ |
D | perf_event.c | 51 unsigned long stack_page = in perf_callchain_kernel() local 53 if (stack_page && sp >= stack_page && in perf_callchain_kernel() 54 sp <= stack_page + THREAD_SIZE - 32) in perf_callchain_kernel()
|
D | stacktrace.c | 46 unsigned long stack_page = in save_context_stack() local 48 if (stack_page && sp >= stack_page && in save_context_stack() 49 sp <= stack_page + THREAD_SIZE - 32) in save_context_stack()
|
D | process.c | 542 unsigned long notrace unwind_stack_by_address(unsigned long stack_page, in unwind_stack_by_address() argument 553 if (!stack_page) in unwind_stack_by_address() 560 low = stack_page; in unwind_stack_by_address() 562 high = stack_page + IRQ_STACK_START; in unwind_stack_by_address() 565 high = stack_page + THREAD_SIZE - 32; in unwind_stack_by_address() 637 unsigned long stack_page = 0; in unwind_stack() local 642 stack_page = (unsigned long)irq_stack[cpu]; in unwind_stack() 647 if (!stack_page) in unwind_stack() 648 stack_page = (unsigned long)task_stack_page(task); in unwind_stack() 650 return unwind_stack_by_address(stack_page, sp, pc, ra); in unwind_stack()
|
/arch/hexagon/kernel/ |
D | process.c | 136 unsigned long stack_page; in get_wchan() local 141 stack_page = (unsigned long)task_stack_page(p); in get_wchan() 144 if (fp < (stack_page + sizeof(struct thread_info)) || in get_wchan() 145 fp >= (THREAD_SIZE - 8 + stack_page)) in get_wchan()
|
/arch/h8300/kernel/ |
D | process.c | 134 unsigned long stack_page; in get_wchan() local 140 stack_page = (unsigned long)p; in get_wchan() 143 if (fp < stack_page+sizeof(struct thread_info) || in get_wchan() 144 fp >= 8184+stack_page) in get_wchan()
|
/arch/m68k/kernel/ |
D | process.c | 269 unsigned long stack_page; in get_wchan() local 274 stack_page = (unsigned long)task_stack_page(p); in get_wchan() 277 if (fp < stack_page+sizeof(struct thread_info) || in get_wchan() 278 fp >= 8184+stack_page) in get_wchan()
|
/arch/powerpc/kernel/ |
D | stacktrace.c | 70 unsigned long stack_page = (unsigned long)task_stack_page(task); in arch_stack_walk_reliable() local 75 stack_end = stack_page + THREAD_SIZE; in arch_stack_walk_reliable() 105 if (sp < stack_page + sizeof(struct thread_struct) || in arch_stack_walk_reliable()
|
D | process.c | 2052 unsigned long stack_page; in valid_irq_stack() local 2055 stack_page = (unsigned long)hardirq_ctx[cpu]; in valid_irq_stack() 2056 if (sp >= stack_page && sp <= stack_page + THREAD_SIZE - nbytes) in valid_irq_stack() 2059 stack_page = (unsigned long)softirq_ctx[cpu]; in valid_irq_stack() 2060 if (sp >= stack_page && sp <= stack_page + THREAD_SIZE - nbytes) in valid_irq_stack() 2070 unsigned long stack_page; in valid_emergency_stack() local 2076 stack_page = (unsigned long)paca_ptrs[cpu]->emergency_sp - THREAD_SIZE; in valid_emergency_stack() 2077 if (sp >= stack_page && sp <= stack_page + THREAD_SIZE - nbytes) in valid_emergency_stack() 2081 stack_page = (unsigned long)paca_ptrs[cpu]->nmi_emergency_sp - THREAD_SIZE; in valid_emergency_stack() 2082 if (sp >= stack_page && sp <= stack_page + THREAD_SIZE - nbytes) in valid_emergency_stack() [all …]
|
/arch/um/kernel/ |
D | process.c | 369 unsigned long stack_page, sp, ip; in get_wchan() local 375 stack_page = (unsigned long) task_stack_page(p); in get_wchan() 377 if (stack_page == 0) in get_wchan() 385 if (sp < stack_page) in get_wchan() 388 while (sp < stack_page + THREAD_SIZE) { in get_wchan()
|
/arch/nios2/kernel/ |
D | process.c | 223 unsigned long stack_page; in get_wchan() local 229 stack_page = (unsigned long)p; in get_wchan() 232 if (fp < stack_page+sizeof(struct task_struct) || in get_wchan() 233 fp >= 8184+stack_page) /* ;dgt2;tmp */ in get_wchan()
|
/arch/xtensa/kernel/ |
D | process.c | 304 unsigned long stack_page = (unsigned long) task_stack_page(p); in get_wchan() local 314 if (sp < stack_page + sizeof(struct task_struct) || in get_wchan() 315 sp >= (stack_page + THREAD_SIZE) || in get_wchan()
|
/arch/arm/kernel/ |
D | process.c | 282 unsigned long stack_page; in get_wchan() local 291 stack_page = (unsigned long)task_stack_page(p); in get_wchan() 293 if (frame.sp < stack_page || in get_wchan() 294 frame.sp >= stack_page + THREAD_SIZE || in get_wchan()
|
/arch/mips/include/asm/ |
D | stacktrace.h | 13 extern unsigned long unwind_stack_by_address(unsigned long stack_page,
|
/arch/arm64/kernel/ |
D | process.c | 537 unsigned long stack_page, ret = 0; in get_wchan() local 542 stack_page = (unsigned long)try_get_task_stack(p); in get_wchan() 543 if (!stack_page) in get_wchan()
|
/arch/arm64/kvm/ |
D | arm.c | 2015 unsigned long stack_page; in init_hyp_mode() local 2017 stack_page = __get_free_page(GFP_KERNEL); in init_hyp_mode() 2018 if (!stack_page) { in init_hyp_mode() 2023 per_cpu(kvm_arm_hyp_stack_page, cpu) = stack_page; in init_hyp_mode() 2098 char *stack_page = (char *)per_cpu(kvm_arm_hyp_stack_page, cpu); in init_hyp_mode() local 2099 err = create_hyp_mappings(stack_page, stack_page + PAGE_SIZE, in init_hyp_mode()
|