/arch/x86/kernel/ |
D | espfix_64.c | 139 void *stack_page; in init_espfix_ap() local 150 stack_page = READ_ONCE(espfix_pages[page]); in init_espfix_ap() 151 if (likely(stack_page)) in init_espfix_ap() 157 stack_page = READ_ONCE(espfix_pages[page]); in init_espfix_ap() 158 if (stack_page) in init_espfix_ap() 189 stack_page = page_address(alloc_pages_node(node, GFP_KERNEL, 0)); in init_espfix_ap() 194 pte = __pte(__pa(stack_page) | ((__PAGE_KERNEL_RO | _PAGE_ENC) & ptemask)); in init_espfix_ap() 199 WRITE_ONCE(espfix_pages[page], stack_page); in init_espfix_ap() 205 per_cpu(espfix_waddr, cpu) = (unsigned long)stack_page in init_espfix_ap()
|
/arch/mips/kernel/ |
D | perf_event.c | 51 unsigned long stack_page = in perf_callchain_kernel() local 53 if (stack_page && sp >= stack_page && in perf_callchain_kernel() 54 sp <= stack_page + THREAD_SIZE - 32) in perf_callchain_kernel()
|
D | stacktrace.c | 46 unsigned long stack_page = in save_context_stack() local 48 if (stack_page && sp >= stack_page && in save_context_stack() 49 sp <= stack_page + THREAD_SIZE - 32) in save_context_stack()
|
D | process.c | 527 unsigned long notrace unwind_stack_by_address(unsigned long stack_page, in unwind_stack_by_address() argument 538 if (!stack_page) in unwind_stack_by_address() 545 low = stack_page; in unwind_stack_by_address() 547 high = stack_page + IRQ_STACK_START; in unwind_stack_by_address() 550 high = stack_page + THREAD_SIZE - 32; in unwind_stack_by_address() 622 unsigned long stack_page = 0; in unwind_stack() local 627 stack_page = (unsigned long)irq_stack[cpu]; in unwind_stack() 632 if (!stack_page) in unwind_stack() 633 stack_page = (unsigned long)task_stack_page(task); in unwind_stack() 635 return unwind_stack_by_address(stack_page, sp, pc, ra); in unwind_stack()
|
/arch/hexagon/kernel/ |
D | process.c | 136 unsigned long stack_page; in get_wchan() local 141 stack_page = (unsigned long)task_stack_page(p); in get_wchan() 144 if (fp < (stack_page + sizeof(struct thread_info)) || in get_wchan() 145 fp >= (THREAD_SIZE - 8 + stack_page)) in get_wchan()
|
/arch/h8300/kernel/ |
D | process.c | 134 unsigned long stack_page; in get_wchan() local 140 stack_page = (unsigned long)p; in get_wchan() 143 if (fp < stack_page+sizeof(struct thread_info) || in get_wchan() 144 fp >= 8184+stack_page) in get_wchan()
|
/arch/m68k/kernel/ |
D | process.c | 269 unsigned long stack_page; in get_wchan() local 274 stack_page = (unsigned long)task_stack_page(p); in get_wchan() 277 if (fp < stack_page+sizeof(struct thread_info) || in get_wchan() 278 fp >= 8184+stack_page) in get_wchan()
|
/arch/um/kernel/ |
D | process.c | 372 unsigned long stack_page, sp, ip; in get_wchan() local 378 stack_page = (unsigned long) task_stack_page(p); in get_wchan() 380 if (stack_page == 0) in get_wchan() 388 if (sp < stack_page) in get_wchan() 391 while (sp < stack_page + THREAD_SIZE) { in get_wchan()
|
/arch/nios2/kernel/ |
D | process.c | 223 unsigned long stack_page; in get_wchan() local 229 stack_page = (unsigned long)p; in get_wchan() 232 if (fp < stack_page+sizeof(struct task_struct) || in get_wchan() 233 fp >= 8184+stack_page) /* ;dgt2;tmp */ in get_wchan()
|
/arch/xtensa/kernel/ |
D | process.c | 304 unsigned long stack_page = (unsigned long) task_stack_page(p); in get_wchan() local 314 if (sp < stack_page + sizeof(struct task_struct) || in get_wchan() 315 sp >= (stack_page + THREAD_SIZE) || in get_wchan()
|
/arch/powerpc/kernel/ |
D | stacktrace.c | 105 unsigned long stack_page = (unsigned long)task_stack_page(tsk); in __save_stack_trace_tsk_reliable() local 110 stack_end = stack_page + THREAD_SIZE; in __save_stack_trace_tsk_reliable() 140 if (sp < stack_page + sizeof(struct thread_struct) || in __save_stack_trace_tsk_reliable()
|
D | process.c | 2039 unsigned long stack_page; in valid_irq_stack() local 2042 stack_page = (unsigned long)hardirq_ctx[cpu]; in valid_irq_stack() 2043 if (sp >= stack_page && sp <= stack_page + THREAD_SIZE - nbytes) in valid_irq_stack() 2046 stack_page = (unsigned long)softirq_ctx[cpu]; in valid_irq_stack() 2047 if (sp >= stack_page && sp <= stack_page + THREAD_SIZE - nbytes) in valid_irq_stack() 2057 unsigned long stack_page; in valid_emergency_stack() local 2060 stack_page = (unsigned long)paca_ptrs[cpu]->emergency_sp - THREAD_SIZE; in valid_emergency_stack() 2061 if (sp >= stack_page && sp <= stack_page + THREAD_SIZE - nbytes) in valid_emergency_stack() 2065 stack_page = (unsigned long)paca_ptrs[cpu]->nmi_emergency_sp - THREAD_SIZE; in valid_emergency_stack() 2066 if (sp >= stack_page && sp <= stack_page + THREAD_SIZE - nbytes) in valid_emergency_stack() [all …]
|
/arch/arm/kernel/ |
D | process.c | 287 unsigned long stack_page; in get_wchan() local 296 stack_page = (unsigned long)task_stack_page(p); in get_wchan() 298 if (frame.sp < stack_page || in get_wchan() 299 frame.sp >= stack_page + THREAD_SIZE || in get_wchan()
|
/arch/mips/include/asm/ |
D | stacktrace.h | 13 extern unsigned long unwind_stack_by_address(unsigned long stack_page,
|
/arch/arm64/kernel/ |
D | process.c | 596 unsigned long stack_page, ret = 0; in get_wchan() local 601 stack_page = (unsigned long)try_get_task_stack(p); in get_wchan() 602 if (!stack_page) in get_wchan()
|
/arch/arm64/kvm/ |
D | arm.c | 1794 unsigned long stack_page; in init_hyp_mode() local 1796 stack_page = __get_free_page(GFP_KERNEL); in init_hyp_mode() 1797 if (!stack_page) { in init_hyp_mode() 1802 per_cpu(kvm_arm_hyp_stack_page, cpu) = stack_page; in init_hyp_mode() 1870 char *stack_page = (char *)per_cpu(kvm_arm_hyp_stack_page, cpu); in init_hyp_mode() local 1871 err = create_hyp_mappings(stack_page, stack_page + PAGE_SIZE, in init_hyp_mode()
|