Searched refs:stack (Results 1 – 10 of 10) sorted by relevance
/kernel/ |
D | fork.c | 218 void *stack; in alloc_thread_stack_node() local 233 tsk->stack = s->addr; in alloc_thread_stack_node() 242 stack = __vmalloc_node_range(THREAD_SIZE, THREAD_ALIGN, in alloc_thread_stack_node() 253 if (stack) { in alloc_thread_stack_node() 254 tsk->stack_vm_area = find_vm_area(stack); in alloc_thread_stack_node() 255 tsk->stack = stack; in alloc_thread_stack_node() 257 return stack; in alloc_thread_stack_node() 263 tsk->stack = page_address(page); in alloc_thread_stack_node() 264 return tsk->stack; in alloc_thread_stack_node() 294 vfree_atomic(tsk->stack); in free_thread_stack() [all …]
|
D | Makefile | 33 CFLAGS_kcov.o := $(call cc-option, -fno-conserve-stack -fno-stack-protector)
|
/kernel/trace/ |
D | trace_stack.c | 155 static void check_stack(unsigned long ip, unsigned long *stack) in check_stack() argument 162 this_size = ((unsigned long)stack) & (THREAD_SIZE-1); in check_stack() 171 if (!object_is_on_stack(stack)) in check_stack() 212 start = stack; in check_stack() 249 tracer_frame = (p - stack) * in check_stack() 295 unsigned long stack; in stack_trace_call() local 310 check_stack(ip, &stack); in stack_trace_call()
|
D | Kconfig | 170 address on the current task structure into a stack of calls. 410 bool "Trace max stack" 416 This special tracer records the maximum stack footprint of the 420 kernel executes, and keeping a maximum stack depth value and 421 stack-trace saved. If this is configured with DYNAMIC_FTRACE 422 then it will not have any overhead while the stack tracer 425 To enable the stack tracer on bootup, pass in 'stacktrace' 428 The stack tracer can also be enabled or disabled via the
|
/kernel/bpf/ |
D | verifier.c | 506 if (state->stack[i].slot_type[j] != STACK_INVALID) in print_verifier_state() 509 state->stack[i].slot_type[j]]; in print_verifier_state() 515 print_liveness(env, state->stack[i].spilled_ptr.live); in print_verifier_state() 516 if (state->stack[i].slot_type[0] == STACK_SPILL) { in print_verifier_state() 517 reg = &state->stack[i].spilled_ptr; in print_verifier_state() 555 COPY_STATE_FN(stack, allocated_stack, stack, BPF_REG_SIZE) in COPY_STATE_FN() argument 595 REALLOC_STATE_FN(stack, allocated_stack, stack, BPF_REG_SIZE) 671 kfree(state->stack); in free_func_state() 1674 if (func->stack[j].slot_type[0] != STACK_SPILL) in mark_all_scalars_precise() 1676 reg = &func->stack[j].spilled_ptr; in mark_all_scalars_precise() [all …]
|
D | core.c | 1317 static u64 ___bpf_prog_run(u64 *regs, const struct bpf_insn *insn, u64 *stack) in ___bpf_prog_run() argument 1623 u64 stack[stack_size / sizeof(u64)]; \ 1626 FP = (u64) (unsigned long) &stack[ARRAY_SIZE(stack)]; \ 1628 return ___bpf_prog_run(regs, insn, stack); \ 1636 u64 stack[stack_size / sizeof(u64)]; \ 1639 FP = (u64) (unsigned long) &stack[ARRAY_SIZE(stack)]; \ 1645 return ___bpf_prog_run(regs, insn, stack); \
|
D | btf.c | 252 struct resolve_vertex stack[MAX_RESOLVE_DEPTH]; member 1008 v = &env->stack[env->top_stack++]; in env_stack_push() 1026 env->stack[env->top_stack - 1].next_member = next_member; in env_stack_set_next_member() 1033 u32 type_id = env->stack[--(env->top_stack)].type_id; in env_stack_pop_resolved() 1043 return env->top_stack ? &env->stack[env->top_stack - 1] : NULL; in env_stack_peak()
|
/kernel/debug/kdb/ |
D | kdb_cmds | 3 # processes. Commands that require process data (including stack or
|
/kernel/dma/ |
D | debug.c | 1115 addr = (u8 *)current->stack + i * PAGE_SIZE + offset; in check_for_stack()
|
/kernel/events/ |
D | uprobes.c | 1920 ri->stack = user_stack_pointer(regs); in prepare_uretprobe()
|