Searched refs:stack (Results 1 – 10 of 10) sorted by relevance
/kernel/ |
D | fork.c | 218 void *stack; in alloc_thread_stack_node() local 233 tsk->stack = s->addr; in alloc_thread_stack_node() 242 stack = __vmalloc_node_range(THREAD_SIZE, THREAD_ALIGN, in alloc_thread_stack_node() 253 if (stack) { in alloc_thread_stack_node() 254 tsk->stack_vm_area = find_vm_area(stack); in alloc_thread_stack_node() 255 tsk->stack = stack; in alloc_thread_stack_node() 257 return stack; in alloc_thread_stack_node() 263 tsk->stack = page_address(page); in alloc_thread_stack_node() 264 return tsk->stack; in alloc_thread_stack_node() 294 vfree_atomic(tsk->stack); in free_thread_stack() [all …]
|
D | Makefile | 33 CFLAGS_kcov.o := $(call cc-option, -fno-conserve-stack -fno-stack-protector)
|
/kernel/trace/ |
D | trace_stack.c | 155 static void check_stack(unsigned long ip, unsigned long *stack) in check_stack() argument 162 this_size = ((unsigned long)stack) & (THREAD_SIZE-1); in check_stack() 171 if (!object_is_on_stack(stack)) in check_stack() 212 start = stack; in check_stack() 249 tracer_frame = (p - stack) * in check_stack() 295 unsigned long stack; in stack_trace_call() local 310 check_stack(ip, &stack); in stack_trace_call()
|
D | Kconfig | 170 address on the current task structure into a stack of calls. 411 bool "Trace max stack" 417 This special tracer records the maximum stack footprint of the 421 kernel executes, and keeping a maximum stack depth value and 422 stack-trace saved. If this is configured with DYNAMIC_FTRACE 423 then it will not have any overhead while the stack tracer 426 To enable the stack tracer on bootup, pass in 'stacktrace' 429 The stack tracer can also be enabled or disabled via the
|
/kernel/bpf/ |
D | verifier.c | 507 if (state->stack[i].slot_type[j] != STACK_INVALID) in print_verifier_state() 510 state->stack[i].slot_type[j]]; in print_verifier_state() 516 print_liveness(env, state->stack[i].spilled_ptr.live); in print_verifier_state() 517 if (state->stack[i].slot_type[0] == STACK_SPILL) { in print_verifier_state() 518 reg = &state->stack[i].spilled_ptr; in print_verifier_state() 556 COPY_STATE_FN(stack, allocated_stack, stack, BPF_REG_SIZE) in COPY_STATE_FN() argument 596 REALLOC_STATE_FN(stack, allocated_stack, stack, BPF_REG_SIZE) 672 kfree(state->stack); in free_func_state() 1662 if (func->stack[j].slot_type[0] != STACK_SPILL) in mark_all_scalars_precise() 1664 reg = &func->stack[j].spilled_ptr; in mark_all_scalars_precise() [all …]
|
D | core.c | 1311 static u64 __no_fgcse ___bpf_prog_run(u64 *regs, const struct bpf_insn *insn, u64 *stack) in ___bpf_prog_run() argument 1579 u64 stack[stack_size / sizeof(u64)]; \ 1582 FP = (u64) (unsigned long) &stack[ARRAY_SIZE(stack)]; \ 1584 return ___bpf_prog_run(regs, insn, stack); \ 1592 u64 stack[stack_size / sizeof(u64)]; \ 1595 FP = (u64) (unsigned long) &stack[ARRAY_SIZE(stack)]; \ 1601 return ___bpf_prog_run(regs, insn, stack); \
|
D | btf.c | 252 struct resolve_vertex stack[MAX_RESOLVE_DEPTH]; member 1008 v = &env->stack[env->top_stack++]; in env_stack_push() 1026 env->stack[env->top_stack - 1].next_member = next_member; in env_stack_set_next_member() 1033 u32 type_id = env->stack[--(env->top_stack)].type_id; in env_stack_pop_resolved() 1043 return env->top_stack ? &env->stack[env->top_stack - 1] : NULL; in env_stack_peak()
|
/kernel/debug/kdb/ |
D | kdb_cmds | 3 # processes. Commands that require process data (including stack or
|
/kernel/dma/ |
D | debug.c | 1111 addr = (u8 *)current->stack + i * PAGE_SIZE + offset; in check_for_stack()
|
/kernel/events/ |
D | uprobes.c | 1915 ri->stack = user_stack_pointer(regs); in prepare_uretprobe()
|