/arch/sparc/kernel/ |
D | sigutil_32.c | 18 if (test_tsk_thread_flag(current, TIF_USEDFPU)) { in save_fpu_state() 20 fpsave(¤t->thread.float_regs[0], ¤t->thread.fsr, in save_fpu_state() 21 ¤t->thread.fpqueue[0], ¤t->thread.fpqdepth); in save_fpu_state() 23 clear_tsk_thread_flag(current, TIF_USEDFPU); in save_fpu_state() 26 if (current == last_task_used_math) { in save_fpu_state() 28 fpsave(¤t->thread.float_regs[0], ¤t->thread.fsr, in save_fpu_state() 29 ¤t->thread.fpqueue[0], ¤t->thread.fpqdepth); in save_fpu_state() 35 ¤t->thread.float_regs[0], in save_fpu_state() 37 err |= __put_user(current->thread.fsr, &fpu->si_fsr); in save_fpu_state() 38 err |= __put_user(current->thread.fpqdepth, &fpu->si_fpqdepth); in save_fpu_state() [all …]
|
D | process_32.c | 155 tsk = current; in show_stack() 157 if (tsk == current && !_ksp) in show_stack() 211 if(last_task_used_math == current) { 217 fpsave(¤t->thread.float_regs[0], ¤t->thread.fsr, 218 ¤t->thread.fpqueue[0], ¤t->thread.fpqdepth); 227 if (current->thread.flags & SPARC_FLAG_KTHREAD) { 228 current->thread.flags &= ~SPARC_FLAG_KTHREAD; 232 current->thread.kregs = (struct pt_regs *) 233 (task_stack_page(current) + (THREAD_SIZE - TRACEREG_SZ)); 316 if(last_task_used_math == current) { [all …]
|
/arch/s390/math-emu/ |
D | math.c | 110 instr, current->comm, current->pid); in display_emulation_not_implemented() 152 mode = current->thread.fp_regs.fpc & 3; in emu_axbr() 153 cvt.w.high = current->thread.fp_regs.fprs[rx].ui; in emu_axbr() 154 cvt.w.low = current->thread.fp_regs.fprs[rx+2].ui; in emu_axbr() 156 cvt.w.high = current->thread.fp_regs.fprs[ry].ui; in emu_axbr() 157 cvt.w.low = current->thread.fp_regs.fprs[ry+2].ui; in emu_axbr() 161 current->thread.fp_regs.fprs[rx].ui = cvt.w.high; in emu_axbr() 162 current->thread.fp_regs.fprs[rx+2].ui = cvt.w.low; in emu_axbr() 173 mode = current->thread.fp_regs.fpc & 3; in emu_adbr() 174 FP_UNPACK_DP(DA, ¤t->thread.fp_regs.fprs[rx].d); in emu_adbr() [all …]
|
/arch/m68k/mm/ |
D | fault.c | 25 siginfo.si_signo = current->thread.signo; in send_fault_sig() 26 siginfo.si_code = current->thread.code; in send_fault_sig() 27 siginfo.si_addr = (void *)current->thread.faddr; in send_fault_sig() 33 &siginfo, current); in send_fault_sig() 72 struct mm_struct *mm = current->mm; in do_page_fault() 142 if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) in do_page_fault() 162 current->maj_flt++; in do_page_fault() 164 current->min_flt++; in do_page_fault() 196 current->thread.signo = SIGBUS; in do_page_fault() 197 current->thread.faddr = address; in do_page_fault() [all …]
|
/arch/parisc/kernel/ |
D | ftrace.c | 29 if (!current->ret_stack) in push_return_trace() 33 if (current->curr_ret_stack == FTRACE_RETFUNC_DEPTH - 1) { in push_return_trace() 34 atomic_inc(¤t->trace_overrun); in push_return_trace() 38 index = ++current->curr_ret_stack; in push_return_trace() 40 current->ret_stack[index].ret = ret; in push_return_trace() 41 current->ret_stack[index].func = func; in push_return_trace() 42 current->ret_stack[index].calltime = time; in push_return_trace() 53 index = current->curr_ret_stack; in pop_return_trace() 64 *ret = current->ret_stack[index].ret; in pop_return_trace() 65 trace->func = current->ret_stack[index].func; in pop_return_trace() [all …]
|
/arch/powerpc/kernel/ |
D | signal_64.c | 102 if (current->thread.used_vr) { in setup_sigcontext() 103 flush_altivec_to_thread(current); in setup_sigcontext() 105 err |= __copy_to_user(v_regs, ¤t->thread.vr_state, in setup_sigcontext() 116 current->thread.vrsave = mfspr(SPRN_VRSAVE); in setup_sigcontext() 117 err |= __put_user(current->thread.vrsave, (u32 __user *)&v_regs[33]); in setup_sigcontext() 121 flush_fp_to_thread(current); in setup_sigcontext() 123 err |= copy_fpr_to_user(&sc->fp_regs, current); in setup_sigcontext() 136 if (current->thread.used_vsr && ctx_has_vsx_region) { in setup_sigcontext() 137 __giveup_vsx(current); in setup_sigcontext() 139 err |= copy_vsx_to_user(v_regs, current); in setup_sigcontext() [all …]
|
D | uprobes.c | 66 struct arch_uprobe_task *autask = ¤t->utask->autask; in arch_uprobe_pre_xol() 68 autask->saved_trap_nr = current->thread.trap_nr; in arch_uprobe_pre_xol() 69 current->thread.trap_nr = UPROBE_TRAP_NR; in arch_uprobe_pre_xol() 70 regs->nip = current->utask->xol_vaddr; in arch_uprobe_pre_xol() 72 user_enable_single_step(current); in arch_uprobe_pre_xol() 114 struct uprobe_task *utask = current->utask; in arch_uprobe_post_xol() 116 WARN_ON_ONCE(current->thread.trap_nr != UPROBE_TRAP_NR); in arch_uprobe_post_xol() 118 current->thread.trap_nr = utask->autask.saved_trap_nr; in arch_uprobe_post_xol() 129 user_disable_single_step(current); in arch_uprobe_post_xol() 169 struct uprobe_task *utask = current->utask; in arch_uprobe_abort_xol() [all …]
|
D | signal_32.c | 415 flush_fp_to_thread(current); in save_user_regs() 423 if (current->thread.used_vr) { in save_user_regs() 424 flush_altivec_to_thread(current); in save_user_regs() 425 if (__copy_to_user(&frame->mc_vregs, ¤t->thread.vr_state, in save_user_regs() 441 current->thread.vrsave = mfspr(SPRN_VRSAVE); in save_user_regs() 442 if (__put_user(current->thread.vrsave, (u32 __user *)&frame->mc_vregs[32])) in save_user_regs() 445 if (copy_fpr_to_user(&frame->mc_fregs, current)) in save_user_regs() 460 if (current->thread.used_vsr && ctx_has_vsx_region) { in save_user_regs() 461 __giveup_vsx(current); in save_user_regs() 462 if (copy_vsx_to_user(&frame->mc_vsregs, current)) in save_user_regs() [all …]
|
D | swsusp.c | 23 flush_fp_to_thread(current); in save_processor_state() 24 flush_altivec_to_thread(current); in save_processor_state() 25 flush_spe_to_thread(current); in save_processor_state() 36 switch_mmu_context(current->active_mm, current->active_mm); in restore_processor_state()
|
D | process.c | 86 if (tsk == current && tsk->thread.regs && in giveup_fpu_maybe_transactional() 104 if (tsk == current && tsk->thread.regs && in giveup_altivec_maybe_transactional() 145 BUG_ON(tsk != current); in flush_fp_to_thread() 160 if (current->thread.regs && (current->thread.regs->msr & MSR_FP)) in enable_kernel_fp() 161 giveup_fpu_maybe_transactional(current); in enable_kernel_fp() 176 if (current->thread.regs && (current->thread.regs->msr & MSR_VEC)) in enable_kernel_altivec() 177 giveup_altivec_maybe_transactional(current); in enable_kernel_altivec() 196 BUG_ON(tsk != current); in flush_altivec_to_thread() 214 if (current->thread.regs && (current->thread.regs->msr & MSR_VSX)) 215 giveup_vsx(current); [all …]
|
/arch/s390/kernel/ |
D | runtime_instr.c | 28 struct pt_regs *regs = task_pt_regs(current); in disable_runtime_instr() 54 struct task_struct *task = current; in exit_thread_runtime_instr() 74 if (!current->thread.ri_cb) in runtime_instr_int_handler() 76 if (current->thread.ri_signum < SIGRTMIN || in runtime_instr_int_handler() 77 current->thread.ri_signum > SIGRTMAX) { in runtime_instr_int_handler() 83 info.si_signo = current->thread.ri_signum; in runtime_instr_int_handler() 92 send_sig_info(current->thread.ri_signum, &info, current); in runtime_instr_int_handler() 113 if (!current->thread.ri_cb) { in SYSCALL_DEFINE2() 118 cb = current->thread.ri_cb; in SYSCALL_DEFINE2() 123 current->thread.ri_signum = signum; in SYSCALL_DEFINE2() [all …]
|
D | signal.c | 107 save_access_regs(current->thread.acrs); in store_sigregs() 108 save_fp_ctl(¤t->thread.fp_regs.fpc); in store_sigregs() 110 if (current->thread.vxrs) { in store_sigregs() 113 save_vx_regs(current->thread.vxrs); in store_sigregs() 115 current->thread.fp_regs.fprs[i] = in store_sigregs() 116 *(freg_t *)(current->thread.vxrs + i); in store_sigregs() 119 save_fp_regs(current->thread.fp_regs.fprs); in store_sigregs() 125 restore_access_regs(current->thread.acrs); in load_sigregs() 128 if (current->thread.vxrs) { in load_sigregs() 132 *(freg_t *)(current->thread.vxrs + i) = in load_sigregs() [all …]
|
D | compat_signal.c | 171 save_access_regs(current->thread.acrs); in store_sigregs() 172 save_fp_ctl(¤t->thread.fp_regs.fpc); in store_sigregs() 173 if (current->thread.vxrs) { in store_sigregs() 174 save_vx_regs(current->thread.vxrs); in store_sigregs() 176 current->thread.fp_regs.fprs[i] = in store_sigregs() 177 *(freg_t *)(current->thread.vxrs + i); in store_sigregs() 179 save_fp_regs(current->thread.fp_regs.fprs); in store_sigregs() 187 restore_access_regs(current->thread.acrs); in load_sigregs() 189 if (current->thread.vxrs) { in load_sigregs() 191 *(freg_t *)(current->thread.vxrs + i) = in load_sigregs() [all …]
|
/arch/um/kernel/ |
D | trap.c | 26 struct mm_struct *mm = current->mm; in handle_page_fault() 77 if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) in handle_page_fault() 93 current->maj_flt++; in handle_page_fault() 95 current->min_flt++; in handle_page_fault() 142 struct task_struct *tsk = current; in show_segv_info() 168 current->thread.arch.faultinfo = fi; in bad_segv() 169 force_sig_info(SIGSEGV, &si, current); in bad_segv() 174 force_sigsegv(SIGSEGV, current); in fatal_sigsegv() 212 current->thread.segv_regs = container_of(regs, struct pt_regs, regs); in segv() 218 else if (current->mm == NULL) { in segv() [all …]
|
D | exec.c | 25 arch_flush_thread(¤t->thread.arch); in flush_thread() 27 ret = unmap(¤t->mm->context.id, 0, STUB_START, 0, &data); in flush_thread() 28 ret = ret || unmap(¤t->mm->context.id, STUB_END, in flush_thread() 33 force_sig(SIGKILL, current); in flush_thread() 38 __switch_mm(¤t->mm->context.id); in flush_thread() 45 current->ptrace &= ~PT_DTRACE; in start_thread()
|
D | signal.c | 28 if ((current->ptrace & PT_DTRACE) && (current->ptrace & PT_PTRACED)) in handle_signal() 55 sp = current->sas_ss_sp + current->sas_ss_size; in handle_signal() 103 if (current->ptrace & PT_DTRACE) in kern_do_signal() 104 current->thread.singlestep_syscall = in kern_do_signal() 105 is_syscall(PT_REGS_IP(¤t->thread.regs)); in kern_do_signal() 118 return kern_do_signal(¤t->thread.regs); in do_signal()
|
D | process.c | 86 arch_switch_to(current); in __switch_to() 88 return current->thread.prev_sched; in __switch_to() 98 tracehook_notify_resume(¤t->thread.regs); in interrupt_end() 103 return task_pid_nr(current); in get_current_pid() 115 if (current->thread.prev_sched != NULL) in new_thread_handler() 116 schedule_tail(current->thread.prev_sched); in new_thread_handler() 117 current->thread.prev_sched = NULL; in new_thread_handler() 119 fn = current->thread.request.u.thread.proc; in new_thread_handler() 120 arg = current->thread.request.u.thread.arg; in new_thread_handler() 126 userspace(¤t->thread.regs.regs); in new_thread_handler() [all …]
|
/arch/score/kernel/ |
D | traps.c | 147 current->comm, current->pid, (unsigned long) current); in show_registers() 178 force_sig(SIGBUS, current); in do_adelinsn() 192 force_sig(SIGBUS, current); in do_adedata() 198 force_sig(SIGFPE, current); in do_pel() 204 force_sig(SIGFPE, current); in do_cee() 210 force_sig(SIGFPE, current); in do_cpe() 216 force_sig(SIGBUS, current); in do_be() 229 force_sig_info(SIGFPE, &info, current); in do_ov() 235 force_sig(SIGTRAP, current); in do_tr() 243 read_tsk_long(current, epc, &epc_insn); in do_ri() [all …]
|
/arch/mips/include/asm/ |
D | fpu.h | 135 KSTK_STATUS(current) |= ST0_CU1; in __own_fpu() 137 KSTK_STATUS(current) |= ST0_FR; in __own_fpu() 139 KSTK_STATUS(current) &= ~ST0_FR; in __own_fpu() 152 _restore_fp(current); in own_fpu_inatomic() 172 save_msa(current); in lose_fpu() 173 current->thread.fpu.fcr31 = in lose_fpu() 181 _save_fp(current); in lose_fpu() 184 KSTK_STATUS(current) &= ~ST0_CU1; in lose_fpu() 191 unsigned int fcr31 = current->thread.fpu.fcr31; in init_fpu() 241 if (tsk == current) { in get_fpu_regs() [all …]
|
/arch/mips/math-emu/ |
D | dsemul.c | 74 mm_context_t *mm_ctx = ¤t->mm->context; in alloc_emuframe() 116 pr_debug("allocate emuframe %d to %d\n", idx, current->pid); in alloc_emuframe() 128 pr_debug("free emuframe %d from %d\n", idx, current->pid); in free_emuframe() 180 fr_idx = atomic_read(¤t->thread.bd_emu_frame); in dsemul_thread_rollback() 193 regs->cp0_epc = current->thread.bd_emu_branch_pc; in dsemul_thread_rollback() 195 regs->cp0_epc = current->thread.bd_emu_cont_pc; in dsemul_thread_rollback() 197 atomic_set(¤t->thread.bd_emu_frame, BD_EMUFRAME_NONE); in dsemul_thread_rollback() 198 free_emuframe(fr_idx, current->mm); in dsemul_thread_rollback() 243 fr_idx = atomic_read(¤t->thread.bd_emu_frame); in mips_dsemul() 263 free_emuframe(fr_idx, current->mm); in mips_dsemul() [all …]
|
/arch/arm64/kernel/ |
D | fpsimd.c | 125 send_sig_info(SIGFPE, &info, current); in do_fpsimd_exc() 135 if (current->mm && !test_thread_flag(TIF_FOREIGN_FPSTATE)) in fpsimd_thread_switch() 136 fpsimd_save_state(¤t->thread.fpsimd_state); in fpsimd_thread_switch() 161 memset(¤t->thread.fpsimd_state, 0, sizeof(struct fpsimd_state)); in fpsimd_flush_thread() 162 fpsimd_flush_task_state(current); in fpsimd_flush_thread() 175 fpsimd_save_state(¤t->thread.fpsimd_state); in fpsimd_preserve_current_state() 188 struct fpsimd_state *st = ¤t->thread.fpsimd_state; in fpsimd_restore_current_state() 207 struct fpsimd_state *st = ¤t->thread.fpsimd_state; in fpsimd_update_current_state() 247 if (current->mm && in kernel_neon_begin_partial() 249 fpsimd_save_state(¤t->thread.fpsimd_state); in kernel_neon_begin_partial() [all …]
|
/arch/xtensa/mm/ |
D | fault.c | 40 struct mm_struct *mm = current->mm; in do_page_fault() 71 printk("[%s:%d:%08x:%d:%08x:%s%s]\n", current->comm, current->pid, in do_page_fault() 114 if ((fault & VM_FAULT_RETRY) && fatal_signal_pending(current)) in do_page_fault() 128 current->maj_flt++; in do_page_fault() 130 current->min_flt++; in do_page_fault() 153 current->thread.bad_vaddr = address; in do_page_fault() 154 current->thread.error_code = is_write; in do_page_fault() 159 force_sig_info(SIGSEGV, &info, current); in do_page_fault() 183 current->thread.bad_vaddr = address; in do_page_fault() 188 force_sig_info(SIGBUS, &info, current); in do_page_fault() [all …]
|
/arch/mips/loongson/loongson-3/ |
D | cop2-ex.c | 37 KSTK_STATUS(current) |= (ST0_CU1 | ST0_CU2); in loongson_cu2_call() 39 KSTK_STATUS(current) |= ST0_FR; in loongson_cu2_call() 41 KSTK_STATUS(current) &= ~ST0_FR; in loongson_cu2_call() 46 _init_fpu(current->thread.fpu.fcr31); in loongson_cu2_call() 49 _restore_fp(current); in loongson_cu2_call()
|
/arch/ia64/include/asm/ |
D | switch_to.h | 68 if (unlikely((current->thread.flags & IA64_THREAD_MIGRATION) && \ 69 (task_cpu(current) != \ 70 task_thread_info(current)->last_cpu))) { \ 71 platform_migrate(current); \ 72 task_thread_info(current)->last_cpu = task_cpu(current); \
|
/arch/mips/power/ |
D | cpu.c | 22 save_fp(current); in save_processor_state() 24 save_dsp(current); in save_processor_state() 32 restore_fp(current); in restore_processor_state() 34 restore_dsp(current); in restore_processor_state()
|