/arch/mips/include/asm/ |
D | asmmacro-32.h | 15 .macro fpu_save_single thread tmp=t0 19 s.d $f0, THREAD_FPR0(\thread) 20 s.d $f2, THREAD_FPR2(\thread) 21 s.d $f4, THREAD_FPR4(\thread) 22 s.d $f6, THREAD_FPR6(\thread) 23 s.d $f8, THREAD_FPR8(\thread) 24 s.d $f10, THREAD_FPR10(\thread) 25 s.d $f12, THREAD_FPR12(\thread) 26 s.d $f14, THREAD_FPR14(\thread) 27 s.d $f16, THREAD_FPR16(\thread) [all …]
|
D | asmmacro-64.h | 16 .macro cpu_save_nonscratch thread 17 LONG_S s0, THREAD_REG16(\thread) 18 LONG_S s1, THREAD_REG17(\thread) 19 LONG_S s2, THREAD_REG18(\thread) 20 LONG_S s3, THREAD_REG19(\thread) 21 LONG_S s4, THREAD_REG20(\thread) 22 LONG_S s5, THREAD_REG21(\thread) 23 LONG_S s6, THREAD_REG22(\thread) 24 LONG_S s7, THREAD_REG23(\thread) 25 LONG_S sp, THREAD_REG29(\thread) [all …]
|
D | dsp.h | 45 tsk->thread.dsp.dspr[0] = mfhi1(); \ 46 tsk->thread.dsp.dspr[1] = mflo1(); \ 47 tsk->thread.dsp.dspr[2] = mfhi2(); \ 48 tsk->thread.dsp.dspr[3] = mflo2(); \ 49 tsk->thread.dsp.dspr[4] = mfhi3(); \ 50 tsk->thread.dsp.dspr[5] = mflo3(); \ 51 tsk->thread.dsp.dspcontrol = rddsp(DSP_MASK); \ 62 mthi1(tsk->thread.dsp.dspr[0]); \ 63 mtlo1(tsk->thread.dsp.dspr[1]); \ 64 mthi2(tsk->thread.dsp.dspr[2]); \ [all …]
|
D | asmmacro.h | 84 .macro fpu_save_16even thread tmp=t0 88 sdc1 $f0, THREAD_FPR0(\thread) 89 sdc1 $f2, THREAD_FPR2(\thread) 90 sdc1 $f4, THREAD_FPR4(\thread) 91 sdc1 $f6, THREAD_FPR6(\thread) 92 sdc1 $f8, THREAD_FPR8(\thread) 93 sdc1 $f10, THREAD_FPR10(\thread) 94 sdc1 $f12, THREAD_FPR12(\thread) 95 sdc1 $f14, THREAD_FPR14(\thread) 96 sdc1 $f16, THREAD_FPR16(\thread) [all …]
|
/arch/score/kernel/ |
D | ptrace.c | 182 child->thread.single_step = 1; in user_enable_single_step() 183 child->thread.ss_nextcnt = 1; in user_enable_single_step() 193 child->thread.ss_nextcnt = 2; in user_enable_single_step() 200 child->thread.ss_nextcnt = 2; in user_enable_single_step() 213 child->thread.ss_nextcnt = 2; in user_enable_single_step() 222 child->thread.ss_nextcnt = 2; in user_enable_single_step() 230 if (child->thread.ss_nextcnt == 1) { in user_enable_single_step() 242 child->thread.insn1_type = 0; in user_enable_single_step() 243 child->thread.addr1 = epc; in user_enable_single_step() 245 child->thread.insn1 = (short)epc_insn; in user_enable_single_step() [all …]
|
/arch/metag/kernel/ |
D | process.c | 191 memset(&tsk->thread.kernel_context, 0, in copy_thread() 192 sizeof(tsk->thread.kernel_context)); in copy_thread() 194 tsk->thread.kernel_context = __TBISwitchInit(kernel_context, in copy_thread() 211 tsk->thread.int_depth = 2; in copy_thread() 225 tsk->thread.int_depth = 1; in copy_thread() 232 tsk->thread.tls_ptr = in copy_thread() 236 if (tsk->thread.fpu_context) { in copy_thread() 239 ctx = kmemdup(tsk->thread.fpu_context, in copy_thread() 241 tsk->thread.fpu_context = ctx; in copy_thread() 246 if (tsk->thread.dsp_context) { in copy_thread() [all …]
|
/arch/parisc/kernel/ |
D | asm-offsets.c | 66 DEFINE(TASK_REGS, offsetof(struct task_struct, thread.regs)); in main() 67 DEFINE(TASK_PT_PSW, offsetof(struct task_struct, thread.regs.gr[ 0])); in main() 68 DEFINE(TASK_PT_GR1, offsetof(struct task_struct, thread.regs.gr[ 1])); in main() 69 DEFINE(TASK_PT_GR2, offsetof(struct task_struct, thread.regs.gr[ 2])); in main() 70 DEFINE(TASK_PT_GR3, offsetof(struct task_struct, thread.regs.gr[ 3])); in main() 71 DEFINE(TASK_PT_GR4, offsetof(struct task_struct, thread.regs.gr[ 4])); in main() 72 DEFINE(TASK_PT_GR5, offsetof(struct task_struct, thread.regs.gr[ 5])); in main() 73 DEFINE(TASK_PT_GR6, offsetof(struct task_struct, thread.regs.gr[ 6])); in main() 74 DEFINE(TASK_PT_GR7, offsetof(struct task_struct, thread.regs.gr[ 7])); in main() 75 DEFINE(TASK_PT_GR8, offsetof(struct task_struct, thread.regs.gr[ 8])); in main() [all …]
|
/arch/powerpc/kernel/ |
D | process.c | 85 if (tsk == current && tsk->thread.regs && in check_if_tm_restore_required() 86 MSR_TM_ACTIVE(tsk->thread.regs->msr) && in check_if_tm_restore_required() 88 tsk->thread.ckpt_regs.msr = tsk->thread.regs->msr; in check_if_tm_restore_required() 155 msr = tsk->thread.regs->msr; in __giveup_fpu() 161 tsk->thread.regs->msr = msr; in __giveup_fpu() 180 if (tsk->thread.regs) { in flush_fp_to_thread() 190 if (tsk->thread.regs->msr & MSR_FP) { in flush_fp_to_thread() 214 if (current->thread.regs && (current->thread.regs->msr & MSR_FP)) { in enable_kernel_fp() 223 if(!msr_tm_active(cpumsr) && msr_tm_active(current->thread.regs->msr)) in enable_kernel_fp() 231 if (tsk->thread.load_fp || msr_tm_active(tsk->thread.regs->msr)) { in restore_fp() [all …]
|
D | ptrace.c | 141 tm_save_sprs(&(tsk->thread)); in flush_tmregs_to_thread() 205 return task->thread.regs->msr | task->thread.fpexc_mode; in get_user_msr() 210 task->thread.regs->msr &= ~MSR_DEBUGCHANGE; in set_user_msr() 211 task->thread.regs->msr |= msr & MSR_DEBUGCHANGE; in set_user_msr() 218 return task->thread.ckpt_regs.msr | task->thread.fpexc_mode; in get_user_ckpt_msr() 223 task->thread.ckpt_regs.msr &= ~MSR_DEBUGCHANGE; in set_user_ckpt_msr() 224 task->thread.ckpt_regs.msr |= msr & MSR_DEBUGCHANGE; in set_user_ckpt_msr() 230 task->thread.ckpt_regs.trap = trap & 0xfff0; in set_user_ckpt_trap() 238 *data = task->thread.dscr; in get_user_dscr() 244 task->thread.dscr = dscr; in set_user_dscr() [all …]
|
/arch/um/kernel/ |
D | process.c | 85 to->thread.prev_sched = from; in __switch_to() 88 switch_threads(&from->thread.switch_buf, &to->thread.switch_buf); in __switch_to() 91 return current->thread.prev_sched; in __switch_to() 96 struct pt_regs *regs = ¤t->thread.regs; in interrupt_end() 120 if (current->thread.prev_sched != NULL) in new_thread_handler() 121 schedule_tail(current->thread.prev_sched); in new_thread_handler() 122 current->thread.prev_sched = NULL; in new_thread_handler() 124 fn = current->thread.request.u.thread.proc; in new_thread_handler() 125 arg = current->thread.request.u.thread.arg; in new_thread_handler() 131 userspace(¤t->thread.regs.regs); in new_thread_handler() [all …]
|
/arch/sh/kernel/cpu/sh4/ |
D | fpu.c | 88 :"0"((char *)(&tsk->thread.xstate->hardfpu.status)), in save_fpu() 138 :"0" (tsk->thread.xstate), "r" (FPSCR_RCHG) in restore_fpu() 234 if ((tsk->thread.xstate->hardfpu.fpscr & FPSCR_CAUSE_ERROR)) in ieee_fpe_handler() 236 denormal_to_double(&tsk->thread.xstate->hardfpu, in ieee_fpe_handler() 252 hx = tsk->thread.xstate->hardfpu.fp_regs[n]; in ieee_fpe_handler() 253 hy = tsk->thread.xstate->hardfpu.fp_regs[m]; in ieee_fpe_handler() 254 fpscr = tsk->thread.xstate->hardfpu.fpscr; in ieee_fpe_handler() 264 | tsk->thread.xstate->hardfpu.fp_regs[n + 1]; in ieee_fpe_handler() 266 | tsk->thread.xstate->hardfpu.fp_regs[m + 1]; in ieee_fpe_handler() 268 tsk->thread.xstate->hardfpu.fp_regs[n] = llx >> 32; in ieee_fpe_handler() [all …]
|
/arch/sparc/kernel/ |
D | sigutil_32.c | 20 fpsave(¤t->thread.float_regs[0], ¤t->thread.fsr, in save_fpu_state() 21 ¤t->thread.fpqueue[0], ¤t->thread.fpqdepth); in save_fpu_state() 28 fpsave(¤t->thread.float_regs[0], ¤t->thread.fsr, in save_fpu_state() 29 ¤t->thread.fpqueue[0], ¤t->thread.fpqdepth); in save_fpu_state() 35 ¤t->thread.float_regs[0], in save_fpu_state() 37 err |= __put_user(current->thread.fsr, &fpu->si_fsr); in save_fpu_state() 38 err |= __put_user(current->thread.fpqdepth, &fpu->si_fpqdepth); in save_fpu_state() 39 if (current->thread.fpqdepth != 0) in save_fpu_state() 41 ¤t->thread.fpqueue[0], in save_fpu_state() 70 err = __copy_from_user(¤t->thread.float_regs[0], &fpu->si_float_regs[0], in restore_fpu_state() [all …]
|
D | process_32.c | 196 fpsave(&tsk->thread.float_regs[0], &tsk->thread.fsr, 197 &tsk->thread.fpqueue[0], &tsk->thread.fpqdepth); 217 fpsave(¤t->thread.float_regs[0], ¤t->thread.fsr, 218 ¤t->thread.fpqueue[0], ¤t->thread.fpqdepth); 227 if (current->thread.flags & SPARC_FLAG_KTHREAD) { 228 current->thread.flags &= ~SPARC_FLAG_KTHREAD; 232 current->thread.kregs = (struct pt_regs *) 321 fpsave(&p->thread.float_regs[0], &p->thread.fsr, 322 &p->thread.fpqueue[0], &p->thread.fpqdepth); 343 p->thread.kregs = childregs; [all …]
|
/arch/m68k/kernel/ |
D | process.c | 46 struct switch_stack *sw = (struct switch_stack *)tsk->thread.ksp; in thread_saved_pc() 105 current->thread.fs = __USER_DS; in flush_thread() 142 p->thread.ksp = (unsigned long)frame; in copy_thread() 143 p->thread.esp0 = (unsigned long)&frame->regs; in copy_thread() 149 p->thread.fs = get_fs().seg; in copy_thread() 158 p->thread.usp = 0; in copy_thread() 165 p->thread.usp = usp ?: rdusp(); in copy_thread() 173 asm volatile ("fsave %0" : : "m" (p->thread.fpstate[0]) : "memory"); in copy_thread() 175 if (!CPU_IS_060 ? p->thread.fpstate[0] : p->thread.fpstate[2]) { in copy_thread() 182 : "m" (p->thread.fp[0]), in copy_thread() [all …]
|
/arch/mn10300/kernel/ |
D | fpu.c | 52 info.si_addr = (void *) tsk->thread.uregs->pc; in fpu_exception() 57 fpcr = tsk->thread.fpu_state.fpcr; in fpu_exception() 88 if (tsk->thread.fpu_flags & THREAD_HAS_FPU) { in fpu_setup_sigcontext() 89 fpu_save(&tsk->thread.fpu_state); in fpu_setup_sigcontext() 90 tsk->thread.uregs->epsw &= ~EPSW_FE; in fpu_setup_sigcontext() 91 tsk->thread.fpu_flags &= ~THREAD_HAS_FPU; in fpu_setup_sigcontext() 95 fpu_save(&tsk->thread.fpu_state); in fpu_setup_sigcontext() 96 fpu_state_owner->thread.uregs->epsw &= ~EPSW_FE; in fpu_setup_sigcontext() 108 &tsk->thread.fpu_state, in fpu_setup_sigcontext() 125 if (tsk->thread.fpu_flags & THREAD_HAS_FPU) { in fpu_kill_state() [all …]
|
/arch/frv/kernel/ |
D | process.c | 140 p->thread.frame = childregs; in copy_thread() 141 p->thread.curr = p; in copy_thread() 142 p->thread.sp = (unsigned long) childregs; in copy_thread() 143 p->thread.fp = 0; in copy_thread() 144 p->thread.lr = 0; in copy_thread() 145 p->thread.frame0 = childregs; in copy_thread() 150 p->thread.pc = (unsigned long) ret_from_kernel_thread; in copy_thread() 151 save_user_regs(p->thread.user); in copy_thread() 158 p->thread.pc = (unsigned long) ret_from_fork; in copy_thread() 164 save_user_regs(p->thread.user); in copy_thread() [all …]
|
/arch/s390/kernel/ |
D | process.c | 57 sf = (struct stack_frame *) tsk->thread.ksp; in thread_saved_pc() 98 dst->thread.fpu.regs = dst->thread.fpu.fprs; in arch_dup_task_struct() 113 p->thread.ksp = (unsigned long) frame; in copy_thread() 115 save_access_regs(&p->thread.acrs[0]); in copy_thread() 117 p->thread.mm_segment = get_fs(); in copy_thread() 119 memset(&p->thread.per_user, 0, sizeof(p->thread.per_user)); in copy_thread() 120 memset(&p->thread.per_event, 0, sizeof(p->thread.per_event)); in copy_thread() 122 p->thread.per_flags = 0; in copy_thread() 156 p->thread.ri_cb = NULL; in copy_thread() 163 p->thread.acrs[0] = (unsigned int)tls; in copy_thread() [all …]
|
/arch/mips/kernel/ |
D | asm-offsets.c | 113 OFFSET(THREAD_REG16, task_struct, thread.reg16); in output_thread_defines() 114 OFFSET(THREAD_REG17, task_struct, thread.reg17); in output_thread_defines() 115 OFFSET(THREAD_REG18, task_struct, thread.reg18); in output_thread_defines() 116 OFFSET(THREAD_REG19, task_struct, thread.reg19); in output_thread_defines() 117 OFFSET(THREAD_REG20, task_struct, thread.reg20); in output_thread_defines() 118 OFFSET(THREAD_REG21, task_struct, thread.reg21); in output_thread_defines() 119 OFFSET(THREAD_REG22, task_struct, thread.reg22); in output_thread_defines() 120 OFFSET(THREAD_REG23, task_struct, thread.reg23); in output_thread_defines() 121 OFFSET(THREAD_REG29, task_struct, thread.reg29); in output_thread_defines() 122 OFFSET(THREAD_REG30, task_struct, thread.reg30); in output_thread_defines() [all …]
|
/arch/mn10300/include/asm/ |
D | switch_to.h | 24 if ((prev)->thread.fpu_flags & THREAD_HAS_FPU) { \ 25 (prev)->thread.fpu_flags &= ~THREAD_HAS_FPU; \ 26 (prev)->thread.uregs->epsw &= ~EPSW_FE; \ 27 fpu_save(&(prev)->thread.fpu_state); \ 43 current->thread.wchan = (u_long) __builtin_return_address(0); \ 44 (last) = __switch_to(&(prev)->thread, &(next)->thread, (prev)); \ 46 current->thread.wchan = 0; \
|
D | fpu.h | 41 : "i"(THREAD_USING_FPU), "a"(&tsk->thread.fpu_flags) in set_using_fpu() 50 : "i"(THREAD_USING_FPU), "a"(&tsk->thread.fpu_flags) in clear_using_fpu() 54 #define is_using_fpu(tsk) ((tsk)->thread.fpu_flags & THREAD_USING_FPU) 67 if (tsk->thread.fpu_flags & THREAD_HAS_FPU) { in unlazy_fpu() 68 fpu_save(&tsk->thread.fpu_state); in unlazy_fpu() 69 tsk->thread.fpu_flags &= ~THREAD_HAS_FPU; in unlazy_fpu() 70 tsk->thread.uregs->epsw &= ~EPSW_FE; in unlazy_fpu() 74 fpu_save(&tsk->thread.fpu_state); in unlazy_fpu() 95 if (tsk->thread.fpu_flags & THREAD_HAS_FPU) { in flush_fpu() 96 tsk->thread.fpu_flags &= ~THREAD_HAS_FPU; in flush_fpu() [all …]
|
/arch/sh/kernel/ |
D | process.c | 28 if (src->thread.xstate) { in arch_dup_task_struct() 29 dst->thread.xstate = kmem_cache_alloc(task_xstate_cachep, in arch_dup_task_struct() 31 if (!dst->thread.xstate) in arch_dup_task_struct() 33 memcpy(dst->thread.xstate, src->thread.xstate, xstate_size); in arch_dup_task_struct() 41 if (tsk->thread.xstate) { in free_thread_xstate() 42 kmem_cache_free(task_xstate_cachep, tsk->thread.xstate); in free_thread_xstate() 43 tsk->thread.xstate = NULL; in free_thread_xstate()
|
/arch/m68k/mm/ |
D | fault.c | 25 siginfo.si_signo = current->thread.signo; in send_fault_sig() 26 siginfo.si_code = current->thread.code; in send_fault_sig() 27 siginfo.si_addr = (void *)current->thread.faddr; in send_fault_sig() 196 current->thread.signo = SIGBUS; in do_page_fault() 197 current->thread.faddr = address; in do_page_fault() 201 current->thread.signo = SIGBUS; in do_page_fault() 202 current->thread.code = BUS_ADRERR; in do_page_fault() 203 current->thread.faddr = address; in do_page_fault() 207 current->thread.signo = SIGSEGV; in do_page_fault() 208 current->thread.code = SEGV_MAPERR; in do_page_fault() [all …]
|
/arch/x86/kernel/ |
D | process_64.c | 183 prev_p->thread.fsbase = 0; in save_base_legacy() 185 prev_p->thread.gsbase = 0; in save_base_legacy() 191 savesegment(fs, task->thread.fsindex); in save_fsgs() 192 savesegment(gs, task->thread.gsindex); in save_fsgs() 193 save_base_legacy(task, task->thread.fsindex, FS); in save_fsgs() 194 save_base_legacy(task, task->thread.gsindex, GS); in save_fsgs() 265 p->thread.sp0 = (unsigned long)task_stack_page(p) + THREAD_SIZE; in copy_thread_tls() 271 p->thread.sp = (unsigned long) fork_frame; in copy_thread_tls() 272 p->thread.io_bitmap_ptr = NULL; in copy_thread_tls() 274 savesegment(gs, p->thread.gsindex); in copy_thread_tls() [all …]
|
/arch/powerpc/include/asm/ |
D | switch_to.h | 82 t->thread.ebbrr = 0; in clear_task_ebb() 83 t->thread.ebbhr = 0; in clear_task_ebb() 84 t->thread.bescr = 0; in clear_task_ebb() 85 t->thread.mmcr2 = 0; in clear_task_ebb() 86 t->thread.mmcr0 = 0; in clear_task_ebb() 87 t->thread.siar = 0; in clear_task_ebb() 88 t->thread.sdar = 0; in clear_task_ebb() 89 t->thread.sier = 0; in clear_task_ebb() 90 t->thread.used_ebb = 0; in clear_task_ebb()
|
/arch/arm/vfp/ |
D | vfpmodule.c | 65 static bool vfp_state_in_hw(unsigned int cpu, struct thread_info *thread) in vfp_state_in_hw() argument 68 if (thread->vfpstate.hard.cpu != cpu) in vfp_state_in_hw() 71 return vfp_current_hw_state[cpu] == &thread->vfpstate; in vfp_state_in_hw() 79 static void vfp_force_reload(unsigned int cpu, struct thread_info *thread) in vfp_force_reload() argument 81 if (vfp_state_in_hw(cpu, thread)) { in vfp_force_reload() 86 thread->vfpstate.hard.cpu = NR_CPUS; in vfp_force_reload() 93 static void vfp_thread_flush(struct thread_info *thread) in vfp_thread_flush() argument 95 union vfp_state *vfp = &thread->vfpstate; in vfp_thread_flush() 121 static void vfp_thread_exit(struct thread_info *thread) in vfp_thread_exit() argument 124 union vfp_state *vfp = &thread->vfpstate; in vfp_thread_exit() [all …]
|