Home
last modified time | relevance | path

Searched refs:thread (Results 1 – 25 of 385) sorted by relevance

12345678910>>...16

/arch/mips/include/asm/
Dasmmacro-32.h16 .macro fpu_save_single thread tmp=t0
20 s.d $f0, THREAD_FPR0(\thread)
21 s.d $f2, THREAD_FPR2(\thread)
22 s.d $f4, THREAD_FPR4(\thread)
23 s.d $f6, THREAD_FPR6(\thread)
24 s.d $f8, THREAD_FPR8(\thread)
25 s.d $f10, THREAD_FPR10(\thread)
26 s.d $f12, THREAD_FPR12(\thread)
27 s.d $f14, THREAD_FPR14(\thread)
28 s.d $f16, THREAD_FPR16(\thread)
[all …]
Dasmmacro-64.h17 .macro cpu_save_nonscratch thread
18 LONG_S s0, THREAD_REG16(\thread)
19 LONG_S s1, THREAD_REG17(\thread)
20 LONG_S s2, THREAD_REG18(\thread)
21 LONG_S s3, THREAD_REG19(\thread)
22 LONG_S s4, THREAD_REG20(\thread)
23 LONG_S s5, THREAD_REG21(\thread)
24 LONG_S s6, THREAD_REG22(\thread)
25 LONG_S s7, THREAD_REG23(\thread)
26 LONG_S sp, THREAD_REG29(\thread)
[all …]
Ddsp.h41 tsk->thread.dsp.dspr[0] = mfhi1(); \
42 tsk->thread.dsp.dspr[1] = mflo1(); \
43 tsk->thread.dsp.dspr[2] = mfhi2(); \
44 tsk->thread.dsp.dspr[3] = mflo2(); \
45 tsk->thread.dsp.dspr[4] = mfhi3(); \
46 tsk->thread.dsp.dspr[5] = mflo3(); \
47 tsk->thread.dsp.dspcontrol = rddsp(DSP_MASK); \
58 mthi1(tsk->thread.dsp.dspr[0]); \
59 mtlo1(tsk->thread.dsp.dspr[1]); \
60 mthi2(tsk->thread.dsp.dspr[2]); \
[all …]
Dasmmacro.h84 .macro fpu_save_16even thread tmp=t0
88 sdc1 $f0, THREAD_FPR0(\thread)
89 sdc1 $f2, THREAD_FPR2(\thread)
90 sdc1 $f4, THREAD_FPR4(\thread)
91 sdc1 $f6, THREAD_FPR6(\thread)
92 sdc1 $f8, THREAD_FPR8(\thread)
93 sdc1 $f10, THREAD_FPR10(\thread)
94 sdc1 $f12, THREAD_FPR12(\thread)
95 sdc1 $f14, THREAD_FPR14(\thread)
96 sdc1 $f16, THREAD_FPR16(\thread)
[all …]
/arch/riscv/kernel/
Dasm-offsets.c16 OFFSET(TASK_THREAD_RA, task_struct, thread.ra); in asm_offsets()
17 OFFSET(TASK_THREAD_SP, task_struct, thread.sp); in asm_offsets()
18 OFFSET(TASK_THREAD_S0, task_struct, thread.s[0]); in asm_offsets()
19 OFFSET(TASK_THREAD_S1, task_struct, thread.s[1]); in asm_offsets()
20 OFFSET(TASK_THREAD_S2, task_struct, thread.s[2]); in asm_offsets()
21 OFFSET(TASK_THREAD_S3, task_struct, thread.s[3]); in asm_offsets()
22 OFFSET(TASK_THREAD_S4, task_struct, thread.s[4]); in asm_offsets()
23 OFFSET(TASK_THREAD_S5, task_struct, thread.s[5]); in asm_offsets()
24 OFFSET(TASK_THREAD_S6, task_struct, thread.s[6]); in asm_offsets()
25 OFFSET(TASK_THREAD_S7, task_struct, thread.s[7]); in asm_offsets()
[all …]
/arch/parisc/kernel/
Dasm-offsets.c53 DEFINE(TASK_REGS, offsetof(struct task_struct, thread.regs)); in main()
54 DEFINE(TASK_PT_PSW, offsetof(struct task_struct, thread.regs.gr[ 0])); in main()
55 DEFINE(TASK_PT_GR1, offsetof(struct task_struct, thread.regs.gr[ 1])); in main()
56 DEFINE(TASK_PT_GR2, offsetof(struct task_struct, thread.regs.gr[ 2])); in main()
57 DEFINE(TASK_PT_GR3, offsetof(struct task_struct, thread.regs.gr[ 3])); in main()
58 DEFINE(TASK_PT_GR4, offsetof(struct task_struct, thread.regs.gr[ 4])); in main()
59 DEFINE(TASK_PT_GR5, offsetof(struct task_struct, thread.regs.gr[ 5])); in main()
60 DEFINE(TASK_PT_GR6, offsetof(struct task_struct, thread.regs.gr[ 6])); in main()
61 DEFINE(TASK_PT_GR7, offsetof(struct task_struct, thread.regs.gr[ 7])); in main()
62 DEFINE(TASK_PT_GR8, offsetof(struct task_struct, thread.regs.gr[ 8])); in main()
[all …]
/arch/s390/kernel/
Dguarded_storage.c17 kfree(tsk->thread.gs_cb); in guarded_storage_release()
18 kfree(tsk->thread.gs_bc_cb); in guarded_storage_release()
25 if (!current->thread.gs_cb) { in gs_enable()
33 current->thread.gs_cb = gs_cb; in gs_enable()
41 if (current->thread.gs_cb) { in gs_disable()
43 kfree(current->thread.gs_cb); in gs_disable()
44 current->thread.gs_cb = NULL; in gs_disable()
55 gs_cb = current->thread.gs_bc_cb; in gs_set_bc_cb()
60 current->thread.gs_bc_cb = gs_cb; in gs_set_bc_cb()
71 gs_cb = current->thread.gs_bc_cb; in gs_clear_bc_cb()
[all …]
Dprocess.c78 dst->thread.fpu.regs = dst->thread.fpu.fprs; in arch_dup_task_struct()
92 p->thread.ksp = (unsigned long) frame; in copy_thread_tls()
94 save_access_regs(&p->thread.acrs[0]); in copy_thread_tls()
96 p->thread.mm_segment = get_fs(); in copy_thread_tls()
98 memset(&p->thread.per_user, 0, sizeof(p->thread.per_user)); in copy_thread_tls()
99 memset(&p->thread.per_event, 0, sizeof(p->thread.per_event)); in copy_thread_tls()
101 p->thread.per_flags = 0; in copy_thread_tls()
103 p->thread.user_timer = 0; in copy_thread_tls()
104 p->thread.guest_timer = 0; in copy_thread_tls()
105 p->thread.system_timer = 0; in copy_thread_tls()
[all …]
Dptrace.c48 struct thread_struct *thread = &task->thread; in update_cr_regs() local
62 if (task->thread.per_flags & PER_FLAG_NO_TE) in update_cr_regs()
66 if (task->thread.per_flags & PER_FLAG_TE_ABORT_RAND) { in update_cr_regs()
67 if (task->thread.per_flags & PER_FLAG_TE_ABORT_RAND_TEND) in update_cr_regs()
76 if (task->thread.gs_cb) in update_cr_regs()
87 new.control = thread->per_user.control; in update_cr_regs()
88 new.start = thread->per_user.start; in update_cr_regs()
89 new.end = thread->per_user.end; in update_cr_regs()
142 memset(&task->thread.per_user, 0, sizeof(task->thread.per_user)); in ptrace_disable()
143 memset(&task->thread.per_event, 0, sizeof(task->thread.per_event)); in ptrace_disable()
[all …]
/arch/powerpc/kernel/
Dprocess.c96 if (tsk == current && tsk->thread.regs && in check_if_tm_restore_required()
97 MSR_TM_ACTIVE(tsk->thread.regs->msr) && in check_if_tm_restore_required()
99 tsk->thread.ckpt_regs.msr = tsk->thread.regs->msr; in check_if_tm_restore_required()
164 msr = tsk->thread.regs->msr; in __giveup_fpu()
170 tsk->thread.regs->msr = msr; in __giveup_fpu()
189 if (tsk->thread.regs) { in flush_fp_to_thread()
199 if (tsk->thread.regs->msr & MSR_FP) { in flush_fp_to_thread()
223 if (current->thread.regs && (current->thread.regs->msr & MSR_FP)) { in enable_kernel_fp()
233 MSR_TM_ACTIVE(current->thread.regs->msr)) in enable_kernel_fp()
242 if (tsk->thread.load_fp) { in restore_fp()
[all …]
Dptrace.c145 tm_save_sprs(&(tsk->thread)); in flush_tmregs_to_thread()
209 return task->thread.regs->msr | task->thread.fpexc_mode; in get_user_msr()
214 task->thread.regs->msr &= ~MSR_DEBUGCHANGE; in set_user_msr()
215 task->thread.regs->msr |= msr & MSR_DEBUGCHANGE; in set_user_msr()
222 return task->thread.ckpt_regs.msr | task->thread.fpexc_mode; in get_user_ckpt_msr()
227 task->thread.ckpt_regs.msr &= ~MSR_DEBUGCHANGE; in set_user_ckpt_msr()
228 task->thread.ckpt_regs.msr |= msr & MSR_DEBUGCHANGE; in set_user_ckpt_msr()
234 task->thread.ckpt_regs.trap = trap & 0xfff0; in set_user_ckpt_trap()
242 *data = task->thread.dscr; in get_user_dscr()
248 task->thread.dscr = dscr; in set_user_dscr()
[all …]
Dsignal.c116 check_syscall_restart(tsk->thread.regs, &ksig.ka, ksig.sig > 0); in do_signal()
121 tsk->thread.regs->trap = 0; in do_signal()
131 if (tsk->thread.hw_brk.address && tsk->thread.hw_brk.type) in do_signal()
132 __set_breakpoint(&tsk->thread.hw_brk); in do_signal()
135 thread_change_pc(tsk, tsk->thread.regs); in do_signal()
137 rseq_signal_deliver(&ksig, tsk->thread.regs); in do_signal()
148 tsk->thread.regs->trap = 0; in do_signal()
166 BUG_ON(regs != current->thread.regs); in do_notify_resume()
206 if (MSR_TM_ACTIVE(tsk->thread.regs->msr)) { in get_tm_stackpointer()
208 if (MSR_TM_TRANSACTIONAL(tsk->thread.regs->msr)) in get_tm_stackpointer()
[all …]
Dsignal_32.c247 buf[i] = task->thread.TS_FPR(i); in copy_fpr_to_user()
248 buf[i] = task->thread.fp_state.fpscr; in copy_fpr_to_user()
261 task->thread.TS_FPR(i) = buf[i]; in copy_fpr_from_user()
262 task->thread.fp_state.fpscr = buf[i]; in copy_fpr_from_user()
275 buf[i] = task->thread.fp_state.fpr[i][TS_VSRLOWOFFSET]; in copy_vsx_to_user()
288 task->thread.fp_state.fpr[i][TS_VSRLOWOFFSET] = buf[i]; in copy_vsx_from_user()
301 buf[i] = task->thread.TS_CKFPR(i); in copy_ckfpr_to_user()
302 buf[i] = task->thread.ckfp_state.fpscr; in copy_ckfpr_to_user()
315 task->thread.TS_CKFPR(i) = buf[i]; in copy_ckfpr_from_user()
316 task->thread.ckfp_state.fpscr = buf[i]; in copy_ckfpr_from_user()
[all …]
/arch/sh/kernel/cpu/sh4/
Dfpu.c84 :"0"((char *)(&tsk->thread.xstate->hardfpu.status)), in save_fpu()
134 :"0" (tsk->thread.xstate), "r" (FPSCR_RCHG) in restore_fpu()
230 if ((tsk->thread.xstate->hardfpu.fpscr & FPSCR_CAUSE_ERROR)) in ieee_fpe_handler()
232 denormal_to_double(&tsk->thread.xstate->hardfpu, in ieee_fpe_handler()
248 hx = tsk->thread.xstate->hardfpu.fp_regs[n]; in ieee_fpe_handler()
249 hy = tsk->thread.xstate->hardfpu.fp_regs[m]; in ieee_fpe_handler()
250 fpscr = tsk->thread.xstate->hardfpu.fpscr; in ieee_fpe_handler()
260 | tsk->thread.xstate->hardfpu.fp_regs[n + 1]; in ieee_fpe_handler()
262 | tsk->thread.xstate->hardfpu.fp_regs[m + 1]; in ieee_fpe_handler()
264 tsk->thread.xstate->hardfpu.fp_regs[n] = llx >> 32; in ieee_fpe_handler()
[all …]
/arch/um/kernel/
Dprocess.c88 to->thread.prev_sched = from; in __switch_to()
91 switch_threads(&from->thread.switch_buf, &to->thread.switch_buf); in __switch_to()
94 return current->thread.prev_sched; in __switch_to()
99 struct pt_regs *regs = &current->thread.regs; in interrupt_end()
123 if (current->thread.prev_sched != NULL) in new_thread_handler()
124 schedule_tail(current->thread.prev_sched); in new_thread_handler()
125 current->thread.prev_sched = NULL; in new_thread_handler()
127 fn = current->thread.request.u.thread.proc; in new_thread_handler()
128 arg = current->thread.request.u.thread.arg; in new_thread_handler()
134 userspace(&current->thread.regs.regs, current_thread_info()->aux_fp_regs); in new_thread_handler()
[all …]
/arch/sparc/kernel/
Dsigutil_32.c21 fpsave(&current->thread.float_regs[0], &current->thread.fsr, in save_fpu_state()
22 &current->thread.fpqueue[0], &current->thread.fpqdepth); in save_fpu_state()
29 fpsave(&current->thread.float_regs[0], &current->thread.fsr, in save_fpu_state()
30 &current->thread.fpqueue[0], &current->thread.fpqdepth); in save_fpu_state()
36 &current->thread.float_regs[0], in save_fpu_state()
38 err |= __put_user(current->thread.fsr, &fpu->si_fsr); in save_fpu_state()
39 err |= __put_user(current->thread.fpqdepth, &fpu->si_fpqdepth); in save_fpu_state()
40 if (current->thread.fpqdepth != 0) in save_fpu_state()
42 &current->thread.fpqueue[0], in save_fpu_state()
71 err = __copy_from_user(&current->thread.float_regs[0], &fpu->si_float_regs[0], in restore_fpu_state()
[all …]
Dprocess_32.c192 fpsave(&tsk->thread.float_regs[0], &tsk->thread.fsr,
193 &tsk->thread.fpqueue[0], &tsk->thread.fpqdepth);
213 fpsave(&current->thread.float_regs[0], &current->thread.fsr,
214 &current->thread.fpqueue[0], &current->thread.fpqdepth);
223 if (current->thread.flags & SPARC_FLAG_KTHREAD) {
224 current->thread.flags &= ~SPARC_FLAG_KTHREAD;
228 current->thread.kregs = (struct pt_regs *)
317 fpsave(&p->thread.float_regs[0], &p->thread.fsr,
318 &p->thread.fpqueue[0], &p->thread.fpqdepth);
339 p->thread.kregs = childregs;
[all …]
/arch/m68k/kernel/
Dprocess.c95 current->thread.fs = __USER_DS; in flush_thread()
132 p->thread.ksp = (unsigned long)frame; in copy_thread()
133 p->thread.esp0 = (unsigned long)&frame->regs; in copy_thread()
139 p->thread.fs = get_fs().seg; in copy_thread()
148 p->thread.usp = 0; in copy_thread()
155 p->thread.usp = usp ?: rdusp(); in copy_thread()
163 asm volatile ("fsave %0" : : "m" (p->thread.fpstate[0]) : "memory"); in copy_thread()
165 if (!CPU_IS_060 ? p->thread.fpstate[0] : p->thread.fpstate[2]) { in copy_thread()
172 : "m" (p->thread.fp[0]), in copy_thread()
173 "m" (p->thread.fpcntl[0]), in copy_thread()
[all …]
/arch/mips/kernel/
Dasm-offsets.c113 OFFSET(THREAD_REG16, task_struct, thread.reg16); in output_thread_defines()
114 OFFSET(THREAD_REG17, task_struct, thread.reg17); in output_thread_defines()
115 OFFSET(THREAD_REG18, task_struct, thread.reg18); in output_thread_defines()
116 OFFSET(THREAD_REG19, task_struct, thread.reg19); in output_thread_defines()
117 OFFSET(THREAD_REG20, task_struct, thread.reg20); in output_thread_defines()
118 OFFSET(THREAD_REG21, task_struct, thread.reg21); in output_thread_defines()
119 OFFSET(THREAD_REG22, task_struct, thread.reg22); in output_thread_defines()
120 OFFSET(THREAD_REG23, task_struct, thread.reg23); in output_thread_defines()
121 OFFSET(THREAD_REG29, task_struct, thread.reg29); in output_thread_defines()
122 OFFSET(THREAD_REG30, task_struct, thread.reg30); in output_thread_defines()
[all …]
/arch/arm/vfp/
Dvfpmodule.c62 static bool vfp_state_in_hw(unsigned int cpu, struct thread_info *thread) in vfp_state_in_hw() argument
65 if (thread->vfpstate.hard.cpu != cpu) in vfp_state_in_hw()
68 return vfp_current_hw_state[cpu] == &thread->vfpstate; in vfp_state_in_hw()
76 static void vfp_force_reload(unsigned int cpu, struct thread_info *thread) in vfp_force_reload() argument
78 if (vfp_state_in_hw(cpu, thread)) { in vfp_force_reload()
83 thread->vfpstate.hard.cpu = NR_CPUS; in vfp_force_reload()
90 static void vfp_thread_flush(struct thread_info *thread) in vfp_thread_flush() argument
92 union vfp_state *vfp = &thread->vfpstate; in vfp_thread_flush()
118 static void vfp_thread_exit(struct thread_info *thread) in vfp_thread_exit() argument
121 union vfp_state *vfp = &thread->vfpstate; in vfp_thread_exit()
[all …]
/arch/s390/include/asm/
Dswitch_to.h39 save_access_regs(&prev->thread.acrs[0]); \
40 save_ri_cb(prev->thread.ri_cb); \
41 save_gs_cb(prev->thread.gs_cb); \
43 restore_access_regs(&next->thread.acrs[0]); \
44 restore_ri_cb(next->thread.ri_cb, prev->thread.ri_cb); \
45 restore_gs_cb(next->thread.gs_cb); \
/arch/m68k/mm/
Dfault.c27 signo = current->thread.signo; in send_fault_sig()
28 si_code = current->thread.code; in send_fault_sig()
29 addr = (void __user *)current->thread.faddr; in send_fault_sig()
195 current->thread.signo = SIGBUS; in do_page_fault()
196 current->thread.faddr = address; in do_page_fault()
200 current->thread.signo = SIGBUS; in do_page_fault()
201 current->thread.code = BUS_ADRERR; in do_page_fault()
202 current->thread.faddr = address; in do_page_fault()
206 current->thread.signo = SIGSEGV; in do_page_fault()
207 current->thread.code = SEGV_MAPERR; in do_page_fault()
[all …]
/arch/sh/kernel/
Dprocess.c31 if (src->thread.xstate) { in arch_dup_task_struct()
32 dst->thread.xstate = kmem_cache_alloc(task_xstate_cachep, in arch_dup_task_struct()
34 if (!dst->thread.xstate) in arch_dup_task_struct()
36 memcpy(dst->thread.xstate, src->thread.xstate, xstate_size); in arch_dup_task_struct()
44 if (tsk->thread.xstate) { in free_thread_xstate()
45 kmem_cache_free(task_xstate_cachep, tsk->thread.xstate); in free_thread_xstate()
46 tsk->thread.xstate = NULL; in free_thread_xstate()
/arch/x86/kernel/
Dprocess_64.c193 prev_p->thread.fsbase = 0; in save_base_legacy()
195 prev_p->thread.gsbase = 0; in save_base_legacy()
201 savesegment(fs, task->thread.fsindex); in save_fsgs()
202 savesegment(gs, task->thread.gsindex); in save_fsgs()
203 save_base_legacy(task, task->thread.fsindex, FS); in save_fsgs()
204 save_base_legacy(task, task->thread.gsindex, GS); in save_fsgs()
307 base = get_desc_base(&task->thread.tls_array[idx]); in x86_fsgsbase_read_task()
338 else if (task->thread.fsindex == 0) in x86_fsbase_read_task()
339 fsbase = task->thread.fsbase; in x86_fsbase_read_task()
341 fsbase = x86_fsgsbase_read_task(task, task->thread.fsindex); in x86_fsbase_read_task()
[all …]
/arch/powerpc/include/asm/
Dswitch_to.h82 t->thread.ebbrr = 0; in clear_task_ebb()
83 t->thread.ebbhr = 0; in clear_task_ebb()
84 t->thread.bescr = 0; in clear_task_ebb()
85 t->thread.mmcr2 = 0; in clear_task_ebb()
86 t->thread.mmcr0 = 0; in clear_task_ebb()
87 t->thread.siar = 0; in clear_task_ebb()
88 t->thread.sdar = 0; in clear_task_ebb()
89 t->thread.sier = 0; in clear_task_ebb()
90 t->thread.used_ebb = 0; in clear_task_ebb()

12345678910>>...16