• Home
  • Raw
  • Download

Lines Matching refs:thread

44 	struct thread_struct *thread = &task->thread;  in update_cr_regs()  local
58 if (task->thread.per_flags & PER_FLAG_NO_TE) in update_cr_regs()
62 if (task->thread.per_flags & PER_FLAG_TE_ABORT_RAND) { in update_cr_regs()
63 if (task->thread.per_flags & PER_FLAG_TE_ABORT_RAND_TEND) in update_cr_regs()
72 if (task->thread.gs_cb) in update_cr_regs()
83 new.control = thread->per_user.control; in update_cr_regs()
84 new.start = thread->per_user.start; in update_cr_regs()
85 new.end = thread->per_user.end; in update_cr_regs()
138 memset(&task->thread.per_user, 0, sizeof(task->thread.per_user)); in ptrace_disable()
139 memset(&task->thread.per_event, 0, sizeof(task->thread.per_event)); in ptrace_disable()
142 task->thread.per_flags = 0; in ptrace_disable()
155 PER_EVENT_IFETCH : child->thread.per_user.control; in __peek_user_per()
159 0 : child->thread.per_user.start; in __peek_user_per()
163 -1UL : child->thread.per_user.end; in __peek_user_per()
170 return child->thread.per_user.start; in __peek_user_per()
173 return child->thread.per_user.end; in __peek_user_per()
177 child->thread.per_event.cause << (BITS_PER_LONG - 16); in __peek_user_per()
180 return child->thread.per_event.address; in __peek_user_per()
184 child->thread.per_event.paid << (BITS_PER_LONG - 8); in __peek_user_per()
224 tmp = ((unsigned long) child->thread.acrs[15]) << 32; in __peek_user()
226 tmp = *(addr_t *)((addr_t) &child->thread.acrs + offset); in __peek_user()
245 tmp = child->thread.fpu.fpc; in __peek_user()
256 ((addr_t) child->thread.fpu.vxrs + 2*offset); in __peek_user()
259 ((addr_t) child->thread.fpu.fprs + offset); in __peek_user()
313 child->thread.per_user.control = in __poke_user_per()
317 child->thread.per_user.start = data; in __poke_user_per()
320 child->thread.per_user.end = data; in __poke_user_per()
374 child->thread.acrs[15] = (unsigned int) (data >> 32); in __poke_user()
376 *(addr_t *)((addr_t) &child->thread.acrs + offset) = data; in __poke_user()
399 child->thread.fpu.fpc = data >> (BITS_PER_LONG - 32); in __poke_user()
409 child->thread.fpu.vxrs + 2*offset) = data; in __poke_user()
412 child->thread.fpu.fprs + offset) = data; in __poke_user()
485 return put_user(child->thread.last_break, (unsigned long __user *)data); in arch_ptrace()
489 child->thread.per_flags &= ~PER_FLAG_NO_TE; in arch_ptrace()
494 child->thread.per_flags |= PER_FLAG_NO_TE; in arch_ptrace()
495 child->thread.per_flags &= ~PER_FLAG_TE_ABORT_RAND; in arch_ptrace()
498 if (!MACHINE_HAS_TE || (child->thread.per_flags & PER_FLAG_NO_TE)) in arch_ptrace()
502 child->thread.per_flags &= ~PER_FLAG_TE_ABORT_RAND; in arch_ptrace()
505 child->thread.per_flags |= PER_FLAG_TE_ABORT_RAND; in arch_ptrace()
506 child->thread.per_flags |= PER_FLAG_TE_ABORT_RAND_TEND; in arch_ptrace()
509 child->thread.per_flags |= PER_FLAG_TE_ABORT_RAND; in arch_ptrace()
510 child->thread.per_flags &= ~PER_FLAG_TE_ABORT_RAND_TEND; in arch_ptrace()
547 PER_EVENT_IFETCH : child->thread.per_user.control; in __peek_user_per_compat()
551 0 : child->thread.per_user.start; in __peek_user_per_compat()
555 PSW32_ADDR_INSN : child->thread.per_user.end; in __peek_user_per_compat()
562 return (__u32) child->thread.per_user.start; in __peek_user_per_compat()
565 return (__u32) child->thread.per_user.end; in __peek_user_per_compat()
568 return (__u32) child->thread.per_event.cause << 16; in __peek_user_per_compat()
571 return (__u32) child->thread.per_event.address; in __peek_user_per_compat()
574 return (__u32) child->thread.per_event.paid << 24; in __peek_user_per_compat()
610 tmp = *(__u32*)((addr_t) &child->thread.acrs + offset); in __peek_user_compat()
629 tmp = child->thread.fpu.fpc; in __peek_user_compat()
639 ((addr_t) child->thread.fpu.vxrs + 2*offset); in __peek_user_compat()
642 ((addr_t) child->thread.fpu.fprs + offset); in __peek_user_compat()
679 child->thread.per_user.control = in __poke_user_per_compat()
683 child->thread.per_user.start = data; in __poke_user_per_compat()
686 child->thread.per_user.end = data; in __poke_user_per_compat()
739 *(__u32*)((addr_t) &child->thread.acrs + offset) = tmp; in __poke_user_compat()
761 child->thread.fpu.fpc = data; in __poke_user_compat()
771 child->thread.fpu.vxrs + 2*offset) = tmp; in __poke_user_compat()
774 child->thread.fpu.fprs + offset) = tmp; in __poke_user_compat()
840 return put_user(child->thread.last_break, (unsigned int __user *)data); in compat_arch_ptrace()
856 save_access_regs(target->thread.acrs); in s390_regs_get()
871 save_access_regs(target->thread.acrs); in s390_regs_set()
894 restore_access_regs(target->thread.acrs); in s390_regs_set()
908 fp_regs.fpc = target->thread.fpu.fpc; in s390_fpregs_get()
909 fpregs_store(&fp_regs, &target->thread.fpu); in s390_fpregs_get()
924 convert_vx_to_fp(fprs, target->thread.fpu.vxrs); in s390_fpregs_set()
926 memcpy(&fprs, target->thread.fpu.fprs, sizeof(fprs)); in s390_fpregs_set()
930 u32 ufpc[2] = { target->thread.fpu.fpc, 0 }; in s390_fpregs_set()
937 target->thread.fpu.fpc = ufpc[0]; in s390_fpregs_set()
947 convert_fp_to_vx(target->thread.fpu.vxrs, fprs); in s390_fpregs_set()
949 memcpy(target->thread.fpu.fprs, &fprs, sizeof(fprs)); in s390_fpregs_set()
958 return membuf_store(&to, target->thread.last_break); in s390_last_break_get()
978 size = sizeof(target->thread.trap_tdb.data); in s390_tdb_get()
979 return membuf_write(&to, target->thread.trap_tdb.data, size); in s390_tdb_get()
1002 vxrs[i] = *((__u64 *)(target->thread.fpu.vxrs + i) + 1); in s390_vxrs_low_get()
1020 vxrs[i] = *((__u64 *)(target->thread.fpu.vxrs + i) + 1); in s390_vxrs_low_set()
1025 *((__u64 *)(target->thread.fpu.vxrs + i) + 1) = vxrs[i]; in s390_vxrs_low_set()
1038 return membuf_write(&to, target->thread.fpu.vxrs + __NUM_VXRS_LOW, in s390_vxrs_high_get()
1055 target->thread.fpu.vxrs + __NUM_VXRS_LOW, 0, -1); in s390_vxrs_high_set()
1063 return membuf_store(&to, target->thread.system_call); in s390_system_call_get()
1071 unsigned int *data = &target->thread.system_call; in s390_system_call_set()
1080 struct gs_cb *data = target->thread.gs_cb; in s390_gs_cb_get()
1101 if (!target->thread.gs_cb) { in s390_gs_cb_set()
1106 if (!target->thread.gs_cb) in s390_gs_cb_set()
1111 gs_cb = *target->thread.gs_cb; in s390_gs_cb_set()
1119 if (!target->thread.gs_cb) in s390_gs_cb_set()
1120 target->thread.gs_cb = data; in s390_gs_cb_set()
1121 *target->thread.gs_cb = gs_cb; in s390_gs_cb_set()
1124 restore_gs_cb(target->thread.gs_cb); in s390_gs_cb_set()
1134 struct gs_cb *data = target->thread.gs_bc_cb; in s390_gs_bc_get()
1148 struct gs_cb *data = target->thread.gs_bc_cb; in s390_gs_bc_set()
1156 target->thread.gs_bc_cb = data; in s390_gs_bc_set()
1192 struct runtime_instr_cb *data = target->thread.ri_cb; in s390_runtime_instr_get()
1213 if (!target->thread.ri_cb) { in s390_runtime_instr_set()
1219 if (target->thread.ri_cb) { in s390_runtime_instr_set()
1223 ri_cb = *target->thread.ri_cb; in s390_runtime_instr_set()
1243 if (!target->thread.ri_cb) in s390_runtime_instr_set()
1244 target->thread.ri_cb = data; in s390_runtime_instr_set()
1245 *target->thread.ri_cb = ri_cb; in s390_runtime_instr_set()
1247 load_runtime_instr_cb(target->thread.ri_cb); in s390_runtime_instr_set()
1351 save_access_regs(target->thread.acrs); in s390_compat_regs_get()
1366 save_access_regs(target->thread.acrs); in s390_compat_regs_set()
1389 restore_access_regs(target->thread.acrs); in s390_compat_regs_set()
1444 compat_ulong_t last_break = target->thread.last_break; in s390_compat_last_break_get()