Searched refs:tid (Results 1 – 25 of 474) sorted by relevance
12345678910>>...19
56 ThreadId tid; in VG_() local62 for (tid = 1; tid < VG_N_THREADS; tid++) { in VG_()64 ANNOTATE_BENIGN_RACE_SIZED(&VG_(threads)[tid].status, in VG_()65 sizeof(VG_(threads)[tid].status), "")); in VG_()67 ANNOTATE_BENIGN_RACE_SIZED(&VG_(threads)[tid].os_state.exitcode, in VG_()68 sizeof(VG_(threads)[tid].os_state.exitcode), in VG_()97 ThreadState *VG_(get_ThreadState)(ThreadId tid) in VG_()99 vg_assert(tid >= 0 && tid < VG_N_THREADS); in VG_()100 vg_assert(VG_(threads)[tid].tid == tid); in VG_()101 return &VG_(threads)[tid]; in VG_()[all …]
48 Addr VG_(get_IP) ( ThreadId tid ) { in VG_()49 return INSTR_PTR( VG_(threads)[tid].arch ); in VG_()51 Addr VG_(get_SP) ( ThreadId tid ) { in VG_()52 return STACK_PTR( VG_(threads)[tid].arch ); in VG_()54 Addr VG_(get_FP) ( ThreadId tid ) { in VG_()55 return FRAME_PTR( VG_(threads)[tid].arch ); in VG_()58 void VG_(set_IP) ( ThreadId tid, Addr ip ) { in VG_()59 INSTR_PTR( VG_(threads)[tid].arch ) = ip; in VG_()61 void VG_(set_SP) ( ThreadId tid, Addr sp ) { in VG_()62 STACK_PTR( VG_(threads)[tid].arch ) = sp; in VG_()[all …]
48 static void thread_append_segment(const DrdThreadId tid, Segment* const sg);49 static void thread_discard_segment(const DrdThreadId tid, Segment* const sg);51 const DrdThreadId tid);52 static Bool thread_conflict_set_up_to_date(const DrdThreadId tid);164 DrdThreadId DRD_(VgThreadIdToDrdThreadId)(const ThreadId tid) in DRD_()168 if (tid == VG_INVALID_THREADID) in DRD_()174 && DRD_(g_threadinfo)[i].vg_threadid == tid) in DRD_()184 static DrdThreadId DRD_(VgThreadIdToNewDrdThreadId)(const ThreadId tid) in DRD_()188 tl_assert(DRD_(VgThreadIdToDrdThreadId)(tid) == DRD_INVALID_THREADID); in DRD_()198 DRD_(g_threadinfo)[i].vg_threadid = tid; in DRD_()[all …]
146 DrdThreadId DRD_(VgThreadIdToDrdThreadId)(const ThreadId tid);147 DrdThreadId DRD_(NewVgThreadIdToDrdThreadId)(const ThreadId tid);148 DrdThreadId DRD_(PtThreadIdToDrdThreadId)(const PThreadId tid);149 ThreadId DRD_(DrdThreadIdToVgThreadId)(const DrdThreadId tid);154 void DRD_(thread_delete)(const DrdThreadId tid, Bool detached);155 void DRD_(thread_finished)(const DrdThreadId tid);156 void DRD_(drd_thread_atfork_child)(const DrdThreadId tid);157 void DRD_(thread_pre_cancel)(const DrdThreadId tid);158 void DRD_(thread_set_stack_startup)(const DrdThreadId tid,160 Addr DRD_(thread_get_stack_min)(const DrdThreadId tid);[all …]
70 static void* new_block(ThreadId tid, SizeT size, SizeT align, Bool is_zeroed) in new_block() argument80 DRD_(malloclike_block)(tid, (Addr)p, size); in new_block()89 void DRD_(malloclike_block)(const ThreadId tid, const Addr p, const SizeT size) in DRD_()105 mc->where = VG_(record_ExeContext)(tid, 0); in DRD_()109 static void handle_free(ThreadId tid, void* p) in handle_free() argument114 success = DRD_(freelike_block)(tid, (Addr)p, True); in handle_free()122 Bool DRD_(freelike_block)(const ThreadId tid, const Addr p, const Bool dealloc) in DRD_()146 static void* drd_malloc(ThreadId tid, SizeT n) in drd_malloc() argument148 return new_block(tid, n, VG_(clo_alignment), /*is_zeroed*/False); in drd_malloc()152 static void* drd_memalign(ThreadId tid, SizeT align, SizeT n) in drd_memalign() argument[all …]
90 int tid = flattenedThreadId(); in yota() local91 value += tid; in yota()93 for(OutIt t = beg + tid; t < end; t += STRIDE, value += STRIDE) in yota()134 int tid = flattenedThreadId(); in reduce() local135 T val = buffer[tid]; in reduce()137 …if (CTA_SIZE >= 1024) { if (tid < 512) buffer[tid] = val = op(val, buffer[tid + 512]); __syncthrea… in reduce()138 …if (CTA_SIZE >= 512) { if (tid < 256) buffer[tid] = val = op(val, buffer[tid + 256]); __syncthrea… in reduce()139 …if (CTA_SIZE >= 256) { if (tid < 128) buffer[tid] = val = op(val, buffer[tid + 128]); __syncthrea… in reduce()140 …if (CTA_SIZE >= 128) { if (tid < 64) buffer[tid] = val = op(val, buffer[tid + 64]); __syncthrea… in reduce()142 if (tid < 32) in reduce()[all …]
55 __device__ __forceinline__ T warp_reduce(volatile T *ptr , const unsigned int tid = threadIdx.x) in warp_reduce() argument57 const unsigned int lane = tid & 31; // index of thread in warp (0..31) in warp_reduce()61 T partial = ptr[tid]; in warp_reduce()63 ptr[tid] = partial = partial + ptr[tid + 16]; in warp_reduce()64 ptr[tid] = partial = partial + ptr[tid + 8]; in warp_reduce()65 ptr[tid] = partial = partial + ptr[tid + 4]; in warp_reduce()66 ptr[tid] = partial = partial + ptr[tid + 2]; in warp_reduce()67 ptr[tid] = partial = partial + ptr[tid + 1]; in warp_reduce()70 return ptr[tid - lane]; in warp_reduce()
83 __device__ __forceinline__ unsigned int index(const unsigned int tid) in index()85 return tid; in index()117 __device__ __forceinline__ unsigned int index(const unsigned int tid) in index()119 return (tid >> warp_log) * warp_smem_stride + 16 + (tid & warp_mask); in index()142 const unsigned int tid = threadIdx.x; in operator ()() local143 const unsigned int lane = tid & warp_mask; in operator ()()144 const unsigned int warp = tid >> warp_log; in operator ()()148 const unsigned int idx = scan.index(tid); in operator ()()180 __device__ T warpScanInclusive(T idata, volatile T* s_Data, unsigned int tid) in warpScanInclusive() argument196 unsigned int pos = 2 * tid - (tid & (OPENCV_CUDA_WARP_SIZE - 1)); in warpScanInclusive()[all …]
147 int tid = (cell_y * CELLS_PER_BLOCK_Y + cell_x) * 12 + cell_thread_x;148 if ((tid < cblock_hist_size) && (gid < blocks_total))152 block_hist[tid] = final_hist[tid];162 const int tid = get_local_id(0);164 const int bid = tid / 36; /* block-hist id, (0 - 6) */166 const int hid = tid - boffset; /* histogram bin id, (0 - 35) */169 squares[tid] = elem * elem;189 squares[tid] = elem * elem;212 unsigned int tid = get_local_id(0);213 float sum = smem[tid];[all …]
125 static void do_client_request ( ThreadId tid );126 static void scheduler_sanity ( ThreadId tid );127 static void mostly_clear_thread_record ( ThreadId tid );176 void print_sched_event ( ThreadId tid, const HChar* what ) in print_sched_event() argument178 VG_(message)(Vg_DebugMsg, " SCHED[%u]: %s\n", tid, what ); in print_sched_event()265 void VG_(acquire_BigLock)(ThreadId tid, const HChar* who) in VG_()273 print_sched_event(tid, buf); in VG_()282 tst = VG_(get_ThreadState)(tid); in VG_()289 VG_(printf)("tid %u found %u running\n", tid, VG_(running_tid)); in VG_()291 VG_(running_tid) = tid; in VG_()[all …]
19 ThreadContextBase::ThreadContextBase(u32 tid) in ThreadContextBase() argument20 : tid(tid), unique_id(0), reuse_count(), os_id(0), user_id(0), in ThreadContextBase()75 if (tid != 0) in SetCreated()124 u32 tid = kUnknownTid; in CreateThread() local127 tid = tctx->tid; in CreateThread()130 tid = n_contexts_++; in CreateThread()131 tctx = context_factory_(tid); in CreateThread()132 threads_[tid] = tctx; in CreateThread()144 CHECK_NE(tid, kUnknownTid); in CreateThread()145 CHECK_LT(tid, max_threads_); in CreateThread()[all …]
82 … __device__ static void loadToSmem(const PointerTuple& smem, const ReferenceTuple& data, uint tid) in loadToSmem()84 get<I>(smem)[tid] = get<I>(data); in loadToSmem()86 For<I + 1, N>::loadToSmem(smem, data, tid); in loadToSmem()90 …__device__ static void loadFromSmem(const PointerTuple& smem, const ReferenceTuple& data, uint tid) in loadFromSmem()92 get<I>(data) = get<I>(smem)[tid]; in loadFromSmem()94 For<I + 1, N>::loadFromSmem(smem, data, tid); in loadFromSmem()98 …__device__ static void copy(const PointerTuple& svals, const ReferenceTuple& val, uint tid, uint d… in copy()100 get<I>(svals)[tid] = get<I>(val) = get<I>(svals)[tid + delta]; in copy()102 For<I + 1, N>::copy(svals, val, tid, delta); in copy()109 uint tid, uint delta) in merge()[all …]
83 __device__ static void loadToSmem(const PointerTuple& smem, const ValTuple& val, uint tid) in loadToSmem()85 get<I>(smem)[tid] = get<I>(val); in loadToSmem()87 For<I + 1, N>::loadToSmem(smem, val, tid); in loadToSmem()91 __device__ static void loadFromSmem(const PointerTuple& smem, const ValTuple& val, uint tid) in loadFromSmem()93 get<I>(val) = get<I>(smem)[tid]; in loadFromSmem()95 For<I + 1, N>::loadFromSmem(smem, val, tid); in loadFromSmem()99 …__device__ static void merge(const PointerTuple& smem, const ValTuple& val, uint tid, uint delta, … in merge()101 …ename GetType<typename tuple_element<I, PointerTuple>::type>::type reg = get<I>(smem)[tid + delta]; in merge()102 get<I>(smem)[tid] = get<I>(val) = get<I>(op)(get<I>(val), reg); in merge()104 For<I + 1, N>::merge(smem, val, tid, delta, op); in merge()[all …]
86 void ML_(setup_start_thread_context)(ThreadId tid, vki_ucontext_t *uc) in ML_()88 ThreadState *tst = VG_(get_ThreadState)(tid); in ML_()110 void ML_(save_machine_context)(ThreadId tid, vki_ucontext_t *uc, in ML_()113 ThreadState *tst = VG_(get_ThreadState)(tid); in ML_()121 VG_TRACK(copy_reg_to_mem, part, tid, OFFSET_amd64_RIP, in ML_()124 VG_TRACK(copy_reg_to_mem, part, tid, OFFSET_amd64_RAX, in ML_()127 VG_TRACK(copy_reg_to_mem, part, tid, OFFSET_amd64_RBX, in ML_()130 VG_TRACK(copy_reg_to_mem, part, tid, OFFSET_amd64_RCX, in ML_()133 VG_TRACK(copy_reg_to_mem, part, tid, OFFSET_amd64_RDX, in ML_()136 VG_TRACK(copy_reg_to_mem, part, tid, OFFSET_amd64_RBP, in ML_()[all …]
92 void ML_(setup_start_thread_context)(ThreadId tid, vki_ucontext_t *uc) in ML_()94 ThreadState *tst = VG_(get_ThreadState)(tid); in ML_()134 void ML_(save_machine_context)(ThreadId tid, vki_ucontext_t *uc, in ML_()137 ThreadState *tst = VG_(get_ThreadState)(tid); in ML_()145 VG_TRACK(copy_reg_to_mem, part, tid, OFFSET_x86_EIP, in ML_()148 VG_TRACK(copy_reg_to_mem, part, tid, OFFSET_x86_EAX, in ML_()151 VG_TRACK(copy_reg_to_mem, part, tid, OFFSET_x86_EBX, in ML_()154 VG_TRACK(copy_reg_to_mem, part, tid, OFFSET_x86_ECX, in ML_()157 VG_TRACK(copy_reg_to_mem, part, tid, OFFSET_x86_EDX, in ML_()160 VG_TRACK(copy_reg_to_mem, part, tid, OFFSET_x86_EBP, in ML_()[all …]
50 inline int calc_lut(__local int* smem, int val, int tid)52 smem[tid] = val;55 if (tid == 0)60 return smem[tid];64 inline void reduce(volatile __local int* smem, int val, int tid)66 smem[tid] = val;69 if (tid < 128)70 smem[tid] = val += smem[tid + 128];73 if (tid < 64)74 smem[tid] = val += smem[tid + 64];[all …]
... 2 {"cat": "PERF", "pid": 22630, "tid": 22630, "ts": 826, "ph": "C ...
74 … static __device__ void loadToSmem(const PointerTuple& smem, const ValTuple& val, unsigned int tid) in loadToSmem()76 thrust::get<I>(smem)[tid] = thrust::get<I>(val); in loadToSmem()78 For<I + 1, N>::loadToSmem(smem, val, tid); in loadToSmem()81 …tatic __device__ void loadFromSmem(const PointerTuple& smem, const ValTuple& val, unsigned int tid) in loadFromSmem()83 thrust::get<I>(val) = thrust::get<I>(smem)[tid]; in loadFromSmem()85 For<I + 1, N>::loadFromSmem(smem, val, tid); in loadFromSmem()89 …e__ void merge(const PointerTuple& smem, const ValTuple& val, unsigned int tid, unsigned int delta… in merge()91 …pename thrust::tuple_element<I, PointerTuple>::type>::type reg = thrust::get<I>(smem)[tid + delta]; in merge()92 … thrust::get<I>(smem)[tid] = thrust::get<I>(val) = thrust::get<I>(op)(thrust::get<I>(val), reg); in merge()94 For<I + 1, N>::merge(smem, val, tid, delta, op); in merge()[all …]
74 … __device__ void loadToSmem(const PointerTuple& smem, const ReferenceTuple& data, unsigned int tid) in loadToSmem()76 thrust::get<I>(smem)[tid] = thrust::get<I>(data); in loadToSmem()78 For<I + 1, N>::loadToSmem(smem, data, tid); in loadToSmem()81 …_device__ void loadFromSmem(const PointerTuple& smem, const ReferenceTuple& data, unsigned int tid) in loadFromSmem()83 thrust::get<I>(data) = thrust::get<I>(smem)[tid]; in loadFromSmem()85 For<I + 1, N>::loadFromSmem(smem, data, tid); in loadFromSmem()96 …id copy(const PointerTuple& svals, const ReferenceTuple& val, unsigned int tid, unsigned int delta) in copy()98 … thrust::get<I>(svals)[tid] = thrust::get<I>(val) = thrust::get<I>(svals)[tid + delta]; in copy()100 For<I + 1, N>::copy(svals, val, tid, delta); in copy()120 unsigned int tid, unsigned int delta) in merge()[all …]
80 … __device__ static void loadToSmem(const PointerTuple& smem, const ReferenceTuple& data, uint tid) in loadToSmem()82 get<I>(smem)[tid] = get<I>(data); in loadToSmem()84 For<I + 1, N>::loadToSmem(smem, data, tid); in loadToSmem()88 …__device__ static void copy(const PointerTuple& svals, const ReferenceTuple& val, uint tid, uint d… in copy()90 get<I>(svals)[tid] = get<I>(val) = get<I>(svals)[tid + delta]; in copy()92 For<I + 1, N>::copy(svals, val, tid, delta); in copy()99 uint tid, uint delta) in merge()101 …e GetType<typename tuple_element<I, KeyPointerTuple>::type>::type reg = get<I>(skeys)[tid + delta]; in merge()105 get<I>(skeys)[tid] = get<I>(key) = reg; in merge()106 get<I>(svals)[tid] = get<I>(val) = get<I>(svals)[tid + delta]; in merge()[all …]
1267 lg{"Acre-tid"}1271 ls{"Afghansk tid"}1274 ls{"Centralafrikansk tid"}1277 ls{"Østafrikansk tid"}1280 ls{"Sydafrikansk tid"}1284 lg{"Vestafrikansk tid"}1289 lg{"Alaska-tid"}1294 lg{"Almaty-tid"}1299 lg{"Amazonas-tid"}1304 lg{"Central-tid"}[all …]
57 void VG_(sigframe_create)(ThreadId tid, Bool on_altstack, in VG_()63 ThreadState *tst = VG_(get_ThreadState)(tid); in VG_()85 VG_(save_context)(tid, &frame->ucontext, Vg_CoreSignal); in VG_()132 frame->siginfo.si_addr = (void*)VG_(get_IP)(tid); in VG_()135 frame->siginfo.si_faddr = (void*)VG_(get_IP)(tid); in VG_()140 VG_TRACK(post_mem_write, Vg_CoreSignal, tid, (Addr)&frame->siginfo, in VG_()153 VG_TRACK(post_mem_write, Vg_CoreSignal, tid, in VG_()158 VG_TRACK(post_mem_write, Vg_CoreSignal, tid, in VG_()175 VG_TRACK(post_mem_write, Vg_CoreSignal, tid, (Addr)&frame->a1_signo, in VG_()179 VG_TRACK(post_mem_write, Vg_CoreSignal, tid, (Addr)&frame->a3_ucontext, in VG_()[all …]
12 %tid = call i32 @llvm.r600.read.tidig.x() #013 %gep.in = getelementptr i32, i32 addrspace(1)* %in, i32 %tid14 %gep.out = getelementptr i32, i32 addrspace(1)* %out, i32 %tid25 %tid = call i32 @llvm.r600.read.tidig.x() #026 %gep.in = getelementptr i32, i32 addrspace(1)* %in, i32 %tid27 %gep.out = getelementptr i32, i32 addrspace(1)* %out, i32 %tid40 %tid = call i32 @llvm.r600.read.tidig.x() #041 %gep.in = getelementptr i32, i32 addrspace(1)* %in, i32 %tid42 %gep.out = getelementptr i32, i32 addrspace(1)* %out, i32 %tid53 %tid = call i32 @llvm.r600.read.tidig.x() #0[all …]
13 %tid = tail call i32 @llvm.r600.read.tidig.x() #014 %tid.ext = sext i32 %tid to i6415 %gep0 = getelementptr float, float addrspace(1)* %ptr, i64 %tid.ext16 %add1 = add i64 %tid.ext, 118 %add2 = add i64 %tid.ext, 220 %outgep = getelementptr float, float addrspace(1)* %out, i64 %tid.ext37 %tid = tail call i32 @llvm.r600.read.tidig.x() #038 %tid.ext = sext i32 %tid to i6439 %gep0 = getelementptr float, float addrspace(1)* %ptr, i64 %tid.ext40 %add1 = add i64 %tid.ext, 1[all …]
58 __device__ T blockScanInclusive(T data, volatile T* smem, uint tid) in blockScanInclusive() argument63 T warpResult = warpScanInclusive(data, smem, tid); in blockScanInclusive()69 if ((tid & (WARP_SIZE - 1)) == (WARP_SIZE - 1)) in blockScanInclusive()71 smem[tid >> LOG_WARP_SIZE] = warpResult; in blockScanInclusive()76 if (tid < (THREADS_NUM / WARP_SIZE)) in blockScanInclusive()79 T val = smem[tid]; in blockScanInclusive()82 smem[tid] = warpScanExclusive(val, smem, tid); in blockScanInclusive()88 return warpResult + smem[tid >> LOG_WARP_SIZE]; in blockScanInclusive()92 return warpScanInclusive(data, smem, tid); in blockScanInclusive()97 __device__ __forceinline__ T blockScanExclusive(T data, volatile T* smem, uint tid) in blockScanExclusive() argument[all …]