• Home
  • Raw
  • Download

Lines Matching refs:tid

47     enum barrier_type bt, kmp_info_t *this_thr, int gtid, int tid,  in __kmp_linear_barrier_gather_template()  argument
57 gtid, team->t.t_id, tid, bt)); in __kmp_linear_barrier_gather_template()
69 if (!KMP_MASTER_TID(tid)) { in __kmp_linear_barrier_gather_template()
73 gtid, team->t.t_id, tid, __kmp_gtid_from_tid(0, team), in __kmp_linear_barrier_gather_template()
99 gtid, team->t.t_id, tid, __kmp_gtid_from_tid(i, team), in __kmp_linear_barrier_gather_template()
126 gtid, team->t.t_id, tid, __kmp_gtid_from_tid(i, team), in __kmp_linear_barrier_gather_template()
142 gtid, team->t.t_id, tid, team->t.t_id, &team_bar->b_arrived, in __kmp_linear_barrier_gather_template()
148 gtid, team->t.t_id, tid, bt)); in __kmp_linear_barrier_gather_template()
154 enum barrier_type bt, kmp_info_t *this_thr, int gtid, int tid, in __kmp_linear_barrier_release_template() argument
160 if (KMP_MASTER_TID(tid)) { in __kmp_linear_barrier_release_template()
171 gtid, team->t.t_id, tid, bt)); in __kmp_linear_barrier_release_template()
201 gtid, team->t.t_id, tid, other_threads[i]->th.th_info.ds.ds_gtid, in __kmp_linear_barrier_release_template()
245 tid = __kmp_tid_from_gtid(gtid); in __kmp_linear_barrier_release_template()
252 gtid, team->t.t_id, tid, &thr_bar->b_go, KMP_INIT_BARRIER_STATE)); in __kmp_linear_barrier_release_template()
258 gtid, team->t.t_id, tid, bt)); in __kmp_linear_barrier_release_template()
263 enum barrier_type bt, kmp_info_t *this_thr, int gtid, int tid, in __kmp_linear_barrier_gather() argument
266 bt, this_thr, gtid, tid, reduce USE_ITT_BUILD_ARG(itt_sync_obj)); in __kmp_linear_barrier_gather()
270 enum barrier_type bt, kmp_info_t *this_thr, int gtid, int tid, in __kmp_linear_barrier_gather_cancellable() argument
273 bt, this_thr, gtid, tid, reduce USE_ITT_BUILD_ARG(itt_sync_obj)); in __kmp_linear_barrier_gather_cancellable()
277 enum barrier_type bt, kmp_info_t *this_thr, int gtid, int tid, in __kmp_linear_barrier_release() argument
280 bt, this_thr, gtid, tid, propagate_icvs USE_ITT_BUILD_ARG(itt_sync_obj)); in __kmp_linear_barrier_release()
284 enum barrier_type bt, kmp_info_t *this_thr, int gtid, int tid, in __kmp_linear_barrier_release_cancellable() argument
287 bt, this_thr, gtid, tid, propagate_icvs USE_ITT_BUILD_ARG(itt_sync_obj)); in __kmp_linear_barrier_release_cancellable()
293 int tid, void (*reduce)(void *, void *) in __kmp_tree_barrier_gather() argument
308 gtid, team->t.t_id, tid, bt)); in __kmp_tree_barrier_gather()
320 child_tid = (tid << branch_bits) + 1; in __kmp_tree_barrier_gather()
337 gtid, team->t.t_id, tid, __kmp_gtid_from_tid(child_tid, team), in __kmp_tree_barrier_gather()
354 gtid, team->t.t_id, tid, __kmp_gtid_from_tid(child_tid, team), in __kmp_tree_barrier_gather()
370 if (!KMP_MASTER_TID(tid)) { // Worker threads in __kmp_tree_barrier_gather()
371 kmp_int32 parent_tid = (tid - 1) >> branch_bits; in __kmp_tree_barrier_gather()
376 gtid, team->t.t_id, tid, __kmp_gtid_from_tid(parent_tid, team), in __kmp_tree_barrier_gather()
395 gtid, team->t.t_id, tid, team->t.t_id, in __kmp_tree_barrier_gather()
400 gtid, team->t.t_id, tid, bt)); in __kmp_tree_barrier_gather()
404 enum barrier_type bt, kmp_info_t *this_thr, int gtid, int tid, in __kmp_tree_barrier_release() argument
417 tid)) { // Handle fork barrier workers who aren't part of a team yet in __kmp_tree_barrier_release()
448 tid = __kmp_tid_from_gtid(gtid); in __kmp_tree_barrier_release()
453 team->t.t_id, tid, &thr_bar->b_go, KMP_INIT_BARRIER_STATE)); in __kmp_tree_barrier_release()
460 gtid, team->t.t_id, tid, bt)); in __kmp_tree_barrier_release()
463 child_tid = (tid << branch_bits) + 1; in __kmp_tree_barrier_release()
494 gtid, team->t.t_id, tid, __kmp_gtid_from_tid(child_tid, team), in __kmp_tree_barrier_release()
507 gtid, team->t.t_id, tid, bt)); in __kmp_tree_barrier_release()
513 int tid, void (*reduce)(void *, void *) in __kmp_hyper_barrier_gather() argument
529 gtid, team->t.t_id, tid, bt)); in __kmp_hyper_barrier_gather()
547 if (((tid >> level) & (branch_factor - 1)) != 0) { in __kmp_hyper_barrier_gather()
548 kmp_int32 parent_tid = tid & ~((1 << (level + branch_bits)) - 1); in __kmp_hyper_barrier_gather()
554 gtid, team->t.t_id, tid, __kmp_gtid_from_tid(parent_tid, team), in __kmp_hyper_barrier_gather()
571 for (child = 1, child_tid = tid + (1 << level); in __kmp_hyper_barrier_gather()
586 gtid, team->t.t_id, tid, __kmp_gtid_from_tid(child_tid, team), in __kmp_hyper_barrier_gather()
604 gtid, team->t.t_id, tid, __kmp_gtid_from_tid(child_tid, team), in __kmp_hyper_barrier_gather()
618 if (KMP_MASTER_TID(tid)) { in __kmp_hyper_barrier_gather()
626 gtid, team->t.t_id, tid, team->t.t_id, in __kmp_hyper_barrier_gather()
631 gtid, team->t.t_id, tid, bt)); in __kmp_hyper_barrier_gather()
637 enum barrier_type bt, kmp_info_t *this_thr, int gtid, int tid, in __kmp_hyper_barrier_release() argument
655 if (KMP_MASTER_TID(tid)) { // master in __kmp_hyper_barrier_release()
660 gtid, team->t.t_id, tid, bt)); in __kmp_hyper_barrier_release()
664 &team->t.t_implicit_task_taskdata[tid].td_icvs); in __kmp_hyper_barrier_release()
697 tid = __kmp_tid_from_gtid(gtid); in __kmp_hyper_barrier_release()
702 gtid, team->t.t_id, tid, &thr_bar->b_go, KMP_INIT_BARRIER_STATE)); in __kmp_hyper_barrier_release()
711 offset < num_threads && (((tid >> level) & (branch_factor - 1)) == 0); in __kmp_hyper_barrier_release()
729 child_tid = tid + (child << level); in __kmp_hyper_barrier_release()
732 if (((tid >> level) & (branch_factor - 1)) != 0) in __kmp_hyper_barrier_release()
737 for (child = 1, child_tid = tid + (1 << level); in __kmp_hyper_barrier_release()
768 gtid, team->t.t_id, tid, __kmp_gtid_from_tid(child_tid, team), in __kmp_hyper_barrier_release()
780 !KMP_MASTER_TID(tid)) { // copy ICVs locally to final dest in __kmp_hyper_barrier_release()
781 __kmp_init_implicit_task(team->t.t_ident, team->t.t_threads[tid], team, tid, in __kmp_hyper_barrier_release()
783 copy_icvs(&team->t.t_implicit_task_taskdata[tid].td_icvs, in __kmp_hyper_barrier_release()
790 gtid, team->t.t_id, tid, bt)); in __kmp_hyper_barrier_release()
806 int tid, kmp_team_t *team) { in __kmp_init_hierarchical_barrier_thread() argument
811 bool tid_changed = tid != thr_bar->old_tid; in __kmp_init_hierarchical_barrier_thread()
822 tid)) { // if not master, find parent thread in hierarchy in __kmp_init_hierarchical_barrier_thread()
831 } else if ((rem = tid % thr_bar->skip_per_level[d + 1]) != in __kmp_init_hierarchical_barrier_thread()
834 thr_bar->parent_tid = tid - rem; in __kmp_init_hierarchical_barrier_thread()
841 thr_bar->offset = 7 - (tid - thr_bar->parent_tid - 1); in __kmp_init_hierarchical_barrier_thread()
842 thr_bar->old_tid = tid; in __kmp_init_hierarchical_barrier_thread()
859 if (thr_bar->leaf_kids && (kmp_uint32)tid + thr_bar->leaf_kids + 1 > nproc) in __kmp_init_hierarchical_barrier_thread()
860 thr_bar->leaf_kids = nproc - tid - 1; in __kmp_init_hierarchical_barrier_thread()
869 enum barrier_type bt, kmp_info_t *this_thr, int gtid, int tid, in __kmp_hierarchical_barrier_gather() argument
890 gtid, team->t.t_id, tid, bt)); in __kmp_hierarchical_barrier_gather()
900 (void)__kmp_init_hierarchical_barrier_thread(bt, thr_bar, nproc, gtid, tid, in __kmp_hierarchical_barrier_gather()
912 KMP_MASTER_TID(tid) in __kmp_hierarchical_barrier_gather()
917 gtid, team->t.t_id, tid)); in __kmp_hierarchical_barrier_gather()
924 for (child_tid = tid + 1; child_tid <= tid + thr_bar->leaf_kids; in __kmp_hierarchical_barrier_gather()
928 gtid, team->t.t_id, tid, in __kmp_hierarchical_barrier_gather()
945 kmp_uint32 last = tid + thr_bar->skip_per_level[d + 1], in __kmp_hierarchical_barrier_gather()
949 for (child_tid = tid + skip; child_tid < (int)last; child_tid += skip) { in __kmp_hierarchical_barrier_gather()
955 gtid, team->t.t_id, tid, in __kmp_hierarchical_barrier_gather()
964 gtid, team->t.t_id, tid, in __kmp_hierarchical_barrier_gather()
978 kmp_uint32 last = tid + thr_bar->skip_per_level[d + 1], in __kmp_hierarchical_barrier_gather()
982 for (child_tid = tid + skip; child_tid < (int)last; child_tid += skip) { in __kmp_hierarchical_barrier_gather()
988 gtid, team->t.t_id, tid, in __kmp_hierarchical_barrier_gather()
997 gtid, team->t.t_id, tid, in __kmp_hierarchical_barrier_gather()
1012 if (!KMP_MASTER_TID(tid)) { // worker threads release parent in hierarchy in __kmp_hierarchical_barrier_gather()
1015 gtid, team->t.t_id, tid, in __kmp_hierarchical_barrier_gather()
1040 gtid, team->t.t_id, tid, team->t.t_id, in __kmp_hierarchical_barrier_gather()
1046 gtid, team->t.t_id, tid, bt)); in __kmp_hierarchical_barrier_gather()
1050 enum barrier_type bt, kmp_info_t *this_thr, int gtid, int tid, in __kmp_hierarchical_barrier_release() argument
1058 if (KMP_MASTER_TID(tid)) { in __kmp_hierarchical_barrier_release()
1063 gtid, team->t.t_id, tid, bt)); in __kmp_hierarchical_barrier_release()
1100 tid = __kmp_tid_from_gtid(gtid); in __kmp_hierarchical_barrier_release()
1105 gtid, team->t.t_id, tid, &thr_bar->b_go, KMP_INIT_BARRIER_STATE)); in __kmp_hierarchical_barrier_release()
1129 tid, team); in __kmp_hierarchical_barrier_release()
1136 __kmp_init_implicit_task(team->t.t_ident, team->t.t_threads[tid], team, tid, in __kmp_hierarchical_barrier_release()
1139 tid)) { // master already has copy in final destination; copy in __kmp_hierarchical_barrier_release()
1141 &team->t.t_implicit_task_taskdata[tid].td_icvs); in __kmp_hierarchical_barrier_release()
1147 copy_icvs(&team->t.t_implicit_task_taskdata[tid].td_icvs, in __kmp_hierarchical_barrier_release()
1155 copy_icvs(&team->t.t_implicit_task_taskdata[tid].td_icvs, in __kmp_hierarchical_barrier_release()
1167 if (KMP_MASTER_TID(tid)) { // do a flat release in __kmp_hierarchical_barrier_release()
1183 gtid, team->t.t_id, tid, in __kmp_hierarchical_barrier_release()
1204 last = tid + thr_bar->skip_per_level[1]; in __kmp_hierarchical_barrier_release()
1207 for (child_tid = tid + 1 + old_leaf_kids; child_tid < (int)last; in __kmp_hierarchical_barrier_release()
1215 gtid, team->t.t_id, tid, __kmp_gtid_from_tid(child_tid, team), in __kmp_hierarchical_barrier_release()
1231 last = tid + thr_bar->skip_per_level[d + 1]; in __kmp_hierarchical_barrier_release()
1235 for (child_tid = tid + skip; child_tid < (int)last; child_tid += skip) { in __kmp_hierarchical_barrier_release()
1240 gtid, team->t.t_id, tid, in __kmp_hierarchical_barrier_release()
1252 if (propagate_icvs && !KMP_MASTER_TID(tid)) in __kmp_hierarchical_barrier_release()
1254 copy_icvs(&team->t.t_implicit_task_taskdata[tid].td_icvs, in __kmp_hierarchical_barrier_release()
1260 gtid, team->t.t_id, tid, bt)); in __kmp_hierarchical_barrier_release()
1298 int tid = __kmp_tid_from_gtid(gtid); in __kmp_barrier_template() local
1362 team->t.t_implicit_task_taskdata[tid].td_icvs.bt_intervals; in __kmp_barrier_template()
1364 team->t.t_implicit_task_taskdata[tid].td_icvs.bt_set; in __kmp_barrier_template()
1366 this_thr->th.th_team_bt_intervals = KMP_BLOCKTIME_INTERVAL(team, tid); in __kmp_barrier_template()
1376 if (KMP_MASTER_TID(tid)) { // Master counter is stored in team structure. in __kmp_barrier_template()
1387 if (KMP_MASTER_TID(tid) && __kmp_tasking_mode != tskm_immediate_exec) in __kmp_barrier_template()
1393 bt, this_thr, gtid, tid, reduce USE_ITT_BUILD_ARG(itt_sync_obj)); in __kmp_barrier_template()
1399 __kmp_hyper_barrier_gather(bt, this_thr, gtid, tid, in __kmp_barrier_template()
1405 bt, this_thr, gtid, tid, reduce USE_ITT_BUILD_ARG(itt_sync_obj)); in __kmp_barrier_template()
1411 __kmp_tree_barrier_gather(bt, this_thr, gtid, tid, in __kmp_barrier_template()
1416 __kmp_linear_barrier_gather(bt, this_thr, gtid, tid, in __kmp_barrier_template()
1424 if (KMP_MASTER_TID(tid)) { in __kmp_barrier_template()
1506 bt, this_thr, gtid, tid, FALSE USE_ITT_BUILD_ARG(itt_sync_obj)); in __kmp_barrier_template()
1511 __kmp_hyper_barrier_release(bt, this_thr, gtid, tid, in __kmp_barrier_template()
1517 bt, this_thr, gtid, tid, FALSE USE_ITT_BUILD_ARG(itt_sync_obj)); in __kmp_barrier_template()
1522 __kmp_tree_barrier_release(bt, this_thr, gtid, tid, in __kmp_barrier_template()
1527 __kmp_linear_barrier_release(bt, this_thr, gtid, tid, in __kmp_barrier_template()
1611 int tid = __kmp_tid_from_gtid(gtid); in __kmp_barrier_gomp_cancel() local
1613 if (KMP_MASTER_TID(tid)) { in __kmp_barrier_gomp_cancel()
1631 int tid = __kmp_tid_from_gtid(gtid); in __kmp_end_split_barrier() local
1641 __kmp_hyper_barrier_release(bt, this_thr, gtid, tid, in __kmp_end_split_barrier()
1646 __kmp_hierarchical_barrier_release(bt, this_thr, gtid, tid, in __kmp_end_split_barrier()
1652 __kmp_tree_barrier_release(bt, this_thr, gtid, tid, in __kmp_end_split_barrier()
1657 __kmp_linear_barrier_release(bt, this_thr, gtid, tid, in __kmp_end_split_barrier()
1676 int tid; in __kmp_join_barrier() local
1694 tid = __kmp_tid_from_gtid(gtid); in __kmp_join_barrier()
1711 KMP_DEBUG_ASSERT(this_thr == team->t.t_threads[tid]); in __kmp_join_barrier()
1713 gtid, team_id, tid)); in __kmp_join_barrier()
1749 team_id, tid)); in __kmp_join_barrier()
1771 team->t.t_implicit_task_taskdata[tid].td_icvs.bt_intervals; in __kmp_join_barrier()
1773 team->t.t_implicit_task_taskdata[tid].td_icvs.bt_set; in __kmp_join_barrier()
1775 this_thr->th.th_team_bt_intervals = KMP_BLOCKTIME_INTERVAL(team, tid); in __kmp_join_barrier()
1787 __kmp_hyper_barrier_gather(bs_forkjoin_barrier, this_thr, gtid, tid, in __kmp_join_barrier()
1792 __kmp_hierarchical_barrier_gather(bs_forkjoin_barrier, this_thr, gtid, tid, in __kmp_join_barrier()
1798 __kmp_tree_barrier_gather(bs_forkjoin_barrier, this_thr, gtid, tid, in __kmp_join_barrier()
1803 __kmp_linear_barrier_gather(bs_forkjoin_barrier, this_thr, gtid, tid, in __kmp_join_barrier()
1812 if (KMP_MASTER_TID(tid)) { in __kmp_join_barrier()
1890 if (KMP_MASTER_TID(tid)) { in __kmp_join_barrier()
1894 gtid, team_id, tid, nproc)); in __kmp_join_barrier()
1901 ("__kmp_join_barrier: T#%d(%d:%d) leaving\n", gtid, team_id, tid)); in __kmp_join_barrier()
1908 void __kmp_fork_barrier(int gtid, int tid) { in __kmp_fork_barrier() argument
1912 kmp_team_t *team = (tid == 0) ? this_thr->th.th_team : NULL; in __kmp_fork_barrier()
1920 (team != NULL) ? team->t.t_id : -1, tid)); in __kmp_fork_barrier()
1923 if (KMP_MASTER_TID(tid)) { in __kmp_fork_barrier()
1967 team->t.t_implicit_task_taskdata[tid].td_icvs.bt_intervals; in __kmp_fork_barrier()
1969 team->t.t_implicit_task_taskdata[tid].td_icvs.bt_set; in __kmp_fork_barrier()
1971 this_thr->th.th_team_bt_intervals = KMP_BLOCKTIME_INTERVAL(team, tid); in __kmp_fork_barrier()
1979 __kmp_hyper_barrier_release(bs_forkjoin_barrier, this_thr, gtid, tid, in __kmp_fork_barrier()
1984 __kmp_hierarchical_barrier_release(bs_forkjoin_barrier, this_thr, gtid, tid, in __kmp_fork_barrier()
1990 __kmp_tree_barrier_release(bs_forkjoin_barrier, this_thr, gtid, tid, in __kmp_fork_barrier()
1995 __kmp_linear_barrier_release(bs_forkjoin_barrier, this_thr, gtid, tid, in __kmp_fork_barrier()
2038 if (!KMP_MASTER_TID(tid)) { in __kmp_fork_barrier()
2055 tid = __kmp_tid_from_gtid(gtid); in __kmp_fork_barrier()
2066 if (!KMP_MASTER_TID(tid)) { // master thread already has ICVs in __kmp_fork_barrier()
2070 ("__kmp_fork_barrier: T#%d(%d) is PULLing ICVs\n", gtid, tid)); in __kmp_fork_barrier()
2071 __kmp_init_implicit_task(team->t.t_ident, team->t.t_threads[tid], team, in __kmp_fork_barrier()
2072 tid, FALSE); in __kmp_fork_barrier()
2073 copy_icvs(&team->t.t_implicit_task_taskdata[tid].td_icvs, in __kmp_fork_barrier()
2115 if (!KMP_MASTER_TID(tid)) in __kmp_fork_barrier()
2120 if (!KMP_MASTER_TID(tid)) { in __kmp_fork_barrier()
2129 team->t.t_id, tid)); in __kmp_fork_barrier()