/external/llvm-project/openmp/runtime/src/ |
D | ompt-specific.h | 72 kmp_info_t *thr = __kmp_threads[gtid]; in __ompt_load_return_address() 87 ((ompt_enabled.enabled && gtid >= 0 && __kmp_threads[gtid] && \ 88 __kmp_threads[gtid]->th.ompt_thread_info.return_address)? \ 120 if (ompt_enabled.enabled && Gtid >= 0 && __kmp_threads[Gtid] && in OmptReturnAddressGuard() 121 !__kmp_threads[Gtid]->th.ompt_thread_info.return_address) { in OmptReturnAddressGuard() 123 __kmp_threads[Gtid]->th.ompt_thread_info.return_address = ReturnAddress; in OmptReturnAddressGuard() 128 __kmp_threads[Gtid]->th.ompt_thread_info.return_address = NULL; in ~OmptReturnAddressGuard()
|
D | kmp_threadprivate.cpp | 181 if (__kmp_threads[gtid]) { in __kmp_common_destroy() 185 __kmp_threads[gtid]->th.th_pri_common, gtid, in __kmp_common_destroy() 200 if (__kmp_threads[gtid]) { in __kmp_common_destroy() 204 __kmp_threads[gtid]->th.th_pri_common, gtid, in __kmp_common_destroy() 243 for (tn = __kmp_threads[gtid]->th.th_pri_head; tn; tn = tn->link) { in __kmp_common_destroy_gtid() 277 if (!__kmp_threads[p]) in dump_list() 280 if (__kmp_threads[p]->th.th_pri_common->data[q]) { in dump_list() 285 for (tn = __kmp_threads[p]->th.th_pri_common->data[q]; tn; in dump_list() 301 KMP_DEBUG_ASSERT(__kmp_threads[gtid] && in kmp_threadprivate_insert_private_data() 302 __kmp_threads[gtid]->th.th_root->r.r_active == 0); in kmp_threadprivate_insert_private_data() [all …]
|
D | kmp_error.cpp | 189 struct cons_header *p = __kmp_threads[gtid]->th.th_cons; in __kmp_push_parallel() 191 KMP_DEBUG_ASSERT(__kmp_threads[gtid]->th.th_cons); in __kmp_push_parallel() 207 struct cons_header *p = __kmp_threads[gtid]->th.th_cons; in __kmp_check_workshare() 209 KMP_DEBUG_ASSERT(__kmp_threads[gtid]->th.th_cons); in __kmp_check_workshare() 229 struct cons_header *p = __kmp_threads[gtid]->th.th_cons; in __kmp_push_workshare() 249 struct cons_header *p = __kmp_threads[gtid]->th.th_cons; 336 struct cons_header *p = __kmp_threads[gtid]->th.th_cons; 359 struct cons_header *p = __kmp_threads[gtid]->th.th_cons; 380 struct cons_header *p = __kmp_threads[gtid]->th.th_cons; 407 struct cons_header *p = __kmp_threads[gtid]->th.th_cons; [all …]
|
D | kmp_runtime.cpp | 141 other_threads = __kmp_threads; in __kmp_get_global_thread_id() 298 kmp_info_t *f_th = (kmp_info_t *)TCR_SYNC_PTR(__kmp_threads[f]); in __kmp_check_stack_overlap() 570 if (!__kmp_threads) in __kmp_reset_locks_on_process_detach() 572 kmp_info_t *th = __kmp_threads[i]; in __kmp_reset_locks_on_process_detach() 672 if (__kmp_threads[gtid]->th.th_root->r.r_active) in __kmp_parallel_deo() 698 if (__kmp_threads[gtid]->th.th_root->r.r_active) in __kmp_parallel_dxo() 726 th = __kmp_threads[gtid]; in __kmp_enter_single() 915 if (TCR_PTR(__kmp_threads[0]) == NULL) { in __kmp_reserve_threads() 1156 this_thr = __kmp_threads[global_tid]; in __kmp_serialized_parallel() 1441 master_th = __kmp_threads[gtid]; // AC: potentially unsafe, not in sync with in __kmp_fork_call() [all …]
|
D | kmp_cancel.cpp | 30 kmp_info_t *this_thr = __kmp_threads[gtid]; in __kmpc_cancel() 137 kmp_info_t *this_thr = __kmp_threads[gtid]; in __kmpc_cancellationpoint() 245 kmp_info_t *this_thr = __kmp_threads[gtid]; in __kmpc_cancel_barrier()
|
D | kmp_csupport.cpp | 289 kmp_info_t *master_th = __kmp_threads[gtid]; in __kmpc_fork_call() 367 kmp_info_t *this_thr = __kmp_threads[gtid]; in __kmpc_fork_teams() 503 this_thr = __kmp_threads[global_tid]; in __kmpc_end_serialized_parallel() 719 __kmp_threads[global_tid]->th.th_ident = loc; in __kmpc_barrier() 762 kmp_info_t *this_thr = __kmp_threads[global_tid]; in __kmpc_master() 806 kmp_info_t *this_thr = __kmp_threads[global_tid]; in __kmpc_end_master() 848 th = __kmp_threads[gtid]; in __kmpc_ordered() 915 th = __kmp_threads[gtid]; in __kmpc_end_ordered() 1190 ti = __kmp_threads[global_tid]->th.ompt_thread_info; in __kmpc_critical() 1405 ti = __kmp_threads[global_tid]->th.ompt_thread_info; in __kmpc_critical_with_hint() [all …]
|
D | kmp_dispatch.h | 324 kmp_info_t *th = __kmp_threads[gtid]; in __kmp_dispatch_deo() 355 struct cons_header *p = __kmp_threads[gtid]->th.th_cons; in __kmp_dispatch_deo() 400 kmp_info_t *th = __kmp_threads[gtid]; in __kmp_dispatch_dxo() 426 struct cons_header *p = __kmp_threads[gtid]->th.th_cons; in __kmp_dispatch_dxo()
|
D | kmp_barrier.cpp | 165 team = __kmp_threads[gtid]->th.th_team; in __kmp_linear_barrier_release_template() 246 team = __kmp_threads[gtid]->th.th_team; in __kmp_linear_barrier_release_template() 446 team = __kmp_threads[gtid]->th.th_team; in __kmp_tree_barrier_release() 456 team = __kmp_threads[gtid]->th.th_team; in __kmp_tree_barrier_release() 656 team = __kmp_threads[gtid]->th.th_team; in __kmp_hyper_barrier_release() 695 team = __kmp_threads[gtid]->th.th_team; in __kmp_hyper_barrier_release() 1059 team = __kmp_threads[gtid]->th.th_team; in __kmp_hierarchical_barrier_release() 1098 team = __kmp_threads[gtid]->th.th_team; in __kmp_hierarchical_barrier_release() 1299 kmp_info_t *this_thr = __kmp_threads[gtid]; in __kmp_barrier_template() 1458 ident_t *loc = __kmp_threads[gtid]->th.th_ident; in __kmp_barrier_template() [all …]
|
D | kmp_tasking.cpp | 326 kmp_info_t *thread = __kmp_threads[gtid]; in __kmp_push_task() 500 kmp_info_t *thread = __kmp_threads[gtid]; in __kmp_task_start() 564 if (__kmp_threads[gtid]->th.ompt_thread_info.ompt_task_yielded) { in __ompt_task_start() 566 __kmp_threads[gtid]->th.ompt_thread_info.ompt_task_yielded = 0; in __ompt_task_start() 603 kmp_taskdata_t *current_task = __kmp_threads[gtid]->th.th_current_task; in __kmpc_omp_task_begin_if0_template() 682 kmp_taskdata_t *current_task = __kmp_threads[gtid]->th.th_current_task; in __kmpc_omp_task_begin() 809 kmp_info_t *thread = __kmp_threads[gtid]; in __kmp_task_finish() 1173 kmp_info_t *thread = __kmp_threads[gtid]; in __kmp_task_alloc() 1460 thread = __kmp_threads[gtid]; in __kmp_invoke_task() 1480 thread = __kmp_threads[gtid]; in __kmp_invoke_task() [all …]
|
D | kmp_gsupport.cpp | 196 kmp_info_t *this_thr = __kmp_threads[gtid]; in KMP_EXPAND_NAME() 360 thr = __kmp_threads[*gtid]; in __kmp_GOMP_microtask_wrapper() 406 thr = __kmp_threads[*gtid]; in __kmp_GOMP_parallel_microtask_wrapper() 435 kmp_info_t *thr = __kmp_threads[gtid]; in __kmp_GOMP_fork_call() 506 thr = __kmp_threads[gtid]; in KMP_EXPAND_NAME() 644 if (!status && __kmp_threads[gtid]->th.th_dispatch->th_doacross_flags) { \ 1234 current_task = __kmp_threads[gtid]->th.th_current_task; 1259 thread = __kmp_threads[gtid]; 1790 kmp_info_t *th = __kmp_threads[gtid]; in __kmp_GOMP_doacross_post() 1813 kmp_info_t *th = __kmp_threads[gtid]; in __kmp_GOMP_doacross_wait()
|
D | kmp_taskdeps.cpp | 231 sink_data = &__kmp_threads[gtid]->th.ompt_thread_info.task_data; in __kmp_track_dependence() 300 kmp_info_t *thread = __kmp_threads[gtid]; in __kmp_process_deps() 518 kmp_info_t *thread = __kmp_threads[gtid]; in __kmpc_omp_task_with_deps() 680 kmp_info_t *thread = __kmp_threads[gtid]; in __kmpc_omp_wait_deps()
|
D | kmp_taskdeps.h | 89 kmp_info_t *thread = __kmp_threads[gtid]; in __kmp_release_deps()
|
D | kmp_ftn_entry.h | 127 team = __kmp_threads[gtid]->th.th_team; in FTN_GET_BLOCKTIME() 343 thread = __kmp_threads[gtid]; in KMP_EXPAND_NAME() 358 kmp_info_t *this_thr = __kmp_threads[__kmp_entry_gtid()]; in FTN_CONTROL_TOOL() 745 thread = __kmp_threads[gtid]; in KMP_EXPAND_NAME()
|
D | kmp_dispatch.cpp | 48 th = __kmp_threads[*gtid_ref]; in __kmp_dispatch_deo_error() 64 th = __kmp_threads[*gtid_ref]; in __kmp_dispatch_dxo_error() 139 th = __kmp_threads[gtid]; in __kmp_dispatch_init_algorithm() 808 th = __kmp_threads[gtid]; in __kmp_dispatch_init() 1010 kmp_info_t *th = __kmp_threads[gtid]; in __kmp_dispatch_finish() 1074 kmp_info_t *th = __kmp_threads[gtid]; in __kmp_dispatch_finish_chunk() 1165 kmp_info_t *th = __kmp_threads[gtid]; in __kmp_dispatch_next_algorithm() 1915 kmp_info_t *th = __kmp_threads[gtid]; in __kmp_dispatch_next() 2209 th = __kmp_threads[gtid]; in __kmp_dist_get_bounds()
|
D | kmp.h | 3113 extern kmp_info_t **__kmp_threads; /* Descriptors for the threads */ 3133 #define __kmp_get_team() (__kmp_threads[(__kmp_get_gtid())]->th.th_team) 3140 (__kmp_threads[(gtid)]->th.th_team->t.t_nproc) 3145 return (gtid >= 0 && __kmp_root[gtid] && __kmp_threads[gtid] && in KMP_UBER_GTID() 3146 __kmp_threads[gtid] == __kmp_root[gtid]->r.r_uber_thread); in KMP_UBER_GTID() 3151 return __kmp_threads[gtid]->th.th_info.ds.ds_tid; in __kmp_tid_from_gtid() 3166 return __kmp_threads[gtid]; in __kmp_thread_from_gtid() 3171 return __kmp_threads[gtid]->th.th_team; in __kmp_team_from_gtid() 3254 return __kmp_threads[gtid]; in __kmp_entry_thread()
|
D | kmp_sched.cpp | 95 kmp_info_t *th = __kmp_threads[gtid]; in __kmp_for_static_init() 485 th = __kmp_threads[gtid]; in __kmp_dist_for_static_init() 728 th = __kmp_threads[gtid]; in __kmp_team_static_init()
|
D | kmp_debugger.cpp | 69 addr_and_size_of(__kmp_threads),
|
D | kmp_global.cpp | 425 kmp_info_t **__kmp_threads = NULL; variable
|
D | kmp_dispatch_hier.h | 482 kmp_info_t *th = __kmp_threads[gtid]; in next_recurse() 714 kmp_info_t *th = __kmp_threads[gtid]; in next() 946 th = __kmp_threads[gtid]; in __kmp_dispatch_init_hierarchy()
|
D | z_Windows_NT_util.cpp | 358 kmp_info_t *th = __kmp_threads[th_gtid]; in __kmp_suspend_template() 464 kmp_info_t *th = __kmp_threads[target_gtid]; in __kmp_resume_template() 884 kmp_info_t *th = __kmp_threads[gtid]; in __kmp_terminate_thread()
|
D | kmp_alloc.cpp | 1429 __kmp_threads[gtid]->th.th_def_allocator = allocator; in __kmpc_set_default_allocator() 1433 return __kmp_threads[gtid]->th.th_def_allocator; in __kmpc_get_default_allocator() 1454 allocator = __kmp_threads[gtid]->th.th_def_allocator; in __kmpc_alloc() 1591 allocator = __kmp_threads[gtid]->th.th_def_allocator; in __kmpc_calloc()
|
D | z_Linux_util.cpp | 452 kmp_info_t *th = __kmp_threads[gtid]; in __kmp_terminate_thread() 1454 kmp_info_t *th = __kmp_threads[th_gtid]; in __kmp_suspend_template() 1603 kmp_info_t *th = __kmp_threads[target_gtid]; in __kmp_resume_template()
|
D | kmp_wait_release.h | 500 kmp_info_t *th = __kmp_threads[th_gtid]; in __kmp_mwait_template()
|
D | kmp_affinity.cpp | 4719 kmp_info_t *th = (kmp_info_t *)TCR_SYNC_PTR(__kmp_threads[gtid]); in __kmp_affinity_set_init_mask() 4818 kmp_info_t *th = (kmp_info_t *)TCR_SYNC_PTR(__kmp_threads[gtid]); in __kmp_affinity_set_place() 4900 th = __kmp_threads[gtid]; in __kmp_aux_set_affinity() 4928 th = __kmp_threads[gtid]; in __kmp_aux_get_affinity()
|
D | kmp_lock.cpp | 1063 t = __kmp_threads[lck->lk.head_id - 1]->th.th_next_waiting; in __kmp_dump_queuing_lock() 1066 t = __kmp_threads[t - 1]->th.th_next_waiting; in __kmp_dump_queuing_lock()
|