/external/llvm-project/openmp/runtime/src/ |
D | ompt-specific.h | 26 void __ompt_lw_taskteam_init(ompt_lw_taskteam_t *lwt, kmp_info_t *thr, 29 void __ompt_lw_taskteam_link(ompt_lw_taskteam_t *lwt, kmp_info_t *thr, 32 void __ompt_lw_taskteam_unlink(kmp_info_t *thr); 54 ompt_sync_region_t __ompt_get_barrier_kind(enum barrier_type, kmp_info_t *); 72 kmp_info_t *thr = __kmp_threads[gtid]; in __ompt_load_return_address() 96 inline kmp_info_t *ompt_get_thread_gtid(int gtid) { in ompt_get_thread_gtid() 100 inline kmp_info_t *ompt_get_thread() { in ompt_get_thread() 105 inline void ompt_set_thread_state(kmp_info_t *thread, ompt_state_t state) { in ompt_set_thread_state()
|
D | ompt-specific.cpp | 52 kmp_info_t *thr = ompt_get_thread(); in __ompt_get_teaminfo() 105 kmp_info_t *thr = ompt_get_thread(); in __ompt_get_task_info_object() 145 kmp_info_t *thr = ompt_get_thread(); in __ompt_get_scheduling_taskinfo() 198 kmp_info_t *thread = ompt_get_thread(); in __ompt_get_thread_data_internal() 211 kmp_info_t *ti = ompt_get_thread(); in __ompt_thread_assign_wait_id() 218 kmp_info_t *ti = ompt_get_thread(); in __ompt_get_state_internal() 255 void __ompt_lw_taskteam_init(ompt_lw_taskteam_t *lwt, kmp_info_t *thr, int gtid, in __ompt_lw_taskteam_init() 269 void __ompt_lw_taskteam_link(ompt_lw_taskteam_t *lwt, kmp_info_t *thr, in __ompt_lw_taskteam_link() 303 void __ompt_lw_taskteam_unlink(kmp_info_t *thr) { in __ompt_lw_taskteam_unlink() 342 kmp_info_t *thr = ompt_get_thread(); in __ompt_get_task_info_internal() [all …]
|
D | kmp_wait_release.h | 130 static void __ompt_implicit_task_end(kmp_info_t *this_thr, in __ompt_implicit_task_end() 174 __kmp_wait_template(kmp_info_t *this_thr, in __kmp_wait_template() 500 kmp_info_t *th = __kmp_threads[th_gtid]; in __kmp_mwait_template() 594 kmp_info_t *waiter = flag->get_waiter(i); in __kmp_release_template() 646 kmp_info_t 653 kmp_basic_flag_native(volatile FlagType *p, kmp_info_t *thr) 664 kmp_info_t *get_waiter(kmp_uint32 i) { 677 void set_waiter(kmp_info_t *thr) { 750 kmp_info_t 757 kmp_basic_flag(std::atomic<FlagType> *p, kmp_info_t *thr) [all …]
|
D | kmp.h | 2651 } kmp_info_t; typedef 2740 KMP_ALIGN_CACHE kmp_info_t **t_threads; 2825 kmp_info_t *r_uber_thread; 3113 extern kmp_info_t **__kmp_threads; /* Descriptors for the threads */ 3116 extern volatile kmp_info_t *__kmp_thread_pool; 3117 extern kmp_info_t *__kmp_thread_pool_insert_pt; 3159 static inline int __kmp_gtid_from_thread(const kmp_info_t *thr) { in __kmp_gtid_from_thread() 3164 static inline kmp_info_t *__kmp_thread_from_gtid(int gtid) { in __kmp_thread_from_gtid() 3190 extern kmp_info_t __kmp_monitor; 3251 static inline kmp_info_t *__kmp_entry_thread() { in __kmp_entry_thread() [all …]
|
D | kmp_taskdeps.h | 23 static inline void __kmp_node_deref(kmp_info_t *thread, kmp_depnode_t *node) { in __kmp_node_deref() 38 static inline void __kmp_depnode_list_free(kmp_info_t *thread, in __kmp_depnode_list_free() 54 static inline void __kmp_dephash_free_entries(kmp_info_t *thread, in __kmp_dephash_free_entries() 79 static inline void __kmp_dephash_free(kmp_info_t *thread, kmp_dephash_t *h) { in __kmp_dephash_free() 89 kmp_info_t *thread = __kmp_threads[gtid]; in __kmp_release_deps()
|
D | kmp_alloc.cpp | 46 static void bpool(kmp_info_t *th, void *buffer, bufsize len); 47 static void *bget(kmp_info_t *th, bufsize size); 48 static void *bgetz(kmp_info_t *th, bufsize size); 49 static void *bgetr(kmp_info_t *th, void *buffer, bufsize newsize); 50 static void brel(kmp_info_t *th, void *buf); 51 static void bectl(kmp_info_t *th, bget_compact_t compact, 138 kmp_info_t *bthr; /* The thread which owns the buffer pool */ 227 static void set_thr_data(kmp_info_t *th) { in set_thr_data() 253 static thr_data_t *get_thr_data(kmp_info_t *th) { in get_thr_data() 264 static void __kmp_bget_dequeue(kmp_info_t *th) { in __kmp_bget_dequeue() [all …]
|
D | kmp_runtime.cpp | 71 kmp_info_t __kmp_monitor; 78 static void __kmp_initialize_info(kmp_info_t *, kmp_team_t *, int tid, 101 static void __kmp_reap_thread(kmp_info_t *thread, int is_root); 102 kmp_info_t *__kmp_thread_pool_insert_pt = NULL; 109 kmp_info_t **other_threads; in __kmp_get_global_thread_id() 156 kmp_info_t *thr = (kmp_info_t *)TCR_SYNC_PTR(other_threads[i]); in __kmp_get_global_thread_id() 260 void __kmp_check_stack_overlap(kmp_info_t *th) { in __kmp_check_stack_overlap() 298 kmp_info_t *f_th = (kmp_info_t *)TCR_SYNC_PTR(__kmp_threads[f]); in __kmp_check_stack_overlap() 461 static void __kmp_print_thread_storage_map(kmp_info_t *thr, int gtid) { in __kmp_print_thread_storage_map() 462 __kmp_print_storage_map_gtid(gtid, thr, thr + 1, sizeof(kmp_info_t), "th_%d", in __kmp_print_thread_storage_map() [all …]
|
D | z_Windows_NT_util.cpp | 198 kmp_info_t *th, C *flag) { in __kmp_win32_cond_wait() 312 void __kmp_suspend_initialize_thread(kmp_info_t *th) { in __kmp_suspend_initialize_thread() 332 void __kmp_suspend_uninitialize_thread(kmp_info_t *th) { in __kmp_suspend_uninitialize_thread() 342 int __kmp_try_suspend_mx(kmp_info_t *th) { in __kmp_try_suspend_mx() 346 void __kmp_lock_suspend_mx(kmp_info_t *th) { in __kmp_lock_suspend_mx() 350 void __kmp_unlock_suspend_mx(kmp_info_t *th) { in __kmp_unlock_suspend_mx() 358 kmp_info_t *th = __kmp_threads[th_gtid]; in __kmp_suspend_template() 464 kmp_info_t *th = __kmp_threads[target_gtid]; in __kmp_resume_template() 884 kmp_info_t *th = __kmp_threads[gtid]; in __kmp_terminate_thread() 958 kmp_info_t *this_thr = (kmp_info_t *)arg; in __kmp_launch_worker() [all …]
|
D | kmp_barrier.cpp | 47 enum barrier_type bt, kmp_info_t *this_thr, int gtid, int tid, in __kmp_linear_barrier_gather_template() 52 kmp_info_t **other_threads = team->t.t_threads; in __kmp_linear_barrier_gather_template() 154 enum barrier_type bt, kmp_info_t *this_thr, int gtid, int tid, in __kmp_linear_barrier_release_template() 163 kmp_info_t **other_threads; in __kmp_linear_barrier_release_template() 263 enum barrier_type bt, kmp_info_t *this_thr, int gtid, int tid, in __kmp_linear_barrier_gather() 270 enum barrier_type bt, kmp_info_t *this_thr, int gtid, int tid, in __kmp_linear_barrier_gather_cancellable() 277 enum barrier_type bt, kmp_info_t *this_thr, int gtid, int tid, in __kmp_linear_barrier_release() 284 enum barrier_type bt, kmp_info_t *this_thr, int gtid, int tid, in __kmp_linear_barrier_release_cancellable() 292 __kmp_tree_barrier_gather(enum barrier_type bt, kmp_info_t *this_thr, int gtid, in __kmp_tree_barrier_gather() 298 kmp_info_t **other_threads = team->t.t_threads; in __kmp_tree_barrier_gather() [all …]
|
D | kmp_tasking.cpp | 28 kmp_info_t *this_thr); 29 static void __kmp_alloc_task_deque(kmp_info_t *thread, 31 static int __kmp_realloc_task_threads_data(kmp_info_t *thread, 148 static void __kmp_push_task_stack(kmp_int32 gtid, kmp_info_t *thread, in __kmp_push_task_stack() 207 static void __kmp_pop_task_stack(kmp_int32 gtid, kmp_info_t *thread, in __kmp_pop_task_stack() 298 static void __kmp_realloc_task_deque(kmp_info_t *thread, in __kmp_realloc_task_deque() 326 kmp_info_t *thread = __kmp_threads[gtid]; in __kmp_push_task() 439 void __kmp_pop_current_task_from_thread(kmp_info_t *this_thr) { in __kmp_pop_current_task_from_thread() 461 void __kmp_push_current_task_to_thread(kmp_info_t *this_thr, kmp_team_t *team, in __kmp_push_current_task_to_thread() 500 kmp_info_t *thread = __kmp_threads[gtid]; in __kmp_task_start() [all …]
|
D | kmp_ftn_entry.h | 106 kmp_info_t *thread; in FTN_SET_BLOCKTIME() 121 kmp_info_t *thread; in FTN_GET_BLOCKTIME() 338 kmp_info_t *thread; in KMP_EXPAND_NAME() 358 kmp_info_t *this_thr = __kmp_threads[__kmp_entry_gtid()]; in FTN_CONTROL_TOOL() 418 kmp_info_t *th; 599 kmp_info_t *thread; in KMP_EXPAND_NAME() 616 kmp_info_t *thread; in KMP_EXPAND_NAME() 626 kmp_info_t *thread; in KMP_EXPAND_NAME() 639 kmp_info_t *thread; in KMP_EXPAND_NAME() 649 kmp_info_t *th = __kmp_entry_thread(); in KMP_EXPAND_NAME() [all …]
|
D | kmp_cancel.cpp | 30 kmp_info_t *this_thr = __kmp_threads[gtid]; in __kmpc_cancel() 137 kmp_info_t *this_thr = __kmp_threads[gtid]; in __kmpc_cancellationpoint() 245 kmp_info_t *this_thr = __kmp_threads[gtid]; in __kmpc_cancel_barrier() 311 kmp_info_t *this_thr = __kmp_entry_thread(); in __kmp_get_cancellation_status()
|
D | z_Linux_util.cpp | 452 kmp_info_t *th = __kmp_threads[gtid]; in __kmp_terminate_thread() 472 static kmp_int32 __kmp_set_stack_info(int gtid, kmp_info_t *th) { in __kmp_set_stack_info() 534 gtid = ((kmp_info_t *)thr)->th.th_info.ds.ds_gtid; in __kmp_launch_worker() 541 __kmp_stats_thread_ptr = ((kmp_info_t *)thr)->th.th_stats; in __kmp_launch_worker() 585 __kmp_set_stack_info(gtid, (kmp_info_t *)thr); in __kmp_launch_worker() 587 __kmp_check_stack_overlap((kmp_info_t *)thr); in __kmp_launch_worker() 589 exit_val = __kmp_launch_thread((kmp_info_t *)thr); in __kmp_launch_worker() 626 __kmp_set_stack_info(((kmp_info_t *)thr)->th.th_info.ds.ds_gtid, in __kmp_launch_monitor() 627 (kmp_info_t *)thr); in __kmp_launch_monitor() 629 __kmp_check_stack_overlap((kmp_info_t *)thr); in __kmp_launch_monitor() [all …]
|
D | kmp_taskdeps.cpp | 66 static kmp_dephash_t *__kmp_dephash_extend(kmp_info_t *thread, in __kmp_dephash_extend() 115 static kmp_dephash_t *__kmp_dephash_create(kmp_info_t *thread, in __kmp_dephash_create() 151 __kmp_dephash_find(kmp_info_t *thread, kmp_dephash_t **hash, kmp_intptr_t addr) { in __kmp_dephash_find() 189 static kmp_depnode_list_t *__kmp_add_node(kmp_info_t *thread, in __kmp_add_node() 240 __kmp_depnode_link_successor(kmp_int32 gtid, kmp_info_t *thread, in __kmp_depnode_link_successor() 267 kmp_info_t *thread, in __kmp_depnode_link_successor() 300 kmp_info_t *thread = __kmp_threads[gtid]; in __kmp_process_deps() 518 kmp_info_t *thread = __kmp_threads[gtid]; in __kmpc_omp_task_with_deps() 680 kmp_info_t *thread = __kmp_threads[gtid]; in __kmpc_omp_wait_deps()
|
D | kmp_csupport.cpp | 289 kmp_info_t *master_th = __kmp_threads[gtid]; in __kmpc_fork_call() 367 kmp_info_t *this_thr = __kmp_threads[gtid]; in __kmpc_fork_teams() 485 kmp_info_t *this_thr; in __kmpc_end_serialized_parallel() 762 kmp_info_t *this_thr = __kmp_threads[global_tid]; in __kmpc_master() 806 kmp_info_t *this_thr = __kmp_threads[global_tid]; in __kmpc_end_master() 832 kmp_info_t *th; in __kmpc_ordered() 905 kmp_info_t *th; in __kmpc_end_ordered() 1725 kmp_info_t *this_thr = __kmp_threads[global_tid]; in __kmpc_single() 1773 kmp_info_t *this_thr = __kmp_threads[global_tid]; in __kmpc_end_single() 1835 kmp_info_t *thread; in ompc_set_dynamic() [all …]
|
D | kmp_wait_release.cpp | 15 void __kmp_wait_64(kmp_info_t *this_thr, kmp_flag_64<> *flag, in __kmp_wait_64()
|
D | ompt-general.cpp | 470 kmp_info_t *root_thread = ompt_get_thread(); 710 kmp_info_t *thread; 731 kmp_info_t *thread;
|
D | kmp_global.cpp | 425 kmp_info_t **__kmp_threads = NULL; 432 volatile kmp_info_t *__kmp_thread_pool = NULL;
|
D | kmp_dispatch.cpp | 43 kmp_info_t *th; in __kmp_dispatch_deo_error() 61 kmp_info_t *th; in __kmp_dispatch_dxo_error() 118 kmp_info_t *th; in __kmp_dispatch_init_algorithm() 774 kmp_info_t *th; in __kmp_dispatch_init() 1010 kmp_info_t *th = __kmp_threads[gtid]; in __kmp_dispatch_finish() 1074 kmp_info_t *th = __kmp_threads[gtid]; in __kmp_dispatch_finish_chunk() 1165 kmp_info_t *th = __kmp_threads[gtid]; in __kmp_dispatch_next_algorithm() 1221 kmp_info_t **other_threads = team->t.t_threads; in __kmp_dispatch_next_algorithm() 1320 kmp_info_t **other_threads = team->t.t_threads; in __kmp_dispatch_next_algorithm() 1915 kmp_info_t *th = __kmp_threads[gtid]; in __kmp_dispatch_next() [all …]
|
D | kmp_gsupport.cpp | 196 kmp_info_t *this_thr = __kmp_threads[gtid]; in KMP_EXPAND_NAME() 354 kmp_info_t *thr; in __kmp_GOMP_microtask_wrapper() 401 kmp_info_t *thr; in __kmp_GOMP_parallel_microtask_wrapper() 435 kmp_info_t *thr = __kmp_threads[gtid]; in __kmp_GOMP_fork_call() 504 kmp_info_t *thr; in KMP_EXPAND_NAME() 1255 kmp_info_t *thread; 1790 kmp_info_t *th = __kmp_threads[gtid]; in __kmp_GOMP_doacross_post() 1813 kmp_info_t *th = __kmp_threads[gtid]; in __kmp_GOMP_doacross_wait()
|
D | kmp_itt.inl | 482 kmp_info_t *thr = __kmp_thread_from_gtid(gtid); 625 kmp_info_t *thread = __kmp_thread_from_gtid(gtid); 637 kmp_info_t *thread = __kmp_thread_from_gtid(gtid); 859 kmp_info_t *thr = __kmp_thread_from_gtid((gtid)); 903 kmp_info_t *thr = __kmp_thread_from_gtid(gtid); 917 kmp_info_t *th = __kmp_thread_from_gtid(gtid); 929 kmp_info_t *th = __kmp_thread_from_gtid(gtid); 941 kmp_info_t *th = __kmp_thread_from_gtid(gtid);
|
D | kmp_dispatch.h | 324 kmp_info_t *th = __kmp_threads[gtid]; in __kmp_dispatch_deo() 400 kmp_info_t *th = __kmp_threads[gtid]; in __kmp_dispatch_dxo()
|
D | kmp_sched.cpp | 95 kmp_info_t *th = __kmp_threads[gtid]; in __kmp_for_static_init() 445 kmp_info_t *th; in __kmp_dist_for_static_init() 689 kmp_info_t *th; in __kmp_team_static_init()
|
D | kmp_dispatch_hier.h | 482 kmp_info_t *th = __kmp_threads[gtid]; in next_recurse() 714 kmp_info_t *th = __kmp_threads[gtid]; in next() 922 kmp_info_t *th; in __kmp_dispatch_init_hierarchy()
|
D | kmp_lock.cpp | 1039 static void __kmp_dump_queuing_lock(kmp_info_t *this_thr, kmp_int32 gtid, in __kmp_dump_queuing_lock() 1090 kmp_info_t *this_thr = __kmp_thread_from_gtid(gtid); in __kmp_acquire_queuing_lock_timed_template() 1230 kmp_info_t *tail_thr = __kmp_thread_from_gtid(tail - 1); in __kmp_acquire_queuing_lock_timed_template() 1315 kmp_info_t *this_thr; in __kmp_test_queuing_lock() 1363 kmp_info_t *this_thr; in __kmp_release_queuing_lock() 1439 kmp_info_t *head_thr = __kmp_thread_from_gtid(head - 1); in __kmp_release_queuing_lock() 1464 kmp_info_t *head_thr = __kmp_thread_from_gtid(head - 1); in __kmp_release_queuing_lock()
|