/external/llvm-project/openmp/runtime/src/ |
D | kmp_atomic.h | 362 kmp_int32 gtid) { in __kmp_acquire_atomic_lock() argument 371 __kmp_acquire_queuing_lock(lck, gtid); in __kmp_acquire_atomic_lock() 382 kmp_int32 gtid) { in __kmp_test_atomic_lock() argument 383 return __kmp_test_queuing_lock(lck, gtid); in __kmp_test_atomic_lock() 387 kmp_int32 gtid) { in __kmp_release_atomic_lock() argument 388 __kmp_release_queuing_lock(lck, gtid); in __kmp_release_atomic_lock() 448 void __kmpc_atomic_fixed1_add(ident_t *id_ref, int gtid, char *lhs, char rhs); 449 void __kmpc_atomic_fixed1_andb(ident_t *id_ref, int gtid, char *lhs, char rhs); 450 void __kmpc_atomic_fixed1_div(ident_t *id_ref, int gtid, char *lhs, char rhs); 451 void __kmpc_atomic_fixed1u_div(ident_t *id_ref, int gtid, unsigned char *lhs, [all …]
|
D | kmp_gsupport.cpp | 97 int gtid = __kmp_entry_gtid(); in KMP_EXPAND_NAME() local 99 KA_TRACE(20, ("GOMP_barrier: T#%d\n", gtid)); in KMP_EXPAND_NAME() 106 OMPT_STORE_RETURN_ADDRESS(gtid); in KMP_EXPAND_NAME() 108 __kmpc_barrier(&loc, gtid); in KMP_EXPAND_NAME() 128 int gtid = __kmp_entry_gtid(); in KMP_EXPAND_NAME() local 130 KA_TRACE(20, ("GOMP_critical_start: T#%d\n", gtid)); in KMP_EXPAND_NAME() 132 OMPT_STORE_RETURN_ADDRESS(gtid); in KMP_EXPAND_NAME() 134 __kmpc_critical(&loc, gtid, __kmp_unnamed_critical_addr); in KMP_EXPAND_NAME() 138 int gtid = __kmp_get_gtid(); in KMP_EXPAND_NAME() local 140 KA_TRACE(20, ("GOMP_critical_end: T#%d\n", gtid)); in KMP_EXPAND_NAME() [all …]
|
D | kmp_error.cpp | 50 static void __kmp_expand_cons_stack(int gtid, struct cons_header *p) { in __kmp_expand_cons_stack() argument 55 if (gtid < 0) in __kmp_expand_cons_stack() 58 KE_TRACE(10, ("expand cons_stack (%d %d)\n", gtid, __kmp_get_gtid())); in __kmp_expand_cons_stack() 127 struct cons_header *__kmp_allocate_cons_stack(int gtid) { in __kmp_allocate_cons_stack() argument 131 if (gtid < 0) { in __kmp_allocate_cons_stack() 134 KE_TRACE(10, ("allocate cons_stack (%d)\n", gtid)); in __kmp_allocate_cons_stack() 159 static void dump_cons_stack(int gtid, struct cons_header *p) { in dump_cons_stack() argument 169 tos, gtid); in dump_cons_stack() 178 __kmp_str_buf_print(&buffer, "End construct stack for thread %d\n", gtid); in dump_cons_stack() 187 void __kmp_push_parallel(int gtid, ident_t const *ident) { in __kmp_push_parallel() argument [all …]
|
D | kmp_lock.cpp | 81 __kmp_acquire_tas_lock_timed_template(kmp_tas_lock_t *lck, kmp_int32 gtid) { in __kmp_acquire_tas_lock_timed_template() argument 86 if ((curr != 0) && (curr != gtid + 1)) in __kmp_acquire_tas_lock_timed_template() 92 kmp_int32 tas_busy = KMP_LOCK_BUSY(gtid + 1, tas); in __kmp_acquire_tas_lock_timed_template() 113 int __kmp_acquire_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid) { in __kmp_acquire_tas_lock() argument 114 int retval = __kmp_acquire_tas_lock_timed_template(lck, gtid); in __kmp_acquire_tas_lock() 120 kmp_int32 gtid) { in __kmp_acquire_tas_lock_with_checks() argument 126 if ((gtid >= 0) && (__kmp_get_tas_lock_owner(lck) == gtid)) { in __kmp_acquire_tas_lock_with_checks() 129 return __kmp_acquire_tas_lock(lck, gtid); in __kmp_acquire_tas_lock_with_checks() 132 int __kmp_test_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid) { in __kmp_test_tas_lock() argument 134 kmp_int32 tas_busy = KMP_LOCK_BUSY(gtid + 1, tas); in __kmp_test_tas_lock() [all …]
|
D | kmp_threadprivate.cpp | 21 void kmp_threadprivate_insert_private_data(int gtid, void *pc_addr, 23 struct private_common *kmp_threadprivate_insert(int gtid, void *pc_addr, 34 __kmp_threadprivate_find_task_common(struct common_table *tbl, int gtid, in __kmp_threadprivate_find_task_common() argument 43 gtid, pc_addr)); in __kmp_threadprivate_find_task_common() 52 gtid, pc_addr)); in __kmp_threadprivate_find_task_common() 65 __kmp_find_shared_task_common(struct shared_table *tbl, int gtid, in __kmp_find_shared_task_common() argument 75 gtid, pc_addr)); in __kmp_find_shared_task_common() 134 int gtid; in __kmp_common_initialize() local 141 for (gtid = 0; gtid < __kmp_threads_capacity; gtid++) in __kmp_common_initialize() 142 if (__kmp_root[gtid]) { in __kmp_common_initialize() [all …]
|
D | ompt-specific.h | 27 int gtid, ompt_data_t *ompt_pid, void *codeptr); 71 inline void *__ompt_load_return_address(int gtid) { in __ompt_load_return_address() argument 72 kmp_info_t *thr = __kmp_threads[gtid]; in __ompt_load_return_address() 83 #define OMPT_STORE_RETURN_ADDRESS(gtid) \ argument 84 OmptReturnAddressGuard ReturnAddressGuard{gtid, __builtin_return_address(0)}; 85 #define OMPT_LOAD_RETURN_ADDRESS(gtid) __ompt_load_return_address(gtid) argument 86 #define OMPT_LOAD_OR_GET_RETURN_ADDRESS(gtid) \ argument 87 ((ompt_enabled.enabled && gtid >= 0 && __kmp_threads[gtid] && \ 88 __kmp_threads[gtid]->th.ompt_thread_info.return_address)? \ 89 __ompt_load_return_address(gtid): \ [all …]
|
D | kmp_barrier.cpp | 47 enum barrier_type bt, kmp_info_t *this_thr, int gtid, int tid, in __kmp_linear_barrier_gather_template() argument 57 gtid, team->t.t_id, tid, bt)); in __kmp_linear_barrier_gather_template() 73 gtid, team->t.t_id, tid, __kmp_gtid_from_tid(0, team), in __kmp_linear_barrier_gather_template() 99 gtid, team->t.t_id, tid, __kmp_gtid_from_tid(i, team), in __kmp_linear_barrier_gather_template() 126 gtid, team->t.t_id, tid, __kmp_gtid_from_tid(i, team), in __kmp_linear_barrier_gather_template() 129 OMPT_REDUCTION_DECL(this_thr, gtid); in __kmp_linear_barrier_gather_template() 142 gtid, team->t.t_id, tid, team->t.t_id, &team_bar->b_arrived, in __kmp_linear_barrier_gather_template() 148 gtid, team->t.t_id, tid, bt)); in __kmp_linear_barrier_gather_template() 154 enum barrier_type bt, kmp_info_t *this_thr, int gtid, int tid, in __kmp_linear_barrier_release_template() argument 165 team = __kmp_threads[gtid]->th.th_team; in __kmp_linear_barrier_release_template() [all …]
|
D | kmp_tasking.cpp | 33 static void __kmp_bottom_half_finish_proxy(kmp_int32 gtid, kmp_task_t *ptask); 44 static void __kmp_trace_task_stack(kmp_int32 gtid, in __kmp_trace_task_stack() argument 56 location, gtid, entries, task_stack->ts_first_block, stack_top)); in __kmp_trace_task_stack() 83 location, gtid, entries, stack_top, tied_task)); in __kmp_trace_task_stack() 89 location, gtid)); in __kmp_trace_task_stack() 98 static void __kmp_init_task_stack(kmp_int32 gtid, in __kmp_init_task_stack() argument 119 static void __kmp_free_task_stack(kmp_int32 gtid, in __kmp_free_task_stack() argument 148 static void __kmp_push_task_stack(kmp_int32 gtid, kmp_info_t *thread, in __kmp_push_task_stack() argument 152 &thread->th.th_task_team->tt.tt_threads_data[__kmp_tid_from_gtid(gtid)]; in __kmp_push_task_stack() 164 gtid, thread, tied_task)); in __kmp_push_task_stack() [all …]
|
D | kmp_ftn_entry.h | 105 int gtid, tid; in FTN_SET_BLOCKTIME() 108 gtid = __kmp_entry_gtid(); in FTN_SET_BLOCKTIME() 109 tid = __kmp_tid_from_gtid(gtid); in FTN_SET_BLOCKTIME() 110 thread = __kmp_thread_from_gtid(gtid); in FTN_SET_BLOCKTIME() 120 int gtid, tid; in FTN_GET_BLOCKTIME() 124 gtid = __kmp_entry_gtid(); in FTN_GET_BLOCKTIME() 125 tid = __kmp_tid_from_gtid(gtid); in FTN_GET_BLOCKTIME() 126 thread = __kmp_thread_from_gtid(gtid); in FTN_GET_BLOCKTIME() 127 team = __kmp_threads[gtid]->th.th_team; in FTN_GET_BLOCKTIME() 131 KF_TRACE(10, ("kmp_get_blocktime: T#%d(%d:%d), blocktime=%d\n", gtid, in FTN_GET_BLOCKTIME() [all …]
|
D | kmp_lock.h | 144 extern int __kmp_acquire_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid); 145 extern int __kmp_test_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid); 146 extern int __kmp_release_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid); 150 extern int __kmp_acquire_nested_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid); 151 extern int __kmp_test_nested_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid); 152 extern int __kmp_release_nested_tas_lock(kmp_tas_lock_t *lck, kmp_int32 gtid); 203 extern int __kmp_acquire_futex_lock(kmp_futex_lock_t *lck, kmp_int32 gtid); 204 extern int __kmp_test_futex_lock(kmp_futex_lock_t *lck, kmp_int32 gtid); 205 extern int __kmp_release_futex_lock(kmp_futex_lock_t *lck, kmp_int32 gtid); 210 kmp_int32 gtid); [all …]
|
D | kmp_error.h | 28 struct cons_header *__kmp_allocate_cons_stack(int gtid); 31 void __kmp_push_parallel(int gtid, ident_t const *ident); 32 void __kmp_push_workshare(int gtid, enum cons_type ct, ident_t const *ident); 34 void __kmp_push_sync(int gtid, enum cons_type ct, ident_t const *ident, 37 void __kmp_push_sync(int gtid, enum cons_type ct, ident_t const *ident, 41 void __kmp_check_workshare(int gtid, enum cons_type ct, ident_t const *ident); 43 void __kmp_check_sync(int gtid, enum cons_type ct, ident_t const *ident, 46 void __kmp_check_sync(int gtid, enum cons_type ct, ident_t const *ident, 50 void __kmp_pop_parallel(int gtid, ident_t const *ident); 51 enum cons_type __kmp_pop_workshare(int gtid, enum cons_type ct, [all …]
|
D | kmp_dispatch.cpp | 104 void __kmp_dispatch_init_algorithm(ident_t *loc, int gtid, in __kmp_dispatch_init_algorithm() argument 134 KD_TRACE(10, (buff, gtid, pr, lb, ub, st, schedule, chunk, nproc, tid)); in __kmp_dispatch_init_algorithm() 139 th = __kmp_threads[gtid]; in __kmp_dispatch_init_algorithm() 146 KMP_MASTER_GTID(gtid) && th->th.th_teams_microtask == NULL && in __kmp_dispatch_init_algorithm() 211 KD_TRACE(10, (buff, gtid, schedule, chunk)); in __kmp_dispatch_init_algorithm() 235 KD_TRACE(10, (buff, gtid, schedule, chunk)); in __kmp_dispatch_init_algorithm() 280 KD_TRACE(10, (buff, gtid, schedule, chunk)); in __kmp_dispatch_init_algorithm() 324 if (KMP_MASTER_GTID(gtid)) { in __kmp_dispatch_init_algorithm() 355 gtid)); in __kmp_dispatch_init_algorithm() 394 gtid)); in __kmp_dispatch_init_algorithm() [all …]
|
D | kmp_itt.h | 55 __kmp_itt_region_forking(int gtid, int team_size, 58 __kmp_itt_region_joined(int gtid); // Master only, after joining threads. 63 __kmp_inline void __kmp_itt_frame_submit(int gtid, __itt_timestamp begin, 71 __kmp_inline void __kmp_itt_metadata_imbalance(int gtid, kmp_uint64 begin, 83 __kmp_inline void *__kmp_itt_barrier_object(int gtid, int bt, int set_name = 0, 85 __kmp_inline void __kmp_itt_barrier_starting(int gtid, void *object); 86 __kmp_inline void __kmp_itt_barrier_middle(int gtid, void *object); 87 __kmp_inline void __kmp_itt_barrier_finished(int gtid, void *object); 90 __kmp_inline void *__kmp_itt_taskwait_object(int gtid); 91 __kmp_inline void __kmp_itt_taskwait_starting(int gtid, void *object); [all …]
|
D | kmp_csupport.cpp | 99 kmp_int32 gtid = __kmp_entry_gtid(); in __kmpc_global_thread_num() local 101 KC_TRACE(10, ("__kmpc_global_thread_num: T#%d\n", gtid)); in __kmpc_global_thread_num() 103 return gtid; in __kmpc_global_thread_num() 262 int gtid = __kmp_entry_gtid(); in __kmpc_fork_call() local 289 kmp_info_t *master_th = __kmp_threads[gtid]; in __kmpc_fork_call() 295 int tid = __kmp_tid_from_gtid(gtid); in __kmpc_fork_call() 301 OMPT_STORE_RETURN_ADDRESS(gtid); in __kmpc_fork_call() 307 __kmp_fork_call(loc, gtid, fork_context_intel, argc, in __kmpc_fork_call() 314 __kmp_join_call(loc, gtid in __kmpc_fork_call() 366 int gtid = __kmp_entry_gtid(); in __kmpc_fork_teams() local [all …]
|
D | kmp_taskdeps.h | 20 #define KMP_ACQUIRE_DEPNODE(gtid, n) __kmp_acquire_lock(&(n)->dn.lock, (gtid)) argument 21 #define KMP_RELEASE_DEPNODE(gtid, n) __kmp_release_lock(&(n)->dn.lock, (gtid)) argument 88 static inline void __kmp_release_deps(kmp_int32 gtid, kmp_taskdata_t *task) { in __kmp_release_deps() argument 89 kmp_info_t *thread = __kmp_threads[gtid]; in __kmp_release_deps() 98 __kmp_release_lock(node->dn.mtx_locks[i], gtid); in __kmp_release_deps() 105 gtid, task)); in __kmp_release_deps() 114 gtid, task)); in __kmp_release_deps() 116 KMP_ACQUIRE_DEPNODE(gtid, node); in __kmp_release_deps() 119 KMP_RELEASE_DEPNODE(gtid, node); in __kmp_release_deps() 133 gtid, successor->dn.task, task)); in __kmp_release_deps() [all …]
|
D | kmp.h | 962 extern omp_allocator_handle_t __kmpc_init_allocator(int gtid, 966 extern void __kmpc_destroy_allocator(int gtid, omp_allocator_handle_t al); 967 extern void __kmpc_set_default_allocator(int gtid, omp_allocator_handle_t al); 968 extern omp_allocator_handle_t __kmpc_get_default_allocator(int gtid); 969 extern void *__kmpc_alloc(int gtid, size_t sz, omp_allocator_handle_t al); 970 extern void *__kmpc_calloc(int gtid, size_t nmemb, size_t sz, 972 extern void *__kmpc_realloc(int gtid, void *ptr, size_t sz, 975 extern void __kmpc_free(int gtid, void *ptr, omp_allocator_handle_t al); 1119 #define KMP_MASTER_GTID(gtid) (0 == __kmp_tid_from_gtid((gtid))) argument 1120 #define KMP_WORKER_GTID(gtid) (0 != __kmp_tid_from_gtid((gtid))) argument [all …]
|
D | kmp_runtime.cpp | 79 int gtid); 88 void __kmp_fork_barrier(int gtid, int tid); 89 void __kmp_join_barrier(int gtid); 99 static int __kmp_unregister_root_other_thread(int gtid); 218 int gtid; in __kmp_get_global_thread_id_reg() local 221 gtid = KMP_GTID_DNE; in __kmp_get_global_thread_id_reg() 226 gtid = __kmp_gtid; in __kmp_get_global_thread_id_reg() 231 gtid = __kmp_gtid_get_specific(); in __kmp_get_global_thread_id_reg() 235 gtid = __kmp_get_global_thread_id(); in __kmp_get_global_thread_id_reg() 239 if (gtid == KMP_GTID_DNE) { in __kmp_get_global_thread_id_reg() [all …]
|
D | kmp_taskdeps.cpp | 208 static inline void __kmp_track_dependence(kmp_int32 gtid, kmp_depnode_t *source, in __kmp_track_dependence() argument 231 sink_data = &__kmp_threads[gtid]->th.ompt_thread_info.task_data; in __kmp_track_dependence() 240 __kmp_depnode_link_successor(kmp_int32 gtid, kmp_info_t *thread, in __kmp_depnode_link_successor() argument 250 KMP_ACQUIRE_DEPNODE(gtid, dep); in __kmp_depnode_link_successor() 252 __kmp_track_dependence(gtid, dep, node, task); in __kmp_depnode_link_successor() 256 gtid, KMP_TASK_TO_TASKDATA(dep->dn.task), in __kmp_depnode_link_successor() 260 KMP_RELEASE_DEPNODE(gtid, dep); in __kmp_depnode_link_successor() 266 static inline kmp_int32 __kmp_depnode_link_successor(kmp_int32 gtid, in __kmp_depnode_link_successor() argument 276 KMP_ACQUIRE_DEPNODE(gtid, sink); in __kmp_depnode_link_successor() 278 __kmp_track_dependence(gtid, sink, source, task); in __kmp_depnode_link_successor() [all …]
|
D | kmp_cancel.cpp | 29 kmp_int32 __kmpc_cancel(ident_t *loc_ref, kmp_int32 gtid, kmp_int32 cncl_kind) { in __kmpc_cancel() argument 30 kmp_info_t *this_thr = __kmp_threads[gtid]; in __kmpc_cancel() 32 KC_TRACE(10, ("__kmpc_cancel: T#%d request %d OMP_CANCELLATION=%d\n", gtid, in __kmpc_cancel() 39 KMP_DEBUG_ASSERT(__kmp_get_gtid() == gtid); in __kmpc_cancel() 135 kmp_int32 __kmpc_cancellationpoint(ident_t *loc_ref, kmp_int32 gtid, in __kmpc_cancellationpoint() argument 137 kmp_info_t *this_thr = __kmp_threads[gtid]; in __kmpc_cancellationpoint() 141 gtid, cncl_kind, __kmp_omp_cancellation)); in __kmpc_cancellationpoint() 147 KMP_DEBUG_ASSERT(__kmp_get_gtid() == gtid); in __kmpc_cancellationpoint() 243 kmp_int32 __kmpc_cancel_barrier(ident_t *loc, kmp_int32 gtid) { in __kmpc_cancel_barrier() argument 245 kmp_info_t *this_thr = __kmp_threads[gtid]; in __kmpc_cancel_barrier() [all …]
|
/external/llvm-project/openmp/runtime/test/tasking/ |
D | kmp_task_reduction_nest.cpp | 76 extern void* __kmpc_task_reduction_get_th_data(int gtid, void* tg, void* item); 77 extern void* __kmpc_task_reduction_init(int gtid, int num, void* data); 200 int gtid = __kmpc_global_thread_num(NULL); in main() local 201 void* tg1 = __kmpc_task_reduction_init(gtid, 3, red_data); in main() 208 int gtid = __kmpc_global_thread_num(NULL); in main() local 209 int *p_ip = (int*)__kmpc_task_reduction_get_th_data(gtid, tg1, &ip); in main() 211 gtid, tg1, &xp); in main() 212 if (!ptrs[gtid]) ptrs[gtid] = p_xp; in main() 245 int gtid = __kmpc_global_thread_num(NULL); in main() local 246 void* tg2 = __kmpc_task_reduction_init(gtid, 3, red_data); in main() [all …]
|
D | kmp_task_modifier_simple_par_old.cpp | 23 extern void *__kmpc_task_reduction_get_th_data(int gtid, void *tg, void *item); 24 extern void *__kmpc_task_reduction_modifier_init(void *loc, int gtid, int is_ws, 26 extern void __kmpc_task_reduction_modifier_fini(void *loc, int gtid, int is_ws); 54 int gtid = __kmpc_global_thread_num(NULL); in main() local 64 gtid, in main() 72 int gtid = __kmpc_global_thread_num(NULL); in main() local 73 int *p_var = (int *)__kmpc_task_reduction_get_th_data(gtid, tg, &var); in main() 79 int gtid = __kmpc_global_thread_num(NULL); in main() local 80 int *p_var = (int *)__kmpc_task_reduction_get_th_data(gtid, tg, &var); in main() 84 __kmpc_task_reduction_modifier_fini(NULL, gtid, 0); in main()
|
D | kmp_task_modifier_simple_par_new.cpp | 23 extern void *__kmpc_task_reduction_get_th_data(int gtid, void *tg, void *item); 26 extern void *__kmpc_taskred_modifier_init(void *loc, int gtid, int is_ws, 28 extern void __kmpc_task_reduction_modifier_fini(void *loc, int gtid, int is_ws); 58 int gtid = __kmpc_global_thread_num(NULL); in main() local 70 gtid, in main() 78 int gtid = __kmpc_global_thread_num(NULL); in main() local 79 int *p_var = (int *)__kmpc_task_reduction_get_th_data(gtid, tg, &var); in main() 85 int gtid = __kmpc_global_thread_num(NULL); in main() local 86 int *p_var = (int *)__kmpc_task_reduction_get_th_data(gtid, tg, &var); in main() 90 __kmpc_task_reduction_modifier_fini(NULL, gtid, 0); in main()
|
D | kmp_task_modifier_simple_ws_old.cpp | 23 extern void *__kmpc_task_reduction_get_th_data(int gtid, void *tg, void *item); 24 extern void *__kmpc_task_reduction_modifier_init(void *loc, int gtid, int is_ws, 26 extern void __kmpc_task_reduction_modifier_fini(void *loc, int gtid, int is_ws); 60 int gtid = __kmpc_global_thread_num(NULL); in main() local 70 gtid, in main() 82 int gtid = __kmpc_global_thread_num(NULL); in main() local 83 int *p_var = (int *)__kmpc_task_reduction_get_th_data(gtid, tg, &var); in main() 89 int gtid = __kmpc_global_thread_num(NULL); in main() local 90 int *p_var = (int *)__kmpc_task_reduction_get_th_data(gtid, tg, &var); in main() 98 __kmpc_task_reduction_modifier_fini(NULL, gtid, 1); in main()
|
D | kmp_task_modifier_simple_ws_new.cpp | 23 extern void *__kmpc_task_reduction_get_th_data(int gtid, void *tg, void *item); 26 extern void *__kmpc_taskred_modifier_init(void *loc, int gtid, int is_ws, 28 extern void __kmpc_task_reduction_modifier_fini(void *loc, int gtid, int is_ws); 64 int gtid = __kmpc_global_thread_num(NULL); in main() local 76 gtid, in main() 88 int gtid = __kmpc_global_thread_num(NULL); in main() local 89 int *p_var = (int *)__kmpc_task_reduction_get_th_data(gtid, tg, &var); in main() 95 int gtid = __kmpc_global_thread_num(NULL); in main() local 96 int *p_var = (int *)__kmpc_task_reduction_get_th_data(gtid, tg, &var); in main() 104 __kmpc_task_reduction_modifier_fini(NULL, gtid, 1); in main()
|
D | kmp_detach_tasks_t3.c | 68 extern int** __kmpc_omp_task_alloc(id *loc, int gtid, int flags, 70 extern int __kmpc_omp_task_with_deps(id *loc, int gtid, ptask task, int nd, 72 extern int __kmpc_omp_task(id *loc, int gtid, kmp_task_t *task); 74 ident_t *loc_ref, int gtid, kmp_task_t *task); 82 int task_entry(int gtid, ptask task) { in task_entry() argument 88 int i, j, gtid = __kmpc_global_thread_num(NULL); in main() local 102 int gtid = __kmpc_global_thread_num(NULL); in main() local 108 task = (ptask)__kmpc_omp_task_alloc(NULL,gtid,PTASK_FLAG_DETACHABLE, in main() 111 evt = (omp_event_handle_t)__kmpc_task_allow_completion_event(NULL,gtid,task); in main() 119 __kmpc_omp_task_with_deps(NULL,gtid,task,1,&sdep,0,0); in main()
|