Lines Matching refs:gtid
33 static void __kmp_bottom_half_finish_proxy(kmp_int32 gtid, kmp_task_t *ptask);
44 static void __kmp_trace_task_stack(kmp_int32 gtid, in __kmp_trace_task_stack() argument
56 location, gtid, entries, task_stack->ts_first_block, stack_top)); in __kmp_trace_task_stack()
83 location, gtid, entries, stack_top, tied_task)); in __kmp_trace_task_stack()
89 location, gtid)); in __kmp_trace_task_stack()
98 static void __kmp_init_task_stack(kmp_int32 gtid, in __kmp_init_task_stack() argument
119 static void __kmp_free_task_stack(kmp_int32 gtid, in __kmp_free_task_stack() argument
148 static void __kmp_push_task_stack(kmp_int32 gtid, kmp_info_t *thread, in __kmp_push_task_stack() argument
152 &thread->th.th_task_team->tt.tt_threads_data[__kmp_tid_from_gtid(gtid)]; in __kmp_push_task_stack()
164 gtid, thread, tied_task)); in __kmp_push_task_stack()
193 gtid, tied_task, new_block)); in __kmp_push_task_stack()
196 KA_TRACE(20, ("__kmp_push_task_stack(exit): GTID: %d; TASK: %p\n", gtid, in __kmp_push_task_stack()
207 static void __kmp_pop_task_stack(kmp_int32 gtid, kmp_info_t *thread, in __kmp_pop_task_stack() argument
211 &thread->th.th_task_team->tt_threads_data[__kmp_tid_from_gtid(gtid)]; in __kmp_pop_task_stack()
223 KA_TRACE(20, ("__kmp_pop_task_stack(enter): GTID: %d; THREAD: %p\n", gtid, in __kmp_pop_task_stack()
244 KA_TRACE(20, ("__kmp_pop_task_stack(exit): GTID: %d; TASK: %p\n", gtid, in __kmp_pop_task_stack()
253 static bool __kmp_task_is_allowed(int gtid, const kmp_int32 is_constrained, in __kmp_task_is_allowed() argument
281 if (__kmp_test_lock(node->dn.mtx_locks[i], gtid)) in __kmp_task_is_allowed()
285 __kmp_release_lock(node->dn.mtx_locks[j], gtid); in __kmp_task_is_allowed()
325 static kmp_int32 __kmp_push_task(kmp_int32 gtid, kmp_task_t *task) { in __kmp_push_task() argument
326 kmp_info_t *thread = __kmp_threads[gtid]; in __kmp_push_task()
329 kmp_int32 tid = __kmp_tid_from_gtid(gtid); in __kmp_push_task()
333 ("__kmp_push_task: T#%d trying to push task %p.\n", gtid, taskdata)); in __kmp_push_task()
343 gtid, counter, taskdata)); in __kmp_push_task()
350 gtid, taskdata)); in __kmp_push_task()
376 __kmp_task_is_allowed(gtid, __kmp_task_stealing_constraint, taskdata, in __kmp_push_task()
380 gtid, taskdata)); in __kmp_push_task()
399 __kmp_task_is_allowed(gtid, __kmp_task_stealing_constraint, taskdata, in __kmp_push_task()
404 gtid, taskdata)); in __kmp_push_task()
427 gtid, taskdata, thread_data->td.td_deque_ntasks, in __kmp_push_task()
497 static void __kmp_task_start(kmp_int32 gtid, kmp_task_t *task, in __kmp_task_start() argument
500 kmp_info_t *thread = __kmp_threads[gtid]; in __kmp_task_start()
504 gtid, taskdata, current_task)); in __kmp_task_start()
516 __kmp_push_task_stack(gtid, thread, taskdata); in __kmp_task_start()
537 KA_TRACE(10, ("__kmp_task_start(exit): T#%d task=%p\n", gtid, taskdata)); in __kmp_task_start()
561 kmp_int32 gtid) { in __ompt_task_start() argument
564 if (__kmp_threads[gtid]->th.ompt_thread_info.ompt_task_yielded) { in __ompt_task_start()
566 __kmp_threads[gtid]->th.ompt_thread_info.ompt_task_yielded = 0; in __ompt_task_start()
598 static void __kmpc_omp_task_begin_if0_template(ident_t *loc_ref, kmp_int32 gtid, in __kmpc_omp_task_begin_if0_template() argument
603 kmp_taskdata_t *current_task = __kmp_threads[gtid]->th.th_current_task; in __kmpc_omp_task_begin_if0_template()
607 gtid, loc_ref, taskdata, current_task)); in __kmpc_omp_task_begin_if0_template()
616 gtid, counter, taskdata)); in __kmpc_omp_task_begin_if0_template()
621 __kmp_task_start(gtid, task, current_task); in __kmpc_omp_task_begin_if0_template()
639 __ompt_task_start(task, current_task, gtid); in __kmpc_omp_task_begin_if0_template()
643 KA_TRACE(10, ("__kmpc_omp_task_begin_if0(exit): T#%d loc=%p task=%p,\n", gtid, in __kmpc_omp_task_begin_if0_template()
649 static void __kmpc_omp_task_begin_if0_ompt(ident_t *loc_ref, kmp_int32 gtid, in __kmpc_omp_task_begin_if0_ompt() argument
653 __kmpc_omp_task_begin_if0_template<true>(loc_ref, gtid, task, frame_address, in __kmpc_omp_task_begin_if0_ompt()
664 void __kmpc_omp_task_begin_if0(ident_t *loc_ref, kmp_int32 gtid, in __kmpc_omp_task_begin_if0() argument
668 OMPT_STORE_RETURN_ADDRESS(gtid); in __kmpc_omp_task_begin_if0()
669 __kmpc_omp_task_begin_if0_ompt(loc_ref, gtid, task, in __kmpc_omp_task_begin_if0()
671 OMPT_LOAD_RETURN_ADDRESS(gtid)); in __kmpc_omp_task_begin_if0()
675 __kmpc_omp_task_begin_if0_template<false>(loc_ref, gtid, task, NULL, NULL); in __kmpc_omp_task_begin_if0()
681 void __kmpc_omp_task_begin(ident_t *loc_ref, kmp_int32 gtid, kmp_task_t *task) { in __kmpc_omp_task_begin() argument
682 kmp_taskdata_t *current_task = __kmp_threads[gtid]->th.th_current_task; in __kmpc_omp_task_begin()
687 gtid, loc_ref, KMP_TASK_TO_TASKDATA(task), current_task)); in __kmpc_omp_task_begin()
689 __kmp_task_start(gtid, task, current_task); in __kmpc_omp_task_begin()
691 KA_TRACE(10, ("__kmpc_omp_task_begin(exit): T#%d loc=%p task=%p,\n", gtid, in __kmpc_omp_task_begin()
702 static void __kmp_free_task(kmp_int32 gtid, kmp_taskdata_t *taskdata, in __kmp_free_task() argument
704 KA_TRACE(30, ("__kmp_free_task: T#%d freeing data from task %p\n", gtid, in __kmp_free_task()
725 KA_TRACE(20, ("__kmp_free_task: T#%d freed task %p\n", gtid, taskdata)); in __kmp_free_task()
734 static void __kmp_free_task_and_ancestors(kmp_int32 gtid, in __kmp_free_task_and_ancestors() argument
753 gtid, taskdata)); in __kmp_free_task_and_ancestors()
756 __kmp_free_task(gtid, taskdata, thread); in __kmp_free_task_and_ancestors()
777 gtid, taskdata)); in __kmp_free_task_and_ancestors()
793 gtid, taskdata, children)); in __kmp_free_task_and_ancestors()
806 static void __kmp_task_finish(kmp_int32 gtid, kmp_task_t *task, in __kmp_task_finish() argument
809 kmp_info_t *thread = __kmp_threads[gtid]; in __kmp_task_finish()
816 gtid, taskdata, resumed_task)); in __kmp_task_finish()
823 __kmp_pop_task_stack(gtid, thread, taskdata); in __kmp_task_finish()
834 gtid, counter, taskdata)); in __kmp_task_finish()
847 gtid, taskdata, resumed_task)); in __kmp_task_finish()
876 destr_thunk(gtid, task); in __kmp_task_finish()
888 __kmp_acquire_tas_lock(&taskdata->td_allow_completion_event.lock, gtid); in __kmp_task_finish()
909 __kmp_release_tas_lock(&taskdata->td_allow_completion_event.lock, gtid); in __kmp_task_finish()
932 __kmp_release_deps(gtid, taskdata); in __kmp_task_finish()
936 __kmp_release_deps(gtid, taskdata); in __kmp_task_finish()
949 gtid, taskdata, children)); in __kmp_task_finish()
957 __kmp_free_task_and_ancestors(gtid, taskdata, thread); in __kmp_task_finish()
965 gtid, taskdata, resumed_task)); in __kmp_task_finish()
972 kmp_int32 gtid, in __kmpc_omp_task_complete_if0_template() argument
975 gtid, loc_ref, KMP_TASK_TO_TASKDATA(task))); in __kmpc_omp_task_complete_if0_template()
976 __kmp_assert_valid_gtid(gtid); in __kmpc_omp_task_complete_if0_template()
978 __kmp_task_finish<ompt>(gtid, task, NULL); in __kmpc_omp_task_complete_if0_template()
981 gtid, loc_ref, KMP_TASK_TO_TASKDATA(task))); in __kmpc_omp_task_complete_if0_template()
997 void __kmpc_omp_task_complete_if0_ompt(ident_t *loc_ref, kmp_int32 gtid, in __kmpc_omp_task_complete_if0_ompt() argument
999 __kmpc_omp_task_complete_if0_template<true>(loc_ref, gtid, task); in __kmpc_omp_task_complete_if0_ompt()
1008 void __kmpc_omp_task_complete_if0(ident_t *loc_ref, kmp_int32 gtid, in __kmpc_omp_task_complete_if0() argument
1012 __kmpc_omp_task_complete_if0_ompt(loc_ref, gtid, task); in __kmpc_omp_task_complete_if0()
1016 __kmpc_omp_task_complete_if0_template<false>(loc_ref, gtid, task); in __kmpc_omp_task_complete_if0()
1022 void __kmpc_omp_task_complete(ident_t *loc_ref, kmp_int32 gtid, in __kmpc_omp_task_complete() argument
1024 KA_TRACE(10, ("__kmpc_omp_task_complete(enter): T#%d loc=%p task=%p\n", gtid, in __kmpc_omp_task_complete()
1027 __kmp_task_finish<false>(gtid, task, in __kmpc_omp_task_complete()
1030 KA_TRACE(10, ("__kmpc_omp_task_complete(exit): T#%d loc=%p task=%p\n", gtid, in __kmpc_omp_task_complete()
1167 kmp_task_t *__kmp_task_alloc(ident_t *loc_ref, kmp_int32 gtid, in __kmp_task_alloc() argument
1173 kmp_info_t *thread = __kmp_threads[gtid]; in __kmp_task_alloc()
1183 gtid, loc_ref, *((kmp_int32 *)flags), sizeof_kmp_task_t, in __kmp_task_alloc()
1214 gtid)); in __kmp_task_alloc()
1226 ("T#%d enabling tasking in __kmp_task_alloc for proxy task\n", gtid)); in __kmp_task_alloc()
1246 KA_TRACE(30, ("__kmp_task_alloc: T#%d First malloc size: %ld\n", gtid, in __kmp_task_alloc()
1248 KA_TRACE(30, ("__kmp_task_alloc: T#%d Second malloc size: %ld\n", gtid, in __kmp_task_alloc()
1343 __ompt_task_init(taskdata, gtid); in __kmp_task_alloc()
1362 gtid, taskdata, taskdata->td_parent)); in __kmp_task_alloc()
1368 kmp_task_t *__kmpc_omp_task_alloc(ident_t *loc_ref, kmp_int32 gtid, in __kmpc_omp_task_alloc() argument
1374 __kmp_assert_valid_gtid(gtid); in __kmpc_omp_task_alloc()
1379 gtid, loc_ref, input_flags->tiedness ? "tied " : "untied", in __kmpc_omp_task_alloc()
1384 retval = __kmp_task_alloc(loc_ref, gtid, input_flags, sizeof_kmp_task_t, in __kmpc_omp_task_alloc()
1387 KA_TRACE(20, ("__kmpc_omp_task_alloc(exit): T#%d retval %p\n", gtid, retval)); in __kmpc_omp_task_alloc()
1392 kmp_task_t *__kmpc_omp_target_task_alloc(ident_t *loc_ref, kmp_int32 gtid, in __kmpc_omp_target_task_alloc() argument
1398 return __kmpc_omp_task_alloc(loc_ref, gtid, flags, sizeof_kmp_task_t, in __kmpc_omp_target_task_alloc()
1416 __kmpc_omp_reg_task_with_affinity(ident_t *loc_ref, kmp_int32 gtid, in __kmpc_omp_reg_task_with_affinity() argument
1427 static void __kmp_invoke_task(kmp_int32 gtid, kmp_task_t *task, in __kmp_invoke_task() argument
1434 gtid, taskdata, current_task)); in __kmp_invoke_task()
1443 gtid, taskdata)); in __kmp_invoke_task()
1445 __kmp_bottom_half_finish_proxy(gtid, task); in __kmp_invoke_task()
1449 gtid, taskdata, current_task)); in __kmp_invoke_task()
1460 thread = __kmp_threads[gtid]; in __kmp_invoke_task()
1473 __kmp_task_start(gtid, task, current_task); // OMPT only if not discarded in __kmp_invoke_task()
1480 thread = __kmp_threads[gtid]; in __kmp_invoke_task()
1537 __ompt_task_start(task, current_task, gtid); in __kmp_invoke_task()
1546 thread = __kmp_threads[gtid]; in __kmp_invoke_task()
1562 (*(task->routine))(gtid, task); in __kmp_invoke_task()
1586 __kmp_task_finish<true>(gtid, task, current_task); in __kmp_invoke_task()
1589 __kmp_task_finish<false>(gtid, task, current_task); in __kmp_invoke_task()
1595 gtid, taskdata, current_task)); in __kmp_invoke_task()
1609 kmp_int32 __kmpc_omp_task_parts(ident_t *loc_ref, kmp_int32 gtid, in __kmpc_omp_task_parts() argument
1613 KA_TRACE(10, ("__kmpc_omp_task_parts(enter): T#%d loc=%p task=%p\n", gtid, in __kmpc_omp_task_parts()
1634 if (__kmp_push_task(gtid, new_task) == TASK_NOT_PUSHED) // if cannot defer in __kmpc_omp_task_parts()
1636 kmp_taskdata_t *current_task = __kmp_threads[gtid]->th.th_current_task; in __kmpc_omp_task_parts()
1638 __kmp_invoke_task(gtid, new_task, current_task); in __kmpc_omp_task_parts()
1645 gtid, loc_ref, new_taskdata)); in __kmpc_omp_task_parts()
1667 kmp_int32 __kmp_omp_task(kmp_int32 gtid, kmp_task_t *new_task, in __kmp_omp_task() argument
1674 __kmp_push_task(gtid, new_task) == TASK_NOT_PUSHED) // if cannot defer in __kmp_omp_task()
1676 kmp_taskdata_t *current_task = __kmp_threads[gtid]->th.th_current_task; in __kmp_omp_task()
1679 __kmp_invoke_task(gtid, new_task, current_task); in __kmp_omp_task()
1698 kmp_int32 __kmpc_omp_task(ident_t *loc_ref, kmp_int32 gtid, in __kmpc_omp_task() argument
1706 KA_TRACE(10, ("__kmpc_omp_task(enter): T#%d loc=%p task=%p\n", gtid, loc_ref, in __kmpc_omp_task()
1708 __kmp_assert_valid_gtid(gtid); in __kmpc_omp_task()
1714 OMPT_STORE_RETURN_ADDRESS(gtid); in __kmpc_omp_task()
1726 OMPT_LOAD_RETURN_ADDRESS(gtid)); in __kmpc_omp_task()
1739 res = __kmp_omp_task(gtid, new_task, true); in __kmpc_omp_task()
1743 gtid, loc_ref, new_taskdata)); in __kmpc_omp_task()
1765 kmp_int32 __kmp_omp_taskloop_task(ident_t *loc_ref, kmp_int32 gtid, in __kmp_omp_taskloop_task() argument
1773 KA_TRACE(10, ("__kmpc_omp_task(enter): T#%d loc=%p task=%p\n", gtid, loc_ref, in __kmp_omp_taskloop_task()
1794 res = __kmp_omp_task(gtid, new_task, true); in __kmp_omp_taskloop_task()
1798 gtid, loc_ref, new_taskdata)); in __kmp_omp_taskloop_task()
1808 static kmp_int32 __kmpc_omp_taskwait_template(ident_t *loc_ref, kmp_int32 gtid, in __kmpc_omp_taskwait_template() argument
1816 KA_TRACE(10, ("__kmpc_omp_taskwait(enter): T#%d loc=%p\n", gtid, loc_ref)); in __kmpc_omp_taskwait_template()
1817 __kmp_assert_valid_gtid(gtid); in __kmpc_omp_taskwait_template()
1820 thread = __kmp_threads[gtid]; in __kmpc_omp_taskwait_template()
1854 taskdata->td_taskwait_thread = gtid + 1; in __kmpc_omp_taskwait_template()
1857 void *itt_sync_obj = __kmp_itt_taskwait_object(gtid); in __kmpc_omp_taskwait_template()
1859 __kmp_itt_taskwait_starting(gtid, itt_sync_obj); in __kmpc_omp_taskwait_template()
1873 flag.execute_tasks(thread, gtid, FALSE, in __kmpc_omp_taskwait_template()
1880 __kmp_itt_taskwait_finished(gtid, itt_sync_obj); in __kmpc_omp_taskwait_template()
1909 gtid, taskdata)); in __kmpc_omp_taskwait_template()
1916 static kmp_int32 __kmpc_omp_taskwait_ompt(ident_t *loc_ref, kmp_int32 gtid, in __kmpc_omp_taskwait_ompt() argument
1919 return __kmpc_omp_taskwait_template<true>(loc_ref, gtid, frame_address, in __kmpc_omp_taskwait_ompt()
1926 kmp_int32 __kmpc_omp_taskwait(ident_t *loc_ref, kmp_int32 gtid) { in __kmpc_omp_taskwait() argument
1929 OMPT_STORE_RETURN_ADDRESS(gtid); in __kmpc_omp_taskwait()
1930 return __kmpc_omp_taskwait_ompt(loc_ref, gtid, OMPT_GET_FRAME_ADDRESS(0), in __kmpc_omp_taskwait()
1931 OMPT_LOAD_RETURN_ADDRESS(gtid)); in __kmpc_omp_taskwait()
1934 return __kmpc_omp_taskwait_template<false>(loc_ref, gtid, NULL, NULL); in __kmpc_omp_taskwait()
1938 kmp_int32 __kmpc_omp_taskyield(ident_t *loc_ref, kmp_int32 gtid, int end_part) { in __kmpc_omp_taskyield() argument
1947 gtid, loc_ref, end_part)); in __kmpc_omp_taskyield()
1948 __kmp_assert_valid_gtid(gtid); in __kmpc_omp_taskyield()
1951 thread = __kmp_threads[gtid]; in __kmpc_omp_taskyield()
1961 taskdata->td_taskwait_thread = gtid + 1; in __kmpc_omp_taskyield()
1964 void *itt_sync_obj = __kmp_itt_taskwait_object(gtid); in __kmpc_omp_taskyield()
1966 __kmp_itt_taskwait_starting(gtid, itt_sync_obj); in __kmpc_omp_taskyield()
1977 thread, gtid, (kmp_flag_32<> *)NULL, FALSE, in __kmpc_omp_taskyield()
1989 __kmp_itt_taskwait_finished(gtid, itt_sync_obj); in __kmpc_omp_taskyield()
1999 gtid, taskdata)); in __kmpc_omp_taskyield()
2103 void *__kmp_task_reduction_init(int gtid, int num, T *data) { in __kmp_task_reduction_init() argument
2104 __kmp_assert_valid_gtid(gtid); in __kmp_task_reduction_init()
2105 kmp_info_t *thread = __kmp_threads[gtid]; in __kmp_task_reduction_init()
2116 gtid, tg)); in __kmp_task_reduction_init()
2120 gtid, tg, num)); in __kmp_task_reduction_init()
2171 void *__kmpc_task_reduction_init(int gtid, int num, void *data) { in __kmpc_task_reduction_init() argument
2172 return __kmp_task_reduction_init(gtid, num, (kmp_task_red_input_t *)data); in __kmpc_task_reduction_init()
2187 void *__kmpc_taskred_init(int gtid, int num, void *data) { in __kmpc_taskred_init() argument
2188 return __kmp_task_reduction_init(gtid, num, (kmp_taskred_input_t *)data); in __kmpc_taskred_init()
2219 void *__kmpc_task_reduction_get_th_data(int gtid, void *tskgrp, void *data) { in __kmpc_task_reduction_get_th_data() argument
2220 __kmp_assert_valid_gtid(gtid); in __kmpc_task_reduction_get_th_data()
2221 kmp_info_t *thread = __kmp_threads[gtid]; in __kmpc_task_reduction_get_th_data()
2324 void *__kmp_task_reduction_modifier_init(ident_t *loc, int gtid, int is_ws, in __kmp_task_reduction_modifier_init() argument
2326 __kmp_assert_valid_gtid(gtid); in __kmp_task_reduction_modifier_init()
2327 kmp_info_t *thr = __kmp_threads[gtid]; in __kmp_task_reduction_modifier_init()
2329 __kmpc_taskgroup(loc, gtid); // form new taskgroup first in __kmp_task_reduction_modifier_init()
2333 gtid, thr->th.th_current_task->td_taskgroup)); in __kmp_task_reduction_modifier_init()
2346 tg = (kmp_taskgroup_t *)__kmp_task_reduction_init<T>(gtid, num, data); in __kmp_task_reduction_modifier_init()
2382 void *__kmpc_task_reduction_modifier_init(ident_t *loc, int gtid, int is_ws, in __kmpc_task_reduction_modifier_init() argument
2384 return __kmp_task_reduction_modifier_init(loc, gtid, is_ws, num, in __kmpc_task_reduction_modifier_init()
2402 void *__kmpc_taskred_modifier_init(ident_t *loc, int gtid, int is_ws, int num, in __kmpc_taskred_modifier_init() argument
2404 return __kmp_task_reduction_modifier_init(loc, gtid, is_ws, num, in __kmpc_taskred_modifier_init()
2416 void __kmpc_task_reduction_modifier_fini(ident_t *loc, int gtid, int is_ws) { in __kmpc_task_reduction_modifier_fini() argument
2417 __kmpc_end_taskgroup(loc, gtid); in __kmpc_task_reduction_modifier_fini()
2421 void __kmpc_taskgroup(ident_t *loc, int gtid) { in __kmpc_taskgroup() argument
2422 __kmp_assert_valid_gtid(gtid); in __kmpc_taskgroup()
2423 kmp_info_t *thread = __kmp_threads[gtid]; in __kmpc_taskgroup()
2427 KA_TRACE(10, ("__kmpc_taskgroup: T#%d loc=%p group=%p\n", gtid, loc, tg_new)); in __kmpc_taskgroup()
2437 void *codeptr = OMPT_LOAD_RETURN_ADDRESS(gtid); in __kmpc_taskgroup()
2454 void __kmpc_end_taskgroup(ident_t *loc, int gtid) { in __kmpc_end_taskgroup() argument
2455 __kmp_assert_valid_gtid(gtid); in __kmpc_end_taskgroup()
2456 kmp_info_t *thread = __kmp_threads[gtid]; in __kmpc_end_taskgroup()
2471 codeptr = OMPT_LOAD_RETURN_ADDRESS(gtid); in __kmpc_end_taskgroup()
2477 KA_TRACE(10, ("__kmpc_end_taskgroup(enter): T#%d loc=%p\n", gtid, loc)); in __kmpc_end_taskgroup()
2485 taskdata->td_taskwait_thread = gtid + 1; in __kmpc_end_taskgroup()
2489 void *itt_sync_obj = __kmp_itt_taskwait_object(gtid); in __kmpc_end_taskgroup()
2491 __kmp_itt_taskwait_starting(gtid, itt_sync_obj); in __kmpc_end_taskgroup()
2508 flag.execute_tasks(thread, gtid, FALSE, in __kmpc_end_taskgroup()
2525 __kmp_itt_taskwait_finished(gtid, itt_sync_obj); in __kmpc_end_taskgroup()
2584 gtid, taskdata)); in __kmpc_end_taskgroup()
2597 static kmp_task_t *__kmp_remove_my_task(kmp_info_t *thread, kmp_int32 gtid, in __kmp_remove_my_task() argument
2609 thread_data = &task_team->tt.tt_threads_data[__kmp_tid_from_gtid(gtid)]; in __kmp_remove_my_task()
2612 gtid, thread_data->td.td_deque_ntasks, in __kmp_remove_my_task()
2619 gtid, thread_data->td.td_deque_ntasks, in __kmp_remove_my_task()
2631 gtid, thread_data->td.td_deque_ntasks, in __kmp_remove_my_task()
2640 if (!__kmp_task_is_allowed(gtid, is_constrained, taskdata, in __kmp_remove_my_task()
2647 gtid, thread_data->td.td_deque_ntasks, in __kmp_remove_my_task()
2659 gtid, taskdata, thread_data->td.td_deque_ntasks, in __kmp_remove_my_task()
2669 static kmp_task_t *__kmp_steal_task(kmp_info_t *victim_thr, kmp_int32 gtid, in __kmp_steal_task() argument
2691 gtid, __kmp_gtid_from_thread(victim_thr), task_team, in __kmp_steal_task()
2698 gtid, __kmp_gtid_from_thread(victim_thr), task_team, in __kmp_steal_task()
2712 gtid, __kmp_gtid_from_thread(victim_thr), task_team, ntasks, in __kmp_steal_task()
2718 current = __kmp_threads[gtid]->th.th_current_task; in __kmp_steal_task()
2720 if (__kmp_task_is_allowed(gtid, is_constrained, taskdata, current)) { in __kmp_steal_task()
2730 gtid, __kmp_gtid_from_thread(victim_thr), task_team, ntasks, in __kmp_steal_task()
2741 if (__kmp_task_is_allowed(gtid, is_constrained, taskdata, current)) { in __kmp_steal_task()
2752 gtid, __kmp_gtid_from_thread(victim_thr), task_team, ntasks, in __kmp_steal_task()
2779 gtid, count + 1, task_team)); in __kmp_steal_task()
2791 gtid, taskdata, __kmp_gtid_from_thread(victim_thr), task_team, in __kmp_steal_task()
2809 kmp_info_t *thread, kmp_int32 gtid, C *flag, int final_spin, in __kmp_execute_tasks_template() argument
2822 KMP_DEBUG_ASSERT(thread == __kmp_threads[gtid]); in __kmp_execute_tasks_template()
2829 gtid, final_spin, *thread_finished)); in __kmp_execute_tasks_template()
2845 task = __kmp_remove_my_task(thread, gtid, task_team, is_constrained); in __kmp_execute_tasks_template()
2899 task = __kmp_steal_task(other_thread, gtid, task_team, in __kmp_execute_tasks_template()
2925 itt_sync_obj = __kmp_itt_barrier_object(gtid, bs_forkjoin_barrier); in __kmp_execute_tasks_template()
2930 __kmp_invoke_task(gtid, task, current_task); in __kmp_execute_tasks_template()
2944 gtid)); in __kmp_execute_tasks_template()
2956 gtid)); in __kmp_execute_tasks_template()
2976 gtid, count, task_team)); in __kmp_execute_tasks_template()
2989 gtid)); in __kmp_execute_tasks_template()
2998 ("__kmp_execute_tasks_template: T#%d no more tasks\n", gtid)); in __kmp_execute_tasks_template()
3008 ("__kmp_execute_tasks_template: T#%d can't find work\n", gtid)); in __kmp_execute_tasks_template()
3016 kmp_info_t *thread, kmp_int32 gtid, kmp_flag_32<C, S> *flag, int final_spin, in __kmp_execute_tasks_32() argument
3020 thread, gtid, flag, final_spin, in __kmp_execute_tasks_32()
3026 kmp_info_t *thread, kmp_int32 gtid, kmp_flag_64<C, S> *flag, int final_spin, in __kmp_execute_tasks_64() argument
3030 thread, gtid, flag, final_spin, in __kmp_execute_tasks_64()
3035 kmp_info_t *thread, kmp_int32 gtid, kmp_flag_oncore *flag, int final_spin, in __kmp_execute_tasks_oncore() argument
3039 thread, gtid, flag, final_spin, in __kmp_execute_tasks_oncore()
3208 __kmp_free_task_stack(__kmp_thread_from_gtid(gtid), thread_data); in __kmp_free_task_deque()
3638 void __kmp_tasking_barrier(kmp_team_t *team, kmp_info_t *thread, int gtid) { in __kmp_tasking_barrier() argument
3649 while (!spin_flag.execute_tasks(thread, gtid, TRUE, in __kmp_tasking_barrier()
3797 static void __kmp_bottom_half_finish_proxy(kmp_int32 gtid, kmp_task_t *ptask) { in __kmp_bottom_half_finish_proxy() argument
3799 kmp_info_t *thread = __kmp_threads[gtid]; in __kmp_bottom_half_finish_proxy()
3810 __kmp_release_deps(gtid, taskdata); in __kmp_bottom_half_finish_proxy()
3811 __kmp_free_task_and_ancestors(gtid, taskdata, thread); in __kmp_bottom_half_finish_proxy()
3822 void __kmpc_proxy_task_completed(kmp_int32 gtid, kmp_task_t *ptask) { in __kmpc_proxy_task_completed() argument
3827 gtid, taskdata)); in __kmpc_proxy_task_completed()
3828 __kmp_assert_valid_gtid(gtid); in __kmpc_proxy_task_completed()
3833 __kmp_bottom_half_finish_proxy(gtid, ptask); in __kmpc_proxy_task_completed()
3837 gtid, taskdata)); in __kmpc_proxy_task_completed()
3891 kmp_event_t *__kmpc_task_allow_completion_event(ident_t *loc_ref, int gtid, in __kmpc_task_allow_completion_event() argument
3907 int gtid = __kmp_get_gtid(); in __kmp_fulfill_event() local
3912 __kmp_acquire_tas_lock(&event->lock, gtid); in __kmp_fulfill_event()
3924 __kmp_release_tas_lock(&event->lock, gtid); in __kmp_fulfill_event()
3934 if (gtid >= 0) { in __kmp_fulfill_event()
3938 __kmpc_proxy_task_completed(gtid, ptask); in __kmp_fulfill_event()
4148 void __kmp_taskloop_linear(ident_t *loc, int gtid, kmp_task_t *task, in __kmp_taskloop_linear() argument
4165 kmp_info_t *thread = __kmp_threads[gtid]; in __kmp_taskloop_linear()
4175 gtid, num_tasks, grainsize, extras, lower, upper, ub_glob, st, in __kmp_taskloop_linear()
4222 gtid, i, next_task, lower, upper, st, in __kmp_taskloop_linear()
4226 __kmp_omp_taskloop_task(NULL, gtid, next_task, in __kmp_taskloop_linear()
4229 __kmp_omp_task(gtid, next_task, true); // schedule new task in __kmp_taskloop_linear()
4234 __kmp_task_start(gtid, task, current_task); // make internal bookkeeping in __kmp_taskloop_linear()
4236 __kmp_task_finish<false>(gtid, task, current_task); in __kmp_taskloop_linear()
4267 int __kmp_taskloop_task(int gtid, void *ptask) { in __kmp_taskloop_task() argument
4290 gtid, taskdata, num_tasks, grainsize, extras, *lb, *ub, st, in __kmp_taskloop_task()
4295 __kmp_taskloop_recur(NULL, gtid, task, lb, ub, st, ub_glob, num_tasks, in __kmp_taskloop_task()
4302 __kmp_taskloop_linear(NULL, gtid, task, lb, ub, st, ub_glob, num_tasks, in __kmp_taskloop_task()
4309 KA_TRACE(40, ("__kmp_taskloop_task(exit): T#%d\n", gtid)); in __kmp_taskloop_task()
4330 void __kmp_taskloop_recur(ident_t *loc, int gtid, kmp_task_t *task, in __kmp_taskloop_recur() argument
4344 gtid, taskdata, num_tasks, grainsize, extras, *lb, *ub, st, in __kmp_taskloop_recur()
4348 kmp_info_t *thread = __kmp_threads[gtid]; in __kmp_taskloop_recur()
4393 __kmpc_omp_task_alloc(loc, gtid, 1, 3 * sizeof(void *), in __kmp_taskloop_recur()
4415 __kmp_omp_taskloop_task(NULL, gtid, new_task, codeptr_ra); in __kmp_taskloop_recur()
4417 __kmp_omp_task(gtid, new_task, true); // schedule new task in __kmp_taskloop_recur()
4422 __kmp_taskloop_recur(loc, gtid, task, lb, ub, st, ub_glob, n_tsk0, gr_size0, in __kmp_taskloop_recur()
4429 __kmp_taskloop_linear(loc, gtid, task, lb, ub, st, ub_glob, n_tsk0, in __kmp_taskloop_recur()
4436 KA_TRACE(40, ("__kmpc_taskloop_recur(exit): T#%d\n", gtid)); in __kmp_taskloop_recur()
4455 void __kmpc_taskloop(ident_t *loc, int gtid, kmp_task_t *task, int if_val, in __kmpc_taskloop() argument
4460 __kmp_assert_valid_gtid(gtid); in __kmpc_taskloop()
4463 OMPT_STORE_RETURN_ADDRESS(gtid); in __kmpc_taskloop()
4465 __kmpc_taskgroup(loc, gtid); in __kmpc_taskloop()
4478 kmp_info_t *thread = __kmp_threads[gtid]; in __kmpc_taskloop()
4483 gtid, taskdata, lower, upper, st, grainsize, sched, task_dup)); in __kmpc_taskloop()
4494 KA_TRACE(20, ("__kmpc_taskloop(exit): T#%d zero-trip loop\n", gtid)); in __kmpc_taskloop()
4496 __kmp_task_start(gtid, task, current_task); in __kmpc_taskloop()
4498 __kmp_task_finish<false>(gtid, task, current_task); in __kmpc_taskloop()
4560 __kmp_taskloop_linear(loc, gtid, task, lb, ub, st, ub_glob, num_tasks, in __kmpc_taskloop()
4571 gtid, tc, num_tasks, num_tasks_min, grainsize, extras)); in __kmpc_taskloop()
4572 __kmp_taskloop_recur(loc, gtid, task, lb, ub, st, ub_glob, num_tasks, in __kmpc_taskloop()
4581 gtid, tc, num_tasks, num_tasks_min, grainsize, extras)); in __kmpc_taskloop()
4582 __kmp_taskloop_linear(loc, gtid, task, lb, ub, st, ub_glob, num_tasks, in __kmpc_taskloop()
4600 OMPT_STORE_RETURN_ADDRESS(gtid); in __kmpc_taskloop()
4602 __kmpc_end_taskgroup(loc, gtid); in __kmpc_taskloop()
4604 KA_TRACE(20, ("__kmpc_taskloop(exit): T#%d\n", gtid)); in __kmpc_taskloop()