• Home
  • Raw
  • Download

Lines Matching refs:master_th

967                                     kmp_info_t *master_th, int master_gtid) {  in __kmp_fork_team_threads()  argument
976 master_th->th.th_info.ds.ds_tid = 0; in __kmp_fork_team_threads()
977 master_th->th.th_team = team; in __kmp_fork_team_threads()
978 master_th->th.th_team_nproc = team->t.t_nproc; in __kmp_fork_team_threads()
979 master_th->th.th_team_master = master_th; in __kmp_fork_team_threads()
980 master_th->th.th_team_serialized = FALSE; in __kmp_fork_team_threads()
981 master_th->th.th_dispatch = &team->t.t_dispatch[0]; in __kmp_fork_team_threads()
986 kmp_hot_team_ptr_t *hot_teams = master_th->th.th_hot_teams; in __kmp_fork_team_threads()
990 if (master_th->th.th_teams_microtask) { // are we inside the teams? in __kmp_fork_team_threads()
991 if (master_th->th.th_teams_size.nteams > 1) { in __kmp_fork_team_threads()
996 master_th->th.th_teams_level == team->t.t_level) { in __kmp_fork_team_threads()
1021 team->t.t_threads[0] = master_th; in __kmp_fork_team_threads()
1022 __kmp_initialize_info(master_th, team, 0, master_gtid); in __kmp_fork_team_threads()
1039 thr->th.th_teams_microtask = master_th->th.th_teams_microtask; in __kmp_fork_team_threads()
1040 thr->th.th_teams_level = master_th->th.th_teams_level; in __kmp_fork_team_threads()
1041 thr->th.th_teams_size = master_th->th.th_teams_size; in __kmp_fork_team_threads()
1408 kmp_info_t *master_th; in __kmp_fork_call() local
1441 master_th = __kmp_threads[gtid]; // AC: potentially unsafe, not in sync with in __kmp_fork_call()
1443 parent_team = master_th->th.th_team; in __kmp_fork_call()
1444 master_tid = master_th->th.th_info.ds.ds_tid; in __kmp_fork_call()
1445 master_this_cons = master_th->th.th_local.this_construct; in __kmp_fork_call()
1446 root = master_th->th.th_root; in __kmp_fork_call()
1448 master_set_numthreads = master_th->th.th_set_nproc; in __kmp_fork_call()
1469 teams_level = master_th->th.th_teams_level; in __kmp_fork_call()
1471 p_hot_teams = &master_th->th.th_hot_teams; in __kmp_fork_call()
1495 master_th->th.ompt_thread_info.state = ompt_state_overhead; in __kmp_fork_call()
1499 master_th->th.th_ident = loc; in __kmp_fork_call()
1501 if (master_th->th.th_teams_microtask && ap && in __kmp_fork_call()
1513 if (parent_team == master_th->th.th_serial_team) { in __kmp_fork_call()
1532 __ompt_lw_taskteam_init(&lw_taskteam, master_th, gtid, in __kmp_fork_call()
1536 __ompt_lw_taskteam_link(&lw_taskteam, master_th, 0); in __kmp_fork_call()
1540 implicit_task_data = OMPT_CUR_TASK_DATA(master_th); in __kmp_fork_call()
1542 OMPT_CUR_TASK_INFO(master_th) in __kmp_fork_call()
1545 ompt_scope_begin, OMPT_CUR_TEAM_DATA(master_th), in __kmp_fork_call()
1547 OMPT_CUR_TASK_INFO(master_th)->thread_num, ompt_task_implicit); in __kmp_fork_call()
1551 master_th->th.ompt_thread_info.state = ompt_state_work_parallel; in __kmp_fork_call()
1574 OMPT_CUR_TASK_INFO(master_th)->frame.exit_frame = ompt_data_none; in __kmp_fork_call()
1578 OMPT_CUR_TASK_INFO(master_th)->thread_num, ompt_task_implicit); in __kmp_fork_call()
1580 ompt_parallel_data = *OMPT_CUR_TEAM_DATA(master_th); in __kmp_fork_call()
1581 __ompt_lw_taskteam_unlink(master_th); in __kmp_fork_call()
1584 &ompt_parallel_data, OMPT_CUR_TASK_DATA(master_th), in __kmp_fork_call()
1588 master_th->th.ompt_thread_info.state = ompt_state_overhead; in __kmp_fork_call()
1599 parent_team->t.t_def_allocator = master_th->th.th_def_allocator; // save in __kmp_fork_call()
1604 __ompt_lw_taskteam_init(&lw_taskteam, master_th, gtid, in __kmp_fork_call()
1606 __ompt_lw_taskteam_link(&lw_taskteam, master_th, 1, true); in __kmp_fork_call()
1612 if (master_set_numthreads < master_th->th.th_teams_size.nth) { in __kmp_fork_call()
1621 master_th->th.th_set_nproc = 0; in __kmp_fork_call()
1638 && master_th->th.th_teams_size.nteams == 1) { in __kmp_fork_call()
1640 master_th->th.th_frame_time = tmp_time; in __kmp_fork_call()
1651 root, parent_team, master_th, gtid)); in __kmp_fork_call()
1655 root, parent_team, master_th, gtid)); in __kmp_fork_call()
1678 KMP_DEBUG_ASSERT(master_th->th.th_task_team == in __kmp_fork_call()
1679 parent_team->t.t_task_team[master_th->th.th_task_state]); in __kmp_fork_call()
1684 master_th->th.th_current_task->td_icvs.max_active_levels) { in __kmp_fork_call()
1700 if ((get__max_active_levels(master_th) == 1 && in __kmp_fork_call()
1729 master_th->th.th_set_nproc = 0; in __kmp_fork_call()
1749 master_th->th.th_serial_team->t.t_ident = loc; in __kmp_fork_call()
1752 master_th->th.th_serial_team->t.t_level--; in __kmp_fork_call()
1763 __ompt_lw_taskteam_init(&lw_taskteam, master_th, gtid, in __kmp_fork_call()
1766 __ompt_lw_taskteam_link(&lw_taskteam, master_th, 0); in __kmp_fork_call()
1769 task_info = OMPT_CUR_TASK_INFO(master_th); in __kmp_fork_call()
1772 OMPT_CUR_TASK_INFO(master_th) in __kmp_fork_call()
1775 ompt_scope_begin, OMPT_CUR_TEAM_DATA(master_th), in __kmp_fork_call()
1777 OMPT_CUR_TASK_INFO(master_th)->thread_num, in __kmp_fork_call()
1782 master_th->th.ompt_thread_info.state = ompt_state_work_parallel; in __kmp_fork_call()
1806 OMPT_CUR_TASK_INFO(master_th)->thread_num, in __kmp_fork_call()
1809 ompt_parallel_data = *OMPT_CUR_TEAM_DATA(master_th); in __kmp_fork_call()
1810 __ompt_lw_taskteam_unlink(master_th); in __kmp_fork_call()
1817 master_th->th.ompt_thread_info.state = ompt_state_overhead; in __kmp_fork_call()
1821 KMP_DEBUG_ASSERT(master_th->th.th_team == in __kmp_fork_call()
1822 master_th->th.th_serial_team); in __kmp_fork_call()
1823 team = master_th->th.th_team; in __kmp_fork_call()
1844 ompt_task_info_t *task_info = OMPT_CUR_TASK_INFO(master_th); in __kmp_fork_call()
1848 OMPT_CUR_TASK_INFO(master_th)->thread_num, ompt_task_initial); in __kmp_fork_call()
1856 master_th->th.ompt_thread_info.state = ompt_state_overhead; in __kmp_fork_call()
1873 __ompt_lw_taskteam_init(&lw_taskteam, master_th, gtid, in __kmp_fork_call()
1875 __ompt_lw_taskteam_link(&lw_taskteam, master_th, 0); in __kmp_fork_call()
1877 task_info = OMPT_CUR_TASK_INFO(master_th); in __kmp_fork_call()
1881 implicit_task_data = OMPT_CUR_TASK_DATA(master_th); in __kmp_fork_call()
1884 ompt_scope_begin, OMPT_CUR_TEAM_DATA(master_th), in __kmp_fork_call()
1887 OMPT_CUR_TASK_INFO(master_th) in __kmp_fork_call()
1892 master_th->th.ompt_thread_info.state = ompt_state_work_parallel; in __kmp_fork_call()
1915 OMPT_CUR_TASK_INFO(master_th)->thread_num, in __kmp_fork_call()
1919 ompt_parallel_data = *OMPT_CUR_TEAM_DATA(master_th); in __kmp_fork_call()
1920 __ompt_lw_taskteam_unlink(master_th); in __kmp_fork_call()
1927 master_th->th.ompt_thread_info.state = ompt_state_overhead; in __kmp_fork_call()
1934 __ompt_lw_taskteam_init(&lwt, master_th, gtid, &ompt_parallel_data, in __kmp_fork_call()
1938 __ompt_lw_taskteam_link(&lwt, master_th, 1); in __kmp_fork_call()
1959 parent_team->t.t_active_level, master_th, in __kmp_fork_call()
1960 master_th->th.th_current_task, in __kmp_fork_call()
1961 master_th->th.th_current_task->td_icvs.max_active_levels)); in __kmp_fork_call()
1965 master_th->th.th_current_task->td_flags.executing = 0; in __kmp_fork_call()
1967 if (!master_th->th.th_teams_microtask || level > teams_level) { in __kmp_fork_call()
1973 int nthreads_icv = master_th->th.th_current_task->td_icvs.nproc; in __kmp_fork_call()
1982 kmp_proc_bind_t proc_bind = master_th->th.th_set_proc_bind; in __kmp_fork_call()
1985 if (master_th->th.th_current_task->td_icvs.proc_bind == proc_bind_false) { in __kmp_fork_call()
1991 proc_bind = master_th->th.th_current_task->td_icvs.proc_bind; in __kmp_fork_call()
1999 master_th->th.th_current_task->td_icvs.proc_bind)) { in __kmp_fork_call()
2005 master_th->th.th_set_proc_bind = proc_bind_default; in __kmp_fork_call()
2009 copy_icvs(&new_icvs, &master_th->th.th_current_task->td_icvs); in __kmp_fork_call()
2025 argc USE_NESTED_HOT_ARG(master_th)); in __kmp_fork_call()
2034 &master_th->th.th_current_task->td_icvs, in __kmp_fork_call()
2035 argc USE_NESTED_HOT_ARG(master_th)); in __kmp_fork_call()
2052 if (!master_th->th.th_teams_microtask || level > teams_level) { in __kmp_fork_call()
2069 KMP_CHECK_UPDATE(team->t.t_def_allocator, master_th->th.th_def_allocator); in __kmp_fork_call()
2077 KMP_DEBUG_ASSERT(master_th->th.th_task_team == in __kmp_fork_call()
2078 parent_team->t.t_task_team[master_th->th.th_task_state]); in __kmp_fork_call()
2081 __kmp_gtid_from_thread(master_th), in __kmp_fork_call()
2082 master_th->th.th_task_team, parent_team, in __kmp_fork_call()
2083 team->t.t_task_team[master_th->th.th_task_state], team)); in __kmp_fork_call()
2085 if (active_level || master_th->th.th_task_team) { in __kmp_fork_call()
2087 KMP_DEBUG_ASSERT(master_th->th.th_task_state_memo_stack); in __kmp_fork_call()
2088 if (master_th->th.th_task_state_top >= in __kmp_fork_call()
2089 master_th->th.th_task_state_stack_sz) { // increase size in __kmp_fork_call()
2090 kmp_uint32 new_size = 2 * master_th->th.th_task_state_stack_sz; in __kmp_fork_call()
2094 for (i = 0; i < master_th->th.th_task_state_stack_sz; ++i) { in __kmp_fork_call()
2095 new_stack[i] = master_th->th.th_task_state_memo_stack[i]; in __kmp_fork_call()
2097 for (i = master_th->th.th_task_state_stack_sz; i < new_size; in __kmp_fork_call()
2101 old_stack = master_th->th.th_task_state_memo_stack; in __kmp_fork_call()
2102 master_th->th.th_task_state_memo_stack = new_stack; in __kmp_fork_call()
2103 master_th->th.th_task_state_stack_sz = new_size; in __kmp_fork_call()
2107 master_th->th in __kmp_fork_call()
2108 .th_task_state_memo_stack[master_th->th.th_task_state_top] = in __kmp_fork_call()
2109 master_th->th.th_task_state; in __kmp_fork_call()
2110 master_th->th.th_task_state_top++; in __kmp_fork_call()
2112 if (master_th->th.th_hot_teams && in __kmp_fork_call()
2114 team == master_th->th.th_hot_teams[active_level].hot_team) { in __kmp_fork_call()
2116 master_th->th.th_task_state = in __kmp_fork_call()
2117 master_th->th in __kmp_fork_call()
2118 .th_task_state_memo_stack[master_th->th.th_task_state_top]; in __kmp_fork_call()
2121 master_th->th.th_task_state = 0; in __kmp_fork_call()
2127 KMP_DEBUG_ASSERT((master_th->th.th_task_team == NULL) || in __kmp_fork_call()
2163 __kmp_fork_team_threads(root, team, master_th, gtid); in __kmp_fork_call()
2165 &master_th->th.th_current_task->td_icvs, loc); in __kmp_fork_call()
2168 master_th->th.ompt_thread_info.state = ompt_state_work_parallel; in __kmp_fork_call()
2175 && !master_th->th.th_teams_microtask) { // not in teams construct in __kmp_fork_call()
2184 master_th->th.th_frame_time = tmp_time; in __kmp_fork_call()
2203 root, team, master_th, gtid)); in __kmp_fork_call()
2219 root, team, master_th, gtid)); in __kmp_fork_call()
2259 master_th->th.ompt_thread_info.state = ompt_state_overhead; in __kmp_fork_call()
2299 kmp_info_t *master_th; in __kmp_join_call() local
2306 master_th = __kmp_threads[gtid]; in __kmp_join_call()
2307 root = master_th->th.th_root; in __kmp_join_call()
2308 team = master_th->th.th_team; in __kmp_join_call()
2311 master_th->th.th_ident = loc; in __kmp_join_call()
2320 master_th->th.ompt_thread_info.state = ompt_state_overhead; in __kmp_join_call()
2328 __kmp_gtid_from_thread(master_th), team, in __kmp_join_call()
2329 team->t.t_task_team[master_th->th.th_task_state], in __kmp_join_call()
2330 master_th->th.th_task_team)); in __kmp_join_call()
2331 KMP_DEBUG_ASSERT(master_th->th.th_task_team == in __kmp_join_call()
2332 team->t.t_task_team[master_th->th.th_task_state]); in __kmp_join_call()
2337 if (master_th->th.th_teams_microtask) { in __kmp_join_call()
2340 int tlevel = master_th->th.th_teams_level; in __kmp_join_call()
2356 __kmp_join_restore_state(master_th, parent_team); in __kmp_join_call()
2370 master_th->th.th_task_state = in __kmp_join_call()
2388 (!master_th->th.th_teams_microtask || /* not in teams construct */ in __kmp_join_call()
2389 master_th->th.th_teams_size.nteams == 1)) { in __kmp_join_call()
2390 master_th->th.th_ident = loc; in __kmp_join_call()
2396 master_th->th.th_frame_time, 0, loc, in __kmp_join_call()
2397 master_th->th.th_team_nproc, 1); in __kmp_join_call()
2404 if (master_th->th.th_teams_microtask && !exit_teams && in __kmp_join_call()
2406 team->t.t_level == master_th->th.th_teams_level + 1) { in __kmp_join_call()
2418 OMPT_CUR_TASK_INFO(master_th)->thread_num, ompt_task_implicit); in __kmp_join_call()
2422 ompt_parallel_data = *OMPT_CUR_TEAM_DATA(master_th); in __kmp_join_call()
2423 __ompt_lw_taskteam_unlink(master_th); in __kmp_join_call()
2435 if (master_th->th.th_team_nproc < master_th->th.th_teams_size.nth) { in __kmp_join_call()
2436 int old_num = master_th->th.th_team_nproc; in __kmp_join_call()
2437 int new_num = master_th->th.th_teams_size.nth; in __kmp_join_call()
2457 other_threads[i]->th.th_task_state = master_th->th.th_task_state; in __kmp_join_call()
2464 __kmp_join_ompt(gtid, master_th, parent_team, &ompt_parallel_data, in __kmp_join_call()
2473 master_th->th.th_info.ds.ds_tid = team->t.t_master_tid; in __kmp_join_call()
2474 master_th->th.th_local.this_construct = team->t.t_master_this_cons; in __kmp_join_call()
2476 master_th->th.th_dispatch = &parent_team->t.t_dispatch[team->t.t_master_tid]; in __kmp_join_call()
2483 if (!master_th->th.th_teams_microtask || in __kmp_join_call()
2484 team->t.t_level > master_th->th.th_teams_level) { in __kmp_join_call()
2500 OMPT_CUR_TASK_INFO(master_th)->thread_num, flags); in __kmp_join_call()
2508 master_th, team)); in __kmp_join_call()
2509 __kmp_pop_current_task_from_thread(master_th); in __kmp_join_call()
2513 master_th->th.th_first_place = team->t.t_first_place; in __kmp_join_call()
2514 master_th->th.th_last_place = team->t.t_last_place; in __kmp_join_call()
2516 master_th->th.th_def_allocator = team->t.t_def_allocator; in __kmp_join_call()
2524 master_th)); // this will free worker threads in __kmp_join_call()
2531 master_th->th.th_team = parent_team; in __kmp_join_call()
2532 master_th->th.th_team_nproc = parent_team->t.t_nproc; in __kmp_join_call()
2533 master_th->th.th_team_master = parent_team->t.t_threads[0]; in __kmp_join_call()
2534 master_th->th.th_team_serialized = parent_team->t.t_serialized; in __kmp_join_call()
2538 parent_team != master_th->th.th_serial_team && in __kmp_join_call()
2541 master_th->th.th_serial_team USE_NESTED_HOT_ARG(NULL)); in __kmp_join_call()
2542 master_th->th.th_serial_team = parent_team; in __kmp_join_call()
2546 if (master_th->th.th_task_state_top > in __kmp_join_call()
2548 KMP_DEBUG_ASSERT(master_th->th.th_task_state_memo_stack); in __kmp_join_call()
2550 master_th->th.th_task_state_memo_stack[master_th->th.th_task_state_top] = in __kmp_join_call()
2551 master_th->th.th_task_state; in __kmp_join_call()
2552 --master_th->th.th_task_state_top; // pop in __kmp_join_call()
2554 master_th->th.th_task_state = in __kmp_join_call()
2555 master_th->th in __kmp_join_call()
2556 .th_task_state_memo_stack[master_th->th.th_task_state_top]; in __kmp_join_call()
2559 master_th->th.th_task_team = in __kmp_join_call()
2560 parent_team->t.t_task_team[master_th->th.th_task_state]; in __kmp_join_call()
2563 __kmp_gtid_from_thread(master_th), master_th->th.th_task_team, in __kmp_join_call()
2570 master_th->th.th_current_task->td_flags.executing = 1; in __kmp_join_call()
2580 __kmp_join_ompt(gtid, master_th, parent_team, parallel_data, flags, in __kmp_join_call()
4562 kmp_info_t *master_th = team->t.t_threads[0]; in __kmp_partition_places() local
4563 KMP_DEBUG_ASSERT(master_th != NULL); in __kmp_partition_places()
4565 int first_place = master_th->th.th_first_place; in __kmp_partition_places()
4566 int last_place = master_th->th.th_last_place; in __kmp_partition_places()
4567 int masters_place = master_th->th.th_current_place; in __kmp_partition_places()