Lines Matching refs:parent_team
766 static int __kmp_reserve_threads(kmp_root_t *root, kmp_team_t *parent_team, in __kmp_reserve_threads() argument
772 KMP_DEBUG_ASSERT(root && parent_team); in __kmp_reserve_threads()
773 kmp_info_t *this_thr = parent_team->t.t_threads[master_tid]; in __kmp_reserve_threads()
778 if (!get__dynamic_2(parent_team, master_tid)) { in __kmp_reserve_threads()
815 new_nthreads = __kmp_get_random(parent_team->t.t_threads[master_tid]); in __kmp_reserve_threads()
844 if (!get__dynamic_2(parent_team, master_tid) && (!__kmp_reserve_warn)) { in __kmp_reserve_threads()
875 if (!get__dynamic_2(parent_team, master_tid) && (!__kmp_reserve_warn)) { in __kmp_reserve_threads()
920 if (!get__dynamic_2(parent_team, master_tid) && (!__kmp_reserve_warn)) { in __kmp_reserve_threads()
1465 __kmp_fork_in_teams(ident_t *loc, int gtid, kmp_team_t *parent_team, in __kmp_fork_in_teams() argument
1476 parent_team->t.t_ident = loc; in __kmp_fork_in_teams()
1477 __kmp_alloc_argv_entries(argc, parent_team, TRUE); in __kmp_fork_in_teams()
1478 parent_team->t.t_argc = argc; in __kmp_fork_in_teams()
1479 argv = (void **)parent_team->t.t_argv; in __kmp_fork_in_teams()
1484 if (parent_team == master_th->th.th_serial_team) { in __kmp_fork_in_teams()
1487 KMP_DEBUG_ASSERT(parent_team->t.t_serialized > 1); in __kmp_fork_in_teams()
1492 parent_team->t.t_serialized--; in __kmp_fork_in_teams()
1497 parent_team->t.t_pkfn = microtask; in __kmp_fork_in_teams()
1532 parent_team->t.t_serialized--; in __kmp_fork_in_teams()
1537 __kmp_invoke_microtask(microtask, gtid, 0, argc, parent_team->t.t_argv in __kmp_fork_in_teams()
1567 parent_team->t.t_pkfn = microtask; in __kmp_fork_in_teams()
1568 parent_team->t.t_invoke = invoker; in __kmp_fork_in_teams()
1570 parent_team->t.t_active_level++; in __kmp_fork_in_teams()
1571 parent_team->t.t_level++; in __kmp_fork_in_teams()
1572 parent_team->t.t_def_allocator = master_th->th.th_def_allocator; // save in __kmp_fork_in_teams()
1579 master_th->th.th_teams_size.nth = parent_team->t.t_nproc; in __kmp_fork_in_teams()
1594 kmp_info_t **other_threads = parent_team->t.t_threads; in __kmp_fork_in_teams()
1599 __kmp_resize_dist_barrier(parent_team, old_proc, master_set_numthreads); in __kmp_fork_in_teams()
1600 __kmp_add_threads_to_team(parent_team, master_set_numthreads); in __kmp_fork_in_teams()
1602 parent_team->t.t_nproc = master_set_numthreads; in __kmp_fork_in_teams()
1641 KMP_CHECK_UPDATE(parent_team->t.t_proc_bind, proc_bind); in __kmp_fork_in_teams()
1645 kmp_info_t **other_threads = parent_team->t.t_threads; in __kmp_fork_in_teams()
1657 parent_team->t.t_active_level == 1 // only report frames at level 1 in __kmp_fork_in_teams()
1661 parent_team->t.t_region_time = tmp_time; in __kmp_fork_in_teams()
1664 KMP_DEBUG_ASSERT(parent_team->t.t_stack_id == NULL); in __kmp_fork_in_teams()
1666 parent_team->t.t_stack_id = __kmp_itt_stack_caller_create(); in __kmp_fork_in_teams()
1670 __kmp_partition_places(parent_team); in __kmp_fork_in_teams()
1675 root, parent_team, master_th, gtid)); in __kmp_fork_in_teams()
1676 __kmp_internal_fork(loc, gtid, parent_team); in __kmp_fork_in_teams()
1679 root, parent_team, master_th, gtid)); in __kmp_fork_in_teams()
1686 parent_team->t.t_id, parent_team->t.t_pkfn)); in __kmp_fork_in_teams()
1688 if (!parent_team->t.t_invoke(gtid)) { in __kmp_fork_in_teams()
1692 parent_team->t.t_id, parent_team->t.t_pkfn)); in __kmp_fork_in_teams()
1704 kmp_info_t *master_th, kmp_team_t *parent_team, in __kmp_serial_fork_call() argument
1772 __kmp_invoke_microtask(microtask, gtid, 0, argc, parent_team->t.t_argv in __kmp_serial_fork_call()
1936 kmp_team_t *parent_team; in __kmp_fork_call() local
1975 parent_team = master_th->th.th_team; in __kmp_fork_call()
2001 level = parent_team->t.t_level; in __kmp_fork_call()
2003 active_level = parent_team->t.t_active_level; in __kmp_fork_call()
2022 : get__nproc_2(parent_team, master_tid); in __kmp_fork_call()
2039 return __kmp_fork_in_teams(loc, gtid, parent_team, argc, master_th, root, in __kmp_fork_call()
2052 KMP_DEBUG_ASSERT_TASKTEAM_INVARIANT(parent_team, master_th); in __kmp_fork_call()
2058 (parent_team->t.t_active_level >= in __kmp_fork_call()
2067 : get__nproc_2(parent_team, master_tid); in __kmp_fork_call()
2082 nthreads = __kmp_reserve_threads(root, parent_team, master_tid, in __kmp_fork_call()
2099 invoker, master_th, parent_team, in __kmp_fork_call()
2111 parent_team->t.t_active_level, master_th, in __kmp_fork_call()
2128 (level + 1 < parent_team->t.t_nested_nth->used) && in __kmp_fork_call()
2129 (parent_team->t.t_nested_nth->nth[level + 1] != nthreads_icv)) { in __kmp_fork_call()
2130 nthreads_icv = parent_team->t.t_nested_nth->nth[level + 1]; in __kmp_fork_call()
2219 KMP_CHECK_UPDATE(team->t.t_parent, parent_team); in __kmp_fork_call()
2228 int new_level = parent_team->t.t_level + 1; in __kmp_fork_call()
2230 new_level = parent_team->t.t_active_level + 1; in __kmp_fork_call()
2234 int new_level = parent_team->t.t_level; in __kmp_fork_call()
2236 new_level = parent_team->t.t_active_level; in __kmp_fork_call()
2239 kmp_r_sched_t new_sched = get__sched_2(parent_team, master_tid); in __kmp_fork_call()
2248 team->t.t_nested_nth != parent_team->t.t_nested_nth) { in __kmp_fork_call()
2253 team->t.t_nested_nth = parent_team->t.t_nested_nth; in __kmp_fork_call()
2274 gtid, parent_team->t.t_id, team->t.t_master_tid, team->t.t_id, in __kmp_fork_call()
2350 } else if (parent_team->t.t_serialized) { in __kmp_fork_call()
2355 KMP_DEBUG_ASSERT(parent_team->t.t_stack_id == NULL); in __kmp_fork_call()
2356 parent_team->t.t_stack_id = __kmp_itt_stack_caller_create(); in __kmp_fork_call()
2445 kmp_team_t *parent_team; in __kmp_join_call() local
2456 parent_team = team->t.t_parent; in __kmp_join_call()
2505 __kmp_join_restore_state(master_th, parent_team); in __kmp_join_call()
2530 if (__itt_stack_caller_create_ptr && parent_team->t.t_serialized) { in __kmp_join_call()
2531 KMP_DEBUG_ASSERT(parent_team->t.t_stack_id != NULL); in __kmp_join_call()
2535 __kmp_itt_stack_caller_destroy((__itt_caller)parent_team->t.t_stack_id); in __kmp_join_call()
2536 parent_team->t.t_stack_id = NULL; in __kmp_join_call()
2635 __kmp_join_ompt(gtid, master_th, parent_team, &ompt_parallel_data, in __kmp_join_call()
2647 master_th->th.th_dispatch = &parent_team->t.t_dispatch[team->t.t_master_tid]; in __kmp_join_call()
2701 master_th->th.th_team = parent_team; in __kmp_join_call()
2702 master_th->th.th_team_nproc = parent_team->t.t_nproc; in __kmp_join_call()
2703 master_th->th.th_team_master = parent_team->t.t_threads[0]; in __kmp_join_call()
2704 master_th->th.th_team_serialized = parent_team->t.t_serialized; in __kmp_join_call()
2707 if (parent_team->t.t_serialized && in __kmp_join_call()
2708 parent_team != master_th->th.th_serial_team && in __kmp_join_call()
2709 parent_team != root->r.r_root_team) { in __kmp_join_call()
2712 master_th->th.th_serial_team = parent_team; in __kmp_join_call()
2723 parent_team->t.t_task_team[master_th->th.th_task_state]; in __kmp_join_call()
2727 parent_team)); in __kmp_join_call()
2748 __kmp_join_ompt(gtid, master_th, parent_team, parallel_data, flags, in __kmp_join_call()