Lines Matching refs:t_nproc

689     team->t.t_ordered.dt.t_value = ((tid + 1) % team->t.t_nproc);  in __kmp_parallel_dxo()
799 (root->r.r_active ? 1 : root->r.r_hot_team->t.t_nproc); in __kmp_reserve_threads()
835 (root->r.r_active ? 1 : root->r.r_hot_team->t.t_nproc) > in __kmp_reserve_threads()
838 (root->r.r_active ? 1 : root->r.r_hot_team->t.t_nproc); in __kmp_reserve_threads()
866 (root->r.r_active ? 1 : root->r.r_hot_team->t.t_nproc) > in __kmp_reserve_threads()
869 (root->r.r_active ? 1 : root->r.r_hot_team->t.t_nproc); in __kmp_reserve_threads()
907 (root->r.r_active ? 1 : root->r.r_hot_team->t.t_nproc) > in __kmp_reserve_threads()
911 (root->r.r_active ? 1 : root->r.r_hot_team->t.t_nproc) - in __kmp_reserve_threads()
965 KA_TRACE(10, ("__kmp_fork_team_threads: new_nprocs = %d\n", team->t.t_nproc)); in __kmp_fork_team_threads()
972 master_th->th.th_team_nproc = team->t.t_nproc; in __kmp_fork_team_threads()
1003 hot_teams[level].hot_team_nth = team->t.t_nproc; in __kmp_fork_team_threads()
1019 for (i = 1; i < team->t.t_nproc; i++) { in __kmp_fork_team_threads()
1058 if (team->t.t_nproc > 1 && in __kmp_fork_team_threads()
1060 team->t.b->update_num_threads(team->t.t_nproc); in __kmp_fork_team_threads()
1061 __kmp_add_threads_to_team(team, team->t.t_nproc); in __kmp_fork_team_threads()
1083 if (team->t.t_nproc > 1) { in __kmp_fork_team_threads()
1101 for (i = 0; i < team->t.t_nproc; i++) { in __kmp_fork_team_threads()
1103 if (thr->th.th_prev_num_threads != team->t.t_nproc || in __kmp_fork_team_threads()
1286 serial_team->t.t_nproc = 1; in __kmp_serialized_parallel()
1579 master_th->th.th_teams_size.nth = parent_team->t.t_nproc; in __kmp_fork_in_teams()
1602 parent_team->t.t_nproc = master_set_numthreads; in __kmp_fork_in_teams()
2275 team->t.t_nproc)); in __kmp_fork_call()
2332 __kmp_itt_region_forking(gtid, team->t.t_nproc, 0); in __kmp_fork_call()
2586 int ompt_team_size = team->t.t_nproc; in __kmp_join_call()
2610 team->t.t_nproc = new_num; in __kmp_join_call()
2668 int ompt_team_size = (flags == ompt_task_initial) ? 0 : team->t.t_nproc; in __kmp_join_call()
2702 master_th->th.th_team_nproc = parent_team->t.t_nproc; in __kmp_join_call()
2817 (root->r.r_hot_team->t.t_nproc > new_nth) in __kmp_set_num_threads()
2828 __kmp_resize_dist_barrier(hot_team, hot_team->t.t_nproc, new_nth); in __kmp_set_num_threads()
2831 for (f = new_nth; f < hot_team->t.t_nproc; f++) { in __kmp_set_num_threads()
2841 hot_team->t.t_nproc = new_nth; in __kmp_set_num_threads()
3156 return team->t.t_nproc; in __kmp_get_team_size()
3292 team->t.t_nproc * sizeof(kmp_info_t *)); in __kmp_reallocate_team_arrays()
3393 root_team->t.t_nproc = 1; in __kmp_initialize_root()
3431 hot_team->t.t_nproc = 1; in __kmp_initialize_root()
3603 __kmp_printf(" Number threads: %2d\n", team->t.t_nproc); in __kmp_print_structure()
3604 for (i = 0; i < team->t.t_nproc; ++i) { in __kmp_print_structure()
4099 int n = hot_team->t.t_nproc; in __kmp_reset_root()
4112 for (i = 0; i < hot_team->t.t_nproc; ++i) { in __kmp_reset_root()
4309 this_thr->th.th_team_nproc = team->t.t_nproc; in __kmp_initialize_info()
4759 team->t.t_nproc = new_nproc; in __kmp_initialize_team()
4842 KMP_DEBUG_ASSERT(team->t.t_nproc == 1); in __kmp_partition_places()
4847 int n_th = team->t.t_nproc; in __kmp_partition_places()
4862 int n_th = team->t.t_nproc; in __kmp_partition_places()
4942 int n_th = team->t.t_nproc; in __kmp_partition_places()
5204 if (team->t.t_nproc != new_nproc && in __kmp_allocate_team()
5207 int old_nthr = team->t.t_nproc; in __kmp_allocate_team()
5218 if (team->t.t_nproc == new_nproc) { // Check changes in number of threads in __kmp_allocate_team()
5262 } else if (team->t.t_nproc > new_nproc) { in __kmp_allocate_team()
5276 for (f = new_nproc; f < team->t.t_nproc; f++) { in __kmp_allocate_team()
5286 KMP_DEBUG_ASSERT(hot_teams[level].hot_team_nth == team->t.t_nproc); in __kmp_allocate_team()
5290 for (f = new_nproc; f < team->t.t_nproc; f++) { in __kmp_allocate_team()
5300 for (f = new_nproc; f < team->t.t_nproc; ++f) { in __kmp_allocate_team()
5312 team->t.t_nproc = new_nproc; in __kmp_allocate_team()
5331 for (f = 0; f < team->t.t_nproc; f++) { in __kmp_allocate_team()
5334 team->t.t_nproc); in __kmp_allocate_team()
5349 int old_nproc = team->t.t_nproc; // save old value and use to update only in __kmp_allocate_team()
5357 for (f = team->t.t_nproc; f < avail_threads; ++f) { in __kmp_allocate_team()
5374 team->t.t_nproc = new_nproc; // just get reserved threads involved in __kmp_allocate_team()
5378 team->t.t_nproc = hot_teams[level].hot_team_nth; in __kmp_allocate_team()
5398 for (f = team->t.t_nproc; f < new_nproc; f++) { in __kmp_allocate_team()
5443 KMP_DEBUG_ASSERT(team->t.t_nproc == new_nproc); in __kmp_allocate_team()
5444 for (f = 0; f < team->t.t_nproc; ++f) in __kmp_allocate_team()
5450 for (f = old_nproc; f < team->t.t_nproc; ++f) in __kmp_allocate_team()
5454 for (f = 0; f < team->t.t_nproc; ++f) { in __kmp_allocate_team()
5457 team->t.t_nproc); in __kmp_allocate_team()
5669 KMP_DEBUG_ASSERT(team->t.t_nproc <= team->t.t_max_nproc); in __kmp_free_team()
5710 for (f = 1; f < team->t.t_nproc; ++f) { in __kmp_free_team()
5735 for (f = 0; f < team->t.t_nproc; ++f) { // threads unref task teams in __kmp_free_team()
5765 for (f = 1; f < team->t.t_nproc; ++f) { in __kmp_free_team()
5779 for (f = 1; f < team->t.t_nproc; ++f) { in __kmp_free_team()
5788 for (int f = 1; f < team->t.t_nproc; ++f) { in __kmp_free_team()
5795 for (f = 1; f < team->t.t_nproc; ++f) { in __kmp_free_team()
5814 for (f = 1; f < team->t.t_nproc; ++f) { in __kmp_free_team()
7721 ompt_team_size = team->t.t_nproc; in __kmp_invoke_task_func()
7839 ompt_scope_begin, parallel_data, task_data, team->t.t_nproc, tid, in __kmp_invoke_teams_master()
8092 for (f = 0; f < team->t.t_nproc; f++) { in __kmp_internal_fork()
8094 team->t.t_threads[f]->th.th_team_nproc == team->t.t_nproc); in __kmp_internal_fork()
8114 __kmp_threads[gtid]->th.th_team_nproc != team->t.t_nproc) { in __kmp_internal_join()
8120 team->t.t_nproc); in __kmp_internal_join()
8124 __kmp_threads[gtid]->th.th_team_nproc == team->t.t_nproc); in __kmp_internal_join()
8183 return hot_team->t.t_nproc - 1; // Don't count primary thread in __kmp_active_hot_team_nproc()
8188 for (i = 1; i < hot_team->t.t_nproc; i++) { in __kmp_active_hot_team_nproc()
8241 (root->r.r_active ? 1 : root->r.r_hot_team->t.t_nproc); in __kmp_load_balance_nproc()
8571 return team->t.t_parent->t.t_nproc; in __kmp_aux_get_num_teams()
8761 rc = __kmp_str_buf_print(field_buffer, format, th->th.th_team->t.t_nproc); in __kmp_aux_capture_affinity_field()