Lines Matching refs:gtid
1131 extern omp_allocator_handle_t __kmpc_init_allocator(int gtid,
1135 extern void __kmpc_destroy_allocator(int gtid, omp_allocator_handle_t al);
1136 extern void __kmpc_set_default_allocator(int gtid, omp_allocator_handle_t al);
1137 extern omp_allocator_handle_t __kmpc_get_default_allocator(int gtid);
1139 extern void *__kmpc_alloc(int gtid, size_t sz, omp_allocator_handle_t al);
1140 extern void *__kmpc_aligned_alloc(int gtid, size_t align, size_t sz,
1142 extern void *__kmpc_calloc(int gtid, size_t nmemb, size_t sz,
1144 extern void *__kmpc_realloc(int gtid, void *ptr, size_t sz,
1147 extern void __kmpc_free(int gtid, void *ptr, omp_allocator_handle_t al);
1149 extern void *__kmp_alloc(int gtid, size_t align, size_t sz,
1151 extern void *__kmp_calloc(int gtid, size_t align, size_t nmemb, size_t sz,
1153 extern void *__kmp_realloc(int gtid, void *ptr, size_t sz,
1156 extern void ___kmpc_free(int gtid, void *ptr, omp_allocator_handle_t al);
1327 #define KMP_MASTER_GTID(gtid) (0 == __kmp_tid_from_gtid((gtid))) argument
1328 #define KMP_WORKER_GTID(gtid) (0 != __kmp_tid_from_gtid((gtid))) argument
1329 #define KMP_INITIAL_GTID(gtid) (0 == (gtid)) argument
2079 void (*th_deo_fcn)(int *gtid, int *cid, ident_t *);
2081 void (*th_dxo_fcn)(int *gtid, int *cid, ident_t *);
3109 typedef int (*launch_t)(int gtid);
3618 #define __kmp_get_team_num_threads(gtid) \ argument
3619 (__kmp_threads[(gtid)]->th.th_team->t.t_nproc)
3621 static inline bool KMP_UBER_GTID(int gtid) { in KMP_UBER_GTID() argument
3622 KMP_DEBUG_ASSERT(gtid >= KMP_GTID_MIN); in KMP_UBER_GTID()
3623 KMP_DEBUG_ASSERT(gtid < __kmp_threads_capacity); in KMP_UBER_GTID()
3624 return (gtid >= 0 && __kmp_root[gtid] && __kmp_threads[gtid] && in KMP_UBER_GTID()
3625 __kmp_threads[gtid] == __kmp_root[gtid]->r.r_uber_thread); in KMP_UBER_GTID()
3628 static inline int __kmp_tid_from_gtid(int gtid) { in __kmp_tid_from_gtid() argument
3629 KMP_DEBUG_ASSERT(gtid >= 0); in __kmp_tid_from_gtid()
3630 return __kmp_threads[gtid]->th.th_info.ds.ds_tid; in __kmp_tid_from_gtid()
3643 static inline kmp_info_t *__kmp_thread_from_gtid(int gtid) { in __kmp_thread_from_gtid() argument
3644 KMP_DEBUG_ASSERT(gtid >= 0); in __kmp_thread_from_gtid()
3645 return __kmp_threads[gtid]; in __kmp_thread_from_gtid()
3648 static inline kmp_team_t *__kmp_team_from_gtid(int gtid) { in __kmp_team_from_gtid() argument
3649 KMP_DEBUG_ASSERT(gtid >= 0); in __kmp_team_from_gtid()
3650 return __kmp_threads[gtid]->th.th_team; in __kmp_team_from_gtid()
3653 static inline void __kmp_assert_valid_gtid(kmp_int32 gtid) { in __kmp_assert_valid_gtid() argument
3654 if (UNLIKELY(gtid < 0 || gtid >= __kmp_threads_capacity)) in __kmp_assert_valid_gtid()
3694 extern void __kmp_print_storage_map_gtid(int gtid, void *p1, void *p2,
3702 extern void __kmp_internal_end_library(int gtid);
3703 extern void __kmp_internal_end_thread(int gtid);
3709 extern void __kmp_unregister_root(int gtid);
3715 extern int __kmp_enter_single(int gtid, ident_t *id_ref, int push_ws);
3716 extern void __kmp_exit_single(int gtid);
3733 extern void __kmp_set_num_threads(int new_nth, int gtid);
3741 int gtid = __kmp_entry_gtid(); in __kmp_entry_thread() local
3743 return __kmp_threads[gtid]; in __kmp_entry_thread()
3746 extern void __kmp_set_max_active_levels(int gtid, int new_max_active_levels);
3747 extern int __kmp_get_max_active_levels(int gtid);
3748 extern int __kmp_get_ancestor_thread_num(int gtid, int level);
3749 extern int __kmp_get_team_size(int gtid, int level);
3750 extern void __kmp_set_schedule(int gtid, kmp_sched_t new_sched, int chunk);
3751 extern void __kmp_get_schedule(int gtid, kmp_sched_t *sched, int *chunk);
3794 extern void __kmp_push_num_threads(ident_t *loc, int gtid, int num_threads);
3795 extern void __kmp_push_num_threads_list(ident_t *loc, int gtid,
3798 extern void __kmp_set_strict_num_threads(ident_t *loc, int gtid, int sev,
3801 extern void __kmp_push_proc_bind(ident_t *loc, int gtid,
3803 extern void __kmp_push_num_teams(ident_t *loc, int gtid, int num_teams,
3805 extern void __kmp_push_num_teams_51(ident_t *loc, int gtid, int num_teams_lb,
3810 extern void __kmpc_dispatch_init_4(ident_t *loc, kmp_int32 gtid,
3813 extern void __kmpc_dispatch_init_4u(ident_t *loc, kmp_int32 gtid,
3817 extern void __kmpc_dispatch_init_8(ident_t *loc, kmp_int32 gtid,
3820 extern void __kmpc_dispatch_init_8u(ident_t *loc, kmp_int32 gtid,
3825 extern int __kmpc_dispatch_next_4(ident_t *loc, kmp_int32 gtid,
3828 extern int __kmpc_dispatch_next_4u(ident_t *loc, kmp_int32 gtid,
3831 extern int __kmpc_dispatch_next_8(ident_t *loc, kmp_int32 gtid,
3834 extern int __kmpc_dispatch_next_8u(ident_t *loc, kmp_int32 gtid,
3838 extern void __kmpc_dispatch_fini_4(ident_t *loc, kmp_int32 gtid);
3839 extern void __kmpc_dispatch_fini_8(ident_t *loc, kmp_int32 gtid);
3840 extern void __kmpc_dispatch_fini_4u(ident_t *loc, kmp_int32 gtid);
3841 extern void __kmpc_dispatch_fini_8u(ident_t *loc, kmp_int32 gtid);
3843 extern void __kmpc_dispatch_deinit(ident_t *loc, kmp_int32 gtid);
3847 extern void __kmp_aux_dispatch_init_4(ident_t *loc, kmp_int32 gtid,
3851 extern void __kmp_aux_dispatch_init_4u(ident_t *loc, kmp_int32 gtid,
3855 extern void __kmp_aux_dispatch_init_8(ident_t *loc, kmp_int32 gtid,
3859 extern void __kmp_aux_dispatch_init_8u(ident_t *loc, kmp_int32 gtid,
3863 extern void __kmp_aux_dispatch_fini_chunk_4(ident_t *loc, kmp_int32 gtid);
3864 extern void __kmp_aux_dispatch_fini_chunk_8(ident_t *loc, kmp_int32 gtid);
3865 extern void __kmp_aux_dispatch_fini_chunk_4u(ident_t *loc, kmp_int32 gtid);
3866 extern void __kmp_aux_dispatch_fini_chunk_8u(ident_t *loc, kmp_int32 gtid);
3925 int gtid, int isa_root); /* set affinity according to KMP_AFFINITY */
3926 void __kmp_affinity_bind_init_mask(int gtid);
3927 extern void __kmp_affinity_bind_place(int gtid);
3944 int gtid = __kmp_entry_gtid(); in __kmp_assign_root_init_mask() local
3945 kmp_root_t *r = __kmp_threads[gtid]->th.th_root; in __kmp_assign_root_init_mask()
3946 if (r->r.r_uber_thread == __kmp_threads[gtid] && !r->r.r_affinity_assigned) { in __kmp_assign_root_init_mask()
3947 __kmp_affinity_set_init_mask(gtid, /*isa_root=*/TRUE); in __kmp_assign_root_init_mask()
3948 __kmp_affinity_bind_init_mask(gtid); in __kmp_assign_root_init_mask()
3952 static inline void __kmp_reset_root_init_mask(int gtid) { in __kmp_reset_root_init_mask() argument
3955 kmp_info_t *th = __kmp_threads[gtid]; in __kmp_reset_root_init_mask()
3965 static inline void __kmp_reset_root_init_mask(int gtid) {} in __kmp_reset_root_init_mask() argument
3970 extern size_t __kmp_aux_capture_affinity(int gtid, const char *format,
3972 extern void __kmp_aux_display_affinity(int gtid, const char *format);
3983 extern void __kmp_gtid_set_specific(int gtid);
3996 extern void __kmp_create_worker(int gtid, kmp_info_t *th, size_t stack_size);
4008 extern void __kmp_terminate_thread(int gtid);
4024 extern void __kmp_common_destroy_gtid(int gtid);
4060 extern int __kmp_barrier(enum barrier_type bt, int gtid, int is_split,
4063 extern void __kmp_end_split_barrier(enum barrier_type bt, int gtid);
4064 extern int __kmp_barrier_gomp_cancel(int gtid);
4076 extern int __kmp_fork_call(ident_t *loc, int gtid,
4081 extern void __kmp_join_call(ident_t *loc, int gtid
4089 extern void __kmp_serialized_parallel(ident_t *id, kmp_int32 gtid);
4090 extern void __kmp_internal_fork(ident_t *id, int gtid, kmp_team_t *team);
4091 extern void __kmp_internal_join(ident_t *id, int gtid, kmp_team_t *team);
4092 extern int __kmp_invoke_task_func(int gtid);
4093 extern void __kmp_run_before_invoked_task(int gtid, int tid,
4096 extern void __kmp_run_after_invoked_task(int gtid, int tid,
4101 KMP_EXPORT int __kmpc_invoke_task_func(int gtid);
4102 extern int __kmp_invoke_teams_master(int gtid);
4103 extern void __kmp_teams_master(int gtid);
4122 extern kmp_task_t *__kmp_task_alloc(ident_t *loc_ref, kmp_int32 gtid,
4134 int gtid,
4154 int gtid);
4176 extern int __kmp_invoke_microtask(microtask_t pkfn, int gtid, int npr, int argc,
4258 KMP_EXPORT void *__kmpc_copyprivate_light(ident_t *loc, kmp_int32 gtid,
4266 KMP_EXPORT kmp_int32 __kmpc_omp_task(ident_t *loc_ref, kmp_int32 gtid,
4268 KMP_EXPORT kmp_task_t *__kmpc_omp_task_alloc(ident_t *loc_ref, kmp_int32 gtid,
4274 ident_t *loc_ref, kmp_int32 gtid, kmp_int32 flags, size_t sizeof_kmp_task_t,
4276 KMP_EXPORT void __kmpc_omp_task_begin_if0(ident_t *loc_ref, kmp_int32 gtid,
4278 KMP_EXPORT void __kmpc_omp_task_complete_if0(ident_t *loc_ref, kmp_int32 gtid,
4280 KMP_EXPORT kmp_int32 __kmpc_omp_task_parts(ident_t *loc_ref, kmp_int32 gtid,
4282 KMP_EXPORT kmp_int32 __kmpc_omp_taskwait(ident_t *loc_ref, kmp_int32 gtid);
4283 KMP_EXPORT kmp_int32 __kmpc_omp_taskyield(ident_t *loc_ref, kmp_int32 gtid,
4287 void __kmpc_omp_task_begin(ident_t *loc_ref, kmp_int32 gtid, kmp_task_t *task);
4288 void __kmpc_omp_task_complete(ident_t *loc_ref, kmp_int32 gtid,
4294 KMP_EXPORT void __kmpc_taskgroup(ident_t *loc, int gtid);
4295 KMP_EXPORT void __kmpc_end_taskgroup(ident_t *loc, int gtid);
4298 ident_t *loc_ref, kmp_int32 gtid, kmp_task_t *new_task, kmp_int32 ndeps,
4306 KMP_EXPORT void __kmpc_omp_wait_deps(ident_t *loc_ref, kmp_int32 gtid,
4313 KMP_EXPORT void __kmpc_omp_taskwait_deps_51(ident_t *loc_ref, kmp_int32 gtid,
4320 extern kmp_int32 __kmp_omp_task(kmp_int32 gtid, kmp_task_t *new_task,
4323 KMP_EXPORT kmp_int32 __kmpc_cancel(ident_t *loc_ref, kmp_int32 gtid,
4325 KMP_EXPORT kmp_int32 __kmpc_cancellationpoint(ident_t *loc_ref, kmp_int32 gtid,
4327 KMP_EXPORT kmp_int32 __kmpc_cancel_barrier(ident_t *loc_ref, kmp_int32 gtid);
4330 KMP_EXPORT void __kmpc_proxy_task_completed(kmp_int32 gtid, kmp_task_t *ptask);
4332 KMP_EXPORT void __kmpc_taskloop(ident_t *loc, kmp_int32 gtid, kmp_task_t *task,
4337 KMP_EXPORT void __kmpc_taskloop_5(ident_t *loc, kmp_int32 gtid,
4343 KMP_EXPORT void *__kmpc_task_reduction_init(int gtid, int num_data, void *data);
4344 KMP_EXPORT void *__kmpc_taskred_init(int gtid, int num_data, void *data);
4345 KMP_EXPORT void *__kmpc_task_reduction_get_th_data(int gtid, void *tg, void *d);
4346 KMP_EXPORT void *__kmpc_task_reduction_modifier_init(ident_t *loc, int gtid,
4349 KMP_EXPORT void *__kmpc_taskred_modifier_init(ident_t *loc, int gtid, int is_ws,
4351 KMP_EXPORT void __kmpc_task_reduction_modifier_fini(ident_t *loc, int gtid,
4354 ident_t *loc_ref, kmp_int32 gtid, kmp_task_t *new_task, kmp_int32 naffins,
4362 KMP_EXPORT void **__kmpc_omp_get_target_async_handle_ptr(kmp_int32 gtid);
4363 KMP_EXPORT bool __kmpc_omp_has_task_team(kmp_int32 gtid);
4366 KMP_EXPORT void __kmpc_init_lock(ident_t *loc, kmp_int32 gtid,
4368 KMP_EXPORT void __kmpc_init_nest_lock(ident_t *loc, kmp_int32 gtid,
4370 KMP_EXPORT void __kmpc_destroy_lock(ident_t *loc, kmp_int32 gtid,
4372 KMP_EXPORT void __kmpc_destroy_nest_lock(ident_t *loc, kmp_int32 gtid,
4374 KMP_EXPORT void __kmpc_set_lock(ident_t *loc, kmp_int32 gtid, void **user_lock);
4375 KMP_EXPORT void __kmpc_set_nest_lock(ident_t *loc, kmp_int32 gtid,
4377 KMP_EXPORT void __kmpc_unset_lock(ident_t *loc, kmp_int32 gtid,
4379 KMP_EXPORT void __kmpc_unset_nest_lock(ident_t *loc, kmp_int32 gtid,
4381 KMP_EXPORT int __kmpc_test_lock(ident_t *loc, kmp_int32 gtid, void **user_lock);
4382 KMP_EXPORT int __kmpc_test_nest_lock(ident_t *loc, kmp_int32 gtid,
4385 KMP_EXPORT void __kmpc_init_lock_with_hint(ident_t *loc, kmp_int32 gtid,
4387 KMP_EXPORT void __kmpc_init_nest_lock_with_hint(ident_t *loc, kmp_int32 gtid,
4399 KMP_EXPORT kmp_int32 __kmpc_start_record_task(ident_t *loc, kmp_int32 gtid,
4402 KMP_EXPORT void __kmpc_end_record_task(ident_t *loc, kmp_int32 gtid,
4472 KMP_EXPORT void __kmpc_doacross_init(ident_t *loc, kmp_int32 gtid,
4475 KMP_EXPORT void __kmpc_doacross_wait(ident_t *loc, kmp_int32 gtid,
4477 KMP_EXPORT void __kmpc_doacross_post(ident_t *loc, kmp_int32 gtid,
4479 KMP_EXPORT void __kmpc_doacross_fini(ident_t *loc, kmp_int32 gtid);
4487 void kmp_threadprivate_insert_private_data(int gtid, void *pc_addr,
4489 struct private_common *kmp_threadprivate_insert(int gtid, void *pc_addr,
4603 #define KMP_HIDDEN_HELPER_THREAD(gtid) \ argument
4604 ((gtid) >= 1 && (gtid) <= __kmp_hidden_helper_threads_num)
4606 #define KMP_HIDDEN_HELPER_WORKER_THREAD(gtid) \ argument
4607 ((gtid) > 1 && (gtid) <= __kmp_hidden_helper_threads_num)
4609 #define KMP_HIDDEN_HELPER_MAIN_THREAD(gtid) \ argument
4610 ((gtid) == 1 && (gtid) <= __kmp_hidden_helper_threads_num)
4617 #define KMP_GTID_TO_SHADOW_GTID(gtid) \ argument
4618 ((gtid) % (__kmp_hidden_helper_threads_num - 1) + 2)
4623 static inline int __kmp_adjust_gtid_for_hidden_helpers(int gtid) { in __kmp_adjust_gtid_for_hidden_helpers() argument
4624 int adjusted_gtid = gtid; in __kmp_adjust_gtid_for_hidden_helpers()
4625 if (__kmp_hidden_helper_threads_num > 0 && gtid > 0 && in __kmp_adjust_gtid_for_hidden_helpers()
4626 gtid - __kmp_hidden_helper_threads_num >= 0) { in __kmp_adjust_gtid_for_hidden_helpers()
4640 KMP_EXPORT void __kmpc_scope(ident_t *loc, kmp_int32 gtid, void *reserved);
4641 KMP_EXPORT void __kmpc_end_scope(ident_t *loc, kmp_int32 gtid, void *reserved);
4674 int __kmp_execute_tasks_32(kmp_info_t *thread, kmp_int32 gtid,
4682 int __kmp_execute_tasks_64(kmp_info_t *thread, kmp_int32 gtid,
4690 int __kmp_atomic_execute_tasks_64(kmp_info_t *thread, kmp_int32 gtid,
4697 int __kmp_execute_tasks_oncore(kmp_info_t *thread, kmp_int32 gtid,