/* * ompt-specific.h - header of OMPT internal functions implementation */ //===----------------------------------------------------------------------===// // // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. // See https://llvm.org/LICENSE.txt for license information. // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception // //===----------------------------------------------------------------------===// #ifndef OMPT_SPECIFIC_H #define OMPT_SPECIFIC_H #include "kmp.h" #if OMPT_SUPPORT /***************************************************************************** * forward declarations ****************************************************************************/ /// Entrypoint used by libomptarget to register callbacks in libomp, if not /// done already void __ompt_force_initialization(); void __ompt_team_assign_id(kmp_team_t *team, ompt_data_t ompt_pid); void __ompt_thread_assign_wait_id(void *variable); void __ompt_lw_taskteam_init(ompt_lw_taskteam_t *lwt, kmp_info_t *thr, int gtid, ompt_data_t *ompt_pid, void *codeptr); void __ompt_lw_taskteam_link(ompt_lw_taskteam_t *lwt, kmp_info_t *thr, int on_heap, bool always = false); void __ompt_lw_taskteam_unlink(kmp_info_t *thr); ompt_team_info_t *__ompt_get_teaminfo(int depth, int *size); ompt_data_t *__ompt_get_task_data(); ompt_data_t *__ompt_get_target_task_data(); ompt_task_info_t *__ompt_get_task_info_object(int depth); int __ompt_get_parallel_info_internal(int ancestor_level, ompt_data_t **parallel_data, int *team_size); int __ompt_get_task_info_internal(int ancestor_level, int *type, ompt_data_t **task_data, ompt_frame_t **task_frame, ompt_data_t **parallel_data, int *thread_num); ompt_data_t *__ompt_get_thread_data_internal(); /* * Unused currently static uint64_t __ompt_get_get_unique_id_internal(); */ ompt_sync_region_t __ompt_get_barrier_kind(enum barrier_type, kmp_info_t *); /***************************************************************************** * macros ****************************************************************************/ #define OMPT_CUR_TASK_INFO(thr) (&((thr)->th.th_current_task->ompt_task_info)) #define OMPT_CUR_TASK_DATA(thr) \ (&((thr)->th.th_current_task->ompt_task_info.task_data)) #define OMPT_CUR_TEAM_INFO(thr) (&((thr)->th.th_team->t.ompt_team_info)) #define OMPT_CUR_TEAM_DATA(thr) \ (&((thr)->th.th_team->t.ompt_team_info.parallel_data)) #define OMPT_HAVE_WEAK_ATTRIBUTE KMP_HAVE_WEAK_ATTRIBUTE #define OMPT_HAVE_PSAPI KMP_HAVE_PSAPI #define OMPT_STR_MATCH(haystack, needle) __kmp_str_match(haystack, 0, needle) inline void *__ompt_load_return_address(int gtid) { kmp_info_t *thr = __kmp_threads[gtid]; void *return_address = thr->th.ompt_thread_info.return_address; thr->th.ompt_thread_info.return_address = NULL; return return_address; } /*#define OMPT_STORE_RETURN_ADDRESS(gtid) \ if (ompt_enabled.enabled && gtid >= 0 && __kmp_threads[gtid] && \ !__kmp_threads[gtid]->th.ompt_thread_info.return_address) \ __kmp_threads[gtid]->th.ompt_thread_info.return_address = \ __builtin_return_address(0)*/ #define OMPT_STORE_RETURN_ADDRESS(gtid) \ OmptReturnAddressGuard ReturnAddressGuard{gtid, __builtin_return_address(0)}; #define OMPT_LOAD_RETURN_ADDRESS(gtid) __ompt_load_return_address(gtid) #define OMPT_LOAD_OR_GET_RETURN_ADDRESS(gtid) \ ((ompt_enabled.enabled && gtid >= 0 && __kmp_threads[gtid] && \ __kmp_threads[gtid]->th.ompt_thread_info.return_address) \ ? __ompt_load_return_address(gtid) \ : __builtin_return_address(0)) #define OMPT_GET_DISPATCH_CHUNK(chunk, lb, ub, incr) \ do { \ if (incr > 0) { \ chunk.start = static_cast(lb); \ chunk.iterations = static_cast(((ub) - (lb)) / (incr) + 1); \ } else { \ chunk.start = static_cast(ub); \ chunk.iterations = static_cast(((lb) - (ub)) / -(incr) + 1); \ } \ } while (0) //****************************************************************************** // inline functions //****************************************************************************** inline kmp_info_t *ompt_get_thread_gtid(int gtid) { return (gtid >= 0) ? __kmp_thread_from_gtid(gtid) : NULL; } inline kmp_info_t *ompt_get_thread() { int gtid = __kmp_get_gtid(); return ompt_get_thread_gtid(gtid); } inline void ompt_set_thread_state(kmp_info_t *thread, ompt_state_t state) { if (thread) thread->th.ompt_thread_info.state = state; } inline const char *ompt_get_runtime_version() { return &__kmp_version_lib_ver[KMP_VERSION_MAGIC_LEN]; } inline ompt_work_t ompt_get_work_schedule(enum sched_type schedule) { switch (SCHEDULE_WITHOUT_MODIFIERS(schedule)) { case kmp_sch_static_chunked: case kmp_sch_static_balanced: case kmp_sch_static_greedy: return ompt_work_loop_static; case kmp_sch_dynamic_chunked: case kmp_sch_static_steal: return ompt_work_loop_dynamic; case kmp_sch_guided_iterative_chunked: case kmp_sch_guided_analytical_chunked: case kmp_sch_guided_chunked: case kmp_sch_guided_simd: return ompt_work_loop_guided; default: return ompt_work_loop_other; } } class OmptReturnAddressGuard { private: bool SetAddress{false}; int Gtid; public: OmptReturnAddressGuard(int Gtid, void *ReturnAddress) : Gtid(Gtid) { if (ompt_enabled.enabled && Gtid >= 0 && __kmp_threads[Gtid] && !__kmp_threads[Gtid]->th.ompt_thread_info.return_address) { SetAddress = true; __kmp_threads[Gtid]->th.ompt_thread_info.return_address = ReturnAddress; } } ~OmptReturnAddressGuard() { if (SetAddress) __kmp_threads[Gtid]->th.ompt_thread_info.return_address = NULL; } }; #endif // OMPT_SUPPORT // macros providing the OMPT callbacks for reduction clause #if OMPT_SUPPORT && OMPT_OPTIONAL #define OMPT_REDUCTION_DECL(this_thr, gtid) \ ompt_data_t *my_task_data = OMPT_CUR_TASK_DATA(this_thr); \ ompt_data_t *my_parallel_data = OMPT_CUR_TEAM_DATA(this_thr); \ void *return_address = OMPT_LOAD_RETURN_ADDRESS(gtid); #define OMPT_REDUCTION_BEGIN \ if (ompt_enabled.enabled && ompt_enabled.ompt_callback_reduction) { \ ompt_callbacks.ompt_callback(ompt_callback_reduction)( \ ompt_sync_region_reduction, ompt_scope_begin, my_parallel_data, \ my_task_data, return_address); \ } #define OMPT_REDUCTION_END \ if (ompt_enabled.enabled && ompt_enabled.ompt_callback_reduction) { \ ompt_callbacks.ompt_callback(ompt_callback_reduction)( \ ompt_sync_region_reduction, ompt_scope_end, my_parallel_data, \ my_task_data, return_address); \ } #else // OMPT_SUPPORT && OMPT_OPTIONAL #define OMPT_REDUCTION_DECL(this_thr, gtid) #define OMPT_REDUCTION_BEGIN #define OMPT_REDUCTION_END #endif // ! OMPT_SUPPORT && OMPT_OPTIONAL #endif