1 /* 2 * ompt-specific.h - header of OMPT internal functions implementation 3 */ 4 5 //===----------------------------------------------------------------------===// 6 // 7 // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. 8 // See https://llvm.org/LICENSE.txt for license information. 9 // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception 10 // 11 //===----------------------------------------------------------------------===// 12 13 #ifndef OMPT_SPECIFIC_H 14 #define OMPT_SPECIFIC_H 15 16 #include "kmp.h" 17 18 #if OMPT_SUPPORT 19 /***************************************************************************** 20 * forward declarations 21 ****************************************************************************/ 22 23 void __ompt_team_assign_id(kmp_team_t *team, ompt_data_t ompt_pid); 24 void __ompt_thread_assign_wait_id(void *variable); 25 26 void __ompt_lw_taskteam_init(ompt_lw_taskteam_t *lwt, kmp_info_t *thr, int gtid, 27 ompt_data_t *ompt_pid, void *codeptr); 28 29 void __ompt_lw_taskteam_link(ompt_lw_taskteam_t *lwt, kmp_info_t *thr, 30 int on_heap, bool always = false); 31 32 void __ompt_lw_taskteam_unlink(kmp_info_t *thr); 33 34 ompt_team_info_t *__ompt_get_teaminfo(int depth, int *size); 35 36 ompt_task_info_t *__ompt_get_task_info_object(int depth); 37 38 int __ompt_get_parallel_info_internal(int ancestor_level, 39 ompt_data_t **parallel_data, 40 int *team_size); 41 42 int __ompt_get_task_info_internal(int ancestor_level, int *type, 43 ompt_data_t **task_data, 44 ompt_frame_t **task_frame, 45 ompt_data_t **parallel_data, int *thread_num); 46 47 ompt_data_t *__ompt_get_thread_data_internal(); 48 49 /* 50 * Unused currently 51 static uint64_t __ompt_get_get_unique_id_internal(); 52 */ 53 54 ompt_sync_region_t __ompt_get_barrier_kind(enum barrier_type, kmp_info_t *); 55 56 /***************************************************************************** 57 * macros 58 ****************************************************************************/ 59 60 #define OMPT_CUR_TASK_INFO(thr) (&(thr->th.th_current_task->ompt_task_info)) 61 #define OMPT_CUR_TASK_DATA(thr) \ 62 (&(thr->th.th_current_task->ompt_task_info.task_data)) 63 #define OMPT_CUR_TEAM_INFO(thr) (&(thr->th.th_team->t.ompt_team_info)) 64 #define OMPT_CUR_TEAM_DATA(thr) \ 65 (&(thr->th.th_team->t.ompt_team_info.parallel_data)) 66 67 #define OMPT_HAVE_WEAK_ATTRIBUTE KMP_HAVE_WEAK_ATTRIBUTE 68 #define OMPT_HAVE_PSAPI KMP_HAVE_PSAPI 69 #define OMPT_STR_MATCH(haystack, needle) __kmp_str_match(haystack, 0, needle) 70 71 inline void *__ompt_load_return_address(int gtid) { 72 kmp_info_t *thr = __kmp_threads[gtid]; 73 void *return_address = thr->th.ompt_thread_info.return_address; 74 thr->th.ompt_thread_info.return_address = NULL; 75 return return_address; 76 } 77 78 /*#define OMPT_STORE_RETURN_ADDRESS(gtid) \ 79 if (ompt_enabled.enabled && gtid >= 0 && __kmp_threads[gtid] && \ 80 !__kmp_threads[gtid]->th.ompt_thread_info.return_address) \ 81 __kmp_threads[gtid]->th.ompt_thread_info.return_address = \ 82 __builtin_return_address(0)*/ 83 #define OMPT_STORE_RETURN_ADDRESS(gtid) \ 84 OmptReturnAddressGuard ReturnAddressGuard{gtid, __builtin_return_address(0)}; 85 #define OMPT_LOAD_RETURN_ADDRESS(gtid) __ompt_load_return_address(gtid) 86 #define OMPT_LOAD_OR_GET_RETURN_ADDRESS(gtid) \ 87 ((ompt_enabled.enabled && gtid >= 0 && __kmp_threads[gtid] && \ 88 __kmp_threads[gtid]->th.ompt_thread_info.return_address) \ 89 ? __ompt_load_return_address(gtid) \ 90 : __builtin_return_address(0)) 91 92 //****************************************************************************** 93 // inline functions 94 //****************************************************************************** 95 96 inline kmp_info_t *ompt_get_thread_gtid(int gtid) { 97 return (gtid >= 0) ? __kmp_thread_from_gtid(gtid) : NULL; 98 } 99 100 inline kmp_info_t *ompt_get_thread() { 101 int gtid = __kmp_get_gtid(); 102 return ompt_get_thread_gtid(gtid); 103 } 104 105 inline void ompt_set_thread_state(kmp_info_t *thread, ompt_state_t state) { 106 if (thread) 107 thread->th.ompt_thread_info.state = state; 108 } 109 110 inline const char *ompt_get_runtime_version() { 111 return &__kmp_version_lib_ver[KMP_VERSION_MAGIC_LEN]; 112 } 113 114 class OmptReturnAddressGuard { 115 private: 116 bool SetAddress{false}; 117 int Gtid; 118 119 public: 120 OmptReturnAddressGuard(int Gtid, void *ReturnAddress) : Gtid(Gtid) { 121 if (ompt_enabled.enabled && Gtid >= 0 && __kmp_threads[Gtid] && 122 !__kmp_threads[Gtid]->th.ompt_thread_info.return_address) { 123 SetAddress = true; 124 __kmp_threads[Gtid]->th.ompt_thread_info.return_address = ReturnAddress; 125 } 126 } 127 ~OmptReturnAddressGuard() { 128 if (SetAddress) 129 __kmp_threads[Gtid]->th.ompt_thread_info.return_address = NULL; 130 } 131 }; 132 133 #endif // OMPT_SUPPORT 134 135 // macros providing the OMPT callbacks for reduction clause 136 #if OMPT_SUPPORT && OMPT_OPTIONAL 137 #define OMPT_REDUCTION_DECL(this_thr, gtid) \ 138 ompt_data_t *my_task_data = OMPT_CUR_TASK_DATA(this_thr); \ 139 ompt_data_t *my_parallel_data = OMPT_CUR_TEAM_DATA(this_thr); \ 140 void *return_address = OMPT_LOAD_RETURN_ADDRESS(gtid); 141 #define OMPT_REDUCTION_BEGIN \ 142 if (ompt_enabled.enabled && ompt_enabled.ompt_callback_reduction) { \ 143 ompt_callbacks.ompt_callback(ompt_callback_reduction)( \ 144 ompt_sync_region_reduction, ompt_scope_begin, my_parallel_data, \ 145 my_task_data, return_address); \ 146 } 147 #define OMPT_REDUCTION_END \ 148 if (ompt_enabled.enabled && ompt_enabled.ompt_callback_reduction) { \ 149 ompt_callbacks.ompt_callback(ompt_callback_reduction)( \ 150 ompt_sync_region_reduction, ompt_scope_end, my_parallel_data, \ 151 my_task_data, return_address); \ 152 } 153 #else // OMPT_SUPPORT && OMPT_OPTIONAL 154 #define OMPT_REDUCTION_DECL(this_thr, gtid) 155 #define OMPT_REDUCTION_BEGIN 156 #define OMPT_REDUCTION_END 157 #endif // ! OMPT_SUPPORT && OMPT_OPTIONAL 158 159 #endif 160