/freebsd/contrib/llvm-project/openmp/runtime/src/ |
H A D | z_Windows_NT-586_util.cpp | 47 kmp_uint32 __kmp_test_then_or32(volatile kmp_uint32 *p, kmp_uint32 d) { in __kmp_test_then_or32() 48 kmp_uint32 old_value, new_value; in __kmp_test_then_or32() 62 kmp_uint32 __kmp_test_then_and32(volatile kmp_uint32 *p, kmp_uint32 d) { in __kmp_test_then_and32() 63 kmp_uint32 old_value, new_value; in __kmp_test_then_and32()
|
H A D | kmp_os.h | 136 typedef unsigned int kmp_uint32; typedef 172 typedef unsigned int kmp_uint32; typedef 215 typedef kmp_uint32 kmp_uint; 514 extern kmp_uint32 __kmp_test_then_or32(volatile kmp_uint32 *p, kmp_uint32 v); 515 extern kmp_uint32 __kmp_test_then_and32(volatile kmp_uint32 *p, kmp_uint32 v); 784 __sync_fetch_and_or((volatile kmp_uint32 *)(p), (kmp_uint32)(v)) 786 __sync_fetch_and_and((volatile kmp_uint32 *)(p), (kmp_uint32)(v)) 814 __sync_bool_compare_and_swap((volatile kmp_uint32 *)(p), (kmp_uint32)(cv), \ 815 (kmp_uint32)(sv)) 817 __sync_bool_compare_and_swap((volatile kmp_uint32 *)(p), (kmp_uint32)(cv), \ [all …]
|
H A D | kmp_lock.h | 68 typedef kmp_uint32 kmp_lock_flags_t; 73 typedef kmp_uint32 kmp_lock_index_t; 324 kmp_uint32 successfulSpeculations; 325 kmp_uint32 hardFailedSpeculations; 326 kmp_uint32 softFailedSpeculations; 327 kmp_uint32 nonSpeculativeAcquires; 328 kmp_uint32 nonSpeculativeAcquireAttempts; 329 kmp_uint32 lemmingYields; 344 kmp_uint32 volatile badness; 345 kmp_uint32 volatil [all...] |
H A D | kmp.h | 1384 kmp_uint32 eax; 1385 kmp_uint32 ebx; 1386 kmp_uint32 ecx; 1387 kmp_uint32 edx; 1450 static inline void __kmp_load_mxcsr(const kmp_uint32 *p) { _mm_setcsr(*p); } in __kmp_load_mxcsr() 1451 static inline void __kmp_store_mxcsr(kmp_uint32 *p) { *p = _mm_getcsr(); } in __kmp_store_mxcsr() 1453 static inline void __kmp_load_mxcsr(const kmp_uint32 *p) {} in __kmp_load_mxcsr() 1454 static inline void __kmp_store_mxcsr(kmp_uint32 *p) { *p = 0; } in __kmp_store_mxcsr() 1462 static inline void __kmp_load_mxcsr(const kmp_uint32 *p) { _mm_setcsr(*p); } in __kmp_load_mxcsr() 1463 static inline void __kmp_store_mxcsr(kmp_uint32 *p) { *p = _mm_getcsr(); } in __kmp_store_mxcsr() [all …]
|
H A D | kmp_dispatch.cpp | 444 kmp_uint32 old = UNUSED; in __kmp_dispatch_init_algorithm() 919 __kmp_dispatch_init_hier_runtime<kmp_uint32>(ident_t *loc, kmp_uint32 lb, in __kmp_dispatch_init_hier_runtime() 920 kmp_uint32 ub, kmp_int32 st) { in __kmp_dispatch_init_hier_runtime() 921 __kmp_dispatch_init_hierarchy<kmp_uint32>( in __kmp_dispatch_init_hier_runtime() 970 kmp_uint32 my_buffer_index; in __kmp_dispatch_init() 1075 __kmp_wait<kmp_uint32>(&sh->buffer_index, my_buffer_index, in __kmp_dispatch_init() 1076 __kmp_eq<kmp_uint32> USE_ITT_BUILD_ARG(NULL)); in __kmp_dispatch_init() 1420 kmp_uint32 old = UNUSED; in __kmp_dispatch_next_algorithm() 1563 kmp_uint32 old = UNUSED; in __kmp_dispatch_next_algorithm() 2472 kmp_uint32 my_buffer_index; in __kmpc_sections_init() [all …]
|
H A D | kmp_collapse.cpp | 32 kmp_uint32 __kmp_abs(const kmp_uint32 val) { return val; } in __kmp_abs() 123 kmp_canonicalize_one_loop_XX<kmp_uint32>( in kmp_canonicalize_loop_nest() 125 /*in/out*/ (bounds_infoXX_template<kmp_uint32> *)(bounds)); in kmp_canonicalize_loop_nest() 199 trip_count = kmp_calculate_trip_count_XX<kmp_uint32>( in kmp_calculate_trip_count() 200 /*in/out*/ (bounds_infoXX_template<kmp_uint32> *)(bounds)); in kmp_calculate_trip_count() 242 res = static_cast<kmp_uint64>(static_cast<kmp_uint32>(original_iv)); in kmp_fix_iv() 285 res = static_cast<kmp_uint32>(original_iv1) == in kmp_ivs_eq() 286 static_cast<kmp_uint32>(original_iv2); in kmp_ivs_eq() 370 return kmp_calc_one_iv_XX<kmp_uint32>( in kmp_calc_one_iv() 371 (bounds_infoXX_template<kmp_uint32> *)(bounds), in kmp_calc_one_iv() [all …]
|
H A D | kmp_affinity.h | 1258 static const kmp_uint32 maxLeaves = 4; 1259 static const kmp_uint32 minBranch = 4; 1265 kmp_uint32 maxLevels; 1271 kmp_uint32 depth; 1272 kmp_uint32 base_num_threads; 1282 kmp_uint32 *numPerLevel; 1283 kmp_uint32 *skipPerLevel; 1321 (kmp_uint32 *)__kmp_allocate(maxLevels * 2 * sizeof(kmp_uint32)); in init() 1323 for (kmp_uint32 i = 0; i < maxLevels; in init() 1345 kmp_uint32 branch = minBranch; in init() [all …]
|
H A D | kmp_dispatch.h | 142 std::atomic<kmp_uint32> steal_flag; // static_steal only, state of a buffer 143 kmp_uint32 ordered_bumped; 145 kmp_uint32 type_size; in get_parent() 178 volatile kmp_uint32 buffer_index; 180 kmp_uint32 *doacross_flags; // array of iteration flags (0/1) 267 template <typename T> kmp_uint32 __kmp_ge(T value, T checker) { 270 template <typename T> kmp_uint32 __kmp_eq(T value, T checker) { 295 kmp_uint32 (*pred)(UT, UT) USE_ITT_BUILD_ARG(void *obj)) { in __kmp_wait() 299 kmp_uint32 spins; in __kmp_wait() 300 kmp_uint32 (* in __kmp_wait() [all...] |
H A D | kmp_sched.cpp | 96 kmp_uint32 tid; in __kmp_for_static_init() 97 kmp_uint32 nth; in __kmp_for_static_init() 495 kmp_uint32 tid; in __kmp_dist_for_static_init() 496 kmp_uint32 nth; in __kmp_dist_for_static_init() 497 kmp_uint32 team_id; in __kmp_dist_for_static_init() 498 kmp_uint32 nteams; in __kmp_dist_for_static_init() 547 KMP_DEBUG_ASSERT(nteams == (kmp_uint32)team->t.t_parent->t.t_nproc); in __kmp_dist_for_static_init() 764 kmp_uint32 team_id; in __kmp_team_static_init() 765 kmp_uint32 nteams; in __kmp_team_static_init() 815 KMP_DEBUG_ASSERT(nteams == (kmp_uint32)team->t.t_parent->t.t_nproc); in __kmp_team_static_init() [all …]
|
H A D | kmp_global.cpp | 80 kmp_uint32 __kmp_barrier_gather_bb_dflt = 2; 82 kmp_uint32 __kmp_barrier_release_bb_dflt = 2; 90 kmp_uint32 __kmp_barrier_gather_branch_bits[bs_last_barrier] = {0}; 91 kmp_uint32 __kmp_barrier_release_branch_bits[bs_last_barrier] = {0}; 193 kmp_uint32 __kmp_init_mxcsr = 0; 439 kmp_uint32 __kmp_yield_init = KMP_INIT_WAIT; 440 kmp_uint32 __kmp_yield_next = KMP_NEXT_WAIT;
|
H A D | kmp_wait_release.h | 44 typedef kmp_uint32 flag_t; 108 kmp_uint32 num_waiting_threads; /**< Num threads sleeping on this thread. */ 123 kmp_info_t *get_waiter(kmp_uint32 i) { 128 kmp_uint32 get_num_waiters() { return num_waiting_threads; } 373 kmp_uint32 spins; 380 kmp_uint32 hibernate; 826 class kmp_flag_32 : public kmp_flag_atomic<kmp_uint32, flag32, Sleepable> { 828 kmp_flag_32(std::atomic<kmp_uint32> *p) 829 : kmp_flag_atomic<kmp_uint32, flag32, Sleepable>(p) {} 830 kmp_flag_32(std::atomic<kmp_uint32> *p, kmp_info_t *thr) [all …]
|
H A D | kmp_barrier.cpp | 234 kmp_uint32 nproc; in __kmp_dist_barrier_gather() 667 kmp_uint32 nproc = this_thr->th.th_team_nproc; in __kmp_linear_barrier_release_template() 801 kmp_uint32 nproc = this_thr->th.th_team_nproc; in __kmp_tree_barrier_gather() 802 kmp_uint32 branch_bits = __kmp_barrier_gather_branch_bits[bt]; in __kmp_tree_barrier_gather() 803 kmp_uint32 branch_factor = 1 << branch_bits; in __kmp_tree_barrier_gather() 804 kmp_uint32 child; in __kmp_tree_barrier_gather() 805 kmp_uint32 child_tid; in __kmp_tree_barrier_gather() 906 kmp_uint32 nproc; in __kmp_tree_barrier_release() 907 kmp_uint32 branch_bits = __kmp_barrier_release_branch_bits[bt]; in __kmp_tree_barrier_release() 908 kmp_uint32 branch_factor = 1 << branch_bits; in __kmp_tree_barrier_release() [all …]
|
H A D | kmp_atomic.h | 509 void __kmpc_atomic_fixed4u_div(ident_t *id_ref, int gtid, kmp_uint32 *lhs, 510 kmp_uint32 rhs); 519 void __kmpc_atomic_fixed4u_shr(ident_t *id_ref, int gtid, kmp_uint32 *lhs, 520 kmp_uint32 rhs); 734 void __kmpc_atomic_fixed4u_div_rev(ident_t *id_ref, int gtid, kmp_uint32 *lhs, 735 kmp_uint32 rhs); 740 void __kmpc_atomic_fixed4u_shr_rev(ident_t *id_ref, int gtid, kmp_uint32 *lhs, 741 kmp_uint32 rhs); 873 void __kmpc_atomic_fixed4u_add_fp(ident_t *id_ref, int gtid, kmp_uint32 *lhs, 877 void __kmpc_atomic_fixed4u_sub_fp(ident_t *id_ref, int gtid, kmp_uint32 *lhs, [all …]
|
H A D | kmp_error.h | 35 kmp_user_lock_p name, kmp_uint32); 44 kmp_user_lock_p name, kmp_uint32);
|
H A D | kmp_lock.cpp | 45 kmp_uint32 x, y; in __kmp_validate_locks() 48 x = ~((kmp_uint32)0) - 2; in __kmp_validate_locks() 52 kmp_uint32 z = (x - y); in __kmp_validate_locks() 83 kmp_uint32 curr = KMP_LOCK_STRIP(lck->lk.poll); in __kmp_acquire_tas_lock_timed_template() 98 kmp_uint32 spins; in __kmp_acquire_tas_lock_timed_template() 326 kmp_uint32 curr = KMP_LOCK_STRIP(TCR_4(lck->lk.poll)); in __kmp_acquire_futex_lock_timed_template() 618 static kmp_uint32 __kmp_bakery_check(void *now_serving, kmp_uint32 my_ticket) { in __kmp_bakery_check() 626 kmp_uint32 my_ticket = std::atomic_fetch_add_explicit( in __kmp_acquire_ticket_lock_timed_template() 675 kmp_uint32 my_ticke in __kmp_test_ticket_lock() [all...] |
H A D | kmp_atomic.cpp | 1031 ATOMIC_CMPXCHG(fixed4u, div, kmp_uint32, 32, /, 4i, 3, 1041 ATOMIC_CMPXCHG(fixed4u, shr, kmp_uint32, 32, >>, 4i, 3, 1523 ATOMIC_CMPXCHG_REV(fixed4u, div, kmp_uint32, 32, /, 4i, 1529 ATOMIC_CMPXCHG_REV(fixed4u, shr, kmp_uint32, 32, >>, 4i, 1756 ATOMIC_CMPXCHG_MIX(fixed4u, kmp_uint32, add, 32, +, fp, _Quad, 4i, 3, 1760 ATOMIC_CMPXCHG_MIX(fixed4u, kmp_uint32, sub, 32, -, fp, _Quad, 4i, 3, 1764 ATOMIC_CMPXCHG_MIX(fixed4u, kmp_uint32, mul, 32, *, fp, _Quad, 4i, 3, 1768 ATOMIC_CMPXCHG_MIX(fixed4u, kmp_uint32, div, 32, /, fp, _Quad, 4i, 3, 1837 ATOMIC_CMPXCHG_REV_MIX(fixed4u, kmp_uint32, sub_rev, 32, -, fp, _Quad, 4i, 3, 1841 ATOMIC_CMPXCHG_REV_MIX(fixed4u, kmp_uint32, div_rev, 32, /, fp, _Quad, 4i, 3, [all …]
|
H A D | kmp_csupport.cpp | 261 kmp_uint32 list_length, in __kmpc_push_num_threads_list() 267 for (kmp_uint32 i = 1; i < list_length; ++i) in __kmpc_push_num_threads_list() 277 kmp_uint32 list_length, in __kmpc_push_num_threads_list_strict() 1178 kmp_uint32 spins; \ 4152 kmp_uint32 *flags; in __kmpc_doacross_init() 4228 __kmp_wait_4((volatile kmp_uint32 *)&sh_buf->doacross_buf_idx, idx, in __kmpc_doacross_init() 4235 flags = (kmp_uint32 *)KMP_COMPARE_AND_STORE_RET32( in __kmpc_doacross_init() 4238 flags = (kmp_uint32 *)KMP_COMPARE_AND_STORE_RET64( in __kmpc_doacross_init() 4245 flags = (kmp_uint32 *)__kmp_thread_calloc(th, size, 1); in __kmpc_doacross_init() 4248 } else if (flags == (kmp_uint32 *)1) { in __kmpc_doacross_init() [all …]
|
H A D | z_Linux_util.cpp | 354 kmp_uint32 __kmp_test_then_or32(volatile kmp_uint32 *p, kmp_uint32 d) { in __kmp_test_then_or32() 355 kmp_uint32 old_value, new_value; in __kmp_test_then_or32() 368 kmp_uint32 __kmp_test_then_and32(volatile kmp_uint32 *p, kmp_uint32 d) { in __kmp_test_then_and32() 369 kmp_uint32 old_value, new_value; in __kmp_test_then_and32() 1016 KMP_DEBUG_ASSERT(sizeof(kmp_uint32) == in __kmp_create_monitor() 1018 __kmp_wait_4((kmp_uint32 volatile *)&__kmp_global.g.g_time.dt.t_value, -1, in __kmp_create_monitor()
|
H A D | kmp_barrier.h | 61 kmp_uint32 volatile KMP_FOURLINE_ALIGN_CACHE stillNeed;
|
H A D | kmp_tasking.cpp | 2294 RCAST(std::atomic<kmp_uint32> *, 2529 kmp_uint32 nth = thread->th.th_team_nproc; 2969 RCAST(std::atomic<kmp_uint32> *, &(taskgroup->count)), 0U); 3149 (kmp_uint32)((target + 1) & TASK_DEQUE_MASK(thread_data->td))); 3165 kmp_uint32 tail; 3330 (kmp_uint32)((target + 1) & TASK_DEQUE_MASK(victim_td->td))); 4006 CCAST(kmp_uint32 *, &task_team->tt.tt_active), 4108 kmp_uint32 spins; 4297 RCAST(std::atomic<kmp_uint32> *, 4325 std::atomic<kmp_uint32> *spin = RCAST( [all …]
|
H A D | kmp_error.cpp | 250 …eck_sync( int gtid, enum cons_type ct, ident_t const * ident, kmp_user_lock_p lck, kmp_uint32 seq ) 336 …ush_sync( int gtid, enum cons_type ct, ident_t const * ident, kmp_user_lock_p lck, kmp_uint32 seq )
|
H A D | kmp_utility.cpp | 83 kmp_uint32 t, data[4]; in __kmp_query_cpuid()
|
H A D | kmp_affinity.cpp | 78 void __kmp_get_hierarchy(kmp_uint32 nproc, kmp_bstate_t *thr_bar) { in __kmp_get_hierarchy() 79 kmp_uint32 depth; in __kmp_get_hierarchy() 2078 template <kmp_uint32 LSB, kmp_uint32 MSB> 2079 static inline unsigned __kmp_extract_bits(kmp_uint32 v) { in __kmp_extract_bits() 2080 const kmp_uint32 SHIFT_LEFT = sizeof(kmp_uint32) * 8 - 1 - MSB; in __kmp_extract_bits() 2081 const kmp_uint32 SHIFT_RIGHT = LSB; in __kmp_extract_bits() 2082 kmp_uint32 retval = v; in __kmp_extract_bits()
|
H A D | kmp_settings.cpp | 1706 (kmp_uint32)__kmp_str_to_int(value, ','); in __kmp_stg_parse_barrier_branch_bit() 1712 (kmp_uint32)__kmp_str_to_int(comma + 1, 0); in __kmp_stg_parse_barrier_branch_bit() 4845 kmp_uint32 max_backoff = __kmp_spin_backoff_params.max_backoff; in __kmp_stg_parse_spin_backoff_params() 4846 kmp_uint32 min_tick = __kmp_spin_backoff_params.min_tick; in __kmp_stg_parse_spin_backoff_params()
|
H A D | kmp_taskdeps.cpp | 1026 (std::atomic<kmp_uint32> *)&node.dn.npredecessors, 0U); in __kmpc_omp_taskwait_deps_51()
|