Lines Matching refs:kmp_uint32

1384   kmp_uint32 eax;
1385 kmp_uint32 ebx;
1386 kmp_uint32 ecx;
1387 kmp_uint32 edx;
1450 static inline void __kmp_load_mxcsr(const kmp_uint32 *p) { _mm_setcsr(*p); } in __kmp_load_mxcsr()
1451 static inline void __kmp_store_mxcsr(kmp_uint32 *p) { *p = _mm_getcsr(); } in __kmp_store_mxcsr()
1453 static inline void __kmp_load_mxcsr(const kmp_uint32 *p) {} in __kmp_load_mxcsr()
1454 static inline void __kmp_store_mxcsr(kmp_uint32 *p) { *p = 0; } in __kmp_store_mxcsr()
1462 static inline void __kmp_load_mxcsr(const kmp_uint32 *p) { _mm_setcsr(*p); } in __kmp_load_mxcsr()
1463 static inline void __kmp_store_mxcsr(kmp_uint32 *p) { *p = _mm_getcsr(); } in __kmp_store_mxcsr()
1905 kmp_uint32 ordered_lower;
1906 kmp_uint32 ordered_upper;
1922 kmp_uint32 pchunks;
1923 kmp_uint32 num_procs_with_pcore;
1989 kmp_uint32 ordered_lower;
1990 kmp_uint32 ordered_upper;
2025 std::atomic<kmp_uint32> steal_flag; // static_steal only, state of a buffer
2040 volatile kmp_uint32 iteration;
2042 volatile kmp_uint32 ordered_iteration;
2062 volatile kmp_uint32 buffer_index;
2064 volatile kmp_uint32 *doacross_flags; // shared array of iteration flags (0/1)
2087 kmp_uint32 th_disp_index;
2089 volatile kmp_uint32 *th_doacross_flags; // pointer to shared array of flags
2210 kmp_uint32 *skip_per_level;
2211 kmp_uint32 my_level;
2214 kmp_uint32 depth;
2218 kmp_uint32 nproc;
2555 kmp_uint32 id;
2582 kmp_uint32 nelements;
2583 kmp_uint32 nconflicts;
2767 kmp_uint32 td_taskwait_counter;
2816 kmp_uint32 td_deque_head; // Head of deque (will wrap)
2817 kmp_uint32 td_deque_tail; // Tail of deque (will wrap)
2875 volatile kmp_uint32
3019 volatile kmp_uint32 th_spin_here; /* thread-local location for spinning */
3033 kmp_uint32 th_reap_state; // Non-zero indicates thread is not
3040 std::atomic<kmp_uint32> th_used_in_team; // Flag indicating use in team
3094 volatile kmp_uint32 t_value;
3122 sizeof(kmp_int16) + sizeof(kmp_uint32)) % \
3176 kmp_uint32 t_mxcsr;
3218 std::atomic<kmp_uint32> t_copyin_counter;
3390 extern kmp_uint32 __kmp_barrier_gather_bb_dflt;
3391 extern kmp_uint32 __kmp_barrier_release_bb_dflt;
3394 extern kmp_uint32 __kmp_barrier_gather_branch_bits[bs_last_barrier];
3395 extern kmp_uint32 __kmp_barrier_release_branch_bits[bs_last_barrier];
3452 extern kmp_uint32 __kmp_yield_init;
3453 extern kmp_uint32 __kmp_yield_next;
3532 extern kmp_uint32 __kmp_init_mxcsr; /* init thread's mxscr */
3566 kmp_uint32 max_soft_retries;
3569 kmp_uint32 max_badness;
3796 kmp_uint32 list_length,
3814 enum sched_type schedule, kmp_uint32 lb,
3815 kmp_uint32 ub, kmp_int32 st,
3829 kmp_int32 *p_last, kmp_uint32 *p_lb,
3830 kmp_uint32 *p_ub, kmp_int32 *p_st);
3852 enum sched_type schedule, kmp_uint32 lb,
3853 kmp_uint32 ub, kmp_int32 st,
3870 extern kmp_uint32 __kmp_eq_4(kmp_uint32 value, kmp_uint32 checker);
3871 extern kmp_uint32 __kmp_neq_4(kmp_uint32 value, kmp_uint32 checker);
3872 extern kmp_uint32 __kmp_lt_4(kmp_uint32 value, kmp_uint32 checker);
3873 extern kmp_uint32 __kmp_ge_4(kmp_uint32 value, kmp_uint32 checker);
3874 extern kmp_uint32 __kmp_le_4(kmp_uint32 value, kmp_uint32 checker);
3875 extern kmp_uint32 __kmp_wait_4(kmp_uint32 volatile *spinner, kmp_uint32 checker,
3876 kmp_uint32 (*pred)(kmp_uint32, kmp_uint32),
3878 extern void __kmp_wait_4_ptr(void *spinner, kmp_uint32 checker,
3879 kmp_uint32 (*pred)(void *, kmp_uint32), void *obj);
3975 extern void __kmp_get_hierarchy(kmp_uint32 nproc, kmp_bstate_t *thr_bar);
4447 kmp_uint32 list_length,
4450 ident_t *loc, kmp_int32 global_tid, kmp_uint32 list_length,