Lines Matching refs:old_value
792 TYPE old_value, new_value; \
793 old_value = *(TYPE volatile *)lhs; \
794 new_value = (TYPE)(old_value OP rhs); \
796 (kmp_int##BITS *)lhs, *VOLATILE_CAST(kmp_int##BITS *) & old_value, \
800 old_value = *(TYPE volatile *)lhs; \
801 new_value = (TYPE)(old_value OP rhs); \
818 struct _sss old_value, new_value; \
819 old_value.vvv = (kmp_int##BITS *)&old_value.cmp; \
821 *old_value.vvv = *(volatile kmp_int##BITS *)lhs; \
822 new_value.cmp = (TYPE)(old_value.cmp OP rhs); \
824 (kmp_int##BITS *)lhs, *VOLATILE_CAST(kmp_int##BITS *) old_value.vvv, \
828 *old_value.vvv = *(volatile kmp_int##BITS *)lhs; \
829 new_value.cmp = (TYPE)(old_value.cmp OP rhs); \
845 struct _sss old_value, new_value; \
846 old_value.vvv = (kmp_int##BITS *)&old_value.cmp; \
848 *old_value.vvv = *(volatile kmp_int##BITS *)lhs; \
849 new_value.cmp = old_value.cmp OP rhs; \
851 (kmp_int##BITS *)lhs, *VOLATILE_CAST(kmp_int##BITS *) old_value.vvv, \
855 *old_value.vvv = *(volatile kmp_int##BITS *)lhs; \
856 new_value.cmp = old_value.cmp OP rhs; \
1160 TYPE old_value; \
1162 old_value = temp_val; \
1163 while (old_value OP rhs && /* still need actions? */ \
1166 *VOLATILE_CAST(kmp_int##BITS *) & old_value, \
1169 old_value = temp_val; \
1459 TYPE old_value, new_value; \
1461 old_value = temp_val; \
1462 new_value = (TYPE)(rhs OP old_value); \
1464 (kmp_int##BITS *)lhs, *VOLATILE_CAST(kmp_int##BITS *) & old_value, \
1469 old_value = temp_val; \
1470 new_value = (TYPE)(rhs OP old_value); \
1950 union f_i_union old_value; \
1952 old_value.f_val = temp_val; \
1953 old_value.i_val = KMP_COMPARE_AND_STORE_RET##BITS( \
1955 *VOLATILE_CAST(kmp_int##BITS *) & old_value.i_val, \
1956 *VOLATILE_CAST(kmp_int##BITS *) & old_value.i_val); \
1957 new_value = old_value.f_val; \
2124 TYPE old_value, new_value; \
2126 old_value = temp_val; \
2129 (kmp_int##BITS *)lhs, *VOLATILE_CAST(kmp_int##BITS *) & old_value, \
2132 old_value = temp_val; \
2272 TYPE old_value, new_value; \
2274 old_value = temp_val; \
2275 new_value = (TYPE)(old_value OP rhs); \
2277 (kmp_int##BITS *)lhs, *VOLATILE_CAST(kmp_int##BITS *) & old_value, \
2280 old_value = temp_val; \
2281 new_value = (TYPE)(old_value OP rhs); \
2286 return old_value; \
2301 TYPE old_value, new_value; \
2305 old_value = KMP_TEST_THEN_ADD##BITS(lhs, OP rhs); \
2307 return old_value OP rhs; \
2309 return old_value; \
2640 old_value = *lhs; \
2645 new_value = old_value; \
2669 old_value = temp_val; \
2670 while (old_value OP rhs && /* still need actions? */ \
2673 *VOLATILE_CAST(kmp_int##BITS *) & old_value, \
2676 old_value = temp_val; \
2681 return old_value; \
2688 TYPE new_value, old_value; \
2698 TYPE new_value, old_value; \
2969 TYPE old_value, new_value; \
2971 old_value = temp_val; \
2972 new_value = (TYPE)(rhs OP old_value); \
2974 (kmp_int##BITS *)lhs, *VOLATILE_CAST(kmp_int##BITS *) & old_value, \
2977 old_value = temp_val; \
2978 new_value = (TYPE)(rhs OP old_value); \
2983 return old_value; \
3256 old_value = (*lhs); \
3260 return old_value;
3275 TYPE old_value; \
3277 old_value = KMP_XCHG_FIXED##BITS(lhs, rhs); \
3278 return old_value; \
3283 TYPE old_value; \
3285 old_value = KMP_XCHG_REAL##BITS(lhs, rhs); \
3286 return old_value; \
3293 TYPE old_value, new_value; \
3295 old_value = temp_val; \
3298 (kmp_int##BITS *)lhs, *VOLATILE_CAST(kmp_int##BITS *) & old_value, \
3301 old_value = temp_val; \
3304 return old_value; \
3310 TYPE old_value; \
3311 (void)old_value; \
3339 TYPE old_value; \
3426 kmp_int8 old_value, new_value; in __kmpc_atomic_1() local
3428 old_value = *(kmp_int8 *)lhs; in __kmpc_atomic_1()
3429 (*f)(&new_value, &old_value, rhs); in __kmpc_atomic_1()
3432 while (!KMP_COMPARE_AND_STORE_ACQ8((kmp_int8 *)lhs, *(kmp_int8 *)&old_value, in __kmpc_atomic_1()
3436 old_value = *(kmp_int8 *)lhs; in __kmpc_atomic_1()
3437 (*f)(&new_value, &old_value, rhs); in __kmpc_atomic_1()
3473 kmp_int16 old_value, new_value; in __kmpc_atomic_2() local
3475 old_value = *(kmp_int16 *)lhs; in __kmpc_atomic_2()
3476 (*f)(&new_value, &old_value, rhs); in __kmpc_atomic_2()
3480 (kmp_int16 *)lhs, *(kmp_int16 *)&old_value, *(kmp_int16 *)&new_value)) { in __kmpc_atomic_2()
3483 old_value = *(kmp_int16 *)lhs; in __kmpc_atomic_2()
3484 (*f)(&new_value, &old_value, rhs); in __kmpc_atomic_2()
3522 kmp_int32 old_value, new_value; in __kmpc_atomic_4() local
3524 old_value = *(kmp_int32 *)lhs; in __kmpc_atomic_4()
3525 (*f)(&new_value, &old_value, rhs); in __kmpc_atomic_4()
3529 (kmp_int32 *)lhs, *(kmp_int32 *)&old_value, *(kmp_int32 *)&new_value)) { in __kmpc_atomic_4()
3532 old_value = *(kmp_int32 *)lhs; in __kmpc_atomic_4()
3533 (*f)(&new_value, &old_value, rhs); in __kmpc_atomic_4()
3572 kmp_int64 old_value, new_value; in __kmpc_atomic_8() local
3574 old_value = *(kmp_int64 *)lhs; in __kmpc_atomic_8()
3575 (*f)(&new_value, &old_value, rhs); in __kmpc_atomic_8()
3578 (kmp_int64 *)lhs, *(kmp_int64 *)&old_value, *(kmp_int64 *)&new_value)) { in __kmpc_atomic_8()
3581 old_value = *(kmp_int64 *)lhs; in __kmpc_atomic_8()
3582 (*f)(&new_value, &old_value, rhs); in __kmpc_atomic_8()