Lines Matching refs:__j
65 // represent. However, we then shift this value to the right by __j, which is
66 // at least __j >= 115, so the result is guaranteed to fit into 179 - 115 = 64
73 // lower result to the higher result, and shift by __j - 64 bits.
95 [[nodiscard]] _LIBCPP_HIDE_FROM_ABI inline uint64_t __mulShift(const uint64_t __m, const uint64_t* const __mul, const int32_t __j) {
105 return __ryu_shiftright128(__sum, __high1, static_cast<uint32_t>(__j - 64));
108 [[nodiscard]] _LIBCPP_HIDE_FROM_ABI inline uint64_t __mulShiftAll(const uint64_t __m, const uint64_t* const __mul, const int32_t __j,
110 *__vp = __mulShift(4 * __m + 2, __mul, __j);
111 *__vm = __mulShift(4 * __m - 1 - __mmShift, __mul, __j);
112 return __mulShift(4 * __m, __mul, __j);
117 [[nodiscard]] _LIBCPP_HIDE_FROM_ABI inline _LIBCPP_ALWAYS_INLINE uint64_t __mulShiftAll(uint64_t __m, const uint64_t* const __mul, const int32_t __j,
130 *__vp = __ryu_shiftright128(__mid2, __hi2, static_cast<uint32_t>(__j - 64 - 1));
136 *__vm = __ryu_shiftright128(__mid3, __hi3, static_cast<uint32_t>(__j - 64 - 1));
144 *__vm = __ryu_shiftright128(__mid4, __hi4, static_cast<uint32_t>(__j - 64));
147 return __ryu_shiftright128(__mid, __hi, static_cast<uint32_t>(__j - 64 - 1));
241 const int32_t __j = static_cast<int32_t>(__q) - __k;
242 __vr = __mulShiftAll(__m2, __DOUBLE_POW5_SPLIT[__i], __j, &__vp, &__vm, __mmShift);