Searched refs:IsScaled (Results 1 – 4 of 4) sorted by relevance
/freebsd/contrib/llvm-project/llvm/lib/Target/AArch64/ |
H A D | AArch64LoadStoreOptimizer.cpp | 755 bool IsScaled = !TII->hasUnscaledLdStOffset(Opc); in mergeNarrowZeroStores() local 757 int OffsetStride = IsScaled ? TII->getMemScale(*I) : 1; in mergeNarrowZeroStores()
|
H A D | AArch64ISelLowering.cpp | 6140 unsigned getGatherVecOpcode(bool IsScaled, bool IsSigned, bool NeedsExtend) { in getGatherVecOpcode() argument 6159 auto Key = std::make_tuple(IsScaled, IsSigned, NeedsExtend); in getGatherVecOpcode() 6212 bool IsScaled = MGT->isIndexScaled(); in LowerMGATHER() local 6218 if (IsScaled && ScaleVal != MemVT.getScalarStoreSize()) { in LowerMGATHER() 6300 bool IsScaled = MSC->isIndexScaled(); in LowerMSCATTER() local 6306 if (IsScaled && ScaleVal != MemVT.getScalarStoreSize()) { in LowerMSCATTER()
|
/freebsd/contrib/llvm-project/llvm/lib/Target/AMDGPU/ |
H A D | AMDGPUISelLowering.cpp | 2718 auto [ScaledInput, IsScaled] = getScaledLogInput(DAG, DL, X, Flags); in LowerFLOGCommon() 2775 if (IsScaled) { in LowerFLOGCommon() 2780 DAG.getNode(ISD::SELECT, DL, VT, IsScaled, ShiftK, Zero, Flags); in LowerFLOGCommon() 2804 auto [ScaledInput, IsScaled] = getScaledLogInput(DAG, SL, Src, Flags); in LowerFLOGUnsafe() 2812 SDValue ResultOffset = DAG.getNode(ISD::SELECT, SL, VT, IsScaled, in LowerFLOGUnsafe()
|
H A D | AMDGPULegalizerInfo.cpp | 3374 auto [ScaledInput, IsScaled] = getScaledLogInput(B, X, Flags); in legalizeFlogCommon() 3439 auto Shift = B.buildSelect(Ty, IsScaled, ShiftK, Zero, Flags); in legalizeFlogCommon() 3458 auto [ScaledInput, IsScaled] = getScaledLogInput(B, Src, Flags); in legalizeFlogUnsafe() 3466 B.buildSelect(Ty, IsScaled, ScaledResultOffset, Zero, Flags); in legalizeFlogUnsafe()
|