/freebsd/contrib/llvm-project/llvm/lib/Transforms/Utils/ |
H A D | VNCoercion.cpp | 18 bool canCoerceMustAliasedValueToLoad(Value *StoredVal, Type *LoadTy, in canCoerceMustAliasedValueToLoad() argument 22 if (StoredTy == LoadTy) in canCoerceMustAliasedValueToLoad() 27 if (isFirstClassAggregateOrScalableType(LoadTy) || in canCoerceMustAliasedValueToLoad() 38 if (StoreSize < DL.getTypeSizeInBits(LoadTy).getFixedValue()) in canCoerceMustAliasedValueToLoad() 42 bool LoadNI = DL.isNonIntegralPointerType(LoadTy->getScalarType()); in canCoerceMustAliasedValueToLoad() 53 LoadTy->getPointerAddressSpace()) { in canCoerceMustAliasedValueToLoad() 61 if (StoredNI && StoreSize != DL.getTypeSizeInBits(LoadTy).getFixedValue()) in canCoerceMustAliasedValueToLoad() 64 if (StoredTy->isTargetExtTy() || LoadTy->isTargetExtTy()) in canCoerceMustAliasedValueToLoad() 173 static int analyzeLoadFromClobberingWrite(Type *LoadTy, Value *LoadPtr, in analyzeLoadFromClobberingWrite() argument 179 if (isFirstClassAggregateOrScalableType(LoadTy)) in analyzeLoadFromClobberingWrite() [all …]
|
/freebsd/contrib/llvm-project/llvm/include/llvm/Transforms/Utils/ |
H A D | VNCoercion.h | 37 bool canCoerceMustAliasedValueToLoad(Value *StoredVal, Type *LoadTy, 54 int analyzeLoadFromClobberingStore(Type *LoadTy, Value *LoadPtr, 62 int analyzeLoadFromClobberingLoad(Type *LoadTy, Value *LoadPtr, LoadInst *DepLI, 70 int analyzeLoadFromClobberingMemInst(Type *LoadTy, Value *LoadPtr, 77 Value *getValueForLoad(Value *SrcVal, unsigned Offset, Type *LoadTy, 82 Type *LoadTy, const DataLayout &DL); 89 Type *LoadTy, Instruction *InsertPt, 94 Type *LoadTy, const DataLayout &DL);
|
/freebsd/contrib/llvm-project/llvm/lib/Analysis/ |
H A D | ConstantFolding.cpp | 546 Constant *FoldReinterpretLoadFromConst(Constant *C, Type *LoadTy, in FoldReinterpretLoadFromConst() argument 549 if (isa<ScalableVectorType>(LoadTy)) in FoldReinterpretLoadFromConst() 552 auto *IntType = dyn_cast<IntegerType>(LoadTy); in FoldReinterpretLoadFromConst() 560 if (!LoadTy->isFloatingPointTy() && !LoadTy->isPointerTy() && in FoldReinterpretLoadFromConst() 561 !LoadTy->isVectorTy()) in FoldReinterpretLoadFromConst() 565 DL.getTypeSizeInBits(LoadTy).getFixedValue()); in FoldReinterpretLoadFromConst() 567 if (Res->isNullValue() && !LoadTy->isX86_MMXTy() && in FoldReinterpretLoadFromConst() 568 !LoadTy->isX86_AMXTy()) in FoldReinterpretLoadFromConst() 570 return Constant::getNullValue(LoadTy); in FoldReinterpretLoadFromConst() 571 Type *CastTy = LoadTy->isPtrOrPtrVectorTy() ? DL.getIntPtrType(LoadTy) : LoadTy; in FoldReinterpretLoadFromConst() [all …]
|
H A D | Loads.cpp | 480 Type *LoadTy, in areNonOverlapSameBaseLoadAndStore() argument 492 auto LoadAccessSize = LocationSize::precise(DL.getTypeStoreSize(LoadTy)); in areNonOverlapSameBaseLoadAndStore()
|
/freebsd/contrib/llvm-project/llvm/lib/Target/ARM/ |
H A D | ARMParallelDSP.cpp | 226 LoadInst* CreateWideLoad(MemInstList &Loads, IntegerType *LoadTy); 730 IntegerType *LoadTy) { in CreateWideLoad() argument 769 LoadInst *WideLoad = IRB.CreateAlignedLoad(LoadTy, VecPtr, Base->getAlign()); in CreateWideLoad() 783 Value *ShiftVal = ConstantInt::get(LoadTy, OffsetTy->getBitWidth()); in CreateWideLoad()
|
/freebsd/contrib/llvm-project/llvm/lib/Transforms/Scalar/ |
H A D | ScalarizeMaskedMemIntrin.cpp | 1028 Type *LoadTy = CI->getType(); in optimizeCallInst() local 1030 LoadTy->getScalarType()); in optimizeCallInst() 1031 if (TTI.isLegalMaskedGather(LoadTy, Alignment) && in optimizeCallInst() 1032 !TTI.forceScalarizeMaskedGather(cast<VectorType>(LoadTy), Alignment)) in optimizeCallInst()
|
H A D | GVN.cpp | 1074 Type *LoadTy = Load->getType(); in MaterializeAdjustedValue() local 1078 if (Res->getType() != LoadTy) { in MaterializeAdjustedValue() 1079 Res = getValueForLoad(Res, Offset, LoadTy, InsertPt, DL); in MaterializeAdjustedValue() 1088 if (CoercedLoad->getType() == LoadTy && Offset == 0) { in MaterializeAdjustedValue() 1092 Res = getValueForLoad(CoercedLoad, Offset, LoadTy, InsertPt, DL); in MaterializeAdjustedValue() 1113 Res = getMemInstValueForLoad(getMemIntrinValue(), Offset, LoadTy, in MaterializeAdjustedValue() 1213 static Value *findDominatingValue(const MemoryLocation &Loc, Type *LoadTy, in findDominatingValue() argument 1227 if (LI->getPointerOperand() == Loc.Ptr && LI->getType() == LoadTy) in findDominatingValue()
|
H A D | SROA.cpp | 1607 Type *LoadTy = SomeLoad->getType(); in speculatePHINodeLoads() local 1609 PHINode *NewPN = IRB.CreatePHI(LoadTy, PN.getNumIncomingValues(), in speculatePHINodeLoads() 1643 LoadTy, InVal, Alignment, in speculatePHINodeLoads()
|
/freebsd/contrib/llvm-project/llvm/lib/Transforms/AggressiveInstCombine/ |
H A D | AggressiveInstCombine.cpp | 900 Type *LoadTy = LI->getType(); in foldPatternedLoads() 913 Constant *Ca = ConstantFoldLoadFromConst(C, LoadTy, ConstOffset, DL); in foldPatternedLoads() 917 unsigned E = GVSize - DL.getTypeStoreSize(LoadTy); in foldPatternedLoads() 919 if (Ca != ConstantFoldLoadFromConst(C, LoadTy, ConstOffset, DL)) in foldPatternedLoads() 901 Type *LoadTy = LI->getType(); foldPatternedLoads() local
|
/freebsd/contrib/llvm-project/llvm/lib/Transforms/InstCombine/ |
H A D | InstCombineLoadStoreAlloca.cpp | 668 Type *LoadTy = Load.getType(); in combineLoadToOperationType() local 670 assert(!LoadTy->isX86_AMXTy() && "Load from x86_amx* should not happen!"); in combineLoadToOperationType() 678 LoadTy->isPtrOrPtrVectorTy() == DestTy->isPtrOrPtrVectorTy() && in combineLoadToOperationType()
|
/freebsd/contrib/llvm-project/llvm/lib/Target/AMDGPU/ |
H A D | AMDGPURegisterBankInfo.cpp | 1060 const LLT LoadTy = MRI.getType(DstReg); in applyMappingLoad() local 1061 unsigned LoadSize = LoadTy.getSizeInBits(); in applyMappingLoad() 1079 (MemSize == 32 || LoadTy.isVector() || !isScalarLoadLegal(MI))) in applyMappingLoad() 1114 std::tie(Part64, Part32) = splitUnequalType(LoadTy, 64); in applyMappingLoad() 1120 LLT WiderTy = widen96To128(LoadTy); in applyMappingLoad() 1152 unsigned NumSplitParts = LoadTy.getSizeInBits() / MaxNonSmrdLoadSize; in applyMappingLoad() 1153 const LLT LoadSplitTy = LoadTy.divide(NumSplitParts); in applyMappingLoad() 1157 if (LoadTy.isVector()) { in applyMappingLoad()
|
H A D | AMDGPULegalizerInfo.cpp | 3037 LLT LoadTy = Ty.getSizeInBits() == 32 ? PtrTy : Ty; in legalizeGlobalValue() local 3042 LoadTy, Align(8)); in legalizeGlobalValue() 6011 LLT LoadTy = LLT::fixed_vector(NumLoadDWords, S32); in legalizeBufferLoad() local 6012 Register LoadDstReg = B.getMRI()->createGenericVirtualRegister(LoadTy); in legalizeBufferLoad()
|
H A D | AMDGPUISelLowering.cpp | 856 bool AMDGPUTargetLowering::isLoadBitCastBeneficial(EVT LoadTy, EVT CastTy, in isLoadBitCastBeneficial() argument 860 assert(LoadTy.getSizeInBits() == CastTy.getSizeInBits()); in isLoadBitCastBeneficial() 862 if (LoadTy.getScalarType() == MVT::i32) in isLoadBitCastBeneficial() 865 unsigned LScalarSize = LoadTy.getScalarSizeInBits(); in isLoadBitCastBeneficial()
|
/freebsd/contrib/llvm-project/llvm/lib/Transforms/Vectorize/ |
H A D | VectorCombine.cpp | 243 Type *LoadTy = Load->getType(); in vectorizeLoadInsert() local 246 TTI.getMemoryOpCost(Instruction::Load, LoadTy, Alignment, AS); in vectorizeLoadInsert() 320 Type *LoadTy = Load->getType(); in widenSubvectorLoad() local 328 TTI.getMemoryOpCost(Instruction::Load, LoadTy, Alignment, AS); in widenSubvectorLoad()
|
H A D | SLPVectorizer.cpp | 8354 auto *LoadTy = getWidenedType(VL.front()->getType(), VF); in getBuildVectorCost() local 8360 ? TTI.getMemoryOpCost(Instruction::Load, LoadTy, Alignment, in getBuildVectorCost() 8364 Instruction::Load, LoadTy, LI->getPointerOperand(), in getBuildVectorCost() 8372 Instruction::Load, CostKind, LI->getType(), LoadTy); in getBuildVectorCost() 8380 Instruction::Load, LoadTy, LI0->getPointerOperand(), in getBuildVectorCost() 8415 ShuffleMask, CostKind, I, LoadTy); in getBuildVectorCost()
|
/freebsd/contrib/llvm-project/clang/include/clang/StaticAnalyzer/Core/PathSensitive/ |
H A D | ExprEngine.h | 677 QualType LoadTy = QualType());
|
/freebsd/contrib/llvm-project/llvm/lib/Target/Hexagon/ |
H A D | HexagonISelLowering.cpp | 3173 MVT LoadTy = ty(Op); in LowerUnalignedLoad() local 3174 unsigned NeedAlign = Subtarget.getTypeAlignment(LoadTy).value(); in LowerUnalignedLoad() 3213 assert(LoadTy.getSizeInBits() == 8*NeedAlign); in LowerUnalignedLoad() 3246 SDValue Load0 = DAG.getLoad(LoadTy, dl, Chain, Base0, WideMMO); in LowerUnalignedLoad() 3247 SDValue Load1 = DAG.getLoad(LoadTy, dl, Chain, Base1, WideMMO); in LowerUnalignedLoad() 3249 SDValue Aligned = DAG.getNode(HexagonISD::VALIGN, dl, LoadTy, in LowerUnalignedLoad()
|
H A D | HexagonISelLoweringHVX.cpp | 3067 MVT LoadTy = MVT::getVectorVT(MVT::i8, HwLen); in WidenHvxLoad() local 3071 SDValue Load = DAG.getMaskedLoad(LoadTy, dl, Chain, Base, Offset, Mask, in WidenHvxLoad() 3072 DAG.getUNDEF(LoadTy), LoadTy, MemOp, in WidenHvxLoad()
|
/freebsd/contrib/llvm-project/llvm/lib/CodeGen/GlobalISel/ |
H A D | MachineIRBuilder.cpp | 454 LLT LoadTy = Dst.getLLTTy(*getMRI()); in buildLoadFromOffset() local 456 getMF().getMachineMemOperand(&BaseMMO, Offset, LoadTy); in buildLoadFromOffset()
|
H A D | LegalizerHelper.cpp | 3492 LLT LoadTy = DstTy; in lowerLoad() local 3497 LoadTy = WideMemTy; in lowerLoad() 3502 auto NewLoad = MIRBuilder.buildLoad(LoadTy, PtrReg, *NewMMO); in lowerLoad() 3504 } else if (isa<GZExtLoad>(LoadMI) || WideMemTy == LoadTy) { in lowerLoad() 3505 auto NewLoad = MIRBuilder.buildLoad(LoadTy, PtrReg, *NewMMO); in lowerLoad() 3513 if (DstTy != LoadTy) in lowerLoad()
|
/freebsd/contrib/llvm-project/clang/lib/StaticAnalyzer/Core/ |
H A D | ExprEngine.cpp | 3668 QualType LoadTy) { in evalLoad() argument 3689 if (LoadTy.isNull()) in evalLoad() 3690 LoadTy = BoundEx->getType(); in evalLoad() 3691 V = state->getSVal(location.castAs<Loc>(), LoadTy); in evalLoad()
|
/freebsd/contrib/llvm-project/llvm/lib/CodeGen/SelectionDAG/ |
H A D | LegalizeDAG.cpp | 1583 MVT LoadTy = TLI.getRegisterType(MVT::i8); in getSignAsIntValue() local 1585 SDValue StackPtr = DAG.CreateStackTemporary(FloatVT, LoadTy); in getSignAsIntValue() 1610 State.IntValue = DAG.getExtLoad(ISD::EXTLOAD, DL, LoadTy, State.Chain, IntPtr, in getSignAsIntValue() 1612 State.SignMask = APInt::getOneBitSet(LoadTy.getScalarSizeInBits(), 7); in getSignAsIntValue()
|
H A D | SelectionDAGBuilder.cpp | 8859 Type *LoadTy = in getMemCmpLoad() local 8862 LoadTy = FixedVectorType::get(LoadTy, LoadVT.getVectorNumElements()); in getMemCmpLoad() 8865 PointerType::getUnqual(LoadTy)); in getMemCmpLoad() 8869 LoadTy, Builder.DAG.getDataLayout())) in getMemCmpLoad()
|
/freebsd/contrib/llvm-project/clang/lib/CodeGen/ |
H A D | CGBuiltin.cpp | 18837 llvm::Type *LoadTy = ConvertType(E->getType()); in EmitAMDGPUBuiltinExpr() local 18839 llvm::Function *F = CGM.getIntrinsic(IID, {LoadTy}); in EmitAMDGPUBuiltinExpr()
|