| /freebsd/contrib/llvm-project/llvm/lib/Transforms/Utils/ |
| H A D | VNCoercion.cpp | 17 bool canCoerceMustAliasedValueToLoad(Value *StoredVal, Type *LoadTy, in canCoerceMustAliasedValueToLoad() argument 20 if (StoredTy == LoadTy) in canCoerceMustAliasedValueToLoad() 25 TypeSize LoadSize = DL.getTypeSizeInBits(LoadTy); in canCoerceMustAliasedValueToLoad() 26 if (isa<ScalableVectorType>(StoredTy) && isa<ScalableVectorType>(LoadTy) && in canCoerceMustAliasedValueToLoad() 33 if (isa<ScalableVectorType>(StoredTy) && isa<FixedVectorType>(LoadTy)) { in canCoerceMustAliasedValueToLoad() 34 if (StoredTy->getScalarType() != LoadTy->getScalarType()) in canCoerceMustAliasedValueToLoad() 43 } else if (isFirstClassAggregateOrScalableType(LoadTy) || in canCoerceMustAliasedValueToLoad() 57 bool LoadNI = DL.isNonIntegralPointerType(LoadTy->getScalarType()); in canCoerceMustAliasedValueToLoad() 68 LoadTy->getPointerAddressSpace()) { in canCoerceMustAliasedValueToLoad() 78 if (StoredTy->isTargetExtTy() || LoadTy->isTargetExtTy()) in canCoerceMustAliasedValueToLoad() [all …]
|
| /freebsd/contrib/llvm-project/llvm/include/llvm/Transforms/Utils/ |
| H A D | VNCoercion.h | 38 bool canCoerceMustAliasedValueToLoad(Value *StoredVal, Type *LoadTy, 55 int analyzeLoadFromClobberingStore(Type *LoadTy, Value *LoadPtr, 63 int analyzeLoadFromClobberingLoad(Type *LoadTy, Value *LoadPtr, LoadInst *DepLI, 71 int analyzeLoadFromClobberingMemInst(Type *LoadTy, Value *LoadPtr, 78 Value *getValueForLoad(Value *SrcVal, unsigned Offset, Type *LoadTy, 83 Type *LoadTy, const DataLayout &DL); 90 Type *LoadTy, Instruction *InsertPt, 95 Type *LoadTy, const DataLayout &DL);
|
| /freebsd/contrib/llvm-project/llvm/lib/Target/ARM/ |
| H A D | ARMParallelDSP.cpp | 225 LoadInst* CreateWideLoad(MemInstList &Loads, IntegerType *LoadTy); 735 IntegerType *LoadTy) { in CreateWideLoad() argument 774 LoadInst *WideLoad = IRB.CreateAlignedLoad(LoadTy, VecPtr, Base->getAlign()); in CreateWideLoad() 788 Value *ShiftVal = ConstantInt::get(LoadTy, OffsetTy->getBitWidth()); in CreateWideLoad()
|
| /freebsd/contrib/llvm-project/llvm/lib/Target/AMDGPU/ |
| H A D | AMDGPUPreloadKernelArguments.cpp | 212 Type *LoadTy = Load->getType(); in tryAllocHiddenArgPreloadSGPRs() local 214 if (HA == END_HIDDEN_ARGS || LoadTy != getHiddenArgType(Ctx, HA)) in tryAllocHiddenArgPreloadSGPRs()
|
| H A D | AMDGPURegisterBankInfo.cpp | 1060 const LLT LoadTy = MRI.getType(DstReg); in applyMappingLoad() local 1061 unsigned LoadSize = LoadTy.getSizeInBits(); in applyMappingLoad() 1079 (MemSize == 32 || LoadTy.isVector() || !isScalarLoadLegal(MI))) in applyMappingLoad() 1114 std::tie(Part64, Part32) = splitUnequalType(LoadTy, 64); in applyMappingLoad() 1120 LLT WiderTy = widen96To128(LoadTy); in applyMappingLoad() 1154 unsigned NumSplitParts = LoadTy.getSizeInBits() / MaxNonSmrdLoadSize; in applyMappingLoad() 1155 const LLT LoadSplitTy = LoadTy.divide(NumSplitParts); in applyMappingLoad() 1158 if (LoadTy.isVector()) { in applyMappingLoad()
|
| H A D | AMDGPUISelLowering.cpp | 871 bool AMDGPUTargetLowering::isLoadBitCastBeneficial(EVT LoadTy, EVT CastTy, in isLoadBitCastBeneficial() argument 875 assert(LoadTy.getSizeInBits() == CastTy.getSizeInBits()); in isLoadBitCastBeneficial() 877 if (LoadTy.getScalarType() == MVT::i32) in isLoadBitCastBeneficial() 880 unsigned LScalarSize = LoadTy.getScalarSizeInBits(); in isLoadBitCastBeneficial()
|
| H A D | AMDGPULegalizerInfo.cpp | 3095 LLT LoadTy = Ty.getSizeInBits() == 32 ? PtrTy : Ty; in legalizeGlobalValue() local 3100 LoadTy, Align(8)); in legalizeGlobalValue() 6141 LLT LoadTy = LLT::fixed_vector(NumLoadDWords, S32); in legalizeBufferLoad() local 6142 Register LoadDstReg = B.getMRI()->createGenericVirtualRegister(LoadTy); in legalizeBufferLoad()
|
| /freebsd/contrib/llvm-project/llvm/lib/Analysis/ |
| H A D | ConstantFolding.cpp | 563 Constant *FoldReinterpretLoadFromConst(Constant *C, Type *LoadTy, in FoldReinterpretLoadFromConst() argument 566 if (isa<ScalableVectorType>(LoadTy)) in FoldReinterpretLoadFromConst() 569 auto *IntType = dyn_cast<IntegerType>(LoadTy); in FoldReinterpretLoadFromConst() 577 if (!LoadTy->isFloatingPointTy() && !LoadTy->isPointerTy() && in FoldReinterpretLoadFromConst() 578 !LoadTy->isVectorTy()) in FoldReinterpretLoadFromConst() 582 DL.getTypeSizeInBits(LoadTy).getFixedValue()); in FoldReinterpretLoadFromConst() 584 if (Res->isNullValue() && !LoadTy->isX86_AMXTy()) in FoldReinterpretLoadFromConst() 586 return Constant::getNullValue(LoadTy); in FoldReinterpretLoadFromConst() 587 Type *CastTy = LoadTy->isPtrOrPtrVectorTy() ? DL.getIntPtrType(LoadTy) : LoadTy; in FoldReinterpretLoadFromConst() 589 if (LoadTy->isPtrOrPtrVectorTy()) { in FoldReinterpretLoadFromConst() [all …]
|
| H A D | Loads.cpp | 555 Type *LoadTy, in areNonOverlapSameBaseLoadAndStore() argument 567 auto LoadAccessSize = LocationSize::precise(DL.getTypeStoreSize(LoadTy)); in areNonOverlapSameBaseLoadAndStore()
|
| /freebsd/contrib/llvm-project/llvm/lib/Transforms/Scalar/ |
| H A D | ScalarizeMaskedMemIntrin.cpp | 1147 Type *LoadTy = CI->getType(); in optimizeCallInst() local 1149 LoadTy->getScalarType()); in optimizeCallInst() 1150 if (TTI.isLegalMaskedGather(LoadTy, Alignment) && in optimizeCallInst() 1151 !TTI.forceScalarizeMaskedGather(cast<VectorType>(LoadTy), Alignment)) in optimizeCallInst()
|
| H A D | GVN.cpp | 1151 Type *LoadTy = Load->getType(); in MaterializeAdjustedValue() local 1155 if (Res->getType() != LoadTy) { in MaterializeAdjustedValue() 1156 Res = getValueForLoad(Res, Offset, LoadTy, InsertPt, Load->getFunction()); in MaterializeAdjustedValue() 1165 if (CoercedLoad->getType() == LoadTy && Offset == 0) { in MaterializeAdjustedValue() 1169 Res = getValueForLoad(CoercedLoad, Offset, LoadTy, InsertPt, in MaterializeAdjustedValue() 1191 Res = getMemInstValueForLoad(getMemIntrinValue(), Offset, LoadTy, in MaterializeAdjustedValue() 1306 static Value *findDominatingValue(const MemoryLocation &Loc, Type *LoadTy, in findDominatingValue() argument 1320 if (LI->getPointerOperand() == Loc.Ptr && LI->getType() == LoadTy) in findDominatingValue()
|
| H A D | SROA.cpp | 1634 Type *LoadTy = SomeLoad->getType(); in speculatePHINodeLoads() local 1636 PHINode *NewPN = IRB.CreatePHI(LoadTy, PN.getNumIncomingValues(), in speculatePHINodeLoads() 1670 LoadTy, InVal, Alignment, in speculatePHINodeLoads()
|
| /freebsd/contrib/llvm-project/llvm/lib/Transforms/AggressiveInstCombine/ |
| H A D | AggressiveInstCombine.cpp | 963 Type *LoadTy = LI->getType(); in foldPatternedLoads() local 976 Constant *Ca = ConstantFoldLoadFromConst(C, LoadTy, ConstOffset, DL); in foldPatternedLoads() 980 unsigned E = GVSize - DL.getTypeStoreSize(LoadTy); in foldPatternedLoads() 982 if (Ca != ConstantFoldLoadFromConst(C, LoadTy, ConstOffset, DL)) in foldPatternedLoads()
|
| /freebsd/contrib/llvm-project/llvm/lib/Transforms/InstCombine/ |
| H A D | InstCombineLoadStoreAlloca.cpp | 683 Type *LoadTy = Load.getType(); in combineLoadToOperationType() local 685 assert(!LoadTy->isX86_AMXTy() && "Load from x86_amx* should not happen!"); in combineLoadToOperationType() 693 LoadTy->isPtrOrPtrVectorTy() == DestTy->isPtrOrPtrVectorTy() && in combineLoadToOperationType()
|
| /freebsd/contrib/llvm-project/llvm/lib/Target/Hexagon/ |
| H A D | HexagonISelLowering.cpp | 3195 MVT LoadTy = ty(Op); in LowerUnalignedLoad() local 3196 unsigned NeedAlign = Subtarget.getTypeAlignment(LoadTy).value(); in LowerUnalignedLoad() 3235 assert(LoadTy.getSizeInBits() == 8*NeedAlign); in LowerUnalignedLoad() 3268 SDValue Load0 = DAG.getLoad(LoadTy, dl, Chain, Base0, WideMMO); in LowerUnalignedLoad() 3269 SDValue Load1 = DAG.getLoad(LoadTy, dl, Chain, Base1, WideMMO); in LowerUnalignedLoad() 3271 SDValue Aligned = DAG.getNode(HexagonISD::VALIGN, dl, LoadTy, in LowerUnalignedLoad()
|
| H A D | HexagonISelLoweringHVX.cpp | 3104 MVT LoadTy = MVT::getVectorVT(MVT::i8, HwLen); in WidenHvxLoad() local 3108 SDValue Load = DAG.getMaskedLoad(LoadTy, dl, Chain, Base, Offset, Mask, in WidenHvxLoad() 3109 DAG.getUNDEF(LoadTy), LoadTy, MemOp, in WidenHvxLoad()
|
| /freebsd/contrib/llvm-project/clang/lib/CodeGen/TargetBuiltins/ |
| H A D | AMDGPU.cpp | 622 llvm::Type *LoadTy = ConvertType(E->getType()); in EmitAMDGPUBuiltinExpr() local 624 llvm::Function *F = CGM.getIntrinsic(IID, {LoadTy}); in EmitAMDGPUBuiltinExpr()
|
| /freebsd/contrib/llvm-project/clang/include/clang/StaticAnalyzer/Core/PathSensitive/ |
| H A D | ExprEngine.h | 690 QualType LoadTy = QualType());
|
| /freebsd/contrib/llvm-project/clang/lib/StaticAnalyzer/Core/ |
| H A D | ExprEngine.cpp | 3793 QualType LoadTy) { in evalLoad() argument 3814 if (LoadTy.isNull()) in evalLoad() 3815 LoadTy = BoundEx->getType(); in evalLoad() 3816 V = state->getSVal(location.castAs<Loc>(), LoadTy); in evalLoad()
|
| /freebsd/contrib/llvm-project/llvm/lib/CodeGen/GlobalISel/ |
| H A D | MachineIRBuilder.cpp | 456 LLT LoadTy = Dst.getLLTTy(*getMRI()); in buildLoadFromOffset() local 458 getMF().getMachineMemOperand(&BaseMMO, Offset, LoadTy); in buildLoadFromOffset()
|
| H A D | LegalizerHelper.cpp | 4069 LLT LoadTy = DstTy; in lowerLoad() local 4074 LoadTy = WideMemTy; in lowerLoad() 4079 auto NewLoad = MIRBuilder.buildLoad(LoadTy, PtrReg, *NewMMO); in lowerLoad() 4081 } else if (isa<GZExtLoad>(LoadMI) || WideMemTy == LoadTy) { in lowerLoad() 4082 auto NewLoad = MIRBuilder.buildLoad(LoadTy, PtrReg, *NewMMO); in lowerLoad() 4090 if (DstTy != LoadTy) in lowerLoad()
|
| /freebsd/contrib/llvm-project/llvm/lib/Transforms/Vectorize/ |
| H A D | VectorCombine.cpp | 272 Type *LoadTy = Load->getType(); in vectorizeLoadInsert() local 275 TTI.getMemoryOpCost(Instruction::Load, LoadTy, Alignment, AS, CostKind); in vectorizeLoadInsert() 349 Type *LoadTy = Load->getType(); in widenSubvectorLoad() local 357 TTI.getMemoryOpCost(Instruction::Load, LoadTy, Alignment, AS, CostKind); in widenSubvectorLoad()
|
| /freebsd/contrib/llvm-project/llvm/lib/CodeGen/SelectionDAG/ |
| H A D | LegalizeDAG.cpp | 1649 MVT LoadTy = TLI.getRegisterType(MVT::i8); in getSignAsIntValue() local 1651 SDValue StackPtr = DAG.CreateStackTemporary(FloatVT, LoadTy); in getSignAsIntValue() 1676 State.IntValue = DAG.getExtLoad(ISD::EXTLOAD, DL, LoadTy, State.Chain, IntPtr, in getSignAsIntValue() 1678 State.SignMask = APInt::getOneBitSet(LoadTy.getScalarSizeInBits(), 7); in getSignAsIntValue()
|
| H A D | SelectionDAGBuilder.cpp | 9012 Type *LoadTy = in getMemCmpLoad() local 9015 LoadTy = FixedVectorType::get(LoadTy, LoadVT.getVectorNumElements()); in getMemCmpLoad() 9018 LoadTy, Builder.DAG.getDataLayout())) in getMemCmpLoad()
|