Home
last modified time | relevance | path

Searched refs:LoadTy (Results 1 – 24 of 24) sorted by relevance

/freebsd/contrib/llvm-project/llvm/lib/Transforms/Utils/
H A DVNCoercion.cpp17 bool canCoerceMustAliasedValueToLoad(Value *StoredVal, Type *LoadTy, in canCoerceMustAliasedValueToLoad() argument
20 if (StoredTy == LoadTy) in canCoerceMustAliasedValueToLoad()
25 TypeSize LoadSize = DL.getTypeSizeInBits(LoadTy); in canCoerceMustAliasedValueToLoad()
26 if (isa<ScalableVectorType>(StoredTy) && isa<ScalableVectorType>(LoadTy) && in canCoerceMustAliasedValueToLoad()
33 if (isa<ScalableVectorType>(StoredTy) && isa<FixedVectorType>(LoadTy)) { in canCoerceMustAliasedValueToLoad()
34 if (StoredTy->getScalarType() != LoadTy->getScalarType()) in canCoerceMustAliasedValueToLoad()
43 } else if (isFirstClassAggregateOrScalableType(LoadTy) || in canCoerceMustAliasedValueToLoad()
57 bool LoadNI = DL.isNonIntegralPointerType(LoadTy->getScalarType()); in canCoerceMustAliasedValueToLoad()
68 LoadTy->getPointerAddressSpace()) { in canCoerceMustAliasedValueToLoad()
78 if (StoredTy->isTargetExtTy() || LoadTy->isTargetExtTy()) in canCoerceMustAliasedValueToLoad()
[all …]
/freebsd/contrib/llvm-project/llvm/include/llvm/Transforms/Utils/
H A DVNCoercion.h38 bool canCoerceMustAliasedValueToLoad(Value *StoredVal, Type *LoadTy,
55 int analyzeLoadFromClobberingStore(Type *LoadTy, Value *LoadPtr,
63 int analyzeLoadFromClobberingLoad(Type *LoadTy, Value *LoadPtr, LoadInst *DepLI,
71 int analyzeLoadFromClobberingMemInst(Type *LoadTy, Value *LoadPtr,
78 Value *getValueForLoad(Value *SrcVal, unsigned Offset, Type *LoadTy,
83 Type *LoadTy, const DataLayout &DL);
90 Type *LoadTy, Instruction *InsertPt,
95 Type *LoadTy, const DataLayout &DL);
/freebsd/contrib/llvm-project/llvm/lib/Target/ARM/
H A DARMParallelDSP.cpp225 LoadInst* CreateWideLoad(MemInstList &Loads, IntegerType *LoadTy);
735 IntegerType *LoadTy) { in CreateWideLoad() argument
774 LoadInst *WideLoad = IRB.CreateAlignedLoad(LoadTy, VecPtr, Base->getAlign()); in CreateWideLoad()
788 Value *ShiftVal = ConstantInt::get(LoadTy, OffsetTy->getBitWidth()); in CreateWideLoad()
/freebsd/contrib/llvm-project/llvm/lib/Target/AMDGPU/
H A DAMDGPUPreloadKernelArguments.cpp212 Type *LoadTy = Load->getType(); in tryAllocHiddenArgPreloadSGPRs() local
214 if (HA == END_HIDDEN_ARGS || LoadTy != getHiddenArgType(Ctx, HA)) in tryAllocHiddenArgPreloadSGPRs()
H A DAMDGPURegisterBankInfo.cpp1060 const LLT LoadTy = MRI.getType(DstReg); in applyMappingLoad() local
1061 unsigned LoadSize = LoadTy.getSizeInBits(); in applyMappingLoad()
1079 (MemSize == 32 || LoadTy.isVector() || !isScalarLoadLegal(MI))) in applyMappingLoad()
1114 std::tie(Part64, Part32) = splitUnequalType(LoadTy, 64); in applyMappingLoad()
1120 LLT WiderTy = widen96To128(LoadTy); in applyMappingLoad()
1154 unsigned NumSplitParts = LoadTy.getSizeInBits() / MaxNonSmrdLoadSize; in applyMappingLoad()
1155 const LLT LoadSplitTy = LoadTy.divide(NumSplitParts); in applyMappingLoad()
1158 if (LoadTy.isVector()) { in applyMappingLoad()
H A DAMDGPUISelLowering.cpp871 bool AMDGPUTargetLowering::isLoadBitCastBeneficial(EVT LoadTy, EVT CastTy, in isLoadBitCastBeneficial() argument
875 assert(LoadTy.getSizeInBits() == CastTy.getSizeInBits()); in isLoadBitCastBeneficial()
877 if (LoadTy.getScalarType() == MVT::i32) in isLoadBitCastBeneficial()
880 unsigned LScalarSize = LoadTy.getScalarSizeInBits(); in isLoadBitCastBeneficial()
H A DAMDGPULegalizerInfo.cpp3095 LLT LoadTy = Ty.getSizeInBits() == 32 ? PtrTy : Ty; in legalizeGlobalValue() local
3100 LoadTy, Align(8)); in legalizeGlobalValue()
6141 LLT LoadTy = LLT::fixed_vector(NumLoadDWords, S32); in legalizeBufferLoad() local
6142 Register LoadDstReg = B.getMRI()->createGenericVirtualRegister(LoadTy); in legalizeBufferLoad()
/freebsd/contrib/llvm-project/llvm/lib/Analysis/
H A DConstantFolding.cpp563 Constant *FoldReinterpretLoadFromConst(Constant *C, Type *LoadTy, in FoldReinterpretLoadFromConst() argument
566 if (isa<ScalableVectorType>(LoadTy)) in FoldReinterpretLoadFromConst()
569 auto *IntType = dyn_cast<IntegerType>(LoadTy); in FoldReinterpretLoadFromConst()
577 if (!LoadTy->isFloatingPointTy() && !LoadTy->isPointerTy() && in FoldReinterpretLoadFromConst()
578 !LoadTy->isVectorTy()) in FoldReinterpretLoadFromConst()
582 DL.getTypeSizeInBits(LoadTy).getFixedValue()); in FoldReinterpretLoadFromConst()
584 if (Res->isNullValue() && !LoadTy->isX86_AMXTy()) in FoldReinterpretLoadFromConst()
586 return Constant::getNullValue(LoadTy); in FoldReinterpretLoadFromConst()
587 Type *CastTy = LoadTy->isPtrOrPtrVectorTy() ? DL.getIntPtrType(LoadTy) : LoadTy; in FoldReinterpretLoadFromConst()
589 if (LoadTy->isPtrOrPtrVectorTy()) { in FoldReinterpretLoadFromConst()
[all …]
H A DLoads.cpp555 Type *LoadTy, in areNonOverlapSameBaseLoadAndStore() argument
567 auto LoadAccessSize = LocationSize::precise(DL.getTypeStoreSize(LoadTy)); in areNonOverlapSameBaseLoadAndStore()
/freebsd/contrib/llvm-project/llvm/lib/Transforms/Scalar/
H A DScalarizeMaskedMemIntrin.cpp1147 Type *LoadTy = CI->getType(); in optimizeCallInst() local
1149 LoadTy->getScalarType()); in optimizeCallInst()
1150 if (TTI.isLegalMaskedGather(LoadTy, Alignment) && in optimizeCallInst()
1151 !TTI.forceScalarizeMaskedGather(cast<VectorType>(LoadTy), Alignment)) in optimizeCallInst()
H A DGVN.cpp1151 Type *LoadTy = Load->getType(); in MaterializeAdjustedValue() local
1155 if (Res->getType() != LoadTy) { in MaterializeAdjustedValue()
1156 Res = getValueForLoad(Res, Offset, LoadTy, InsertPt, Load->getFunction()); in MaterializeAdjustedValue()
1165 if (CoercedLoad->getType() == LoadTy && Offset == 0) { in MaterializeAdjustedValue()
1169 Res = getValueForLoad(CoercedLoad, Offset, LoadTy, InsertPt, in MaterializeAdjustedValue()
1191 Res = getMemInstValueForLoad(getMemIntrinValue(), Offset, LoadTy, in MaterializeAdjustedValue()
1306 static Value *findDominatingValue(const MemoryLocation &Loc, Type *LoadTy, in findDominatingValue() argument
1320 if (LI->getPointerOperand() == Loc.Ptr && LI->getType() == LoadTy) in findDominatingValue()
H A DSROA.cpp1634 Type *LoadTy = SomeLoad->getType(); in speculatePHINodeLoads() local
1636 PHINode *NewPN = IRB.CreatePHI(LoadTy, PN.getNumIncomingValues(), in speculatePHINodeLoads()
1670 LoadTy, InVal, Alignment, in speculatePHINodeLoads()
/freebsd/contrib/llvm-project/llvm/lib/Transforms/AggressiveInstCombine/
H A DAggressiveInstCombine.cpp963 Type *LoadTy = LI->getType(); in foldPatternedLoads() local
976 Constant *Ca = ConstantFoldLoadFromConst(C, LoadTy, ConstOffset, DL); in foldPatternedLoads()
980 unsigned E = GVSize - DL.getTypeStoreSize(LoadTy); in foldPatternedLoads()
982 if (Ca != ConstantFoldLoadFromConst(C, LoadTy, ConstOffset, DL)) in foldPatternedLoads()
/freebsd/contrib/llvm-project/llvm/lib/Transforms/InstCombine/
H A DInstCombineLoadStoreAlloca.cpp683 Type *LoadTy = Load.getType(); in combineLoadToOperationType() local
685 assert(!LoadTy->isX86_AMXTy() && "Load from x86_amx* should not happen!"); in combineLoadToOperationType()
693 LoadTy->isPtrOrPtrVectorTy() == DestTy->isPtrOrPtrVectorTy() && in combineLoadToOperationType()
/freebsd/contrib/llvm-project/llvm/lib/Target/Hexagon/
H A DHexagonISelLowering.cpp3195 MVT LoadTy = ty(Op); in LowerUnalignedLoad() local
3196 unsigned NeedAlign = Subtarget.getTypeAlignment(LoadTy).value(); in LowerUnalignedLoad()
3235 assert(LoadTy.getSizeInBits() == 8*NeedAlign); in LowerUnalignedLoad()
3268 SDValue Load0 = DAG.getLoad(LoadTy, dl, Chain, Base0, WideMMO); in LowerUnalignedLoad()
3269 SDValue Load1 = DAG.getLoad(LoadTy, dl, Chain, Base1, WideMMO); in LowerUnalignedLoad()
3271 SDValue Aligned = DAG.getNode(HexagonISD::VALIGN, dl, LoadTy, in LowerUnalignedLoad()
H A DHexagonISelLoweringHVX.cpp3104 MVT LoadTy = MVT::getVectorVT(MVT::i8, HwLen); in WidenHvxLoad() local
3108 SDValue Load = DAG.getMaskedLoad(LoadTy, dl, Chain, Base, Offset, Mask, in WidenHvxLoad()
3109 DAG.getUNDEF(LoadTy), LoadTy, MemOp, in WidenHvxLoad()
/freebsd/contrib/llvm-project/clang/lib/CodeGen/TargetBuiltins/
H A DAMDGPU.cpp622 llvm::Type *LoadTy = ConvertType(E->getType()); in EmitAMDGPUBuiltinExpr() local
624 llvm::Function *F = CGM.getIntrinsic(IID, {LoadTy}); in EmitAMDGPUBuiltinExpr()
/freebsd/contrib/llvm-project/clang/include/clang/StaticAnalyzer/Core/PathSensitive/
H A DExprEngine.h690 QualType LoadTy = QualType());
/freebsd/contrib/llvm-project/clang/lib/StaticAnalyzer/Core/
H A DExprEngine.cpp3793 QualType LoadTy) { in evalLoad() argument
3814 if (LoadTy.isNull()) in evalLoad()
3815 LoadTy = BoundEx->getType(); in evalLoad()
3816 V = state->getSVal(location.castAs<Loc>(), LoadTy); in evalLoad()
/freebsd/contrib/llvm-project/llvm/lib/CodeGen/GlobalISel/
H A DMachineIRBuilder.cpp456 LLT LoadTy = Dst.getLLTTy(*getMRI()); in buildLoadFromOffset() local
458 getMF().getMachineMemOperand(&BaseMMO, Offset, LoadTy); in buildLoadFromOffset()
H A DLegalizerHelper.cpp4069 LLT LoadTy = DstTy; in lowerLoad() local
4074 LoadTy = WideMemTy; in lowerLoad()
4079 auto NewLoad = MIRBuilder.buildLoad(LoadTy, PtrReg, *NewMMO); in lowerLoad()
4081 } else if (isa<GZExtLoad>(LoadMI) || WideMemTy == LoadTy) { in lowerLoad()
4082 auto NewLoad = MIRBuilder.buildLoad(LoadTy, PtrReg, *NewMMO); in lowerLoad()
4090 if (DstTy != LoadTy) in lowerLoad()
/freebsd/contrib/llvm-project/llvm/lib/Transforms/Vectorize/
H A DVectorCombine.cpp272 Type *LoadTy = Load->getType(); in vectorizeLoadInsert() local
275 TTI.getMemoryOpCost(Instruction::Load, LoadTy, Alignment, AS, CostKind); in vectorizeLoadInsert()
349 Type *LoadTy = Load->getType(); in widenSubvectorLoad() local
357 TTI.getMemoryOpCost(Instruction::Load, LoadTy, Alignment, AS, CostKind); in widenSubvectorLoad()
/freebsd/contrib/llvm-project/llvm/lib/CodeGen/SelectionDAG/
H A DLegalizeDAG.cpp1649 MVT LoadTy = TLI.getRegisterType(MVT::i8); in getSignAsIntValue() local
1651 SDValue StackPtr = DAG.CreateStackTemporary(FloatVT, LoadTy); in getSignAsIntValue()
1676 State.IntValue = DAG.getExtLoad(ISD::EXTLOAD, DL, LoadTy, State.Chain, IntPtr, in getSignAsIntValue()
1678 State.SignMask = APInt::getOneBitSet(LoadTy.getScalarSizeInBits(), 7); in getSignAsIntValue()
H A DSelectionDAGBuilder.cpp9012 Type *LoadTy = in getMemCmpLoad() local
9015 LoadTy = FixedVectorType::get(LoadTy, LoadVT.getVectorNumElements()); in getMemCmpLoad()
9018 LoadTy, Builder.DAG.getDataLayout())) in getMemCmpLoad()