Lines Matching refs:StoreInst
147 using UnspeculatableStore = StoreInst *;
1123 void visitStoreInst(StoreInst &SI) { in visitStoreInst()
1300 if (StoreInst *SI = dyn_cast<StoreInst>(I)) { in hasUnsafePHIOrSelectUse()
1481 } else if (StoreInst *SI = dyn_cast<StoreInst>(U->getUser())) { in findCommonType()
1707 if (auto *Store = dyn_cast<StoreInst>(U)) { in isSafeSelectToSpeculate()
1787 assert((isa<LoadInst>(I) || isa<StoreInst>(I)) && "Only for load and store!"); in rewriteMemOpOfSelect()
1845 else if (auto *SI = dyn_cast<StoreInst>(&I)) in rewriteMemOpOfSelect()
2071 } else if (StoreInst *SI = dyn_cast<StoreInst>(U->getUser())) { in isVectorPromotionViableForSlice()
2300 else if (auto *SI = dyn_cast<StoreInst>(S.getUse()->getUser())) in isVectorPromotionViable()
2386 } else if (StoreInst *SI = dyn_cast<StoreInst>(U->getUser())) { in isIntegerWideningViableForSlice()
2968 bool rewriteVectorizedStoreInst(Value *V, StoreInst &SI, Value *OldOp, in rewriteVectorizedStoreInst()
2991 StoreInst *Store = IRB.CreateAlignedStore(V, &NewAI, NewAI.getAlign()); in rewriteVectorizedStoreInst()
3006 bool rewriteIntegerStore(Value *V, StoreInst &SI, AAMDNodes AATags) { in rewriteIntegerStore()
3019 StoreInst *Store = IRB.CreateAlignedStore(V, &NewAI, NewAI.getAlign()); in rewriteIntegerStore()
3035 bool visitStoreInst(StoreInst &SI) { in visitStoreInst()
3065 StoreInst *NewSI; in visitStoreInst()
3264 StoreInst *New = in visitMemSetInst()
3493 StoreInst *Store = cast<StoreInst>( in visitMemTransferInst()
3582 if (StoreInst *SI = dyn_cast<StoreInst>(I)) { in fixLoadStoreAlign()
3854 AAMDNodes AATags, StoreInst *AggStore, Align BaseAlign, in StoreOpSplitter()
3860 StoreInst *AggStore;
3873 StoreInst *Store = in emitFunc()
3906 bool visitStoreInst(StoreInst &SI) { in visitStoreInst()
4313 SmallVector<StoreInst *, 4> Stores; in presplitLoadsAndStores()
4348 else if (auto *SI = dyn_cast<StoreInst>(I)) in presplitLoadsAndStores()
4365 auto *SI = dyn_cast<StoreInst>(LU); in presplitLoadsAndStores()
4377 } else if (auto *SI = dyn_cast<StoreInst>(I)) { in presplitLoadsAndStores()
4428 llvm::erase_if(Stores, [&UnsplittableLoads, &SplitOffsetsMap](StoreInst *SI) { in presplitLoadsAndStores()
4466 llvm::erase_if(Stores, [&UnsplittableLoads](StoreInst *SI) { in presplitLoadsAndStores()
4566 StoreInst *SI = cast<StoreInst>(LU); in presplitLoadsAndStores()
4586 StoreInst *PStore = IRB.CreateAlignedStore( in presplitLoadsAndStores()
4633 for (StoreInst *SI : Stores) { in presplitLoadsAndStores()
4691 StoreInst *PStore = IRB.CreateAlignedStore( in presplitLoadsAndStores()
5245 isa<StoreInst>(S.getUse()->getUser())) { in splitAlloca()
5261 isa<StoreInst>(S.getUse()->getUser())) { in splitAlloca()