xref: /freebsd/contrib/llvm-project/llvm/lib/Transforms/Utils/CloneFunction.cpp (revision 770cf0a5f02dc8983a89c6568d741fbc25baa999)
1 //===- CloneFunction.cpp - Clone a function into another function ---------===//
2 //
3 // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4 // See https://llvm.org/LICENSE.txt for license information.
5 // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6 //
7 //===----------------------------------------------------------------------===//
8 //
9 // This file implements the CloneFunctionInto interface, which is used as the
10 // low-level function cloner.  This is used by the CloneFunction and function
11 // inliner to do the dirty work of copying the body of a function around.
12 //
13 //===----------------------------------------------------------------------===//
14 
15 #include "llvm/ADT/SmallVector.h"
16 #include "llvm/ADT/Statistic.h"
17 #include "llvm/Analysis/ConstantFolding.h"
18 #include "llvm/Analysis/DomTreeUpdater.h"
19 #include "llvm/Analysis/InstructionSimplify.h"
20 #include "llvm/Analysis/LoopInfo.h"
21 #include "llvm/IR/AttributeMask.h"
22 #include "llvm/IR/CFG.h"
23 #include "llvm/IR/Constants.h"
24 #include "llvm/IR/DebugInfo.h"
25 #include "llvm/IR/DerivedTypes.h"
26 #include "llvm/IR/Function.h"
27 #include "llvm/IR/InstIterator.h"
28 #include "llvm/IR/Instructions.h"
29 #include "llvm/IR/IntrinsicInst.h"
30 #include "llvm/IR/LLVMContext.h"
31 #include "llvm/IR/MDBuilder.h"
32 #include "llvm/IR/Metadata.h"
33 #include "llvm/IR/Module.h"
34 #include "llvm/Transforms/Utils/BasicBlockUtils.h"
35 #include "llvm/Transforms/Utils/Cloning.h"
36 #include "llvm/Transforms/Utils/Local.h"
37 #include "llvm/Transforms/Utils/ValueMapper.h"
38 #include <map>
39 #include <optional>
40 using namespace llvm;
41 
42 #define DEBUG_TYPE "clone-function"
43 
44 STATISTIC(RemappedAtomMax, "Highest global NextAtomGroup (after mapping)");
45 
46 void llvm::mapAtomInstance(const DebugLoc &DL, ValueToValueMapTy &VMap) {
47   auto CurGroup = DL->getAtomGroup();
48   if (!CurGroup)
49     return;
50 
51   // Try inserting a new entry. If there's already a mapping for this atom
52   // then there's nothing to do.
53   auto [It, Inserted] = VMap.AtomMap.insert({{DL.getInlinedAt(), CurGroup}, 0});
54   if (!Inserted)
55     return;
56 
57   // Map entry to a new atom group.
58   uint64_t NewGroup = DL->getContext().incNextDILocationAtomGroup();
59   assert(NewGroup > CurGroup && "Next should always be greater than current");
60   It->second = NewGroup;
61 
62   RemappedAtomMax = std::max<uint64_t>(NewGroup, RemappedAtomMax);
63 }
64 
65 namespace {
66 void collectDebugInfoFromInstructions(const Function &F,
67                                       DebugInfoFinder &DIFinder) {
68   const Module *M = F.getParent();
69   if (M) {
70     // Inspect instructions to process e.g. DILexicalBlocks of inlined functions
71     for (const auto &I : instructions(F))
72       DIFinder.processInstruction(*M, I);
73   }
74 }
75 
76 // Create a predicate that matches the metadata that should be identity mapped
77 // during function cloning.
78 MetadataPredicate createIdentityMDPredicate(const Function &F,
79                                             CloneFunctionChangeType Changes) {
80   if (Changes >= CloneFunctionChangeType::DifferentModule)
81     return [](const Metadata *MD) { return false; };
82 
83   DISubprogram *SPClonedWithinModule = F.getSubprogram();
84 
85   // Don't clone inlined subprograms.
86   auto ShouldKeep = [SPClonedWithinModule](const DISubprogram *SP) -> bool {
87     return SP != SPClonedWithinModule;
88   };
89 
90   return [=](const Metadata *MD) {
91     // Avoid cloning types, compile units, and (other) subprograms.
92     if (isa<DICompileUnit>(MD) || isa<DIType>(MD))
93       return true;
94 
95     if (auto *SP = dyn_cast<DISubprogram>(MD))
96       return ShouldKeep(SP);
97 
98     // If a subprogram isn't going to be cloned skip its lexical blocks as well.
99     if (auto *LScope = dyn_cast<DILocalScope>(MD))
100       return ShouldKeep(LScope->getSubprogram());
101 
102     // Avoid cloning local variables of subprograms that won't be cloned.
103     if (auto *DV = dyn_cast<DILocalVariable>(MD))
104       if (auto *S = dyn_cast_or_null<DILocalScope>(DV->getScope()))
105         return ShouldKeep(S->getSubprogram());
106 
107     return false;
108   };
109 }
110 } // namespace
111 
112 /// See comments in Cloning.h.
113 BasicBlock *llvm::CloneBasicBlock(const BasicBlock *BB, ValueToValueMapTy &VMap,
114                                   const Twine &NameSuffix, Function *F,
115                                   ClonedCodeInfo *CodeInfo, bool MapAtoms) {
116   BasicBlock *NewBB = BasicBlock::Create(BB->getContext(), "", F);
117   if (BB->hasName())
118     NewBB->setName(BB->getName() + NameSuffix);
119 
120   bool hasCalls = false, hasDynamicAllocas = false, hasMemProfMetadata = false;
121 
122   // Loop over all instructions, and copy them over.
123   for (const Instruction &I : *BB) {
124     Instruction *NewInst = I.clone();
125     if (I.hasName())
126       NewInst->setName(I.getName() + NameSuffix);
127 
128     NewInst->insertBefore(*NewBB, NewBB->end());
129     NewInst->cloneDebugInfoFrom(&I);
130 
131     VMap[&I] = NewInst; // Add instruction map to value.
132 
133     if (MapAtoms) {
134       if (const DebugLoc &DL = NewInst->getDebugLoc())
135         mapAtomInstance(DL.get(), VMap);
136     }
137 
138     if (isa<CallInst>(I) && !I.isDebugOrPseudoInst()) {
139       hasCalls = true;
140       hasMemProfMetadata |= I.hasMetadata(LLVMContext::MD_memprof);
141       hasMemProfMetadata |= I.hasMetadata(LLVMContext::MD_callsite);
142     }
143     if (const AllocaInst *AI = dyn_cast<AllocaInst>(&I)) {
144       if (!AI->isStaticAlloca()) {
145         hasDynamicAllocas = true;
146       }
147     }
148   }
149 
150   if (CodeInfo) {
151     CodeInfo->ContainsCalls |= hasCalls;
152     CodeInfo->ContainsMemProfMetadata |= hasMemProfMetadata;
153     CodeInfo->ContainsDynamicAllocas |= hasDynamicAllocas;
154   }
155   return NewBB;
156 }
157 
158 void llvm::CloneFunctionAttributesInto(Function *NewFunc,
159                                        const Function *OldFunc,
160                                        ValueToValueMapTy &VMap,
161                                        bool ModuleLevelChanges,
162                                        ValueMapTypeRemapper *TypeMapper,
163                                        ValueMaterializer *Materializer) {
164   // Copy all attributes other than those stored in Function's AttributeList
165   // which holds e.g. parameters and return value attributes.
166   AttributeList NewAttrs = NewFunc->getAttributes();
167   NewFunc->copyAttributesFrom(OldFunc);
168   NewFunc->setAttributes(NewAttrs);
169 
170   const RemapFlags FuncGlobalRefFlags =
171       ModuleLevelChanges ? RF_None : RF_NoModuleLevelChanges;
172 
173   // Fix up the personality function that got copied over.
174   if (OldFunc->hasPersonalityFn())
175     NewFunc->setPersonalityFn(MapValue(OldFunc->getPersonalityFn(), VMap,
176                                        FuncGlobalRefFlags, TypeMapper,
177                                        Materializer));
178 
179   if (OldFunc->hasPrefixData()) {
180     NewFunc->setPrefixData(MapValue(OldFunc->getPrefixData(), VMap,
181                                     FuncGlobalRefFlags, TypeMapper,
182                                     Materializer));
183   }
184 
185   if (OldFunc->hasPrologueData()) {
186     NewFunc->setPrologueData(MapValue(OldFunc->getPrologueData(), VMap,
187                                       FuncGlobalRefFlags, TypeMapper,
188                                       Materializer));
189   }
190 
191   SmallVector<AttributeSet, 4> NewArgAttrs(NewFunc->arg_size());
192   AttributeList OldAttrs = OldFunc->getAttributes();
193 
194   // Clone any argument attributes that are present in the VMap.
195   for (const Argument &OldArg : OldFunc->args()) {
196     if (Argument *NewArg = dyn_cast<Argument>(VMap[&OldArg])) {
197       // Remap the parameter indices.
198       NewArgAttrs[NewArg->getArgNo()] =
199           OldAttrs.getParamAttrs(OldArg.getArgNo());
200     }
201   }
202 
203   NewFunc->setAttributes(
204       AttributeList::get(NewFunc->getContext(), OldAttrs.getFnAttrs(),
205                          OldAttrs.getRetAttrs(), NewArgAttrs));
206 }
207 
208 void llvm::CloneFunctionMetadataInto(Function &NewFunc, const Function &OldFunc,
209                                      ValueToValueMapTy &VMap,
210                                      RemapFlags RemapFlag,
211                                      ValueMapTypeRemapper *TypeMapper,
212                                      ValueMaterializer *Materializer,
213                                      const MetadataPredicate *IdentityMD) {
214   SmallVector<std::pair<unsigned, MDNode *>, 1> MDs;
215   OldFunc.getAllMetadata(MDs);
216   for (auto MD : MDs) {
217     NewFunc.addMetadata(MD.first,
218                         *MapMetadata(MD.second, VMap, RemapFlag, TypeMapper,
219                                      Materializer, IdentityMD));
220   }
221 }
222 
223 void llvm::CloneFunctionBodyInto(Function &NewFunc, const Function &OldFunc,
224                                  ValueToValueMapTy &VMap, RemapFlags RemapFlag,
225                                  SmallVectorImpl<ReturnInst *> &Returns,
226                                  const char *NameSuffix,
227                                  ClonedCodeInfo *CodeInfo,
228                                  ValueMapTypeRemapper *TypeMapper,
229                                  ValueMaterializer *Materializer,
230                                  const MetadataPredicate *IdentityMD) {
231   if (OldFunc.isDeclaration())
232     return;
233 
234   // Loop over all of the basic blocks in the function, cloning them as
235   // appropriate.  Note that we save BE this way in order to handle cloning of
236   // recursive functions into themselves.
237   for (const BasicBlock &BB : OldFunc) {
238 
239     // Create a new basic block and copy instructions into it!
240     BasicBlock *CBB =
241         CloneBasicBlock(&BB, VMap, NameSuffix, &NewFunc, CodeInfo);
242 
243     // Add basic block mapping.
244     VMap[&BB] = CBB;
245 
246     // It is only legal to clone a function if a block address within that
247     // function is never referenced outside of the function.  Given that, we
248     // want to map block addresses from the old function to block addresses in
249     // the clone. (This is different from the generic ValueMapper
250     // implementation, which generates an invalid blockaddress when
251     // cloning a function.)
252     if (BB.hasAddressTaken()) {
253       Constant *OldBBAddr = BlockAddress::get(const_cast<Function *>(&OldFunc),
254                                               const_cast<BasicBlock *>(&BB));
255       VMap[OldBBAddr] = BlockAddress::get(&NewFunc, CBB);
256     }
257 
258     // Note return instructions for the caller.
259     if (ReturnInst *RI = dyn_cast<ReturnInst>(CBB->getTerminator()))
260       Returns.push_back(RI);
261   }
262 
263   // Loop over all of the instructions in the new function, fixing up operand
264   // references as we go. This uses VMap to do all the hard work.
265   for (Function::iterator
266            BB = cast<BasicBlock>(VMap[&OldFunc.front()])->getIterator(),
267            BE = NewFunc.end();
268        BB != BE; ++BB)
269     // Loop over all instructions, fixing each one as we find it, and any
270     // attached debug-info records.
271     for (Instruction &II : *BB) {
272       RemapInstruction(&II, VMap, RemapFlag, TypeMapper, Materializer,
273                        IdentityMD);
274       RemapDbgRecordRange(II.getModule(), II.getDbgRecordRange(), VMap,
275                           RemapFlag, TypeMapper, Materializer, IdentityMD);
276     }
277 }
278 
279 // Clone OldFunc into NewFunc, transforming the old arguments into references to
280 // VMap values.
281 void llvm::CloneFunctionInto(Function *NewFunc, const Function *OldFunc,
282                              ValueToValueMapTy &VMap,
283                              CloneFunctionChangeType Changes,
284                              SmallVectorImpl<ReturnInst *> &Returns,
285                              const char *NameSuffix, ClonedCodeInfo *CodeInfo,
286                              ValueMapTypeRemapper *TypeMapper,
287                              ValueMaterializer *Materializer) {
288   assert(NameSuffix && "NameSuffix cannot be null!");
289 
290 #ifndef NDEBUG
291   for (const Argument &I : OldFunc->args())
292     assert(VMap.count(&I) && "No mapping from source argument specified!");
293 #endif
294 
295   bool ModuleLevelChanges = Changes > CloneFunctionChangeType::LocalChangesOnly;
296 
297   CloneFunctionAttributesInto(NewFunc, OldFunc, VMap, ModuleLevelChanges,
298                               TypeMapper, Materializer);
299 
300   // Everything else beyond this point deals with function instructions,
301   // so if we are dealing with a function declaration, we're done.
302   if (OldFunc->isDeclaration())
303     return;
304 
305   if (Changes < CloneFunctionChangeType::DifferentModule) {
306     assert((NewFunc->getParent() == nullptr ||
307             NewFunc->getParent() == OldFunc->getParent()) &&
308            "Expected NewFunc to have the same parent, or no parent");
309   } else {
310     assert((NewFunc->getParent() == nullptr ||
311             NewFunc->getParent() != OldFunc->getParent()) &&
312            "Expected NewFunc to have different parents, or no parent");
313 
314     if (Changes == CloneFunctionChangeType::DifferentModule) {
315       assert(NewFunc->getParent() &&
316              "Need parent of new function to maintain debug info invariants");
317     }
318   }
319 
320   MetadataPredicate IdentityMD = createIdentityMDPredicate(*OldFunc, Changes);
321 
322   // Cloning is always a Module level operation, since Metadata needs to be
323   // cloned.
324   const auto RemapFlag = RF_None;
325 
326   CloneFunctionMetadataInto(*NewFunc, *OldFunc, VMap, RemapFlag, TypeMapper,
327                             Materializer, &IdentityMD);
328 
329   CloneFunctionBodyInto(*NewFunc, *OldFunc, VMap, RemapFlag, Returns,
330                         NameSuffix, CodeInfo, TypeMapper, Materializer,
331                         &IdentityMD);
332 
333   // Only update !llvm.dbg.cu for DifferentModule (not CloneModule). In the
334   // same module, the compile unit will already be listed (or not). When
335   // cloning a module, CloneModule() will handle creating the named metadata.
336   if (Changes != CloneFunctionChangeType::DifferentModule)
337     return;
338 
339   // Update !llvm.dbg.cu with compile units added to the new module if this
340   // function is being cloned in isolation.
341   //
342   // FIXME: This is making global / module-level changes, which doesn't seem
343   // like the right encapsulation  Consider dropping the requirement to update
344   // !llvm.dbg.cu (either obsoleting the node, or restricting it to
345   // non-discardable compile units) instead of discovering compile units by
346   // visiting the metadata attached to global values, which would allow this
347   // code to be deleted. Alternatively, perhaps give responsibility for this
348   // update to CloneFunctionInto's callers.
349   auto *NewModule = NewFunc->getParent();
350   auto *NMD = NewModule->getOrInsertNamedMetadata("llvm.dbg.cu");
351   // Avoid multiple insertions of the same DICompileUnit to NMD.
352   SmallPtrSet<const void *, 8> Visited(llvm::from_range, NMD->operands());
353 
354   // Collect and clone all the compile units referenced from the instructions in
355   // the function (e.g. as instructions' scope).
356   DebugInfoFinder DIFinder;
357   collectDebugInfoFromInstructions(*OldFunc, DIFinder);
358   for (auto *Unit : DIFinder.compile_units()) {
359     MDNode *MappedUnit =
360         MapMetadata(Unit, VMap, RF_None, TypeMapper, Materializer);
361     if (Visited.insert(MappedUnit).second)
362       NMD->addOperand(MappedUnit);
363   }
364 }
365 
366 /// Return a copy of the specified function and add it to that function's
367 /// module.  Also, any references specified in the VMap are changed to refer to
368 /// their mapped value instead of the original one.  If any of the arguments to
369 /// the function are in the VMap, the arguments are deleted from the resultant
370 /// function.  The VMap is updated to include mappings from all of the
371 /// instructions and basicblocks in the function from their old to new values.
372 ///
373 Function *llvm::CloneFunction(Function *F, ValueToValueMapTy &VMap,
374                               ClonedCodeInfo *CodeInfo) {
375   std::vector<Type *> ArgTypes;
376 
377   // The user might be deleting arguments to the function by specifying them in
378   // the VMap.  If so, we need to not add the arguments to the arg ty vector
379   //
380   for (const Argument &I : F->args())
381     if (VMap.count(&I) == 0) // Haven't mapped the argument to anything yet?
382       ArgTypes.push_back(I.getType());
383 
384   // Create a new function type...
385   FunctionType *FTy =
386       FunctionType::get(F->getFunctionType()->getReturnType(), ArgTypes,
387                         F->getFunctionType()->isVarArg());
388 
389   // Create the new function...
390   Function *NewF = Function::Create(FTy, F->getLinkage(), F->getAddressSpace(),
391                                     F->getName(), F->getParent());
392 
393   // Loop over the arguments, copying the names of the mapped arguments over...
394   Function::arg_iterator DestI = NewF->arg_begin();
395   for (const Argument &I : F->args())
396     if (VMap.count(&I) == 0) {     // Is this argument preserved?
397       DestI->setName(I.getName()); // Copy the name over...
398       VMap[&I] = &*DestI++;        // Add mapping to VMap
399     }
400 
401   SmallVector<ReturnInst *, 8> Returns; // Ignore returns cloned.
402   CloneFunctionInto(NewF, F, VMap, CloneFunctionChangeType::LocalChangesOnly,
403                     Returns, "", CodeInfo);
404 
405   return NewF;
406 }
407 
408 namespace {
409 /// This is a private class used to implement CloneAndPruneFunctionInto.
410 struct PruningFunctionCloner {
411   Function *NewFunc;
412   const Function *OldFunc;
413   ValueToValueMapTy &VMap;
414   bool ModuleLevelChanges;
415   const char *NameSuffix;
416   ClonedCodeInfo *CodeInfo;
417   bool HostFuncIsStrictFP;
418 
419   Instruction *cloneInstruction(BasicBlock::const_iterator II);
420 
421 public:
422   PruningFunctionCloner(Function *newFunc, const Function *oldFunc,
423                         ValueToValueMapTy &valueMap, bool moduleLevelChanges,
424                         const char *nameSuffix, ClonedCodeInfo *codeInfo)
425       : NewFunc(newFunc), OldFunc(oldFunc), VMap(valueMap),
426         ModuleLevelChanges(moduleLevelChanges), NameSuffix(nameSuffix),
427         CodeInfo(codeInfo) {
428     HostFuncIsStrictFP =
429         newFunc->getAttributes().hasFnAttr(Attribute::StrictFP);
430   }
431 
432   /// The specified block is found to be reachable, clone it and
433   /// anything that it can reach.
434   void CloneBlock(const BasicBlock *BB, BasicBlock::const_iterator StartingInst,
435                   std::vector<const BasicBlock *> &ToClone);
436 };
437 } // namespace
438 
439 Instruction *
440 PruningFunctionCloner::cloneInstruction(BasicBlock::const_iterator II) {
441   const Instruction &OldInst = *II;
442   Instruction *NewInst = nullptr;
443   if (HostFuncIsStrictFP) {
444     Intrinsic::ID CIID = getConstrainedIntrinsicID(OldInst);
445     if (CIID != Intrinsic::not_intrinsic) {
446       // Instead of cloning the instruction, a call to constrained intrinsic
447       // should be created.
448       // Assume the first arguments of constrained intrinsics are the same as
449       // the operands of original instruction.
450 
451       // Determine overloaded types of the intrinsic.
452       SmallVector<Type *, 2> TParams;
453       SmallVector<Intrinsic::IITDescriptor, 8> Descriptor;
454       getIntrinsicInfoTableEntries(CIID, Descriptor);
455       for (unsigned I = 0, E = Descriptor.size(); I != E; ++I) {
456         Intrinsic::IITDescriptor Operand = Descriptor[I];
457         switch (Operand.Kind) {
458         case Intrinsic::IITDescriptor::Argument:
459           if (Operand.getArgumentKind() !=
460               Intrinsic::IITDescriptor::AK_MatchType) {
461             if (I == 0)
462               TParams.push_back(OldInst.getType());
463             else
464               TParams.push_back(OldInst.getOperand(I - 1)->getType());
465           }
466           break;
467         case Intrinsic::IITDescriptor::SameVecWidthArgument:
468           ++I;
469           break;
470         default:
471           break;
472         }
473       }
474 
475       // Create intrinsic call.
476       LLVMContext &Ctx = NewFunc->getContext();
477       Function *IFn = Intrinsic::getOrInsertDeclaration(NewFunc->getParent(),
478                                                         CIID, TParams);
479       SmallVector<Value *, 4> Args;
480       unsigned NumOperands = OldInst.getNumOperands();
481       if (isa<CallInst>(OldInst))
482         --NumOperands;
483       for (unsigned I = 0; I < NumOperands; ++I) {
484         Value *Op = OldInst.getOperand(I);
485         Args.push_back(Op);
486       }
487       if (const auto *CmpI = dyn_cast<FCmpInst>(&OldInst)) {
488         FCmpInst::Predicate Pred = CmpI->getPredicate();
489         StringRef PredName = FCmpInst::getPredicateName(Pred);
490         Args.push_back(MetadataAsValue::get(Ctx, MDString::get(Ctx, PredName)));
491       }
492 
493       // The last arguments of a constrained intrinsic are metadata that
494       // represent rounding mode (absents in some intrinsics) and exception
495       // behavior. The inlined function uses default settings.
496       if (Intrinsic::hasConstrainedFPRoundingModeOperand(CIID))
497         Args.push_back(
498             MetadataAsValue::get(Ctx, MDString::get(Ctx, "round.tonearest")));
499       Args.push_back(
500           MetadataAsValue::get(Ctx, MDString::get(Ctx, "fpexcept.ignore")));
501 
502       NewInst = CallInst::Create(IFn, Args, OldInst.getName() + ".strict");
503     }
504   }
505   if (!NewInst)
506     NewInst = II->clone();
507   return NewInst;
508 }
509 
510 /// The specified block is found to be reachable, clone it and
511 /// anything that it can reach.
512 void PruningFunctionCloner::CloneBlock(
513     const BasicBlock *BB, BasicBlock::const_iterator StartingInst,
514     std::vector<const BasicBlock *> &ToClone) {
515   WeakTrackingVH &BBEntry = VMap[BB];
516 
517   // Have we already cloned this block?
518   if (BBEntry)
519     return;
520 
521   // Nope, clone it now.
522   BasicBlock *NewBB;
523   Twine NewName(BB->hasName() ? Twine(BB->getName()) + NameSuffix : "");
524   BBEntry = NewBB = BasicBlock::Create(BB->getContext(), NewName, NewFunc);
525 
526   // It is only legal to clone a function if a block address within that
527   // function is never referenced outside of the function.  Given that, we
528   // want to map block addresses from the old function to block addresses in
529   // the clone. (This is different from the generic ValueMapper
530   // implementation, which generates an invalid blockaddress when
531   // cloning a function.)
532   //
533   // Note that we don't need to fix the mapping for unreachable blocks;
534   // the default mapping there is safe.
535   if (BB->hasAddressTaken()) {
536     Constant *OldBBAddr = BlockAddress::get(const_cast<Function *>(OldFunc),
537                                             const_cast<BasicBlock *>(BB));
538     VMap[OldBBAddr] = BlockAddress::get(NewFunc, NewBB);
539   }
540 
541   bool hasCalls = false, hasDynamicAllocas = false, hasStaticAllocas = false;
542   bool hasMemProfMetadata = false;
543 
544   // Keep a cursor pointing at the last place we cloned debug-info records from.
545   BasicBlock::const_iterator DbgCursor = StartingInst;
546   auto CloneDbgRecordsToHere =
547       [&DbgCursor](Instruction *NewInst, BasicBlock::const_iterator II) {
548         // Clone debug-info records onto this instruction. Iterate through any
549         // source-instructions we've cloned and then subsequently optimised
550         // away, so that their debug-info doesn't go missing.
551         for (; DbgCursor != II; ++DbgCursor)
552           NewInst->cloneDebugInfoFrom(&*DbgCursor, std::nullopt, false);
553         NewInst->cloneDebugInfoFrom(&*II);
554         DbgCursor = std::next(II);
555       };
556 
557   // Loop over all instructions, and copy them over, DCE'ing as we go.  This
558   // loop doesn't include the terminator.
559   for (BasicBlock::const_iterator II = StartingInst, IE = --BB->end(); II != IE;
560        ++II) {
561 
562     // Don't clone fake_use as it may suppress many optimizations
563     // due to inlining, especially SROA.
564     if (auto *IntrInst = dyn_cast<IntrinsicInst>(II))
565       if (IntrInst->getIntrinsicID() == Intrinsic::fake_use)
566         continue;
567 
568     Instruction *NewInst = cloneInstruction(II);
569     NewInst->insertInto(NewBB, NewBB->end());
570 
571     if (HostFuncIsStrictFP) {
572       // All function calls in the inlined function must get 'strictfp'
573       // attribute to prevent undesirable optimizations.
574       if (auto *Call = dyn_cast<CallInst>(NewInst))
575         Call->addFnAttr(Attribute::StrictFP);
576     }
577 
578     // Eagerly remap operands to the newly cloned instruction, except for PHI
579     // nodes for which we defer processing until we update the CFG. Also defer
580     // debug intrinsic processing because they may contain use-before-defs.
581     if (!isa<PHINode>(NewInst) && !isa<DbgVariableIntrinsic>(NewInst)) {
582       RemapInstruction(NewInst, VMap,
583                        ModuleLevelChanges ? RF_None : RF_NoModuleLevelChanges);
584 
585       // Eagerly constant fold the newly cloned instruction. If successful, add
586       // a mapping to the new value. Non-constant operands may be incomplete at
587       // this stage, thus instruction simplification is performed after
588       // processing phi-nodes.
589       if (Value *V = ConstantFoldInstruction(
590               NewInst, BB->getDataLayout())) {
591         if (isInstructionTriviallyDead(NewInst)) {
592           VMap[&*II] = V;
593           NewInst->eraseFromParent();
594           continue;
595         }
596       }
597     }
598 
599     if (II->hasName())
600       NewInst->setName(II->getName() + NameSuffix);
601     VMap[&*II] = NewInst; // Add instruction map to value.
602     if (isa<CallInst>(II) && !II->isDebugOrPseudoInst()) {
603       hasCalls = true;
604       hasMemProfMetadata |= II->hasMetadata(LLVMContext::MD_memprof);
605       hasMemProfMetadata |= II->hasMetadata(LLVMContext::MD_callsite);
606     }
607 
608     CloneDbgRecordsToHere(NewInst, II);
609 
610     if (CodeInfo) {
611       CodeInfo->OrigVMap[&*II] = NewInst;
612       if (auto *CB = dyn_cast<CallBase>(&*II))
613         if (CB->hasOperandBundles())
614           CodeInfo->OperandBundleCallSites.push_back(NewInst);
615     }
616 
617     if (const AllocaInst *AI = dyn_cast<AllocaInst>(II)) {
618       if (isa<ConstantInt>(AI->getArraySize()))
619         hasStaticAllocas = true;
620       else
621         hasDynamicAllocas = true;
622     }
623   }
624 
625   // Finally, clone over the terminator.
626   const Instruction *OldTI = BB->getTerminator();
627   bool TerminatorDone = false;
628   if (const BranchInst *BI = dyn_cast<BranchInst>(OldTI)) {
629     if (BI->isConditional()) {
630       // If the condition was a known constant in the callee...
631       ConstantInt *Cond = dyn_cast<ConstantInt>(BI->getCondition());
632       // Or is a known constant in the caller...
633       if (!Cond) {
634         Value *V = VMap.lookup(BI->getCondition());
635         Cond = dyn_cast_or_null<ConstantInt>(V);
636       }
637 
638       // Constant fold to uncond branch!
639       if (Cond) {
640         BasicBlock *Dest = BI->getSuccessor(!Cond->getZExtValue());
641         auto *NewBI = BranchInst::Create(Dest, NewBB);
642         NewBI->setDebugLoc(BI->getDebugLoc());
643         VMap[OldTI] = NewBI;
644         ToClone.push_back(Dest);
645         TerminatorDone = true;
646       }
647     }
648   } else if (const SwitchInst *SI = dyn_cast<SwitchInst>(OldTI)) {
649     // If switching on a value known constant in the caller.
650     ConstantInt *Cond = dyn_cast<ConstantInt>(SI->getCondition());
651     if (!Cond) { // Or known constant after constant prop in the callee...
652       Value *V = VMap.lookup(SI->getCondition());
653       Cond = dyn_cast_or_null<ConstantInt>(V);
654     }
655     if (Cond) { // Constant fold to uncond branch!
656       SwitchInst::ConstCaseHandle Case = *SI->findCaseValue(Cond);
657       BasicBlock *Dest = const_cast<BasicBlock *>(Case.getCaseSuccessor());
658       auto *NewBI = BranchInst::Create(Dest, NewBB);
659       NewBI->setDebugLoc(SI->getDebugLoc());
660       VMap[OldTI] = NewBI;
661       ToClone.push_back(Dest);
662       TerminatorDone = true;
663     }
664   }
665 
666   if (!TerminatorDone) {
667     Instruction *NewInst = OldTI->clone();
668     if (OldTI->hasName())
669       NewInst->setName(OldTI->getName() + NameSuffix);
670     NewInst->insertInto(NewBB, NewBB->end());
671 
672     CloneDbgRecordsToHere(NewInst, OldTI->getIterator());
673 
674     VMap[OldTI] = NewInst; // Add instruction map to value.
675 
676     if (CodeInfo) {
677       CodeInfo->OrigVMap[OldTI] = NewInst;
678       if (auto *CB = dyn_cast<CallBase>(OldTI))
679         if (CB->hasOperandBundles())
680           CodeInfo->OperandBundleCallSites.push_back(NewInst);
681     }
682 
683     // Recursively clone any reachable successor blocks.
684     append_range(ToClone, successors(BB->getTerminator()));
685   } else {
686     // If we didn't create a new terminator, clone DbgVariableRecords from the
687     // old terminator onto the new terminator.
688     Instruction *NewInst = NewBB->getTerminator();
689     assert(NewInst);
690 
691     CloneDbgRecordsToHere(NewInst, OldTI->getIterator());
692   }
693 
694   if (CodeInfo) {
695     CodeInfo->ContainsCalls |= hasCalls;
696     CodeInfo->ContainsMemProfMetadata |= hasMemProfMetadata;
697     CodeInfo->ContainsDynamicAllocas |= hasDynamicAllocas;
698     CodeInfo->ContainsDynamicAllocas |=
699         hasStaticAllocas && BB != &BB->getParent()->front();
700   }
701 }
702 
703 /// This works like CloneAndPruneFunctionInto, except that it does not clone the
704 /// entire function. Instead it starts at an instruction provided by the caller
705 /// and copies (and prunes) only the code reachable from that instruction.
706 void llvm::CloneAndPruneIntoFromInst(Function *NewFunc, const Function *OldFunc,
707                                      const Instruction *StartingInst,
708                                      ValueToValueMapTy &VMap,
709                                      bool ModuleLevelChanges,
710                                      SmallVectorImpl<ReturnInst *> &Returns,
711                                      const char *NameSuffix,
712                                      ClonedCodeInfo *CodeInfo) {
713   assert(NameSuffix && "NameSuffix cannot be null!");
714 
715   ValueMapTypeRemapper *TypeMapper = nullptr;
716   ValueMaterializer *Materializer = nullptr;
717 
718 #ifndef NDEBUG
719   // If the cloning starts at the beginning of the function, verify that
720   // the function arguments are mapped.
721   if (!StartingInst)
722     for (const Argument &II : OldFunc->args())
723       assert(VMap.count(&II) && "No mapping from source argument specified!");
724 #endif
725 
726   PruningFunctionCloner PFC(NewFunc, OldFunc, VMap, ModuleLevelChanges,
727                             NameSuffix, CodeInfo);
728   const BasicBlock *StartingBB;
729   if (StartingInst)
730     StartingBB = StartingInst->getParent();
731   else {
732     StartingBB = &OldFunc->getEntryBlock();
733     StartingInst = &StartingBB->front();
734   }
735 
736   // Collect debug intrinsics for remapping later.
737   SmallVector<const DbgVariableIntrinsic *, 8> DbgIntrinsics;
738   for (const auto &BB : *OldFunc) {
739     for (const auto &I : BB) {
740       if (const auto *DVI = dyn_cast<DbgVariableIntrinsic>(&I))
741         DbgIntrinsics.push_back(DVI);
742     }
743   }
744 
745   // Clone the entry block, and anything recursively reachable from it.
746   std::vector<const BasicBlock *> CloneWorklist;
747   PFC.CloneBlock(StartingBB, StartingInst->getIterator(), CloneWorklist);
748   while (!CloneWorklist.empty()) {
749     const BasicBlock *BB = CloneWorklist.back();
750     CloneWorklist.pop_back();
751     PFC.CloneBlock(BB, BB->begin(), CloneWorklist);
752   }
753 
754   // Loop over all of the basic blocks in the old function.  If the block was
755   // reachable, we have cloned it and the old block is now in the value map:
756   // insert it into the new function in the right order.  If not, ignore it.
757   //
758   // Defer PHI resolution until rest of function is resolved.
759   SmallVector<const PHINode *, 16> PHIToResolve;
760   for (const BasicBlock &BI : *OldFunc) {
761     Value *V = VMap.lookup(&BI);
762     BasicBlock *NewBB = cast_or_null<BasicBlock>(V);
763     if (!NewBB)
764       continue; // Dead block.
765 
766     // Move the new block to preserve the order in the original function.
767     NewBB->moveBefore(NewFunc->end());
768 
769     // Handle PHI nodes specially, as we have to remove references to dead
770     // blocks.
771     for (const PHINode &PN : BI.phis()) {
772       // PHI nodes may have been remapped to non-PHI nodes by the caller or
773       // during the cloning process.
774       if (isa<PHINode>(VMap[&PN]))
775         PHIToResolve.push_back(&PN);
776       else
777         break;
778     }
779 
780     // Finally, remap the terminator instructions, as those can't be remapped
781     // until all BBs are mapped.
782     RemapInstruction(NewBB->getTerminator(), VMap,
783                      ModuleLevelChanges ? RF_None : RF_NoModuleLevelChanges,
784                      TypeMapper, Materializer);
785   }
786 
787   // Defer PHI resolution until rest of function is resolved, PHI resolution
788   // requires the CFG to be up-to-date.
789   for (unsigned phino = 0, e = PHIToResolve.size(); phino != e;) {
790     const PHINode *OPN = PHIToResolve[phino];
791     unsigned NumPreds = OPN->getNumIncomingValues();
792     const BasicBlock *OldBB = OPN->getParent();
793     BasicBlock *NewBB = cast<BasicBlock>(VMap[OldBB]);
794 
795     // Map operands for blocks that are live and remove operands for blocks
796     // that are dead.
797     for (; phino != PHIToResolve.size() &&
798            PHIToResolve[phino]->getParent() == OldBB;
799          ++phino) {
800       OPN = PHIToResolve[phino];
801       PHINode *PN = cast<PHINode>(VMap[OPN]);
802       for (unsigned pred = 0, e = NumPreds; pred != e; ++pred) {
803         Value *V = VMap.lookup(PN->getIncomingBlock(pred));
804         if (BasicBlock *MappedBlock = cast_or_null<BasicBlock>(V)) {
805           Value *InVal =
806               MapValue(PN->getIncomingValue(pred), VMap,
807                        ModuleLevelChanges ? RF_None : RF_NoModuleLevelChanges);
808           assert(InVal && "Unknown input value?");
809           PN->setIncomingValue(pred, InVal);
810           PN->setIncomingBlock(pred, MappedBlock);
811         } else {
812           PN->removeIncomingValue(pred, false);
813           --pred; // Revisit the next entry.
814           --e;
815         }
816       }
817     }
818 
819     // The loop above has removed PHI entries for those blocks that are dead
820     // and has updated others.  However, if a block is live (i.e. copied over)
821     // but its terminator has been changed to not go to this block, then our
822     // phi nodes will have invalid entries.  Update the PHI nodes in this
823     // case.
824     PHINode *PN = cast<PHINode>(NewBB->begin());
825     NumPreds = pred_size(NewBB);
826     if (NumPreds != PN->getNumIncomingValues()) {
827       assert(NumPreds < PN->getNumIncomingValues());
828       // Count how many times each predecessor comes to this block.
829       std::map<BasicBlock *, unsigned> PredCount;
830       for (BasicBlock *Pred : predecessors(NewBB))
831         --PredCount[Pred];
832 
833       // Figure out how many entries to remove from each PHI.
834       for (unsigned i = 0, e = PN->getNumIncomingValues(); i != e; ++i)
835         ++PredCount[PN->getIncomingBlock(i)];
836 
837       // At this point, the excess predecessor entries are positive in the
838       // map.  Loop over all of the PHIs and remove excess predecessor
839       // entries.
840       BasicBlock::iterator I = NewBB->begin();
841       for (; (PN = dyn_cast<PHINode>(I)); ++I) {
842         for (const auto &PCI : PredCount) {
843           BasicBlock *Pred = PCI.first;
844           for (unsigned NumToRemove = PCI.second; NumToRemove; --NumToRemove)
845             PN->removeIncomingValue(Pred, false);
846         }
847       }
848     }
849 
850     // If the loops above have made these phi nodes have 0 or 1 operand,
851     // replace them with poison or the input value.  We must do this for
852     // correctness, because 0-operand phis are not valid.
853     PN = cast<PHINode>(NewBB->begin());
854     if (PN->getNumIncomingValues() == 0) {
855       BasicBlock::iterator I = NewBB->begin();
856       BasicBlock::const_iterator OldI = OldBB->begin();
857       while ((PN = dyn_cast<PHINode>(I++))) {
858         Value *NV = PoisonValue::get(PN->getType());
859         PN->replaceAllUsesWith(NV);
860         assert(VMap[&*OldI] == PN && "VMap mismatch");
861         VMap[&*OldI] = NV;
862         PN->eraseFromParent();
863         ++OldI;
864       }
865     }
866   }
867 
868   // Drop all incompatible return attributes that cannot be applied to NewFunc
869   // during cloning, so as to allow instruction simplification to reason on the
870   // old state of the function. The original attributes are restored later.
871   AttributeList Attrs = NewFunc->getAttributes();
872   AttributeMask IncompatibleAttrs = AttributeFuncs::typeIncompatible(
873       OldFunc->getReturnType(), Attrs.getRetAttrs());
874   NewFunc->removeRetAttrs(IncompatibleAttrs);
875 
876   // As phi-nodes have been now remapped, allow incremental simplification of
877   // newly-cloned instructions.
878   const DataLayout &DL = NewFunc->getDataLayout();
879   for (const auto &BB : *OldFunc) {
880     for (const auto &I : BB) {
881       auto *NewI = dyn_cast_or_null<Instruction>(VMap.lookup(&I));
882       if (!NewI)
883         continue;
884 
885       if (Value *V = simplifyInstruction(NewI, DL)) {
886         NewI->replaceAllUsesWith(V);
887 
888         if (isInstructionTriviallyDead(NewI)) {
889           NewI->eraseFromParent();
890         } else {
891           // Did not erase it? Restore the new instruction into VMap previously
892           // dropped by `ValueIsRAUWd`.
893           VMap[&I] = NewI;
894         }
895       }
896     }
897   }
898 
899   // Restore attributes.
900   NewFunc->setAttributes(Attrs);
901 
902   // Remap debug intrinsic operands now that all values have been mapped.
903   // Doing this now (late) preserves use-before-defs in debug intrinsics. If
904   // we didn't do this, ValueAsMetadata(use-before-def) operands would be
905   // replaced by empty metadata. This would signal later cleanup passes to
906   // remove the debug intrinsics, potentially causing incorrect locations.
907   for (const auto *DVI : DbgIntrinsics) {
908     if (DbgVariableIntrinsic *NewDVI =
909             cast_or_null<DbgVariableIntrinsic>(VMap.lookup(DVI)))
910       RemapInstruction(NewDVI, VMap,
911                        ModuleLevelChanges ? RF_None : RF_NoModuleLevelChanges,
912                        TypeMapper, Materializer);
913   }
914 
915   // Do the same for DbgVariableRecords, touching all the instructions in the
916   // cloned range of blocks.
917   Function::iterator Begin = cast<BasicBlock>(VMap[StartingBB])->getIterator();
918   for (BasicBlock &BB : make_range(Begin, NewFunc->end())) {
919     for (Instruction &I : BB) {
920       RemapDbgRecordRange(I.getModule(), I.getDbgRecordRange(), VMap,
921                           ModuleLevelChanges ? RF_None
922                                              : RF_NoModuleLevelChanges,
923                           TypeMapper, Materializer);
924     }
925   }
926 
927   // Simplify conditional branches and switches with a constant operand. We try
928   // to prune these out when cloning, but if the simplification required
929   // looking through PHI nodes, those are only available after forming the full
930   // basic block. That may leave some here, and we still want to prune the dead
931   // code as early as possible.
932   for (BasicBlock &BB : make_range(Begin, NewFunc->end()))
933     ConstantFoldTerminator(&BB);
934 
935   // Some blocks may have become unreachable as a result. Find and delete them.
936   {
937     SmallPtrSet<BasicBlock *, 16> ReachableBlocks;
938     SmallVector<BasicBlock *, 16> Worklist;
939     Worklist.push_back(&*Begin);
940     while (!Worklist.empty()) {
941       BasicBlock *BB = Worklist.pop_back_val();
942       if (ReachableBlocks.insert(BB).second)
943         append_range(Worklist, successors(BB));
944     }
945 
946     SmallVector<BasicBlock *, 16> UnreachableBlocks;
947     for (BasicBlock &BB : make_range(Begin, NewFunc->end()))
948       if (!ReachableBlocks.contains(&BB))
949         UnreachableBlocks.push_back(&BB);
950     DeleteDeadBlocks(UnreachableBlocks);
951   }
952 
953   // Now that the inlined function body has been fully constructed, go through
954   // and zap unconditional fall-through branches. This happens all the time when
955   // specializing code: code specialization turns conditional branches into
956   // uncond branches, and this code folds them.
957   Function::iterator I = Begin;
958   while (I != NewFunc->end()) {
959     BranchInst *BI = dyn_cast<BranchInst>(I->getTerminator());
960     if (!BI || BI->isConditional()) {
961       ++I;
962       continue;
963     }
964 
965     BasicBlock *Dest = BI->getSuccessor(0);
966     if (!Dest->getSinglePredecessor() || Dest->hasAddressTaken()) {
967       ++I;
968       continue;
969     }
970 
971     // We shouldn't be able to get single-entry PHI nodes here, as instsimplify
972     // above should have zapped all of them..
973     assert(!isa<PHINode>(Dest->begin()));
974 
975     // We know all single-entry PHI nodes in the inlined function have been
976     // removed, so we just need to splice the blocks.
977     BI->eraseFromParent();
978 
979     // Make all PHI nodes that referred to Dest now refer to I as their source.
980     Dest->replaceAllUsesWith(&*I);
981 
982     // Move all the instructions in the succ to the pred.
983     I->splice(I->end(), Dest);
984 
985     // Remove the dest block.
986     Dest->eraseFromParent();
987 
988     // Do not increment I, iteratively merge all things this block branches to.
989   }
990 
991   // Make a final pass over the basic blocks from the old function to gather
992   // any return instructions which survived folding. We have to do this here
993   // because we can iteratively remove and merge returns above.
994   for (Function::iterator I = cast<BasicBlock>(VMap[StartingBB])->getIterator(),
995                           E = NewFunc->end();
996        I != E; ++I)
997     if (ReturnInst *RI = dyn_cast<ReturnInst>(I->getTerminator()))
998       Returns.push_back(RI);
999 }
1000 
1001 /// This works exactly like CloneFunctionInto,
1002 /// except that it does some simple constant prop and DCE on the fly.  The
1003 /// effect of this is to copy significantly less code in cases where (for
1004 /// example) a function call with constant arguments is inlined, and those
1005 /// constant arguments cause a significant amount of code in the callee to be
1006 /// dead.  Since this doesn't produce an exact copy of the input, it can't be
1007 /// used for things like CloneFunction or CloneModule.
1008 void llvm::CloneAndPruneFunctionInto(
1009     Function *NewFunc, const Function *OldFunc, ValueToValueMapTy &VMap,
1010     bool ModuleLevelChanges, SmallVectorImpl<ReturnInst *> &Returns,
1011     const char *NameSuffix, ClonedCodeInfo *CodeInfo) {
1012   CloneAndPruneIntoFromInst(NewFunc, OldFunc, &OldFunc->front().front(), VMap,
1013                             ModuleLevelChanges, Returns, NameSuffix, CodeInfo);
1014 }
1015 
1016 /// Remaps instructions in \p Blocks using the mapping in \p VMap.
1017 void llvm::remapInstructionsInBlocks(ArrayRef<BasicBlock *> Blocks,
1018                                      ValueToValueMapTy &VMap) {
1019   // Rewrite the code to refer to itself.
1020   for (auto *BB : Blocks) {
1021     for (auto &Inst : *BB) {
1022       RemapDbgRecordRange(Inst.getModule(), Inst.getDbgRecordRange(), VMap,
1023                           RF_NoModuleLevelChanges | RF_IgnoreMissingLocals);
1024       RemapInstruction(&Inst, VMap,
1025                        RF_NoModuleLevelChanges | RF_IgnoreMissingLocals);
1026     }
1027   }
1028 }
1029 
1030 /// Clones a loop \p OrigLoop.  Returns the loop and the blocks in \p
1031 /// Blocks.
1032 ///
1033 /// Updates LoopInfo and DominatorTree assuming the loop is dominated by block
1034 /// \p LoopDomBB.  Insert the new blocks before block specified in \p Before.
1035 Loop *llvm::cloneLoopWithPreheader(BasicBlock *Before, BasicBlock *LoopDomBB,
1036                                    Loop *OrigLoop, ValueToValueMapTy &VMap,
1037                                    const Twine &NameSuffix, LoopInfo *LI,
1038                                    DominatorTree *DT,
1039                                    SmallVectorImpl<BasicBlock *> &Blocks) {
1040   Function *F = OrigLoop->getHeader()->getParent();
1041   Loop *ParentLoop = OrigLoop->getParentLoop();
1042   DenseMap<Loop *, Loop *> LMap;
1043 
1044   Loop *NewLoop = LI->AllocateLoop();
1045   LMap[OrigLoop] = NewLoop;
1046   if (ParentLoop)
1047     ParentLoop->addChildLoop(NewLoop);
1048   else
1049     LI->addTopLevelLoop(NewLoop);
1050 
1051   BasicBlock *OrigPH = OrigLoop->getLoopPreheader();
1052   assert(OrigPH && "No preheader");
1053   BasicBlock *NewPH = CloneBasicBlock(OrigPH, VMap, NameSuffix, F);
1054   // To rename the loop PHIs.
1055   VMap[OrigPH] = NewPH;
1056   Blocks.push_back(NewPH);
1057 
1058   // Update LoopInfo.
1059   if (ParentLoop)
1060     ParentLoop->addBasicBlockToLoop(NewPH, *LI);
1061 
1062   // Update DominatorTree.
1063   DT->addNewBlock(NewPH, LoopDomBB);
1064 
1065   for (Loop *CurLoop : OrigLoop->getLoopsInPreorder()) {
1066     Loop *&NewLoop = LMap[CurLoop];
1067     if (!NewLoop) {
1068       NewLoop = LI->AllocateLoop();
1069 
1070       // Establish the parent/child relationship.
1071       Loop *OrigParent = CurLoop->getParentLoop();
1072       assert(OrigParent && "Could not find the original parent loop");
1073       Loop *NewParentLoop = LMap[OrigParent];
1074       assert(NewParentLoop && "Could not find the new parent loop");
1075 
1076       NewParentLoop->addChildLoop(NewLoop);
1077     }
1078   }
1079 
1080   for (BasicBlock *BB : OrigLoop->getBlocks()) {
1081     Loop *CurLoop = LI->getLoopFor(BB);
1082     Loop *&NewLoop = LMap[CurLoop];
1083     assert(NewLoop && "Expecting new loop to be allocated");
1084 
1085     BasicBlock *NewBB = CloneBasicBlock(BB, VMap, NameSuffix, F);
1086     VMap[BB] = NewBB;
1087 
1088     // Update LoopInfo.
1089     NewLoop->addBasicBlockToLoop(NewBB, *LI);
1090 
1091     // Add DominatorTree node. After seeing all blocks, update to correct
1092     // IDom.
1093     DT->addNewBlock(NewBB, NewPH);
1094 
1095     Blocks.push_back(NewBB);
1096   }
1097 
1098   for (BasicBlock *BB : OrigLoop->getBlocks()) {
1099     // Update loop headers.
1100     Loop *CurLoop = LI->getLoopFor(BB);
1101     if (BB == CurLoop->getHeader())
1102       LMap[CurLoop]->moveToHeader(cast<BasicBlock>(VMap[BB]));
1103 
1104     // Update DominatorTree.
1105     BasicBlock *IDomBB = DT->getNode(BB)->getIDom()->getBlock();
1106     DT->changeImmediateDominator(cast<BasicBlock>(VMap[BB]),
1107                                  cast<BasicBlock>(VMap[IDomBB]));
1108   }
1109 
1110   // Move them physically from the end of the block list.
1111   F->splice(Before->getIterator(), F, NewPH->getIterator());
1112   F->splice(Before->getIterator(), F, NewLoop->getHeader()->getIterator(),
1113             F->end());
1114 
1115   return NewLoop;
1116 }
1117 
1118 /// Duplicate non-Phi instructions from the beginning of block up to
1119 /// StopAt instruction into a split block between BB and its predecessor.
1120 BasicBlock *llvm::DuplicateInstructionsInSplitBetween(
1121     BasicBlock *BB, BasicBlock *PredBB, Instruction *StopAt,
1122     ValueToValueMapTy &ValueMapping, DomTreeUpdater &DTU) {
1123 
1124   assert(count(successors(PredBB), BB) == 1 &&
1125          "There must be a single edge between PredBB and BB!");
1126   // We are going to have to map operands from the original BB block to the new
1127   // copy of the block 'NewBB'.  If there are PHI nodes in BB, evaluate them to
1128   // account for entry from PredBB.
1129   BasicBlock::iterator BI = BB->begin();
1130   for (; PHINode *PN = dyn_cast<PHINode>(BI); ++BI)
1131     ValueMapping[PN] = PN->getIncomingValueForBlock(PredBB);
1132 
1133   BasicBlock *NewBB = SplitEdge(PredBB, BB);
1134   NewBB->setName(PredBB->getName() + ".split");
1135   Instruction *NewTerm = NewBB->getTerminator();
1136 
1137   // FIXME: SplitEdge does not yet take a DTU, so we include the split edge
1138   //        in the update set here.
1139   DTU.applyUpdates({{DominatorTree::Delete, PredBB, BB},
1140                     {DominatorTree::Insert, PredBB, NewBB},
1141                     {DominatorTree::Insert, NewBB, BB}});
1142 
1143   // Clone the non-phi instructions of BB into NewBB, keeping track of the
1144   // mapping and using it to remap operands in the cloned instructions.
1145   // Stop once we see the terminator too. This covers the case where BB's
1146   // terminator gets replaced and StopAt == BB's terminator.
1147   for (; StopAt != &*BI && BB->getTerminator() != &*BI; ++BI) {
1148     Instruction *New = BI->clone();
1149     New->setName(BI->getName());
1150     New->insertBefore(NewTerm->getIterator());
1151     New->cloneDebugInfoFrom(&*BI);
1152     ValueMapping[&*BI] = New;
1153 
1154     // Remap operands to patch up intra-block references.
1155     for (unsigned i = 0, e = New->getNumOperands(); i != e; ++i)
1156       if (Instruction *Inst = dyn_cast<Instruction>(New->getOperand(i))) {
1157         auto I = ValueMapping.find(Inst);
1158         if (I != ValueMapping.end())
1159           New->setOperand(i, I->second);
1160       }
1161 
1162     // Remap debug variable operands.
1163     remapDebugVariable(ValueMapping, New);
1164   }
1165 
1166   return NewBB;
1167 }
1168 
1169 void llvm::cloneNoAliasScopes(ArrayRef<MDNode *> NoAliasDeclScopes,
1170                               DenseMap<MDNode *, MDNode *> &ClonedScopes,
1171                               StringRef Ext, LLVMContext &Context) {
1172   MDBuilder MDB(Context);
1173 
1174   for (auto *ScopeList : NoAliasDeclScopes) {
1175     for (const auto &MDOperand : ScopeList->operands()) {
1176       if (MDNode *MD = dyn_cast<MDNode>(MDOperand)) {
1177         AliasScopeNode SNANode(MD);
1178 
1179         std::string Name;
1180         auto ScopeName = SNANode.getName();
1181         if (!ScopeName.empty())
1182           Name = (Twine(ScopeName) + ":" + Ext).str();
1183         else
1184           Name = std::string(Ext);
1185 
1186         MDNode *NewScope = MDB.createAnonymousAliasScope(
1187             const_cast<MDNode *>(SNANode.getDomain()), Name);
1188         ClonedScopes.insert(std::make_pair(MD, NewScope));
1189       }
1190     }
1191   }
1192 }
1193 
1194 void llvm::adaptNoAliasScopes(Instruction *I,
1195                               const DenseMap<MDNode *, MDNode *> &ClonedScopes,
1196                               LLVMContext &Context) {
1197   auto CloneScopeList = [&](const MDNode *ScopeList) -> MDNode * {
1198     bool NeedsReplacement = false;
1199     SmallVector<Metadata *, 8> NewScopeList;
1200     for (const auto &MDOp : ScopeList->operands()) {
1201       if (MDNode *MD = dyn_cast<MDNode>(MDOp)) {
1202         if (auto *NewMD = ClonedScopes.lookup(MD)) {
1203           NewScopeList.push_back(NewMD);
1204           NeedsReplacement = true;
1205           continue;
1206         }
1207         NewScopeList.push_back(MD);
1208       }
1209     }
1210     if (NeedsReplacement)
1211       return MDNode::get(Context, NewScopeList);
1212     return nullptr;
1213   };
1214 
1215   if (auto *Decl = dyn_cast<NoAliasScopeDeclInst>(I))
1216     if (auto *NewScopeList = CloneScopeList(Decl->getScopeList()))
1217       Decl->setScopeList(NewScopeList);
1218 
1219   auto replaceWhenNeeded = [&](unsigned MD_ID) {
1220     if (const MDNode *CSNoAlias = I->getMetadata(MD_ID))
1221       if (auto *NewScopeList = CloneScopeList(CSNoAlias))
1222         I->setMetadata(MD_ID, NewScopeList);
1223   };
1224   replaceWhenNeeded(LLVMContext::MD_noalias);
1225   replaceWhenNeeded(LLVMContext::MD_alias_scope);
1226 }
1227 
1228 void llvm::cloneAndAdaptNoAliasScopes(ArrayRef<MDNode *> NoAliasDeclScopes,
1229                                       ArrayRef<BasicBlock *> NewBlocks,
1230                                       LLVMContext &Context, StringRef Ext) {
1231   if (NoAliasDeclScopes.empty())
1232     return;
1233 
1234   DenseMap<MDNode *, MDNode *> ClonedScopes;
1235   LLVM_DEBUG(dbgs() << "cloneAndAdaptNoAliasScopes: cloning "
1236                     << NoAliasDeclScopes.size() << " node(s)\n");
1237 
1238   cloneNoAliasScopes(NoAliasDeclScopes, ClonedScopes, Ext, Context);
1239   // Identify instructions using metadata that needs adaptation
1240   for (BasicBlock *NewBlock : NewBlocks)
1241     for (Instruction &I : *NewBlock)
1242       adaptNoAliasScopes(&I, ClonedScopes, Context);
1243 }
1244 
1245 void llvm::cloneAndAdaptNoAliasScopes(ArrayRef<MDNode *> NoAliasDeclScopes,
1246                                       Instruction *IStart, Instruction *IEnd,
1247                                       LLVMContext &Context, StringRef Ext) {
1248   if (NoAliasDeclScopes.empty())
1249     return;
1250 
1251   DenseMap<MDNode *, MDNode *> ClonedScopes;
1252   LLVM_DEBUG(dbgs() << "cloneAndAdaptNoAliasScopes: cloning "
1253                     << NoAliasDeclScopes.size() << " node(s)\n");
1254 
1255   cloneNoAliasScopes(NoAliasDeclScopes, ClonedScopes, Ext, Context);
1256   // Identify instructions using metadata that needs adaptation
1257   assert(IStart->getParent() == IEnd->getParent() && "different basic block ?");
1258   auto ItStart = IStart->getIterator();
1259   auto ItEnd = IEnd->getIterator();
1260   ++ItEnd; // IEnd is included, increment ItEnd to get the end of the range
1261   for (auto &I : llvm::make_range(ItStart, ItEnd))
1262     adaptNoAliasScopes(&I, ClonedScopes, Context);
1263 }
1264 
1265 void llvm::identifyNoAliasScopesToClone(
1266     ArrayRef<BasicBlock *> BBs, SmallVectorImpl<MDNode *> &NoAliasDeclScopes) {
1267   for (BasicBlock *BB : BBs)
1268     for (Instruction &I : *BB)
1269       if (auto *Decl = dyn_cast<NoAliasScopeDeclInst>(&I))
1270         NoAliasDeclScopes.push_back(Decl->getScopeList());
1271 }
1272 
1273 void llvm::identifyNoAliasScopesToClone(
1274     BasicBlock::iterator Start, BasicBlock::iterator End,
1275     SmallVectorImpl<MDNode *> &NoAliasDeclScopes) {
1276   for (Instruction &I : make_range(Start, End))
1277     if (auto *Decl = dyn_cast<NoAliasScopeDeclInst>(&I))
1278       NoAliasDeclScopes.push_back(Decl->getScopeList());
1279 }
1280