xref: /freebsd/contrib/llvm-project/llvm/lib/Analysis/AliasAnalysis.cpp (revision 700637cbb5e582861067a11aaca4d053546871d2)
1 //==- AliasAnalysis.cpp - Generic Alias Analysis Interface Implementation --==//
2 //
3 // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4 // See https://llvm.org/LICENSE.txt for license information.
5 // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6 //
7 //===----------------------------------------------------------------------===//
8 //
9 // This file implements the generic AliasAnalysis interface which is used as the
10 // common interface used by all clients and implementations of alias analysis.
11 //
12 // This file also implements the default version of the AliasAnalysis interface
13 // that is to be used when no other implementation is specified.  This does some
14 // simple tests that detect obvious cases: two different global pointers cannot
15 // alias, a global cannot alias a malloc, two different mallocs cannot alias,
16 // etc.
17 //
18 // This alias analysis implementation really isn't very good for anything, but
19 // it is very fast, and makes a nice clean default implementation.  Because it
20 // handles lots of little corner cases, other, more complex, alias analysis
21 // implementations may choose to rely on this pass to resolve these simple and
22 // easy cases.
23 //
24 //===----------------------------------------------------------------------===//
25 
26 #include "llvm/Analysis/AliasAnalysis.h"
27 #include "llvm/ADT/Statistic.h"
28 #include "llvm/Analysis/BasicAliasAnalysis.h"
29 #include "llvm/Analysis/CaptureTracking.h"
30 #include "llvm/Analysis/GlobalsModRef.h"
31 #include "llvm/Analysis/MemoryLocation.h"
32 #include "llvm/Analysis/ScalarEvolutionAliasAnalysis.h"
33 #include "llvm/Analysis/ScopedNoAliasAA.h"
34 #include "llvm/Analysis/TargetLibraryInfo.h"
35 #include "llvm/Analysis/TypeBasedAliasAnalysis.h"
36 #include "llvm/Analysis/ValueTracking.h"
37 #include "llvm/IR/Argument.h"
38 #include "llvm/IR/Attributes.h"
39 #include "llvm/IR/BasicBlock.h"
40 #include "llvm/IR/Instruction.h"
41 #include "llvm/IR/Instructions.h"
42 #include "llvm/IR/Type.h"
43 #include "llvm/IR/Value.h"
44 #include "llvm/InitializePasses.h"
45 #include "llvm/Pass.h"
46 #include "llvm/Support/AtomicOrdering.h"
47 #include "llvm/Support/Casting.h"
48 #include "llvm/Support/CommandLine.h"
49 #include <cassert>
50 #include <functional>
51 #include <iterator>
52 
53 #define DEBUG_TYPE "aa"
54 
55 using namespace llvm;
56 
57 STATISTIC(NumNoAlias,   "Number of NoAlias results");
58 STATISTIC(NumMayAlias,  "Number of MayAlias results");
59 STATISTIC(NumMustAlias, "Number of MustAlias results");
60 
61 /// Allow disabling BasicAA from the AA results. This is particularly useful
62 /// when testing to isolate a single AA implementation.
63 static cl::opt<bool> DisableBasicAA("disable-basic-aa", cl::Hidden,
64                                     cl::init(false));
65 
66 #ifndef NDEBUG
67 /// Print a trace of alias analysis queries and their results.
68 static cl::opt<bool> EnableAATrace("aa-trace", cl::Hidden, cl::init(false));
69 #else
70 static const bool EnableAATrace = false;
71 #endif
72 
AAResults(const TargetLibraryInfo & TLI)73 AAResults::AAResults(const TargetLibraryInfo &TLI) : TLI(TLI) {}
74 
AAResults(AAResults && Arg)75 AAResults::AAResults(AAResults &&Arg)
76     : TLI(Arg.TLI), AAs(std::move(Arg.AAs)), AADeps(std::move(Arg.AADeps)) {}
77 
~AAResults()78 AAResults::~AAResults() {}
79 
invalidate(Function & F,const PreservedAnalyses & PA,FunctionAnalysisManager::Invalidator & Inv)80 bool AAResults::invalidate(Function &F, const PreservedAnalyses &PA,
81                            FunctionAnalysisManager::Invalidator &Inv) {
82   // AAResults preserves the AAManager by default, due to the stateless nature
83   // of AliasAnalysis. There is no need to check whether it has been preserved
84   // explicitly. Check if any module dependency was invalidated and caused the
85   // AAManager to be invalidated. Invalidate ourselves in that case.
86   auto PAC = PA.getChecker<AAManager>();
87   if (!PAC.preservedWhenStateless())
88     return true;
89 
90   // Check if any of the function dependencies were invalidated, and invalidate
91   // ourselves in that case.
92   for (AnalysisKey *ID : AADeps)
93     if (Inv.invalidate(ID, F, PA))
94       return true;
95 
96   // Everything we depend on is still fine, so are we. Nothing to invalidate.
97   return false;
98 }
99 
100 //===----------------------------------------------------------------------===//
101 // Default chaining methods
102 //===----------------------------------------------------------------------===//
103 
alias(const MemoryLocation & LocA,const MemoryLocation & LocB)104 AliasResult AAResults::alias(const MemoryLocation &LocA,
105                              const MemoryLocation &LocB) {
106   SimpleAAQueryInfo AAQIP(*this);
107   return alias(LocA, LocB, AAQIP, nullptr);
108 }
109 
alias(const MemoryLocation & LocA,const MemoryLocation & LocB,AAQueryInfo & AAQI,const Instruction * CtxI)110 AliasResult AAResults::alias(const MemoryLocation &LocA,
111                              const MemoryLocation &LocB, AAQueryInfo &AAQI,
112                              const Instruction *CtxI) {
113   assert(LocA.Ptr->getType()->isPointerTy() &&
114          LocB.Ptr->getType()->isPointerTy() &&
115          "Can only call alias() on pointers");
116   AliasResult Result = AliasResult::MayAlias;
117 
118   if (EnableAATrace) {
119     for (unsigned I = 0; I < AAQI.Depth; ++I)
120       dbgs() << "  ";
121     dbgs() << "Start " << *LocA.Ptr << " @ " << LocA.Size << ", "
122            << *LocB.Ptr << " @ " << LocB.Size << "\n";
123   }
124 
125   AAQI.Depth++;
126   for (const auto &AA : AAs) {
127     Result = AA->alias(LocA, LocB, AAQI, CtxI);
128     if (Result != AliasResult::MayAlias)
129       break;
130   }
131   AAQI.Depth--;
132 
133   if (EnableAATrace) {
134     for (unsigned I = 0; I < AAQI.Depth; ++I)
135       dbgs() << "  ";
136     dbgs() << "End " << *LocA.Ptr << " @ " << LocA.Size << ", "
137            << *LocB.Ptr << " @ " << LocB.Size << " = " << Result << "\n";
138   }
139 
140   if (AAQI.Depth == 0) {
141     if (Result == AliasResult::NoAlias)
142       ++NumNoAlias;
143     else if (Result == AliasResult::MustAlias)
144       ++NumMustAlias;
145     else
146       ++NumMayAlias;
147   }
148   return Result;
149 }
150 
getModRefInfoMask(const MemoryLocation & Loc,bool IgnoreLocals)151 ModRefInfo AAResults::getModRefInfoMask(const MemoryLocation &Loc,
152                                         bool IgnoreLocals) {
153   SimpleAAQueryInfo AAQIP(*this);
154   return getModRefInfoMask(Loc, AAQIP, IgnoreLocals);
155 }
156 
getModRefInfoMask(const MemoryLocation & Loc,AAQueryInfo & AAQI,bool IgnoreLocals)157 ModRefInfo AAResults::getModRefInfoMask(const MemoryLocation &Loc,
158                                         AAQueryInfo &AAQI, bool IgnoreLocals) {
159   ModRefInfo Result = ModRefInfo::ModRef;
160 
161   for (const auto &AA : AAs) {
162     Result &= AA->getModRefInfoMask(Loc, AAQI, IgnoreLocals);
163 
164     // Early-exit the moment we reach the bottom of the lattice.
165     if (isNoModRef(Result))
166       return ModRefInfo::NoModRef;
167   }
168 
169   return Result;
170 }
171 
getArgModRefInfo(const CallBase * Call,unsigned ArgIdx)172 ModRefInfo AAResults::getArgModRefInfo(const CallBase *Call, unsigned ArgIdx) {
173   ModRefInfo Result = ModRefInfo::ModRef;
174 
175   for (const auto &AA : AAs) {
176     Result &= AA->getArgModRefInfo(Call, ArgIdx);
177 
178     // Early-exit the moment we reach the bottom of the lattice.
179     if (isNoModRef(Result))
180       return ModRefInfo::NoModRef;
181   }
182 
183   return Result;
184 }
185 
getModRefInfo(const Instruction * I,const CallBase * Call2)186 ModRefInfo AAResults::getModRefInfo(const Instruction *I,
187                                     const CallBase *Call2) {
188   SimpleAAQueryInfo AAQIP(*this);
189   return getModRefInfo(I, Call2, AAQIP);
190 }
191 
getModRefInfo(const Instruction * I,const CallBase * Call2,AAQueryInfo & AAQI)192 ModRefInfo AAResults::getModRefInfo(const Instruction *I, const CallBase *Call2,
193                                     AAQueryInfo &AAQI) {
194   // We may have two calls.
195   if (const auto *Call1 = dyn_cast<CallBase>(I)) {
196     // Check if the two calls modify the same memory.
197     return getModRefInfo(Call1, Call2, AAQI);
198   }
199   // If this is a fence, just return ModRef.
200   if (I->isFenceLike())
201     return ModRefInfo::ModRef;
202   // Otherwise, check if the call modifies or references the
203   // location this memory access defines.  The best we can say
204   // is that if the call references what this instruction
205   // defines, it must be clobbered by this location.
206   const MemoryLocation DefLoc = MemoryLocation::get(I);
207   ModRefInfo MR = getModRefInfo(Call2, DefLoc, AAQI);
208   if (isModOrRefSet(MR))
209     return ModRefInfo::ModRef;
210   return ModRefInfo::NoModRef;
211 }
212 
getModRefInfo(const CallBase * Call,const MemoryLocation & Loc,AAQueryInfo & AAQI)213 ModRefInfo AAResults::getModRefInfo(const CallBase *Call,
214                                     const MemoryLocation &Loc,
215                                     AAQueryInfo &AAQI) {
216   ModRefInfo Result = ModRefInfo::ModRef;
217 
218   for (const auto &AA : AAs) {
219     Result &= AA->getModRefInfo(Call, Loc, AAQI);
220 
221     // Early-exit the moment we reach the bottom of the lattice.
222     if (isNoModRef(Result))
223       return ModRefInfo::NoModRef;
224   }
225 
226   // Apply the ModRef mask. This ensures that if Loc is a constant memory
227   // location, we take into account the fact that the call definitely could not
228   // modify the memory location.
229   if (!isNoModRef(Result))
230     Result &= getModRefInfoMask(Loc);
231 
232   return Result;
233 }
234 
getModRefInfo(const CallBase * Call1,const CallBase * Call2,AAQueryInfo & AAQI)235 ModRefInfo AAResults::getModRefInfo(const CallBase *Call1,
236                                     const CallBase *Call2, AAQueryInfo &AAQI) {
237   ModRefInfo Result = ModRefInfo::ModRef;
238 
239   for (const auto &AA : AAs) {
240     Result &= AA->getModRefInfo(Call1, Call2, AAQI);
241 
242     // Early-exit the moment we reach the bottom of the lattice.
243     if (isNoModRef(Result))
244       return ModRefInfo::NoModRef;
245   }
246 
247   // Try to refine the mod-ref info further using other API entry points to the
248   // aggregate set of AA results.
249 
250   // If Call1 or Call2 are readnone, they don't interact.
251   auto Call1B = getMemoryEffects(Call1, AAQI);
252   if (Call1B.doesNotAccessMemory())
253     return ModRefInfo::NoModRef;
254 
255   auto Call2B = getMemoryEffects(Call2, AAQI);
256   if (Call2B.doesNotAccessMemory())
257     return ModRefInfo::NoModRef;
258 
259   // If they both only read from memory, there is no dependence.
260   if (Call1B.onlyReadsMemory() && Call2B.onlyReadsMemory())
261     return ModRefInfo::NoModRef;
262 
263   // If Call1 only reads memory, the only dependence on Call2 can be
264   // from Call1 reading memory written by Call2.
265   if (Call1B.onlyReadsMemory())
266     Result &= ModRefInfo::Ref;
267   else if (Call1B.onlyWritesMemory())
268     Result &= ModRefInfo::Mod;
269 
270   // If Call2 only access memory through arguments, accumulate the mod/ref
271   // information from Call1's references to the memory referenced by
272   // Call2's arguments.
273   if (Call2B.onlyAccessesArgPointees()) {
274     if (!Call2B.doesAccessArgPointees())
275       return ModRefInfo::NoModRef;
276     ModRefInfo R = ModRefInfo::NoModRef;
277     for (auto I = Call2->arg_begin(), E = Call2->arg_end(); I != E; ++I) {
278       const Value *Arg = *I;
279       if (!Arg->getType()->isPointerTy())
280         continue;
281       unsigned Call2ArgIdx = std::distance(Call2->arg_begin(), I);
282       auto Call2ArgLoc =
283           MemoryLocation::getForArgument(Call2, Call2ArgIdx, TLI);
284 
285       // ArgModRefC2 indicates what Call2 might do to Call2ArgLoc, and the
286       // dependence of Call1 on that location is the inverse:
287       // - If Call2 modifies location, dependence exists if Call1 reads or
288       //   writes.
289       // - If Call2 only reads location, dependence exists if Call1 writes.
290       ModRefInfo ArgModRefC2 = getArgModRefInfo(Call2, Call2ArgIdx);
291       ModRefInfo ArgMask = ModRefInfo::NoModRef;
292       if (isModSet(ArgModRefC2))
293         ArgMask = ModRefInfo::ModRef;
294       else if (isRefSet(ArgModRefC2))
295         ArgMask = ModRefInfo::Mod;
296 
297       // ModRefC1 indicates what Call1 might do to Call2ArgLoc, and we use
298       // above ArgMask to update dependence info.
299       ArgMask &= getModRefInfo(Call1, Call2ArgLoc, AAQI);
300 
301       R = (R | ArgMask) & Result;
302       if (R == Result)
303         break;
304     }
305 
306     return R;
307   }
308 
309   // If Call1 only accesses memory through arguments, check if Call2 references
310   // any of the memory referenced by Call1's arguments. If not, return NoModRef.
311   if (Call1B.onlyAccessesArgPointees()) {
312     if (!Call1B.doesAccessArgPointees())
313       return ModRefInfo::NoModRef;
314     ModRefInfo R = ModRefInfo::NoModRef;
315     for (auto I = Call1->arg_begin(), E = Call1->arg_end(); I != E; ++I) {
316       const Value *Arg = *I;
317       if (!Arg->getType()->isPointerTy())
318         continue;
319       unsigned Call1ArgIdx = std::distance(Call1->arg_begin(), I);
320       auto Call1ArgLoc =
321           MemoryLocation::getForArgument(Call1, Call1ArgIdx, TLI);
322 
323       // ArgModRefC1 indicates what Call1 might do to Call1ArgLoc; if Call1
324       // might Mod Call1ArgLoc, then we care about either a Mod or a Ref by
325       // Call2. If Call1 might Ref, then we care only about a Mod by Call2.
326       ModRefInfo ArgModRefC1 = getArgModRefInfo(Call1, Call1ArgIdx);
327       ModRefInfo ModRefC2 = getModRefInfo(Call2, Call1ArgLoc, AAQI);
328       if ((isModSet(ArgModRefC1) && isModOrRefSet(ModRefC2)) ||
329           (isRefSet(ArgModRefC1) && isModSet(ModRefC2)))
330         R = (R | ArgModRefC1) & Result;
331 
332       if (R == Result)
333         break;
334     }
335 
336     return R;
337   }
338 
339   return Result;
340 }
341 
getModRefInfo(const Instruction * I1,const Instruction * I2)342 ModRefInfo AAResults::getModRefInfo(const Instruction *I1,
343                                     const Instruction *I2) {
344   SimpleAAQueryInfo AAQIP(*this);
345   return getModRefInfo(I1, I2, AAQIP);
346 }
347 
getModRefInfo(const Instruction * I1,const Instruction * I2,AAQueryInfo & AAQI)348 ModRefInfo AAResults::getModRefInfo(const Instruction *I1,
349                                     const Instruction *I2, AAQueryInfo &AAQI) {
350   // Early-exit if either instruction does not read or write memory.
351   if (!I1->mayReadOrWriteMemory() || !I2->mayReadOrWriteMemory())
352     return ModRefInfo::NoModRef;
353 
354   if (const auto *Call2 = dyn_cast<CallBase>(I2))
355     return getModRefInfo(I1, Call2, AAQI);
356 
357   // FIXME: We can have a more precise result.
358   ModRefInfo MR = getModRefInfo(I1, MemoryLocation::getOrNone(I2), AAQI);
359   return isModOrRefSet(MR) ? ModRefInfo::ModRef : ModRefInfo::NoModRef;
360 }
361 
getMemoryEffects(const CallBase * Call,AAQueryInfo & AAQI)362 MemoryEffects AAResults::getMemoryEffects(const CallBase *Call,
363                                           AAQueryInfo &AAQI) {
364   MemoryEffects Result = MemoryEffects::unknown();
365 
366   for (const auto &AA : AAs) {
367     Result &= AA->getMemoryEffects(Call, AAQI);
368 
369     // Early-exit the moment we reach the bottom of the lattice.
370     if (Result.doesNotAccessMemory())
371       return Result;
372   }
373 
374   return Result;
375 }
376 
getMemoryEffects(const CallBase * Call)377 MemoryEffects AAResults::getMemoryEffects(const CallBase *Call) {
378   SimpleAAQueryInfo AAQI(*this);
379   return getMemoryEffects(Call, AAQI);
380 }
381 
getMemoryEffects(const Function * F)382 MemoryEffects AAResults::getMemoryEffects(const Function *F) {
383   MemoryEffects Result = MemoryEffects::unknown();
384 
385   for (const auto &AA : AAs) {
386     Result &= AA->getMemoryEffects(F);
387 
388     // Early-exit the moment we reach the bottom of the lattice.
389     if (Result.doesNotAccessMemory())
390       return Result;
391   }
392 
393   return Result;
394 }
395 
operator <<(raw_ostream & OS,AliasResult AR)396 raw_ostream &llvm::operator<<(raw_ostream &OS, AliasResult AR) {
397   switch (AR) {
398   case AliasResult::NoAlias:
399     OS << "NoAlias";
400     break;
401   case AliasResult::MustAlias:
402     OS << "MustAlias";
403     break;
404   case AliasResult::MayAlias:
405     OS << "MayAlias";
406     break;
407   case AliasResult::PartialAlias:
408     OS << "PartialAlias";
409     if (AR.hasOffset())
410       OS << " (off " << AR.getOffset() << ")";
411     break;
412   }
413   return OS;
414 }
415 
416 //===----------------------------------------------------------------------===//
417 // Helper method implementation
418 //===----------------------------------------------------------------------===//
419 
getModRefInfo(const LoadInst * L,const MemoryLocation & Loc,AAQueryInfo & AAQI)420 ModRefInfo AAResults::getModRefInfo(const LoadInst *L,
421                                     const MemoryLocation &Loc,
422                                     AAQueryInfo &AAQI) {
423   // Be conservative in the face of atomic.
424   if (isStrongerThan(L->getOrdering(), AtomicOrdering::Unordered))
425     return ModRefInfo::ModRef;
426 
427   // If the load address doesn't alias the given address, it doesn't read
428   // or write the specified memory.
429   if (Loc.Ptr) {
430     AliasResult AR = alias(MemoryLocation::get(L), Loc, AAQI, L);
431     if (AR == AliasResult::NoAlias)
432       return ModRefInfo::NoModRef;
433   }
434   // Otherwise, a load just reads.
435   return ModRefInfo::Ref;
436 }
437 
getModRefInfo(const StoreInst * S,const MemoryLocation & Loc,AAQueryInfo & AAQI)438 ModRefInfo AAResults::getModRefInfo(const StoreInst *S,
439                                     const MemoryLocation &Loc,
440                                     AAQueryInfo &AAQI) {
441   // Be conservative in the face of atomic.
442   if (isStrongerThan(S->getOrdering(), AtomicOrdering::Unordered))
443     return ModRefInfo::ModRef;
444 
445   if (Loc.Ptr) {
446     AliasResult AR = alias(MemoryLocation::get(S), Loc, AAQI, S);
447     // If the store address cannot alias the pointer in question, then the
448     // specified memory cannot be modified by the store.
449     if (AR == AliasResult::NoAlias)
450       return ModRefInfo::NoModRef;
451 
452     // Examine the ModRef mask. If Mod isn't present, then return NoModRef.
453     // This ensures that if Loc is a constant memory location, we take into
454     // account the fact that the store definitely could not modify the memory
455     // location.
456     if (!isModSet(getModRefInfoMask(Loc)))
457       return ModRefInfo::NoModRef;
458   }
459 
460   // Otherwise, a store just writes.
461   return ModRefInfo::Mod;
462 }
463 
getModRefInfo(const FenceInst * S,const MemoryLocation & Loc,AAQueryInfo & AAQI)464 ModRefInfo AAResults::getModRefInfo(const FenceInst *S,
465                                     const MemoryLocation &Loc,
466                                     AAQueryInfo &AAQI) {
467   // All we know about a fence instruction is what we get from the ModRef
468   // mask: if Loc is a constant memory location, the fence definitely could
469   // not modify it.
470   if (Loc.Ptr)
471     return getModRefInfoMask(Loc);
472   return ModRefInfo::ModRef;
473 }
474 
getModRefInfo(const VAArgInst * V,const MemoryLocation & Loc,AAQueryInfo & AAQI)475 ModRefInfo AAResults::getModRefInfo(const VAArgInst *V,
476                                     const MemoryLocation &Loc,
477                                     AAQueryInfo &AAQI) {
478   if (Loc.Ptr) {
479     AliasResult AR = alias(MemoryLocation::get(V), Loc, AAQI, V);
480     // If the va_arg address cannot alias the pointer in question, then the
481     // specified memory cannot be accessed by the va_arg.
482     if (AR == AliasResult::NoAlias)
483       return ModRefInfo::NoModRef;
484 
485     // If the pointer is a pointer to invariant memory, then it could not have
486     // been modified by this va_arg.
487     return getModRefInfoMask(Loc, AAQI);
488   }
489 
490   // Otherwise, a va_arg reads and writes.
491   return ModRefInfo::ModRef;
492 }
493 
getModRefInfo(const CatchPadInst * CatchPad,const MemoryLocation & Loc,AAQueryInfo & AAQI)494 ModRefInfo AAResults::getModRefInfo(const CatchPadInst *CatchPad,
495                                     const MemoryLocation &Loc,
496                                     AAQueryInfo &AAQI) {
497   if (Loc.Ptr) {
498     // If the pointer is a pointer to invariant memory,
499     // then it could not have been modified by this catchpad.
500     return getModRefInfoMask(Loc, AAQI);
501   }
502 
503   // Otherwise, a catchpad reads and writes.
504   return ModRefInfo::ModRef;
505 }
506 
getModRefInfo(const CatchReturnInst * CatchRet,const MemoryLocation & Loc,AAQueryInfo & AAQI)507 ModRefInfo AAResults::getModRefInfo(const CatchReturnInst *CatchRet,
508                                     const MemoryLocation &Loc,
509                                     AAQueryInfo &AAQI) {
510   if (Loc.Ptr) {
511     // If the pointer is a pointer to invariant memory,
512     // then it could not have been modified by this catchpad.
513     return getModRefInfoMask(Loc, AAQI);
514   }
515 
516   // Otherwise, a catchret reads and writes.
517   return ModRefInfo::ModRef;
518 }
519 
getModRefInfo(const AtomicCmpXchgInst * CX,const MemoryLocation & Loc,AAQueryInfo & AAQI)520 ModRefInfo AAResults::getModRefInfo(const AtomicCmpXchgInst *CX,
521                                     const MemoryLocation &Loc,
522                                     AAQueryInfo &AAQI) {
523   // Acquire/Release cmpxchg has properties that matter for arbitrary addresses.
524   if (isStrongerThanMonotonic(CX->getSuccessOrdering()))
525     return ModRefInfo::ModRef;
526 
527   if (Loc.Ptr) {
528     AliasResult AR = alias(MemoryLocation::get(CX), Loc, AAQI, CX);
529     // If the cmpxchg address does not alias the location, it does not access
530     // it.
531     if (AR == AliasResult::NoAlias)
532       return ModRefInfo::NoModRef;
533   }
534 
535   return ModRefInfo::ModRef;
536 }
537 
getModRefInfo(const AtomicRMWInst * RMW,const MemoryLocation & Loc,AAQueryInfo & AAQI)538 ModRefInfo AAResults::getModRefInfo(const AtomicRMWInst *RMW,
539                                     const MemoryLocation &Loc,
540                                     AAQueryInfo &AAQI) {
541   // Acquire/Release atomicrmw has properties that matter for arbitrary addresses.
542   if (isStrongerThanMonotonic(RMW->getOrdering()))
543     return ModRefInfo::ModRef;
544 
545   if (Loc.Ptr) {
546     AliasResult AR = alias(MemoryLocation::get(RMW), Loc, AAQI, RMW);
547     // If the atomicrmw address does not alias the location, it does not access
548     // it.
549     if (AR == AliasResult::NoAlias)
550       return ModRefInfo::NoModRef;
551   }
552 
553   return ModRefInfo::ModRef;
554 }
555 
getModRefInfo(const Instruction * I,const std::optional<MemoryLocation> & OptLoc,AAQueryInfo & AAQIP)556 ModRefInfo AAResults::getModRefInfo(const Instruction *I,
557                                     const std::optional<MemoryLocation> &OptLoc,
558                                     AAQueryInfo &AAQIP) {
559   if (OptLoc == std::nullopt) {
560     if (const auto *Call = dyn_cast<CallBase>(I))
561       return getMemoryEffects(Call, AAQIP).getModRef();
562   }
563 
564   const MemoryLocation &Loc = OptLoc.value_or(MemoryLocation());
565 
566   switch (I->getOpcode()) {
567   case Instruction::VAArg:
568     return getModRefInfo((const VAArgInst *)I, Loc, AAQIP);
569   case Instruction::Load:
570     return getModRefInfo((const LoadInst *)I, Loc, AAQIP);
571   case Instruction::Store:
572     return getModRefInfo((const StoreInst *)I, Loc, AAQIP);
573   case Instruction::Fence:
574     return getModRefInfo((const FenceInst *)I, Loc, AAQIP);
575   case Instruction::AtomicCmpXchg:
576     return getModRefInfo((const AtomicCmpXchgInst *)I, Loc, AAQIP);
577   case Instruction::AtomicRMW:
578     return getModRefInfo((const AtomicRMWInst *)I, Loc, AAQIP);
579   case Instruction::Call:
580   case Instruction::CallBr:
581   case Instruction::Invoke:
582     return getModRefInfo((const CallBase *)I, Loc, AAQIP);
583   case Instruction::CatchPad:
584     return getModRefInfo((const CatchPadInst *)I, Loc, AAQIP);
585   case Instruction::CatchRet:
586     return getModRefInfo((const CatchReturnInst *)I, Loc, AAQIP);
587   default:
588     assert(!I->mayReadOrWriteMemory() &&
589            "Unhandled memory access instruction!");
590     return ModRefInfo::NoModRef;
591   }
592 }
593 
594 /// Return information about whether a particular call site modifies
595 /// or reads the specified memory location \p MemLoc before instruction \p I
596 /// in a BasicBlock.
597 /// FIXME: this is really just shoring-up a deficiency in alias analysis.
598 /// BasicAA isn't willing to spend linear time determining whether an alloca
599 /// was captured before or after this particular call, while we are. However,
600 /// with a smarter AA in place, this test is just wasting compile time.
callCapturesBefore(const Instruction * I,const MemoryLocation & MemLoc,DominatorTree * DT,AAQueryInfo & AAQI)601 ModRefInfo AAResults::callCapturesBefore(const Instruction *I,
602                                          const MemoryLocation &MemLoc,
603                                          DominatorTree *DT,
604                                          AAQueryInfo &AAQI) {
605   if (!DT)
606     return ModRefInfo::ModRef;
607 
608   const Value *Object = getUnderlyingObject(MemLoc.Ptr);
609   if (!isIdentifiedFunctionLocal(Object))
610     return ModRefInfo::ModRef;
611 
612   const auto *Call = dyn_cast<CallBase>(I);
613   if (!Call || Call == Object)
614     return ModRefInfo::ModRef;
615 
616   if (capturesAnything(PointerMayBeCapturedBefore(
617           Object, /* ReturnCaptures */ true, I, DT,
618           /* include Object */ true, CaptureComponents::Provenance)))
619     return ModRefInfo::ModRef;
620 
621   unsigned ArgNo = 0;
622   ModRefInfo R = ModRefInfo::NoModRef;
623   // Set flag only if no May found and all operands processed.
624   for (auto CI = Call->data_operands_begin(), CE = Call->data_operands_end();
625        CI != CE; ++CI, ++ArgNo) {
626     // Only look at the no-capture or byval pointer arguments.  If this
627     // pointer were passed to arguments that were neither of these, then it
628     // couldn't be no-capture.
629     if (!(*CI)->getType()->isPointerTy())
630       continue;
631 
632     // Make sure we still check captures(ret: address, provenance) and
633     // captures(address) arguments, as these wouldn't be treated as a capture
634     // at the call-site.
635     CaptureInfo Captures = Call->getCaptureInfo(ArgNo);
636     if (capturesAnyProvenance(Captures.getOtherComponents()))
637       continue;
638 
639     AliasResult AR =
640         alias(MemoryLocation::getBeforeOrAfter(*CI),
641               MemoryLocation::getBeforeOrAfter(Object), AAQI, Call);
642     // If this is a no-capture pointer argument, see if we can tell that it
643     // is impossible to alias the pointer we're checking.  If not, we have to
644     // assume that the call could touch the pointer, even though it doesn't
645     // escape.
646     if (AR == AliasResult::NoAlias)
647       continue;
648     if (Call->doesNotAccessMemory(ArgNo))
649       continue;
650     if (Call->onlyReadsMemory(ArgNo)) {
651       R = ModRefInfo::Ref;
652       continue;
653     }
654     return ModRefInfo::ModRef;
655   }
656   return R;
657 }
658 
659 /// canBasicBlockModify - Return true if it is possible for execution of the
660 /// specified basic block to modify the location Loc.
661 ///
canBasicBlockModify(const BasicBlock & BB,const MemoryLocation & Loc)662 bool AAResults::canBasicBlockModify(const BasicBlock &BB,
663                                     const MemoryLocation &Loc) {
664   return canInstructionRangeModRef(BB.front(), BB.back(), Loc, ModRefInfo::Mod);
665 }
666 
667 /// canInstructionRangeModRef - Return true if it is possible for the
668 /// execution of the specified instructions to mod\ref (according to the
669 /// mode) the location Loc. The instructions to consider are all
670 /// of the instructions in the range of [I1,I2] INCLUSIVE.
671 /// I1 and I2 must be in the same basic block.
canInstructionRangeModRef(const Instruction & I1,const Instruction & I2,const MemoryLocation & Loc,const ModRefInfo Mode)672 bool AAResults::canInstructionRangeModRef(const Instruction &I1,
673                                           const Instruction &I2,
674                                           const MemoryLocation &Loc,
675                                           const ModRefInfo Mode) {
676   assert(I1.getParent() == I2.getParent() &&
677          "Instructions not in same basic block!");
678   BasicBlock::const_iterator I = I1.getIterator();
679   BasicBlock::const_iterator E = I2.getIterator();
680   ++E;  // Convert from inclusive to exclusive range.
681 
682   for (; I != E; ++I) // Check every instruction in range
683     if (isModOrRefSet(getModRefInfo(&*I, Loc) & Mode))
684       return true;
685   return false;
686 }
687 
688 // Provide a definition for the root virtual destructor.
689 AAResults::Concept::~Concept() = default;
690 
691 // Provide a definition for the static object used to identify passes.
692 AnalysisKey AAManager::Key;
693 
ExternalAAWrapperPass()694 ExternalAAWrapperPass::ExternalAAWrapperPass() : ImmutablePass(ID) {}
695 
ExternalAAWrapperPass(CallbackT CB,bool RunEarly)696 ExternalAAWrapperPass::ExternalAAWrapperPass(CallbackT CB, bool RunEarly)
697     : ImmutablePass(ID), CB(std::move(CB)), RunEarly(RunEarly) {}
698 
699 char ExternalAAWrapperPass::ID = 0;
700 
701 INITIALIZE_PASS(ExternalAAWrapperPass, "external-aa", "External Alias Analysis",
702                 false, true)
703 
704 ImmutablePass *
createExternalAAWrapperPass(ExternalAAWrapperPass::CallbackT Callback)705 llvm::createExternalAAWrapperPass(ExternalAAWrapperPass::CallbackT Callback) {
706   return new ExternalAAWrapperPass(std::move(Callback));
707 }
708 
AAResultsWrapperPass()709 AAResultsWrapperPass::AAResultsWrapperPass() : FunctionPass(ID) {}
710 
711 char AAResultsWrapperPass::ID = 0;
712 
713 INITIALIZE_PASS_BEGIN(AAResultsWrapperPass, "aa",
714                       "Function Alias Analysis Results", false, true)
INITIALIZE_PASS_DEPENDENCY(BasicAAWrapperPass)715 INITIALIZE_PASS_DEPENDENCY(BasicAAWrapperPass)
716 INITIALIZE_PASS_DEPENDENCY(ExternalAAWrapperPass)
717 INITIALIZE_PASS_DEPENDENCY(GlobalsAAWrapperPass)
718 INITIALIZE_PASS_DEPENDENCY(SCEVAAWrapperPass)
719 INITIALIZE_PASS_DEPENDENCY(ScopedNoAliasAAWrapperPass)
720 INITIALIZE_PASS_DEPENDENCY(TypeBasedAAWrapperPass)
721 INITIALIZE_PASS_END(AAResultsWrapperPass, "aa",
722                     "Function Alias Analysis Results", false, true)
723 
724 /// Run the wrapper pass to rebuild an aggregation over known AA passes.
725 ///
726 /// This is the legacy pass manager's interface to the new-style AA results
727 /// aggregation object. Because this is somewhat shoe-horned into the legacy
728 /// pass manager, we hard code all the specific alias analyses available into
729 /// it. While the particular set enabled is configured via commandline flags,
730 /// adding a new alias analysis to LLVM will require adding support for it to
731 /// this list.
732 bool AAResultsWrapperPass::runOnFunction(Function &F) {
733   // NB! This *must* be reset before adding new AA results to the new
734   // AAResults object because in the legacy pass manager, each instance
735   // of these will refer to the *same* immutable analyses, registering and
736   // unregistering themselves with them. We need to carefully tear down the
737   // previous object first, in this case replacing it with an empty one, before
738   // registering new results.
739   AAR.reset(
740       new AAResults(getAnalysis<TargetLibraryInfoWrapperPass>().getTLI(F)));
741 
742   // Add any target-specific alias analyses that should be run early.
743   auto *ExtWrapperPass = getAnalysisIfAvailable<ExternalAAWrapperPass>();
744   if (ExtWrapperPass && ExtWrapperPass->RunEarly && ExtWrapperPass->CB) {
745     LLVM_DEBUG(dbgs() << "AAResults register Early ExternalAA: "
746                       << ExtWrapperPass->getPassName() << "\n");
747     ExtWrapperPass->CB(*this, F, *AAR);
748   }
749 
750   // BasicAA is always available for function analyses. Also, we add it first
751   // so that it can trump TBAA results when it proves MustAlias.
752   // FIXME: TBAA should have an explicit mode to support this and then we
753   // should reconsider the ordering here.
754   if (!DisableBasicAA) {
755     LLVM_DEBUG(dbgs() << "AAResults register BasicAA\n");
756     AAR->addAAResult(getAnalysis<BasicAAWrapperPass>().getResult());
757   }
758 
759   // Populate the results with the currently available AAs.
760   if (auto *WrapperPass =
761           getAnalysisIfAvailable<ScopedNoAliasAAWrapperPass>()) {
762     LLVM_DEBUG(dbgs() << "AAResults register ScopedNoAliasAA\n");
763     AAR->addAAResult(WrapperPass->getResult());
764   }
765   if (auto *WrapperPass = getAnalysisIfAvailable<TypeBasedAAWrapperPass>()) {
766     LLVM_DEBUG(dbgs() << "AAResults register TypeBasedAA\n");
767     AAR->addAAResult(WrapperPass->getResult());
768   }
769   if (auto *WrapperPass = getAnalysisIfAvailable<GlobalsAAWrapperPass>()) {
770     LLVM_DEBUG(dbgs() << "AAResults register GlobalsAA\n");
771     AAR->addAAResult(WrapperPass->getResult());
772   }
773   if (auto *WrapperPass = getAnalysisIfAvailable<SCEVAAWrapperPass>()) {
774     LLVM_DEBUG(dbgs() << "AAResults register SCEVAA\n");
775     AAR->addAAResult(WrapperPass->getResult());
776   }
777 
778   // If available, run an external AA providing callback over the results as
779   // well.
780   if (ExtWrapperPass && !ExtWrapperPass->RunEarly && ExtWrapperPass->CB) {
781     LLVM_DEBUG(dbgs() << "AAResults register Late ExternalAA: "
782                       << ExtWrapperPass->getPassName() << "\n");
783     ExtWrapperPass->CB(*this, F, *AAR);
784   }
785 
786   // Analyses don't mutate the IR, so return false.
787   return false;
788 }
789 
getAnalysisUsage(AnalysisUsage & AU) const790 void AAResultsWrapperPass::getAnalysisUsage(AnalysisUsage &AU) const {
791   AU.setPreservesAll();
792   AU.addRequiredTransitive<BasicAAWrapperPass>();
793   AU.addRequiredTransitive<TargetLibraryInfoWrapperPass>();
794 
795   // We also need to mark all the alias analysis passes we will potentially
796   // probe in runOnFunction as used here to ensure the legacy pass manager
797   // preserves them. This hard coding of lists of alias analyses is specific to
798   // the legacy pass manager.
799   AU.addUsedIfAvailable<ScopedNoAliasAAWrapperPass>();
800   AU.addUsedIfAvailable<TypeBasedAAWrapperPass>();
801   AU.addUsedIfAvailable<GlobalsAAWrapperPass>();
802   AU.addUsedIfAvailable<SCEVAAWrapperPass>();
803   AU.addUsedIfAvailable<ExternalAAWrapperPass>();
804 }
805 
run(Function & F,FunctionAnalysisManager & AM)806 AAManager::Result AAManager::run(Function &F, FunctionAnalysisManager &AM) {
807   Result R(AM.getResult<TargetLibraryAnalysis>(F));
808   for (auto &Getter : ResultGetters)
809     (*Getter)(F, AM, R);
810   return R;
811 }
812 
isNoAliasCall(const Value * V)813 bool llvm::isNoAliasCall(const Value *V) {
814   if (const auto *Call = dyn_cast<CallBase>(V))
815     return Call->hasRetAttr(Attribute::NoAlias);
816   return false;
817 }
818 
isNoAliasOrByValArgument(const Value * V)819 static bool isNoAliasOrByValArgument(const Value *V) {
820   if (const Argument *A = dyn_cast<Argument>(V))
821     return A->hasNoAliasAttr() || A->hasByValAttr();
822   return false;
823 }
824 
isIdentifiedObject(const Value * V)825 bool llvm::isIdentifiedObject(const Value *V) {
826   if (isa<AllocaInst>(V))
827     return true;
828   if (isa<GlobalValue>(V) && !isa<GlobalAlias>(V))
829     return true;
830   if (isNoAliasCall(V))
831     return true;
832   if (isNoAliasOrByValArgument(V))
833     return true;
834   return false;
835 }
836 
isIdentifiedFunctionLocal(const Value * V)837 bool llvm::isIdentifiedFunctionLocal(const Value *V) {
838   return isa<AllocaInst>(V) || isNoAliasCall(V) || isNoAliasOrByValArgument(V);
839 }
840 
isBaseOfObject(const Value * V)841 bool llvm::isBaseOfObject(const Value *V) {
842   // TODO: We can handle other cases here
843   // 1) For GC languages, arguments to functions are often required to be
844   //    base pointers.
845   // 2) Result of allocation routines are often base pointers.  Leverage TLI.
846   return (isa<AllocaInst>(V) || isa<GlobalVariable>(V));
847 }
848 
isEscapeSource(const Value * V)849 bool llvm::isEscapeSource(const Value *V) {
850   if (auto *CB = dyn_cast<CallBase>(V)) {
851     if (isIntrinsicReturningPointerAliasingArgumentWithoutCapturing(CB, true))
852       return false;
853 
854     // The return value of a function with a captures(ret: address, provenance)
855     // attribute is not necessarily an escape source. The return value may
856     // alias with a non-escaping object.
857     return !CB->hasArgumentWithAdditionalReturnCaptureComponents();
858   }
859 
860   // The load case works because isNotCapturedBefore considers all
861   // stores to be escapes (it passes true for the StoreCaptures argument
862   // to PointerMayBeCaptured).
863   if (isa<LoadInst>(V))
864     return true;
865 
866   // The inttoptr case works because isNotCapturedBefore considers all
867   // means of converting or equating a pointer to an int (ptrtoint, ptr store
868   // which could be followed by an integer load, ptr<->int compare) as
869   // escaping, and objects located at well-known addresses via platform-specific
870   // means cannot be considered non-escaping local objects.
871   if (isa<IntToPtrInst>(V))
872     return true;
873 
874   // Capture tracking considers insertions into aggregates and vectors as
875   // captures. As such, extractions from aggregates and vectors are escape
876   // sources.
877   if (isa<ExtractValueInst, ExtractElementInst>(V))
878     return true;
879 
880   // Same for inttoptr constant expressions.
881   if (auto *CE = dyn_cast<ConstantExpr>(V))
882     if (CE->getOpcode() == Instruction::IntToPtr)
883       return true;
884 
885   return false;
886 }
887 
isNotVisibleOnUnwind(const Value * Object,bool & RequiresNoCaptureBeforeUnwind)888 bool llvm::isNotVisibleOnUnwind(const Value *Object,
889                                 bool &RequiresNoCaptureBeforeUnwind) {
890   RequiresNoCaptureBeforeUnwind = false;
891 
892   // Alloca goes out of scope on unwind.
893   if (isa<AllocaInst>(Object))
894     return true;
895 
896   // Byval goes out of scope on unwind.
897   if (auto *A = dyn_cast<Argument>(Object))
898     return A->hasByValAttr() || A->hasAttribute(Attribute::DeadOnUnwind);
899 
900   // A noalias return is not accessible from any other code. If the pointer
901   // does not escape prior to the unwind, then the caller cannot access the
902   // memory either.
903   if (isNoAliasCall(Object)) {
904     RequiresNoCaptureBeforeUnwind = true;
905     return true;
906   }
907 
908   return false;
909 }
910 
911 // We don't consider globals as writable: While the physical memory is writable,
912 // we may not have provenance to perform the write.
isWritableObject(const Value * Object,bool & ExplicitlyDereferenceableOnly)913 bool llvm::isWritableObject(const Value *Object,
914                             bool &ExplicitlyDereferenceableOnly) {
915   ExplicitlyDereferenceableOnly = false;
916 
917   // TODO: Alloca might not be writable after its lifetime ends.
918   // See https://github.com/llvm/llvm-project/issues/51838.
919   if (isa<AllocaInst>(Object))
920     return true;
921 
922   if (auto *A = dyn_cast<Argument>(Object)) {
923     // Also require noalias, otherwise writability at function entry cannot be
924     // generalized to writability at other program points, even if the pointer
925     // does not escape.
926     if (A->hasAttribute(Attribute::Writable) && A->hasNoAliasAttr()) {
927       ExplicitlyDereferenceableOnly = true;
928       return true;
929     }
930 
931     return A->hasByValAttr();
932   }
933 
934   // TODO: Noalias shouldn't imply writability, this should check for an
935   // allocator function instead.
936   return isNoAliasCall(Object);
937 }
938